Optoelectronics Letters, Volume. 21, Issue 8, 499(2025)

Robust human motion prediction via integration of spatial and temporal cues

Shaobo ZHANG, Sheng LIU, Fei GAO, and Yuan FENG

Research on human motion prediction has made significant progress due to its importance in the development of various artificial intelligence applications. However, effectively capturing spatio-temporal features for smoother and more precise human motion prediction remains a challenge. To address these issues, a robust human motion prediction method via integration of spatial and temporal cues (RISTC) has been proposed. This method captures sufficient spatio-temporal correlation of the observable sequence of human poses by utilizing the spatio-temporal mixed feature extractor (MFE). In multi-layer MFEs, the channel-graph united attention blocks extract the augmented spatial features of the human poses in the channel and spatial dimension. Additionally, multi-scale temporal blocks have been designed to effectively capture complicated and highly dynamic temporal information. Our experiments on the Human3.6M and Carnegie Mellon University motion capture (CMU Mocap) datasets show that the proposed network yields higher prediction accuracy than the state-of-the-art methods.

Tools

Get Citation

Copy Citation Text

ZHANG Shaobo, LIU Sheng, GAO Fei, FENG Yuan. Robust human motion prediction via integration of spatial and temporal cues[J]. Optoelectronics Letters, 2025, 21(8): 499

Download Citation

EndNote(RIS)BibTexPlain Text
Save article for my favorites
Paper Information

Received: May. 14, 2024

Accepted: Jul. 24, 2025

Published Online: Jul. 24, 2025

The Author Email:

DOI:10.1007/s11801-025-4119-4

Topics