Research Output
NeFT-Net: N-window Extended Frequency Transformer for Rhythmic Motion Prediction
  Advancements in prediction of human motion sequences are critical for enabling online virtual reality (VR) users to dance and move in ways that accurately mirror real-world actions, delivering a more immersive and connected experience. However, latency in networked motion tracking remains a significant challenge, disrupting engagement and necessitating predictive solutions to achieve real-time synchronization of remote motions. To address this issue, we propose a novel approach leveraging a synthetically generated dataset based on supervised foot
anchor placement timings for rhythmic motions, ensuring periodicity and reducing prediction errors. Our model integrates a discrete cosine transform (DCT) to encode motion, refine highfrequency components, and smooth motion sequences, mitigating jittery artifacts. Additionally,
we introduce a feed-forward attention mechanism designed to learn from N-window pairs of 3D key-point pose histories for precise future motion prediction. Quantitative and qualitative evaluations on the Human3.6M dataset highlight significant improvements in mean per joint position error (MPJPE) metrics, demonstrating the superiority of our technique over state-ofthe-art approaches. We further introduce novel result pose visualizations through the use of generative AI methods.

  • Date:

    24 March 2025

  • Publication Status:

    Accepted

  • ISSN:

    0097-8493

  • Funders:

    European Commission

Citation

Âé¶¹ÉçÇø

Ademola, A., Sinclair, D., Koniaris, B., Hannah, S., & Mitchell, K. (in press). NeFT-Net: N-window Extended Frequency Transformer for Rhythmic Motion Prediction. Computers and Graphics,

Authors

Monthly Views:

Linked Projects

Available Documents