Ademola, AdeyemiSinclair, DavidKoniaris, BabisHannah, SamanthaMitchell, KennyHunter, DavidSlingsby, Aidan2024-09-092024-09-092024978-3-03868-249-3https://doi.org/10.2312/cgvc.20241220https://diglib.eg.org/handle/10.2312/cgvc20241220Enabling online virtual reality (VR) users to dance and move in a way that mirrors the real-world necessitates improvements in the accuracy of predicting human motion sequences paving way for an immersive and connected experience. However, the drawbacks of latency in networked motion tracking present a critical detriment in creating a sense of complete engagement, requiring prediction for online synchronization of remote motions. To address this challenge, we propose a novel approach that leverages a synthetically generated dataset based on supervised foot anchor placement timings of rhythmic motions to ensure periodicity resulting in reduced prediction error. Specifically, our model compromises a discrete cosine transform (DCT) to encode motion, refine high frequencies and smooth motion sequences and prevent jittery motions. We introduce a feed-forward attention mechanism to learn based on dual-window pairs of 3D key points pose histories to predict future motions. Quantitative and qualitative experiments validating on the Human3.6m dataset result in observed improvements in the MPJPE evaluation metrics protocol compared with prior state-of-the-art.Attribution 4.0 International LicenseCCS Concepts: Computing methodologies → Machine Learning; Motion Processing; Virtual RealityComputing methodologies → Machine LearningMotion ProcessingVirtual RealityDeFT-Net: Dual-Window Extended Frequency Transformer for Rhythmic Motion Prediction10.2312/cgvc.202412207 pages