Milef, NicholasSueda, ShinjiroKalantari, Nima KhademiMyszkowski, KarolNiessner, Matthias2023-05-032023-05-0320231467-8659https://doi.org/10.1111/cgf.14767https://diglib.eg.org:443/handle/10.1111/cgf14767We propose a learning-based approach for full-body pose reconstruction from extremely sparse upper body tracking data, obtained from a virtual reality (VR) device. We leverage a conditional variational autoencoder with gated recurrent units to synthesize plausible and temporally coherent motions from 4-point tracking (head, hands, and waist positions and orientations). To avoid synthesizing implausible poses, we propose a novel sample selection and interpolation strategy along with an anomaly detection algorithm. Specifically, we monitor the quality of our generated poses using the anomaly detection algorithm and smoothly transition to better samples when the quality falls below a statistically defined threshold. Moreover, we demonstrate that our sample selection and interpolation method can be used for other applications, such as target hitting and collision avoidance, where the generated motions should adhere to the constraints of the virtual environment. Our system is lightweight, operates in real-time, and is able to produce temporally coherent and realistic motions.Attribution 4.0 International LicenseCCS Concepts: Computing methodologies -> Neural networks; Motion processing; Virtual realityComputing methodologiesNeural networksMotion processingVirtual realityVariational Pose Prediction with Dynamic Sample Selection from Sparse Tracking Signals10.1111/cgf.14767359-36911 pages