Rondón, Miguel Fabian RomeroSassatelli, LucilePardo, Ramon AparicioPrecioso, FrédéricChristie, Marc and Wu, Hui-Yin and Li, Tsai-Yen and Gandhi, Vineet2020-05-242020-05-242020978-3-03868-127-42411-9733https://doi.org/10.2312/wiced.20201125https://diglib.eg.org:443/handle/10.2312/wiced20201125Designing editing cuts for cinematic Virtual Reality (VR) has been under active investigation. Recently, the connection has been made between cuts in VR and adaptive streaming logics for 360º videos, with the introduction of rotational snap-cuts. Snapcuts can benefit the user's experience both by improving the streamed quality in the FoV and ensuring the user sees important elements for the plot. However, snap-cuts should not be too frequent and may be avoided when not beneficial to the streamed quality. We formulate the dynamic decision problem of snap-change triggering as a model-free Reinforcement Learning. We express the optimum cut triggering decisions computed offline with dynamic programming and investigate possible gains in quality of experience compared to baselines. We design Imitation Learning-based dynamic triggering strategies, and show that only knowing the past user's motion and video content, it is possible to outperform the controls without and with all cuts.Computing methodologiesVirtual realitySequential decision makingApprenticeship learningInformation systemsMultimedia streamingUser-Adaptive Rotational Snap-Cutting for Streamed 360º Videos10.2312/wiced.202011251-8