Xu, Sen-ZheHu, JunWang, MiaoMu, Tai-JiangHu, Shi-MinFu, Hongbo and Ghosh, Abhijeet and Kopf, Johannes2018-10-072018-10-0720181467-8659https://doi.org/10.1111/cgf.13566https://diglib.eg.org:443/handle/10.1111/cgf13566Video stabilization is necessary for many hand-held shot videos. In the past decades, although various video stabilization methods were proposed based on the smoothing of 2D, 2.5D or 3D camera paths, hardly have there been any deep learning methods to solve this problem. Instead of explicitly estimating and smoothing the camera path, we present a novel online deep learning framework to learn the stabilization transformation for each unsteady frame, given historical steady frames. Our network is composed of a generative network with spatial transformer networks embedded in different layers, and generates a stable frame for the incoming unstable frame by computing an appropriate affine transformation. We also introduce an adversarial network to determine the stability of a piece of video. The network is trained directly using the pair of steady and unsteady videos. Experiments show that our method can produce similar results as traditional methods, moreover, it is capable of handling challenging unsteady video of low quality, where traditional methods fail, such as video with heavy noise or multiple exposures. Our method runs in real time, which is much faster than traditional methods.Computing methodologiesComputer GraphicsDeep Video Stabilization Using Adversarial Networks10.1111/cgf.13566267-276