Kim, BeomseokSon, HyeongseokPark, Seong-JinCho, SunghyunLee, SeungyongFu, Hongbo and Ghosh, Abhijeet and Kopf, Johannes2018-10-072018-10-0720181467-8659https://doi.org/10.1111/cgf.13567https://diglib.eg.org:443/handle/10.1111/cgf13567We propose a novel approach for detecting two kinds of partial blur, defocus and motion blur, by training a deep convolutional neural network. Existing blur detection methods concentrate on designing low-level features, but those features have difficulty in detecting blur in homogeneous regions without enough textures or edges. To handle such regions, we propose a deep encoder-decoder network with long residual skip-connections and multi-scale reconstruction loss functions to exploit high-level contextual features as well as low-level structural features. Another difficulty in partial blur detection is that there are no available datasets with images having both defocus and motion blur together, as most existing approaches concentrate only on either defocus or motion blur. To resolve this issue, we construct a synthetic dataset that consists of complex scenes with both types of blur. Experimental results show that our approach effectively detects and classifies blur, outperforming other state-of-the-art methods. Our method can be used for various applications, such as photo editing, blur magnification, and deblurring.Computing methodologiesImage processingDefocus and Motion Blur Detection with Deep Contextual Features10.1111/cgf.13567277-288