• Medientyp: E-Artikel
  • Titel: Deep end-to-end rolling shutter rectification
  • Beteiligte: Kandula, Praveen; Kumar, T. Lokesh; Rajagopalan, A. N.
  • Erschienen: Optica Publishing Group, 2020
  • Erschienen in: Journal of the Optical Society of America A
  • Sprache: Englisch
  • DOI: 10.1364/josaa.388818
  • ISSN: 1084-7529; 1520-8532
  • Schlagwörter: Computer Vision and Pattern Recognition ; Atomic and Molecular Physics, and Optics ; Electronic, Optical and Magnetic Materials
  • Entstehung:
  • Anmerkungen:
  • Beschreibung: <jats:p>CMOS sensors employ a row-wise acquisition mechanism while imaging a scene, which can result in undesired motion artifacts known as rolling shutter (RS) distortions in the captured image. Existing single image RS rectification methods attempt to account for these distortions by using either algorithms tailored for a specific class of scenes that warrants information of intrinsic camera parameters or a learning-based framework with known ground truth motion parameters. In this paper, we propose an end-to-end deep neural network for the challenging task of single image RS rectification. Our network consists of a motion block, a trajectory module, a row block, an RS rectification module, and an RS regeneration module (which is used only during training). The motion block predicts the camera pose for every row of the input RS distorted image, while the trajectory module fits estimated motion parameters to a third-order polynomial. The row block predicts the camera motion that must be associated with every pixel in the target, i.e., RS rectified image. Finally, the RS rectification module uses motion trajectory and the output of a row block to warp the input RS image to arrive at a distortion-free image. For faster convergence during training, we additionally use an RS regeneration module that compares the input RS image with the ground truth image distorted by estimated motion parameters. The end-to-end formulation in our model does not constrain the estimated motion to ground truth motion parameters, thereby successfully rectifying the RS images with complex real-life camera motion. Experiments on synthetic and real datasets reveal that our network outperforms prior art both qualitatively and quantitatively.</jats:p>