Deep network for rolling shutter rectification
CMOS sensors employ row-wise acquisition mechanism while imaging a scene, which can result in undesired motion artifacts known as rolling shutter (RS) distortions in the captured image. Existing single image RS rectification methods attempt to account for these distortions by either using algorithms...
Saved in:
Main Authors | , , |
---|---|
Format | Journal Article |
Language | English |
Published |
12.12.2021
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | CMOS sensors employ row-wise acquisition mechanism while imaging a scene,
which can result in undesired motion artifacts known as rolling shutter (RS)
distortions in the captured image. Existing single image RS rectification
methods attempt to account for these distortions by either using algorithms
tailored for specific class of scenes which warrants information of intrinsic
camera parameters or a learning-based framework with known ground truth motion
parameters. In this paper, we propose an end-to-end deep neural network for the
challenging task of single image RS rectification. Our network consists of a
motion block, a trajectory module, a row block, an RS rectification module and
an RS regeneration module (which is used only during training). The motion
block predicts camera pose for every row of the input RS distorted image while
the trajectory module fits estimated motion parameters to a third-order
polynomial. The row block predicts the camera motion that must be associated
with every pixel in the target i.e, RS rectified image. Finally, the RS
rectification module uses motion trajectory and the output of row block to warp
the input RS image to arrive at a distortionfree image. For faster convergence
during training, we additionally use an RS regeneration module which compares
the input RS image with the ground truth image distorted by estimated motion
parameters. The end-to-end formulation in our model does not constrain the
estimated motion to ground-truth motion parameters, thereby successfully
rectifying the RS images with complex real-life camera motion. Experiments on
synthetic and real datasets reveal that our network outperforms prior art both
qualitatively and quantitatively. |
---|---|
DOI: | 10.48550/arxiv.2112.06170 |