Abstract
Motion deblurring is a challenging task in vision and graphics. Recent researches aim to deblur by using multiple sub-networks with multi-scale or multi-patch inputs. However, scaling or splitting operations on input images inevitably loses the spatial details of the images. Meanwhile, their models are usually complex and computationally expensive. To address these problems, we propose a novel variant-depth scheme. In particular, we utilize the multiple variant-depth sub-networks with scale-invariant inputs to combine into a variant-depth network (VDN). In our design, different levels of sub-networks accomplish progressive deblurring effects without transforming the inputs, thereby effectively reducing the computational complexity of the model. Extensive experiments have shown that our VDN outperforms the state-of-the-art motion deblurring methods while maintaining a lower computational cost. The source code is publicly available at: https://github.com/CaiGuoHS/VDN.
Original language | English |
---|---|
Article number | e2066 |
Journal | Computer Animation and Virtual Worlds |
Volume | 33 |
Issue number | 3-4 |
DOIs | |
Publication status | Published - 1 Jun 2022 |
Scopus Subject Areas
- Software
- Computer Graphics and Computer-Aided Design
User-Defined Keywords
- motion deblurring
- scale-invariant input
- variant-depth network