VDN: Variant-depth network for motion deblurring

Cai Guo, Qian Wang, Hong Ning Dai*, Ping Li*

*Corresponding author for this work

Research output: Contribution to journalJournal articlepeer-review

3 Citations (Scopus)


Motion deblurring is a challenging task in vision and graphics. Recent researches aim to deblur by using multiple sub-networks with multi-scale or multi-patch inputs. However, scaling or splitting operations on input images inevitably loses the spatial details of the images. Meanwhile, their models are usually complex and computationally expensive. To address these problems, we propose a novel variant-depth scheme. In particular, we utilize the multiple variant-depth sub-networks with scale-invariant inputs to combine into a variant-depth network (VDN). In our design, different levels of sub-networks accomplish progressive deblurring effects without transforming the inputs, thereby effectively reducing the computational complexity of the model. Extensive experiments have shown that our VDN outperforms the state-of-the-art motion deblurring methods while maintaining a lower computational cost. The source code is publicly available at: https://github.com/CaiGuoHS/VDN.

Original languageEnglish
Article numbere2066
JournalComputer Animation and Virtual Worlds
Issue number3-4
Publication statusPublished - 1 Jun 2022

Scopus Subject Areas

  • Software
  • Computer Graphics and Computer-Aided Design

User-Defined Keywords

  • motion deblurring
  • scale-invariant input
  • variant-depth network


Dive into the research topics of 'VDN: Variant-depth network for motion deblurring'. Together they form a unique fingerprint.

Cite this