Event-Enhanced Blurry Video Super-Resolution

Corresponding Author
1University of Science and Technology of China 2Institute of Artificial Intelligence, Hefei Comprehensive National Science Center
3National University of Singapore
AAAI 2025

Video Demos

Abstract

In this paper, we tackle the task of blurry video super-resolution (BVSR), aiming to generate high-resolution (HR) videos from low-resolution (LR) and blurry inputs. Current BVSR methods often fail to restore sharp details at high resolutions, resulting in noticeable artifacts and jitter due to insufficient motion information for deconvolution and the lack of high-frequency details in LR frames. To address these challenges, we introduce event signals into BVSR and propose a novel event-enhanced network, Ev-DeblurVSR. To effectively fuse information from frames and events for feature deblurring, we introduce a reciprocal feature deblurring module that leverages motion information from intra-frame events to deblur frame features while reciprocally using global scene context from the frames to enhance event features. Furthermore, to enhance temporal consistency, we propose a hybrid deformable alignment module that fully exploits the complementary motion information from inter-frame events and optical flow to improve motion estimation in the deformable alignment process. Extensive evaluations demonstrate that Ev-DeblurVSR establishes a new state-of-the-art performance on both synthetic and real-world datasets. Notably, on real data, our method is +2.59dB more accurate and 7.28× faster than the recent best BVSR baseline FMA-Net.

Network Architecture

comparison

Overview of Ev-DeblurVSR. Intra-frame voxels, which record the exposure time information of blurry frames, are fused with blurry frames in the RFD module to deblur frame features and enhance event features with scene context. Inter-frame voxels, which record motion information between frames, are integrated into the HDA module, using continuous motion trajectories to guide deformable alignment. Finally, the aligned features are upsampled to reconstruct sharp HR frames.

Quantitative Results

comparison

Quantitative comparison (PSNR↑/SSIM↑/LPIPS↓) on the GoPro dataset for 4× blurry VSR. More comparisons on BSD and NCER datasets are available in our paper.

Qualitative Results

BibTeX

@inproceedings{kai2025event,
  title={Event-{E}nhanced {B}lurry {V}ideo {S}uper-{R}esolution},
  author={Kai, Dachun and Zhang, Yueyi and Wang, Jin and Xiao, Zeyu and Xiong, Zhiwei and Sun, Xiaoyan},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={39},
  number={4},
  pages={4175--4183},
  year={2025}
}