Face video restoration (FVR) is a challenging but important problem where one seeks to recover a perceptually realistic face videos from a low-quality input. While diffusion probabilistic models (DPMs) have been shown to achieve remarkable performance for face image restoration, they often fail to preserve temporally coherent, high-quality videos, compromising the fidelity of reconstructed faces. We present a new conditional diffusion framework called FLAIR for FVR. FLAIR ensures temporal consistency across frames in a computationally efficient fashion by converting a traditional image DPM into a video DPM. The proposed conversion uses a recurrent video refinement layer and a temporal self-attention at different scales. FLAIR also uses a conditional iterative refinement process to balance the perceptual and distortion quality during inference. This process consists of two key components: a data-consistency module that analytically ensures that the generated video precisely matches its degraded observation and a coarse-to-fine image enhancement module specifically for facial regions. Our extensive experiments show superiority of FLAIR over the current state-of-the-art (SOTA) for video super-resolution, deblurring, JPEG restoration, and space-time frame interpolation on two high-quality face video datasets.
JPEG
X4 Gaussian Blur
Spatial-temporal SR
X4 Motion Blur
X8 Bicubic
Web Video Restoration
Comparing to CodeFormer
Comparing to RestoreFormer++
Comparing to DDNM
Comparing to VRT
@InProceedings{Zou_2025_WACV,
author = {Zou, Zihao and Liu, Jiaming and Shoushtari, Shirin and Wang, Yubo and Kamilov, Ulugbek S.},
title = {FLAIR: A Conditional Diffusion Framework with Applications to Face Video Restoration},
booktitle = {Proceedings of the Winter Conference on Applications of Computer Vision (WACV)},
month = {February},
year = {2025},
pages = {5228-5238}
}