Skip to content

DifFace: Blind Face Restoration with Diffused Error Contraction (TPAMI, 2024)

License

Notifications You must be signed in to change notification settings

zsyOAOA/DifFace

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

36 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

DifFace: Blind Face Restoration with Diffused Error Contraction (TPAMI, 2024)

Zongsheng Yue, Chen Change Loy

Paper

google colab logo Hugging Face visitors

โญ If DifFace is helpful to your images or projects, please help star this repo. Thanks! ๐Ÿค—

Update

  • 2023.12.11: Add the code for image inpainting.
  • 2022.12.19: Add Colab demo google colab logo.
  • 2022.12.17: Add the Hugging Face.
  • 2022.12.13: Create this repo.

Applications

๐Ÿ‘‰ Old Photo Enhancement

๐Ÿ‘‰ Face Restoration

๐Ÿ‘‰ Face Inpainting

Requirements

A suitable conda environment named DifFace can be created and activated with:

conda env create -f environment.yaml
conda activate DifFace

Inference

๐Ÿ‘ฆ Face image restoration (cropped and aligned)

python inference_difface.py -i [image folder/image path] -o [result folder] --task restoration --eta 0.5 --aligned --use_fp16

Note that the hyper-parameter eta controls the fidelity-realness trade-off, you can freely adjust it between 0.0 and 1.0.

๐Ÿ‘ฎ Whole image enhancement

python inference_difface.py -i [image folder/image path] -o [result folder] --task restoration --eta 0.5 --use_fp16

๐Ÿ‘ธ Face image inpainting

python inference_difface.py -i [image folder/image path] -o [result folder] --task inpainting --use_fp16

We assume that the masked area is filled with zeros in the low quality image. Based on such an assumption, the image mask is automatically deteced in our code.

Testing

To reproduce the results in our paper, please follow the following guidelines to prepare the testing data.

  1. Download the FFHQ dataset, and resize them into size 512x512(or 256x256).
python scripts/big2small_face.py -i [Face folder(1024x1024)] -o [Saving folder(512x512)] --pch_size 512 
  1. Make the testing dataset for restoration
python scripts/prepare_testing_restoration.py -i [CelebA folder(512x512)] -o [Saving folder]  
  1. Make the testing dataset for inpainting
python scripts/prepare_testing_inpainting.py -i [CelebA folder(256x256)] -o [Saving folder]  

Training

๐Ÿข Configuration

  1. Modify the data path in data.train and data.val according to your own settings.
  2. Adjust the batch size based on your GPU devices.
    • train.batchsize: [A, B] # A denotes the batch size for training, B denotes the batch size for validation
    • train.microbatch: C # C denotes the batch size on each GPU, A = C * num_gpus * num_grad_accumulation

๐Ÿฌ Train diffusion model with 8 GPUS

torchrun --standalone --nproc_per_node=8 --nnodes=1 main.py --cfg_path configs/training/diffsuion_ffhq512.yaml --save_dir [Logging Folder]  

๐Ÿณ Train diffused estimator for restoration (SwinIR) with 4 GPUS

torchrun --standalone --nproc_per_node=4 --nnodes=1 main.py --cfg_path configs/training/swinir_ffhq512.yaml --save_dir [Logging Folder]  

๐Ÿ‘ Train diffused estimator for restoration (LaMa) with 4 GPUS

torchrun --standalone --nproc_per_node=4 --nnodes=1 main.py --cfg_path configs/training/estimator_lama_inpainting.yaml --save_dir [Logging Folder]  

License

This project is licensed under NTU S-Lab License 1.0. Redistribution and use should follow this license.

Acknowledgement

This project is based on Improved Diffusion Model. Some codes are brought from BasicSR, YOLOv5-face, and FaceXLib. We also adopt Real-ESRGAN to support background image enhancement. Thanks for their awesome works.

Contact

If you have any questions, please feel free to contact me via [email protected].