Skip to content

GeneFace v1.1.0, pretrained models and binarized datasets

Latest
Compare
Choose a tag to compare
@yerfor yerfor released this 16 Mar 06:42
· 39 commits to main since this release

Welcome to GeneFace v1.1.0

We have made GeneFace more practical for industrial usage!

What's new in this release:

  1. We implement RAD-NeRF renderer, which could infer in real-time and be trained in 10 hours.
  2. We turn to pytorch-based deep3d_recon module to extract 3DMM, which is easier to install and is 8x faster than the previous TF-based version.
  3. We provide a pitch-aware audio2motion module, which could generate more lip-sync landmark.
  4. Fix some bugs that cause large memory usage.
  5. We will upload the paper about this release soon.

We release the pre-trained models of GeneFace:

  • lrs3.zip includes the models trained on LRS3-TED dataset (a lm3d_vae_sync to perform the audio2motion transform and a syncnet for measuring the lip-sync), which are generic for all possible target person videos.
  • May.zip includes the models trained on the May.mp4 target person video (a lm3d_postnet_sync for refining the predicted 3d landmark, a lm3d_radnerf for rendering the head image, and a lm3d_radnerf_torso for rendering the torso part). For each target person video, you need to train these three models.
  • How to use the pretrained models: unzip the lrs3.zip and May.zip into the checkpoints directory, then follow the commandlines for inference in README.md

🔥 We also release the binarized datasets:

  • LRS3-TED: We provide the processed lrs3 dataset on Google Drive. Download link: Part1, Part2.
  • Baiduyun Disk for LRS3-TED: link, password lrs3