diff --git a/README.md b/README.md index 19a275d..4fc9a24 100644 --- a/README.md +++ b/README.md @@ -45,6 +45,12 @@ This repository is the official implementation of the [CVPR 2022](https://cvpr20 Lightning Project Page + + Youtube Video + + + Paper PDF +

EMOCA takes a single in-the-wild image as input and reconstructs a 3D face with sufficient facial expression detail to convey the emotional state of the input image. EMOCA advances the state-of-the-art monocular face reconstruction in-the-wild, putting emphasis on accurate capture of emotional content. The official project page is [here](https://emoca.is.tue.mpg.de/index.html). diff --git a/gdl_apps/EMOCA/demos/test_emoca_on_video.py b/gdl_apps/EMOCA/demos/test_emoca_on_video.py index b49d473..420c357 100644 --- a/gdl_apps/EMOCA/demos/test_emoca_on_video.py +++ b/gdl_apps/EMOCA/demos/test_emoca_on_video.py @@ -31,7 +31,7 @@ def main(): parser.add_argument('--input_video', type=str, default=str(Path(gdl.__file__).parents[1] / "data/EMOCA_test_example_data/videos/82-25-854x480_affwild2.mp4"), help="Filename of the video for reconstruction.") parser.add_argument('--output_folder', type=str, default="video_output", help="Output folder to save the results to.") - parser.add_argument('--model_name', type=str, default='EMOCA', help='Name of the model to use.') + parser.add_argument('--model_name', type=str, default='EMOCA', help='Name of the model to use. Currently EMOCA or DECA are available.') parser.add_argument('--path_to_models', type=str, default=str(Path(gdl.__file__).parents[1] / "assets/EMOCA/models")) parser.add_argument('--save_images', type=bool, default=True, help="If true, output images will be saved") parser.add_argument('--save_codes', type=bool, default=False, help="If true, output FLAME values for shape, expression, jaw pose will be saved")