Skip to content

ais-lab/gaussiansp-paindiffusion

Repository files navigation

Driving GaussianAvatars to Express Pain

It's fast, high-quality rendering, with teeth...

Screencast.from.01-22-2025.04.53.07.PM.webm

Based on:

  • PainDiffusion for pain expression
  • GaussianAvatars
    • VHAP for face model tracking
    • GaussianAvatars for animation

Install

To keep the installation simple, each component will be installed in a different environment.

Clone this repository, use the recursive flag to pull all submodules at once. git clone --recursive https://github.com/ais-lab/gaussiansp-paindiffusion

For each component, follow its own README to install its environment.

Copy the following files to their corresponding folders: controll_gui_with_gaussian_avatars.py -> paindiffusion, localviewer.py -> gaussianavatars. These two files communicate through reading and writing to a shared-memory file.

Use

  1. Change the flame model of VHAP and GaussianAvatars at (gaussianAvatars/flame_model/flame.py) and (vhap/model/flame.py) to flame2020, which should be generic_model.pkl from https://flame.is.tue.mpg.de/login.php. Because PainDiffusion uses flame2020.

  2. Use VHAP to track a video or multiview video using the monocular steps and merge the output with the final step in nersemble (check vhap/doc).

  3. After building the point cloud and flame sequence in step 2, train a Gaussian splatting of the point cloud. Please follow the steps of GaussianAvatars.

  4. In the GaussianAvatars dir and using its env, use local_viewer.py in GaussianAvatars to view the results with the path to the output of the previous step.

python local_viewer.py --point_path '/path/to/point_cloud.ply' --driving_mode
  1. Run PainDiffusion by python controll_gui_with_gaussian_avatars.py --conf_file 'path/to/conf_of_paindiffusion.yaml' to drive the avatar according to pain stimuli.

About

Driving GaussianAvatars to Express Pain

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors