EasternJournalist / live_mocap

Single-view real-time motion capture built up upon Google Mediapipe.
Apache License 2.0
197 stars 37 forks source link

Live MoCap

Ganyu-ji

Requirements

How to use

  1. Prepare your character model

    Currently this script uses Blender to load model skeleton and bind animation. Your model should be saved as .blend file.

    You may edit your model to assure that

    • Model must be in rest pose (clear all bone rotation/translation/scale in pose mode). And the rest pose should be close to T pose.

    • Clear previous bone animation data and constraints.

    • Name related bones as below (in lower case). You may refer the mixamo example assets/mixamo.blend to see to name the bones so that they can be recogonzed and binded.

    • Save the model as .blend file somewhere.

  2. Run the script mocap.py.

    python mocap.py --blend your_character_model.blend --video your_video.mp4 [other options] 

    The program will read and capture motion from the video, save the animation data, and then open Blender and bind the animation to your charactor model. After everything is done, you should be able to see the Blender window with your character already animated.

Future work