Official implementation for "Ham2Pose: Animating Sign Language Notation into Pose Sequences". The first method for animating HamNoSys into pose sequences.
Create a virtual environment
$ conda create --name Ham2Pose python=3.7
$ conda activate Ham2Pose
$ pip3 install requirements.txt
Prepare new data: To train the model using data that isn't part of our dataset, download the videos and use OpenPose to extract body, face, and hands keypoints from them (137 in total).
To train a new model using the default configurations run the following command:
$ python train.py
To change specific arguments, pass them as explained in the args.py
file, e.g. to change the number of
gpus to 4 and the batch size to 32 use:
$ python train.py --num_gpus 4 --batch_size 32
To pass a complete yaml configuration file use:
$ python train.py --config_file configs/config.yaml
To test an existing model with the default configuration use:
$ python test.py
The default configuration will test our supplied pretrained model "Ham2Pose". To train and test a different model, either change the model_name
in the configuration, or delete the existing checkpoint from the models
directory.
To change other arguments use one of the options mentioned under Train.
If you find this research useful, please cite the following:
@article{shalev2022ham2pose,
title={Ham2Pose: Animating Sign Language Notation into Pose Sequences},
author={Shalev-Arkushin, Rotem and Moryossef, Amit and Fried, Ohad},
journal={arXiv preprint arXiv:2211.13613},
year={2022}
}