bzhangGo / sltunet

SLTUNET: A Simple Unified Model for Sign Language Translation (ICLR 2023)
27 stars 6 forks source link
dgs3-t gloss2text sign-language-processing sign-language-translation sign2gloss sign2text sltunet

SLTUNET: A Simple Unified Model for Sign Language Translation (ICLR 2023)

Paper | Highlights | Overview | DGS3-T | Training&Eval | Model Performance | Citation

Paper Highlights

Among thousands of languages globally, some are written, some are spoken, while some are signed. Sign languages are unique natural languages widely used in Deaf communities. They express meaning through hand gestures, body movements and facial expressions, and are often in a video form. We refer the readers to Sign language Processing for a better understanding of sign languages.

In this study, we aim at improving sign language translation, i.e. translating information from sign languages (in a video) to spoken languages (in text). We address the video-text modality gap and the training data scarcity issue via multi-task learning and unified modeling.

Briefly,

Model Visualization

Overview of ur proposal

DGS3-T

DGS3-T Licensing

DGS3-T is based on the Public DGS Corpus. The license of the Public DGS Corpus does not allow any computational research except if express permission is given by the University of Hamburg.

Constructing DGS3-T

Please check out dgs3-t for details.

Requirement

The source code is based on older tensorflow.

Training and Evaluation

Training includes two phrase: 1) pretrain sign embeddings; 2) train SLTUNet model.

Please check out example for details.

Performance

Resulst on Phoenix

Check out our paper for more results on CSLDaily and DGS3-T.

Citation

If you draw any inspiration from our study, please consider to cite our paper:

@inproceedings{
zhang2023sltunet,
title={{SLTUNET}: A Simple Unified Model for Sign Language Translation},
author={Biao Zhang and Mathias M{\"u}ller and Rico Sennrich},
booktitle={The Eleventh International Conference on Learning Representations },
year={2023},
url={https://openreview.net/forum?id=EBS4C77p_5S}
}