ShaoXiang23 / CLIP_Hand_Demo

A simple Zero-Shot demo for CLIP-Hand 3D.
MIT License
7 stars 1 forks source link

CLIP_Hand_Demo

Update (11/04/2023): Release the evaluation code on FreiHAND datasets.

:loudspeaker: Update (10/09/2023): Our paper "CLIP-Hand3D: Exploiting 3D Hand Pose Estimation via Context-Aware Prompting" has been accepted at ACM MM 2023! Stay tuned for more updates. :tada:

CLIP-Hand3D: Exploiting 3D Hand Pose Estimation via Context-Aware Prompting

Shaoxiang Guo, Qing Cai*, Lin Qi and Junyu Dong* (*Corresponding Authors)

School of Computer Science and Technology, Ocean University of China, 238 Songling Road, Qingdao, China.

Introduction

In our paper, we introduce CLIP-Hand3D, a novel method for 3D hand pose estimation from monocular images using Contrastive Language-Image Pre-training (CLIP). We bridge the gap between text prompts and the irregular distribution of hand joint positions in 3D space by encoding pose labels into text representations and hand joint spatial distribution into pose-aware features. We maximize the semantic consistency between pose-text features using a CLIP-based contrastive learning paradigm. Our method, which includes a coarse-to-fine mesh regressor, achieves comparable SOTA performance and significantly faster inference speed on several public hand benchmarks. In this github repository, we will release the corresponding codes. First, we release a simple zero-shot demo to show the semantic relations between hand images and pose text prompts.

Install

Download Datasets

FreiHAND Dataset

Run Evaluation on FreiHAND Test Dataset

Run Zero Shot Demo

If you are interested in our work or if it is helpful to you, please consider citing our paper:

@inproceedings{guo_clip_hand,
  title={CLIP-Hand3D: Exploiting 3D Hand Pose Estimation via Context-Aware Prompting},
  author={Guo, Shaoxiang and Cai, Qing and Qi, Lin and Dong, Junyu},
  booktitle={Proceedings of the 31st ACM International Conference on Multimedia},
  year={2023},
  organization={ACM}
}

We also referenced the following codebase, which inspired us with their outstanding work. We would like to thank the authors of the following code bases.