murilovarges / HARBoP

Source code of experiments performed in paper: Human Action Recognition in Videos Based on Spatiotemporal Features and Bag-of-Poses
http://www.recogna.tech
4 stars 1 forks source link
human-action-recognition spatiotemporal-features

Human Action Recognition in Videos Based on Spatiotemporal Features and Bag-of-Poses

[Research Group Page] [Paper Link]

This repository presents a new method based spatiotemporal features and Bag-of-Poses to perform Human Action Recognition from videos.

Ilustration of architeture of proposed method: architeture_har

Abstract

Currently, there is a large number of methods that use 2D poses to represent and recognize human action in videos. Most of these methods use information computed from raw 2D poses based on the straight line segments that form the body parts in a 2D pose model in order to extract features (e.g., angles and trajectories). In our work, we propose a new method of representing 2D poses. Instead of directly using the straight line segments, firstly, the 2D pose is converted to the parameter space in which each segment is mapped to a point. Then, from the parameter space, spatiotemporal features are extracted and encoded using a Bag-of-Poses approach, then used for human action recognition in the video. Experiments on two well-known public datasets, Weizmann and KTH, showed that the proposed method using 2D poses encoded in parameter space can improve the recognition rates, obtaining competitive accuracy rates compared to state-of-the-art methods.

If you find this work helpful for your research, please cite our following paper:

M. Varges and A. N. Marana. Human Action Recognition in Videos Based on Spatiotemporal Features and Bag-of-Poses. Applied Soft Computing, 2020.

@article{VARGESDASILVA2020106513,
    title = "Human action recognition in videos based on spatiotemporal features and bag-of-poses",
    journal = "Applied Soft Computing",
    volume = "95",
    pages = "106513",
    year = "2020",
    issn = "1568-4946",
    doi = "https://doi.org/10.1016/j.asoc.2020.106513",
    url = "http://www.sciencedirect.com/science/article/pii/S156849462030452X",
    author = "Murilo {Varges da Silva} and Aparecido {Nilceu Marana}",
    keywords = "Human action recognition, Surveillance systems, Spatiotemporal features, Video sequences, Bag-of-poses"
}

If you have any question or feedback about the code, please contact: murilo.varges@gmail.com.

Requirements

To run this project you will need to install the follows softwares:

This project requires the following dependencies:

matplotlib==2.2.2
numpy==1.15.0
pandas==0.22.0
seaborn==0.9.0
scikit_learn==0.21.3

Install the requirements with command:

pip install -r requirements.txt

Tutorials

We provide some basic tutorials for you to get familar with the code and tools.

  1. 2D Poses Extraction
  2. Encoding 2D Poses in Parameter Space
  3. Features Extraction
  4. Human Action Recognition

License

This project is Apache 2.0 licensed, as found in the LICENSE file.

Acknowledgements

We thank NVIDIA Corporation for the donation of the GPU used in this study. This study was financed in part by CAPES - Brazil (Finance Code 001).