Paper: https://aclanthology.org/2021.naacl-main.79
Talk: https://www.youtube.com/watch?v=b0UogJP4U5I
MTAG (Modal-Temporal Attention Graph) is a GNN-based machine learning framework that can learn fusion and alignment for unaligned multimodal sequences.
Our code is written as an extension to the awesome PyTorch Geometric library. Users are encouraged to read their installation guide and documentations to understand the basics.
Our main contributions include:
Please refer to the requirement.txt
for setup.
Download the following datasets (please copy and paste the URL to browser, as clicking the link might not work):
and put them into a desired folder (.e.g. <dataroot>
). Then specify in run.sh
the folder containing the data of the desired dataset. For example:
python main.py \
...
--dataroot <dataroot>
...
bash run.sh
To visualize the edges:
jupyter notebook network_inference_visualize.ipynb
A more comprehensive hyperparameter list (along with each setting's performance we obtained) can be found in this Google Sheet. For any parameters that are not specified here, we used the default values in main.py
.
@inproceedings{yang-etal-2021-mtag,
title = "{MTAG}: Modal-Temporal Attention Graph for Unaligned Human Multimodal Language Sequences",
author = "Yang, Jianing and
Wang, Yongxin and
Yi, Ruitao and
Zhu, Yuying and
Rehman, Azaan and
Zadeh, Amir and
Poria, Soujanya and
Morency, Louis-Philippe",
booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
month = jun,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2021.naacl-main.79",
pages = "1009--1021",
abstract = "Human communication is multimodal in nature; it is through multiple modalities such as language, voice, and facial expressions, that opinions and emotions are expressed. Data in this domain exhibits complex multi-relational and temporal interactions. Learning from this data is a fundamentally challenging research problem. In this paper, we propose Modal-Temporal Attention Graph (MTAG). MTAG is an interpretable graph-based neural model that provides a suitable framework for analyzing multimodal sequential data. We first introduce a procedure to convert unaligned multimodal sequence data into a graph with heterogeneous nodes and edges that captures the rich interactions across modalities and through time. Then, a novel graph fusion operation, called MTAG fusion, along with a dynamic pruning and read-out technique, is designed to efficiently process this modal-temporal graph and capture various interactions. By learning to focus only on the important interactions within the graph, MTAG achieves state-of-the-art performance on multimodal sentiment analysis and emotion recognition benchmarks, while utilizing significantly fewer model parameters.",
}