:dancer: VALSE: A Task-Independent Benchmark for Vision and Language Models Centered on Linguistic Phenomena. https://aclanthology.org/2022.acl-long.567/
📰 News: Accepted at ACL 2022 Main Conference.
Please find the data in the data
folder. The dataset is in json
format and contains the following relevant fields:
dataset
and image_file
.caption
.foil
.mturk
.
caption
counts the number of annotators who chose the caption, but/and not the foil, to be the one describing the image.foil
counts how many of the three annotators chose the foil to be (also) describing the image.:bangbang: Please be aware that the jsons are containing both valid (meaning: validated by annotators) and non-validated samples. In order to work only with the valid set, please consider filtering them:
We consider a valid foil to mean: at least two out of three annotators identified the caption, but not the foil, as the text which accurately describes the image.
This means that the valid samples of the dataset are the ones where sample["mturk"]["caption"] >= 2
.
Example instance:
{
"actions_test_0": {
"dataset": "SWiG", # dataset from where the image and caption originate from
"original_split": "test", # the split of the original dataset in which the sample belonged to
"dataset_idx": "exercising_255.jpg", # the sample id in the original dataset
"linguistic_phenomena": "actions", # the linguistic phenomenon targeted
"image_file": "exercising_255.jpg", # the image filename (in the original dataset)
"caption": "A man exercises his torso.", # image caption
"classes": "man", # the word of the caption that was replaced
"classes_foil": "torso", # the foil word / phrase
"mturk": { # Amazon Mechanical Turk annotation (validation) results
"foil": 0, # how many annotators voted that the foil describes the image
"caption": 3, # how many annotators voted that the caption only (and not the foil) to describe the image
"other": 0
},
"foil": "A torso exercises for a man." # foil where one word / phrase is exchanged in the original caption such that the foil caption does not describe the image anymore
}, ...
}
For the images, please follow the downloading instructions of the respective original dataset. The provenance of the original images is mentioned in the json files in the field dataset
.
Please cite our :dancer: VALSE paper if you are using this dataset.
@inproceedings{parcalabescu-etal-2022-valse,
title = "{VALSE}: A Task-Independent Benchmark for Vision and Language Models Centered on Linguistic Phenomena",
author = "Parcalabescu, Letitia and
Cafagna, Michele and
Muradjan, Lilitta and
Frank, Anette and
Calixto, Iacer and
Gatt, Albert",
booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.acl-long.567",
pages = "8253--8280",
abstract = "We propose VALSE (Vision And Language Structured Evaluation), a novel benchmark designed for testing general-purpose pretrained vision and language (V{\&}L) models for their visio-linguistic grounding capabilities on specific linguistic phenomena. VALSE offers a suite of six tests covering various linguistic constructs. Solving these requires models to ground linguistic phenomena in the visual modality, allowing more fine-grained evaluations than hitherto possible. We build VALSE using methods that support the construction of valid foils, and report results from evaluating five widely-used V{\&}L models. Our experiments suggest that current models have considerable difficulty addressing most phenomena. Hence, we expect VALSE to serve as an important benchmark to measure future progress of pretrained V{\&}L models from a linguistic perspective, complementing the canonical task-centred V{\&}L evaluations.",
}