Datasets:
Tasks:
Image-Text-to-Text
Modalities:
Image
Formats:
imagefolder
Languages:
English
Size:
1K - 10K
ArXiv:
License:
Add link to paper and code repository, update task category
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,26 +1,26 @@
|
|
1 |
---
|
2 |
-
license: mit
|
3 |
-
task_categories:
|
4 |
-
- visual-question-answering
|
5 |
language:
|
6 |
- en
|
7 |
-
|
8 |
size_categories:
|
9 |
- 10K<n<100K
|
|
|
|
|
|
|
10 |
---
|
11 |
|
12 |
# PopVQA: Popular Entity Visual Question Answering
|
13 |
|
14 |
PopVQA is a dataset designed to study the performance gap in vision-language models (VLMs) when answering factual questions about entities presented in **images** versus **text**.
|
15 |
|
|
|
|
|
16 |
|
17 |

|
18 |
|
19 |
-
|
20 |
## 🔍 Motivation
|
21 |
<img src="./paper_teaser.png" alt="Motivation" width="700">
|
22 |
|
23 |
-
|
24 |
PopVQA was curated to explore the disparity in model performance when answering factual questions about an entity described in text versus depicted in an image. This is achieved by asking the same questions twice, once with the textual representation (the entity's name), then, with the visual representation (entity image). We include several questions about every entity to allow a more fine grained evaluation.
|
25 |
This dataset was introduced in the paper:
|
26 |
|
@@ -85,3 +85,4 @@ To build the dataset, run:
|
|
85 |
|
86 |
```bash
|
87 |
python scripts/build_dataset.py --base-df path/to/base_entities.csv
|
|
|
|
1 |
---
|
|
|
|
|
|
|
2 |
language:
|
3 |
- en
|
4 |
+
license: mit
|
5 |
size_categories:
|
6 |
- 10K<n<100K
|
7 |
+
task_categories:
|
8 |
+
- image-text-to-text
|
9 |
+
pretty_name: PopVQA
|
10 |
---
|
11 |
|
12 |
# PopVQA: Popular Entity Visual Question Answering
|
13 |
|
14 |
PopVQA is a dataset designed to study the performance gap in vision-language models (VLMs) when answering factual questions about entities presented in **images** versus **text**.
|
15 |
|
16 |
+
Paper: https://huggingface.co/papers/2412.14133
|
17 |
+
Code: https://github.com/idocohen/vlm-modality-gap
|
18 |
|
19 |

|
20 |
|
|
|
21 |
## 🔍 Motivation
|
22 |
<img src="./paper_teaser.png" alt="Motivation" width="700">
|
23 |
|
|
|
24 |
PopVQA was curated to explore the disparity in model performance when answering factual questions about an entity described in text versus depicted in an image. This is achieved by asking the same questions twice, once with the textual representation (the entity's name), then, with the visual representation (entity image). We include several questions about every entity to allow a more fine grained evaluation.
|
25 |
This dataset was introduced in the paper:
|
26 |
|
|
|
85 |
|
86 |
```bash
|
87 |
python scripts/build_dataset.py --base-df path/to/base_entities.csv
|
88 |
+
```
|