image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000004545.jpg
|
4545_evjvqa
|
[{"from": "human", "value": "what color shirt is the person running next to the bike with the flag wearing?\n<image>"}, {"from": "gpt", "value": "wear a blue shirt"}]
|
00000004545.jpg
|
4545_evjvqa
|
[{"from": "human", "value": "where are the children going?\n<image>"}, {"from": "gpt", "value": "the children are marching"}]
|
00000004547.jpg
|
4547_evjvqa
|
[{"from": "human", "value": "what is the woman sitting in front of the store doing?\n<image>"}, {"from": "gpt", "value": "choosing goods"}]
|
00000004547.jpg
|
4547_evjvqa
|
[{"from": "human", "value": "how many people are there behind the woman sitting?\n<image>"}, {"from": "gpt", "value": "there is a person behind"}]
|
00000004549.jpg
|
4549_evjvqa
|
[{"from": "human", "value": "how many cars are about to go through the gate?\n<image>"}, {"from": "gpt", "value": "there are two cars about to pass through the gate"}]
|
00000004549.jpg
|
4549_evjvqa
|
[{"from": "human", "value": "what are the two sides of the road decorated with?\n<image>"}, {"from": "gpt", "value": "there are flags on both sides of the road"}]
|
00000004551.jpg
|
4551_evjvqa
|
[{"from": "human", "value": "what color are the flags everyone holding?\n<image>"}, {"from": "gpt", "value": "the flags are red"}]
|
00000004551.jpg
|
4551_evjvqa
|
[{"from": "human", "value": "what is on display on the yard?\n<image>"}, {"from": "gpt", "value": "the paintings are being displayed on the yard"}]
|
00000004551.jpg
|
4551_evjvqa
|
[{"from": "human", "value": "which side of the paintings row is everyone standing on?\n<image>"}, {"from": "gpt", "value": "everyone is standing to the left of the row of paintings"}]
|
00000004552.jpg
|
4552_evjvqa
|
[{"from": "human", "value": "where is the car parked?\n<image>"}, {"from": "gpt", "value": "the car is parked in front of a supermarket"}]
|
00000004553.jpg
|
4553_evjvqa
|
[{"from": "human", "value": "how big is the alley?\n<image>"}, {"from": "gpt", "value": "the alley is very narrow in size"}]
|
00000004553.jpg
|
4553_evjvqa
|
[{"from": "human", "value": "how many vehicles are placed in front of the alley?\n<image>"}, {"from": "gpt", "value": "there is a motorbike in front of the alley"}]
|
00000004553.jpg
|
4553_evjvqa
|
[{"from": "human", "value": "how many people are standing next to the vehicle in front of the alley?\n<image>"}, {"from": "gpt", "value": "there is a person standing next to the vehicle"}]
|
00000004555.jpg
|
4555_evjvqa
|
[{"from": "human", "value": "how many people are standing in front of the gate?\n<image>"}, {"from": "gpt", "value": "there is a person standing in front of the gate"}]
|
00000004555.jpg
|
4555_evjvqa
|
[{"from": "human", "value": "in which direction is the person leading the bicycle?\n<image>"}, {"from": "gpt", "value": "going out of the street"}]
|
00000004556.jpg
|
4556_evjvqa
|
[{"from": "human", "value": "how is the weather here?\n<image>"}, {"from": "gpt", "value": "it is raining"}]
|
00000004556.jpg
|
4556_evjvqa
|
[{"from": "human", "value": "where are people going?\n<image>"}, {"from": "gpt", "value": "heading towards the castle"}]
|
00000004556.jpg
|
4556_evjvqa
|
[{"from": "human", "value": "how are visitors avoiding getting wet?\n<image>"}, {"from": "gpt", "value": "by using umbrellas"}]
|
00000004557.jpg
|
4557_evjvqa
|
[{"from": "human", "value": "how many main dishes are on the table?\n<image>"}, {"from": "gpt", "value": "there are three main dishes on the table"}]
|
00000004557.jpg
|
4557_evjvqa
|
[{"from": "human", "value": "where is the sauce cup placed?\n<image>"}, {"from": "gpt", "value": "placed below"}]
|
00000004557.jpg
|
4557_evjvqa
|
[{"from": "human", "value": "where is the soup placed?\n<image>"}, {"from": "gpt", "value": "the soup is placed on the left"}]
|
00000004558.jpg
|
4558_evjvqa
|
[{"from": "human", "value": "where are two people standing?\n<image>"}, {"from": "gpt", "value": "standing in front of a church"}]
|
00000004558.jpg
|
4558_evjvqa
|
[{"from": "human", "value": "which side of the girl is the boy standing on?\n<image>"}, {"from": "gpt", "value": "standing in front of the girl"}]
|
00000004558.jpg
|
4558_evjvqa
|
[{"from": "human", "value": "what are two people doing?\n<image>"}, {"from": "gpt", "value": "are about to kiss each other"}]
|
00000004559.jpg
|
4559_evjvqa
|
[{"from": "human", "value": "how many people are sitting on the stairs?\n<image>"}, {"from": "gpt", "value": "there is a person sitting on the stairs"}]
|
00000004559.jpg
|
4559_evjvqa
|
[{"from": "human", "value": "what area are people standing in?\n<image>"}, {"from": "gpt", "value": "people are standing in front of a temple"}]
|
00000004559.jpg
|
4559_evjvqa
|
[{"from": "human", "value": "on which side of the girl sitting is the boy standing?\n<image>"}, {"from": "gpt", "value": "stand to the left of the girl"}]
|
00000004561.jpg
|
4561_evjvqa
|
[{"from": "human", "value": "what color is the basket?\n<image>"}, {"from": "gpt", "value": "the basket is blue"}]
|
00000004561.jpg
|
4561_evjvqa
|
[{"from": "human", "value": "how many handles does the basket have?\n<image>"}, {"from": "gpt", "value": "the basket has two handles"}]
|
00000004562.jpg
|
4562_evjvqa
|
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a market"}]
|
00000004562.jpg
|
4562_evjvqa
|
[{"from": "human", "value": "which stall is the man in the hat standing at?\n<image>"}, {"from": "gpt", "value": "standing at the fruit stall"}]
|
00000004562.jpg
|
4562_evjvqa
|
[{"from": "human", "value": "what kind of fruit is on the left-hand side of the man?\n<image>"}, {"from": "gpt", "value": "on the left are grapefruits"}]
|
00000004566.jpg
|
4566_evjvqa
|
[{"from": "human", "value": "what is everyone doing?\n<image>"}, {"from": "gpt", "value": "getting ready to paddle canoes to sea"}]
|
00000004566.jpg
|
4566_evjvqa
|
[{"from": "human", "value": "how many red canoes are there?\n<image>"}, {"from": "gpt", "value": "there is one red canoe"}]
|
00000004567.jpg
|
4567_evjvqa
|
[{"from": "human", "value": "how many motorbikes are running on the road?\n<image>"}, {"from": "gpt", "value": "there is a motorbike running on the road"}]
|
00000004569.jpg
|
4569_evjvqa
|
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a seaport"}]
|
00000004569.jpg
|
4569_evjvqa
|
[{"from": "human", "value": "what is the weather like at the seaport?\n<image>"}, {"from": "gpt", "value": "it is early in the morning"}]
|
00000004569.jpg
|
4569_evjvqa
|
[{"from": "human", "value": "what are the cranes doing?\n<image>"}, {"from": "gpt", "value": "moving the boxes on the ships"}]
|
00000004570.jpg
|
4570_evjvqa
|
[{"from": "human", "value": "where are the paintings put?\n<image>"}, {"from": "gpt", "value": "put in a museum"}]
|
00000004570.jpg
|
4570_evjvqa
|
[{"from": "human", "value": "what artefacts are exhibited in the showroom?\n<image>"}, {"from": "gpt", "value": "there are famous paintings exhibited in the room"}]
|
00000004570.jpg
|
4570_evjvqa
|
[{"from": "human", "value": "how many drawings are exhibited next to the entrance?\n<image>"}, {"from": "gpt", "value": "there are two drawings hanging on either side of the entrance"}]
|
00000004572.jpg
|
4572_evjvqa
|
[{"from": "human", "value": "what color top is the tour guide wearing?\n<image>"}, {"from": "gpt", "value": "wearing a blue top"}]
|
00000004572.jpg
|
4572_evjvqa
|
[{"from": "human", "value": "what are the students doing?\n<image>"}, {"from": "gpt", "value": "raising hands up high following the guide"}]
|
00000004574.jpg
|
4574_evjvqa
|
[{"from": "human", "value": "how many circular windows does the house have?\n<image>"}, {"from": "gpt", "value": "the house has two circular windows"}]
|
00000004574.jpg
|
4574_evjvqa
|
[{"from": "human", "value": "how many sets of tables and chairs are there in the kitchen of the house?\n<image>"}, {"from": "gpt", "value": "there are three sets of tables and chairs in the kitchen"}]
|
00000004574.jpg
|
4574_evjvqa
|
[{"from": "human", "value": "what is the main color of the house?\n<image>"}, {"from": "gpt", "value": "the main color of the house is yellow"}]
|
00000004578.jpg
|
4578_evjvqa
|
[{"from": "human", "value": "what products does the store sell?\n<image>"}, {"from": "gpt", "value": "the store sells all kinds of sports equipment"}]
|
00000004578.jpg
|
4578_evjvqa
|
[{"from": "human", "value": "how many motorbikes are parked next to the shop?\n<image>"}, {"from": "gpt", "value": "there is a motorbike parked next to the shop"}]
|
00000004578.jpg
|
4578_evjvqa
|
[{"from": "human", "value": "what kind of products are displayed on the left side of the store?\n<image>"}, {"from": "gpt", "value": "the left side of the store displays all kinds of treadmills"}]
|
00000004579.jpg
|
4579_evjvqa
|
[{"from": "human", "value": "what products does the store sell?\n<image>"}, {"from": "gpt", "value": "the store sells all kinds of clothes and fabrics"}]
|
00000004579.jpg
|
4579_evjvqa
|
[{"from": "human", "value": "on which side are the clothes sold?\n<image>"}, {"from": "gpt", "value": "the clothes are sold on the right side"}]
|
00000004579.jpg
|
4579_evjvqa
|
[{"from": "human", "value": "where is the fabric located in the store?\n<image>"}, {"from": "gpt", "value": "the fabric is placed on the left side of the store"}]
|
00000004589.jpg
|
4589_evjvqa
|
[{"from": "human", "value": "what color shirt is the man buying?\n<image>"}, {"from": "gpt", "value": "wearing a gray shirt"}]
|
00000004589.jpg
|
4589_evjvqa
|
[{"from": "human", "value": "where is the male adult standing?\n<image>"}, {"from": "gpt", "value": "standing in a supermarket"}]
|
00000004590.jpg
|
4590_evjvqa
|
[{"from": "human", "value": "where is the group of people standing?\n<image>"}, {"from": "gpt", "value": "standing in a museum"}]
|
00000004590.jpg
|
4590_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "giving a presentation to the students"}]
|
00000004590.jpg
|
4590_evjvqa
|
[{"from": "human", "value": "what is the little girl using her phone to take pictures of?\n<image>"}, {"from": "gpt", "value": "take a picture of an ancient statue"}]
|
00000004591.jpg
|
4591_evjvqa
|
[{"from": "human", "value": "wwhat is the bag made of?\n<image>"}, {"from": "gpt", "value": "made of leather"}]
|
00000004591.jpg
|
4591_evjvqa
|
[{"from": "human", "value": "what color is the bag?\n<image>"}, {"from": "gpt", "value": "brown"}]
|
00000004594.jpg
|
4594_evjvqa
|
[{"from": "human", "value": "how many people are standing on the chairs?\n<image>"}, {"from": "gpt", "value": "there are two people standing on the chairs"}]
|
00000004594.jpg
|
4594_evjvqa
|
[{"from": "human", "value": "what is the woman standing on the stool doing?\n<image>"}, {"from": "gpt", "value": "pulling out the roof of the shop"}]
|
00000001800.jpg
|
1800_evjvqa
|
[{"from": "human", "value": "where are people posing for pictures?\n<image>"}, {"from": "gpt", "value": "next to the painted wall"}]
|
00000001800.jpg
|
1800_evjvqa
|
[{"from": "human", "value": "how many motorbikes are parked on the sidewalk opposite the wall?\n<image>"}, {"from": "gpt", "value": "five"}]
|
00000001800.jpg
|
1800_evjvqa
|
[{"from": "human", "value": "how many motorbikes are there with people sitting on them?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000001800.jpg
|
1800_evjvqa
|
[{"from": "human", "value": "how many bicycles are there in the painting on the wall?\n<image>"}, {"from": "gpt", "value": "one"}]
|
00000001801.jpg
|
1801_evjvqa
|
[{"from": "human", "value": "what color is the car parked in the distance?\n<image>"}, {"from": "gpt", "value": "black"}]
|
00000001801.jpg
|
1801_evjvqa
|
[{"from": "human", "value": "what is the road surface covered with?\n<image>"}, {"from": "gpt", "value": "leaves"}]
|
00000001801.jpg
|
1801_evjvqa
|
[{"from": "human", "value": "how many lampposts are there on the road?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001805.jpg
|
1805_evjvqa
|
[{"from": "human", "value": "which direction is the cyclo going?\n<image>"}, {"from": "gpt", "value": "to the city gate"}]
|
00000001805.jpg
|
1805_evjvqa
|
[{"from": "human", "value": "who are selling around the sidewalk on the left?\n<image>"}, {"from": "gpt", "value": "there are three women selling"}]
|
00000001805.jpg
|
1805_evjvqa
|
[{"from": "human", "value": "which sidewalk is the woman with the pedlar selling goods sitting on?\n<image>"}, {"from": "gpt", "value": "right sidewalk"}]
|
00000001805.jpg
|
1805_evjvqa
|
[{"from": "human", "value": "how many cyclos are going to the city gate?\n<image>"}, {"from": "gpt", "value": "there is a cyclo going to the city gate"}]
|
00000001807.jpg
|
1807_evjvqa
|
[{"from": "human", "value": "how many people are standing in the gallery?\n<image>"}, {"from": "gpt", "value": "two people"}]
|
00000001807.jpg
|
1807_evjvqa
|
[{"from": "human", "value": "what objects does the museum display?\n<image>"}, {"from": "gpt", "value": "bones and fossils"}]
|
00000001807.jpg
|
1807_evjvqa
|
[{"from": "human", "value": "what objects does this room display?\n<image>"}, {"from": "gpt", "value": "bones and fossils"}]
|
00000001807.jpg
|
1807_evjvqa
|
[{"from": "human", "value": "how many people are watching in the gallery?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001808.jpg
|
1808_evjvqa
|
[{"from": "human", "value": "how many models of boat are there?\n<image>"}, {"from": "gpt", "value": "two models"}]
|
00000001808.jpg
|
1808_evjvqa
|
[{"from": "human", "value": "how many model boats carry flower pots?\n<image>"}, {"from": "gpt", "value": "one"}]
|
00000001809.jpg
|
1809_evjvqa
|
[{"from": "human", "value": "what stall is the man in the blue shirt standing at?\n<image>"}, {"from": "gpt", "value": "watermelon"}]
|
00000001809.jpg
|
1809_evjvqa
|
[{"from": "human", "value": "where is the man in the blue shirt standing?\n<image>"}, {"from": "gpt", "value": "watermelon stall"}]
|
00000001809.jpg
|
1809_evjvqa
|
[{"from": "human", "value": "what is the man in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "taking pictures"}]
|
00000001809.jpg
|
1809_evjvqa
|
[{"from": "human", "value": "who is the man in the blue shirt standing near?\n<image>"}, {"from": "gpt", "value": "a woman wearing a hat and a watch"}]
|
00000001814.jpg
|
1814_evjvqa
|
[{"from": "human", "value": "where are the two girls standing?\n<image>"}, {"from": "gpt", "value": "orchard"}]
|
00000001814.jpg
|
1814_evjvqa
|
[{"from": "human", "value": "what is the girl in the blue shirt holding in her hand?\n<image>"}, {"from": "gpt", "value": "conical hat"}]
|
00000001814.jpg
|
1814_evjvqa
|
[{"from": "human", "value": "where does the girl in the red shirt put the conical hat?\n<image>"}, {"from": "gpt", "value": "on her head"}]
|
00000001814.jpg
|
1814_evjvqa
|
[{"from": "human", "value": "how many girls stand here in the orchard?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001814.jpg
|
1814_evjvqa
|
[{"from": "human", "value": "how are the two girls in the orchard dressed?\n<image>"}, {"from": "gpt", "value": "blouse"}]
|
00000001815.jpg
|
1815_evjvqa
|
[{"from": "human", "value": "what kind of products do people set up in front of the store?\n<image>"}, {"from": "gpt", "value": "fruit"}]
|
00000001815.jpg
|
1815_evjvqa
|
[{"from": "human", "value": "how many people are coming to buy fruit?\n<image>"}, {"from": "gpt", "value": "no one is coming to buy fruit"}]
|
00000001815.jpg
|
1815_evjvqa
|
[{"from": "human", "value": "how many people are standing near the fruit stall?\n<image>"}, {"from": "gpt", "value": "one person"}]
|
00000001819.jpg
|
1819_evjvqa
|
[{"from": "human", "value": "what is the counter right behind the counter selling food made from dragon fruit?\n<image>"}, {"from": "gpt", "value": "dragon fruit counter"}]
|
00000001820.jpg
|
1820_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "repairing the roof"}]
|
00000001820.jpg
|
1820_evjvqa
|
[{"from": "human", "value": "where is the man sitting?\n<image>"}, {"from": "gpt", "value": "on the roof"}]
|
00000001820.jpg
|
1820_evjvqa
|
[{"from": "human", "value": "what is the man sitting on the roof doing?\n<image>"}, {"from": "gpt", "value": "repairing the roof"}]
|
00000001821.jpg
|
1821_evjvqa
|
[{"from": "human", "value": "how many boats are on the water?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001821.jpg
|
1821_evjvqa
|
[{"from": "human", "value": "how many people are on the boat nearby?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001821.jpg
|
1821_evjvqa
|
[{"from": "human", "value": "how many people are sitting on the boat in the distance?\n<image>"}, {"from": "gpt", "value": "one"}]
|
00000001823.jpg
|
1823_evjvqa
|
[{"from": "human", "value": "where is the boy standing?\n<image>"}, {"from": "gpt", "value": "on a mountain"}]
|
00000001823.jpg
|
1823_evjvqa
|
[{"from": "human", "value": "where is the boy visiting?\n<image>"}, {"from": "gpt", "value": "on top of a mountain"}]
|
00000001826.jpg
|
1826_evjvqa
|
[{"from": "human", "value": "where is the girl in the black shirt near the woman standing?\n<image>"}, {"from": "gpt", "value": "vegetable counter"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.