image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000001225.jpg
|
1225_evjvqa
|
[{"from": "human", "value": "how many people take a seat on the staircase?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000001228.jpg
|
1228_evjvqa
|
[{"from": "human", "value": "how many people sit in front of the house?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000001228.jpg
|
1228_evjvqa
|
[{"from": "human", "value": "where is the old woman sitting?\n<image>"}, {"from": "gpt", "value": "in front of the house"}]
|
00000001229.jpg
|
1229_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "offering of incense"}]
|
00000001234.jpg
|
1234_evjvqa
|
[{"from": "human", "value": "where is the woman wearing a hat and bag staying?\n<image>"}, {"from": "gpt", "value": "on the brigde"}]
|
00000001235.jpg
|
1235_evjvqa
|
[{"from": "human", "value": "what is the stall selling?\n<image>"}, {"from": "gpt", "value": "food"}]
|
00000001235.jpg
|
1235_evjvqa
|
[{"from": "human", "value": "what is the booth used for?\n<image>"}, {"from": "gpt", "value": "selling food"}]
|
00000001237.jpg
|
1237_evjvqa
|
[{"from": "human", "value": "how many motorbikes are parked in front of the house?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000001237.jpg
|
1237_evjvqa
|
[{"from": "human", "value": "where is the motorbike parked in?\n<image>"}, {"from": "gpt", "value": "in front of the yellow house"}]
|
00000001239.jpg
|
1239_evjvqa
|
[{"from": "human", "value": "what type of hat is the woman riding a bicycle wearing?\n<image>"}, {"from": "gpt", "value": "conical hat"}]
|
00000001239.jpg
|
1239_evjvqa
|
[{"from": "human", "value": "how many people ride the bicycle?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000001239.jpg
|
1239_evjvqa
|
[{"from": "human", "value": "how many people are riding cyclo?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000001239.jpg
|
1239_evjvqa
|
[{"from": "human", "value": "what color is the hawker trolley going on the street?\n<image>"}, {"from": "gpt", "value": "blue"}]
|
00000001241.jpg
|
1241_evjvqa
|
[{"from": "human", "value": "where is the girl standing?\n<image>"}, {"from": "gpt", "value": "in the middle of the street"}]
|
00000001241.jpg
|
1241_evjvqa
|
[{"from": "human", "value": "where are the lanterns being hung?\n<image>"}, {"from": "gpt", "value": "above the girl in orange shirt standing in the middle of the street"}]
|
00000001241.jpg
|
1241_evjvqa
|
[{"from": "human", "value": "what is being hung above the girl's head?\n<image>"}, {"from": "gpt", "value": "the lanterns"}]
|
00000001242.jpg
|
1242_evjvqa
|
[{"from": "human", "value": "what is constructed in front of the building to decorate?\n<image>"}, {"from": "gpt", "value": "the fountain"}]
|
00000001247.jpg
|
1247_evjvqa
|
[{"from": "human", "value": "how many people are cycling through the alley?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000001247.jpg
|
1247_evjvqa
|
[{"from": "human", "value": "what is the woman wearing conical hat doing?\n<image>"}, {"from": "gpt", "value": "she is cycling"}]
|
00000001247.jpg
|
1247_evjvqa
|
[{"from": "human", "value": "what color shirt is woman riding the bicycle wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
|
00000001255.jpg
|
1255_evjvqa
|
[{"from": "human", "value": "what is the family visiting?\n<image>"}, {"from": "gpt", "value": "the museum"}]
|
00000001255.jpg
|
1255_evjvqa
|
[{"from": "human", "value": "where is the family?\n<image>"}, {"from": "gpt", "value": "in the museum"}]
|
00000001257.jpg
|
1257_evjvqa
|
[{"from": "human", "value": "how many people are selling candies?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001257.jpg
|
1257_evjvqa
|
[{"from": "human", "value": "what color t-shirt does the stall owner selling candies wear?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000001261.jpg
|
1261_evjvqa
|
[{"from": "human", "value": "who is standing with two men?\n<image>"}, {"from": "gpt", "value": "the woman is in blue"}]
|
00000001261.jpg
|
1261_evjvqa
|
[{"from": "human", "value": "who is standing with the woman?\n<image>"}, {"from": "gpt", "value": "two men are wearing green shirt"}]
|
00000001261.jpg
|
1261_evjvqa
|
[{"from": "human", "value": "how many people are standing in the house?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000001261.jpg
|
1261_evjvqa
|
[{"from": "human", "value": "who is talking with two men?\n<image>"}, {"from": "gpt", "value": "the woman is in blue "}]
|
00000001262.jpg
|
1262_evjvqa
|
[{"from": "human", "value": "what is the man in green doing?\n<image>"}, {"from": "gpt", "value": "he is cleanig up"}]
|
00000001263.jpg
|
1263_evjvqa
|
[{"from": "human", "value": "how many serving of ice cream on the table?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001268.jpg
|
1268_evjvqa
|
[{"from": "human", "value": "where is the scene taken?\n<image>"}, {"from": "gpt", "value": "at the citadel"}]
|
00000001271.jpg
|
1271_evjvqa
|
[{"from": "human", "value": "what are the stall on both sides of the street selling mainly?\n<image>"}, {"from": "gpt", "value": "vegetable"}]
|
00000001271.jpg
|
1271_evjvqa
|
[{"from": "human", "value": "how many motorcyclists are buying vegetable?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000001273.jpg
|
1273_evjvqa
|
[{"from": "human", "value": "where is this?\n<image>"}, {"from": "gpt", "value": "this is the historical monument"}]
|
00000001276.jpg
|
1276_evjvqa
|
[{"from": "human", "value": "what is the woman in green doing?\n<image>"}, {"from": "gpt", "value": "she is giving the cookie box to the customer"}]
|
00000001276.jpg
|
1276_evjvqa
|
[{"from": "human", "value": "what is the woman in the pink striped shirt doing?\n<image>"}, {"from": "gpt", "value": "buying biscuit"}]
|
00000001276.jpg
|
1276_evjvqa
|
[{"from": "human", "value": "how many people are standing at cookie stall?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001276.jpg
|
1276_evjvqa
|
[{"from": "human", "value": "which stall are two woman standing at?\n<image>"}, {"from": "gpt", "value": "cookie stall"}]
|
00000001276.jpg
|
1276_evjvqa
|
[{"from": "human", "value": "which product is the woman in pink shirt intending to buy?\n<image>"}, {"from": "gpt", "value": "biscuit"}]
|
00000001278.jpg
|
1278_evjvqa
|
[{"from": "human", "value": "how many people are carried by the posterior boat?\n<image>"}, {"from": "gpt", "value": "four"}]
|
00000001278.jpg
|
1278_evjvqa
|
[{"from": "human", "value": "what color shirt is the boatman of posterior boat?\n<image>"}, {"from": "gpt", "value": "she is wearing south vietnamese pajama and conical hat"}]
|
00000001285.jpg
|
1285_evjvqa
|
[{"from": "human", "value": "where are they?\n<image>"}, {"from": "gpt", "value": "at the historical monument"}]
|
00000001285.jpg
|
1285_evjvqa
|
[{"from": "human", "value": "where are the travelers visiting?\n<image>"}, {"from": "gpt", "value": "the historical monument"}]
|
00000001286.jpg
|
1286_evjvqa
|
[{"from": "human", "value": "where is the scene taken?\n<image>"}, {"from": "gpt", "value": "in the restaurant"}]
|
00000001295.jpg
|
1295_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "choosing products"}]
|
00000001295.jpg
|
1295_evjvqa
|
[{"from": "human", "value": "where is the man standing?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}]
|
00000001295.jpg
|
1295_evjvqa
|
[{"from": "human", "value": "which stall is the man standing at?\n<image>"}, {"from": "gpt", "value": "frozen food"}]
|
00000001296.jpg
|
1296_evjvqa
|
[{"from": "human", "value": "how many men are standing?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002202.jpg
|
2202_evjvqa
|
[{"from": "human", "value": "where is this picture taken?\n<image>"}, {"from": "gpt", "value": "at an ancient tower"}]
|
00000002205.jpg
|
2205_evjvqa
|
[{"from": "human", "value": "where are these two youngers standing at?\n<image>"}, {"from": "gpt", "value": "in front of a store"}]
|
00000002205.jpg
|
2205_evjvqa
|
[{"from": "human", "value": "what is decorated above the youngers?\n<image>"}, {"from": "gpt", "value": "lanterns"}]
|
00000002205.jpg
|
2205_evjvqa
|
[{"from": "human", "value": "who is the man standing with in front of the store?\n<image>"}, {"from": "gpt", "value": "with lady"}]
|
00000002205.jpg
|
2205_evjvqa
|
[{"from": "human", "value": "how many people are sitting in the store?\n<image>"}, {"from": "gpt", "value": "nobody is sitting in the store"}]
|
00000002208.jpg
|
2208_evjvqa
|
[{"from": "human", "value": "how many people are walking with the bike?\n<image>"}, {"from": "gpt", "value": "one person"}]
|
00000002208.jpg
|
2208_evjvqa
|
[{"from": "human", "value": "who is behind the bike of the man walking?\n<image>"}, {"from": "gpt", "value": "a girl"}]
|
00000002208.jpg
|
2208_evjvqa
|
[{"from": "human", "value": "where is the man walking with the bike at?\n<image>"}, {"from": "gpt", "value": "in a market"}]
|
00000002209.jpg
|
2209_evjvqa
|
[{"from": "human", "value": "how many people are riding motorbike?\n<image>"}, {"from": "gpt", "value": "one person"}]
|
00000002209.jpg
|
2209_evjvqa
|
[{"from": "human", "value": "what are the two girls doing?\n<image>"}, {"from": "gpt", "value": "watching the phone"}]
|
00000002209.jpg
|
2209_evjvqa
|
[{"from": "human", "value": "how many people are standing with the girl holding the phone?\n<image>"}, {"from": "gpt", "value": "one person"}]
|
00000002210.jpg
|
2210_evjvqa
|
[{"from": "human", "value": "what is the girl in blue jacket doing?\n<image>"}, {"from": "gpt", "value": "choosing candies at the candy aisle"}]
|
00000002212.jpg
|
2212_evjvqa
|
[{"from": "human", "value": "how many motorbike are moving on the road?\n<image>"}, {"from": "gpt", "value": "there is one motorbike moving on the road"}]
|
00000002215.jpg
|
2215_evjvqa
|
[{"from": "human", "value": "how many people are using the trolleys?\n<image>"}, {"from": "gpt", "value": "two people"}]
|
00000002215.jpg
|
2215_evjvqa
|
[{"from": "human", "value": "where are the woman in purple t-shirt buying at?\n<image>"}, {"from": "gpt", "value": "in a supermarket"}]
|
00000002215.jpg
|
2215_evjvqa
|
[{"from": "human", "value": "where is the woman in yellow shirt with her trolley buying at?\n<image>"}, {"from": "gpt", "value": "in a supermarket"}]
|
00000002216.jpg
|
2216_evjvqa
|
[{"from": "human", "value": "how many people are there in the store?\n<image>"}, {"from": "gpt", "value": "four people"}]
|
00000002216.jpg
|
2216_evjvqa
|
[{"from": "human", "value": "how many tables are used?\n<image>"}, {"from": "gpt", "value": "one table"}]
|
00000002216.jpg
|
2216_evjvqa
|
[{"from": "human", "value": "how many customers are there in the store?\n<image>"}, {"from": "gpt", "value": "four customers"}]
|
00000002216.jpg
|
2216_evjvqa
|
[{"from": "human", "value": "how many waiters are standing there?\n<image>"}, {"from": "gpt", "value": "one waiter"}]
|
00000002216.jpg
|
2216_evjvqa
|
[{"from": "human", "value": "how many people are there in this group of customers sitting in the store?\n<image>"}, {"from": "gpt", "value": "four people"}]
|
00000002218.jpg
|
2218_evjvqa
|
[{"from": "human", "value": "how many people are holding the flowers?\n<image>"}, {"from": "gpt", "value": "two people"}]
|
00000002218.jpg
|
2218_evjvqa
|
[{"from": "human", "value": "how many people are wearing green shirt?\n<image>"}, {"from": "gpt", "value": "one person"}]
|
00000002218.jpg
|
2218_evjvqa
|
[{"from": "human", "value": "what are the first two man doing?\n<image>"}, {"from": "gpt", "value": "holding the flowers"}]
|
00000002219.jpg
|
2219_evjvqa
|
[{"from": "human", "value": "where are the woman in black shirt standing at?\n<image>"}, {"from": "gpt", "value": "in a flower garden"}]
|
00000002224.jpg
|
2224_evjvqa
|
[{"from": "human", "value": "where are those people playing at?\n<image>"}, {"from": "gpt", "value": "in a park"}]
|
00000002225.jpg
|
2225_evjvqa
|
[{"from": "human", "value": "how many people are there watching the products?\n<image>"}, {"from": "gpt", "value": "two people"}]
|
00000002225.jpg
|
2225_evjvqa
|
[{"from": "human", "value": "what is the woman wearing red shirt doing?\n<image>"}, {"from": "gpt", "value": "asking about the products"}]
|
00000002225.jpg
|
2225_evjvqa
|
[{"from": "human", "value": "how many staffs are there consulting for the customers?\n<image>"}, {"from": "gpt", "value": "one staff"}]
|
00000002225.jpg
|
2225_evjvqa
|
[{"from": "human", "value": "what is the staff standing next to the woman wearing red shirt doing?\n<image>"}, {"from": "gpt", "value": "consulting about the products"}]
|
00000002227.jpg
|
2227_evjvqa
|
[{"from": "human", "value": "how many people are using the phones?\n<image>"}, {"from": "gpt", "value": "two people are using the phones"}]
|
00000002227.jpg
|
2227_evjvqa
|
[{"from": "human", "value": "what is the woman wearing white shirt doing?\n<image>"}, {"from": "gpt", "value": "holding the phone for photographing"}]
|
00000002234.jpg
|
2234_evjvqa
|
[{"from": "human", "value": "what are in those two cups?\n<image>"}, {"from": "gpt", "value": "two cups of fruit"}]
|
00000002234.jpg
|
2234_evjvqa
|
[{"from": "human", "value": "how many cups of fruit are served on the table?\n<image>"}, {"from": "gpt", "value": "two cups of fruit"}]
|
00000002234.jpg
|
2234_evjvqa
|
[{"from": "human", "value": "what are the cups of fruit served with?\n<image>"}, {"from": "gpt", "value": "with two boxes of yogurt"}]
|
00000002235.jpg
|
2235_evjvqa
|
[{"from": "human", "value": "where are those people visiting at?\n<image>"}, {"from": "gpt", "value": "in a temple"}]
|
00000002238.jpg
|
2238_evjvqa
|
[{"from": "human", "value": "wat is the girl holding on her hand?\n<image>"}, {"from": "gpt", "value": "a bunch of flower"}]
|
00000002238.jpg
|
2238_evjvqa
|
[{"from": "human", "value": "what color is the house next to the girl?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000002238.jpg
|
2238_evjvqa
|
[{"from": "human", "value": "how many walking people are holding the flowers?\n<image>"}, {"from": "gpt", "value": "one person"}]
|
00000002238.jpg
|
2238_evjvqa
|
[{"from": "human", "value": "how many people are walking with the girl holding the flowers?\n<image>"}, {"from": "gpt", "value": "the girl walks alone"}]
|
00000002240.jpg
|
2240_evjvqa
|
[{"from": "human", "value": "how many people are standing to talk?\n<image>"}, {"from": "gpt", "value": "three people"}]
|
00000002240.jpg
|
2240_evjvqa
|
[{"from": "human", "value": "what color is the shirt of the people standing to talk?\n<image>"}, {"from": "gpt", "value": "blue"}]
|
00000002241.jpg
|
2241_evjvqa
|
[{"from": "human", "value": "where are this couple standing at?\n<image>"}, {"from": "gpt", "value": "next to a pond in an historical destination"}]
|
00000002243.jpg
|
2243_evjvqa
|
[{"from": "human", "value": "where is this tree planted at?\n<image>"}, {"from": "gpt", "value": "on a yard of an ancient house"}]
|
00000002249.jpg
|
2249_evjvqa
|
[{"from": "human", "value": "what is the boat of man wearing red shirt loading ?\n<image>"}, {"from": "gpt", "value": "fruits"}]
|
00000002249.jpg
|
2249_evjvqa
|
[{"from": "human", "value": "what is the man wearing cap and sitting on the boat loading?\n<image>"}, {"from": "gpt", "value": "fruits"}]
|
00000002249.jpg
|
2249_evjvqa
|
[{"from": "human", "value": "how many boats are there next to each other?\n<image>"}, {"from": "gpt", "value": "three boats"}]
|
00000002255.jpg
|
2255_evjvqa
|
[{"from": "human", "value": "what is the man wearing helmet about to buy?\n<image>"}, {"from": "gpt", "value": "eggs"}]
|
00000002255.jpg
|
2255_evjvqa
|
[{"from": "human", "value": "how many people are standing to sell eggs?\n<image>"}, {"from": "gpt", "value": "onw person"}]
|
00000002255.jpg
|
2255_evjvqa
|
[{"from": "human", "value": "who is the man selling eggs with?\n<image>"}, {"from": "gpt", "value": "the man are selling alone"}]
|
00000002255.jpg
|
2255_evjvqa
|
[{"from": "human", "value": "how many people are selling eggs with the man?\n<image>"}, {"from": "gpt", "value": "the man are selling eggs alone"}]
|
00000002255.jpg
|
2255_evjvqa
|
[{"from": "human", "value": "how many people are buying eggs?\n<image>"}, {"from": "gpt", "value": "two people"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.