image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000004495.jpg
|
4495_evjvqa
|
[{"from": "human", "value": "where is the place the woman in conical hat selling?\n<image>"}, {"from": "gpt", "value": "on the sidewalk"}]
|
00000004100.jpg
|
4100_evjvqa
|
[{"from": "human", "value": "where is this girl sitting?\n<image>"}, {"from": "gpt", "value": "beside a pink wall"}]
|
00000004100.jpg
|
4100_evjvqa
|
[{"from": "human", "value": "what kind of plants do people plant beside the wall?\n<image>"}, {"from": "gpt", "value": "people plant cactuses"}]
|
00000004100.jpg
|
4100_evjvqa
|
[{"from": "human", "value": "where are these cactuses planted?\n<image>"}, {"from": "gpt", "value": "beside a wall"}]
|
00000004102.jpg
|
4102_evjvqa
|
[{"from": "human", "value": "what is the man in vietnamese silk pajamas doing?\n<image>"}, {"from": "gpt", "value": "pulling a rickshaw"}]
|
00000004102.jpg
|
4102_evjvqa
|
[{"from": "human", "value": "what does the man wearing a conical hat do?\n<image>"}, {"from": "gpt", "value": "towing"}]
|
00000004102.jpg
|
4102_evjvqa
|
[{"from": "human", "value": "where is the man in the conical hat going?\n<image>"}, {"from": "gpt", "value": "going on a street"}]
|
00000004104.jpg
|
4104_evjvqa
|
[{"from": "human", "value": "what is the woman in pink shirt doing?\n<image>"}, {"from": "gpt", "value": "selecting fruits"}]
|
00000004104.jpg
|
4104_evjvqa
|
[{"from": "human", "value": "what item is the woman in pink shirt selecting?\n<image>"}, {"from": "gpt", "value": "selecting fruits"}]
|
00000004110.jpg
|
4110_evjvqa
|
[{"from": "human", "value": "where is the car parked?\n<image>"}, {"from": "gpt", "value": "on the roadside, in front of an apartment complex"}]
|
00000004110.jpg
|
4110_evjvqa
|
[{"from": "human", "value": "how many cars are parked on the roadside?\n<image>"}, {"from": "gpt", "value": "there is one car parked on the roadside"}]
|
00000004110.jpg
|
4110_evjvqa
|
[{"from": "human", "value": "where are the children playing?\n<image>"}, {"from": "gpt", "value": "at the courtyard in front of the apartment complex"}]
|
00000004115.jpg
|
4115_evjvqa
|
[{"from": "human", "value": "what are these woman doing?\n<image>"}, {"from": "gpt", "value": "selecting fruits"}]
|
00000004115.jpg
|
4115_evjvqa
|
[{"from": "human", "value": "how many people are selecting fruits?\n<image>"}, {"from": "gpt", "value": "there are four women selecting fruits"}]
|
00000004115.jpg
|
4115_evjvqa
|
[{"from": "human", "value": "where are these women selecting goods?\n<image>"}, {"from": "gpt", "value": "at a fruit counter"}]
|
00000004116.jpg
|
4116_evjvqa
|
[{"from": "human", "value": "what color are the bags that these people use motorbikes to carry?\n<image>"}, {"from": "gpt", "value": "yellow bags"}]
|
00000004116.jpg
|
4116_evjvqa
|
[{"from": "human", "value": "where do these people use motorbikes to go?\n<image>"}, {"from": "gpt", "value": "to go to the supermarket"}]
|
00000004116.jpg
|
4116_evjvqa
|
[{"from": "human", "value": "what means of transportation do these people go to the supermarket by?\n<image>"}, {"from": "gpt", "value": "by motorbikes"}]
|
00000004119.jpg
|
4119_evjvqa
|
[{"from": "human", "value": "what is this girl doing?\n<image>"}, {"from": "gpt", "value": "selecting fruits"}]
|
00000004119.jpg
|
4119_evjvqa
|
[{"from": "human", "value": "what counter is the girl in black outfit standing at?\n<image>"}, {"from": "gpt", "value": "the girl in black outfit is standing at the apple counter"}]
|
00000004119.jpg
|
4119_evjvqa
|
[{"from": "human", "value": "how many people are selecting apples?\n<image>"}, {"from": "gpt", "value": "there is one girl in black outfit selecting apples"}]
|
00000004120.jpg
|
4120_evjvqa
|
[{"from": "human", "value": "where are these people sitting?\n<image>"}, {"from": "gpt", "value": "in a restaurant"}]
|
00000004121.jpg
|
4121_evjvqa
|
[{"from": "human", "value": "how many people are singing on the stage?\n<image>"}, {"from": "gpt", "value": "there are six people singing on the stage"}]
|
00000004121.jpg
|
4121_evjvqa
|
[{"from": "human", "value": "what repertoire are the people sitting on the road following?\n<image>"}, {"from": "gpt", "value": "they are following an outdoor concert"}]
|
00000004121.jpg
|
4121_evjvqa
|
[{"from": "human", "value": "where is this performance held?\n<image>"}, {"from": "gpt", "value": "outdoor"}]
|
00000004121.jpg
|
4121_evjvqa
|
[{"from": "human", "value": "how do the women singing on stage dress?\n<image>"}, {"from": "gpt", "value": "holding conical hats in hands and wearing brown ao dai"}]
|
00000004126.jpg
|
4126_evjvqa
|
[{"from": "human", "value": "what does this room display?\n<image>"}, {"from": "gpt", "value": "airplane models"}]
|
00000004126.jpg
|
4126_evjvqa
|
[{"from": "human", "value": "where are these airplane models displayed?\n<image>"}, {"from": "gpt", "value": "in a room"}]
|
00000004127.jpg
|
4127_evjvqa
|
[{"from": "human", "value": "what are these people using mobile phones for?\n<image>"}, {"from": "gpt", "value": "for filming and taking pictures"}]
|
00000004127.jpg
|
4127_evjvqa
|
[{"from": "human", "value": "where are these people standing?\n<image>"}, {"from": "gpt", "value": "on a walking street"}]
|
00000004128.jpg
|
4128_evjvqa
|
[{"from": "human", "value": "how many people go behind the girl in pink coat?\n<image>"}, {"from": "gpt", "value": "there is one girl going behind the girl in pink coat"}]
|
00000004128.jpg
|
4128_evjvqa
|
[{"from": "human", "value": "how many friends go ahead of the girl in yellow coat?\n<image>"}, {"from": "gpt", "value": "there is one girl going ahead of the girl in yellow coat"}]
|
00000004128.jpg
|
4128_evjvqa
|
[{"from": "human", "value": "what is the girl going ahead holding?\n<image>"}, {"from": "gpt", "value": "she is holding a phone"}]
|
00000004130.jpg
|
4130_evjvqa
|
[{"from": "human", "value": "where are these people standing?\n<image>"}, {"from": "gpt", "value": "on the courtyard in front of a buddha statue"}]
|
00000004130.jpg
|
4130_evjvqa
|
[{"from": "human", "value": "what are the people who are on the green grass in front of the buddha statue doing?\n<image>"}, {"from": "gpt", "value": "praying in front of a buddha statue"}]
|
00000004130.jpg
|
4130_evjvqa
|
[{"from": "human", "value": "in order to celebrate in front of the buddha statue, where must the people attending the ceremony kneel?\n<image>"}, {"from": "gpt", "value": "on the grass in front of the statue"}]
|
00000004131.jpg
|
4131_evjvqa
|
[{"from": "human", "value": "where does this road lead?\n<image>"}, {"from": "gpt", "value": "into a village"}]
|
00000004131.jpg
|
4131_evjvqa
|
[{"from": "human", "value": "how many people are driving motorbikes on the road?\n<image>"}, {"from": "gpt", "value": "there is one person driving motorbike on the road"}]
|
00000004132.jpg
|
4132_evjvqa
|
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a park"}]
|
00000004132.jpg
|
4132_evjvqa
|
[{"from": "human", "value": "where are the two women sitting?\n<image>"}, {"from": "gpt", "value": "on the bench at the house"}]
|
00000004133.jpg
|
4133_evjvqa
|
[{"from": "human", "value": "how many people is the cyclo carrying?\n<image>"}, {"from": "gpt", "value": "the cyclo is carrying two people"}]
|
00000004133.jpg
|
4133_evjvqa
|
[{"from": "human", "value": "what means of transportation are the man and woman moving on the street by?\n<image>"}, {"from": "gpt", "value": "by a cyclo"}]
|
00000004133.jpg
|
4133_evjvqa
|
[{"from": "human", "value": "where are the woman in white shirt and the man in black shirt sitting on the roadside?\n<image>"}, {"from": "gpt", "value": "sitting in a cafe"}]
|
00000004133.jpg
|
4133_evjvqa
|
[{"from": "human", "value": "where are the lanterns hung for decoration?\n<image>"}, {"from": "gpt", "value": "above a street"}]
|
00000004134.jpg
|
4134_evjvqa
|
[{"from": "human", "value": "where are these items displayed for sale?\n<image>"}, {"from": "gpt", "value": "at a grocery store"}]
|
00000004135.jpg
|
4135_evjvqa
|
[{"from": "human", "value": "what are these two women doing?\n<image>"}, {"from": "gpt", "value": "introducing products of a stall"}]
|
00000004135.jpg
|
4135_evjvqa
|
[{"from": "human", "value": "where are these two women standing?\n<image>"}, {"from": "gpt", "value": "in a stall"}]
|
00000004136.jpg
|
4136_evjvqa
|
[{"from": "human", "value": "in which direction is the motorbike at the citadel gate going?\n<image>"}, {"from": "gpt", "value": "in the direction heading out of the citadel gate"}]
|
00000004136.jpg
|
4136_evjvqa
|
[{"from": "human", "value": "what color is the umbrella that the man pushing the hawker cart uses?\n<image>"}, {"from": "gpt", "value": "a purple umbrella"}]
|
00000004136.jpg
|
4136_evjvqa
|
[{"from": "human", "value": "in which direction is the street vendor pushing the hawker cart?\n<image>"}, {"from": "gpt", "value": "in the direction heading out of the citadel gate"}]
|
00000004138.jpg
|
4138_evjvqa
|
[{"from": "human", "value": "which girl is holding a mobile phone?\n<image>"}, {"from": "gpt", "value": "the girl in white shirt is holding a mobile phone"}]
|
00000004138.jpg
|
4138_evjvqa
|
[{"from": "human", "value": "what is the girl in black shirt doing?\n<image>"}, {"from": "gpt", "value": "taking a photo with the girl standing next to"}]
|
00000004138.jpg
|
4138_evjvqa
|
[{"from": "human", "value": "what are these two girls doing?\n<image>"}, {"from": "gpt", "value": "posing for a photo"}]
|
00000004138.jpg
|
4138_evjvqa
|
[{"from": "human", "value": "whom does the girl holding the mobile phone take the photos with?\n<image>"}, {"from": "gpt", "value": "with the girl in black shirt standing next to"}]
|
00000004140.jpg
|
4140_evjvqa
|
[{"from": "human", "value": "where are these people walking?\n<image>"}, {"from": "gpt", "value": "in a crowded fairground"}]
|
00000004141.jpg
|
4141_evjvqa
|
[{"from": "human", "value": "how many cars are parked by the roadside?\n<image>"}, {"from": "gpt", "value": "there is one car parked by the roadside"}]
|
00000004149.jpg
|
4149_evjvqa
|
[{"from": "human", "value": "how many vehicles are parked by the roadside?\n<image>"}, {"from": "gpt", "value": "there is one vehicle parked by the roadside"}]
|
00000004166.jpg
|
4166_evjvqa
|
[{"from": "human", "value": "what items do these shops sell?\n<image>"}, {"from": "gpt", "value": "these shops sell christmas decorations"}]
|
00000004166.jpg
|
4166_evjvqa
|
[{"from": "human", "value": "what do people ride here to buy?\n<image>"}, {"from": "gpt", "value": "to buy christmas decorations"}]
|
00000004166.jpg
|
4166_evjvqa
|
[{"from": "human", "value": "where is the man holding the stick standing?\n<image>"}, {"from": "gpt", "value": "in front of a shop selling christmas decorations"}]
|
00000004169.jpg
|
4169_evjvqa
|
[{"from": "human", "value": "where is this girl standing?\n<image>"}, {"from": "gpt", "value": "the girl is standing on the balcony of a house"}]
|
00000004169.jpg
|
4169_evjvqa
|
[{"from": "human", "value": "where is the girl standing on the balcony?\n<image>"}, {"from": "gpt", "value": "the girl is standing next to a lamp"}]
|
00000004173.jpg
|
4173_evjvqa
|
[{"from": "human", "value": "what do people build on both sides of the road going into the palace?\n<image>"}, {"from": "gpt", "value": "people build lakes"}]
|
00000004173.jpg
|
4173_evjvqa
|
[{"from": "human", "value": "where does the path in the middle lead?\n<image>"}, {"from": "gpt", "value": "it leads into the courtyard of a palace"}]
|
00000004174.jpg
|
4174_evjvqa
|
[{"from": "human", "value": "how do the woman travel on the street?\n<image>"}, {"from": "gpt", "value": "the woman travel on the street by a bicycle"}]
|
00000004174.jpg
|
4174_evjvqa
|
[{"from": "human", "value": "what does the woman wearing a conical hat do?\n<image>"}, {"from": "gpt", "value": "the woman wearing a conical hat work as a florist"}]
|
00000004174.jpg
|
4174_evjvqa
|
[{"from": "human", "value": "how does the woman wearing a conical hat travel to sell flowers?\n<image>"}, {"from": "gpt", "value": "the woman travel to sell flowers by bicycle"}]
|
00000004174.jpg
|
4174_evjvqa
|
[{"from": "human", "value": "where is the woman selling flower going?\n<image>"}, {"from": "gpt", "value": "by the roadside, past a building"}]
|
00000004175.jpg
|
4175_evjvqa
|
[{"from": "human", "value": "where are these people travelling?\n<image>"}, {"from": "gpt", "value": "these people are travelling at the sea"}]
|
00000004175.jpg
|
4175_evjvqa
|
[{"from": "human", "value": "where are these women standing?\n<image>"}, {"from": "gpt", "value": "they are standing on the beach"}]
|
00000004175.jpg
|
4175_evjvqa
|
[{"from": "human", "value": "how many women are standing on the beach?\n<image>"}, {"from": "gpt", "value": "there are three women standing on the beach"}]
|
00000004177.jpg
|
4177_evjvqa
|
[{"from": "human", "value": "what items does this store specialize in?\n<image>"}, {"from": "gpt", "value": "fitness equipment"}]
|
00000004178.jpg
|
4178_evjvqa
|
[{"from": "human", "value": "what goods are these boats used for transporting?\n<image>"}, {"from": "gpt", "value": "for transporting kinds of fruits, vegetables"}]
|
00000004178.jpg
|
4178_evjvqa
|
[{"from": "human", "value": "what way are these fruit items transported by?\n<image>"}, {"from": "gpt", "value": "transported by river"}]
|
00000004179.jpg
|
4179_evjvqa
|
[{"from": "human", "value": "where is this billboard located?\n<image>"}, {"from": "gpt", "value": "outside a wall of a building under construction"}]
|
00000004180.jpg
|
4180_evjvqa
|
[{"from": "human", "value": "where is this man sitting?\n<image>"}, {"from": "gpt", "value": "this man is sitting in a meeting room"}]
|
00000004180.jpg
|
4180_evjvqa
|
[{"from": "human", "value": "what is this man doing?\n<image>"}, {"from": "gpt", "value": "this man is attending a meeting"}]
|
00000004180.jpg
|
4180_evjvqa
|
[{"from": "human", "value": "how many men are sitting in the room?\n<image>"}, {"from": "gpt", "value": "there is one man sitting in the room"}]
|
00000004181.jpg
|
4181_evjvqa
|
[{"from": "human", "value": "what are the counters here decorated with?\n<image>"}, {"from": "gpt", "value": "with bubbles"}]
|
00000004181.jpg
|
4181_evjvqa
|
[{"from": "human", "value": "what do these counters sell?\n<image>"}, {"from": "gpt", "value": "fruits, agricultural products"}]
|
00000004185.jpg
|
4185_evjvqa
|
[{"from": "human", "value": "what is the woman in yellow shirt standing to do?\n<image>"}, {"from": "gpt", "value": "to sell rice"}]
|
00000004185.jpg
|
4185_evjvqa
|
[{"from": "human", "value": "where is the woman in yellow shirt standing?\n<image>"}, {"from": "gpt", "value": "in a restaurant selling rice"}]
|
00000004191.jpg
|
4191_evjvqa
|
[{"from": "human", "value": "where is this chicken going?\n<image>"}, {"from": "gpt", "value": "on a well mouth"}]
|
00000004193.jpg
|
4193_evjvqa
|
[{"from": "human", "value": "where is this tv located?\n<image>"}, {"from": "gpt", "value": "in a supermarket"}]
|
00000004193.jpg
|
4193_evjvqa
|
[{"from": "human", "value": "what items do these counters sell?\n<image>"}, {"from": "gpt", "value": "kinds of fruits"}]
|
00000004193.jpg
|
4193_evjvqa
|
[{"from": "human", "value": "what items does this area sell?\n<image>"}, {"from": "gpt", "value": "this area sells kinds of fruits"}]
|
00000004194.jpg
|
4194_evjvqa
|
[{"from": "human", "value": "from which direction is this scene taken?\n<image>"}, {"from": "gpt", "value": "the direction looking down from above"}]
|
00000004195.jpg
|
4195_evjvqa
|
[{"from": "human", "value": "where is the woman riding the bicycle running out from?\n<image>"}, {"from": "gpt", "value": "from an alley"}]
|
00000004195.jpg
|
4195_evjvqa
|
[{"from": "human", "value": "in which direction is the woman on the bicycle running?\n<image>"}, {"from": "gpt", "value": "towards the right hand side of the alley"}]
|
00000004195.jpg
|
4195_evjvqa
|
[{"from": "human", "value": "how many cars are parked by the left roadside?\n<image>"}, {"from": "gpt", "value": "there is one car parked"}]
|
00000004199.jpg
|
4199_evjvqa
|
[{"from": "human", "value": "where is this girl standing?\n<image>"}, {"from": "gpt", "value": "in a field of flowers"}]
|
00000004199.jpg
|
4199_evjvqa
|
[{"from": "human", "value": "what is the girl holding in her hand?\n<image>"}, {"from": "gpt", "value": "the girl is holding a bouquet"}]
|
00000001301.jpg
|
1301_evjvqa
|
[{"from": "human", "value": "what is thw woman holding the knife doing?\n<image>"}, {"from": "gpt", "value": "she is cutting pork to sell to the customer"}]
|
00000001301.jpg
|
1301_evjvqa
|
[{"from": "human", "value": "what is the woman standing opposite the woman holding the knife doing?\n<image>"}, {"from": "gpt", "value": "she is buying pork"}]
|
00000001301.jpg
|
1301_evjvqa
|
[{"from": "human", "value": "which product is the woman in brown overcoat buying?\n<image>"}, {"from": "gpt", "value": "pork"}]
|
00000001301.jpg
|
1301_evjvqa
|
[{"from": "human", "value": "what does the stall sell?\n<image>"}, {"from": "gpt", "value": "pork"}]
|
00000001303.jpg
|
1303_evjvqa
|
[{"from": "human", "value": "which direction is the woman looking?\n<image>"}, {"from": "gpt", "value": "menu"}]
|
00000001303.jpg
|
1303_evjvqa
|
[{"from": "human", "value": "who is the girl in green shirt going with?\n<image>"}, {"from": "gpt", "value": "with the man wearing sky blue with horizontal stripes"}]
|
00000001303.jpg
|
1303_evjvqa
|
[{"from": "human", "value": "who is the bald man standing with?\n<image>"}, {"from": "gpt", "value": "the woman is in green shirt"}]
|
00000001305.jpg
|
1305_evjvqa
|
[{"from": "human", "value": "which side are the two girls standing talking compared to the woman in black dress?\n<image>"}, {"from": "gpt", "value": "in the behind"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.