image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000004091.jpg
4091_evjvqa
[{"from": "human", "value": "where is this tower built?\n<image>"}, {"from": "gpt", "value": "above a lake"}]
00000004093.jpg
4093_evjvqa
[{"from": "human", "value": "where is the woman taking the photo?\n<image>"}, {"from": "gpt", "value": "in front of a store selling christmas decorations"}]
00000004093.jpg
4093_evjvqa
[{"from": "human", "value": "what occasion does the woman go out to take pictures on?\n<image>"}, {"from": "gpt", "value": "on the christmas occasion"}]
00000004093.jpg
4093_evjvqa
[{"from": "human", "value": "what color is the neck scarf that the woman going out on the street wears?\n<image>"}, {"from": "gpt", "value": "a red neck scarf"}]
00000004097.jpg
4097_evjvqa
[{"from": "human", "value": "what color is the shirt that the man shown in this drawing wears?\n<image>"}, {"from": "gpt", "value": "an orange shirt"}]
00000003200.jpg
3200_evjvqa
[{"from": "human", "value": "where is this lady standing?\n<image>"}, {"from": "gpt", "value": "near a mailbox"}]
00000003200.jpg
3200_evjvqa
[{"from": "human", "value": "what kind of clothes does the girl wear?\n<image>"}, {"from": "gpt", "value": "wearing a blue t-shirt, a blue cap, carrying a bag"}]
00000003200.jpg
3200_evjvqa
[{"from": "human", "value": "what is the girl wearing a blue top standing near?\n<image>"}, {"from": "gpt", "value": "near a mailbox"}]
00000003200.jpg
3200_evjvqa
[{"from": "human", "value": "where is the mailbox located?\n<image>"}, {"from": "gpt", "value": "on the sidewalk"}]
00000003204.jpg
3204_evjvqa
[{"from": "human", "value": "where is the girl in black standing?\n<image>"}, {"from": "gpt", "value": "at the side of a road"}]
00000003205.jpg
3205_evjvqa
[{"from": "human", "value": "where are these flowers planted?\n<image>"}, {"from": "gpt", "value": "by the roadside"}]
00000003205.jpg
3205_evjvqa
[{"from": "human", "value": "how many cars are moving on the road?\n<image>"}, {"from": "gpt", "value": "there are two cars moving on the road"}]
00000003205.jpg
3205_evjvqa
[{"from": "human", "value": "where do these flowers grow towards?\n<image>"}, {"from": "gpt", "value": "towards the roadway"}]
00000003211.jpg
3211_evjvqa
[{"from": "human", "value": "where is this group of people standing?\n<image>"}, {"from": "gpt", "value": "on a river"}]
00000003211.jpg
3211_evjvqa
[{"from": "human", "value": "where are these people traveling?\n<image>"}, {"from": "gpt", "value": "at a biological recreational area "}]
00000003211.jpg
3211_evjvqa
[{"from": "human", "value": "what color are the people standing beside the river wearing?\n<image>"}, {"from": "gpt", "value": "wearing brown outfits"}]
00000003211.jpg
3211_evjvqa
[{"from": "human", "value": "how many people are sitting on the boat?\n<image>"}, {"from": "gpt", "value": "there are five people sitting on the boat"}]
00000003211.jpg
3211_evjvqa
[{"from": "human", "value": "how many boats are there with people on them?\n<image>"}, {"from": "gpt", "value": "there are three boats with people sitting on them"}]
00000003211.jpg
3211_evjvqa
[{"from": "human", "value": "how many people are in the water?\n<image>"}, {"from": "gpt", "value": "there is a person in the water"}]
00000003211.jpg
3211_evjvqa
[{"from": "human", "value": "how many people are standing on the shore?\n<image>"}, {"from": "gpt", "value": "there are four people standing on the shore"}]
00000003211.jpg
3211_evjvqa
[{"from": "human", "value": "how many women are wearing non la?\n<image>"}, {"from": "gpt", "value": "there is a lady wearing a non la"}]
00000003218.jpg
3218_evjvqa
[{"from": "human", "value": "where are these people?\n<image>"}, {"from": "gpt", "value": "in a clinic"}]
00000003218.jpg
3218_evjvqa
[{"from": "human", "value": "how many people are in the clinic?\n<image>"}, {"from": "gpt", "value": "there are two people in the clinic"}]
00000003218.jpg
3218_evjvqa
[{"from": "human", "value": "what do these two guys come here for?\n<image>"}, {"from": "gpt", "value": "for health check"}]
00000003221.jpg
3221_evjvqa
[{"from": "human", "value": "where is this group of people walking?\n<image>"}, {"from": "gpt", "value": "on a path leading into a historical site"}]
00000003221.jpg
3221_evjvqa
[{"from": "human", "value": "where does this path lead?\n<image>"}, {"from": "gpt", "value": "leads to a historical site"}]
00000003221.jpg
3221_evjvqa
[{"from": "human", "value": "where do these people go sightseeing?\n<image>"}, {"from": "gpt", "value": "in a historical site"}]
00000003222.jpg
3222_evjvqa
[{"from": "human", "value": "what do these people come here for?\n<image>"}, {"from": "gpt", "value": "for eating"}]
00000003222.jpg
3222_evjvqa
[{"from": "human", "value": "how many people does the woman in the black t-shirt eat with?\n<image>"}, {"from": "gpt", "value": "the woman in black eats with two men"}]
00000003222.jpg
3222_evjvqa
[{"from": "human", "value": "where are these people sitting?\n<image>"}, {"from": "gpt", "value": "in an eatery"}]
00000003222.jpg
3222_evjvqa
[{"from": "human", "value": "what are these people doing at the table?\n<image>"}, {"from": "gpt", "value": "eating"}]
00000003222.jpg
3222_evjvqa
[{"from": "human", "value": "how many people are walking into the eatery?\n<image>"}, {"from": "gpt", "value": "there are two people entering the eatery"}]
00000003222.jpg
3222_evjvqa
[{"from": "human", "value": "with whom does the man wearing a white mask go into the eatery?\n<image>"}, {"from": "gpt", "value": "with a woman whose a hand is holding a pink coat"}]
00000003223.jpg
3223_evjvqa
[{"from": "human", "value": "where is the boy in the black shirt standing?\n<image>"}, {"from": "gpt", "value": "standing at the cosmetics booth"}]
00000003223.jpg
3223_evjvqa
[{"from": "human", "value": "who is the male friend in the black shirt standing at the cosmetics booth with?\n<image>"}, {"from": "gpt", "value": "with a girl in a blue shirt carrying a red handbag"}]
00000003223.jpg
3223_evjvqa
[{"from": "human", "value": "what kind of goods is the booth near the man wearing black shirt selling?\n<image>"}, {"from": "gpt", "value": "selling cosmetics"}]
00000003223.jpg
3223_evjvqa
[{"from": "human", "value": "where are the male friend in black and the female companion standing?\n<image>"}, {"from": "gpt", "value": "at a cosmetics booth"}]
00000003223.jpg
3223_evjvqa
[{"from": "human", "value": "where are the blue shirt girl and the black t-shirt guy going?\n<image>"}, {"from": "gpt", "value": "going sightseeing in a shopping mall"}]
00000003225.jpg
3225_evjvqa
[{"from": "human", "value": "what is the woman holding in her hand?\n<image>"}, {"from": "gpt", "value": "the woman is holding a pine tree"}]
00000003225.jpg
3225_evjvqa
[{"from": "human", "value": "where is the woman in the green helmet standing?\n<image>"}, {"from": "gpt", "value": "standing at a store that sells christmas decorations"}]
00000003225.jpg
3225_evjvqa
[{"from": "human", "value": "what is the woman wearing the green helmet doing?\n<image>"}, {"from": "gpt", "value": "choosing to buy a pine tree"}]
00000003225.jpg
3225_evjvqa
[{"from": "human", "value": "what commodities does the woman intend to buy?\n<image>"}, {"from": "gpt", "value": "intend to buy a pine tree"}]
00000003225.jpg
3225_evjvqa
[{"from": "human", "value": "how many people are standing outside the christmas decorations shop?\n<image>"}, {"from": "gpt", "value": "there are two people standing outside the shop"}]
00000003225.jpg
3225_evjvqa
[{"from": "human", "value": "what kind of goods does the shop sell where the pink jacket woman is standing?\n<image>"}, {"from": "gpt", "value": "sell the christmas ornaments"}]
00000003225.jpg
3225_evjvqa
[{"from": "human", "value": "where is the man passing by?\n<image>"}, {"from": "gpt", "value": "passing by a christmas ornaments shop"}]
00000003227.jpg
3227_evjvqa
[{"from": "human", "value": "how many people are using the trollies?\n<image>"}, {"from": "gpt", "value": "there is a man who is using the trolly"}]
00000003228.jpg
3228_evjvqa
[{"from": "human", "value": "what items is this alley decorated with?\n<image>"}, {"from": "gpt", "value": "with star lightbulbs and small bulbs"}]
00000003228.jpg
3228_evjvqa
[{"from": "human", "value": "how many cars are parked in the alley?\n<image>"}, {"from": "gpt", "value": "there is a car parked in the alley"}]
00000003229.jpg
3229_evjvqa
[{"from": "human", "value": "what means of transport are the men in blue shirts traveling by?\n<image>"}, {"from": "gpt", "value": "by cyclos"}]
00000003229.jpg
3229_evjvqa
[{"from": "human", "value": "where is the girl in yellow ao dai standing?\n<image>"}, {"from": "gpt", "value": "on a street"}]
00000003229.jpg
3229_evjvqa
[{"from": "human", "value": "how many people are traveling by bicycle?\n<image>"}, {"from": "gpt", "value": "there are two women traveling by bicycle"}]
00000003229.jpg
3229_evjvqa
[{"from": "human", "value": "what hats are the two women wearing to ride bicycles?\n<image>"}, {"from": "gpt", "value": "wearing non la"}]
00000003239.jpg
3239_evjvqa
[{"from": "human", "value": "what kind of commodities do these boats carry?\n<image>"}, {"from": "gpt", "value": "carry fruits"}]
00000003239.jpg
3239_evjvqa
[{"from": "human", "value": "what do people use these boats for?\n<image>"}, {"from": "gpt", "value": "for carrying fruits"}]
00000003239.jpg
3239_evjvqa
[{"from": "human", "value": "by what means do these people transport fruit?\n<image>"}, {"from": "gpt", "value": "by boats"}]
00000003239.jpg
3239_evjvqa
[{"from": "human", "value": "by what way do these people transport the fruit?\n<image>"}, {"from": "gpt", "value": "river road"}]
00000003239.jpg
3239_evjvqa
[{"from": "human", "value": "what means of transport do these people use?\n<image>"}, {"from": "gpt", "value": "use boats"}]
00000003239.jpg
3239_evjvqa
[{"from": "human", "value": "where are these people moving?\n<image>"}, {"from": "gpt", "value": "on the river"}]
00000003242.jpg
3242_evjvqa
[{"from": "human", "value": "where is the woman sitting?\n<image>"}, {"from": "gpt", "value": "on a boat"}]
00000003242.jpg
3242_evjvqa
[{"from": "human", "value": "where is the saleswoman?\n<image>"}, {"from": "gpt", "value": "on a boat"}]
00000003242.jpg
3242_evjvqa
[{"from": "human", "value": "what products does the woman's boat sell?\n<image>"}, {"from": "gpt", "value": "sell food"}]
00000003242.jpg
3242_evjvqa
[{"from": "human", "value": "where does this woman sell food?\n<image>"}, {"from": "gpt", "value": "on a boat"}]
00000003242.jpg
3242_evjvqa
[{"from": "human", "value": "what is this woman sitting doing?\n<image>"}, {"from": "gpt", "value": "sitting and preparing to sell food"}]
00000003242.jpg
3242_evjvqa
[{"from": "human", "value": "what means of transport does the woman in the blue shirt move by?\n<image>"}, {"from": "gpt", "value": "by the boat"}]
00000003244.jpg
3244_evjvqa
[{"from": "human", "value": "where is this group of people gathered?\n<image>"}, {"from": "gpt", "value": "under a tree"}]
00000003244.jpg
3244_evjvqa
[{"from": "human", "value": "where are these people standing?\n<image>"}, {"from": "gpt", "value": "in the shade"}]
00000003244.jpg
3244_evjvqa
[{"from": "human", "value": "where is everyone standing?\n<image>"}, {"from": "gpt", "value": "in the shade of a tree"}]
00000003245.jpg
3245_evjvqa
[{"from": "human", "value": "where is the car parked?\n<image>"}, {"from": "gpt", "value": "in the garden of a house"}]
00000003247.jpg
3247_evjvqa
[{"from": "human", "value": "how is the wall in this room decorated?\n<image>"}, {"from": "gpt", "value": "the wall is decorated with a drawing"}]
00000003252.jpg
3252_evjvqa
[{"from": "human", "value": "where are these people moving?\n<image>"}, {"from": "gpt", "value": "on a river"}]
00000003252.jpg
3252_evjvqa
[{"from": "human", "value": "what means of transport do these people use?\n<image>"}, {"from": "gpt", "value": "use boats"}]
00000003256.jpg
3256_evjvqa
[{"from": "human", "value": "what do the stalls in this market sell?\n<image>"}, {"from": "gpt", "value": "sell food products"}]
00000003256.jpg
3256_evjvqa
[{"from": "human", "value": "what products does this market sell?\n<image>"}, {"from": "gpt", "value": "sell food products"}]
00000003256.jpg
3256_evjvqa
[{"from": "human", "value": "what do the stalls sell on either side of the aisle?\n<image>"}, {"from": "gpt", "value": "sell food products and beverages"}]
00000003260.jpg
3260_evjvqa
[{"from": "human", "value": "with whom does the blue t-shirt man go?\n<image>"}, {"from": "gpt", "value": "with a woman standing with a camera"}]
00000003260.jpg
3260_evjvqa
[{"from": "human", "value": "who is the woman in the black shirt with?\n<image>"}, {"from": "gpt", "value": "with the man in the blue shirt standing beside"}]
00000003260.jpg
3260_evjvqa
[{"from": "human", "value": "what is this woman doing?\n<image>"}, {"from": "gpt", "value": "using the camera to take a photo"}]
00000003260.jpg
3260_evjvqa
[{"from": "human", "value": "where are these two standing?\n<image>"}, {"from": "gpt", "value": "next to a railway line"}]
00000003260.jpg
3260_evjvqa
[{"from": "human", "value": "what is the woman taking photos of?\n<image>"}, {"from": "gpt", "value": "the woman is taking pictures of the opposite wall"}]
00000003260.jpg
3260_evjvqa
[{"from": "human", "value": "where is this house at number 30 located?\n<image>"}, {"from": "gpt", "value": "next to a railway line"}]
00000003263.jpg
3263_evjvqa
[{"from": "human", "value": "what is this woman holding in her hand?\n<image>"}, {"from": "gpt", "value": "lucky money envelopes"}]
00000003263.jpg
3263_evjvqa
[{"from": "human", "value": "what is this woman wearing?\n<image>"}, {"from": "gpt", "value": "red ao dai"}]
00000003265.jpg
3265_evjvqa
[{"from": "human", "value": "where are these three men standing?\n<image>"}, {"from": "gpt", "value": "standing at the liquor shelves in a supermarket"}]
00000003265.jpg
3265_evjvqa
[{"from": "human", "value": "what are these three men going to buy?\n<image>"}, {"from": "gpt", "value": "buy bottles of wine"}]
00000003265.jpg
3265_evjvqa
[{"from": "human", "value": "what are these three guys doing?\n<image>"}, {"from": "gpt", "value": "watching bottles of wine"}]
00000003265.jpg
3265_evjvqa
[{"from": "human", "value": "at which stall are these three men standing?\n<image>"}, {"from": "gpt", "value": "wine shelves"}]
00000003266.jpg
3266_evjvqa
[{"from": "human", "value": "where is this woman standing?\n<image>"}, {"from": "gpt", "value": "at a postcard shop"}]
00000003266.jpg
3266_evjvqa
[{"from": "human", "value": "what is the woman holding in her hand?\n<image>"}, {"from": "gpt", "value": "a postcard"}]
00000003266.jpg
3266_evjvqa
[{"from": "human", "value": "what is this lady doing?\n<image>"}, {"from": "gpt", "value": "selecting and buying postcards"}]
00000003269.jpg
3269_evjvqa
[{"from": "human", "value": "how many people are parking on the side of the road?\n<image>"}, {"from": "gpt", "value": "there is a man parking on the side of the road"}]
00000003271.jpg
3271_evjvqa
[{"from": "human", "value": "where are these houses built?\n<image>"}, {"from": "gpt", "value": "on a mountainside"}]
00000003271.jpg
3271_evjvqa
[{"from": "human", "value": "where is this building built?\n<image>"}, {"from": "gpt", "value": "along a mountainside"}]
00000003271.jpg
3271_evjvqa
[{"from": "human", "value": "where is this pagoda built?\n<image>"}, {"from": "gpt", "value": "on a mountainside"}]
00000003273.jpg
3273_evjvqa
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is the path in a market"}]
00000003273.jpg
3273_evjvqa
[{"from": "human", "value": "how many people go to the market by motorbike?\n<image>"}, {"from": "gpt", "value": "there are two people going to the market by motorbike"}]
00000003273.jpg
3273_evjvqa
[{"from": "human", "value": "what do the stalls across the street sell?\n<image>"}, {"from": "gpt", "value": "mainly sell clothes"}]
00000003273.jpg
3273_evjvqa
[{"from": "human", "value": "what do the stall owners on the opposite side mainly sell?\n<image>"}, {"from": "gpt", "value": "selling all kinds of clothes"}]
00000003280.jpg
3280_evjvqa
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "in front of a historical site"}]
00000003284.jpg
3284_evjvqa
[{"from": "human", "value": "where are these women standing?\n<image>"}, {"from": "gpt", "value": "at a shop that sells lanterns and fabrics"}]
00000003284.jpg
3284_evjvqa
[{"from": "human", "value": "what kind of goods is the shop selling where these people are standing?\n<image>"}, {"from": "gpt", "value": "selling lanterns and fabrics"}]