image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000002497.jpg
2497_evjvqa
[{"from": "human", "value": "what does this ship carry?\n<image>"}, {"from": "gpt", "value": "containers"}]
00000002497.jpg
2497_evjvqa
[{"from": "human", "value": "what is the ship carrying on the sea?\n<image>"}, {"from": "gpt", "value": "containers"}]
00000004002.jpg
4002_evjvqa
[{"from": "human", "value": "how many motorbikes are parked on the pavement?\n<image>"}, {"from": "gpt", "value": "three motorbikes are parked on the pavement"}]
00000004002.jpg
4002_evjvqa
[{"from": "human", "value": "where are these three motorbikes parked?\n<image>"}, {"from": "gpt", "value": "on the pavement, in front of a cafe"}]
00000004004.jpg
4004_evjvqa
[{"from": "human", "value": "how many people are walking on the pavement?\n<image>"}, {"from": "gpt", "value": "two people are walking on the pavement"}]
00000004004.jpg
4004_evjvqa
[{"from": "human", "value": "how many people are walking on the right pavement?\n<image>"}, {"from": "gpt", "value": "two people are walking"}]
00000004004.jpg
4004_evjvqa
[{"from": "human", "value": "how many cars are parked on the pavement?\n<image>"}, {"from": "gpt", "value": "one car is parked on the pavement"}]
00000004006.jpg
4006_evjvqa
[{"from": "human", "value": "where is the woman in yellow jacket with blue helmet driving?\n<image>"}, {"from": "gpt", "value": "right behind the black car"}]
00000004006.jpg
4006_evjvqa
[{"from": "human", "value": "how many cars are running behind the woman in yellow coat?\n<image>"}, {"from": "gpt", "value": "there is one car running behind the woman in yellow coat"}]
00000004006.jpg
4006_evjvqa
[{"from": "human", "value": "where is the black car running?\n<image>"}, {"from": "gpt", "value": "right behind the white car"}]
00000004006.jpg
4006_evjvqa
[{"from": "human", "value": "how many ways is this road?\n<image>"}, {"from": "gpt", "value": "this is a one way road"}]
00000004007.jpg
4007_evjvqa
[{"from": "human", "value": "where do these people stand gathering?\n<image>"}, {"from": "gpt", "value": "at the cashier counters"}]
00000004008.jpg
4008_evjvqa
[{"from": "human", "value": "what is this area?\n<image>"}, {"from": "gpt", "value": "this area is a food court"}]
00000004008.jpg
4008_evjvqa
[{"from": "human", "value": "what do people come here for?\n<image>"}, {"from": "gpt", "value": "for eating and drinking"}]
00000004009.jpg
4009_evjvqa
[{"from": "human", "value": "what products do people sell on the roadway?\n<image>"}, {"from": "gpt", "value": "kinds of flowers and kinds of fruits"}]
00000004010.jpg
4010_evjvqa
[{"from": "human", "value": "how many men do the picture draw?\n<image>"}, {"from": "gpt", "value": "two men"}]
00000004010.jpg
4010_evjvqa
[{"from": "human", "value": "what are the two men in the picture doing?\n<image>"}, {"from": "gpt", "value": "shaking hands"}]
00000004011.jpg
4011_evjvqa
[{"from": "human", "value": "where is the woman in blue coat going?\n<image>"}, {"from": "gpt", "value": "going to the market"}]
00000004011.jpg
4011_evjvqa
[{"from": "human", "value": "where is the woman in red shirt sitting on the vehicle intending to go?\n<image>"}, {"from": "gpt", "value": "intending to go to the market"}]
00000004012.jpg
4012_evjvqa
[{"from": "human", "value": "what are these girls doing?\n<image>"}, {"from": "gpt", "value": "posing for a photo"}]
00000004012.jpg
4012_evjvqa
[{"from": "human", "value": "where are these girls standing?\n<image>"}, {"from": "gpt", "value": "in front of a house"}]
00000004015.jpg
4015_evjvqa
[{"from": "human", "value": "what is the woman in gray shirt standing to do?\n<image>"}, {"from": "gpt", "value": "standing to talk for recording"}]
00000004015.jpg
4015_evjvqa
[{"from": "human", "value": "how many men are driving motorbikes?\n<image>"}, {"from": "gpt", "value": "three men are driving motorbikes"}]
00000004015.jpg
4015_evjvqa
[{"from": "human", "value": "how many people stand outside the store to look after the vehicles?\n<image>"}, {"from": "gpt", "value": "one man stand outside the store to look after the vehicles"}]
00000004015.jpg
4015_evjvqa
[{"from": "human", "value": "where is the man in red shirt going?\n<image>"}, {"from": "gpt", "value": "to a store to buy food"}]
00000004015.jpg
4015_evjvqa
[{"from": "human", "value": "how many men riding vehicles wear green coats?\n<image>"}, {"from": "gpt", "value": "two men riding vehicles wear green coats"}]
00000004015.jpg
4015_evjvqa
[{"from": "human", "value": "how many men in red coats drive motorbikes?\n<image>"}, {"from": "gpt", "value": "one man in red coat drives motorbike"}]
00000004016.jpg
4016_evjvqa
[{"from": "human", "value": "where does this man take the little boy to?\n<image>"}, {"from": "gpt", "value": "to a supermarket"}]
00000004016.jpg
4016_evjvqa
[{"from": "human", "value": "whom does the little boy go to the supermarket with?\n<image>"}, {"from": "gpt", "value": "with a man in black shirt going beside"}]
00000004016.jpg
4016_evjvqa
[{"from": "human", "value": "how many people are walking on the corridor of the supermarket?\n<image>"}, {"from": "gpt", "value": "a little boy and a man"}]
00000004018.jpg
4018_evjvqa
[{"from": "human", "value": "what means of transportations do people going on this street mainly use?\n<image>"}, {"from": "gpt", "value": "they mainly walk"}]
00000004018.jpg
4018_evjvqa
[{"from": "human", "value": "what is this street decorated with above?\n<image>"}, {"from": "gpt", "value": "decorated with the lanterns"}]
00000004021.jpg
4021_evjvqa
[{"from": "human", "value": "where are these people visiting?\n<image>"}, {"from": "gpt", "value": "in a fairground"}]
00000004022.jpg
4022_evjvqa
[{"from": "human", "value": "where are these people standing?\n<image>"}, {"from": "gpt", "value": "in a supermarket"}]
00000004023.jpg
4023_evjvqa
[{"from": "human", "value": "where is this woman sitting?\n<image>"}, {"from": "gpt", "value": "by a pavement"}]
00000004026.jpg
4026_evjvqa
[{"from": "human", "value": "where are these children sitting?\n<image>"}, {"from": "gpt", "value": "in an antiquities gallery of a museum"}]
00000004026.jpg
4026_evjvqa
[{"from": "human", "value": "what are the kids sitting to do?\n<image>"}, {"from": "gpt", "value": "sitting to listen to the explanations about the antiques displayed in the room"}]
00000004026.jpg
4026_evjvqa
[{"from": "human", "value": "what is the woman standing to do?\n<image>"}, {"from": "gpt", "value": "to present the artifacts on display"}]
00000004027.jpg
4027_evjvqa
[{"from": "human", "value": "what is the supermarket employee in red shirt doing?\n<image>"}, {"from": "gpt", "value": "preparing to line up the shelves"}]
00000004028.jpg
4028_evjvqa
[{"from": "human", "value": "where is the woman standing?\n<image>"}, {"from": "gpt", "value": "beside a tank on display in the open air"}]
00000004028.jpg
4028_evjvqa
[{"from": "human", "value": "where is the tank displayed?\n<image>"}, {"from": "gpt", "value": "outside a building"}]
00000004029.jpg
4029_evjvqa
[{"from": "human", "value": "what items does this trolley mainly contain?\n<image>"}, {"from": "gpt", "value": "fruits and vegetables, soft drinks"}]
00000004030.jpg
4030_evjvqa
[{"from": "human", "value": "what items does this market sell?\n<image>"}, {"from": "gpt", "value": "fabrics, clothes"}]
00000004030.jpg
4030_evjvqa
[{"from": "human", "value": "what items do the stalls on both sides of the aisle sell?\n<image>"}, {"from": "gpt", "value": "clothes, fabrics"}]
00000004035.jpg
4035_evjvqa
[{"from": "human", "value": "where are the three children playing?\n<image>"}, {"from": "gpt", "value": "in front of a cafe"}]
00000004035.jpg
4035_evjvqa
[{"from": "human", "value": "how many children are sitting in front of the cafe?\n<image>"}, {"from": "gpt", "value": "three children are sitting in front of the cafe"}]
00000004044.jpg
4044_evjvqa
[{"from": "human", "value": "what department are the two women looking at?\n<image>"}, {"from": "gpt", "value": "the shower gel department"}]
00000004044.jpg
4044_evjvqa
[{"from": "human", "value": "what is the woman in white shirt holding in her hand?\n<image>"}, {"from": "gpt", "value": "a bottle of shower gel"}]
00000004044.jpg
4044_evjvqa
[{"from": "human", "value": "how many people are using carts?\n<image>"}, {"from": "gpt", "value": "two women are using a cart"}]
00000004044.jpg
4044_evjvqa
[{"from": "human", "value": "what department do the two women push the cart to?\n<image>"}, {"from": "gpt", "value": "the shower gel department"}]
00000004045.jpg
4045_evjvqa
[{"from": "human", "value": "where is this room built?\n<image>"}, {"from": "gpt", "value": "in an antique house"}]
00000004047.jpg
4047_evjvqa
[{"from": "human", "value": "how many bowls of food are there on the table?\n<image>"}, {"from": "gpt", "value": "there are three bowls of food on the table"}]
00000004047.jpg
4047_evjvqa
[{"from": "human", "value": "how many people eat this portion?\n<image>"}, {"from": "gpt", "value": "three people"}]
00000004048.jpg
4048_evjvqa
[{"from": "human", "value": "what items do these women come to buy?\n<image>"}, {"from": "gpt", "value": "buy types of items for tet holiday"}]
00000004048.jpg
4048_evjvqa
[{"from": "human", "value": "what is the woman standing in the stall doing?\n<image>"}, {"from": "gpt", "value": "counselling customers to buy goods"}]
00000004048.jpg
4048_evjvqa
[{"from": "human", "value": "what does the three women come here to buy?\n<image>"}, {"from": "gpt", "value": "to buy types of items for tet holiday"}]
00000004053.jpg
4053_evjvqa
[{"from": "human", "value": "what are these women doing?\n<image>"}, {"from": "gpt", "value": "spreading the cotton plants to dry"}]
00000004053.jpg
4053_evjvqa
[{"from": "human", "value": "where are these women working?\n<image>"}, {"from": "gpt", "value": "on the water"}]
00000004053.jpg
4053_evjvqa
[{"from": "human", "value": "what hats do the women working on the water all wear?\n<image>"}, {"from": "gpt", "value": "conical hats"}]
00000004053.jpg
4053_evjvqa
[{"from": "human", "value": "how many women are working on the water?\n<image>"}, {"from": "gpt", "value": "there are five women working on the water"}]
00000004057.jpg
4057_evjvqa
[{"from": "human", "value": "where is this woman walking?\n<image>"}, {"from": "gpt", "value": "in front of a karaoke"}]
00000004061.jpg
4061_evjvqa
[{"from": "human", "value": "what does this young person park by the side of the road for?\n<image>"}, {"from": "gpt", "value": "to go to the sidewalk eatery"}]
00000004061.jpg
4061_evjvqa
[{"from": "human", "value": "how many people are driving motorbikes on the road?\n<image>"}, {"from": "gpt", "value": "there is one person driving motorbike on the road"}]
00000004061.jpg
4061_evjvqa
[{"from": "human", "value": "what color is the motorbike running on the road painted?\n<image>"}, {"from": "gpt", "value": "red"}]
00000004061.jpg
4061_evjvqa
[{"from": "human", "value": "how many people are walking across the road?\n<image>"}, {"from": "gpt", "value": "there is one man in white shirt crossing across the road"}]
00000004062.jpg
4062_evjvqa
[{"from": "human", "value": "what products do these stalls sell?\n<image>"}, {"from": "gpt", "value": "types of fast food"}]
00000004062.jpg
4062_evjvqa
[{"from": "human", "value": "where are these food stalls displayed for sale?\n<image>"}, {"from": "gpt", "value": "in a fairground"}]
00000004063.jpg
4063_evjvqa
[{"from": "human", "value": "what is the woman in the conical hat doing?\n<image>"}, {"from": "gpt", "value": "pushing the cart on the road"}]
00000004064.jpg
4064_evjvqa
[{"from": "human", "value": "where are the motorbikes allowed to move?\n<image>"}, {"from": "gpt", "value": "on either side of the railway"}]
00000004065.jpg
4065_evjvqa
[{"from": "human", "value": "where is the car being parked?\n<image>"}, {"from": "gpt", "value": "outside an antique house which is painted yellow"}]
00000004067.jpg
4067_evjvqa
[{"from": "human", "value": "how many servings are laid on the table?\n<image>"}, {"from": "gpt", "value": "there are two servings laid on the table"}]
00000004069.jpg
4069_evjvqa
[{"from": "human", "value": "what is the young man doing?\n<image>"}, {"from": "gpt", "value": "sitting and working beside a computer"}]
00000004069.jpg
4069_evjvqa
[{"from": "human", "value": "where is the young man sitting and working?\n<image>"}, {"from": "gpt", "value": "in a cafe"}]
00000004069.jpg
4069_evjvqa
[{"from": "human", "value": "what device does the young man work with?\n<image>"}, {"from": "gpt", "value": "with a computer"}]
00000004069.jpg
4069_evjvqa
[{"from": "human", "value": "whom does the young man sit to work with?\n<image>"}, {"from": "gpt", "value": "the young man sits to work alone"}]
00000004070.jpg
4070_evjvqa
[{"from": "human", "value": "what do these people gather in crowds for?\n<image>"}, {"from": "gpt", "value": "to see a performance by a street band"}]
00000004070.jpg
4070_evjvqa
[{"from": "human", "value": "where do the people gather?\n<image>"}, {"from": "gpt", "value": "around a street band"}]
00000004070.jpg
4070_evjvqa
[{"from": "human", "value": "what do people gather in crowds here for?\n<image>"}, {"from": "gpt", "value": "to see a street music performance by a band"}]
00000004070.jpg
4070_evjvqa
[{"from": "human", "value": "how many people in the band are singing?\n<image>"}, {"from": "gpt", "value": "there is one person in the band singing"}]
00000004070.jpg
4070_evjvqa
[{"from": "human", "value": "what are the people using musical instruments doing?\n<image>"}, {"from": "gpt", "value": "performing for the people who are watching"}]
00000004073.jpg
4073_evjvqa
[{"from": "human", "value": "where are these people walking?\n<image>"}, {"from": "gpt", "value": "in a market"}]
00000004073.jpg
4073_evjvqa
[{"from": "human", "value": "where are these people going?\n<image>"}, {"from": "gpt", "value": "going in a market"}]
00000004077.jpg
4077_evjvqa
[{"from": "human", "value": "what items do the stalls here display?\n<image>"}, {"from": "gpt", "value": "types of books"}]
00000004077.jpg
4077_evjvqa
[{"from": "human", "value": "where are the book stalls displayed?\n<image>"}, {"from": "gpt", "value": "on a street"}]
00000004077.jpg
4077_evjvqa
[{"from": "human", "value": "what do passers-by go to these stalls mainly to see?\n<image>"}, {"from": "gpt", "value": "to see types of books"}]
00000004079.jpg
4079_evjvqa
[{"from": "human", "value": "what does this area display?\n<image>"}, {"from": "gpt", "value": "the war cannons"}]
00000004079.jpg
4079_evjvqa
[{"from": "human", "value": "what do people display in this compartment?\n<image>"}, {"from": "gpt", "value": "cannons"}]
00000004082.jpg
4082_evjvqa
[{"from": "human", "value": "what item are these two women looking at?\n<image>"}, {"from": "gpt", "value": "they are looking at the fresh meat"}]
00000004082.jpg
4082_evjvqa
[{"from": "human", "value": "where are these two people standing?\n<image>"}, {"from": "gpt", "value": "at the fresh meat department in the supermarket"}]
00000004085.jpg
4085_evjvqa
[{"from": "human", "value": "what item is the woman in white coat intending to buy?\n<image>"}, {"from": "gpt", "value": "the woman in white coat is intending to buy some bottles of cooking oil"}]
00000004085.jpg
4085_evjvqa
[{"from": "human", "value": "what item is the woman in the yellow shirt coming to select?\n<image>"}, {"from": "gpt", "value": "the woman in yellow shirt is selecting cooking oil"}]
00000004085.jpg
4085_evjvqa
[{"from": "human", "value": "what do these two women use to go shopping at the supermarket?\n<image>"}, {"from": "gpt", "value": "trolley"}]
00000004085.jpg
4085_evjvqa
[{"from": "human", "value": "what department are these two women standing at?\n<image>"}, {"from": "gpt", "value": "the cooking oil department"}]
00000004086.jpg
4086_evjvqa
[{"from": "human", "value": "what is the man in blue shirt doing?\n<image>"}, {"from": "gpt", "value": "taking the body temperature of the woman sitting in the fruit and vegetable stall"}]
00000004086.jpg
4086_evjvqa
[{"from": "human", "value": "where is the woman in red shirt sitting?\n<image>"}, {"from": "gpt", "value": "in the fruit and vegetable stall"}]
00000004086.jpg
4086_evjvqa
[{"from": "human", "value": "what color is the shirt that the woman sitting in the fruit and vegetable stall wears?\n<image>"}, {"from": "gpt", "value": "a red shirt"}]
00000004086.jpg
4086_evjvqa
[{"from": "human", "value": "what is the woman sitting in the fruit and vegetable stall doing?\n<image>"}, {"from": "gpt", "value": "taking body temperature"}]
00000004086.jpg
4086_evjvqa
[{"from": "human", "value": "who is the person to take the body temperature of the woman selling vegetables and fruits?\n<image>"}, {"from": "gpt", "value": "the man in blue shirt"}]
00000004086.jpg
4086_evjvqa
[{"from": "human", "value": "where does the woman in the red shirt display the stall selling vegetables?\n<image>"}, {"from": "gpt", "value": "in a market"}]
00000004088.jpg
4088_evjvqa
[{"from": "human", "value": "where is this group of friends traveling?\n<image>"}, {"from": "gpt", "value": "in an amusement park"}]