image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000002591.jpg
2591_evjvqa
[{"from": "human", "value": "what is the main color of the outerwear that the man riding a motorbike wears?\n<image>"}, {"from": "gpt", "value": "the main color of the outerwear that the man riding a motorbike wears is green"}]
00000002591.jpg
2591_evjvqa
[{"from": "human", "value": "where is the young man wearing a white face mask standing?\n<image>"}, {"from": "gpt", "value": "the young man wearing a white face mask is standing on the pavement"}]
00000002800.jpg
2800_evjvqa
[{"from": "human", "value": "how many children does the woman in the blue coat take with her to the supermarket?\n<image>"}, {"from": "gpt", "value": "just one"}]
00000002800.jpg
2800_evjvqa
[{"from": "human", "value": "who is going to the supermarket with the woman in the shirt?\n<image>"}, {"from": "gpt", "value": "with a baby in a red shirt sitting in a stroller"}]
00000002800.jpg
2800_evjvqa
[{"from": "human", "value": "where is the woman in the green overcoat with the baby sitting with?\n<image>"}, {"from": "gpt", "value": "in the trolley"}]
00000002800.jpg
2800_evjvqa
[{"from": "human", "value": "how many people use the trolley in the supermarket?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002800.jpg
2800_evjvqa
[{"from": "human", "value": "who is the woman in black using a shopping trolley with?\n<image>"}, {"from": "gpt", "value": "with three children"}]
00000002800.jpg
2800_evjvqa
[{"from": "human", "value": "how many children are in the trolley out of three children going with the woman in black?\n<image>"}, {"from": "gpt", "value": "only one"}]
00000002806.jpg
2806_evjvqa
[{"from": "human", "value": "what is the room?\n<image>"}, {"from": "gpt", "value": "the gallery space of a museum"}]
00000002807.jpg
2807_evjvqa
[{"from": "human", "value": "which stall is the boy in orange standing?\n<image>"}, {"from": "gpt", "value": "the toy stall"}]
00000002807.jpg
2807_evjvqa
[{"from": "human", "value": "which stall is the boy in green standing?\n<image>"}, {"from": "gpt", "value": "the toy stall"}]
00000002807.jpg
2807_evjvqa
[{"from": "human", "value": "how many children are standing at the toy stall?\n<image>"}, {"from": "gpt", "value": "just one"}]
00000002807.jpg
2807_evjvqa
[{"from": "human", "value": "where is the woman in the red t-shirt?\n<image>"}, {"from": "gpt", "value": "in the toy stall"}]
00000002807.jpg
2807_evjvqa
[{"from": "human", "value": "how many women are coming outside the toy stall?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002807.jpg
2807_evjvqa
[{"from": "human", "value": "how many people are sitting at the checkout?\n<image>"}, {"from": "gpt", "value": "noone"}]
00000002813.jpg
2813_evjvqa
[{"from": "human", "value": "what goods do these boats mainly carry?\n<image>"}, {"from": "gpt", "value": "fruits and vegetables"}]
00000002815.jpg
2815_evjvqa
[{"from": "human", "value": "where are the people standing?\n<image>"}, {"from": "gpt", "value": "in front of the entrance of the historical monument "}]
00000002816.jpg
2816_evjvqa
[{"from": "human", "value": "where is the man in blue walking?\n<image>"}, {"from": "gpt", "value": "on the street"}]
00000002816.jpg
2816_evjvqa
[{"from": "human", "value": "what is the man in blue doing?\n<image>"}, {"from": "gpt", "value": "he is walking on the street"}]
00000002816.jpg
2816_evjvqa
[{"from": "human", "value": "who is walking with the man in blue?\n<image>"}, {"from": "gpt", "value": "no one"}]
00000002816.jpg
2816_evjvqa
[{"from": "human", "value": "what is the woman in yellow doing?\n<image>"}, {"from": "gpt", "value": "she is taking a photo"}]
00000002816.jpg
2816_evjvqa
[{"from": "human", "value": "how many children are walking in front of the woman holding the phone to take a photo?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002817.jpg
2817_evjvqa
[{"from": "human", "value": "what items does this stall sell?\n<image>"}, {"from": "gpt", "value": "clothes is main"}]
00000002818.jpg
2818_evjvqa
[{"from": "human", "value": "where is the woman sitting?\n<image>"}, {"from": "gpt", "value": "on the riverbank"}]
00000002818.jpg
2818_evjvqa
[{"from": "human", "value": "who is sitting with the woman?\n<image>"}, {"from": "gpt", "value": "no one"}]
00000002818.jpg
2818_evjvqa
[{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "she is sitting on the riverbank"}]
00000002823.jpg
2823_evjvqa
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "he is stacking the products on the shelves"}]
00000002823.jpg
2823_evjvqa
[{"from": "human", "value": "what color of the t-shirt is the staff wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
00000002823.jpg
2823_evjvqa
[{"from": "human", "value": "where is the man standing?\n<image>"}, {"from": "gpt", "value": "at the stall in the supermarket"}]
00000002823.jpg
2823_evjvqa
[{"from": "human", "value": "what is the staff doing?\n<image>"}, {"from": "gpt", "value": "he is stacking the products on the shelves"}]
00000002828.jpg
2828_evjvqa
[{"from": "human", "value": "where are the people standing?\n<image>"}, {"from": "gpt", "value": "in the fair "}]
00000002832.jpg
2832_evjvqa
[{"from": "human", "value": "where is the woman in orange?\n<image>"}, {"from": "gpt", "value": "on the small canoe"}]
00000002832.jpg
2832_evjvqa
[{"from": "human", "value": "what is the woman in orange wearing on her head?\n<image>"}, {"from": "gpt", "value": "the conical hat"}]
00000002832.jpg
2832_evjvqa
[{"from": "human", "value": "how many tires are on the boat to the right of the woman in the orange?\n<image>"}, {"from": "gpt", "value": "three"}]
00000002838.jpg
2838_evjvqa
[{"from": "human", "value": "who is standing woth the man in white?\n<image>"}, {"from": "gpt", "value": "with the man in blue"}]
00000002839.jpg
2839_evjvqa
[{"from": "human", "value": "what hats do most of the people use?\n<image>"}, {"from": "gpt", "value": "the conical hats"}]
00000002839.jpg
2839_evjvqa
[{"from": "human", "value": "where are the people standing?\n<image>"}, {"from": "gpt", "value": "in the market"}]
00000002839.jpg
2839_evjvqa
[{"from": "human", "value": "where is this?\n<image>"}, {"from": "gpt", "value": "this is the market"}]
00000002840.jpg
2840_evjvqa
[{"from": "human", "value": "where is the man holding the phone standing to take a photo?\n<image>"}, {"from": "gpt", "value": "near a display of apricot trees"}]
00000002840.jpg
2840_evjvqa
[{"from": "human", "value": "what is the man standing near the display of apricot trees doing?\n<image>"}, {"from": "gpt", "value": "he is taking a photo"}]
00000002840.jpg
2840_evjvqa
[{"from": "human", "value": "how many clowns are performing standing near the man taking the photo?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002841.jpg
2841_evjvqa
[{"from": "human", "value": "which side of the rock is the group of three standing on?\n<image>"}, {"from": "gpt", "value": "on the left"}]
00000002841.jpg
2841_evjvqa
[{"from": "human", "value": "which side of the rock is the group of four standing on?\n<image>"}, {"from": "gpt", "value": "on the right"}]
00000002845.jpg
2845_evjvqa
[{"from": "human", "value": "wher is this?\n<image>"}, {"from": "gpt", "value": "this is the market"}]
00000002847.jpg
2847_evjvqa
[{"from": "human", "value": "what are the scenes around the citadel?\n<image>"}, {"from": "gpt", "value": "the flower gardens"}]
00000002853.jpg
2853_evjvqa
[{"from": "human", "value": "what is the woman wearing the conical hat doing?\n<image>"}, {"from": "gpt", "value": "she is trying to push a trolley"}]
00000002853.jpg
2853_evjvqa
[{"from": "human", "value": "what is the woman wearing the conical hat pushing?\n<image>"}, {"from": "gpt", "value": "the trolley"}]
00000002853.jpg
2853_evjvqa
[{"from": "human", "value": "what is the woman standing on the side of the road holding in her hand?\n<image>"}, {"from": "gpt", "value": "a bunch of vegetables"}]
00000002853.jpg
2853_evjvqa
[{"from": "human", "value": "what is the woman standing on the side of the road sale?\n<image>"}, {"from": "gpt", "value": "vegetable"}]
00000002853.jpg
2853_evjvqa
[{"from": "human", "value": "what is the woman on the side of the road doing?\n<image>"}, {"from": "gpt", "value": "preparing products for sale"}]
00000002853.jpg
2853_evjvqa
[{"from": "human", "value": "how many people are parking on the curb?\n<image>"}, {"from": "gpt", "value": "just one"}]
00000002855.jpg
2855_evjvqa
[{"from": "human", "value": "what are these two men doing?\n<image>"}, {"from": "gpt", "value": "checking products"}]
00000002855.jpg
2855_evjvqa
[{"from": "human", "value": "how do the staff here wear uniforms?\n<image>"}, {"from": "gpt", "value": "they are wearing a uniform with an apron over it"}]
00000002855.jpg
2855_evjvqa
[{"from": "human", "value": "where do these two men work?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}]
00000002861.jpg
2861_evjvqa
[{"from": "human", "value": "where is the woman in the blue shirt with the conical hat?\n<image>"}, {"from": "gpt", "value": "on the handlebars of the bicycle"}]
00000002861.jpg
2861_evjvqa
[{"from": "human", "value": "what means of transport does the woman in the blue shirt travel by?\n<image>"}, {"from": "gpt", "value": "the bicycle"}]
00000002861.jpg
2861_evjvqa
[{"from": "human", "value": "what is the woman in blue doing?\n<image>"}, {"from": "gpt", "value": "she is riding on the road"}]
00000002861.jpg
2861_evjvqa
[{"from": "human", "value": "how many motorbikes are there behind the woman in the blue shirt?\n<image>"}, {"from": "gpt", "value": "only one"}]
00000002861.jpg
2861_evjvqa
[{"from": "human", "value": "how many people on the moving motorbike behind the woman riding the bicycle?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002864.jpg
2864_evjvqa
[{"from": "human", "value": "where is the citadel built?\n<image>"}, {"from": "gpt", "value": "next to the riverbank"}]
00000002865.jpg
2865_evjvqa
[{"from": "human", "value": "which direction is the man driving the car going?\n<image>"}, {"from": "gpt", "value": "he is running away from the historical monument"}]
00000002870.jpg
2870_evjvqa
[{"from": "human", "value": "what are two girls standing in the street doing?\n<image>"}, {"from": "gpt", "value": "they are posing to take a photo"}]
00000002870.jpg
2870_evjvqa
[{"from": "human", "value": "what color shirt is the man sitting on the side of the road wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
00000002870.jpg
2870_evjvqa
[{"from": "human", "value": "how many unoccupied roadside seats are there?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002882.jpg
2882_evjvqa
[{"from": "human", "value": "where are the two people in conical hats going?\n<image>"}, {"from": "gpt", "value": "in the middle of the field"}]
00000002882.jpg
2882_evjvqa
[{"from": "human", "value": "what means of transport do two people wearing conical hats use?\n<image>"}, {"from": "gpt", "value": "the one is riding a buffalo and the one is riding a bicycle"}]
00000002882.jpg
2882_evjvqa
[{"from": "human", "value": "how many buffaloes does the person riding on a buffalo bring to the field?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002884.jpg
2884_evjvqa
[{"from": "human", "value": "where is the woman wearing the conical hat sitting?\n<image>"}, {"from": "gpt", "value": "on the basket"}]
00000002884.jpg
2884_evjvqa
[{"from": "human", "value": "what color of shirt is the girl in the conical hat wearing?\n<image>"}, {"from": "gpt", "value": "blue with white polka dot"}]
00000002884.jpg
2884_evjvqa
[{"from": "human", "value": "who is the girl in conical hat sitting with?\n<image>"}, {"from": "gpt", "value": "no one"}]
00000002885.jpg
2885_evjvqa
[{"from": "human", "value": "who is the woman choosing with?\n<image>"}, {"from": "gpt", "value": "the man is wearing the shirt"}]
00000002885.jpg
2885_evjvqa
[{"from": "human", "value": "which hand does the man wear the watch on?\n<image>"}, {"from": "gpt", "value": "on the left hand"}]
00000002885.jpg
2885_evjvqa
[{"from": "human", "value": "how many people are choosing products?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002889.jpg
2889_evjvqa
[{"from": "human", "value": "how many people are treating the patient?\n<image>"}, {"from": "gpt", "value": "three"}]
00000002889.jpg
2889_evjvqa
[{"from": "human", "value": "where does the doctor sit?\n<image>"}, {"from": "gpt", "value": "between two nurses"}]
00000002889.jpg
2889_evjvqa
[{"from": "human", "value": "how many people come to have medical examination?\n<image>"}, {"from": "gpt", "value": "only one"}]
00000002892.jpg
2892_evjvqa
[{"from": "human", "value": "where are the lanterns hung?\n<image>"}, {"from": "gpt", "value": "in front of a house, next to the table"}]
00000002897.jpg
2897_evjvqa
[{"from": "human", "value": "what is this woman wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing the ao dai and the conical hat"}]
00000001206.jpg
1206_evjvqa
[{"from": "human", "value": "what is girl wearing helmet doing?\n<image>"}, {"from": "gpt", "value": "she is condisering types of milk"}]
00000001206.jpg
1206_evjvqa
[{"from": "human", "value": "how many people are standing at stall to buy milk?\n<image>"}, {"from": "gpt", "value": "there are two"}]
00000001206.jpg
1206_evjvqa
[{"from": "human", "value": "what is the stall behind two woman used for?\n<image>"}, {"from": "gpt", "value": "it is used for selling vegetable"}]
00000001206.jpg
1206_evjvqa
[{"from": "human", "value": "what is the stall next to milk stall sell?\n<image>"}, {"from": "gpt", "value": "this stall sells biscuit"}]
00000001208.jpg
1208_evjvqa
[{"from": "human", "value": "how many motorbikes are parked in front of the \"vai ky\" store?\n<image>"}, {"from": "gpt", "value": "there are two"}]
00000001208.jpg
1208_evjvqa
[{"from": "human", "value": "how many motorbikes are parked in front of the store next to vai ky store?\n<image>"}, {"from": "gpt", "value": "there is just one"}]
00000001213.jpg
1213_evjvqa
[{"from": "human", "value": "where does the girl sit?\n<image>"}, {"from": "gpt", "value": "on the footpath"}]
00000001216.jpg
1216_evjvqa
[{"from": "human", "value": "how many people are riding in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "there are two"}]
00000001216.jpg
1216_evjvqa
[{"from": "human", "value": "where does the man in red t-shirt sit?\n<image>"}, {"from": "gpt", "value": "in front of the restaurant"}]
00000001216.jpg
1216_evjvqa
[{"from": "human", "value": "how many people are sitting in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "just one"}]
00000001216.jpg
1216_evjvqa
[{"from": "human", "value": "how many motorbike are being ridden in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "only one"}]
00000001218.jpg
1218_evjvqa
[{"from": "human", "value": "how many girls are taking a seat with the boy?\n<image>"}, {"from": "gpt", "value": "five"}]
00000001218.jpg
1218_evjvqa
[{"from": "human", "value": "how many girls are sitting around the boy? \n<image>"}, {"from": "gpt", "value": "there are five"}]
00000001220.jpg
1220_evjvqa
[{"from": "human", "value": "what are people doing?\n<image>"}, {"from": "gpt", "value": "they are holding a festival"}]
00000001222.jpg
1222_evjvqa
[{"from": "human", "value": "what is a man in white shirt holding?\n<image>"}, {"from": "gpt", "value": "bunch of incense"}]
00000001222.jpg
1222_evjvqa
[{"from": "human", "value": "where is a man in whie shirt standing?\n<image>"}, {"from": "gpt", "value": "incense tall"}]
00000001222.jpg
1222_evjvqa
[{"from": "human", "value": "what is the man in white shirt doing?\n<image>"}, {"from": "gpt", "value": "he is buying incense"}]
00000001223.jpg
1223_evjvqa
[{"from": "human", "value": "how many serving of food on the table?\n<image>"}, {"from": "gpt", "value": "five "}]
00000001224.jpg
1224_evjvqa
[{"from": "human", "value": "how many people stepping on the platform?\n<image>"}, {"from": "gpt", "value": "two"}]
00000001224.jpg
1224_evjvqa
[{"from": "human", "value": "what is the man in blue shirt doing?\n<image>"}, {"from": "gpt", "value": "he is standing on the podium and holding a bunch of flowers"}]
00000001224.jpg
1224_evjvqa
[{"from": "human", "value": "what is the man in blue shirt holding?\n<image>"}, {"from": "gpt", "value": "a bunch of flowers"}]
00000001225.jpg
1225_evjvqa
[{"from": "human", "value": "where does the girl sit?\n<image>"}, {"from": "gpt", "value": "on the stairs"}]