image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000005067.jpg
5067_evjvqa
[{"from": "human", "value": "what color is the scarf of the female driver?\n<image>"}, {"from": "gpt", "value": "it is rose red. "}]
00000005068.jpg
5068_evjvqa
[{"from": "human", "value": "what color is the weighing scales?\n<image>"}, {"from": "gpt", "value": "the scales are green. "}]
00000005068.jpg
5068_evjvqa
[{"from": "human", "value": "what color is the woman's jacket?\n<image>"}, {"from": "gpt", "value": "it is green. "}]
00000005069.jpg
5069_evjvqa
[{"from": "human", "value": "how many children are there in the picture?\n<image>"}, {"from": "gpt", "value": "there are three children. "}]
00000005069.jpg
5069_evjvqa
[{"from": "human", "value": "what color is the top of the children's school uniform?\n<image>"}, {"from": "gpt", "value": "it is white. "}]
00000005069.jpg
5069_evjvqa
[{"from": "human", "value": "what color is the bottom of the children's school uniform?\n<image>"}, {"from": "gpt", "value": "it is blue. "}]
00000005073.jpg
5073_evjvqa
[{"from": "human", "value": "where is the girl in the red polo shirt?\n<image>"}, {"from": "gpt", "value": "she is in the mall. "}]
00000005073.jpg
5073_evjvqa
[{"from": "human", "value": "which hand is she using to hold her phone? \n<image>"}, {"from": "gpt", "value": "she is holding her phone with her left hand. "}]
00000005073.jpg
5073_evjvqa
[{"from": "human", "value": "what color is the girl's collar?\n<image>"}, {"from": "gpt", "value": "her collar is yellow. "}]
00000005077.jpg
5077_evjvqa
[{"from": "human", "value": "what type of vehicle is the man driving on the road?\n<image>"}, {"from": "gpt", "value": "he is driving a motorbike. "}]
00000005080.jpg
5080_evjvqa
[{"from": "human", "value": "how many red plastic containers are there ?\n<image>"}, {"from": "gpt", "value": "there are two red plastic container. "}]
00000005080.jpg
5080_evjvqa
[{"from": "human", "value": "how are the red containers arranged in line with the green ones?\n<image>"}, {"from": "gpt", "value": "the red containers are in between the green ones. "}]
00000005080.jpg
5080_evjvqa
[{"from": "human", "value": "how many pink containers are there on the table?\n<image>"}, {"from": "gpt", "value": "there is one pink container."}]
00000005081.jpg
5081_evjvqa
[{"from": "human", "value": "what time of day was the picture taken?\n<image>"}, {"from": "gpt", "value": "it was taken in the evening."}]
00000005081.jpg
5081_evjvqa
[{"from": "human", "value": "what are they doing in the picture?\n<image>"}, {"from": "gpt", "value": "they are dancing?"}]
00000005081.jpg
5081_evjvqa
[{"from": "human", "value": "what costume are most of the women in the photo wearing?\n<image>"}, {"from": "gpt", "value": "they are wearing the traditional ao dai. "}]
00000005087.jpg
5087_evjvqa
[{"from": "human", "value": "where was the picture taken?\n<image>"}, {"from": "gpt", "value": "at a grocery store. "}]
00000005087.jpg
5087_evjvqa
[{"from": "human", "value": "what color is the soft drink rack?\n<image>"}, {"from": "gpt", "value": "it is red."}]
00000005087.jpg
5087_evjvqa
[{"from": "human", "value": "what does the red rack hold?\n<image>"}, {"from": "gpt", "value": "the red rack holds different types of soft drinks."}]
00000005087.jpg
5087_evjvqa
[{"from": "human", "value": "what color is the ball container?\n<image>"}, {"from": "gpt", "value": "it is red. "}]
00000005095.jpg
5095_evjvqa
[{"from": "human", "value": "what color is the shirt of the man wearing glasses?\n<image>"}, {"from": "gpt", "value": "it is grey. "}]
00000005096.jpg
5096_evjvqa
[{"from": "human", "value": "how many circles are there in the picture?\n<image>"}, {"from": "gpt", "value": "there are eight circles."}]
00000005097.jpg
5097_evjvqa
[{"from": "human", "value": "what are the two women doing?\n<image>"}, {"from": "gpt", "value": "they are shaking hands with each other."}]
00000005097.jpg
5097_evjvqa
[{"from": "human", "value": "what is the man's expression?\n<image>"}, {"from": "gpt", "value": "the man looks cheerful. "}]
00000005098.jpg
5098_evjvqa
[{"from": "human", "value": "where is the woman in the red top?\n<image>"}, {"from": "gpt", "value": "she is at the shopping center. "}]
00000005098.jpg
5098_evjvqa
[{"from": "human", "value": "what color is the mask of the woman in the red top?\n<image>"}, {"from": "gpt", "value": "it is blue. "}]
00000005098.jpg
5098_evjvqa
[{"from": "human", "value": "what color is the trolley of the woman in the red top?\n<image>"}, {"from": "gpt", "value": "it is yellow and black. "}]
00000005099.jpg
5099_evjvqa
[{"from": "human", "value": "what color are the windows?\n<image>"}, {"from": "gpt", "value": "they are white. "}]
00000002700.jpg
2700_evjvqa
[{"from": "human", "value": "what kind of costume is the man pointing to the river wearing?\n<image>"}, {"from": "gpt", "value": "the black vest"}]
00000002700.jpg
2700_evjvqa
[{"from": "human", "value": "where are the men wearing vest standing?\n<image>"}, {"from": "gpt", "value": "on the riverbank"}]
00000002700.jpg
2700_evjvqa
[{"from": "human", "value": "what color of the vest is the man pointing to the river wearing?\n<image>"}, {"from": "gpt", "value": "black"}]
00000002701.jpg
2701_evjvqa
[{"from": "human", "value": "where is the bicycle park?\n<image>"}, {"from": "gpt", "value": "on the side of the road"}]
00000002701.jpg
2701_evjvqa
[{"from": "human", "value": "what color are the chairs at this restaurant?\n<image>"}, {"from": "gpt", "value": "red"}]
00000002704.jpg
2704_evjvqa
[{"from": "human", "value": "what color of pants does the young man standing near the roadside wear?\n<image>"}, {"from": "gpt", "value": "white"}]
00000002704.jpg
2704_evjvqa
[{"from": "human", "value": "what is the young man in blue shirt sitting on the chair doing?\n<image>"}, {"from": "gpt", "value": "parking"}]
00000002704.jpg
2704_evjvqa
[{"from": "human", "value": "how many motorbikes are parked on the sidewalk?\n<image>"}, {"from": "gpt", "value": "three"}]
00000002706.jpg
2706_evjvqa
[{"from": "human", "value": "what is the peddler woman holding in her hand?\n<image>"}, {"from": "gpt", "value": "the leaf fan"}]
00000002706.jpg
2706_evjvqa
[{"from": "human", "value": "what is the peddler woman holding the leaf fan for?\n<image>"}, {"from": "gpt", "value": "to chase away the flies"}]
00000002706.jpg
2706_evjvqa
[{"from": "human", "value": "what is behind the peddler woman?\n<image>"}, {"from": "gpt", "value": "the old tree"}]
00000002707.jpg
2707_evjvqa
[{"from": "human", "value": "how many the trolleys are used in the supermarket?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002707.jpg
2707_evjvqa
[{"from": "human", "value": "what color t-shirt does the boy with the black mask wear?\n<image>"}, {"from": "gpt", "value": "orange"}]
00000002707.jpg
2707_evjvqa
[{"from": "human", "value": "what color t-shirt does the man without a mask wear?\n<image>"}, {"from": "gpt", "value": "white"}]
00000002716.jpg
2716_evjvqa
[{"from": "human", "value": "what is the woman wearing on her head?\n<image>"}, {"from": "gpt", "value": "the conical hat"}]
00000002716.jpg
2716_evjvqa
[{"from": "human", "value": "where is the conical hat?\n<image>"}, {"from": "gpt", "value": "on the woman head"}]
00000002716.jpg
2716_evjvqa
[{"from": "human", "value": "where is the woman wearing conical hat walking?\n<image>"}, {"from": "gpt", "value": "on the sidewalk"}]
00000002719.jpg
2719_evjvqa
[{"from": "human", "value": "what is the man in the black shirt with the blue mask pushing?\n<image>"}, {"from": "gpt", "value": "red"}]
00000002719.jpg
2719_evjvqa
[{"from": "human", "value": "what does the young man in black shirt do with a shopping cart?\n<image>"}, {"from": "gpt", "value": "waiting to pay"}]
00000002719.jpg
2719_evjvqa
[{"from": "human", "value": "what is drawn on the floor to indicate where the people in line stand for payment?\n<image>"}, {"from": "gpt", "value": "the red squares"}]
00000002721.jpg
2721_evjvqa
[{"from": "human", "value": "what color are the chairs that the street vendor uses for guests to sit on?\n<image>"}, {"from": "gpt", "value": "red and blue "}]
00000002721.jpg
2721_evjvqa
[{"from": "human", "value": "what is the biggest pot at this street vendor used for?\n<image>"}, {"from": "gpt", "value": "the broth"}]
00000002721.jpg
2721_evjvqa
[{"from": "human", "value": "where is the green scale?\n<image>"}, {"from": "gpt", "value": "on the back of the peddler woman"}]
00000002723.jpg
2723_evjvqa
[{"from": "human", "value": "what do the young men in white t-shirt wear on his wrists?\n<image>"}, {"from": "gpt", "value": "the watch "}]
00000002723.jpg
2723_evjvqa
[{"from": "human", "value": "where is the woman in red t-shirt standing?\n<image>"}, {"from": "gpt", "value": "between the young man in white t-shite and the woman in white dress"}]
00000002723.jpg
2723_evjvqa
[{"from": "human", "value": "what is the position of the young man in the white t-shirt compared to the woman in the red t-shirt?\n<image>"}, {"from": "gpt", "value": "on the right "}]
00000002725.jpg
2725_evjvqa
[{"from": "human", "value": "what is the woman in black t-shirt holding in her hand?\n<image>"}, {"from": "gpt", "value": "a bag"}]
00000002725.jpg
2725_evjvqa
[{"from": "human", "value": "what kind of store is this?\n<image>"}, {"from": "gpt", "value": "this is a grocery store"}]
00000002725.jpg
2725_evjvqa
[{"from": "human", "value": "who is holding the bag?\n<image>"}, {"from": "gpt", "value": "the woman in the black t-shirt"}]
00000002727.jpg
2727_evjvqa
[{"from": "human", "value": "what grows on both sides of the river?\n<image>"}, {"from": "gpt", "value": "the water coconut trees"}]
00000002727.jpg
2727_evjvqa
[{"from": "human", "value": "what means of transport do people use to move on the water?\n<image>"}, {"from": "gpt", "value": "the boats"}]
00000002727.jpg
2727_evjvqa
[{"from": "human", "value": "what kind of means of transport do people use to move on this river?\n<image>"}, {"from": "gpt", "value": "the boats"}]
00000002728.jpg
2728_evjvqa
[{"from": "human", "value": "where is the glass?\n<image>"}, {"from": "gpt", "value": "on the table, on the right side"}]
00000002728.jpg
2728_evjvqa
[{"from": "human", "value": "what is the girl sitting on?\n<image>"}, {"from": "gpt", "value": "the table"}]
00000002728.jpg
2728_evjvqa
[{"from": "human", "value": "what material are the tables and chairs at this restaurant made of?\n<image>"}, {"from": "gpt", "value": "the woods"}]
00000002730.jpg
2730_evjvqa
[{"from": "human", "value": "where is the girl in black standing?\n<image>"}, {"from": "gpt", "value": "behind the woman in white"}]
00000002730.jpg
2730_evjvqa
[{"from": "human", "value": "what color is the print on the girl's black t-shirt?\n<image>"}, {"from": "gpt", "value": "white"}]
00000002730.jpg
2730_evjvqa
[{"from": "human", "value": "what is the woman in the white shirt wearing?\n<image>"}, {"from": "gpt", "value": "she wears a handbag"}]
00000002733.jpg
2733_evjvqa
[{"from": "human", "value": "what is the woman in the long-sleeved overcoat doing?\n<image>"}, {"from": "gpt", "value": "choosing the products"}]
00000002733.jpg
2733_evjvqa
[{"from": "human", "value": "what does the woman in the blue coat use to contain the products she bought in the supermarket?\n<image>"}, {"from": "gpt", "value": "the trolley"}]
00000002733.jpg
2733_evjvqa
[{"from": "human", "value": "who is behind the woman in the blue overcoat?\n<image>"}, {"from": "gpt", "value": "the woman is wearing the purple t-shirt"}]
00000002734.jpg
2734_evjvqa
[{"from": "human", "value": "what is the tall man in the white shirt doing?\n<image>"}, {"from": "gpt", "value": "he is shopping with relatives"}]
00000002734.jpg
2734_evjvqa
[{"from": "human", "value": "what is the woman with the man in the white shirt planning to buy?\n<image>"}, {"from": "gpt", "value": "she is intend choosing the apples"}]
00000002734.jpg
2734_evjvqa
[{"from": "human", "value": "what kind of fruit was touched by the woman in the gray dress?\n<image>"}, {"from": "gpt", "value": "the apple"}]
00000002737.jpg
2737_evjvqa
[{"from": "human", "value": "what color is the car in front of the motorcycle rider wearing a blue helmet?\n<image>"}, {"from": "gpt", "value": "the black car"}]
00000002737.jpg
2737_evjvqa
[{"from": "human", "value": "which side is the black car going compared to the motorcycle driven by the person wearing the blue helmet?\n<image>"}, {"from": "gpt", "value": "the black car is going ahead"}]
00000002737.jpg
2737_evjvqa
[{"from": "human", "value": "where is the blue helmet?\n<image>"}, {"from": "gpt", "value": "it is worn on the head of the person riding the motorbike after the black car on the street"}]
00000002740.jpg
2740_evjvqa
[{"from": "human", "value": "what color t-shirt is the staff in the milk tea store wearing?\n<image>"}, {"from": "gpt", "value": "black"}]
00000002745.jpg
2745_evjvqa
[{"from": "human", "value": "where are the red flags with yellow stars hung?\n<image>"}, {"from": "gpt", "value": "on the two houses on the left side of the road"}]
00000002745.jpg
2745_evjvqa
[{"from": "human", "value": "how many people sit on the steps?\n<image>"}, {"from": "gpt", "value": "two"}]
00000002745.jpg
2745_evjvqa
[{"from": "human", "value": "on which side of the road are the people sitting on their chairs?\n<image>"}, {"from": "gpt", "value": "they sit on the sidewalk on the right of the road"}]
00000002746.jpg
2746_evjvqa
[{"from": "human", "value": "what color clothes is the woman in the blue shirt holding in her hand?\n<image>"}, {"from": "gpt", "value": "black"}]
00000002746.jpg
2746_evjvqa
[{"from": "human", "value": "what is the woman in black wearing on her waist?\n<image>"}, {"from": "gpt", "value": "the handbag"}]
00000002746.jpg
2746_evjvqa
[{"from": "human", "value": "what did the woman in black come to the store for?\n<image>"}, {"from": "gpt", "value": "to choose the clothes is made of silk fabric"}]
00000002754.jpg
2754_evjvqa
[{"from": "human", "value": "what occasion does the store sell decorations for?\n<image>"}, {"from": "gpt", "value": "the chirstmas"}]
00000002754.jpg
2754_evjvqa
[{"from": "human", "value": "who is standing next to the boy in white?\n<image>"}, {"from": "gpt", "value": "the girl is wearing the pink t-shirt"}]
00000002754.jpg
2754_evjvqa
[{"from": "human", "value": "what can people buy at this store?\n<image>"}, {"from": "gpt", "value": "the decorations for the chirstmas"}]
00000002762.jpg
2762_evjvqa
[{"from": "human", "value": "what is the girl in the white shirt leaning against the railing for?\n<image>"}, {"from": "gpt", "value": "to be photographed"}]
00000002762.jpg
2762_evjvqa
[{"from": "human", "value": "where does the girl in white stand to be photographed by her friend?\n<image>"}, {"from": "gpt", "value": "she is standing by the railing at the bar"}]
00000002762.jpg
2762_evjvqa
[{"from": "human", "value": "what is the girl in the gray shirt holding the phone for?\n<image>"}, {"from": "gpt", "value": "to take the photo "}]
00000002766.jpg
2766_evjvqa
[{"from": "human", "value": "what colors are the street name poles painted?\n<image>"}, {"from": "gpt", "value": "red and white"}]
00000002766.jpg
2766_evjvqa
[{"from": "human", "value": "what is the shape of the road sign?\n<image>"}, {"from": "gpt", "value": "it has rectangular shape, blue background and white text"}]
00000002766.jpg
2766_evjvqa
[{"from": "human", "value": "which side of the road is the motorcycle covered by a gray tarp erected on?\n<image>"}, {"from": "gpt", "value": "on the right road"}]
00000002767.jpg
2767_evjvqa
[{"from": "human", "value": "how many cars are on the road?\n<image>"}, {"from": "gpt", "value": "no one"}]
00000002767.jpg
2767_evjvqa
[{"from": "human", "value": "where is the motorbike parked?\n<image>"}, {"from": "gpt", "value": "next to the roadside light pole"}]
00000002767.jpg
2767_evjvqa
[{"from": "human", "value": "which side of the road are people coming to the park sitting on the benches?\n<image>"}, {"from": "gpt", "value": "on the left"}]
00000002768.jpg
2768_evjvqa
[{"from": "human", "value": "what kind of hat is the woman carrying the yoke wearing?\n<image>"}, {"from": "gpt", "value": "the conical hat"}]
00000002768.jpg
2768_evjvqa
[{"from": "human", "value": "what kind of hat does the woman carrying the pole on the road use?\n<image>"}, {"from": "gpt", "value": "the conical hat"}]
00000002768.jpg
2768_evjvqa
[{"from": "human", "value": "what vehicle is the man in white shirt with black pants use?\n<image>"}, {"from": "gpt", "value": "the bicycle"}]
00000002769.jpg
2769_evjvqa
[{"from": "human", "value": "what is the right below the statue?\n<image>"}, {"from": "gpt", "value": "there is a potted plant"}]
00000002769.jpg
2769_evjvqa
[{"from": "human", "value": "what colors is the building behind the statue painted in?\n<image>"}, {"from": "gpt", "value": "white and yellow"}]
00000002769.jpg
2769_evjvqa
[{"from": "human", "value": "which side of the statue is the building painted in yellow and white on?\n<image>"}, {"from": "gpt", "value": "behing the statue"}]