image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000006089.jpg
6089_evjvqa
[{"from": "human", "value": "what color are the cranes in the harbour?\n<image>"}, {"from": "gpt", "value": "the cranes are yellow"}]
00000006092.jpg
6092_evjvqa
[{"from": "human", "value": "what area is everyone visiting?\n<image>"}, {"from": "gpt", "value": "everyone is visiting a historical relic"}]
00000006092.jpg
6092_evjvqa
[{"from": "human", "value": "what color is the dress that the woman with the tattoo on her back is wearing?\n<image>"}, {"from": "gpt", "value": "the woman with the tattoo on her back is wearing a red dress"}]
00000006092.jpg
6092_evjvqa
[{"from": "human", "value": "what color is the shirt that the man standing next to the girl in black shirt is wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing white shirt"}]
00000005100.jpg
5100_evjvqa
[{"from": "human", "value": "what kind of upper garment is the man riding white bike wearing?\n<image>"}, {"from": "gpt", "value": "shirt"}]
00000005100.jpg
5100_evjvqa
[{"from": "human", "value": "how many motorbikes are running in the street?\n<image>"}, {"from": "gpt", "value": "there are six motorbikes running in the street"}]
00000005102.jpg
5102_evjvqa
[{"from": "human", "value": "what is the woman in black jacket selling?\n<image>"}, {"from": "gpt", "value": "selling flower"}]
00000005102.jpg
5102_evjvqa
[{"from": "human", "value": "what color coat is the woman holding white bouquet wearing?\n<image>"}, {"from": "gpt", "value": "pink"}]
00000005102.jpg
5102_evjvqa
[{"from": "human", "value": "what is the woman in pink coat doing with her hand?\n<image>"}, {"from": "gpt", "value": "holding a bouquet"}]
00000005103.jpg
5103_evjvqa
[{"from": "human", "value": "where are the people sitting?\n<image>"}, {"from": "gpt", "value": "sitting in canoes"}]
00000005103.jpg
5103_evjvqa
[{"from": "human", "value": "what kind of hat is the girl in yellow shirt wearing?\n<image>"}, {"from": "gpt", "value": "wearing non la"}]
00000005103.jpg
5103_evjvqa
[{"from": "human", "value": "what is the girl in yellow shirt doing?\n<image>"}, {"from": "gpt", "value": "paddling"}]
00000005105.jpg
5105_evjvqa
[{"from": "human", "value": "where is the girl standing?\n<image>"}, {"from": "gpt", "value": "on the beach"}]
00000005105.jpg
5105_evjvqa
[{"from": "human", "value": "what color hat is the girl wearing?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005105.jpg
5105_evjvqa
[{"from": "human", "value": "what color dress is the girl wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
00000005106.jpg
5106_evjvqa
[{"from": "human", "value": "what color is the bittermelon?\n<image>"}, {"from": "gpt", "value": "green"}]
00000005106.jpg
5106_evjvqa
[{"from": "human", "value": "what color is the bittermelon box?\n<image>"}, {"from": "gpt", "value": "the bittermelon box is black"}]
00000005107.jpg
5107_evjvqa
[{"from": "human", "value": "what color is the gate of the temple?\n<image>"}, {"from": "gpt", "value": "it is green"}]
00000005107.jpg
5107_evjvqa
[{"from": "human", "value": "how many entries to the temple are there?\n<image>"}, {"from": "gpt", "value": "there are three entries"}]
00000005107.jpg
5107_evjvqa
[{"from": "human", "value": "what color is the letters carved on the two pillars of the temple gate?\n<image>"}, {"from": "gpt", "value": "black"}]
00000005108.jpg
5108_evjvqa
[{"from": "human", "value": "what color shirt is the man standing on the far left wearing?\n<image>"}, {"from": "gpt", "value": "white shirt"}]
00000005108.jpg
5108_evjvqa
[{"from": "human", "value": "what is the emotion of the boy in yellow shirt?\n<image>"}, {"from": "gpt", "value": "the boy is smiling"}]
00000005108.jpg
5108_evjvqa
[{"from": "human", "value": "how many boys in black shirt are there?\n<image>"}, {"from": "gpt", "value": "there is one boy in black shirt"}]
00000005109.jpg
5109_evjvqa
[{"from": "human", "value": "what color is the fruit on the tree?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000005109.jpg
5109_evjvqa
[{"from": "human", "value": "what color is the can hanged on the tree?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000005109.jpg
5109_evjvqa
[{"from": "human", "value": "how many cans are hanged on the tree?\n<image>"}, {"from": "gpt", "value": "there is one can"}]
00000005110.jpg
5110_evjvqa
[{"from": "human", "value": "how many women wearing skirt are there?\n<image>"}, {"from": "gpt", "value": "there is one woman wearing skirt"}]
00000005112.jpg
5112_evjvqa
[{"from": "human", "value": "what color vest is the woman wearing?\n<image>"}, {"from": "gpt", "value": "black"}]
00000005116.jpg
5116_evjvqa
[{"from": "human", "value": "what is the man in the stall doing?\n<image>"}, {"from": "gpt", "value": "sitting"}]
00000005116.jpg
5116_evjvqa
[{"from": "human", "value": "what color bag is the girl in white shirt holding?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000005117.jpg
5117_evjvqa
[{"from": "human", "value": "what is the girl doing?\n<image>"}, {"from": "gpt", "value": "sitting next to the flowerbed"}]
00000005117.jpg
5117_evjvqa
[{"from": "human", "value": "what kind of dress is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a traditional ao dai"}]
00000005117.jpg
5117_evjvqa
[{"from": "human", "value": "what is the main color of the ao dai of the girl?\n<image>"}, {"from": "gpt", "value": "red"}]
00000005118.jpg
5118_evjvqa
[{"from": "human", "value": "how many men are standing there talking with each other?\n<image>"}, {"from": "gpt", "value": "there are three men"}]
00000005118.jpg
5118_evjvqa
[{"from": "human", "value": "what is the girl in black dress doing?\n<image>"}, {"from": "gpt", "value": "talking with other men"}]
00000005118.jpg
5118_evjvqa
[{"from": "human", "value": "what is the man wearing blue mask doing?\n<image>"}, {"from": "gpt", "value": "holding a phone"}]
00000005119.jpg
5119_evjvqa
[{"from": "human", "value": "where is this place?\n<image>"}, {"from": "gpt", "value": "this is a street"}]
00000005119.jpg
5119_evjvqa
[{"from": "human", "value": "what vehicle is the girl in light blue jacket riding?\n<image>"}, {"from": "gpt", "value": "riding a bike"}]
00000005119.jpg
5119_evjvqa
[{"from": "human", "value": "what color helmet is the boy sitting on the white bike wearing?\n<image>"}, {"from": "gpt", "value": "black"}]
00000005120.jpg
5120_evjvqa
[{"from": "human", "value": "where is the girl sitting?\n<image>"}, {"from": "gpt", "value": "the girl is sitting on the roof"}]
00000005120.jpg
5120_evjvqa
[{"from": "human", "value": "what is the girl holding?\n<image>"}, {"from": "gpt", "value": "the girl is holding a camera"}]
00000005121.jpg
5121_evjvqa
[{"from": "human", "value": "what is the security guard doing?\n<image>"}, {"from": "gpt", "value": "spraying disinfectant for the customer"}]
00000005122.jpg
5122_evjvqa
[{"from": "human", "value": "what color shirt is the woman wearing?\n<image>"}, {"from": "gpt", "value": "blue shirt"}]
00000005122.jpg
5122_evjvqa
[{"from": "human", "value": "what is the woman cooking?\n<image>"}, {"from": "gpt", "value": "the women is cooking vermicelli"}]
00000005122.jpg
5122_evjvqa
[{"from": "human", "value": "how many seasoning jars are on the kitchen of the woman?\n<image>"}, {"from": "gpt", "value": "there are three seasoning jars"}]
00000005128.jpg
5128_evjvqa
[{"from": "human", "value": "what color dress is the girl with eyeglasses wearing?\n<image>"}, {"from": "gpt", "value": "red dress"}]
00000005128.jpg
5128_evjvqa
[{"from": "human", "value": "how many girls are there?\n<image>"}, {"from": "gpt", "value": "there are four girls"}]
00000005128.jpg
5128_evjvqa
[{"from": "human", "value": "what color pants is the girl in black shirt wearing?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005129.jpg
5129_evjvqa
[{"from": "human", "value": "what color is the flag hoisted on the roof of the ship?\n<image>"}, {"from": "gpt", "value": "red"}]
00000005132.jpg
5132_evjvqa
[{"from": "human", "value": "what time of the day is it?\n<image>"}, {"from": "gpt", "value": "evening"}]
00000005132.jpg
5132_evjvqa
[{"from": "human", "value": "what vehicle are most of the people here driving?\n<image>"}, {"from": "gpt", "value": "riding motorcycles"}]
00000005133.jpg
5133_evjvqa
[{"from": "human", "value": "how many statues are there?\n<image>"}, {"from": "gpt", "value": "there are three statues"}]
00000005133.jpg
5133_evjvqa
[{"from": "human", "value": "what color are the three statues?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000005133.jpg
5133_evjvqa
[{"from": "human", "value": "what is the man in black suit doing?\n<image>"}, {"from": "gpt", "value": "holding a mic to give a speech"}]
00000005139.jpg
5139_evjvqa
[{"from": "human", "value": "what color face mask is the woman riding motorbike using?\n<image>"}, {"from": "gpt", "value": "the woman riding motorbike is using green face mask"}]
00000005139.jpg
5139_evjvqa
[{"from": "human", "value": "what color suit is the man standing wearing?\n<image>"}, {"from": "gpt", "value": "the man standing is wearing blue suit"}]
00000005139.jpg
5139_evjvqa
[{"from": "human", "value": "what color chair is the boy sitting on?\n<image>"}, {"from": "gpt", "value": "blue chair"}]
00000005143.jpg
5143_evjvqa
[{"from": "human", "value": "what color is the backpack of the girl?\n<image>"}, {"from": "gpt", "value": "brown"}]
00000005143.jpg
5143_evjvqa
[{"from": "human", "value": "what is the main color of the dress of the girl?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005148.jpg
5148_evjvqa
[{"from": "human", "value": "what is the background color of the signboard?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000005153.jpg
5153_evjvqa
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "using the cellphone"}]
00000005153.jpg
5153_evjvqa
[{"from": "human", "value": "what type of product is the man selling?\n<image>"}, {"from": "gpt", "value": "the man is selling fruit"}]
00000005153.jpg
5153_evjvqa
[{"from": "human", "value": "what are the fruit trays made of?\n<image>"}, {"from": "gpt", "value": "the fruit trays are made of paper"}]
00000005156.jpg
5156_evjvqa
[{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "the woman is buying vegetable"}]
00000005156.jpg
5156_evjvqa
[{"from": "human", "value": "what color shirt is the man standing next to the woman in pink shirt wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing white shirt"}]
00000005156.jpg
5156_evjvqa
[{"from": "human", "value": "what color shirt is the woman selecting vegetable wearing?\n<image>"}, {"from": "gpt", "value": "pink shirt"}]
00000005157.jpg
5157_evjvqa
[{"from": "human", "value": "what sport is the man playing?\n<image>"}, {"from": "gpt", "value": "kitesurfing"}]
00000005157.jpg
5157_evjvqa
[{"from": "human", "value": "what color is the shirt of the man?\n<image>"}, {"from": "gpt", "value": "red"}]
00000005157.jpg
5157_evjvqa
[{"from": "human", "value": "what color is the kite of the man?\n<image>"}, {"from": "gpt", "value": "blue"}]
00000005159.jpg
5159_evjvqa
[{"from": "human", "value": "how many men are sitting on the cyclos?\n<image>"}, {"from": "gpt", "value": "there are two men sitting on the cyclos"}]
00000005159.jpg
5159_evjvqa
[{"from": "human", "value": "what is the woman in blue shirt doing?\n<image>"}, {"from": "gpt", "value": "riding a bicycle"}]
00000005159.jpg
5159_evjvqa
[{"from": "human", "value": "how many cyclos are there?\n<image>"}, {"from": "gpt", "value": "there are two cyclos"}]
00000005163.jpg
5163_evjvqa
[{"from": "human", "value": "what is the main color of the wall of the house?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000005163.jpg
5163_evjvqa
[{"from": "human", "value": "what color heels is the girl wearing?\n<image>"}, {"from": "gpt", "value": "beige"}]
00000005165.jpg
5165_evjvqa
[{"from": "human", "value": "what color is the dress of the bride?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005167.jpg
5167_evjvqa
[{"from": "human", "value": "where is this place?\n<image>"}, {"from": "gpt", "value": "this is a café"}]
00000005167.jpg
5167_evjvqa
[{"from": "human", "value": "what color are the door frames painted?\n<image>"}, {"from": "gpt", "value": "brown"}]
00000005169.jpg
5169_evjvqa
[{"from": "human", "value": "what is the woman shopping for?\n<image>"}, {"from": "gpt", "value": "the woman is shopping for avocados"}]
00000005169.jpg
5169_evjvqa
[{"from": "human", "value": "what color face mask is the woman wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
00000005175.jpg
5175_evjvqa
[{"from": "human", "value": "where is this place?\n<image>"}, {"from": "gpt", "value": "this is a clothing store"}]
00000005175.jpg
5175_evjvqa
[{"from": "human", "value": "what color backpack is the woman in white shirt wearing?\n<image>"}, {"from": "gpt", "value": "black backpack"}]
00000005177.jpg
5177_evjvqa
[{"from": "human", "value": "what color is the suit of the youth?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005177.jpg
5177_evjvqa
[{"from": "human", "value": "what color are the sandals of the youth?\n<image>"}, {"from": "gpt", "value": "red"}]
00000005177.jpg
5177_evjvqa
[{"from": "human", "value": "what color is the hat of the youth?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005180.jpg
5180_evjvqa
[{"from": "human", "value": "what color backpack is the woman in light green jacket wearing?\n<image>"}, {"from": "gpt", "value": "red backpack"}]
00000005180.jpg
5180_evjvqa
[{"from": "human", "value": "how many women are there?\n<image>"}, {"from": "gpt", "value": "there are two women"}]
00000005180.jpg
5180_evjvqa
[{"from": "human", "value": "what color helmet is the woman with red backpack wearing?\n<image>"}, {"from": "gpt", "value": "dark red helmet"}]
00000005186.jpg
5186_evjvqa
[{"from": "human", "value": "what color helmet is the man sitting on the bench having?\n<image>"}, {"from": "gpt", "value": "yellow helmet"}]
00000005187.jpg
5187_evjvqa
[{"from": "human", "value": "how many entries to the festival site are there?\n<image>"}, {"from": "gpt", "value": "there are three entries"}]
00000005188.jpg
5188_evjvqa
[{"from": "human", "value": "how many cars are there?\n<image>"}, {"from": "gpt", "value": "there is one car"}]
00000005188.jpg
5188_evjvqa
[{"from": "human", "value": "what color is the car?\n<image>"}, {"from": "gpt", "value": "the car is white"}]
00000005189.jpg
5189_evjvqa
[{"from": "human", "value": "how many rice trucks are there?\n<image>"}, {"from": "gpt", "value": "there is one rice truck"}]
00000005190.jpg
5190_evjvqa
[{"from": "human", "value": "what is the main color of the wall of the store?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005190.jpg
5190_evjvqa
[{"from": "human", "value": "what color is the floor of the store?\n<image>"}, {"from": "gpt", "value": "brown"}]
00000005191.jpg
5191_evjvqa
[{"from": "human", "value": "what is everyone selling?\n<image>"}, {"from": "gpt", "value": "selling vegetable"}]
00000005191.jpg
5191_evjvqa
[{"from": "human", "value": "how many men are sitting on the motorcycle?\n<image>"}, {"from": "gpt", "value": "there is one man sitting on the motorcycle"}]
00000005191.jpg
5191_evjvqa
[{"from": "human", "value": "what is the man in black suit doing?\n<image>"}, {"from": "gpt", "value": "sitting to sell vegetable"}]
00000005192.jpg
5192_evjvqa
[{"from": "human", "value": "how many youths are standing on the car?\n<image>"}, {"from": "gpt", "value": "there is one youth"}]
00000005196.jpg
5196_evjvqa
[{"from": "human", "value": "how many motorcycles are parked in front of the café?\n<image>"}, {"from": "gpt", "value": "there are four motorbikes"}]
00000005198.jpg
5198_evjvqa
[{"from": "human", "value": "what color bag is the girl in black shirt holding?\n<image>"}, {"from": "gpt", "value": "white"}]