image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000006089.jpg
|
6089_evjvqa
|
[{"from": "human", "value": "what color are the cranes in the harbour?\n<image>"}, {"from": "gpt", "value": "the cranes are yellow"}]
|
00000006092.jpg
|
6092_evjvqa
|
[{"from": "human", "value": "what area is everyone visiting?\n<image>"}, {"from": "gpt", "value": "everyone is visiting a historical relic"}]
|
00000006092.jpg
|
6092_evjvqa
|
[{"from": "human", "value": "what color is the dress that the woman with the tattoo on her back is wearing?\n<image>"}, {"from": "gpt", "value": "the woman with the tattoo on her back is wearing a red dress"}]
|
00000006092.jpg
|
6092_evjvqa
|
[{"from": "human", "value": "what color is the shirt that the man standing next to the girl in black shirt is wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing white shirt"}]
|
00000005100.jpg
|
5100_evjvqa
|
[{"from": "human", "value": "what kind of upper garment is the man riding white bike wearing?\n<image>"}, {"from": "gpt", "value": "shirt"}]
|
00000005100.jpg
|
5100_evjvqa
|
[{"from": "human", "value": "how many motorbikes are running in the street?\n<image>"}, {"from": "gpt", "value": "there are six motorbikes running in the street"}]
|
00000005102.jpg
|
5102_evjvqa
|
[{"from": "human", "value": "what is the woman in black jacket selling?\n<image>"}, {"from": "gpt", "value": "selling flower"}]
|
00000005102.jpg
|
5102_evjvqa
|
[{"from": "human", "value": "what color coat is the woman holding white bouquet wearing?\n<image>"}, {"from": "gpt", "value": "pink"}]
|
00000005102.jpg
|
5102_evjvqa
|
[{"from": "human", "value": "what is the woman in pink coat doing with her hand?\n<image>"}, {"from": "gpt", "value": "holding a bouquet"}]
|
00000005103.jpg
|
5103_evjvqa
|
[{"from": "human", "value": "where are the people sitting?\n<image>"}, {"from": "gpt", "value": "sitting in canoes"}]
|
00000005103.jpg
|
5103_evjvqa
|
[{"from": "human", "value": "what kind of hat is the girl in yellow shirt wearing?\n<image>"}, {"from": "gpt", "value": "wearing non la"}]
|
00000005103.jpg
|
5103_evjvqa
|
[{"from": "human", "value": "what is the girl in yellow shirt doing?\n<image>"}, {"from": "gpt", "value": "paddling"}]
|
00000005105.jpg
|
5105_evjvqa
|
[{"from": "human", "value": "where is the girl standing?\n<image>"}, {"from": "gpt", "value": "on the beach"}]
|
00000005105.jpg
|
5105_evjvqa
|
[{"from": "human", "value": "what color hat is the girl wearing?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000005105.jpg
|
5105_evjvqa
|
[{"from": "human", "value": "what color dress is the girl wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
|
00000005106.jpg
|
5106_evjvqa
|
[{"from": "human", "value": "what color is the bittermelon?\n<image>"}, {"from": "gpt", "value": "green"}]
|
00000005106.jpg
|
5106_evjvqa
|
[{"from": "human", "value": "what color is the bittermelon box?\n<image>"}, {"from": "gpt", "value": "the bittermelon box is black"}]
|
00000005107.jpg
|
5107_evjvqa
|
[{"from": "human", "value": "what color is the gate of the temple?\n<image>"}, {"from": "gpt", "value": "it is green"}]
|
00000005107.jpg
|
5107_evjvqa
|
[{"from": "human", "value": "how many entries to the temple are there?\n<image>"}, {"from": "gpt", "value": "there are three entries"}]
|
00000005107.jpg
|
5107_evjvqa
|
[{"from": "human", "value": "what color is the letters carved on the two pillars of the temple gate?\n<image>"}, {"from": "gpt", "value": "black"}]
|
00000005108.jpg
|
5108_evjvqa
|
[{"from": "human", "value": "what color shirt is the man standing on the far left wearing?\n<image>"}, {"from": "gpt", "value": "white shirt"}]
|
00000005108.jpg
|
5108_evjvqa
|
[{"from": "human", "value": "what is the emotion of the boy in yellow shirt?\n<image>"}, {"from": "gpt", "value": "the boy is smiling"}]
|
00000005108.jpg
|
5108_evjvqa
|
[{"from": "human", "value": "how many boys in black shirt are there?\n<image>"}, {"from": "gpt", "value": "there is one boy in black shirt"}]
|
00000005109.jpg
|
5109_evjvqa
|
[{"from": "human", "value": "what color is the fruit on the tree?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000005109.jpg
|
5109_evjvqa
|
[{"from": "human", "value": "what color is the can hanged on the tree?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000005109.jpg
|
5109_evjvqa
|
[{"from": "human", "value": "how many cans are hanged on the tree?\n<image>"}, {"from": "gpt", "value": "there is one can"}]
|
00000005110.jpg
|
5110_evjvqa
|
[{"from": "human", "value": "how many women wearing skirt are there?\n<image>"}, {"from": "gpt", "value": "there is one woman wearing skirt"}]
|
00000005112.jpg
|
5112_evjvqa
|
[{"from": "human", "value": "what color vest is the woman wearing?\n<image>"}, {"from": "gpt", "value": "black"}]
|
00000005116.jpg
|
5116_evjvqa
|
[{"from": "human", "value": "what is the man in the stall doing?\n<image>"}, {"from": "gpt", "value": "sitting"}]
|
00000005116.jpg
|
5116_evjvqa
|
[{"from": "human", "value": "what color bag is the girl in white shirt holding?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000005117.jpg
|
5117_evjvqa
|
[{"from": "human", "value": "what is the girl doing?\n<image>"}, {"from": "gpt", "value": "sitting next to the flowerbed"}]
|
00000005117.jpg
|
5117_evjvqa
|
[{"from": "human", "value": "what kind of dress is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a traditional ao dai"}]
|
00000005117.jpg
|
5117_evjvqa
|
[{"from": "human", "value": "what is the main color of the ao dai of the girl?\n<image>"}, {"from": "gpt", "value": "red"}]
|
00000005118.jpg
|
5118_evjvqa
|
[{"from": "human", "value": "how many men are standing there talking with each other?\n<image>"}, {"from": "gpt", "value": "there are three men"}]
|
00000005118.jpg
|
5118_evjvqa
|
[{"from": "human", "value": "what is the girl in black dress doing?\n<image>"}, {"from": "gpt", "value": "talking with other men"}]
|
00000005118.jpg
|
5118_evjvqa
|
[{"from": "human", "value": "what is the man wearing blue mask doing?\n<image>"}, {"from": "gpt", "value": "holding a phone"}]
|
00000005119.jpg
|
5119_evjvqa
|
[{"from": "human", "value": "where is this place?\n<image>"}, {"from": "gpt", "value": "this is a street"}]
|
00000005119.jpg
|
5119_evjvqa
|
[{"from": "human", "value": "what vehicle is the girl in light blue jacket riding?\n<image>"}, {"from": "gpt", "value": "riding a bike"}]
|
00000005119.jpg
|
5119_evjvqa
|
[{"from": "human", "value": "what color helmet is the boy sitting on the white bike wearing?\n<image>"}, {"from": "gpt", "value": "black"}]
|
00000005120.jpg
|
5120_evjvqa
|
[{"from": "human", "value": "where is the girl sitting?\n<image>"}, {"from": "gpt", "value": "the girl is sitting on the roof"}]
|
00000005120.jpg
|
5120_evjvqa
|
[{"from": "human", "value": "what is the girl holding?\n<image>"}, {"from": "gpt", "value": "the girl is holding a camera"}]
|
00000005121.jpg
|
5121_evjvqa
|
[{"from": "human", "value": "what is the security guard doing?\n<image>"}, {"from": "gpt", "value": "spraying disinfectant for the customer"}]
|
00000005122.jpg
|
5122_evjvqa
|
[{"from": "human", "value": "what color shirt is the woman wearing?\n<image>"}, {"from": "gpt", "value": "blue shirt"}]
|
00000005122.jpg
|
5122_evjvqa
|
[{"from": "human", "value": "what is the woman cooking?\n<image>"}, {"from": "gpt", "value": "the women is cooking vermicelli"}]
|
00000005122.jpg
|
5122_evjvqa
|
[{"from": "human", "value": "how many seasoning jars are on the kitchen of the woman?\n<image>"}, {"from": "gpt", "value": "there are three seasoning jars"}]
|
00000005128.jpg
|
5128_evjvqa
|
[{"from": "human", "value": "what color dress is the girl with eyeglasses wearing?\n<image>"}, {"from": "gpt", "value": "red dress"}]
|
00000005128.jpg
|
5128_evjvqa
|
[{"from": "human", "value": "how many girls are there?\n<image>"}, {"from": "gpt", "value": "there are four girls"}]
|
00000005128.jpg
|
5128_evjvqa
|
[{"from": "human", "value": "what color pants is the girl in black shirt wearing?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000005129.jpg
|
5129_evjvqa
|
[{"from": "human", "value": "what color is the flag hoisted on the roof of the ship?\n<image>"}, {"from": "gpt", "value": "red"}]
|
00000005132.jpg
|
5132_evjvqa
|
[{"from": "human", "value": "what time of the day is it?\n<image>"}, {"from": "gpt", "value": "evening"}]
|
00000005132.jpg
|
5132_evjvqa
|
[{"from": "human", "value": "what vehicle are most of the people here driving?\n<image>"}, {"from": "gpt", "value": "riding motorcycles"}]
|
00000005133.jpg
|
5133_evjvqa
|
[{"from": "human", "value": "how many statues are there?\n<image>"}, {"from": "gpt", "value": "there are three statues"}]
|
00000005133.jpg
|
5133_evjvqa
|
[{"from": "human", "value": "what color are the three statues?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000005133.jpg
|
5133_evjvqa
|
[{"from": "human", "value": "what is the man in black suit doing?\n<image>"}, {"from": "gpt", "value": "holding a mic to give a speech"}]
|
00000005139.jpg
|
5139_evjvqa
|
[{"from": "human", "value": "what color face mask is the woman riding motorbike using?\n<image>"}, {"from": "gpt", "value": "the woman riding motorbike is using green face mask"}]
|
00000005139.jpg
|
5139_evjvqa
|
[{"from": "human", "value": "what color suit is the man standing wearing?\n<image>"}, {"from": "gpt", "value": "the man standing is wearing blue suit"}]
|
00000005139.jpg
|
5139_evjvqa
|
[{"from": "human", "value": "what color chair is the boy sitting on?\n<image>"}, {"from": "gpt", "value": "blue chair"}]
|
00000005143.jpg
|
5143_evjvqa
|
[{"from": "human", "value": "what color is the backpack of the girl?\n<image>"}, {"from": "gpt", "value": "brown"}]
|
00000005143.jpg
|
5143_evjvqa
|
[{"from": "human", "value": "what is the main color of the dress of the girl?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000005148.jpg
|
5148_evjvqa
|
[{"from": "human", "value": "what is the background color of the signboard?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000005153.jpg
|
5153_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "using the cellphone"}]
|
00000005153.jpg
|
5153_evjvqa
|
[{"from": "human", "value": "what type of product is the man selling?\n<image>"}, {"from": "gpt", "value": "the man is selling fruit"}]
|
00000005153.jpg
|
5153_evjvqa
|
[{"from": "human", "value": "what are the fruit trays made of?\n<image>"}, {"from": "gpt", "value": "the fruit trays are made of paper"}]
|
00000005156.jpg
|
5156_evjvqa
|
[{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "the woman is buying vegetable"}]
|
00000005156.jpg
|
5156_evjvqa
|
[{"from": "human", "value": "what color shirt is the man standing next to the woman in pink shirt wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing white shirt"}]
|
00000005156.jpg
|
5156_evjvqa
|
[{"from": "human", "value": "what color shirt is the woman selecting vegetable wearing?\n<image>"}, {"from": "gpt", "value": "pink shirt"}]
|
00000005157.jpg
|
5157_evjvqa
|
[{"from": "human", "value": "what sport is the man playing?\n<image>"}, {"from": "gpt", "value": "kitesurfing"}]
|
00000005157.jpg
|
5157_evjvqa
|
[{"from": "human", "value": "what color is the shirt of the man?\n<image>"}, {"from": "gpt", "value": "red"}]
|
00000005157.jpg
|
5157_evjvqa
|
[{"from": "human", "value": "what color is the kite of the man?\n<image>"}, {"from": "gpt", "value": "blue"}]
|
00000005159.jpg
|
5159_evjvqa
|
[{"from": "human", "value": "how many men are sitting on the cyclos?\n<image>"}, {"from": "gpt", "value": "there are two men sitting on the cyclos"}]
|
00000005159.jpg
|
5159_evjvqa
|
[{"from": "human", "value": "what is the woman in blue shirt doing?\n<image>"}, {"from": "gpt", "value": "riding a bicycle"}]
|
00000005159.jpg
|
5159_evjvqa
|
[{"from": "human", "value": "how many cyclos are there?\n<image>"}, {"from": "gpt", "value": "there are two cyclos"}]
|
00000005163.jpg
|
5163_evjvqa
|
[{"from": "human", "value": "what is the main color of the wall of the house?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000005163.jpg
|
5163_evjvqa
|
[{"from": "human", "value": "what color heels is the girl wearing?\n<image>"}, {"from": "gpt", "value": "beige"}]
|
00000005165.jpg
|
5165_evjvqa
|
[{"from": "human", "value": "what color is the dress of the bride?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000005167.jpg
|
5167_evjvqa
|
[{"from": "human", "value": "where is this place?\n<image>"}, {"from": "gpt", "value": "this is a café"}]
|
00000005167.jpg
|
5167_evjvqa
|
[{"from": "human", "value": "what color are the door frames painted?\n<image>"}, {"from": "gpt", "value": "brown"}]
|
00000005169.jpg
|
5169_evjvqa
|
[{"from": "human", "value": "what is the woman shopping for?\n<image>"}, {"from": "gpt", "value": "the woman is shopping for avocados"}]
|
00000005169.jpg
|
5169_evjvqa
|
[{"from": "human", "value": "what color face mask is the woman wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
|
00000005175.jpg
|
5175_evjvqa
|
[{"from": "human", "value": "where is this place?\n<image>"}, {"from": "gpt", "value": "this is a clothing store"}]
|
00000005175.jpg
|
5175_evjvqa
|
[{"from": "human", "value": "what color backpack is the woman in white shirt wearing?\n<image>"}, {"from": "gpt", "value": "black backpack"}]
|
00000005177.jpg
|
5177_evjvqa
|
[{"from": "human", "value": "what color is the suit of the youth?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000005177.jpg
|
5177_evjvqa
|
[{"from": "human", "value": "what color are the sandals of the youth?\n<image>"}, {"from": "gpt", "value": "red"}]
|
00000005177.jpg
|
5177_evjvqa
|
[{"from": "human", "value": "what color is the hat of the youth?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000005180.jpg
|
5180_evjvqa
|
[{"from": "human", "value": "what color backpack is the woman in light green jacket wearing?\n<image>"}, {"from": "gpt", "value": "red backpack"}]
|
00000005180.jpg
|
5180_evjvqa
|
[{"from": "human", "value": "how many women are there?\n<image>"}, {"from": "gpt", "value": "there are two women"}]
|
00000005180.jpg
|
5180_evjvqa
|
[{"from": "human", "value": "what color helmet is the woman with red backpack wearing?\n<image>"}, {"from": "gpt", "value": "dark red helmet"}]
|
00000005186.jpg
|
5186_evjvqa
|
[{"from": "human", "value": "what color helmet is the man sitting on the bench having?\n<image>"}, {"from": "gpt", "value": "yellow helmet"}]
|
00000005187.jpg
|
5187_evjvqa
|
[{"from": "human", "value": "how many entries to the festival site are there?\n<image>"}, {"from": "gpt", "value": "there are three entries"}]
|
00000005188.jpg
|
5188_evjvqa
|
[{"from": "human", "value": "how many cars are there?\n<image>"}, {"from": "gpt", "value": "there is one car"}]
|
00000005188.jpg
|
5188_evjvqa
|
[{"from": "human", "value": "what color is the car?\n<image>"}, {"from": "gpt", "value": "the car is white"}]
|
00000005189.jpg
|
5189_evjvqa
|
[{"from": "human", "value": "how many rice trucks are there?\n<image>"}, {"from": "gpt", "value": "there is one rice truck"}]
|
00000005190.jpg
|
5190_evjvqa
|
[{"from": "human", "value": "what is the main color of the wall of the store?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000005190.jpg
|
5190_evjvqa
|
[{"from": "human", "value": "what color is the floor of the store?\n<image>"}, {"from": "gpt", "value": "brown"}]
|
00000005191.jpg
|
5191_evjvqa
|
[{"from": "human", "value": "what is everyone selling?\n<image>"}, {"from": "gpt", "value": "selling vegetable"}]
|
00000005191.jpg
|
5191_evjvqa
|
[{"from": "human", "value": "how many men are sitting on the motorcycle?\n<image>"}, {"from": "gpt", "value": "there is one man sitting on the motorcycle"}]
|
00000005191.jpg
|
5191_evjvqa
|
[{"from": "human", "value": "what is the man in black suit doing?\n<image>"}, {"from": "gpt", "value": "sitting to sell vegetable"}]
|
00000005192.jpg
|
5192_evjvqa
|
[{"from": "human", "value": "how many youths are standing on the car?\n<image>"}, {"from": "gpt", "value": "there is one youth"}]
|
00000005196.jpg
|
5196_evjvqa
|
[{"from": "human", "value": "how many motorcycles are parked in front of the café?\n<image>"}, {"from": "gpt", "value": "there are four motorbikes"}]
|
00000005198.jpg
|
5198_evjvqa
|
[{"from": "human", "value": "what color bag is the girl in black shirt holding?\n<image>"}, {"from": "gpt", "value": "white"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.