image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000003943.jpg
|
3943_evjvqa
|
[{"from": "human", "value": "what is the woman in black shirt wearing on her waist?\n<image>"}, {"from": "gpt", "value": "she is wearing a handbag"}]
|
00000003943.jpg
|
3943_evjvqa
|
[{"from": "human", "value": "who stands on the right hand side of the woman in black shirt?\n<image>"}, {"from": "gpt", "value": "the man in white t-shirt stands on the right hand side of the woman in black shirt"}]
|
00000003943.jpg
|
3943_evjvqa
|
[{"from": "human", "value": "what are the man in white t-shirt and the woman next to him doing?\n<image>"}, {"from": "gpt", "value": "they are looking at the items on the shelves"}]
|
00000003945.jpg
|
3945_evjvqa
|
[{"from": "human", "value": "what is the woman in the pink coat doing?\n<image>"}, {"from": "gpt", "value": "the woman in the pink coat is calculating money by calculator"}]
|
00000003945.jpg
|
3945_evjvqa
|
[{"from": "human", "value": "what device is the woman in pink coat holding in her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a calculator"}]
|
00000003945.jpg
|
3945_evjvqa
|
[{"from": "human", "value": "who is holding the calculator to calculate?\n<image>"}, {"from": "gpt", "value": "the woman in pink coat"}]
|
00000003948.jpg
|
3948_evjvqa
|
[{"from": "human", "value": "what pose is the man in the lime green shirt in?\n<image>"}, {"from": "gpt", "value": "he is in the sitting pose"}]
|
00000003948.jpg
|
3948_evjvqa
|
[{"from": "human", "value": "how many people wear green military uniforms?\n<image>"}, {"from": "gpt", "value": "two people wear green military uniforms"}]
|
00000003948.jpg
|
3948_evjvqa
|
[{"from": "human", "value": "what are these people doing?\n<image>"}, {"from": "gpt", "value": "they are raising the flag"}]
|
00000003953.jpg
|
3953_evjvqa
|
[{"from": "human", "value": "what does the man wear?\n<image>"}, {"from": "gpt", "value": "he wears a santa claus costume"}]
|
00000003953.jpg
|
3953_evjvqa
|
[{"from": "human", "value": "what store is the man in the santa claus costume standing at?\n<image>"}, {"from": "gpt", "value": "he is standing in the toy store"}]
|
00000003953.jpg
|
3953_evjvqa
|
[{"from": "human", "value": "what color is the backpack that the person standing near the green plastic chair wears?\n<image>"}, {"from": "gpt", "value": "that person wears a black backpack"}]
|
00000003955.jpg
|
3955_evjvqa
|
[{"from": "human", "value": "what vehicle is the man in purple shirt sitting on?\n<image>"}, {"from": "gpt", "value": "he is sitting on a motorbike"}]
|
00000003957.jpg
|
3957_evjvqa
|
[{"from": "human", "value": "what color is the outer surface of this ship painted?\n<image>"}, {"from": "gpt", "value": "the outer surface of this ship is painted white and orange"}]
|
00000003957.jpg
|
3957_evjvqa
|
[{"from": "human", "value": "what two colors do people paint on the outer surface of this ship?\n<image>"}, {"from": "gpt", "value": "orange and white"}]
|
00000003957.jpg
|
3957_evjvqa
|
[{"from": "human", "value": "what means of transportation are behind these people?\n<image>"}, {"from": "gpt", "value": "a ship"}]
|
00000003960.jpg
|
3960_evjvqa
|
[{"from": "human", "value": "what are piled up on the ground next to the person in red shirt?\n<image>"}, {"from": "gpt", "value": "the black baskets"}]
|
00000003960.jpg
|
3960_evjvqa
|
[{"from": "human", "value": "what is the person in the purple floral dress pushing?\n<image>"}, {"from": "gpt", "value": "that person is pushing a supermarket cart"}]
|
00000003960.jpg
|
3960_evjvqa
|
[{"from": "human", "value": "what is the person in the purple floral dress doing?\n<image>"}, {"from": "gpt", "value": "that person is going shopping"}]
|
00000003961.jpg
|
3961_evjvqa
|
[{"from": "human", "value": "what activity are these people participating in?\n<image>"}, {"from": "gpt", "value": "they are cleaning the beach"}]
|
00000003961.jpg
|
3961_evjvqa
|
[{"from": "human", "value": "what do some people carry sacks for?\n<image>"}, {"from": "gpt", "value": "to contain garbage collected on the beach"}]
|
00000003961.jpg
|
3961_evjvqa
|
[{"from": "human", "value": "whom is the man in gray shirt and black trousers carrying the sack with?\n<image>"}, {"from": "gpt", "value": "with a man in a military uniform wearing a white cap"}]
|
00000003963.jpg
|
3963_evjvqa
|
[{"from": "human", "value": "what is the woman in blue shirt doing?\n<image>"}, {"from": "gpt", "value": "she is taking inventory of the goods"}]
|
00000003963.jpg
|
3963_evjvqa
|
[{"from": "human", "value": "how many people are standing at this supermarket?\n<image>"}, {"from": "gpt", "value": "one person is standing at this supermarket"}]
|
00000003963.jpg
|
3963_evjvqa
|
[{"from": "human", "value": "how does the only person standing next to the shelf dress?\n<image>"}, {"from": "gpt", "value": "that person wears blue t-shirt and black trousers"}]
|
00000003964.jpg
|
3964_evjvqa
|
[{"from": "human", "value": "what is the woman in red apron doing?\n<image>"}, {"from": "gpt", "value": "she is handing the bag of bread to the woman wearing the handbag"}]
|
00000003964.jpg
|
3964_evjvqa
|
[{"from": "human", "value": "whom is the bakery employee giving the bag of bread to?\n<image>"}, {"from": "gpt", "value": "the bakery employee is giving the bag of bread to the woman wearing the handbag"}]
|
00000003964.jpg
|
3964_evjvqa
|
[{"from": "human", "value": "who is the employee of this bakery?\n<image>"}, {"from": "gpt", "value": "the woman wearing red apron"}]
|
00000003970.jpg
|
3970_evjvqa
|
[{"from": "human", "value": "how many bowls are there on the table?\n<image>"}, {"from": "gpt", "value": "there are three bowls on the table"}]
|
00000003970.jpg
|
3970_evjvqa
|
[{"from": "human", "value": "how many plates of dry pancakes are there on the table?\n<image>"}, {"from": "gpt", "value": "there are two plates of dry pancakes on the table"}]
|
00000003970.jpg
|
3970_evjvqa
|
[{"from": "human", "value": "what items are dry pancakes displayed on?\n<image>"}, {"from": "gpt", "value": "the plastic plates"}]
|
00000003972.jpg
|
3972_evjvqa
|
[{"from": "human", "value": "how many boats are there on the river?\n<image>"}, {"from": "gpt", "value": "there are two boats on the river"}]
|
00000003972.jpg
|
3972_evjvqa
|
[{"from": "human", "value": "how many people sit on each boat?\n<image>"}, {"from": "gpt", "value": "one people sits on each boat"}]
|
00000003972.jpg
|
3972_evjvqa
|
[{"from": "human", "value": "what is the weather like in this area?\n<image>"}, {"from": "gpt", "value": "it is sunny in this area"}]
|
00000003973.jpg
|
3973_evjvqa
|
[{"from": "human", "value": "what is the woman sitting on a chair holding?\n<image>"}, {"from": "gpt", "value": "she is holding a hand fan"}]
|
00000003981.jpg
|
3981_evjvqa
|
[{"from": "human", "value": "what vehicles are parked on the pavement?\n<image>"}, {"from": "gpt", "value": "the cars are parked on the pavement"}]
|
00000003981.jpg
|
3981_evjvqa
|
[{"from": "human", "value": "what is the weather like in this quarter of the town?\n<image>"}, {"from": "gpt", "value": "it is sunny in this quarter of the town"}]
|
00000003981.jpg
|
3981_evjvqa
|
[{"from": "human", "value": "what is the scooter rider in white shirt wearing on head?\n<image>"}, {"from": "gpt", "value": "a helmet"}]
|
00000003985.jpg
|
3985_evjvqa
|
[{"from": "human", "value": "where is the boy holding the balloon standing?\n<image>"}, {"from": "gpt", "value": "that boy is standing behind the man in black shirt wearing a black helmet"}]
|
00000003985.jpg
|
3985_evjvqa
|
[{"from": "human", "value": "what is floating above the head of the man wearing a black helmet with a white face mask?\n<image>"}, {"from": "gpt", "value": "a balloon"}]
|
00000003985.jpg
|
3985_evjvqa
|
[{"from": "human", "value": "whom is the woman in brown t-shirt carrying?\n<image>"}, {"from": "gpt", "value": "she is carrying a baby"}]
|
00000003986.jpg
|
3986_evjvqa
|
[{"from": "human", "value": "when are these people eating and drinking?\n<image>"}, {"from": "gpt", "value": "they are eating and drinking in the evening"}]
|
00000003987.jpg
|
3987_evjvqa
|
[{"from": "human", "value": "what is the woman wearing black shirt and white face mask doing?\n<image>"}, {"from": "gpt", "value": "she is selecting apples"}]
|
00000003994.jpg
|
3994_evjvqa
|
[{"from": "human", "value": "how many customers sit in the restaurant?\n<image>"}, {"from": "gpt", "value": "no customer sits in the restaurant"}]
|
00000003995.jpg
|
3995_evjvqa
|
[{"from": "human", "value": "what hat is the man on the motorbike wearing?\n<image>"}, {"from": "gpt", "value": "he is wearing a helmet"}]
|
00000003995.jpg
|
3995_evjvqa
|
[{"from": "human", "value": "what is the man in the red helmet doing?\n<image>"}, {"from": "gpt", "value": "he is driving a motorbike"}]
|
00000003995.jpg
|
3995_evjvqa
|
[{"from": "human", "value": "what is the man in the blue long-sleeve shirt doing?\n<image>"}, {"from": "gpt", "value": "the man in the blue long-sleeve shirt is driving a motorbike"}]
|
00000003999.jpg
|
3999_evjvqa
|
[{"from": "human", "value": "what are the students holding brooms for?\n<image>"}, {"from": "gpt", "value": "sweeping leaves on the yard"}]
|
00000003999.jpg
|
3999_evjvqa
|
[{"from": "human", "value": "what does the student wearing a hat use to sweep the yard?\n<image>"}, {"from": "gpt", "value": "that student uses a broom"}]
|
00000005500.jpg
|
5500_evjvqa
|
[{"from": "human", "value": "what is the girl in the black shirt holding in her right hand?\n<image>"}, {"from": "gpt", "value": "the girl in the black shirt is holding a cucumber"}]
|
00000005500.jpg
|
5500_evjvqa
|
[{"from": "human", "value": "what area is the woman in the red shirt standing in?\n<image>"}, {"from": "gpt", "value": "the woman in the red shirt is standing in the weighing area"}]
|
00000005500.jpg
|
5500_evjvqa
|
[{"from": "human", "value": "what color shirt does the staff in charge of weighing items wear?\n<image>"}, {"from": "gpt", "value": "the staff in charge of weighing items wears yellow shirt"}]
|
00000005500.jpg
|
5500_evjvqa
|
[{"from": "human", "value": "what color is the bag of the girl in the black shirt?\n<image>"}, {"from": "gpt", "value": "the bag is white"}]
|
00000005501.jpg
|
5501_evjvqa
|
[{"from": "human", "value": "how many pairs of chopsticks are on the table?\n<image>"}, {"from": "gpt", "value": "one pair of chopsticks is on the table"}]
|
00000005501.jpg
|
5501_evjvqa
|
[{"from": "human", "value": "what is the dish inside the porcelain bowl?\n<image>"}, {"from": "gpt", "value": "inside the porcelain bowl is the beef pho dish"}]
|
00000005501.jpg
|
5501_evjvqa
|
[{"from": "human", "value": "where are the bean sprouts put on the table?\n<image>"}, {"from": "gpt", "value": "bean sprouts are contained inside a stainless steel bowl"}]
|
00000005502.jpg
|
5502_evjvqa
|
[{"from": "human", "value": "how many people are walking in all?\n<image>"}, {"from": "gpt", "value": "three people are walking"}]
|
00000005502.jpg
|
5502_evjvqa
|
[{"from": "human", "value": "what is the object placed leaning against the fence?\n<image>"}, {"from": "gpt", "value": "the motorbike is placed leaning against the fence"}]
|
00000005502.jpg
|
5502_evjvqa
|
[{"from": "human", "value": "what color shirt are the two men walking together wearing?\n<image>"}, {"from": "gpt", "value": "both of them are wearing white shirts"}]
|
00000005507.jpg
|
5507_evjvqa
|
[{"from": "human", "value": "what color is the roof of the city gate?\n<image>"}, {"from": "gpt", "value": "the roof is orange"}]
|
00000005507.jpg
|
5507_evjvqa
|
[{"from": "human", "value": "where are the flagpoles placed?\n<image>"}, {"from": "gpt", "value": "the flagpoles are erected in front of the city gate"}]
|
00000005508.jpg
|
5508_evjvqa
|
[{"from": "human", "value": "what color is the ship?\n<image>"}, {"from": "gpt", "value": "the ship is white"}]
|
00000005511.jpg
|
5511_evjvqa
|
[{"from": "human", "value": "what is mounted on the sides of the tower?\n<image>"}, {"from": "gpt", "value": "the clock faces are mounted on the sides of the tower"}]
|
00000005511.jpg
|
5511_evjvqa
|
[{"from": "human", "value": "what color is the roof of the market?\n<image>"}, {"from": "gpt", "value": "the roof is red"}]
|
00000005511.jpg
|
5511_evjvqa
|
[{"from": "human", "value": "how many sides does the clock tower of the market have?\n<image>"}, {"from": "gpt", "value": "the tower has four sides"}]
|
00000005512.jpg
|
5512_evjvqa
|
[{"from": "human", "value": "what kind of flag is hung to the left of the entrance?\n<image>"}, {"from": "gpt", "value": "the festival flag is hung to the left of the entrance"}]
|
00000005512.jpg
|
5512_evjvqa
|
[{"from": "human", "value": "where is the trash can placed?\n<image>"}, {"from": "gpt", "value": "the trash can is placed below the festival flag"}]
|
00000005512.jpg
|
5512_evjvqa
|
[{"from": "human", "value": "what are the two sides of the leading stair decorated with?\n<image>"}, {"from": "gpt", "value": "two flower pots are placed on two side of the stair"}]
|
00000005514.jpg
|
5514_evjvqa
|
[{"from": "human", "value": "how many mobile policemen are standing in front of the store?\n<image>"}, {"from": "gpt", "value": "three mobile policemen are standing in front of the store"}]
|
00000005514.jpg
|
5514_evjvqa
|
[{"from": "human", "value": "what color is the store name board?\n<image>"}, {"from": "gpt", "value": "the store name board is blue"}]
|
00000005514.jpg
|
5514_evjvqa
|
[{"from": "human", "value": "how many mobile policemen do not wear masks?\n<image>"}, {"from": "gpt", "value": "one person do not wear a mask"}]
|
00000005515.jpg
|
5515_evjvqa
|
[{"from": "human", "value": "what color hat is the woman in the brown jacket wearing?\n<image>"}, {"from": "gpt", "value": "the woman is wearing a white hat"}]
|
00000005515.jpg
|
5515_evjvqa
|
[{"from": "human", "value": "how many baskets of carrots are on display?\n<image>"}, {"from": "gpt", "value": "two baskets of carrots"}]
|
00000005515.jpg
|
5515_evjvqa
|
[{"from": "human", "value": "which side are the two talking women standing compared to the woman in the white hat?\n<image>"}, {"from": "gpt", "value": "the two women are standing on the left of the woman in the red hat"}]
|
00000005517.jpg
|
5517_evjvqa
|
[{"from": "human", "value": "how many people are smiling?\n<image>"}, {"from": "gpt", "value": "two people are smiling"}]
|
00000005517.jpg
|
5517_evjvqa
|
[{"from": "human", "value": "what is the store selling?\n<image>"}, {"from": "gpt", "value": "the store is selling newspapers"}]
|
00000005518.jpg
|
5518_evjvqa
|
[{"from": "human", "value": "what is the man in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "the man is leading the motorbike"}]
|
00000005519.jpg
|
5519_evjvqa
|
[{"from": "human", "value": "what are the two kids doing?\n<image>"}, {"from": "gpt", "value": "watching the elephants"}]
|
00000005519.jpg
|
5519_evjvqa
|
[{"from": "human", "value": "what color hat is the boy wearing?\n<image>"}, {"from": "gpt", "value": "the boy is wearing a blue hat"}]
|
00000005519.jpg
|
5519_evjvqa
|
[{"from": "human", "value": "what kind of clothes is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a dress"}]
|
00000005522.jpg
|
5522_evjvqa
|
[{"from": "human", "value": "what is the food put in?\n<image>"}, {"from": "gpt", "value": "the foot is put in a porcelain bowl"}]
|
00000005522.jpg
|
5522_evjvqa
|
[{"from": "human", "value": "what item is placed inside the food bowl?\n<image>"}, {"from": "gpt", "value": "a spoon is placed inside the food bowl"}]
|
00000005522.jpg
|
5522_evjvqa
|
[{"from": "human", "value": "what is the food in the bowl?\n<image>"}, {"from": "gpt", "value": "in the bowl is pho"}]
|
00000005523.jpg
|
5523_evjvqa
|
[{"from": "human", "value": "what expression does the man have?\n<image>"}, {"from": "gpt", "value": "the man is smiling"}]
|
00000005523.jpg
|
5523_evjvqa
|
[{"from": "human", "value": "which hand is the man putting on his chest?\n<image>"}, {"from": "gpt", "value": "the man is putting his left hand on his chest"}]
|
00000005523.jpg
|
5523_evjvqa
|
[{"from": "human", "value": "what items does the store sell?\n<image>"}, {"from": "gpt", "value": "the store sells food and drinks"}]
|
00000005523.jpg
|
5523_evjvqa
|
[{"from": "human", "value": "what is the item in front of the man?\n<image>"}, {"from": "gpt", "value": "eggs trays are placed in front of the man"}]
|
00000005525.jpg
|
5525_evjvqa
|
[{"from": "human", "value": "where is the bike parked?\n<image>"}, {"from": "gpt", "value": "the bike is parked right in front of the pharmacy"}]
|
00000005525.jpg
|
5525_evjvqa
|
[{"from": "human", "value": "what color shirt is the girl standing in front of the pharmacy wearing?\n<image>"}, {"from": "gpt", "value": "the girl standing in front of the pharmacy is wearing a black shirt"}]
|
00000005527.jpg
|
5527_evjvqa
|
[{"from": "human", "value": "what is the man in the white shirt doing?\n<image>"}, {"from": "gpt", "value": "the man is walking"}]
|
00000005527.jpg
|
5527_evjvqa
|
[{"from": "human", "value": "what color are the benches painted?\n<image>"}, {"from": "gpt", "value": "the benches are painted black"}]
|
00000005527.jpg
|
5527_evjvqa
|
[{"from": "human", "value": "what is the main color of the exercise equipment painted?\n<image>"}, {"from": "gpt", "value": "the main color painted for the exercise equipment is orange"}]
|
00000005528.jpg
|
5528_evjvqa
|
[{"from": "human", "value": "what is the girl holding in her hand?\n<image>"}, {"from": "gpt", "value": "the girl is holding a milk pack and a shopping cart"}]
|
00000005528.jpg
|
5528_evjvqa
|
[{"from": "human", "value": "what counter is the girl standing at?\n<image>"}, {"from": "gpt", "value": "the girl is standing at the milk products counter"}]
|
00000005528.jpg
|
5528_evjvqa
|
[{"from": "human", "value": "what color is the shopping cart the girl is holding?\n<image>"}, {"from": "gpt", "value": "the shopping cart the girl is holding is blue"}]
|
00000005532.jpg
|
5532_evjvqa
|
[{"from": "human", "value": "what is the man in the red jacket doing?\n<image>"}, {"from": "gpt", "value": "the man in the red jacket is about to light a cigarette"}]
|
00000005533.jpg
|
5533_evjvqa
|
[{"from": "human", "value": "what is the man in the black and yellow shirt holding in his hands?\n<image>"}, {"from": "gpt", "value": "the man in the black and yellow shirt is holding a pair of pants in his hands"}]
|
00000005533.jpg
|
5533_evjvqa
|
[{"from": "human", "value": "what color dress is the woman reaching out to get the clothes wearing?\n<image>"}, {"from": "gpt", "value": "the woman reaching out to get the clothes is wearing the outfit with black and white stripe pattern"}]
|
00000005533.jpg
|
5533_evjvqa
|
[{"from": "human", "value": "what is the woman in black doing?\n<image>"}, {"from": "gpt", "value": "the woman in black is about to take something out of her bag"}]
|
00000005534.jpg
|
5534_evjvqa
|
[{"from": "human", "value": "what is on the back of the man's bicycle?\n<image>"}, {"from": "gpt", "value": "bundles of vegetables are being placed on the back of the man's bicycle"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.