image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000010736.jpg
10736_evjvqa
[{"from": "human", "value": "what products do these stalls display?\n<image>"}, {"from": "gpt", "value": "types of milk"}]
00000010743.jpg
10743_evjvqa
[{"from": "human", "value": "what do these people sell in the market?\n<image>"}, {"from": "gpt", "value": "types of fabrics"}]
00000010746.jpg
10746_evjvqa
[{"from": "human", "value": "what means of transport do these people use to go?\n<image>"}, {"from": "gpt", "value": "the electric bus"}]
00000010747.jpg
10747_evjvqa
[{"from": "human", "value": "how many kids are holding balloons?\n<image>"}, {"from": "gpt", "value": "only one"}]
00000010748.jpg
10748_evjvqa
[{"from": "human", "value": "where is the woman standing?\n<image>"}, {"from": "gpt", "value": "on the balcony of the house"}]
00000010751.jpg
10751_evjvqa
[{"from": "human", "value": "what are these two women doing?\n<image>"}, {"from": "gpt", "value": "đang lựa đồ"}]
00000010753.jpg
10753_evjvqa
[{"from": "human", "value": "what is the man sitting on the chair doing?\n<image>"}, {"from": "gpt", "value": "he has tested covid "}]
00000010754.jpg
10754_evjvqa
[{"from": "human", "value": "what is the woman holding the scoop for?\n<image>"}, {"from": "gpt", "value": "to draw broth to sell to guests to eat"}]
00000010758.jpg
10758_evjvqa
[{"from": "human", "value": "what color shirt is the girl riding a bicycle wearing?\n<image>"}, {"from": "gpt", "value": "orange"}]
00000010760.jpg
10760_evjvqa
[{"from": "human", "value": "where are the price tags hung?\n<image>"}, {"from": "gpt", "value": "on the fruit stall"}]
00000010763.jpg
10763_evjvqa
[{"from": "human", "value": "who are these two women sitting with?\n<image>"}, {"from": "gpt", "value": "two little girls"}]
00000010766.jpg
10766_evjvqa
[{"from": "human", "value": "what are the men with glasses doing?\n<image>"}, {"from": "gpt", "value": "holding phone to take a photo"}]
00000010767.jpg
10767_evjvqa
[{"from": "human", "value": "where is the tank?\n<image>"}, {"from": "gpt", "value": "in front of the building"}]
00000010770.jpg
10770_evjvqa
[{"from": "human", "value": "what kind of food do the stalls on the right hand side sell?\n<image>"}, {"from": "gpt", "value": "types of wine"}]
00000010771.jpg
10771_evjvqa
[{"from": "human", "value": "how many motorcycles are riding on the side of the truck?\n<image>"}, {"from": "gpt", "value": "two"}]
00000010772.jpg
10772_evjvqa
[{"from": "human", "value": "where are these two persons standing?\n<image>"}, {"from": "gpt", "value": "next to a sales shelf"}]
00000010773.jpg
10773_evjvqa
[{"from": "human", "value": "how many women are walking on the street?\n<image>"}, {"from": "gpt", "value": "three"}]
00000010774.jpg
10774_evjvqa
[{"from": "human", "value": "where are these people buying fruit?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}]
00000010775.jpg
10775_evjvqa
[{"from": "human", "value": "how many cars are being parked in the yard?\n<image>"}, {"from": "gpt", "value": "there are the bike and the motorbike"}]
00000010777.jpg
10777_evjvqa
[{"from": "human", "value": "where are the two babies being held?\n<image>"}, {"from": "gpt", "value": "in the hospital lobby"}]
00000010778.jpg
10778_evjvqa
[{"from": "human", "value": "who are these children?\n<image>"}, {"from": "gpt", "value": "they are elementary students"}]
00000010779.jpg
10779_evjvqa
[{"from": "human", "value": "what items does the store next to the pet store sell?\n<image>"}, {"from": "gpt", "value": "the sports shoes"}]
00000010786.jpg
10786_evjvqa
[{"from": "human", "value": "what is the store next to the toad shop where the two men are standing?\n<image>"}, {"from": "gpt", "value": "the clothes shop"}]
00000010787.jpg
10787_evjvqa
[{"from": "human", "value": "how many vehicles are on the street?\n<image>"}, {"from": "gpt", "value": "just the motorbike"}]
00000010788.jpg
10788_evjvqa
[{"from": "human", "value": "what are the two women doing?\n<image>"}, {"from": "gpt", "value": "waiting for temperature measurement"}]
00000010790.jpg
10790_evjvqa
[{"from": "human", "value": "what is this?\n<image>"}, {"from": "gpt", "value": "this is the coffin containing the mummy"}]
00000010791.jpg
10791_evjvqa
[{"from": "human", "value": "what do they sell on these blue baskets?\n<image>"}, {"from": "gpt", "value": "vegetable"}]
00000010792.jpg
10792_evjvqa
[{"from": "human", "value": "how many motorbikes are on the road?\n<image>"}, {"from": "gpt", "value": "just the motorbike"}]
00000010793.jpg
10793_evjvqa
[{"from": "human", "value": "what is the content of the paper?\n<image>"}, {"from": "gpt", "value": "it is the design"}]
00000010801.jpg
10801_evjvqa
[{"from": "human", "value": "what material are the tables and chairs of this restaurant made of?\n<image>"}, {"from": "gpt", "value": "by stainless steel"}]
00000010803.jpg
10803_evjvqa
[{"from": "human", "value": "what time is it in this neighborhood?\n<image>"}, {"from": "gpt", "value": "in the morning"}]
00000010805.jpg
10805_evjvqa
[{"from": "human", "value": "whose hand is the man in the white shirt holding?\n<image>"}, {"from": "gpt", "value": "it is the girl in the white aodai"}]
00000010807.jpg
10807_evjvqa
[{"from": "human", "value": "what are these old men watching?\n<image>"}, {"from": "gpt", "value": "the exhibition"}]
00000010809.jpg
10809_evjvqa
[{"from": "human", "value": "what products does this store offer?\n<image>"}, {"from": "gpt", "value": "type of paints"}]
00000010812.jpg
10812_evjvqa
[{"from": "human", "value": "what is the weather like in this area?\n<image>"}, {"from": "gpt", "value": "it is sunny"}]
00000010813.jpg
10813_evjvqa
[{"from": "human", "value": "what color vehicle is this woman sitting in?\n<image>"}, {"from": "gpt", "value": "she is sitting on the red scooter"}]
00000010817.jpg
10817_evjvqa
[{"from": "human", "value": "how many ceiling fans are installed in this room?\n<image>"}, {"from": "gpt", "value": "four"}]
00000010822.jpg
10822_evjvqa
[{"from": "human", "value": "what color is the roof of the cyclo?\n<image>"}, {"from": "gpt", "value": "green"}]
00000010823.jpg
10823_evjvqa
[{"from": "human", "value": "how many red plastic chairs are there on the sidewalk?\n<image>"}, {"from": "gpt", "value": "four"}]
00000010825.jpg
10825_evjvqa
[{"from": "human", "value": "what is the building?\n<image>"}, {"from": "gpt", "value": "it is the temple"}]
00000010829.jpg
10829_evjvqa
[{"from": "human", "value": "what pose is the man in the green striped shirt in?\n<image>"}, {"from": "gpt", "value": "he is standing"}]
00000010830.jpg
10830_evjvqa
[{"from": "human", "value": "how many noodles are there on the table?\n<image>"}, {"from": "gpt", "value": "four"}]
00000010833.jpg
10833_evjvqa
[{"from": "human", "value": "how many windows does this house have?\n<image>"}, {"from": "gpt", "value": "three"}]
00000010835.jpg
10835_evjvqa
[{"from": "human", "value": "where is the red bowl located?\n<image>"}, {"from": "gpt", "value": "to the right of the green chair"}]
00000010837.jpg
10837_evjvqa
[{"from": "human", "value": "what color are the walls of the houses in this neighborhood painted?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000010838.jpg
10838_evjvqa
[{"from": "human", "value": "what is a person in protective gear doing to the boy in the red shirt?\n<image>"}, {"from": "gpt", "value": "taking sample test for the boy in the red shirt"}]
00000010839.jpg
10839_evjvqa
[{"from": "human", "value": "what kind of red musical instrument is used on stage?\n<image>"}, {"from": "gpt", "value": "the drums"}]
00000010842.jpg
10842_evjvqa
[{"from": "human", "value": "where is the woman in skyblue standing?\n<image>"}, {"from": "gpt", "value": "she stands behind the woman in the black hat"}]
00000010843.jpg
10843_evjvqa
[{"from": "human", "value": "where is the man in gray going?\n<image>"}, {"from": "gpt", "value": "he is going shopping"}]
00000010849.jpg
10849_evjvqa
[{"from": "human", "value": "how many light bulbs does the lamp post have?\n<image>"}, {"from": "gpt", "value": "four"}]
00000010850.jpg
10850_evjvqa
[{"from": "human", "value": "who is driving the white scooter?\n<image>"}, {"from": "gpt", "value": "it is the woman in blue"}]
00000010852.jpg
10852_evjvqa
[{"from": "human", "value": "which arm does the man in jeans have a tattoo on?\n<image>"}, {"from": "gpt", "value": "on the right "}]
00000010857.jpg
10857_evjvqa
[{"from": "human", "value": "what is the woman in the blue shirt buying at this stall?\n<image>"}, {"from": "gpt", "value": "the duck eggs"}]
00000010858.jpg
10858_evjvqa
[{"from": "human", "value": "how many silver cars are parked in front of the supermarket?\n<image>"}, {"from": "gpt", "value": "two"}]
00000010863.jpg
10863_evjvqa
[{"from": "human", "value": "what is the yellow in the supermarket trolley?\n<image>"}, {"from": "gpt", "value": "the cooking oil bottle"}]
00000010864.jpg
10864_evjvqa
[{"from": "human", "value": "who does the man in the red shirt go shopping with?\n<image>"}, {"from": "gpt", "value": "the woman is wearing the yellow pant"}]
00000010865.jpg
10865_evjvqa
[{"from": "human", "value": "what products are on promotion?\n<image>"}, {"from": "gpt", "value": "the houseware"}]
00000010866.jpg
10866_evjvqa
[{"from": "human", "value": "what color shirt is the girl sitting on?\n<image>"}, {"from": "gpt", "value": "red"}]
00000010867.jpg
10867_evjvqa
[{"from": "human", "value": "what is the woman in the blue shirt selling to the man on the motorbike?\n<image>"}, {"from": "gpt", "value": "the lottery tickets"}]
00000010868.jpg
10868_evjvqa
[{"from": "human", "value": "what color clothes is the girl sitting in the sun wearing?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000010869.jpg
10869_evjvqa
[{"from": "human", "value": "where is the man in the green jacket?\n<image>"}, {"from": "gpt", "value": "he is sitting on a motorbike in front of the house on the corner"}]
00000010870.jpg
10870_evjvqa
[{"from": "human", "value": "which road leads through an underpass?\n<image>"}, {"from": "gpt", "value": "the middle"}]
00000010872.jpg
10872_evjvqa
[{"from": "human", "value": "how many floors is the lowest shelf here?\n<image>"}, {"from": "gpt", "value": "three"}]
00000010876.jpg
10876_evjvqa
[{"from": "human", "value": "what color pillow is the girl lying on?\n<image>"}, {"from": "gpt", "value": "pink"}]
00000010878.jpg
10878_evjvqa
[{"from": "human", "value": "what is the girl turning her head looking at?\n<image>"}, {"from": "gpt", "value": "the mot cot pagoda"}]
00000010879.jpg
10879_evjvqa
[{"from": "human", "value": "how many loaves of bread with coriander?h\n<image>"}, {"from": "gpt", "value": "two"}]
00000010880.jpg
10880_evjvqa
[{"from": "human", "value": "what material are the tables and chairs in this room made of?\n<image>"}, {"from": "gpt", "value": "by the wood"}]
00000010881.jpg
10881_evjvqa
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "he is taking care of the rose pots"}]
00000010883.jpg
10883_evjvqa
[{"from": "human", "value": "what is the animal symbol on the sign on the left?\n<image>"}, {"from": "gpt", "value": "the crab"}]
00000010884.jpg
10884_evjvqa
[{"from": "human", "value": "what does the man in the yellow hat carry on his back?\n<image>"}, {"from": "gpt", "value": "the backpack"}]
00000010885.jpg
10885_evjvqa
[{"from": "human", "value": "which side of the boy is the girl sitting on?\n<image>"}, {"from": "gpt", "value": "on the left side"}]
00000010886.jpg
10886_evjvqa
[{"from": "human", "value": "what activity is going on here?\n<image>"}, {"from": "gpt", "value": "there is a book fair "}]
00000010887.jpg
10887_evjvqa
[{"from": "human", "value": "which side does the woman in the white dress stand on compared to the man?\n<image>"}, {"from": "gpt", "value": "she stands behind the man"}]
00000010890.jpg
10890_evjvqa
[{"from": "human", "value": "what does this stall sell?\n<image>"}, {"from": "gpt", "value": "the christmas decorations"}]
00000010891.jpg
10891_evjvqa
[{"from": "human", "value": "what are the mobile police soldiers standing for?\n<image>"}, {"from": "gpt", "value": "they are on duty at the scene"}]
00000010893.jpg
10893_evjvqa
[{"from": "human", "value": "how many boats are on the water?\n<image>"}, {"from": "gpt", "value": "three"}]
00000010895.jpg
10895_evjvqa
[{"from": "human", "value": "what means are the two women traveling?\n<image>"}, {"from": "gpt", "value": "bike"}]
00000010896.jpg
10896_evjvqa
[{"from": "human", "value": "what's the weather like in this neighborhood?\n<image>"}, {"from": "gpt", "value": "it is sunny "}]
00000004800.jpg
4800_evjvqa
[{"from": "human", "value": "which arm is the girl raising?\n<image>"}, {"from": "gpt", "value": "the girl is raising the left arm"}]
00000004800.jpg
4800_evjvqa
[{"from": "human", "value": "where is she sitting?\n<image>"}, {"from": "gpt", "value": "sitting under a tree"}]
00000004800.jpg
4800_evjvqa
[{"from": "human", "value": "what color is the shirt of the girl?\n<image>"}, {"from": "gpt", "value": "the shirt is white"}]
00000004801.jpg
4801_evjvqa
[{"from": "human", "value": "where is the tea set placed?\n<image>"}, {"from": "gpt", "value": "the tea set is placed on a table"}]
00000004801.jpg
4801_evjvqa
[{"from": "human", "value": "how many red boxes are under the fan?\n<image>"}, {"from": "gpt", "value": "there are two red boxes under the fan"}]
00000004801.jpg
4801_evjvqa
[{"from": "human", "value": "what color is the bag on the chair?\n<image>"}, {"from": "gpt", "value": "the bag is black"}]
00000004804.jpg
4804_evjvqa
[{"from": "human", "value": "how many boys are walking?\n<image>"}, {"from": "gpt", "value": "there are three boys walking"}]
00000004804.jpg
4804_evjvqa
[{"from": "human", "value": "what is the girl holding in her hand?\n<image>"}, {"from": "gpt", "value": "the girl is holding a stack of paper in her hand"}]
00000004804.jpg
4804_evjvqa
[{"from": "human", "value": "what color shirt is the boy walking separatedly wearing?\n<image>"}, {"from": "gpt", "value": "wearing a red shirt"}]
00000004806.jpg
4806_evjvqa
[{"from": "human", "value": "what is everyone doing?\n<image>"}, {"from": "gpt", "value": "visiting a relic site"}]
00000004806.jpg
4806_evjvqa
[{"from": "human", "value": "how many paths leading to the top are there?\n<image>"}, {"from": "gpt", "value": "there are two paths leading to the top"}]
00000004806.jpg
4806_evjvqa
[{"from": "human", "value": "how many underneath entries are there?\n<image>"}, {"from": "gpt", "value": "there are five underneath entries"}]
00000004808.jpg
4808_evjvqa
[{"from": "human", "value": "how many girls are standing in front of the stall?\n<image>"}, {"from": "gpt", "value": "there is one girl standing in front of the stall"}]
00000004808.jpg
4808_evjvqa
[{"from": "human", "value": "what color is the hat of the girl?\n<image>"}, {"from": "gpt", "value": "the hat is black"}]
00000004808.jpg
4808_evjvqa
[{"from": "human", "value": "what product is placed on the table?\n<image>"}, {"from": "gpt", "value": "fruit is placed on the table"}]
00000004810.jpg
4810_evjvqa
[{"from": "human", "value": "where are everyone?\n<image>"}, {"from": "gpt", "value": "on a river in a jungle"}]
00000004810.jpg
4810_evjvqa
[{"from": "human", "value": "what color shirt is the boatman wearing?\n<image>"}, {"from": "gpt", "value": "wearing a black shirt"}]
00000004810.jpg
4810_evjvqa
[{"from": "human", "value": "how many people are on a canoe?\n<image>"}, {"from": "gpt", "value": "there are four people on a canoe"}]
00000004811.jpg
4811_evjvqa
[{"from": "human", "value": "what are the man doing?\n<image>"}, {"from": "gpt", "value": "watering vegetables"}]
00000004811.jpg
4811_evjvqa
[{"from": "human", "value": "how many watering cans are the man carrying?\n<image>"}, {"from": "gpt", "value": "carrying two watering cans"}]
00000004811.jpg
4811_evjvqa
[{"from": "human", "value": "what color is the shirt of the man?\n<image>"}, {"from": "gpt", "value": "the shirt is blue"}]
00000004812.jpg
4812_evjvqa
[{"from": "human", "value": "what place is this scene?\n<image>"}, {"from": "gpt", "value": "the scene of a river and a town by the river bank"}]