image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000010736.jpg
|
10736_evjvqa
|
[{"from": "human", "value": "what products do these stalls display?\n<image>"}, {"from": "gpt", "value": "types of milk"}]
|
00000010743.jpg
|
10743_evjvqa
|
[{"from": "human", "value": "what do these people sell in the market?\n<image>"}, {"from": "gpt", "value": "types of fabrics"}]
|
00000010746.jpg
|
10746_evjvqa
|
[{"from": "human", "value": "what means of transport do these people use to go?\n<image>"}, {"from": "gpt", "value": "the electric bus"}]
|
00000010747.jpg
|
10747_evjvqa
|
[{"from": "human", "value": "how many kids are holding balloons?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000010748.jpg
|
10748_evjvqa
|
[{"from": "human", "value": "where is the woman standing?\n<image>"}, {"from": "gpt", "value": "on the balcony of the house"}]
|
00000010751.jpg
|
10751_evjvqa
|
[{"from": "human", "value": "what are these two women doing?\n<image>"}, {"from": "gpt", "value": "đang lựa đồ"}]
|
00000010753.jpg
|
10753_evjvqa
|
[{"from": "human", "value": "what is the man sitting on the chair doing?\n<image>"}, {"from": "gpt", "value": "he has tested covid "}]
|
00000010754.jpg
|
10754_evjvqa
|
[{"from": "human", "value": "what is the woman holding the scoop for?\n<image>"}, {"from": "gpt", "value": "to draw broth to sell to guests to eat"}]
|
00000010758.jpg
|
10758_evjvqa
|
[{"from": "human", "value": "what color shirt is the girl riding a bicycle wearing?\n<image>"}, {"from": "gpt", "value": "orange"}]
|
00000010760.jpg
|
10760_evjvqa
|
[{"from": "human", "value": "where are the price tags hung?\n<image>"}, {"from": "gpt", "value": "on the fruit stall"}]
|
00000010763.jpg
|
10763_evjvqa
|
[{"from": "human", "value": "who are these two women sitting with?\n<image>"}, {"from": "gpt", "value": "two little girls"}]
|
00000010766.jpg
|
10766_evjvqa
|
[{"from": "human", "value": "what are the men with glasses doing?\n<image>"}, {"from": "gpt", "value": "holding phone to take a photo"}]
|
00000010767.jpg
|
10767_evjvqa
|
[{"from": "human", "value": "where is the tank?\n<image>"}, {"from": "gpt", "value": "in front of the building"}]
|
00000010770.jpg
|
10770_evjvqa
|
[{"from": "human", "value": "what kind of food do the stalls on the right hand side sell?\n<image>"}, {"from": "gpt", "value": "types of wine"}]
|
00000010771.jpg
|
10771_evjvqa
|
[{"from": "human", "value": "how many motorcycles are riding on the side of the truck?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000010772.jpg
|
10772_evjvqa
|
[{"from": "human", "value": "where are these two persons standing?\n<image>"}, {"from": "gpt", "value": "next to a sales shelf"}]
|
00000010773.jpg
|
10773_evjvqa
|
[{"from": "human", "value": "how many women are walking on the street?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000010774.jpg
|
10774_evjvqa
|
[{"from": "human", "value": "where are these people buying fruit?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}]
|
00000010775.jpg
|
10775_evjvqa
|
[{"from": "human", "value": "how many cars are being parked in the yard?\n<image>"}, {"from": "gpt", "value": "there are the bike and the motorbike"}]
|
00000010777.jpg
|
10777_evjvqa
|
[{"from": "human", "value": "where are the two babies being held?\n<image>"}, {"from": "gpt", "value": "in the hospital lobby"}]
|
00000010778.jpg
|
10778_evjvqa
|
[{"from": "human", "value": "who are these children?\n<image>"}, {"from": "gpt", "value": "they are elementary students"}]
|
00000010779.jpg
|
10779_evjvqa
|
[{"from": "human", "value": "what items does the store next to the pet store sell?\n<image>"}, {"from": "gpt", "value": "the sports shoes"}]
|
00000010786.jpg
|
10786_evjvqa
|
[{"from": "human", "value": "what is the store next to the toad shop where the two men are standing?\n<image>"}, {"from": "gpt", "value": "the clothes shop"}]
|
00000010787.jpg
|
10787_evjvqa
|
[{"from": "human", "value": "how many vehicles are on the street?\n<image>"}, {"from": "gpt", "value": "just the motorbike"}]
|
00000010788.jpg
|
10788_evjvqa
|
[{"from": "human", "value": "what are the two women doing?\n<image>"}, {"from": "gpt", "value": "waiting for temperature measurement"}]
|
00000010790.jpg
|
10790_evjvqa
|
[{"from": "human", "value": "what is this?\n<image>"}, {"from": "gpt", "value": "this is the coffin containing the mummy"}]
|
00000010791.jpg
|
10791_evjvqa
|
[{"from": "human", "value": "what do they sell on these blue baskets?\n<image>"}, {"from": "gpt", "value": "vegetable"}]
|
00000010792.jpg
|
10792_evjvqa
|
[{"from": "human", "value": "how many motorbikes are on the road?\n<image>"}, {"from": "gpt", "value": "just the motorbike"}]
|
00000010793.jpg
|
10793_evjvqa
|
[{"from": "human", "value": "what is the content of the paper?\n<image>"}, {"from": "gpt", "value": "it is the design"}]
|
00000010801.jpg
|
10801_evjvqa
|
[{"from": "human", "value": "what material are the tables and chairs of this restaurant made of?\n<image>"}, {"from": "gpt", "value": "by stainless steel"}]
|
00000010803.jpg
|
10803_evjvqa
|
[{"from": "human", "value": "what time is it in this neighborhood?\n<image>"}, {"from": "gpt", "value": "in the morning"}]
|
00000010805.jpg
|
10805_evjvqa
|
[{"from": "human", "value": "whose hand is the man in the white shirt holding?\n<image>"}, {"from": "gpt", "value": "it is the girl in the white aodai"}]
|
00000010807.jpg
|
10807_evjvqa
|
[{"from": "human", "value": "what are these old men watching?\n<image>"}, {"from": "gpt", "value": "the exhibition"}]
|
00000010809.jpg
|
10809_evjvqa
|
[{"from": "human", "value": "what products does this store offer?\n<image>"}, {"from": "gpt", "value": "type of paints"}]
|
00000010812.jpg
|
10812_evjvqa
|
[{"from": "human", "value": "what is the weather like in this area?\n<image>"}, {"from": "gpt", "value": "it is sunny"}]
|
00000010813.jpg
|
10813_evjvqa
|
[{"from": "human", "value": "what color vehicle is this woman sitting in?\n<image>"}, {"from": "gpt", "value": "she is sitting on the red scooter"}]
|
00000010817.jpg
|
10817_evjvqa
|
[{"from": "human", "value": "how many ceiling fans are installed in this room?\n<image>"}, {"from": "gpt", "value": "four"}]
|
00000010822.jpg
|
10822_evjvqa
|
[{"from": "human", "value": "what color is the roof of the cyclo?\n<image>"}, {"from": "gpt", "value": "green"}]
|
00000010823.jpg
|
10823_evjvqa
|
[{"from": "human", "value": "how many red plastic chairs are there on the sidewalk?\n<image>"}, {"from": "gpt", "value": "four"}]
|
00000010825.jpg
|
10825_evjvqa
|
[{"from": "human", "value": "what is the building?\n<image>"}, {"from": "gpt", "value": "it is the temple"}]
|
00000010829.jpg
|
10829_evjvqa
|
[{"from": "human", "value": "what pose is the man in the green striped shirt in?\n<image>"}, {"from": "gpt", "value": "he is standing"}]
|
00000010830.jpg
|
10830_evjvqa
|
[{"from": "human", "value": "how many noodles are there on the table?\n<image>"}, {"from": "gpt", "value": "four"}]
|
00000010833.jpg
|
10833_evjvqa
|
[{"from": "human", "value": "how many windows does this house have?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000010835.jpg
|
10835_evjvqa
|
[{"from": "human", "value": "where is the red bowl located?\n<image>"}, {"from": "gpt", "value": "to the right of the green chair"}]
|
00000010837.jpg
|
10837_evjvqa
|
[{"from": "human", "value": "what color are the walls of the houses in this neighborhood painted?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000010838.jpg
|
10838_evjvqa
|
[{"from": "human", "value": "what is a person in protective gear doing to the boy in the red shirt?\n<image>"}, {"from": "gpt", "value": "taking sample test for the boy in the red shirt"}]
|
00000010839.jpg
|
10839_evjvqa
|
[{"from": "human", "value": "what kind of red musical instrument is used on stage?\n<image>"}, {"from": "gpt", "value": "the drums"}]
|
00000010842.jpg
|
10842_evjvqa
|
[{"from": "human", "value": "where is the woman in skyblue standing?\n<image>"}, {"from": "gpt", "value": "she stands behind the woman in the black hat"}]
|
00000010843.jpg
|
10843_evjvqa
|
[{"from": "human", "value": "where is the man in gray going?\n<image>"}, {"from": "gpt", "value": "he is going shopping"}]
|
00000010849.jpg
|
10849_evjvqa
|
[{"from": "human", "value": "how many light bulbs does the lamp post have?\n<image>"}, {"from": "gpt", "value": "four"}]
|
00000010850.jpg
|
10850_evjvqa
|
[{"from": "human", "value": "who is driving the white scooter?\n<image>"}, {"from": "gpt", "value": "it is the woman in blue"}]
|
00000010852.jpg
|
10852_evjvqa
|
[{"from": "human", "value": "which arm does the man in jeans have a tattoo on?\n<image>"}, {"from": "gpt", "value": "on the right "}]
|
00000010857.jpg
|
10857_evjvqa
|
[{"from": "human", "value": "what is the woman in the blue shirt buying at this stall?\n<image>"}, {"from": "gpt", "value": "the duck eggs"}]
|
00000010858.jpg
|
10858_evjvqa
|
[{"from": "human", "value": "how many silver cars are parked in front of the supermarket?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000010863.jpg
|
10863_evjvqa
|
[{"from": "human", "value": "what is the yellow in the supermarket trolley?\n<image>"}, {"from": "gpt", "value": "the cooking oil bottle"}]
|
00000010864.jpg
|
10864_evjvqa
|
[{"from": "human", "value": "who does the man in the red shirt go shopping with?\n<image>"}, {"from": "gpt", "value": "the woman is wearing the yellow pant"}]
|
00000010865.jpg
|
10865_evjvqa
|
[{"from": "human", "value": "what products are on promotion?\n<image>"}, {"from": "gpt", "value": "the houseware"}]
|
00000010866.jpg
|
10866_evjvqa
|
[{"from": "human", "value": "what color shirt is the girl sitting on?\n<image>"}, {"from": "gpt", "value": "red"}]
|
00000010867.jpg
|
10867_evjvqa
|
[{"from": "human", "value": "what is the woman in the blue shirt selling to the man on the motorbike?\n<image>"}, {"from": "gpt", "value": "the lottery tickets"}]
|
00000010868.jpg
|
10868_evjvqa
|
[{"from": "human", "value": "what color clothes is the girl sitting in the sun wearing?\n<image>"}, {"from": "gpt", "value": "yellow"}]
|
00000010869.jpg
|
10869_evjvqa
|
[{"from": "human", "value": "where is the man in the green jacket?\n<image>"}, {"from": "gpt", "value": "he is sitting on a motorbike in front of the house on the corner"}]
|
00000010870.jpg
|
10870_evjvqa
|
[{"from": "human", "value": "which road leads through an underpass?\n<image>"}, {"from": "gpt", "value": "the middle"}]
|
00000010872.jpg
|
10872_evjvqa
|
[{"from": "human", "value": "how many floors is the lowest shelf here?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000010876.jpg
|
10876_evjvqa
|
[{"from": "human", "value": "what color pillow is the girl lying on?\n<image>"}, {"from": "gpt", "value": "pink"}]
|
00000010878.jpg
|
10878_evjvqa
|
[{"from": "human", "value": "what is the girl turning her head looking at?\n<image>"}, {"from": "gpt", "value": "the mot cot pagoda"}]
|
00000010879.jpg
|
10879_evjvqa
|
[{"from": "human", "value": "how many loaves of bread with coriander?h\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000010880.jpg
|
10880_evjvqa
|
[{"from": "human", "value": "what material are the tables and chairs in this room made of?\n<image>"}, {"from": "gpt", "value": "by the wood"}]
|
00000010881.jpg
|
10881_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "he is taking care of the rose pots"}]
|
00000010883.jpg
|
10883_evjvqa
|
[{"from": "human", "value": "what is the animal symbol on the sign on the left?\n<image>"}, {"from": "gpt", "value": "the crab"}]
|
00000010884.jpg
|
10884_evjvqa
|
[{"from": "human", "value": "what does the man in the yellow hat carry on his back?\n<image>"}, {"from": "gpt", "value": "the backpack"}]
|
00000010885.jpg
|
10885_evjvqa
|
[{"from": "human", "value": "which side of the boy is the girl sitting on?\n<image>"}, {"from": "gpt", "value": "on the left side"}]
|
00000010886.jpg
|
10886_evjvqa
|
[{"from": "human", "value": "what activity is going on here?\n<image>"}, {"from": "gpt", "value": "there is a book fair "}]
|
00000010887.jpg
|
10887_evjvqa
|
[{"from": "human", "value": "which side does the woman in the white dress stand on compared to the man?\n<image>"}, {"from": "gpt", "value": "she stands behind the man"}]
|
00000010890.jpg
|
10890_evjvqa
|
[{"from": "human", "value": "what does this stall sell?\n<image>"}, {"from": "gpt", "value": "the christmas decorations"}]
|
00000010891.jpg
|
10891_evjvqa
|
[{"from": "human", "value": "what are the mobile police soldiers standing for?\n<image>"}, {"from": "gpt", "value": "they are on duty at the scene"}]
|
00000010893.jpg
|
10893_evjvqa
|
[{"from": "human", "value": "how many boats are on the water?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000010895.jpg
|
10895_evjvqa
|
[{"from": "human", "value": "what means are the two women traveling?\n<image>"}, {"from": "gpt", "value": "bike"}]
|
00000010896.jpg
|
10896_evjvqa
|
[{"from": "human", "value": "what's the weather like in this neighborhood?\n<image>"}, {"from": "gpt", "value": "it is sunny "}]
|
00000004800.jpg
|
4800_evjvqa
|
[{"from": "human", "value": "which arm is the girl raising?\n<image>"}, {"from": "gpt", "value": "the girl is raising the left arm"}]
|
00000004800.jpg
|
4800_evjvqa
|
[{"from": "human", "value": "where is she sitting?\n<image>"}, {"from": "gpt", "value": "sitting under a tree"}]
|
00000004800.jpg
|
4800_evjvqa
|
[{"from": "human", "value": "what color is the shirt of the girl?\n<image>"}, {"from": "gpt", "value": "the shirt is white"}]
|
00000004801.jpg
|
4801_evjvqa
|
[{"from": "human", "value": "where is the tea set placed?\n<image>"}, {"from": "gpt", "value": "the tea set is placed on a table"}]
|
00000004801.jpg
|
4801_evjvqa
|
[{"from": "human", "value": "how many red boxes are under the fan?\n<image>"}, {"from": "gpt", "value": "there are two red boxes under the fan"}]
|
00000004801.jpg
|
4801_evjvqa
|
[{"from": "human", "value": "what color is the bag on the chair?\n<image>"}, {"from": "gpt", "value": "the bag is black"}]
|
00000004804.jpg
|
4804_evjvqa
|
[{"from": "human", "value": "how many boys are walking?\n<image>"}, {"from": "gpt", "value": "there are three boys walking"}]
|
00000004804.jpg
|
4804_evjvqa
|
[{"from": "human", "value": "what is the girl holding in her hand?\n<image>"}, {"from": "gpt", "value": "the girl is holding a stack of paper in her hand"}]
|
00000004804.jpg
|
4804_evjvqa
|
[{"from": "human", "value": "what color shirt is the boy walking separatedly wearing?\n<image>"}, {"from": "gpt", "value": "wearing a red shirt"}]
|
00000004806.jpg
|
4806_evjvqa
|
[{"from": "human", "value": "what is everyone doing?\n<image>"}, {"from": "gpt", "value": "visiting a relic site"}]
|
00000004806.jpg
|
4806_evjvqa
|
[{"from": "human", "value": "how many paths leading to the top are there?\n<image>"}, {"from": "gpt", "value": "there are two paths leading to the top"}]
|
00000004806.jpg
|
4806_evjvqa
|
[{"from": "human", "value": "how many underneath entries are there?\n<image>"}, {"from": "gpt", "value": "there are five underneath entries"}]
|
00000004808.jpg
|
4808_evjvqa
|
[{"from": "human", "value": "how many girls are standing in front of the stall?\n<image>"}, {"from": "gpt", "value": "there is one girl standing in front of the stall"}]
|
00000004808.jpg
|
4808_evjvqa
|
[{"from": "human", "value": "what color is the hat of the girl?\n<image>"}, {"from": "gpt", "value": "the hat is black"}]
|
00000004808.jpg
|
4808_evjvqa
|
[{"from": "human", "value": "what product is placed on the table?\n<image>"}, {"from": "gpt", "value": "fruit is placed on the table"}]
|
00000004810.jpg
|
4810_evjvqa
|
[{"from": "human", "value": "where are everyone?\n<image>"}, {"from": "gpt", "value": "on a river in a jungle"}]
|
00000004810.jpg
|
4810_evjvqa
|
[{"from": "human", "value": "what color shirt is the boatman wearing?\n<image>"}, {"from": "gpt", "value": "wearing a black shirt"}]
|
00000004810.jpg
|
4810_evjvqa
|
[{"from": "human", "value": "how many people are on a canoe?\n<image>"}, {"from": "gpt", "value": "there are four people on a canoe"}]
|
00000004811.jpg
|
4811_evjvqa
|
[{"from": "human", "value": "what are the man doing?\n<image>"}, {"from": "gpt", "value": "watering vegetables"}]
|
00000004811.jpg
|
4811_evjvqa
|
[{"from": "human", "value": "how many watering cans are the man carrying?\n<image>"}, {"from": "gpt", "value": "carrying two watering cans"}]
|
00000004811.jpg
|
4811_evjvqa
|
[{"from": "human", "value": "what color is the shirt of the man?\n<image>"}, {"from": "gpt", "value": "the shirt is blue"}]
|
00000004812.jpg
|
4812_evjvqa
|
[{"from": "human", "value": "what place is this scene?\n<image>"}, {"from": "gpt", "value": "the scene of a river and a town by the river bank"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.