image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000002591.jpg
|
2591_evjvqa
|
[{"from": "human", "value": "what is the main color of the outerwear that the man riding a motorbike wears?\n<image>"}, {"from": "gpt", "value": "the main color of the outerwear that the man riding a motorbike wears is green"}]
|
00000002591.jpg
|
2591_evjvqa
|
[{"from": "human", "value": "where is the young man wearing a white face mask standing?\n<image>"}, {"from": "gpt", "value": "the young man wearing a white face mask is standing on the pavement"}]
|
00000002800.jpg
|
2800_evjvqa
|
[{"from": "human", "value": "how many children does the woman in the blue coat take with her to the supermarket?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000002800.jpg
|
2800_evjvqa
|
[{"from": "human", "value": "who is going to the supermarket with the woman in the shirt?\n<image>"}, {"from": "gpt", "value": "with a baby in a red shirt sitting in a stroller"}]
|
00000002800.jpg
|
2800_evjvqa
|
[{"from": "human", "value": "where is the woman in the green overcoat with the baby sitting with?\n<image>"}, {"from": "gpt", "value": "in the trolley"}]
|
00000002800.jpg
|
2800_evjvqa
|
[{"from": "human", "value": "how many people use the trolley in the supermarket?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002800.jpg
|
2800_evjvqa
|
[{"from": "human", "value": "who is the woman in black using a shopping trolley with?\n<image>"}, {"from": "gpt", "value": "with three children"}]
|
00000002800.jpg
|
2800_evjvqa
|
[{"from": "human", "value": "how many children are in the trolley out of three children going with the woman in black?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000002806.jpg
|
2806_evjvqa
|
[{"from": "human", "value": "what is the room?\n<image>"}, {"from": "gpt", "value": "the gallery space of a museum"}]
|
00000002807.jpg
|
2807_evjvqa
|
[{"from": "human", "value": "which stall is the boy in orange standing?\n<image>"}, {"from": "gpt", "value": "the toy stall"}]
|
00000002807.jpg
|
2807_evjvqa
|
[{"from": "human", "value": "which stall is the boy in green standing?\n<image>"}, {"from": "gpt", "value": "the toy stall"}]
|
00000002807.jpg
|
2807_evjvqa
|
[{"from": "human", "value": "how many children are standing at the toy stall?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000002807.jpg
|
2807_evjvqa
|
[{"from": "human", "value": "where is the woman in the red t-shirt?\n<image>"}, {"from": "gpt", "value": "in the toy stall"}]
|
00000002807.jpg
|
2807_evjvqa
|
[{"from": "human", "value": "how many women are coming outside the toy stall?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002807.jpg
|
2807_evjvqa
|
[{"from": "human", "value": "how many people are sitting at the checkout?\n<image>"}, {"from": "gpt", "value": "noone"}]
|
00000002813.jpg
|
2813_evjvqa
|
[{"from": "human", "value": "what goods do these boats mainly carry?\n<image>"}, {"from": "gpt", "value": "fruits and vegetables"}]
|
00000002815.jpg
|
2815_evjvqa
|
[{"from": "human", "value": "where are the people standing?\n<image>"}, {"from": "gpt", "value": "in front of the entrance of the historical monument "}]
|
00000002816.jpg
|
2816_evjvqa
|
[{"from": "human", "value": "where is the man in blue walking?\n<image>"}, {"from": "gpt", "value": "on the street"}]
|
00000002816.jpg
|
2816_evjvqa
|
[{"from": "human", "value": "what is the man in blue doing?\n<image>"}, {"from": "gpt", "value": "he is walking on the street"}]
|
00000002816.jpg
|
2816_evjvqa
|
[{"from": "human", "value": "who is walking with the man in blue?\n<image>"}, {"from": "gpt", "value": "no one"}]
|
00000002816.jpg
|
2816_evjvqa
|
[{"from": "human", "value": "what is the woman in yellow doing?\n<image>"}, {"from": "gpt", "value": "she is taking a photo"}]
|
00000002816.jpg
|
2816_evjvqa
|
[{"from": "human", "value": "how many children are walking in front of the woman holding the phone to take a photo?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002817.jpg
|
2817_evjvqa
|
[{"from": "human", "value": "what items does this stall sell?\n<image>"}, {"from": "gpt", "value": "clothes is main"}]
|
00000002818.jpg
|
2818_evjvqa
|
[{"from": "human", "value": "where is the woman sitting?\n<image>"}, {"from": "gpt", "value": "on the riverbank"}]
|
00000002818.jpg
|
2818_evjvqa
|
[{"from": "human", "value": "who is sitting with the woman?\n<image>"}, {"from": "gpt", "value": "no one"}]
|
00000002818.jpg
|
2818_evjvqa
|
[{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "she is sitting on the riverbank"}]
|
00000002823.jpg
|
2823_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "he is stacking the products on the shelves"}]
|
00000002823.jpg
|
2823_evjvqa
|
[{"from": "human", "value": "what color of the t-shirt is the staff wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
|
00000002823.jpg
|
2823_evjvqa
|
[{"from": "human", "value": "where is the man standing?\n<image>"}, {"from": "gpt", "value": "at the stall in the supermarket"}]
|
00000002823.jpg
|
2823_evjvqa
|
[{"from": "human", "value": "what is the staff doing?\n<image>"}, {"from": "gpt", "value": "he is stacking the products on the shelves"}]
|
00000002828.jpg
|
2828_evjvqa
|
[{"from": "human", "value": "where are the people standing?\n<image>"}, {"from": "gpt", "value": "in the fair "}]
|
00000002832.jpg
|
2832_evjvqa
|
[{"from": "human", "value": "where is the woman in orange?\n<image>"}, {"from": "gpt", "value": "on the small canoe"}]
|
00000002832.jpg
|
2832_evjvqa
|
[{"from": "human", "value": "what is the woman in orange wearing on her head?\n<image>"}, {"from": "gpt", "value": "the conical hat"}]
|
00000002832.jpg
|
2832_evjvqa
|
[{"from": "human", "value": "how many tires are on the boat to the right of the woman in the orange?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000002838.jpg
|
2838_evjvqa
|
[{"from": "human", "value": "who is standing woth the man in white?\n<image>"}, {"from": "gpt", "value": "with the man in blue"}]
|
00000002839.jpg
|
2839_evjvqa
|
[{"from": "human", "value": "what hats do most of the people use?\n<image>"}, {"from": "gpt", "value": "the conical hats"}]
|
00000002839.jpg
|
2839_evjvqa
|
[{"from": "human", "value": "where are the people standing?\n<image>"}, {"from": "gpt", "value": "in the market"}]
|
00000002839.jpg
|
2839_evjvqa
|
[{"from": "human", "value": "where is this?\n<image>"}, {"from": "gpt", "value": "this is the market"}]
|
00000002840.jpg
|
2840_evjvqa
|
[{"from": "human", "value": "where is the man holding the phone standing to take a photo?\n<image>"}, {"from": "gpt", "value": "near a display of apricot trees"}]
|
00000002840.jpg
|
2840_evjvqa
|
[{"from": "human", "value": "what is the man standing near the display of apricot trees doing?\n<image>"}, {"from": "gpt", "value": "he is taking a photo"}]
|
00000002840.jpg
|
2840_evjvqa
|
[{"from": "human", "value": "how many clowns are performing standing near the man taking the photo?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002841.jpg
|
2841_evjvqa
|
[{"from": "human", "value": "which side of the rock is the group of three standing on?\n<image>"}, {"from": "gpt", "value": "on the left"}]
|
00000002841.jpg
|
2841_evjvqa
|
[{"from": "human", "value": "which side of the rock is the group of four standing on?\n<image>"}, {"from": "gpt", "value": "on the right"}]
|
00000002845.jpg
|
2845_evjvqa
|
[{"from": "human", "value": "wher is this?\n<image>"}, {"from": "gpt", "value": "this is the market"}]
|
00000002847.jpg
|
2847_evjvqa
|
[{"from": "human", "value": "what are the scenes around the citadel?\n<image>"}, {"from": "gpt", "value": "the flower gardens"}]
|
00000002853.jpg
|
2853_evjvqa
|
[{"from": "human", "value": "what is the woman wearing the conical hat doing?\n<image>"}, {"from": "gpt", "value": "she is trying to push a trolley"}]
|
00000002853.jpg
|
2853_evjvqa
|
[{"from": "human", "value": "what is the woman wearing the conical hat pushing?\n<image>"}, {"from": "gpt", "value": "the trolley"}]
|
00000002853.jpg
|
2853_evjvqa
|
[{"from": "human", "value": "what is the woman standing on the side of the road holding in her hand?\n<image>"}, {"from": "gpt", "value": "a bunch of vegetables"}]
|
00000002853.jpg
|
2853_evjvqa
|
[{"from": "human", "value": "what is the woman standing on the side of the road sale?\n<image>"}, {"from": "gpt", "value": "vegetable"}]
|
00000002853.jpg
|
2853_evjvqa
|
[{"from": "human", "value": "what is the woman on the side of the road doing?\n<image>"}, {"from": "gpt", "value": "preparing products for sale"}]
|
00000002853.jpg
|
2853_evjvqa
|
[{"from": "human", "value": "how many people are parking on the curb?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000002855.jpg
|
2855_evjvqa
|
[{"from": "human", "value": "what are these two men doing?\n<image>"}, {"from": "gpt", "value": "checking products"}]
|
00000002855.jpg
|
2855_evjvqa
|
[{"from": "human", "value": "how do the staff here wear uniforms?\n<image>"}, {"from": "gpt", "value": "they are wearing a uniform with an apron over it"}]
|
00000002855.jpg
|
2855_evjvqa
|
[{"from": "human", "value": "where do these two men work?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}]
|
00000002861.jpg
|
2861_evjvqa
|
[{"from": "human", "value": "where is the woman in the blue shirt with the conical hat?\n<image>"}, {"from": "gpt", "value": "on the handlebars of the bicycle"}]
|
00000002861.jpg
|
2861_evjvqa
|
[{"from": "human", "value": "what means of transport does the woman in the blue shirt travel by?\n<image>"}, {"from": "gpt", "value": "the bicycle"}]
|
00000002861.jpg
|
2861_evjvqa
|
[{"from": "human", "value": "what is the woman in blue doing?\n<image>"}, {"from": "gpt", "value": "she is riding on the road"}]
|
00000002861.jpg
|
2861_evjvqa
|
[{"from": "human", "value": "how many motorbikes are there behind the woman in the blue shirt?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000002861.jpg
|
2861_evjvqa
|
[{"from": "human", "value": "how many people on the moving motorbike behind the woman riding the bicycle?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002864.jpg
|
2864_evjvqa
|
[{"from": "human", "value": "where is the citadel built?\n<image>"}, {"from": "gpt", "value": "next to the riverbank"}]
|
00000002865.jpg
|
2865_evjvqa
|
[{"from": "human", "value": "which direction is the man driving the car going?\n<image>"}, {"from": "gpt", "value": "he is running away from the historical monument"}]
|
00000002870.jpg
|
2870_evjvqa
|
[{"from": "human", "value": "what are two girls standing in the street doing?\n<image>"}, {"from": "gpt", "value": "they are posing to take a photo"}]
|
00000002870.jpg
|
2870_evjvqa
|
[{"from": "human", "value": "what color shirt is the man sitting on the side of the road wearing?\n<image>"}, {"from": "gpt", "value": "blue"}]
|
00000002870.jpg
|
2870_evjvqa
|
[{"from": "human", "value": "how many unoccupied roadside seats are there?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002882.jpg
|
2882_evjvqa
|
[{"from": "human", "value": "where are the two people in conical hats going?\n<image>"}, {"from": "gpt", "value": "in the middle of the field"}]
|
00000002882.jpg
|
2882_evjvqa
|
[{"from": "human", "value": "what means of transport do two people wearing conical hats use?\n<image>"}, {"from": "gpt", "value": "the one is riding a buffalo and the one is riding a bicycle"}]
|
00000002882.jpg
|
2882_evjvqa
|
[{"from": "human", "value": "how many buffaloes does the person riding on a buffalo bring to the field?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002884.jpg
|
2884_evjvqa
|
[{"from": "human", "value": "where is the woman wearing the conical hat sitting?\n<image>"}, {"from": "gpt", "value": "on the basket"}]
|
00000002884.jpg
|
2884_evjvqa
|
[{"from": "human", "value": "what color of shirt is the girl in the conical hat wearing?\n<image>"}, {"from": "gpt", "value": "blue with white polka dot"}]
|
00000002884.jpg
|
2884_evjvqa
|
[{"from": "human", "value": "who is the girl in conical hat sitting with?\n<image>"}, {"from": "gpt", "value": "no one"}]
|
00000002885.jpg
|
2885_evjvqa
|
[{"from": "human", "value": "who is the woman choosing with?\n<image>"}, {"from": "gpt", "value": "the man is wearing the shirt"}]
|
00000002885.jpg
|
2885_evjvqa
|
[{"from": "human", "value": "which hand does the man wear the watch on?\n<image>"}, {"from": "gpt", "value": "on the left hand"}]
|
00000002885.jpg
|
2885_evjvqa
|
[{"from": "human", "value": "how many people are choosing products?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000002889.jpg
|
2889_evjvqa
|
[{"from": "human", "value": "how many people are treating the patient?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000002889.jpg
|
2889_evjvqa
|
[{"from": "human", "value": "where does the doctor sit?\n<image>"}, {"from": "gpt", "value": "between two nurses"}]
|
00000002889.jpg
|
2889_evjvqa
|
[{"from": "human", "value": "how many people come to have medical examination?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000002892.jpg
|
2892_evjvqa
|
[{"from": "human", "value": "where are the lanterns hung?\n<image>"}, {"from": "gpt", "value": "in front of a house, next to the table"}]
|
00000002897.jpg
|
2897_evjvqa
|
[{"from": "human", "value": "what is this woman wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing the ao dai and the conical hat"}]
|
00000001206.jpg
|
1206_evjvqa
|
[{"from": "human", "value": "what is girl wearing helmet doing?\n<image>"}, {"from": "gpt", "value": "she is condisering types of milk"}]
|
00000001206.jpg
|
1206_evjvqa
|
[{"from": "human", "value": "how many people are standing at stall to buy milk?\n<image>"}, {"from": "gpt", "value": "there are two"}]
|
00000001206.jpg
|
1206_evjvqa
|
[{"from": "human", "value": "what is the stall behind two woman used for?\n<image>"}, {"from": "gpt", "value": "it is used for selling vegetable"}]
|
00000001206.jpg
|
1206_evjvqa
|
[{"from": "human", "value": "what is the stall next to milk stall sell?\n<image>"}, {"from": "gpt", "value": "this stall sells biscuit"}]
|
00000001208.jpg
|
1208_evjvqa
|
[{"from": "human", "value": "how many motorbikes are parked in front of the \"vai ky\" store?\n<image>"}, {"from": "gpt", "value": "there are two"}]
|
00000001208.jpg
|
1208_evjvqa
|
[{"from": "human", "value": "how many motorbikes are parked in front of the store next to vai ky store?\n<image>"}, {"from": "gpt", "value": "there is just one"}]
|
00000001213.jpg
|
1213_evjvqa
|
[{"from": "human", "value": "where does the girl sit?\n<image>"}, {"from": "gpt", "value": "on the footpath"}]
|
00000001216.jpg
|
1216_evjvqa
|
[{"from": "human", "value": "how many people are riding in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "there are two"}]
|
00000001216.jpg
|
1216_evjvqa
|
[{"from": "human", "value": "where does the man in red t-shirt sit?\n<image>"}, {"from": "gpt", "value": "in front of the restaurant"}]
|
00000001216.jpg
|
1216_evjvqa
|
[{"from": "human", "value": "how many people are sitting in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000001216.jpg
|
1216_evjvqa
|
[{"from": "human", "value": "how many motorbike are being ridden in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000001218.jpg
|
1218_evjvqa
|
[{"from": "human", "value": "how many girls are taking a seat with the boy?\n<image>"}, {"from": "gpt", "value": "five"}]
|
00000001218.jpg
|
1218_evjvqa
|
[{"from": "human", "value": "how many girls are sitting around the boy? \n<image>"}, {"from": "gpt", "value": "there are five"}]
|
00000001220.jpg
|
1220_evjvqa
|
[{"from": "human", "value": "what are people doing?\n<image>"}, {"from": "gpt", "value": "they are holding a festival"}]
|
00000001222.jpg
|
1222_evjvqa
|
[{"from": "human", "value": "what is a man in white shirt holding?\n<image>"}, {"from": "gpt", "value": "bunch of incense"}]
|
00000001222.jpg
|
1222_evjvqa
|
[{"from": "human", "value": "where is a man in whie shirt standing?\n<image>"}, {"from": "gpt", "value": "incense tall"}]
|
00000001222.jpg
|
1222_evjvqa
|
[{"from": "human", "value": "what is the man in white shirt doing?\n<image>"}, {"from": "gpt", "value": "he is buying incense"}]
|
00000001223.jpg
|
1223_evjvqa
|
[{"from": "human", "value": "how many serving of food on the table?\n<image>"}, {"from": "gpt", "value": "five "}]
|
00000001224.jpg
|
1224_evjvqa
|
[{"from": "human", "value": "how many people stepping on the platform?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000001224.jpg
|
1224_evjvqa
|
[{"from": "human", "value": "what is the man in blue shirt doing?\n<image>"}, {"from": "gpt", "value": "he is standing on the podium and holding a bunch of flowers"}]
|
00000001224.jpg
|
1224_evjvqa
|
[{"from": "human", "value": "what is the man in blue shirt holding?\n<image>"}, {"from": "gpt", "value": "a bunch of flowers"}]
|
00000001225.jpg
|
1225_evjvqa
|
[{"from": "human", "value": "where does the girl sit?\n<image>"}, {"from": "gpt", "value": "on the stairs"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.