image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000001305.jpg
|
1305_evjvqa
|
[{"from": "human", "value": "what is the woman in the black dress doing with the stroller?\n<image>"}, {"from": "gpt", "value": "to shopping"}]
|
00000001305.jpg
|
1305_evjvqa
|
[{"from": "human", "value": "where is the woman in the black dress for the chosen product?\n<image>"}, {"from": "gpt", "value": "in the stroller"}]
|
00000001312.jpg
|
1312_evjvqa
|
[{"from": "human", "value": "how many cars are on this road?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000001312.jpg
|
1312_evjvqa
|
[{"from": "human", "value": "how many directions is this road?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000001312.jpg
|
1312_evjvqa
|
[{"from": "human", "value": "how many directions can vehicles travel on this route?\n<image>"}, {"from": "gpt", "value": "just one"}]
|
00000001313.jpg
|
1313_evjvqa
|
[{"from": "human", "value": "why some people hold the umbrellas?\n<image>"}, {"from": "gpt", "value": "to sundshade"}]
|
00000001313.jpg
|
1313_evjvqa
|
[{"from": "human", "value": "how many people are wearing conical hat?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000001313.jpg
|
1313_evjvqa
|
[{"from": "human", "value": "what color umbrella is the person standing on the right side holding?\n<image>"}, {"from": "gpt", "value": "red"}]
|
00000001318.jpg
|
1318_evjvqa
|
[{"from": "human", "value": "what is the woman wearing conical hat doing?\n<image>"}, {"from": "gpt", "value": "she sitting and scooping food into bowls for the customers"}]
|
00000001318.jpg
|
1318_evjvqa
|
[{"from": "human", "value": "what is the woman wearing conical hat holding?\n<image>"}, {"from": "gpt", "value": "the ladle and the bowl"}]
|
00000001318.jpg
|
1318_evjvqa
|
[{"from": "human", "value": "what is the woman who is scooping food into the bowl for the customers wearing?\n<image>"}, {"from": "gpt", "value": "the conical hat"}]
|
00000001322.jpg
|
1322_evjvqa
|
[{"from": "human", "value": "what is the man behind the podium for?\n<image>"}, {"from": "gpt", "value": "he is speaking at the conference"}]
|
00000001322.jpg
|
1322_evjvqa
|
[{"from": "human", "value": "what do the speakers wearing at the conference wear?\n<image>"}, {"from": "gpt", "value": "vest and tie"}]
|
00000001328.jpg
|
1328_evjvqa
|
[{"from": "human", "value": "where is the shopping activity taken place?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}]
|
00000001328.jpg
|
1328_evjvqa
|
[{"from": "human", "value": "what is the man in sky blue shirt doing at the stall?\n<image>"}, {"from": "gpt", "value": "buying dragon fruits"}]
|
00000001343.jpg
|
1343_evjvqa
|
[{"from": "human", "value": "how do these two women move in the flower garden?\n<image>"}, {"from": "gpt", "value": "the boat"}]
|
00000001343.jpg
|
1343_evjvqa
|
[{"from": "human", "value": "who are on the boat in the middle of the flower garden?\n<image>"}, {"from": "gpt", "value": "two women are wearing conical hats"}]
|
00000001343.jpg
|
1343_evjvqa
|
[{"from": "human", "value": "where are these two women sailing?\n<image>"}, {"from": "gpt", "value": "in the middle of the flower garden"}]
|
00000001354.jpg
|
1354_evjvqa
|
[{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "she is surfing"}]
|
00000001354.jpg
|
1354_evjvqa
|
[{"from": "human", "value": "what sport is the girl playing?\n<image>"}, {"from": "gpt", "value": "surfing"}]
|
00000001354.jpg
|
1354_evjvqa
|
[{"from": "human", "value": "what does this girl surf with?\n<image>"}, {"from": "gpt", "value": "the surfboard"}]
|
00000001357.jpg
|
1357_evjvqa
|
[{"from": "human", "value": "what is the girl in the light blue pants leaning on?\n<image>"}, {"from": "gpt", "value": "the bus model"}]
|
00000001357.jpg
|
1357_evjvqa
|
[{"from": "human", "value": "what is the girl standing on the bus model carrying?\n<image>"}, {"from": "gpt", "value": "the crossbody bag"}]
|
00000001357.jpg
|
1357_evjvqa
|
[{"from": "human", "value": "what color is the bus model painted in?\n<image>"}, {"from": "gpt", "value": "sky blue"}]
|
00000001365.jpg
|
1365_evjvqa
|
[{"from": "human", "value": "what does the discount sign look like?\n<image>"}, {"from": "gpt", "value": "it has rectangle shape, pink background, black border and white text"}]
|
00000001383.jpg
|
1383_evjvqa
|
[{"from": "human", "value": "who is the girl in white standing with?\n<image>"}, {"from": "gpt", "value": "nobody"}]
|
00000001383.jpg
|
1383_evjvqa
|
[{"from": "human", "value": "what does the girl standing look like?\n<image>"}, {"from": "gpt", "value": "she is wearing white shirt, blue dress, white shoes and wide-brimmed hat"}]
|
00000001383.jpg
|
1383_evjvqa
|
[{"from": "human", "value": "what is the place behind the girl?\n<image>"}, {"from": "gpt", "value": "cau bridge, hoi an"}]
|
00000001388.jpg
|
1388_evjvqa
|
[{"from": "human", "value": "what is the man sitting next to the pillar going to do?\n<image>"}, {"from": "gpt", "value": "smoking"}]
|
00000001392.jpg
|
1392_evjvqa
|
[{"from": "human", "value": "what are the main items this store sells?\n<image>"}, {"from": "gpt", "value": "type os shoes"}]
|
00000001392.jpg
|
1392_evjvqa
|
[{"from": "human", "value": "what is the woman with the handbag sitting on?\n<image>"}, {"from": "gpt", "value": "the red chair"}]
|
00000001392.jpg
|
1392_evjvqa
|
[{"from": "human", "value": "where can the red chair be found?\n<image>"}, {"from": "gpt", "value": "it is seated by the woman with the handbag"}]
|
00000001392.jpg
|
1392_evjvqa
|
[{"from": "human", "value": "what is the woman in the red chair wearing?\n<image>"}, {"from": "gpt", "value": "the handbag"}]
|
00000001393.jpg
|
1393_evjvqa
|
[{"from": "human", "value": "who is speaking at the event?\n<image>"}, {"from": "gpt", "value": "the woman is in white, holding microphone"}]
|
00000001393.jpg
|
1393_evjvqa
|
[{"from": "human", "value": "what is the person on the right of the woman in the white shirt holding?\n<image>"}, {"from": "gpt", "value": "bunch of flowers"}]
|
00000001394.jpg
|
1394_evjvqa
|
[{"from": "human", "value": "how do people go to the market?\n<image>"}, {"from": "gpt", "value": "walking or riding a motorbike"}]
|
00000001394.jpg
|
1394_evjvqa
|
[{"from": "human", "value": "what kind of goods do the stalls where the man in the black cap sells?\n<image>"}, {"from": "gpt", "value": "raw fish"}]
|
00000001397.jpg
|
1397_evjvqa
|
[{"from": "human", "value": "how many lanes does this road have?\n<image>"}, {"from": "gpt", "value": "four"}]
|
00000001397.jpg
|
1397_evjvqa
|
[{"from": "human", "value": "which lane has the most cars running?\n<image>"}, {"from": "gpt", "value": "the rightmost lane"}]
|
00000001397.jpg
|
1397_evjvqa
|
[{"from": "human", "value": "which lane can continue going straight?\n<image>"}, {"from": "gpt", "value": "two lanes in the middle"}]
|
00000001397.jpg
|
1397_evjvqa
|
[{"from": "human", "value": "which lane is the manhole cover on?\n<image>"}, {"from": "gpt", "value": "the second lane from left"}]
|
00000003903.jpg
|
3903_evjvqa
|
[{"from": "human", "value": "what is the man in gray shirt raising hand for?\n<image>"}, {"from": "gpt", "value": "he is raising hand for greeting the people in the isolation area"}]
|
00000003903.jpg
|
3903_evjvqa
|
[{"from": "human", "value": "what hat is the woman in pink shirt wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing a purple wide-brimmed hat"}]
|
00000003903.jpg
|
3903_evjvqa
|
[{"from": "human", "value": "who is the one raising hand to greet the people in the isolation area?\n<image>"}, {"from": "gpt", "value": "the man wearing the gray long-sleeve shirt"}]
|
00000003904.jpg
|
3904_evjvqa
|
[{"from": "human", "value": "where is the woman standing?\n<image>"}, {"from": "gpt", "value": "she is standing at the door"}]
|
00000003904.jpg
|
3904_evjvqa
|
[{"from": "human", "value": "what is the weather like here?\n<image>"}, {"from": "gpt", "value": "it is sunny here"}]
|
00000003904.jpg
|
3904_evjvqa
|
[{"from": "human", "value": "what animals are on the yard?\n<image>"}, {"from": "gpt", "value": "chickens"}]
|
00000003905.jpg
|
3905_evjvqa
|
[{"from": "human", "value": "who is sitting next to the young man in blue coat?\n<image>"}, {"from": "gpt", "value": "sitting next to the young man in blue coat is a person in red coat"}]
|
00000003905.jpg
|
3905_evjvqa
|
[{"from": "human", "value": "how many people are sitting on the stone bench?\n<image>"}, {"from": "gpt", "value": "on the right stone bench, someone is sitting"}]
|
00000003905.jpg
|
3905_evjvqa
|
[{"from": "human", "value": "who are sitting on the right stone bench?\n<image>"}, {"from": "gpt", "value": "a person in blue shirt and a person in red shirt are sitting on the right stone bench"}]
|
00000003907.jpg
|
3907_evjvqa
|
[{"from": "human", "value": "what color is the face mask that the woman standing wears?\n<image>"}, {"from": "gpt", "value": "she wears a white face mask"}]
|
00000003909.jpg
|
3909_evjvqa
|
[{"from": "human", "value": "what is the shape of the flag?\n<image>"}, {"from": "gpt", "value": "the flag has a red background and a yellow star"}]
|
00000003909.jpg
|
3909_evjvqa
|
[{"from": "human", "value": "where the conical hat lies?\n<image>"}, {"from": "gpt", "value": "the conical hat lies on the cargo cart"}]
|
00000003909.jpg
|
3909_evjvqa
|
[{"from": "human", "value": "on which side of the road is the electric pole located?\n<image>"}, {"from": "gpt", "value": "the electric pole is located on the left side of the road"}]
|
00000003910.jpg
|
3910_evjvqa
|
[{"from": "human", "value": "what is the woman in blue coat wearing in her hand?\n<image>"}, {"from": "gpt", "value": "she is wearing a red bag"}]
|
00000003910.jpg
|
3910_evjvqa
|
[{"from": "human", "value": "what hat does the woman in blue coat wear?\n<image>"}, {"from": "gpt", "value": "a conical hat"}]
|
00000003911.jpg
|
3911_evjvqa
|
[{"from": "human", "value": "what color are the two balloons?\n<image>"}, {"from": "gpt", "value": "a balloon is orange, the other is yellow"}]
|
00000003911.jpg
|
3911_evjvqa
|
[{"from": "human", "value": "who is standing to the left of the woman in blue t-shirt?\n<image>"}, {"from": "gpt", "value": "standing to the left of the woman in the blue t-shirt is the woman in black outfit"}]
|
00000003911.jpg
|
3911_evjvqa
|
[{"from": "human", "value": "who is standing to the left of the woman in black shirt?\n<image>"}, {"from": "gpt", "value": "standing to the left of the woman in black shirt is the little girl in white hat"}]
|
00000003915.jpg
|
3915_evjvqa
|
[{"from": "human", "value": "what color is the shirt that the woman holding a red bag is wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing a purple t-shirt"}]
|
00000003915.jpg
|
3915_evjvqa
|
[{"from": "human", "value": "what is the man in gray shirt pushing?\n<image>"}, {"from": "gpt", "value": "he is pushing a baby stroller"}]
|
00000003916.jpg
|
3916_evjvqa
|
[{"from": "human", "value": "who is selling directly to the woman in green shirt?\n<image>"}, {"from": "gpt", "value": "the woman in violet shirt"}]
|
00000003916.jpg
|
3916_evjvqa
|
[{"from": "human", "value": "what vehicle is behind the woman in green shirt?\n<image>"}, {"from": "gpt", "value": "a motorbike"}]
|
00000003916.jpg
|
3916_evjvqa
|
[{"from": "human", "value": "what is red hung on the scooter of the woman in black shirt?\n<image>"}, {"from": "gpt", "value": "there is a red helmet hung on the scooter of the woman in black shirt"}]
|
00000003919.jpg
|
3919_evjvqa
|
[{"from": "human", "value": "what color is the car parked on the yard?\n<image>"}, {"from": "gpt", "value": "the car parked on the yard is black"}]
|
00000003919.jpg
|
3919_evjvqa
|
[{"from": "human", "value": "what is the man in black coat doing?\n<image>"}, {"from": "gpt", "value": "he is looking at the exhibits"}]
|
00000003919.jpg
|
3919_evjvqa
|
[{"from": "human", "value": "what are on display on the table?\n<image>"}, {"from": "gpt", "value": "the ceramic wares"}]
|
00000003920.jpg
|
3920_evjvqa
|
[{"from": "human", "value": "what kinds of vehicles are moving on the road?\n<image>"}, {"from": "gpt", "value": "motorbikes and cars are moving on the road"}]
|
00000003920.jpg
|
3920_evjvqa
|
[{"from": "human", "value": "what is the sky like in this locality?\n<image>"}, {"from": "gpt", "value": "the sky is cloudy"}]
|
00000003922.jpg
|
3922_evjvqa
|
[{"from": "human", "value": "what color is the shirt that the woman wears?\n<image>"}, {"from": "gpt", "value": "the woman wears black shirt"}]
|
00000003922.jpg
|
3922_evjvqa
|
[{"from": "human", "value": "what kind of flower tree is the girl standing next to?\n<image>"}, {"from": "gpt", "value": "peach tree"}]
|
00000003922.jpg
|
3922_evjvqa
|
[{"from": "human", "value": "what shirt is the this girl wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing a black halter"}]
|
00000003924.jpg
|
3924_evjvqa
|
[{"from": "human", "value": "how many guests are sitting in this restaurant?\n<image>"}, {"from": "gpt", "value": "there are no guests"}]
|
00000003924.jpg
|
3924_evjvqa
|
[{"from": "human", "value": "where is the air conditioner located in this restaurant?\n<image>"}, {"from": "gpt", "value": "the air conditioner is on the left wall with the same height level as the loft railing"}]
|
00000003928.jpg
|
3928_evjvqa
|
[{"from": "human", "value": "what is the person in red shirt wearing on the hand?\n<image>"}, {"from": "gpt", "value": "that person is wearing rubber gloves"}]
|
00000003928.jpg
|
3928_evjvqa
|
[{"from": "human", "value": "what is the woman in black shirt wearing glasses doing?\n<image>"}, {"from": "gpt", "value": "she is selecting oranges"}]
|
00000003928.jpg
|
3928_evjvqa
|
[{"from": "human", "value": "what type of fruit is being taken by the woman in black shirt with glasses?\n<image>"}, {"from": "gpt", "value": "oranges"}]
|
00000003930.jpg
|
3930_evjvqa
|
[{"from": "human", "value": "what items are hung on the shelf behind the woman wearing a white face mask?\n<image>"}, {"from": "gpt", "value": "pans"}]
|
00000003930.jpg
|
3930_evjvqa
|
[{"from": "human", "value": "which side of the woman wearing white face mask does the woman wearing blue face mask stand?\n<image>"}, {"from": "gpt", "value": "the woman wearing blue face mask stands on the left side of the woman wearing white face mask"}]
|
00000003930.jpg
|
3930_evjvqa
|
[{"from": "human", "value": "what is the woman standing next to the woman wearing a blue face mask wearing on her face?\n<image>"}, {"from": "gpt", "value": "she is wearing a medical face mask and a face shield"}]
|
00000003933.jpg
|
3933_evjvqa
|
[{"from": "human", "value": "which side lane of the median has more traffic?\n<image>"}, {"from": "gpt", "value": "the right side lane of the median"}]
|
00000003933.jpg
|
3933_evjvqa
|
[{"from": "human", "value": "what is the weather like in this area?\n<image>"}, {"from": "gpt", "value": "it is sunny in this area"}]
|
00000003933.jpg
|
3933_evjvqa
|
[{"from": "human", "value": "which side lane of the median has only three moving motorbikes?\n<image>"}, {"from": "gpt", "value": "the left lane of the median"}]
|
00000003935.jpg
|
3935_evjvqa
|
[{"from": "human", "value": "what color is the sea water?\n<image>"}, {"from": "gpt", "value": "the sea water is blue"}]
|
00000003935.jpg
|
3935_evjvqa
|
[{"from": "human", "value": "what vehicles are moving on the sea?\n<image>"}, {"from": "gpt", "value": "the ships"}]
|
00000003935.jpg
|
3935_evjvqa
|
[{"from": "human", "value": "what are on the sea surface?\n<image>"}, {"from": "gpt", "value": "the ships are on the sea surface"}]
|
00000003936.jpg
|
3936_evjvqa
|
[{"from": "human", "value": "how many floors does the row of shelves in the middle of the room have?\n<image>"}, {"from": "gpt", "value": "the row of shelves in the middle of the room has four floors"}]
|
00000003937.jpg
|
3937_evjvqa
|
[{"from": "human", "value": "what are the people standing in between the girls in red ao dai doing?\n<image>"}, {"from": "gpt", "value": "they are cutting the ribbon"}]
|
00000003937.jpg
|
3937_evjvqa
|
[{"from": "human", "value": "who does each person cutting the ribbon stand between?\n<image>"}, {"from": "gpt", "value": "each person cutting the ribbon stands between the two girls in red ao dai"}]
|
00000003937.jpg
|
3937_evjvqa
|
[{"from": "human", "value": "what are the people who do not wear red ao dai doing?\n<image>"}, {"from": "gpt", "value": "the people who do not wear red ao dai is cutting the ribbon"}]
|
00000003938.jpg
|
3938_evjvqa
|
[{"from": "human", "value": "what color is the sweater that the young man wears?\n<image>"}, {"from": "gpt", "value": "the young man wears brown sweater"}]
|
00000003938.jpg
|
3938_evjvqa
|
[{"from": "human", "value": "who does the girl in black shirt stand next to?\n<image>"}, {"from": "gpt", "value": "the girl in black shirt stands next to the young man in brown sweater"}]
|
00000003939.jpg
|
3939_evjvqa
|
[{"from": "human", "value": "where is the woman standing?\n<image>"}, {"from": "gpt", "value": "the woman is standing by the bridge railing"}]
|
00000003939.jpg
|
3939_evjvqa
|
[{"from": "human", "value": "what does the woman wear?\n<image>"}, {"from": "gpt", "value": "the woman wears a red ao dai with a yellow star on the chest"}]
|
00000003939.jpg
|
3939_evjvqa
|
[{"from": "human", "value": "what is the woman in ao dai wearing on her face?\n<image>"}, {"from": "gpt", "value": "she is wearing glasses"}]
|
00000003940.jpg
|
3940_evjvqa
|
[{"from": "human", "value": "where can the green trash can be seen?\n<image>"}, {"from": "gpt", "value": "the green trash can lies next to the head of the orange truck"}]
|
00000003940.jpg
|
3940_evjvqa
|
[{"from": "human", "value": "what is the color of the car in the right front side of the brown car?\n<image>"}, {"from": "gpt", "value": "that car is white"}]
|
00000003941.jpg
|
3941_evjvqa
|
[{"from": "human", "value": "what color are the baubles on the christmas tree?\n<image>"}, {"from": "gpt", "value": "the baubles on the christmas tree are red"}]
|
00000003941.jpg
|
3941_evjvqa
|
[{"from": "human", "value": "how many gift boxes are there under the christmas tree?\n<image>"}, {"from": "gpt", "value": "there are two gift boxes under the christmas tree"}]
|
00000003941.jpg
|
3941_evjvqa
|
[{"from": "human", "value": "where are the red baubles attached?\n<image>"}, {"from": "gpt", "value": "the red baubles attached on the christmas tree"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.