Datasets:
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Error code: UnexpectedError
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
id int64 | image image | image_id string | question_id string | question string | answers list | answer string | image_emb list | question_emb list | full_answer string | structural string | semantic string | detailed string | is_balanced bool | group_global string | group_local string | semantic_str string |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0 | n161313 | 201307251 | Is it overcast? | [
"no"
] | no | [
-0.021881103515625,
0.0168609619140625,
-0.0750732421875,
0.043304443359375,
-0.02545166015625,
0.035308837890625,
0.0210723876953125,
0.0193634033203125,
-0.0220184326171875,
-0.00740814208984375,
-0.001155853271484375,
-0.01430511474609375,
0.032012939453125,
0.03692626953125,
-0.03805... | [
0.018402099609375,
-0.0204010009765625,
-0.01177978515625,
-0.01300811767578125,
0.015228271484375,
0.00621795654296875,
0.026947021484375,
-0.03546142578125,
0.06390380859375,
0.0009732246398925781,
-0.03668212890625,
0.009552001953125,
0.0032100677490234375,
-0.0203399658203125,
0.0171... | No, it is clear. | verify | global | weatherVerifyC | true | null | 01-weather_overcast | select: scene->verify weather: overcast [0] | |
1 | n235859 | 201640614 | Who is wearing the dress? | [
"women"
] | women | [
0.013397216796875,
0.171630859375,
-0.03179931640625,
0.05987548828125,
-0.03289794921875,
0.0170440673828125,
0.01415252685546875,
-0.0032329559326171875,
-0.019287109375,
-0.018524169921875,
0.002857208251953125,
-0.06256103515625,
0.02215576171875,
0.005126953125,
-0.08697509765625,
... | [
-0.004638671875,
0.003322601318359375,
-0.0136871337890625,
-0.09173583984375,
-0.05181884765625,
0.01134490966796875,
-0.004669189453125,
0.0074920654296875,
-0.0069427490234375,
0.00998687744140625,
0.002410888671875,
-0.0631103515625,
-0.0166778564453125,
0.01139068603515625,
-0.02133... | The women are wearing a dress. | query | rel | relS | true | person | 14-dress_wearing,s | select: dress (12)->relate: person,wearing,s (7) [0]->query: name [1] | |
2 | n336443 | 202225914 | Does the utensil on top of the table look clean and black? | [
"no"
] | no | [
0.012481689453125,
0.0758056640625,
-0.0103607177734375,
0.032135009765625,
-0.05938720703125,
-0.008056640625,
-0.031463623046875,
-0.027679443359375,
0.022125244140625,
-0.02008056640625,
-0.0465087890625,
-0.0312042236328125,
-0.01457977294921875,
-0.0733642578125,
-0.0086746215820312... | [
-0.0236663818359375,
0.01311492919921875,
0.060089111328125,
0.01416015625,
-0.0194854736328125,
-0.01512908935546875,
-0.0242156982421875,
0.023345947265625,
-0.00415802001953125,
-0.02154541015625,
-0.0131378173828125,
0.00818634033203125,
0.006011962890625,
-0.01012420654296875,
0.024... | No, the fork is clean but silver. | logical | attr | verifyAttrsC | true | null | 05-black_clean | select: table (2)->relate: utensil,on top of,s (1) [0]->verify color: black [1]->verify cleanliness: clean [1]->and: [2, 3] | |
3 | n179136 | 2062325 | Is the surfer that looks wet wearing a wetsuit? | [
"yes"
] | yes | [
-0.0301666259765625,
0.0020542144775390625,
-0.061920166015625,
0.0167999267578125,
-0.01532745361328125,
0.007541656494140625,
0.0367431640625,
-0.0122222900390625,
0.009033203125,
-0.0303497314453125,
-0.00370025634765625,
0.033843994140625,
0.0007433891296386719,
0.03125,
-0.084716796... | [
0.00952911376953125,
-0.0228118896484375,
-0.02618408203125,
0.028411865234375,
0.0396728515625,
-0.0007081031799316406,
0.01004791259765625,
-0.0144805908203125,
0.0004684925079345703,
-0.0343017578125,
0.006580352783203125,
0.01305389404296875,
0.032318115234375,
-0.006763458251953125,
... | Yes, the surfer is wearing a wetsuit. | verify | rel | relVerify | true | null | 13-surfer_wetsuit | select: surfer (1)->filter: wet [0]->verify rel: wetsuit,wearing,o (12) [1] | |
4 | n518912 | 201303229 | How tall is the chair in the bottom of the photo? | [
"short"
] | short | [
0.0380859375,
0.14599609375,
-0.0806884765625,
0.0094146728515625,
-0.0220794677734375,
0.0012502670288085938,
0.037872314453125,
0.01253509521484375,
-0.0220184326171875,
-0.00823974609375,
0.011138916015625,
-0.0269927978515625,
-0.0364990234375,
-0.0266265869140625,
-0.028823852539062... | [
-0.0193328857421875,
0.0098876953125,
0.0311126708984375,
0.003345489501953125,
-0.052215576171875,
0.0303802490234375,
-0.004364013671875,
-0.043853759765625,
-0.04730224609375,
0.0182037353515625,
0.03607177734375,
0.00916290283203125,
-0.0550537109375,
-0.01111602783203125,
-0.0178222... | The chair is short. | query | attr | how | true | height | 10q-chair_height | select: chair (13)->filter vposition: bottom [0]->query: height [1] | |
5 | n435808 | 201902997 | What kind of device is on top of the desk? | [
"keyboard"
] | keyboard | [
0.01494598388671875,
-0.00305938720703125,
0.00986480712890625,
-0.05340576171875,
-0.01934814453125,
-0.01432037353515625,
0.0200653076171875,
-0.01245880126953125,
0.0183868408203125,
0.07635498046875,
0.06390380859375,
-0.02386474609375,
0.0147705078125,
0.0006957054138183594,
-0.0150... | [
0.004451751708984375,
-0.035430908203125,
0.053375244140625,
-0.003467559814453125,
-0.0460205078125,
0.0277557373046875,
0.00983428955078125,
-0.02911376953125,
0.01494598388671875,
-0.04339599609375,
-0.02197265625,
-0.005401611328125,
0.0094757080078125,
-0.0140380859375,
-0.007156372... | The device is a keyboard. | query | rel | categoryRelS | true | device | 15-desk_on top of,s | select: desk (1)->relate: device,on top of,s (8) [0]->query: name [1] | |
6 | n414992 | 20567512 | What is the airplane flying above? | [
"ocean"
] | ocean | [
-0.0275421142578125,
0.09051513671875,
-0.17822265625,
0.070556640625,
-0.0787353515625,
0.0149993896484375,
-0.051788330078125,
0.0085906982421875,
0.0028209686279296875,
-0.032440185546875,
0.0208892822265625,
-0.01529693603515625,
0.0277252197265625,
-0.041107177734375,
-0.03485107421... | [
-0.01715087890625,
-0.01488494873046875,
-0.038238525390625,
-0.027374267578125,
-0.0325927734375,
0.023651123046875,
-0.03680419921875,
-0.0203094482421875,
-0.0272369384765625,
-0.0521240234375,
-0.0185699462890625,
0.00521087646484375,
0.055450439453125,
-0.032318115234375,
-0.0149230... | The plane is flying above the ocean. | query | rel | relO | true | place | 14-airplane_flying above,o | select: airplane (11)->relate: _,flying above,o (10) [0]->query: name [1] | |
7 | n446242 | 20136592 | What color are the pants? | [
"red"
] | red | [
0.0308837890625,
0.198486328125,
0.0758056640625,
0.0160369873046875,
0.0158233642578125,
-0.0269775390625,
0.01073455810546875,
-0.00423431396484375,
-0.0180816650390625,
0.002712249755859375,
-0.021087646484375,
-0.0233154296875,
-0.01203155517578125,
0.034271240234375,
-0.024963378906... | [
0.0300445556640625,
-0.01073455810546875,
-0.03863525390625,
-0.003597259521484375,
-0.0306549072265625,
0.01178741455078125,
0.0158843994140625,
0.0308837890625,
-0.01233673095703125,
-0.01116180419921875,
-0.02392578125,
-0.000029742717742919922,
-0.0310516357421875,
-0.026702880859375,
... | The pants are red. | query | attr | directOf | true | color | 10q-pants_color | select: pants (3)->query: color [0] | |
8 | n168412 | 20602803 | Is the ground blue or brown? | [
"brown"
] | brown | [
-0.005077362060546875,
0.1363525390625,
-0.05694580078125,
0.08392333984375,
0.0445556640625,
0.01531219482421875,
0.02276611328125,
0.039093017578125,
0.0196533203125,
0.0034694671630859375,
-0.00067138671875,
-0.026611328125,
0.05462646484375,
-0.06671142578125,
0.04638671875,
-0.005... | [
0.0061187744140625,
-0.016815185546875,
0.0018777847290039062,
-0.009002685546875,
-0.055084228515625,
0.052886962890625,
0.006160736083984375,
-0.0189056396484375,
-0.0312347412109375,
0.031036376953125,
-0.0016374588012695312,
0.023468017578125,
0.0149078369140625,
0.0161285400390625,
... | The ground is brown. | choose | attr | chooseAttr | true | color | 10c-ground_color | select: ground (10)->choose color: brown|blue [0] | |
9 | n23181 | 201079951 | What is around the open window? | [
"drapes"
] | drapes | [0.05731201171875,0.112548828125,0.00167083740234375,0.0182647705078125,-0.033447265625,-0.050964355(...TRUNCATED) | [-0.01108551025390625,0.01274871826171875,0.00812530517578125,-0.0206756591796875,-0.023788452148437(...TRUNCATED) | The draperies are around the window. | query | rel | relS | true | textile | 14-window_around,s | select: window (0)->filter: open [0]->relate: _,around,s (12) [1]->query: name [2] |
GQA testdev-balanced (Lance Format)
Lance-formatted version of the canonical GQA testdev_balanced slice — 12,578 compositional VQA questions joined with the matching 398 images — sourced from lmms-lab/GQA.
lmms-lab/GQA exposes instructions and images as separate parquet configs; this Lance dataset joins them on imageId, so each row has the question, the answer, the GQA reasoning-program tags, and the image bytes inline.
Splits
| Split | Rows | Distinct images |
|---|---|---|
testdev.lance |
12,578 | 398 |
Train (
train_balanced_instructions×train_balanced_images, ~943k Q's × 72k images, ~10 GB images) and val splits are not bundled by default — pass--instr-config/--images-configtogqa/dataprep.pyto extend.
Schema
| Column | Type | Notes |
|---|---|---|
id |
int64 |
Row index |
image |
large_binary |
Inline JPEG bytes (image is duplicated across rows that share an image_id) |
image_id |
string |
GQA scene-graph image id |
question_id |
string |
GQA question id |
question |
string |
Compositional natural-language question |
answers |
list<string> |
One-element list (the GQA short answer) |
answer |
string |
Same short answer (canonical / FTS target) |
full_answer |
string? |
Full sentence answer |
structural |
string? |
One of verify, query, compare, choose, logical |
semantic |
string? |
One of attr, cat, global, obj, rel |
detailed |
string? |
Fine-grained type (e.g. weatherVerifyC) |
is_balanced |
bool |
GQA balanced subset flag |
group_global / group_local |
string? |
GQA reasoning-group ids |
semantic_str |
string? |
Compact description of the reasoning program |
image_emb |
fixed_size_list<float32, 512> |
CLIP image embedding (cosine-normalized) |
question_emb |
fixed_size_list<float32, 512> |
CLIP text embedding of the question |
Pre-built indices
IVF_PQonimage_embandquestion_emb—metric=cosineINVERTED(FTS) onquestionandanswerBITMAPonstructural,semantic,detailedBTREEonimage_id,question_id
Quick start
import lance
ds = lance.dataset("hf://datasets/lance-format/gqa-testdev-balanced-lance/data/testdev.lance")
print(ds.count_rows(), ds.schema.names, ds.list_indices())
Load with LanceDB
These tables can also be consumed by LanceDB, the multimodal lakehouse and embedded search library built on top of Lance, for simplified vector search and other queries.
import lancedb
db = lancedb.connect("hf://datasets/lance-format/gqa-testdev-balanced-lance/data")
tbl = db.open_table("testdev")
print(f"LanceDB table opened with {len(tbl)} image-question pairs")
LanceDB vector search
import lancedb
db = lancedb.connect("hf://datasets/lance-format/gqa-testdev-balanced-lance/data")
tbl = db.open_table("testdev")
ref = tbl.search().limit(1).select(["question_emb", "question"]).to_list()[0]
query_embedding = ref["question_emb"]
results = (
tbl.search(query_embedding, vector_column_name="question_emb")
.metric("cosine")
.select(["question", "answer"])
.limit(5)
.to_list()
)
LanceDB full-text search
import lancedb
db = lancedb.connect("hf://datasets/lance-format/gqa-testdev-balanced-lance/data")
tbl = db.open_table("testdev")
results = (
tbl.search("color of the car")
.select(["question", "answer"])
.limit(10)
.to_list()
)
Filter by reasoning type
import lance
ds = lance.dataset("hf://datasets/lance-format/gqa-testdev-balanced-lance/data/testdev.lance")
verify_qs = ds.scanner(filter="structural = 'verify'", columns=["question", "answer"], limit=5).to_table()
Filter with LanceDB
import lancedb
db = lancedb.connect("hf://datasets/lance-format/gqa-testdev-balanced-lance/data")
tbl = db.open_table("testdev")
verify_qs = (
tbl.search()
.where("structural = 'verify'")
.select(["question", "answer"])
.limit(5)
.to_list()
)
Why Lance?
- One dataset for the joined image + question + answer + reasoning-program metadata + dual embeddings + indices — no instructions/images parquet split to keep in sync.
- Schema evolution: add columns (alternate scene graphs, model predictions) without rewriting the data.
Source & license
Converted from lmms-lab/GQA. GQA is released under CC BY 4.0 by Hudson and Manning (Stanford NLP).
Citation
@inproceedings{hudson2019gqa,
title={GQA: A New Dataset for Real-World Visual Reasoning and Compositional Question Answering},
author={Hudson, Drew A. and Manning, Christopher D.},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year={2019}
}
- Downloads last month
- 48