Dataset Preview
Duplicate
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
500 Server Error: Internal Server Error for url: https://huggingface.co/api/datasets/lance-format/gqa-testdev-balanced-lance/refs (Request ID: Root=1-69fe065d-3f3de1967b2a35e6604c4b06;a1956e9d-7e83-40e4-9f95-0360ec2bb205) Internal Error - We're working hard to fix this as soon as possible!
Error code:   UnexpectedError

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

id
int64
image
image
image_id
string
question_id
string
question
string
answers
list
answer
string
image_emb
list
question_emb
list
full_answer
string
structural
string
semantic
string
detailed
string
is_balanced
bool
group_global
string
group_local
string
semantic_str
string
0
n161313
201307251
Is it overcast?
[ "no" ]
no
[ -0.021881103515625, 0.0168609619140625, -0.0750732421875, 0.043304443359375, -0.02545166015625, 0.035308837890625, 0.0210723876953125, 0.0193634033203125, -0.0220184326171875, -0.00740814208984375, -0.001155853271484375, -0.01430511474609375, 0.032012939453125, 0.03692626953125, -0.03805...
[ 0.018402099609375, -0.0204010009765625, -0.01177978515625, -0.01300811767578125, 0.015228271484375, 0.00621795654296875, 0.026947021484375, -0.03546142578125, 0.06390380859375, 0.0009732246398925781, -0.03668212890625, 0.009552001953125, 0.0032100677490234375, -0.0203399658203125, 0.0171...
No, it is clear.
verify
global
weatherVerifyC
true
null
01-weather_overcast
select: scene->verify weather: overcast [0]
1
n235859
201640614
Who is wearing the dress?
[ "women" ]
women
[ 0.013397216796875, 0.171630859375, -0.03179931640625, 0.05987548828125, -0.03289794921875, 0.0170440673828125, 0.01415252685546875, -0.0032329559326171875, -0.019287109375, -0.018524169921875, 0.002857208251953125, -0.06256103515625, 0.02215576171875, 0.005126953125, -0.08697509765625, ...
[ -0.004638671875, 0.003322601318359375, -0.0136871337890625, -0.09173583984375, -0.05181884765625, 0.01134490966796875, -0.004669189453125, 0.0074920654296875, -0.0069427490234375, 0.00998687744140625, 0.002410888671875, -0.0631103515625, -0.0166778564453125, 0.01139068603515625, -0.02133...
The women are wearing a dress.
query
rel
relS
true
person
14-dress_wearing,s
select: dress (12)->relate: person,wearing,s (7) [0]->query: name [1]
2
n336443
202225914
Does the utensil on top of the table look clean and black?
[ "no" ]
no
[ 0.012481689453125, 0.0758056640625, -0.0103607177734375, 0.032135009765625, -0.05938720703125, -0.008056640625, -0.031463623046875, -0.027679443359375, 0.022125244140625, -0.02008056640625, -0.0465087890625, -0.0312042236328125, -0.01457977294921875, -0.0733642578125, -0.0086746215820312...
[ -0.0236663818359375, 0.01311492919921875, 0.060089111328125, 0.01416015625, -0.0194854736328125, -0.01512908935546875, -0.0242156982421875, 0.023345947265625, -0.00415802001953125, -0.02154541015625, -0.0131378173828125, 0.00818634033203125, 0.006011962890625, -0.01012420654296875, 0.024...
No, the fork is clean but silver.
logical
attr
verifyAttrsC
true
null
05-black_clean
select: table (2)->relate: utensil,on top of,s (1) [0]->verify color: black [1]->verify cleanliness: clean [1]->and: [2, 3]
3
n179136
2062325
Is the surfer that looks wet wearing a wetsuit?
[ "yes" ]
yes
[ -0.0301666259765625, 0.0020542144775390625, -0.061920166015625, 0.0167999267578125, -0.01532745361328125, 0.007541656494140625, 0.0367431640625, -0.0122222900390625, 0.009033203125, -0.0303497314453125, -0.00370025634765625, 0.033843994140625, 0.0007433891296386719, 0.03125, -0.084716796...
[ 0.00952911376953125, -0.0228118896484375, -0.02618408203125, 0.028411865234375, 0.0396728515625, -0.0007081031799316406, 0.01004791259765625, -0.0144805908203125, 0.0004684925079345703, -0.0343017578125, 0.006580352783203125, 0.01305389404296875, 0.032318115234375, -0.006763458251953125, ...
Yes, the surfer is wearing a wetsuit.
verify
rel
relVerify
true
null
13-surfer_wetsuit
select: surfer (1)->filter: wet [0]->verify rel: wetsuit,wearing,o (12) [1]
4
n518912
201303229
How tall is the chair in the bottom of the photo?
[ "short" ]
short
[ 0.0380859375, 0.14599609375, -0.0806884765625, 0.0094146728515625, -0.0220794677734375, 0.0012502670288085938, 0.037872314453125, 0.01253509521484375, -0.0220184326171875, -0.00823974609375, 0.011138916015625, -0.0269927978515625, -0.0364990234375, -0.0266265869140625, -0.028823852539062...
[ -0.0193328857421875, 0.0098876953125, 0.0311126708984375, 0.003345489501953125, -0.052215576171875, 0.0303802490234375, -0.004364013671875, -0.043853759765625, -0.04730224609375, 0.0182037353515625, 0.03607177734375, 0.00916290283203125, -0.0550537109375, -0.01111602783203125, -0.0178222...
The chair is short.
query
attr
how
true
height
10q-chair_height
select: chair (13)->filter vposition: bottom [0]->query: height [1]
5
n435808
201902997
What kind of device is on top of the desk?
[ "keyboard" ]
keyboard
[ 0.01494598388671875, -0.00305938720703125, 0.00986480712890625, -0.05340576171875, -0.01934814453125, -0.01432037353515625, 0.0200653076171875, -0.01245880126953125, 0.0183868408203125, 0.07635498046875, 0.06390380859375, -0.02386474609375, 0.0147705078125, 0.0006957054138183594, -0.0150...
[ 0.004451751708984375, -0.035430908203125, 0.053375244140625, -0.003467559814453125, -0.0460205078125, 0.0277557373046875, 0.00983428955078125, -0.02911376953125, 0.01494598388671875, -0.04339599609375, -0.02197265625, -0.005401611328125, 0.0094757080078125, -0.0140380859375, -0.007156372...
The device is a keyboard.
query
rel
categoryRelS
true
device
15-desk_on top of,s
select: desk (1)->relate: device,on top of,s (8) [0]->query: name [1]
6
n414992
20567512
What is the airplane flying above?
[ "ocean" ]
ocean
[ -0.0275421142578125, 0.09051513671875, -0.17822265625, 0.070556640625, -0.0787353515625, 0.0149993896484375, -0.051788330078125, 0.0085906982421875, 0.0028209686279296875, -0.032440185546875, 0.0208892822265625, -0.01529693603515625, 0.0277252197265625, -0.041107177734375, -0.03485107421...
[ -0.01715087890625, -0.01488494873046875, -0.038238525390625, -0.027374267578125, -0.0325927734375, 0.023651123046875, -0.03680419921875, -0.0203094482421875, -0.0272369384765625, -0.0521240234375, -0.0185699462890625, 0.00521087646484375, 0.055450439453125, -0.032318115234375, -0.0149230...
The plane is flying above the ocean.
query
rel
relO
true
place
14-airplane_flying above,o
select: airplane (11)->relate: _,flying above,o (10) [0]->query: name [1]
7
n446242
20136592
What color are the pants?
[ "red" ]
red
[ 0.0308837890625, 0.198486328125, 0.0758056640625, 0.0160369873046875, 0.0158233642578125, -0.0269775390625, 0.01073455810546875, -0.00423431396484375, -0.0180816650390625, 0.002712249755859375, -0.021087646484375, -0.0233154296875, -0.01203155517578125, 0.034271240234375, -0.024963378906...
[ 0.0300445556640625, -0.01073455810546875, -0.03863525390625, -0.003597259521484375, -0.0306549072265625, 0.01178741455078125, 0.0158843994140625, 0.0308837890625, -0.01233673095703125, -0.01116180419921875, -0.02392578125, -0.000029742717742919922, -0.0310516357421875, -0.026702880859375, ...
The pants are red.
query
attr
directOf
true
color
10q-pants_color
select: pants (3)->query: color [0]
8
n168412
20602803
Is the ground blue or brown?
[ "brown" ]
brown
[ -0.005077362060546875, 0.1363525390625, -0.05694580078125, 0.08392333984375, 0.0445556640625, 0.01531219482421875, 0.02276611328125, 0.039093017578125, 0.0196533203125, 0.0034694671630859375, -0.00067138671875, -0.026611328125, 0.05462646484375, -0.06671142578125, 0.04638671875, -0.005...
[ 0.0061187744140625, -0.016815185546875, 0.0018777847290039062, -0.009002685546875, -0.055084228515625, 0.052886962890625, 0.006160736083984375, -0.0189056396484375, -0.0312347412109375, 0.031036376953125, -0.0016374588012695312, 0.023468017578125, 0.0149078369140625, 0.0161285400390625, ...
The ground is brown.
choose
attr
chooseAttr
true
color
10c-ground_color
select: ground (10)->choose color: brown|blue [0]
9
n23181
201079951
What is around the open window?
[ "drapes" ]
drapes
[0.05731201171875,0.112548828125,0.00167083740234375,0.0182647705078125,-0.033447265625,-0.050964355(...TRUNCATED)
[-0.01108551025390625,0.01274871826171875,0.00812530517578125,-0.0206756591796875,-0.023788452148437(...TRUNCATED)
The draperies are around the window.
query
rel
relS
true
textile
14-window_around,s
select: window (0)->filter: open [0]->relate: _,around,s (12) [1]->query: name [2]
End of preview.

GQA testdev-balanced (Lance Format)

Lance-formatted version of the canonical GQA testdev_balanced slice — 12,578 compositional VQA questions joined with the matching 398 images — sourced from lmms-lab/GQA.

lmms-lab/GQA exposes instructions and images as separate parquet configs; this Lance dataset joins them on imageId, so each row has the question, the answer, the GQA reasoning-program tags, and the image bytes inline.

Splits

Split Rows Distinct images
testdev.lance 12,578 398

Train (train_balanced_instructions × train_balanced_images, ~943k Q's × 72k images, ~10 GB images) and val splits are not bundled by default — pass --instr-config/--images-config to gqa/dataprep.py to extend.

Schema

Column Type Notes
id int64 Row index
image large_binary Inline JPEG bytes (image is duplicated across rows that share an image_id)
image_id string GQA scene-graph image id
question_id string GQA question id
question string Compositional natural-language question
answers list<string> One-element list (the GQA short answer)
answer string Same short answer (canonical / FTS target)
full_answer string? Full sentence answer
structural string? One of verify, query, compare, choose, logical
semantic string? One of attr, cat, global, obj, rel
detailed string? Fine-grained type (e.g. weatherVerifyC)
is_balanced bool GQA balanced subset flag
group_global / group_local string? GQA reasoning-group ids
semantic_str string? Compact description of the reasoning program
image_emb fixed_size_list<float32, 512> CLIP image embedding (cosine-normalized)
question_emb fixed_size_list<float32, 512> CLIP text embedding of the question

Pre-built indices

  • IVF_PQ on image_emb and question_embmetric=cosine
  • INVERTED (FTS) on question and answer
  • BITMAP on structural, semantic, detailed
  • BTREE on image_id, question_id

Quick start

import lance
ds = lance.dataset("hf://datasets/lance-format/gqa-testdev-balanced-lance/data/testdev.lance")
print(ds.count_rows(), ds.schema.names, ds.list_indices())

Load with LanceDB

These tables can also be consumed by LanceDB, the multimodal lakehouse and embedded search library built on top of Lance, for simplified vector search and other queries.

import lancedb

db = lancedb.connect("hf://datasets/lance-format/gqa-testdev-balanced-lance/data")
tbl = db.open_table("testdev")
print(f"LanceDB table opened with {len(tbl)} image-question pairs")

LanceDB vector search

import lancedb

db = lancedb.connect("hf://datasets/lance-format/gqa-testdev-balanced-lance/data")
tbl = db.open_table("testdev")

ref = tbl.search().limit(1).select(["question_emb", "question"]).to_list()[0]
query_embedding = ref["question_emb"]

results = (
    tbl.search(query_embedding, vector_column_name="question_emb")
    .metric("cosine")
    .select(["question", "answer"])
    .limit(5)
    .to_list()
)

LanceDB full-text search

import lancedb

db = lancedb.connect("hf://datasets/lance-format/gqa-testdev-balanced-lance/data")
tbl = db.open_table("testdev")

results = (
    tbl.search("color of the car")
    .select(["question", "answer"])
    .limit(10)
    .to_list()
)

Filter by reasoning type

import lance
ds = lance.dataset("hf://datasets/lance-format/gqa-testdev-balanced-lance/data/testdev.lance")
verify_qs = ds.scanner(filter="structural = 'verify'", columns=["question", "answer"], limit=5).to_table()

Filter with LanceDB

import lancedb

db = lancedb.connect("hf://datasets/lance-format/gqa-testdev-balanced-lance/data")
tbl = db.open_table("testdev")
verify_qs = (
    tbl.search()
    .where("structural = 'verify'")
    .select(["question", "answer"])
    .limit(5)
    .to_list()
)

Why Lance?

  • One dataset for the joined image + question + answer + reasoning-program metadata + dual embeddings + indices — no instructions/images parquet split to keep in sync.
  • Schema evolution: add columns (alternate scene graphs, model predictions) without rewriting the data.

Source & license

Converted from lmms-lab/GQA. GQA is released under CC BY 4.0 by Hudson and Manning (Stanford NLP).

Citation

@inproceedings{hudson2019gqa,
  title={GQA: A New Dataset for Real-World Visual Reasoning and Compositional Question Answering},
  author={Hudson, Drew A. and Manning, Christopher D.},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2019}
}
Downloads last month
48