Dataset Preview
Full Screen
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Job manager crashed while running this job (missing heartbeats).
Error code:   JobManagerCrashedError

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

image
image
End of preview.

Dataset Card for GQA-35k

image

The GQA (Visual Reasoning in the Real World) dataset is a large-scale visual question answering dataset that includes scene graph annotations for each image.

This is a FiftyOne dataset with 35000 samples.

Note: This is a 35,000 sample subset which does not contain questions, only the scene graph annotations as detection-level attributes.

You can find the recipe notebook for creating the dataset here

Installation

If you haven't already, install FiftyOne:

pip install -U fiftyone

Usage

import fiftyone as fo
import fiftyone.utils.huggingface as fouh

# Load the dataset
# Note: other available arguments include 'max_samples', etc
dataset = fouh.load_from_hub("Voxel51/GQA-Scene-Graph")

# Launch the App
session = fo.launch_app(dataset)

Dataset Details

Dataset Description

Scene Graph Annotations

  • Each of the 113K images in GQA is associated with a detailed scene graph describing the objects, attributes and relations present.

  • The scene graphs are based on a cleaner version of the Visual Genome scene graphs.

  • For each image, the scene graph is provided as a dictionary (sceneGraph) containing:

    • Image metadata like width, height, location, weather
    • A dictionary (objects) mapping each object ID to its name, bounding box coordinates, attributes, and relations[6]
    • Relations are represented as triples specifying the predicate (e.g. "holding", "on", "left of") and the target object ID[6]
  • Curated by: Drew Hudson & Christopher Manning

  • Shared by: Harpreet Sahota, Hacker-in-Residence at Voxel51

  • Language(s) (NLP): en

  • License:

  • GQA annotations (scene graphs, questions, programs) licensed under CC BY 4.0

  • Images sourced from Visual Genome may have different licensing terms

Dataset Sources

Dataset Structure

Here's the information presented as a markdown table:

Field Type Description
location str Optional. The location of the image, e.g. kitchen, beach.
weather str Optional. The weather in the image, e.g. sunny, cloudy.
objects dict A dictionary from objectId to its object.
    object dict A visual element in the image (node).
        name str The name of the object, e.g. person, apple or sky.
        x int Horizontal position of the object bounding box (top left).
        y int Vertical position of the object bounding box (top left).
        w int The object bounding box width in pixels.
        h int The object bounding box height in pixels.
        attributes [str] A list of all the attributes of the object, e.g. blue, small, running.
        relations [dict] A list of all outgoing relations (edges) from the object (source).
            relation dict A triple representing the relation between source and target objects.

Note: I've used non-breaking spaces ( ) to indent the nested fields in the 'Field' column to represent the hierarchy. This helps to visually distinguish the nested structure within the table.

Citation

BibTeX:

@article{Hudson_2019,
   title={GQA: A New Dataset for Real-World Visual Reasoning and Compositional Question Answering},
   ISBN={9781728132938},
   url={http://dx.doi.org/10.1109/CVPR.2019.00686},
   DOI={10.1109/cvpr.2019.00686},
   journal={2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
   publisher={IEEE},
   author={Hudson, Drew A. and Manning, Christopher D.},
   year={2019},
   month={Jun}
}
Downloads last month
1,901