NLVR2 / README.md
luodian's picture
Update README.md
a101e11 verified
metadata
dataset_info:
  features:
    - name: question_id
      dtype: string
    - name: question
      dtype: string
    - name: left_image
      dtype: image
    - name: right_image
      dtype: image
    - name: answer
      dtype: string
    - name: writer
      dtype: string
    - name: synset
      dtype: string
    - name: query
      dtype: string
    - name: identifier
      dtype: string
    - name: extra_validations
      dtype: string
    - name: left_url
      dtype: string
    - name: right_url
      dtype: string
  splits:
    - name: balanced_dev
      num_bytes: 488008399
      num_examples: 2300
    - name: balanced_test_public
      num_bytes: 507142965.656
      num_examples: 2316
    - name: balanced_test_unseen
      num_bytes: 409992570.832
      num_examples: 2124
    - name: unbalanced_dev
      num_bytes: 762577280.968
      num_examples: 3562
    - name: unbalanced_test_unseen
      num_bytes: 804486738.08
      num_examples: 3668
    - name: unbalanced_test_public
      num_bytes: 731416041.32
      num_examples: 3536
  download_size: 1879819700
  dataset_size: 3703623995.856
configs:
  - config_name: default
    data_files:
      - split: balanced_dev
        path: data/balanced_dev-*
      - split: balanced_test_public
        path: data/balanced_test_public-*
      - split: balanced_test_unseen
        path: data/balanced_test_unseen-*
      - split: unbalanced_dev
        path: data/unbalanced_dev-*
      - split: unbalanced_test_unseen
        path: data/unbalanced_test_unseen-*
      - split: unbalanced_test_public
        path: data/unbalanced_test_public-*

Dataset Card for "nlvr2"

Large-scale Multi-modality Models Evaluation Suite

Accelerating the development of large-scale multi-modality models (LMMs) with lmms-eval

🏠 Homepage | 📚 Documentation | 🤗 Huggingface Datasets

This Dataset

This is a formatted version of NLVR2. It is used in our lmms-eval pipeline to allow for one-click evaluations of large multi-modality models.

@inproceedings{suhr2017corpus,
  title={A corpus of natural language for visual reasoning},
  author={Suhr, Alane and Lewis, Mike and Yeh, James and Artzi, Yoav},
  booktitle={Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)},
  pages={217--223},
  year={2017}
}

More Information needed