Datasets:

License:
Dataset Viewer
Duplicate
The dataset viewer is not available for this split.
Cannot extract the features (columns) for the split 'train' of the config 'default' of the dataset.
Error code:   FeaturesError
Exception:    ArrowInvalid
Message:      JSON parse error: Column() changed from object to string in row 0
Traceback:    Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 174, in _generate_tables
                  df = pandas_read_json(f)
                       ^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 38, in pandas_read_json
                  return pd.read_json(path_or_buf, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pandas/io/json/_json.py", line 815, in read_json
                  return json_reader.read()
                         ^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pandas/io/json/_json.py", line 1014, in read
                  obj = self._get_object_parser(self.data)
                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pandas/io/json/_json.py", line 1040, in _get_object_parser
                  obj = FrameParser(json, **kwargs).parse()
                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pandas/io/json/_json.py", line 1176, in parse
                  self._parse()
                File "/usr/local/lib/python3.12/site-packages/pandas/io/json/_json.py", line 1391, in _parse
                  self.obj = DataFrame(
                             ^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pandas/core/frame.py", line 778, in __init__
                  mgr = dict_to_mgr(data, index, columns, dtype=dtype, copy=copy, typ=manager)
                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pandas/core/internals/construction.py", line 503, in dict_to_mgr
                  return arrays_to_mgr(arrays, columns, index, dtype=dtype, typ=typ, consolidate=copy)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pandas/core/internals/construction.py", line 114, in arrays_to_mgr
                  index = _extract_index(arrays)
                          ^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/pandas/core/internals/construction.py", line 677, in _extract_index
                  raise ValueError("All arrays must be of the same length")
              ValueError: All arrays must be of the same length
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 243, in compute_first_rows_from_streaming_response
                  iterable_dataset = iterable_dataset._resolve_features()
                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 3496, in _resolve_features
                  features = _infer_features_from_batch(self.with_format(None)._head())
                                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2257, in _head
                  return next(iter(self.iter(batch_size=n)))
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2461, in iter
                  for key, example in iterator:
                                      ^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 1952, in __iter__
                  for key, pa_table in self._iter_arrow():
                                       ^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 1974, in _iter_arrow
                  yield from self.ex_iterable._iter_arrow()
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 503, in _iter_arrow
                  for key, pa_table in iterator:
                                       ^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 350, in _iter_arrow
                  for key, pa_table in self.generate_tables_fn(**gen_kwags):
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 177, in _generate_tables
                  raise e
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 151, in _generate_tables
                  pa_table = paj.read_json(
                             ^^^^^^^^^^^^^^
                File "pyarrow/_json.pyx", line 342, in pyarrow._json.read_json
                File "pyarrow/error.pxi", line 155, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status
              pyarrow.lib.ArrowInvalid: JSON parse error: Column() changed from object to string in row 0

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

SportsGrounding Dataset

1. Task Description

The Spatio-Temporal Video Grounding (STVG) task aims to take an untrimmed video and a natural language description as input, and output the spatio-temporal tube corresponding to the natural language description—specifically, locating the start and end frames, as well as the bounding boxes of the target within the located segment.

SportsGrounding is based on the basketball subset of the MultiSports dataset, focusing on basketball-related scenarios. It contains numerous complex interaction scenes between people and between people and objects, with a larger number of instances.

2. Data Overview

The dataset is modified based on the video and annotation data of MultiSports. Six videos with excessive repetitive actions that cannot be distinguished by natural language are removed, resulting in a total of 520 videos. It follows the train/val split of MultiSports, with 374 videos in the training set and 146 videos in the validation set. Unlike other STVG datasets, each video in SportsGrounding contains multiple captions describing different target persons.

Key Statistics of SportsGrounding

Metric Value
Dataset Size 4243 tubes (i.e., 4243 video-text pairs)
Average Video Duration 19.70 seconds
Average Tube Duration 1.49 seconds
Average Description Length 16.89 words

Comparison with Other Datasets

Metric VidSTG HC-STVG v1 HC-STVG v2 SportsGrounding
Data Source VidOR AVA - MultiSports
Dataset Size 99943 video-text pairs 5660 video-text pairs 16544 video-text pairs 4243 video-text pairs
Average Video Duration 28.01 seconds 20 seconds - 19.70 seconds
Average Tube Duration 9.68 seconds 5.37 seconds - 1.49 seconds
Average Description Length Declarative: 11.12; Interrogative: 8.98 17.25 - 16.89

Unique Characteristics of SportsGrounding

  • Some instances have very short durations.
  • More people appear in the videos (other datasets contain fewer people even in multi-person scenarios; 57.2% of videos in HC-STVG have more than 3 people, and the rest have 2 people).
  • More complex interactions between people; many descriptions require certain reasoning and scene information modeling. Example: "The defender in white is blocked by the teammate of this offensive player."

3. Data Format

Annotation files are divided into train.json and val.json. Each annotation entry is structured as follows:

{
  "v_00HRwkvvjtQ_c001.mp4": [
    {
      "bbox": [[x0, y0, w0, h0], [x1, y1, w1, h1], ...],  # Continuous bounding boxes of the target person
      "fps": 25,  # Frames per second of the video
      "st_frame": 7,  # Ground truth start frame (index starts at 1)
      "ed_frame": 28,  # Ground truth end frame
      "caption": "The player in white attempts to gain control of the ball after an official tosses it into the air between him and his opponent in blue.",  # Caption describing the target person
      "width": 1080,  # Image width
      "height": 720  # Image height
    },
    ...
  ],
  ...
}
Downloads last month
1