Dataset Viewer
Duplicate
The dataset viewer is not available for this split.
Cannot extract the features (columns) for the split 'train' of the config 'default' of the dataset.
Error code:   FeaturesError
Exception:    FileNotFoundError
Message:      [Errno 2] No such file or directory: '<datasets.utils.file_utils.FilesIterable object at 0x7f319f4e0fb0>'
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 243, in compute_first_rows_from_streaming_response
                  iterable_dataset = iterable_dataset._resolve_features()
                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 4195, in _resolve_features
                  features = _infer_features_from_batch(self.with_format(None)._head())
                                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2533, in _head
                  return next(iter(self.iter(batch_size=n)))
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2711, in iter
                  for key, pa_table in ex_iterable.iter_arrow():
                                       ^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2249, in _iter_arrow
                  yield from self.ex_iterable._iter_arrow()
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 494, in _iter_arrow
                  for key, pa_table in iterator:
                                       ^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 384, in _iter_arrow
                  for key, pa_table in self.generate_tables_fn(**gen_kwags):
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/xml/xml.py", line 67, in _generate_tables
                  with open(file, encoding=self.config.encoding, errors=self.config.encoding_errors) as f:
                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/streaming.py", line 73, in wrapper
                  return function(*args, download_config=download_config, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 967, in xopen
                  return open(main_hop, mode, *args, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
              FileNotFoundError: [Errno 2] No such file or directory: '<datasets.utils.file_utils.FilesIterable object at 0x7f319f4e0fb0>'

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

Project 02

πŸ“š WikiArticles Dataset by Onyx

This is the WikiArticles dataset by Onyx. This dataset is designed for AI training and for reading too!

Our recommendation is to download the EnglishWiki (en) since English is built-in in almost every AI we have in Hugging Face or the AI industry.


πŸš€ How to Use

This dataset is provided in the original .xml.bz2 format to ensure data integrity and high compression.

  1. For Training: Use an XML parser or the wikiextractor tool to convert this into JSONL for LLM fine-tuning.
  2. For Reading: Use any BZip2 compatible reader to view the raw XML knowledge.

βš–οΈ License

This project is licensed under the Apache License 2.0. You are free to use, modify, and distribute this data as long as you follow the terms of the Apache 2.0 license.

Downloads last month
57

Collections including MegaBites-AI/WikiArticles