Dataset Viewer
Auto-converted to Parquet Duplicate
The dataset viewer is not available for this split.
The info cannot be fetched for the config 'default' of the dataset.
Error code:   InfoError
Exception:    ConnectionError
Message:      Couldn't reach 'a3xrfgb/amharic-sentences-corpus' on the Hub (LocalEntryNotFoundError)
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 223, in compute_first_rows_from_streaming_response
                  info = get_dataset_config_info(path=dataset, config_name=config, token=hf_token)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 268, in get_dataset_config_info
                  builder = load_dataset_builder(
                            ^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/load.py", line 1130, in load_dataset_builder
                  dataset_module = dataset_module_factory(
                                   ^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/load.py", line 1029, in dataset_module_factory
                  raise e1 from None
                File "/usr/local/lib/python3.12/site-packages/datasets/load.py", line 955, in dataset_module_factory
                  raise ConnectionError(f"Couldn't reach '{path}' on the Hub ({e.__class__.__name__})") from e
              ConnectionError: Couldn't reach 'a3xrfgb/amharic-sentences-corpus' on the Hub (LocalEntryNotFoundError)

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

zz7eqhpgTF5HXxf5ybO_9

Amharic Sentences Corpus V1.0

Source: Telegram

This 1.6 million Amharic sentences corpus reflects current Amharic usage as of December 20, 2025, and is designed for anyone interested in:

  • Training Amharic-based LLMs
  • Fine-tuning NLP models
  • Building search, summarization, or generative systems in Amharic

The dataset is heavily cleaned and normalized, but like any serious LLM dataset, it still needs proper tokenization for pre-training.

I recommend using an Amharic-specific tokenizer such as: https://pypi.org/project/amharic-tokenizer/0.2.6

Another useful Amharic Sentences corpus by @rasyosef & @Addis AI

How did I create this text corpus?

This project is fully open-source & community-driven. If you’re building in NLP, AI research, or language technology, this is for you.

Use it. Improve it. Build on top of it.

Downloads last month
23