dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 6242898204.747201
num_examples: 2539940
- name: validation
num_bytes: 328575918.2527994
num_examples: 133682
download_size: 3992353312
dataset_size: 6571474123
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
language:
- fr
Dataset Card for French Wikipedia Text Corpus
Dataset Description
The French Wikipedia Text Corpus is a comprehensive dataset derived from French Wikipedia articles. It is specifically designed for training language models (LLMs). The dataset contains the text of paragraphs from Wikipedia articles, with sections, footnotes, and titles removed to provide a clean and continuous text stream. Dataset Details Features
text: A single attribute containing the full text of Wikipedia articles as a continuous chunk of text.
Languages
The dataset is in French.
Dataset statistics:
Train set: 2539940 examples Validation set: 133682 examples
Dataset Creation
The dataset was created by extracting and processing text from French Wikipedia articles. The processing involved removing article sections, footnotes, and titles to ensure that the text is in a format suitable for training language models.
Intended Uses
This dataset is intended for use in training and evaluating language models. The clean and continuous text format makes it particularly suitable for models that require large volumes of coherent text data.
Example Usage
from datasets import load_dataset
# Load the dataset
dataset = load_dataset('1ou2/fr_wiki_paragraphs')
# Access the training data
train_data = dataset['train']
# Print an example
print(train_data[0]['text'])
Citation Information
If you use this dataset in your research, please cite it as follows:
@misc{french_wikipedia_text_corpus, author = {Gabriel Pastor}, title = {French Wikipedia Text Corpus}, year = {2025}, publisher = {Hugging Face}, journal = {Hugging Face Datasets}, howpublished = {\url{https://huggingface.co/datasets/1ou2/fr_wiki_paragraphs}} }
License
This dataset is licensed under the CC BY-SA 3.0 license, in accordance with Wikipedia's content license.