--- dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 6242898204.747201 num_examples: 2539940 - name: validation num_bytes: 328575918.2527994 num_examples: 133682 download_size: 3992353312 dataset_size: 6571474123 configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* language: - fr --- # Dataset Card for French Wikipedia Text Corpus ## Dataset Description The French Wikipedia Text Corpus is a comprehensive dataset derived from French Wikipedia articles. It is specifically designed for training language models (LLMs). The dataset contains the text of paragraphs from Wikipedia articles, with sections, footnotes, and titles removed to provide a clean and continuous text stream. Dataset Details Features text: A single attribute containing the full text of Wikipedia articles as a continuous chunk of text. ## Languages The dataset is in French. ## Dataset statistics: Train set: 2539940 examples Validation set: 133682 examples ## Dataset Creation The dataset was created by extracting and processing text from French Wikipedia articles. The processing involved removing article sections, footnotes, and titles to ensure that the text is in a format suitable for training language models. ## Intended Uses This dataset is intended for use in training and evaluating language models. The clean and continuous text format makes it particularly suitable for models that require large volumes of coherent text data. ## Example Usage ```python from datasets import load_dataset # Load the dataset dataset = load_dataset('1ou2/fr_wiki_paragraphs') # Access the training data train_data = dataset['train'] # Print an example print(train_data[0]['text']) ``` ## Citation Information If you use this dataset in your research, please cite it as follows: @misc{french_wikipedia_text_corpus, author = {Gabriel Pastor}, title = {French Wikipedia Text Corpus}, year = {2025}, publisher = {Hugging Face}, journal = {Hugging Face Datasets}, howpublished = {\url{https://huggingface.co/datasets/1ou2/fr_wiki_paragraphs}} } ## License This dataset is licensed under the CC BY-SA 3.0 license, in accordance with Wikipedia's content license.