dignity045 commited on
Commit
6f17e5f
Β·
verified Β·
1 Parent(s): 80428f9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +112 -3
README.md CHANGED
@@ -1,3 +1,112 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ ---
5
+ dataset_name: dignity045/Collective-Corpus
6
+ license: mit
7
+ language: multilingual
8
+ size_categories: 500B+ tokens
9
+ task_categories:
10
+ - text-generation
11
+ - fill-mask
12
+ - text-classification
13
+ - summarization
14
+ - question-answering
15
+ pretty_name: Collective Corpus
16
+ tags:
17
+ - pretraining
18
+ - finetuning
19
+ - large-language-model
20
+ - code
21
+ - math
22
+ - instructions
23
+ ---
24
+
25
+ # 🧠 Collective Corpus β€” Universal Pretraining + Finetuning Dataset (500B+ Tokens)
26
+
27
+ [![Hugging Face](https://img.shields.io/badge/πŸ€—-Dataset-yellow)](https://huggingface.co/datasets/dignity045/Collective-Corpus)
28
+ [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT)
29
+ [![Status](https://img.shields.io/badge/Status-In%20Progress-orange)](#-current-status)
30
+
31
+ **`Collective-Corpus`** is a massive-scale, **multi-domain** dataset designed to train Transformer-based language models **from scratch** and **finetune** them across a wide variety of domains β€” all in one place.
32
+
33
+ ## πŸ“š Dataset Scope
34
+
35
+ This dataset aims to **cover the full LLM lifecycle**, from raw pretraining to domain-specialized finetuning.
36
+
37
+ ### 1. Pretraining Corpus
38
+ - Large-scale, diverse multilingual text sources
39
+ - Cleaned, deduplicated, and filtered for quality
40
+ - Inspired by datasets like [C4](https://huggingface.co/datasets/c4) and [FineWeb](https://huggingface.co/datasets/HuggingFaceFW/fineweb)
41
+
42
+ ### 2. Domain-Specific Finetuning
43
+ - **Instruction Following & Dialogue** β€” Chatbots, multi-turn conversations
44
+ - **Code** β€” Python, JavaScript, Java, C++, and more
45
+ - **Math & Logical Reasoning**
46
+ - **Specialized Fields** β€” Research papers, technical documentation
47
+
48
+ ---
49
+
50
+ ## πŸ“Š Scale
51
+
52
+ - **Total Tokens**: **500B+**
53
+ - **Estimated Text Samples**: **700M+**
54
+ - **Target Model Size**: Suitable for training large models **from scratch**
55
+ - Covers **general-purpose** and **domain-specific** training needs
56
+
57
+ ---
58
+
59
+ ## 🎯 Goals
60
+
61
+ 1. Build a **unified corpus** for full-stack LLM development.
62
+ 2. Enable **open and reproducible** large-scale language model research.
63
+ 3. Support **finetuning for high-impact domains** like code, math, and dialogue.
64
+
65
+ ---
66
+
67
+ ## 🚧 Current Status
68
+
69
+ - **Model Pretraining**: Currently training a Transformer model from scratch on the full **500B+ token** dataset.
70
+ - **Public Release**: Planned **after model training completes**.
71
+
72
+ ---
73
+
74
+ ## 🀝 Collaboration
75
+
76
+ We are **actively seeking open-source collaborators** to:
77
+ - Contribute to dataset cleaning, filtering, and deduplication
78
+ - Assist in large-scale model training and evaluation
79
+ - Provide expertise for **specialized domain corpora**
80
+
81
+ We also **offer free guidance** on:
82
+ - Dataset curation best practices
83
+ - Efficient large-scale LLM training pipelines
84
+ - Transformer architecture optimization
85
+
86
+ ---
87
+
88
+ ## πŸ’Ό Job Inquiries
89
+
90
+ Interested in **collaboration, hiring, or consulting** for dataset engineering, large-scale model training, or applied NLP?
91
+
92
+ πŸ“§ **Email**: `your_email@example.com`
93
+
94
+ ---
95
+
96
+ ## πŸ“… Release Timeline
97
+
98
+ | Stage | Status |
99
+ |------------------------|------------------|
100
+ | Data Curation | 🚧 In Progress |
101
+ | Model Pretraining | 🚧 In Progress |
102
+ | Dataset Public Release | ⏳ Post-training |
103
+
104
+ ---
105
+
106
+ ## πŸ“œ License
107
+
108
+ Released under the **MIT License** β€” free for research and commercial use.
109
+
110
+ ---
111
+
112
+ ### 🌍 Let’s build the next generation of **open-source LLMs** β€” together.