Upload config.yaml with huggingface_hub
Browse files- config.yaml +5 -4
config.yaml
CHANGED
|
@@ -39,7 +39,7 @@ data_mix:
|
|
| 39 |
dataset_path: manu/french_librispeech_text_only
|
| 40 |
build_test_set_from_train: true
|
| 41 |
- (): dataset_construction.DatasetConfig
|
| 42 |
-
dataset_path: manu/french_podcasts
|
| 43 |
build_test_set_from_train: true
|
| 44 |
num_test_examples: 100
|
| 45 |
filtering_function:
|
|
@@ -67,9 +67,9 @@ data_mix:
|
|
| 67 |
# num_test_examples: 100
|
| 68 |
# here we should preprocess
|
| 69 |
- (): dataset_construction.DatasetConfig
|
| 70 |
-
dataset_path: manu/ProjectGutenberg_fr
|
| 71 |
build_test_set_from_train: true
|
| 72 |
-
- (): dataset_construction.DatasetConfig
|
| 73 |
dataset_path: manu/bnf_gallica
|
| 74 |
build_test_set_from_train: true
|
| 75 |
- (): dataset_construction.DatasetConfig
|
|
@@ -150,4 +150,5 @@ data_mix:
|
|
| 150 |
(): dataset_collection.french.oscar.OscarFilter
|
| 151 |
|
| 152 |
|
| 153 |
-
tokenizer: "mistralai/Mistral-7B-v0.1"
|
|
|
|
|
|
| 39 |
dataset_path: manu/french_librispeech_text_only
|
| 40 |
build_test_set_from_train: true
|
| 41 |
- (): dataset_construction.DatasetConfig
|
| 42 |
+
dataset_path: manu/french_podcasts # continue scrapping
|
| 43 |
build_test_set_from_train: true
|
| 44 |
num_test_examples: 100
|
| 45 |
filtering_function:
|
|
|
|
| 67 |
# num_test_examples: 100
|
| 68 |
# here we should preprocess
|
| 69 |
- (): dataset_construction.DatasetConfig
|
| 70 |
+
dataset_path: manu/ProjectGutenberg_fr # needs to be filtered - beginning and end of books
|
| 71 |
build_test_set_from_train: true
|
| 72 |
+
- (): dataset_construction.DatasetConfig # needs to be filtered and expanded - perplexity filtering
|
| 73 |
dataset_path: manu/bnf_gallica
|
| 74 |
build_test_set_from_train: true
|
| 75 |
- (): dataset_construction.DatasetConfig
|
|
|
|
| 150 |
(): dataset_collection.french.oscar.OscarFilter
|
| 151 |
|
| 152 |
|
| 153 |
+
# tokenizer: "mistralai/Mistral-7B-v0.1"
|
| 154 |
+
tokenizer: "manu/tok-fr-en-code"
|