Update README.md
Browse files
README.md
CHANGED
|
@@ -11,7 +11,7 @@ Here's a little example:
|
|
| 11 |
``` python
|
| 12 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 13 |
|
| 14 |
-
model_name = "
|
| 15 |
tokenizer = AutoTokenizer.from_pretrained(model_name, device_map="auto")
|
| 16 |
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
|
| 17 |
|
|
@@ -66,7 +66,7 @@ This is not easy. Such data that takes the input as a long text (newspaper artic
|
|
| 66 |
|
| 67 |
After all, we have about 72,101 samples with a ratio of 0.05 (68495 on the train set and 3606 on the test set)
|
| 68 |
|
| 69 |
-
We published the dataset at [here](https://huggingface.co/datasets/
|
| 70 |
|
| 71 |
### Custom Evaluation Data
|
| 72 |
As part of the final evaluation for benchmark, we gathered around 27 Vietnamese children's stories and divided into many samples, accumulating to 118 samples. The dataset can be found [here](/data/eval_set.json)
|
|
@@ -94,7 +94,7 @@ As mentioned earlier, we use [Vistral-7B-Chat](https://huggingface.co/Viet-Mistr
|
|
| 94 |
- Optimizer: Adamw bnb 8bit
|
| 95 |
- Sequence Len: 1096
|
| 96 |
|
| 97 |
-
The weights can be found [here](https://huggingface.co/
|
| 98 |
|
| 99 |
The notebook for training can be found at `notebook/Fine_tune_LLMs_with_Axolotl.ipynb`
|
| 100 |
|
|
|
|
| 11 |
``` python
|
| 12 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 13 |
|
| 14 |
+
model_name = "andythetechnerd03/VistralPoem5"
|
| 15 |
tokenizer = AutoTokenizer.from_pretrained(model_name, device_map="auto")
|
| 16 |
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
|
| 17 |
|
|
|
|
| 66 |
|
| 67 |
After all, we have about 72,101 samples with a ratio of 0.05 (68495 on the train set and 3606 on the test set)
|
| 68 |
|
| 69 |
+
We published the dataset at [here](https://huggingface.co/datasets/andythetechnerd03/Vietnamese-Poem-5words)
|
| 70 |
|
| 71 |
### Custom Evaluation Data
|
| 72 |
As part of the final evaluation for benchmark, we gathered around 27 Vietnamese children's stories and divided into many samples, accumulating to 118 samples. The dataset can be found [here](/data/eval_set.json)
|
|
|
|
| 94 |
- Optimizer: Adamw bnb 8bit
|
| 95 |
- Sequence Len: 1096
|
| 96 |
|
| 97 |
+
The weights can be found [here](https://huggingface.co/andythetechnerd03/VistralPoem5)
|
| 98 |
|
| 99 |
The notebook for training can be found at `notebook/Fine_tune_LLMs_with_Axolotl.ipynb`
|
| 100 |
|