Update README.md
Browse files
README.md
CHANGED
|
@@ -16,14 +16,14 @@ tags:
|
|
| 16 |
|
| 17 |
InfCapital LLama2-7b clone of [Meta's Llama 2 7B Chat](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf).
|
| 18 |
|
| 19 |
-
Adopt for Vietnamese continued pretraining or fine-tuning by extend vocab size from 32,000 to
|
| 20 |
|
| 21 |
|
| 22 |
## Model Architecture
|
| 23 |
```
|
| 24 |
LlamaForCausalLM(
|
| 25 |
(model): LlamaModel(
|
| 26 |
-
(embed_tokens): Embedding(
|
| 27 |
(layers): ModuleList(
|
| 28 |
(0-31): 32 x LlamaDecoderLayer(
|
| 29 |
(self_attn): LlamaAttention(
|
|
@@ -45,6 +45,6 @@ LlamaForCausalLM(
|
|
| 45 |
)
|
| 46 |
(norm): LlamaRMSNorm()
|
| 47 |
)
|
| 48 |
-
(lm_head): Linear(in_features=4096, out_features=
|
| 49 |
)
|
| 50 |
```
|
|
|
|
| 16 |
|
| 17 |
InfCapital LLama2-7b clone of [Meta's Llama 2 7B Chat](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf).
|
| 18 |
|
| 19 |
+
Adopt for Vietnamese continued pretraining or fine-tuning by extend vocab size from 32,000 to 44,800. Vocabs added by training sentencepiece method from dataset vnnews-corpus.
|
| 20 |
|
| 21 |
|
| 22 |
## Model Architecture
|
| 23 |
```
|
| 24 |
LlamaForCausalLM(
|
| 25 |
(model): LlamaModel(
|
| 26 |
+
(embed_tokens): Embedding(44800, 4096)
|
| 27 |
(layers): ModuleList(
|
| 28 |
(0-31): 32 x LlamaDecoderLayer(
|
| 29 |
(self_attn): LlamaAttention(
|
|
|
|
| 45 |
)
|
| 46 |
(norm): LlamaRMSNorm()
|
| 47 |
)
|
| 48 |
+
(lm_head): Linear(in_features=4096, out_features=44800, bias=False)
|
| 49 |
)
|
| 50 |
```
|