Commit ·
91b458a
1
Parent(s): 813660a
Update README.md
Browse files
README.md
CHANGED
|
@@ -10,18 +10,18 @@ tags:
|
|
| 10 |
# Mistral-7B-v0.1 for Italian Language Text Generation
|
| 11 |
|
| 12 |
## Overview
|
| 13 |
-
`Mistral-7B-v0.1` is a state-of-the-art Large Language Model (LLM) specifically pre-trained for generating text. With its 7 billion parameters, it's built to excel in benchmarks and outperforms even some larger models like the Llama 2 13B
|
| 14 |
|
| 15 |
## Model Architecture
|
| 16 |
-
The Mistral-7B-v0.1 model is a transformer-based model that can handle a variety of tasks including but not limited to translation, summarization, and text completion. It's particularly designed for the Italian language and can be fine-tuned for specific tasks
|
| 17 |
|
| 18 |
## Capabilities
|
| 19 |
- **Vocabulary Size**: 32,000 tokens, allowing for a broad range of inputs and outputs.
|
| 20 |
- **Hidden Size**: 4,096 dimensions, providing rich internal representations.
|
| 21 |
-
- **Intermediate Size**: 14,336 dimensions, which contributes to the model's ability to process and generate complex sentences
|
| 22 |
|
| 23 |
## Performance
|
| 24 |
-
Mistral-7B has been demonstrated to perform exceptionally well across a range of benchmarks, making it a reliable choice for developers and researchers working with the Italian language
|
| 25 |
|
| 26 |
|
| 27 |
## How to Use
|
|
|
|
| 10 |
# Mistral-7B-v0.1 for Italian Language Text Generation
|
| 11 |
|
| 12 |
## Overview
|
| 13 |
+
`Mistral-7B-v0.1` is a state-of-the-art Large Language Model (LLM) specifically pre-trained for generating text. With its 7 billion parameters, it's built to excel in benchmarks and outperforms even some larger models like the Llama 2 13B.
|
| 14 |
|
| 15 |
## Model Architecture
|
| 16 |
+
The Mistral-7B-v0.1 model is a transformer-based model that can handle a variety of tasks including but not limited to translation, summarization, and text completion. It's particularly designed for the Italian language and can be fine-tuned for specific tasks.
|
| 17 |
|
| 18 |
## Capabilities
|
| 19 |
- **Vocabulary Size**: 32,000 tokens, allowing for a broad range of inputs and outputs.
|
| 20 |
- **Hidden Size**: 4,096 dimensions, providing rich internal representations.
|
| 21 |
+
- **Intermediate Size**: 14,336 dimensions, which contributes to the model's ability to process and generate complex sentences.
|
| 22 |
|
| 23 |
## Performance
|
| 24 |
+
Mistral-7B has been demonstrated to perform exceptionally well across a range of benchmarks, making it a reliable choice for developers and researchers working with the Italian language.
|
| 25 |
|
| 26 |
|
| 27 |
## How to Use
|