Update README.md
#3
by ivas-tri - opened
README.md
CHANGED
|
@@ -93,7 +93,7 @@ We follow their training recipe and release our version of Mamba-7B.
|
|
| 93 |
|
| 94 |
## Training Details
|
| 95 |
- Mamba-7B was trained using AWS SageMaker on 128 H100 80GB GPUs.
|
| 96 |
-
- Training began in March 2024 and lasted
|
| 97 |
| **Hyperparameter** | **Value** |
|
| 98 |
|--------------------|------------|
|
| 99 |
| Precision | `bfloat16` |
|
|
|
|
| 93 |
|
| 94 |
## Training Details
|
| 95 |
- Mamba-7B was trained using AWS SageMaker on 128 H100 80GB GPUs.
|
| 96 |
+
- Training began in March 2024 and lasted three weeks.
|
| 97 |
| **Hyperparameter** | **Value** |
|
| 98 |
|--------------------|------------|
|
| 99 |
| Precision | `bfloat16` |
|