File size: 1,327 Bytes
eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 eaa0e07 3d1e3e5 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 |
---
library_name: transformers
license: apache-2.0
base_model: training_from_scratch
tags:
- generated_from_trainer
- trainging_from_scratch
datasets:
- Self-GRIT/wikitext-2-raw-v1-preprocessed
metrics:
- accuracy
model-index:
- name: Capybara
results:
- task:
name: Causal Language Modeling
type: text-generation
dataset:
name: Self-GRIT/wikitext-2-raw-v1-preprocessed
type: Self-GRIT/wikitext-2-raw-v1-preprocessed
metrics:
- name: Accuracy
type: accuracy
value: 0.21399413489736072
---
# Capybara
This model is training from scratch on the Self-GRIT/wikitext-2-raw-v1-preprocessed dataset.
It achieves the following results on the evaluation set:
- Loss: 5.9824
- Accuracy: 0.2140
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
### Framework versions
- Transformers 4.45.0.dev0
- Pytorch 2.1.2+cu118
- Datasets 3.0.0
- Tokenizers 0.19.1
|