File size: 1,327 Bytes
eaa0e07
 
3d1e3e5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
 
 
 
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
eaa0e07
3d1e3e5
 
 
 
 
 
 
 
eaa0e07
3d1e3e5
eaa0e07
 
 
3d1e3e5
eaa0e07
3d1e3e5
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
---
library_name: transformers
license: apache-2.0
base_model: training_from_scratch
tags:
- generated_from_trainer
- trainging_from_scratch
datasets:
- Self-GRIT/wikitext-2-raw-v1-preprocessed
metrics:
- accuracy
model-index:
- name: Capybara
  results:
  - task:
      name: Causal Language Modeling
      type: text-generation
    dataset:
      name: Self-GRIT/wikitext-2-raw-v1-preprocessed
      type: Self-GRIT/wikitext-2-raw-v1-preprocessed
    metrics:
    - name: Accuracy
      type: accuracy
      value: 0.21399413489736072
---
# Capybara

This model is training from scratch on the Self-GRIT/wikitext-2-raw-v1-preprocessed dataset.
It achieves the following results on the evaluation set:
- Loss: 5.9824
- Accuracy: 0.2140

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0

### Training results



### Framework versions

- Transformers 4.45.0.dev0
- Pytorch 2.1.2+cu118
- Datasets 3.0.0
- Tokenizers 0.19.1