File size: 2,887 Bytes
730ae20
 
2fd2807
730ae20
 
 
 
 
 
 
 
 
 
 
 
2fd2807
730ae20
d82c6e8
 
b0a498d
d82c6e8
730ae20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2fd2807
730ae20
 
 
 
ecb21b6
 
2fd2807
fe404e2
b408ea7
a43b6f1
2e0a93b
ea3aa1b
c7d662c
ab33078
326d774
 
 
8be1524
294512e
bf8b375
 
1a0c0a1
eb42471
30f119f
b0a498d
d82c6e8
730ae20
 
 
 
603b8da
730ae20
 
f2b8e3e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
---
license: apache-2.0
base_model: ayshi/basic_distil
tags:
- generated_from_keras_callback
model-index:
- name: ayshi/basic_distil
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# ayshi/basic_distil

This model is a fine-tuned version of [ayshi/basic_distil](https://huggingface.co/ayshi/basic_distil) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.0114
- Validation Loss: 1.0035
- Train Accuracy: 0.7911
- Epoch: 19

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 640, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
- training_precision: float32

### Training results

| Train Loss | Validation Loss | Train Accuracy | Epoch |
|:----------:|:---------------:|:--------------:|:-----:|
| 0.3282     | 0.7966          | 0.7644         | 0     |
| 0.2163     | 0.8219          | 0.7778         | 1     |
| 0.1323     | 0.8099          | 0.7778         | 2     |
| 0.0933     | 0.8337          | 0.7956         | 3     |
| 0.0619     | 0.9082          | 0.7689         | 4     |
| 0.0461     | 0.9380          | 0.7778         | 5     |
| 0.0495     | 0.9502          | 0.7556         | 6     |
| 0.0301     | 0.9445          | 0.7733         | 7     |
| 0.0249     | 0.9578          | 0.8            | 8     |
| 0.0209     | 0.9663          | 0.7911         | 9     |
| 0.0200     | 0.9828          | 0.7778         | 10    |
| 0.0159     | 0.9987          | 0.7689         | 11    |
| 0.0163     | 1.0120          | 0.7689         | 12    |
| 0.0142     | 1.0020          | 0.7956         | 13    |
| 0.0153     | 1.0270          | 0.7911         | 14    |
| 0.0142     | 1.0159          | 0.7822         | 15    |
| 0.0130     | 1.0049          | 0.7911         | 16    |
| 0.0129     | 1.0085          | 0.7956         | 17    |
| 0.0099     | 1.0033          | 0.7911         | 18    |
| 0.0114     | 1.0035          | 0.7911         | 19    |


### Framework versions

- Transformers 4.34.0
- TensorFlow 2.13.0
- Datasets 2.14.5
- Tokenizers 0.14.1