Wasifjafri commited on
Commit
9cd0ade
·
verified ·
1 Parent(s): 48f9671

Upload TFDistilBertForSequenceClassification

Browse files
Files changed (3) hide show
  1. README.md +5 -8
  2. config.json +23 -24
  3. tf_model.h5 +3 -0
README.md CHANGED
@@ -1,12 +1,9 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: distilbert-base-uncased
4
- library_name: transformers
5
  tags:
6
- - text-classification
7
- datasets:
8
- - imdb
9
- inference: true
10
  model-index:
11
  - name: test
12
  results: []
@@ -47,6 +44,6 @@ The following hyperparameters were used during training:
47
 
48
  ### Framework versions
49
 
50
- - Transformers 4.37.0
51
- - TensorFlow 2.10.0
52
- - Tokenizers 0.15.2
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: distilbert-base-uncased
 
5
  tags:
6
+ - generated_from_keras_callback
 
 
 
7
  model-index:
8
  - name: test
9
  results: []
 
44
 
45
  ### Framework versions
46
 
47
+ - Transformers 4.48.3
48
+ - TensorFlow 2.18.0
49
+ - Tokenizers 0.21.0
config.json CHANGED
@@ -1,24 +1,23 @@
1
- {
2
- "_name_or_path": "Wasifjafri/test",
3
- "activation": "gelu",
4
- "architectures": [
5
- "DistilBertForSequenceClassification"
6
- ],
7
- "attention_dropout": 0.1,
8
- "dim": 768,
9
- "dropout": 0.1,
10
- "hidden_dim": 3072,
11
- "initializer_range": 0.02,
12
- "max_position_embeddings": 512,
13
- "model_type": "distilbert",
14
- "n_heads": 12,
15
- "n_layers": 6,
16
- "pad_token_id": 0,
17
- "qa_dropout": 0.1,
18
- "seq_classif_dropout": 0.2,
19
- "sinusoidal_pos_embds": false,
20
- "tie_weights_": true,
21
- "torch_dtype": "float32",
22
- "transformers_version": "4.37.0",
23
- "vocab_size": 30522
24
- }
 
1
+ {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "pad_token_id": 0,
17
+ "qa_dropout": 0.1,
18
+ "seq_classif_dropout": 0.2,
19
+ "sinusoidal_pos_embds": false,
20
+ "tie_weights_": true,
21
+ "transformers_version": "4.48.3",
22
+ "vocab_size": 30522
23
+ }
 
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efa58ca0e33ba94961659d4649bbd9cb5e725e1c5d3c674f3f2a7f13d7234564
3
+ size 267951808