Sayan01 commited on
Commit
d20a130
·
verified ·
1 Parent(s): 312e23d

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/data/projects/punim0478/sayantand/Checkpoints/BERT/minilm/12_384",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
@@ -7,12 +7,12 @@
7
  "classifier_dropout": null,
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
- "hidden_size": 384,
11
  "id2label": {
12
  "0": "LABEL_0"
13
  },
14
  "initializer_range": 0.02,
15
- "intermediate_size": 1536,
16
  "label2id": {
17
  "LABEL_0": 0
18
  },
@@ -20,7 +20,7 @@
20
  "max_position_embeddings": 512,
21
  "model_type": "bert",
22
  "num_attention_heads": 12,
23
- "num_hidden_layers": 12,
24
  "pad_token_id": 0,
25
  "position_embedding_type": "absolute",
26
  "problem_type": "regression",
 
1
  {
2
+ "_name_or_path": "/data/projects/punim0478/sayantand/Checkpoints/BERT/minilm/6_768",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
 
7
  "classifier_dropout": null,
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
  "id2label": {
12
  "0": "LABEL_0"
13
  },
14
  "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
  "label2id": {
17
  "LABEL_0": 0
18
  },
 
20
  "max_position_embeddings": 512,
21
  "model_type": "bert",
22
  "num_attention_heads": 12,
23
+ "num_hidden_layers": 6,
24
  "pad_token_id": 0,
25
  "position_embedding_type": "absolute",
26
  "problem_type": "regression",
logs/events.out.tfevents.1724690907.spartan-gpgpu095.hpc.unimelb.edu.au.75681.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfbfc631f0cad5806925e07ab94c1699f8d5d976fac69f9d01ac73cce6b9792a
3
+ size 5607
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f7e3adbe3bd6c5080c5ff3b536f28ad7e704e809b92b2ddf7c95a262496af39
3
- size 133464836
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9855da089994c741fdeb7785a8611ea032c1849eb4a4e67d366b7f9467ac361f
3
+ size 267835308