alinashrestha commited on
Commit
2b73a65
·
verified ·
1 Parent(s): 9b0bc2d

alinashrestha/patentsbert-green-gold

Browse files
Files changed (4) hide show
  1. README.md +58 -0
  2. config.json +26 -0
  3. model.safetensors +3 -0
  4. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - generated_from_trainer
5
+ model-index:
6
+ - name: tmp_trainer
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
+
13
+ # tmp_trainer
14
+
15
+ This model was trained from scratch on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - eval_loss: 0.7463
18
+ - eval_model_preparation_time: 0.0036
19
+ - eval_accuracy: 0.49
20
+ - eval_precision: 0.0577
21
+ - eval_recall: 0.6
22
+ - eval_f1: 0.1053
23
+ - eval_runtime: 18.6792
24
+ - eval_samples_per_second: 5.354
25
+ - eval_steps_per_second: 0.696
26
+ - step: 0
27
+
28
+ ## Model description
29
+
30
+ More information needed
31
+
32
+ ## Intended uses & limitations
33
+
34
+ More information needed
35
+
36
+ ## Training and evaluation data
37
+
38
+ More information needed
39
+
40
+ ## Training procedure
41
+
42
+ ### Training hyperparameters
43
+
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 5e-05
46
+ - train_batch_size: 8
47
+ - eval_batch_size: 8
48
+ - seed: 42
49
+ - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
50
+ - lr_scheduler_type: linear
51
+ - num_epochs: 3.0
52
+
53
+ ### Framework versions
54
+
55
+ - Transformers 5.2.0
56
+ - Pytorch 2.9.1+cu128
57
+ - Datasets 4.5.0
58
+ - Tokenizers 0.22.1
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MPNetForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "dtype": "float32",
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "mpnet",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "problem_type": "single_label_classification",
21
+ "relative_attention_num_buckets": 32,
22
+ "tie_word_embeddings": true,
23
+ "transformers_version": "5.2.0",
24
+ "use_cache": false,
25
+ "vocab_size": 30527
26
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fef43a89f5d855f8331d1aff2cb3e30f3115d1318f3fd6c105a2b68954e25751
3
+ size 437975176
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba8262f63c0ec1be09dd9d7a275daebcd9bc2ac7af8c61e7e8b2dc1dad42e4cd
3
+ size 5201