Ory999 commited on
Commit
e94c4e3
·
verified ·
1 Parent(s): d3b1aa8

Upload 14 files

Browse files
final_classifier.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02467c4ebc63c582e4624fdadfb424cd2906ca732e0ca4160bad480eb5eafbf7
3
+ size 7007
hitl_green_100_final.csv ADDED
The diff for this file is too large to render. See raw diff
 
patentsberta_finetuned_final/1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
patentsberta_finetuned_final/README.md ADDED
@@ -0,0 +1,446 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - generated_from_trainer
7
+ - dataset_size:2000
8
+ - loss:CosineSimilarityLoss
9
+ base_model: AI-Growth-Lab/PatentSBERTa
10
+ widget:
11
+ - source_sentence: '1. A method, comprising: adjusting a wastegate actuator coupled
12
+ to a wastegate valve in an engine exhaust to control an engine boost level of
13
+ an engine, the adjustment based on a magnetic field of a magnet in the wastegate
14
+ actuator corrected based on a magnet temperature; wherein the adjustment includes
15
+ adjusting a current supplied to the actuator, and wherein the adjustment is further
16
+ based on an actuator winding resistance, the resistance based on the magnet temperature.'
17
+ sentences:
18
+ - '1. A beverage bottle for a portable electronic device, comprising: a hand-held
19
+ container which comprises a body unit having a fluid compartment for containing
20
+ beverage therein, and a spout unit, having a mouth piece, detachably coupled at
21
+ said body unit to enclose said fluid compartment; and an accessible station which
22
+ comprises a supporting frame integrated with one of said body unit and said spout
23
+ unit of said hand-held container for stably supporting the portable electronic
24
+ device at a position that the '
25
+ - '1. A composition that comprises an immunotherapeutic agent in admixture with
26
+ a compound having a structure: wherein n is 1, 2, or 3; wherein the immunotherapeutic
27
+ agent comprises a cancer vaccine; a cancer antigen; or at least one antibody selected
28
+ from the group consisting of an anti-programmed cell death protein 1 (PD-1) antibody
29
+ and an anti-programmed cell death ligand (PDL-1) antibody.'
30
+ - '1. A first terminal apparatus of a plurality of terminal apparatuses that are
31
+ mutually connected via a network, the plurality of terminal apparatuses being
32
+ allotted unique first identification information, the first terminal apparatus
33
+ carrying out a transmission and reception of contents information among the plurality
34
+ of terminal apparatuses via an overlay network that is formed by the plurality
35
+ of terminal apparatuses, the first terminal apparatus comprising: a routing table
36
+ storage section configured to '
37
+ - source_sentence: '1. A fuel pellet comprising: petroleum coke, having a sulfur content
38
+ up to 5.5%; a biomass constituent; an alkali constituent adapted to capture SO2
39
+ emissions by reacting with sulfur of the petroleum coke upon burn of the pellet;
40
+ and an iron oxide catalyst capturing in the range of 90 to 95.3 percent of SO2
41
+ emissions;'
42
+ sentences:
43
+ - '1. A firewall apparatus comprising: a memory storing a plurality of firewall
44
+ configurations defined by a plurality of different customers, each firewall configuration
45
+ of the plurality of firewall configurations comprising (i) a production profile
46
+ comprising a different first set of firewall rules or policies actively protecting
47
+ customer content and services against real-world traffic and (ii) an audit profile
48
+ comprising a different second set of firewall rules or policies testing new rules
49
+ or policies again'
50
+ - '1. An apparatus for creating and managing security policies for data leakage
51
+ prevention (DLP), the apparatus comprising: a computer server; server code on
52
+ the computer server; a database which is accessible by the server code and which
53
+ stores a three-layer structure of objects comprising a first layer of digital
54
+ asset objects which describe sensitive information contained in a file, a second
55
+ layer of content template objects which are each associated with at least one
56
+ digital asset object, and a third layer'
57
+ - '1. An imaging device comprising: a plurality of pixels disposed to form a matrix
58
+ having pixel rows, the pixels including a driving section configured to apply
59
+ an electric potential to said photoelectric conversion film on each of said pixel
60
+ rows at least having read timings different from each other with a predetermined
61
+ timing outside an exposure period of said pixels in a direction opposite to that
62
+ of an electric potential applied to said photoelectric conversion film during
63
+ said exposure period of said pi'
64
+ - source_sentence: '1. An energy absorber, comprising: a plurality of crush lobes
65
+ that deform plastically upon impact to absorb energy, wherein the crush lobes
66
+ include a base and sides extending from the base to an outer wall, wherein the
67
+ base, sides and outer wall comprise a first thermoplastic material; and a composite
68
+ insert in the energy absorber, wherein the insert comprises a second plastic material
69
+ and reinforcement, wherein the second plastic material is different than the first
70
+ thermoplastic material, wherein the inse'
71
+ sentences:
72
+ - '1. A method of tiling a roof, comprising: positioning at least one row of tiles
73
+ across a surface in a row direction, a first tile of the at least one row of tiles
74
+ comprising a rain lock adapted to interlock with a second tile adjacent to and
75
+ within the same row as the first tile, said rain lock having a lower end, an upper
76
+ end, and a weakened portion proximate the lower end to allow for removal of a
77
+ predetermined portion of the rain lock; removing the predetermined portion of
78
+ the rain lock; and staggering t'
79
+ - '1. A vending machine for scrubs comprising: a housing configured to store a plurality
80
+ of different sized scrubs; a scrub dispenser carried by said housing; a labeler
81
+ carried by said housing; and a controller carried by said housing and configured
82
+ to wherein said labeler is configured to print indicia on a first badge substrate
83
+ to be attachable to a second badge substrate carried by the selected scrub to
84
+ define the time expiring badge.'
85
+ - 1. A method of treatment comprising the administering of a pharmaceutical composition
86
+ comprising a purified extract with the secondary fractionation (ATC2) from the
87
+ extract of Pseudolysimachion rotundum var subintegrum , comprising 30%-60% (w/w)
88
+ verproside, 0.5%-10% (w/w) veratric acid, 2%-20% (w/w) catalposide, 1%-10% (w/w)
89
+ picroside II, 1%-10% (w/w) isovanilloyl catalpol and 2%-20% (w/w) 6-O-veratroyl
90
+ catalpol based on the weight of total extract (100%) of Pseudolysimachion rotundum
91
+ var subintegrum to a s
92
+ - source_sentence: '1. A board reinforcing structure comprising: a reinforcing member
93
+ configured to reinforce a circuit board having an electronic component mounted
94
+ thereon, wherein the electronic component has a plurality of electrodes arranged
95
+ within a rectangular bonding region on a first surface of the circuit board, wherein
96
+ the reinforcing member is bonded on a second surface of the circuit board opposite
97
+ to the first surface at positions corresponding to four corners of the rectangular
98
+ bonding region; wherein the reinfor'
99
+ sentences:
100
+ - '1. A refrigeration unit comprising: (a) a countermeasure to prevent heat exchange
101
+ efficiency from decreasing due to a temperature glide in a heat exchanger; and
102
+ (b) a refrigerant composition comprising a refrigerant mixture, the refrigerant
103
+ mixture consisting essentially of 30 to 35 mass % of difluoromethane (HFC32) and
104
+ 70 to 65 mass % of 2,3,3,3-tetrafluoropropene (HFO1234yf), based on a total of
105
+ HFC32 and HFO1234yf in the refrigerant composition wherein the total amount of
106
+ HFC32 and HFO1234yf is taken as '
107
+ - '1. A style masking engine for a web server, comprising: a receiving unit programmed
108
+ to receive, at the web server from a client device responsive to user selection
109
+ within a user interface of a source application output window and a target application
110
+ output window of a web page displayed by the client device, a user request to
111
+ modify a style of content displayed within the user-selected target application
112
+ output window of the web page displayed by the client device; a web application
113
+ aggregator programmed t'
114
+ - '1. A steam reforming system comprising: a) a kiln, comprising a susceptor tube;
115
+ a kiln inlet for receiving a feedstock; a conveyor for transporting said feedstock
116
+ through said kiln; b) a steam reforming reactor comprising a reformer tube; a
117
+ reactor inlet in fluid communication with said first kiln outlet for gaseous product
118
+ of said kiln; and a reactor outlet for a gaseous product; c) an inductive heating
119
+ means contiguous to said reformer tube for providing heat to said steam reforming
120
+ reactor; and d) said s'
121
+ - source_sentence: 1. An industrial truck with an electric travel drive ( 14 ) and
122
+ controls ( 20 ) that can switch the electric travel drive ( 14 ) to regenerative
123
+ operation to brake the industrial truck, characterized in that an eddy current
124
+ brake ( wherein the controls ( further characterized in that the controls ( further
125
+ wherein the controls determine a setpoint for the eddy current brake in a manner
126
+ such that the braking torque resulting from the regenerative operation of the
127
+ travel drive and the eddy current brake is co
128
+ sentences:
129
+ - '1. A method for reallocating primary and secondary destinations in a virtual
130
+ server for one or more segmented servers, the method comprising the steps of:
131
+ a computer allocating buckets to at least two data structures designated to the
132
+ one or more segmented servers, wherein one of the at least two data structures
133
+ includes one less bucket than the other data structure; the computer allocating
134
+ primary destinations to the buckets, wherein a number, the number identifying
135
+ the bucket destinations, of each primary'
136
+ - '1. An apparatus including a system for enabling communications via a power line
137
+ conveying DC power from multiple DC power sources, comprising: first and second
138
+ power line electrodes for connecting to first and second ends of a power line
139
+ conveying DC power from a plurality of serially coupled DC power sources; receiver
140
+ circuitry coupled to said first and second power line electrodes, and responsive
141
+ to one or more power line voltages at said first and second power line electrodes
142
+ by providing first and secon'
143
+ - '1. A hydraulic control system, comprising: a hydraulic circuit; a pump configured
144
+ to supply pressurized fluid to the hydraulic circuit; a first fluid actuator fluidly
145
+ connected to receive pressurized fluid from the hydraulic circuit; a first valve
146
+ arrangement movable to control a flow of fluid to the first fluid actuator; a
147
+ second fluid actuator fluidly connected to receive pressurized fluid from the
148
+ hydraulic circuit; a second valve arrangement movable to control a flow of fluid
149
+ to the second fluid actuato'
150
+ pipeline_tag: sentence-similarity
151
+ library_name: sentence-transformers
152
+ ---
153
+
154
+ # SentenceTransformer based on AI-Growth-Lab/PatentSBERTa
155
+
156
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [AI-Growth-Lab/PatentSBERTa](https://huggingface.co/AI-Growth-Lab/PatentSBERTa). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
157
+
158
+ ## Model Details
159
+
160
+ ### Model Description
161
+ - **Model Type:** Sentence Transformer
162
+ - **Base model:** [AI-Growth-Lab/PatentSBERTa](https://huggingface.co/AI-Growth-Lab/PatentSBERTa) <!-- at revision 3ff1d553c861d8f5bfd902333d97fc95eb6b4c8f -->
163
+ - **Maximum Sequence Length:** 512 tokens
164
+ - **Output Dimensionality:** 768 dimensions
165
+ - **Similarity Function:** Cosine Similarity
166
+ <!-- - **Training Dataset:** Unknown -->
167
+ <!-- - **Language:** Unknown -->
168
+ <!-- - **License:** Unknown -->
169
+
170
+ ### Model Sources
171
+
172
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
173
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
174
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
175
+
176
+ ### Full Model Architecture
177
+
178
+ ```
179
+ SentenceTransformer(
180
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel
181
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
182
+ )
183
+ ```
184
+
185
+ ## Usage
186
+
187
+ ### Direct Usage (Sentence Transformers)
188
+
189
+ First install the Sentence Transformers library:
190
+
191
+ ```bash
192
+ pip install -U sentence-transformers
193
+ ```
194
+
195
+ Then you can load this model and run inference.
196
+ ```python
197
+ from sentence_transformers import SentenceTransformer
198
+
199
+ # Download from the 🤗 Hub
200
+ model = SentenceTransformer("sentence_transformers_model_id")
201
+ # Run inference
202
+ sentences = [
203
+ '1. An industrial truck with an electric travel drive ( 14 ) and controls ( 20 ) that can switch the electric travel drive ( 14 ) to regenerative operation to brake the industrial truck, characterized in that an eddy current brake ( wherein the controls ( further characterized in that the controls ( further wherein the controls determine a setpoint for the eddy current brake in a manner such that the braking torque resulting from the regenerative operation of the travel drive and the eddy current brake is co',
204
+ '1. A method for reallocating primary and secondary destinations in a virtual server for one or more segmented servers, the method comprising the steps of: a computer allocating buckets to at least two data structures designated to the one or more segmented servers, wherein one of the at least two data structures includes one less bucket than the other data structure; the computer allocating primary destinations to the buckets, wherein a number, the number identifying the bucket destinations, of each primary',
205
+ '1. A hydraulic control system, comprising: a hydraulic circuit; a pump configured to supply pressurized fluid to the hydraulic circuit; a first fluid actuator fluidly connected to receive pressurized fluid from the hydraulic circuit; a first valve arrangement movable to control a flow of fluid to the first fluid actuator; a second fluid actuator fluidly connected to receive pressurized fluid from the hydraulic circuit; a second valve arrangement movable to control a flow of fluid to the second fluid actuato',
206
+ ]
207
+ embeddings = model.encode(sentences)
208
+ print(embeddings.shape)
209
+ # [3, 768]
210
+
211
+ # Get the similarity scores for the embeddings
212
+ similarities = model.similarity(embeddings, embeddings)
213
+ print(similarities.shape)
214
+ # [3, 3]
215
+ ```
216
+
217
+ <!--
218
+ ### Direct Usage (Transformers)
219
+
220
+ <details><summary>Click to see the direct usage in Transformers</summary>
221
+
222
+ </details>
223
+ -->
224
+
225
+ <!--
226
+ ### Downstream Usage (Sentence Transformers)
227
+
228
+ You can finetune this model on your own dataset.
229
+
230
+ <details><summary>Click to expand</summary>
231
+
232
+ </details>
233
+ -->
234
+
235
+ <!--
236
+ ### Out-of-Scope Use
237
+
238
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
239
+ -->
240
+
241
+ <!--
242
+ ## Bias, Risks and Limitations
243
+
244
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
245
+ -->
246
+
247
+ <!--
248
+ ### Recommendations
249
+
250
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
251
+ -->
252
+
253
+ ## Training Details
254
+
255
+ ### Training Dataset
256
+
257
+ #### Unnamed Dataset
258
+
259
+ * Size: 2,000 training samples
260
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
261
+ * Approximate statistics based on the first 1000 samples:
262
+ | | sentence_0 | sentence_1 | label |
263
+ |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:---------------------------------------------------------------|
264
+ | type | string | string | float |
265
+ | details | <ul><li>min: 17 tokens</li><li>mean: 98.38 tokens</li><li>max: 313 tokens</li></ul> | <ul><li>min: 14 tokens</li><li>mean: 98.92 tokens</li><li>max: 313 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.52</li><li>max: 1.0</li></ul> |
266
+ * Samples:
267
+ | sentence_0 | sentence_1 | label |
268
+ |:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------|
269
+ | <code>1. A cell capacity adjusting device for reducing fluctuations in state of charge (SOC) among cells of a battery pack, which is formed by connecting a plurality of cells in series, during suspension of operation of electrically-powered equipment whose main power source is the battery pack, the cell capacity adjusting device comprising: cell capacity target value setting means for setting a cell capacity adjustment target voltage; cell selection means for detecting an open-circuit voltage of each of the cells</code> | <code>1. A semiconductor package comprising: an encapsulant comprising: a semiconductor substrate in the encapsulant, the substrate comprising: a plurality of electrically conductive interconnects extending to the lower side of the encapsulant for electrically connecting the die contacts of the substrate to the surface; and a heat spreader in thermal contact with the backside of the semiconductor substrate for aiding thermal connection to the surface.</code> | <code>0.0</code> |
270
+ | <code>1. A method comprising: receiving a request from a user of a mobile device to track the mobile device; generating, in response to the request, a control signal to remotely activate an application on the mobile device for controlling an audio interface or an imaging interface of the mobile device to capture a signal from the audio interface or the imaging interface; receiving one or more signals captured from the audio interface or the imaging interface of the mobile device; determining to present the receiv</code> | <code>1. A method for monitoring the condition of an article comprising the steps of: a. affixing a covering made of a substrate that is coated with a layer of an electrically conductive material and forming a single electrically conductive surface that extends over the entire covering and has an electrical resistance, said covering being configured to at least partially encapsulate the article such that the article cannot be tampered with, without modifying the electrical resistance of said covering; b. producin</code> | <code>1.0</code> |
271
+ | <code>1. A system to store and to transmit electrical power, comprising a storage system comprising a central storage system used to store electrical power of at least one power source, an HVDC power transmission system to which the central storage system is connected, a first bidirectional converter connected at one side to the HVDC power transmission system, at least one an AC network connected to another side of the first bidirectional converter, a first load connected to the AC network and adapted to both, re</code> | <code>1. An elliptically polarized dielectric block antenna comprising: a linearly polarized dielectric block antenna; and a wave polarizer integrated with the linearly polarized dielectric block antenna, wherein the wave polarizer converts the linearly polarized wave of the linearly polarized dielectric block antenna into an elliptically polarized wave; wherein the wave polarizer is integrated with the dielectric block of the linearly polarized dielectric block antenna; and wherein the wave polarizer comprises o</code> | <code>0.0</code> |
272
+ * Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
273
+ ```json
274
+ {
275
+ "loss_fct": "torch.nn.modules.loss.MSELoss"
276
+ }
277
+ ```
278
+
279
+ ### Training Hyperparameters
280
+ #### Non-Default Hyperparameters
281
+
282
+ - `per_device_train_batch_size`: 64
283
+ - `per_device_eval_batch_size`: 64
284
+ - `multi_dataset_batch_sampler`: round_robin
285
+
286
+ #### All Hyperparameters
287
+ <details><summary>Click to expand</summary>
288
+
289
+ - `overwrite_output_dir`: False
290
+ - `do_predict`: False
291
+ - `eval_strategy`: no
292
+ - `prediction_loss_only`: True
293
+ - `per_device_train_batch_size`: 64
294
+ - `per_device_eval_batch_size`: 64
295
+ - `per_gpu_train_batch_size`: None
296
+ - `per_gpu_eval_batch_size`: None
297
+ - `gradient_accumulation_steps`: 1
298
+ - `eval_accumulation_steps`: None
299
+ - `torch_empty_cache_steps`: None
300
+ - `learning_rate`: 5e-05
301
+ - `weight_decay`: 0.0
302
+ - `adam_beta1`: 0.9
303
+ - `adam_beta2`: 0.999
304
+ - `adam_epsilon`: 1e-08
305
+ - `max_grad_norm`: 1
306
+ - `num_train_epochs`: 3
307
+ - `max_steps`: -1
308
+ - `lr_scheduler_type`: linear
309
+ - `lr_scheduler_kwargs`: {}
310
+ - `warmup_ratio`: 0.0
311
+ - `warmup_steps`: 0
312
+ - `log_level`: passive
313
+ - `log_level_replica`: warning
314
+ - `log_on_each_node`: True
315
+ - `logging_nan_inf_filter`: True
316
+ - `save_safetensors`: True
317
+ - `save_on_each_node`: False
318
+ - `save_only_model`: False
319
+ - `restore_callback_states_from_checkpoint`: False
320
+ - `no_cuda`: False
321
+ - `use_cpu`: False
322
+ - `use_mps_device`: False
323
+ - `seed`: 42
324
+ - `data_seed`: None
325
+ - `jit_mode_eval`: False
326
+ - `use_ipex`: False
327
+ - `bf16`: False
328
+ - `fp16`: False
329
+ - `fp16_opt_level`: O1
330
+ - `half_precision_backend`: auto
331
+ - `bf16_full_eval`: False
332
+ - `fp16_full_eval`: False
333
+ - `tf32`: None
334
+ - `local_rank`: 0
335
+ - `ddp_backend`: None
336
+ - `tpu_num_cores`: None
337
+ - `tpu_metrics_debug`: False
338
+ - `debug`: []
339
+ - `dataloader_drop_last`: False
340
+ - `dataloader_num_workers`: 0
341
+ - `dataloader_prefetch_factor`: None
342
+ - `past_index`: -1
343
+ - `disable_tqdm`: False
344
+ - `remove_unused_columns`: True
345
+ - `label_names`: None
346
+ - `load_best_model_at_end`: False
347
+ - `ignore_data_skip`: False
348
+ - `fsdp`: []
349
+ - `fsdp_min_num_params`: 0
350
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
351
+ - `fsdp_transformer_layer_cls_to_wrap`: None
352
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
353
+ - `deepspeed`: None
354
+ - `label_smoothing_factor`: 0.0
355
+ - `optim`: adamw_torch
356
+ - `optim_args`: None
357
+ - `adafactor`: False
358
+ - `group_by_length`: False
359
+ - `length_column_name`: length
360
+ - `ddp_find_unused_parameters`: None
361
+ - `ddp_bucket_cap_mb`: None
362
+ - `ddp_broadcast_buffers`: False
363
+ - `dataloader_pin_memory`: True
364
+ - `dataloader_persistent_workers`: False
365
+ - `skip_memory_metrics`: True
366
+ - `use_legacy_prediction_loop`: False
367
+ - `push_to_hub`: False
368
+ - `resume_from_checkpoint`: None
369
+ - `hub_model_id`: None
370
+ - `hub_strategy`: every_save
371
+ - `hub_private_repo`: False
372
+ - `hub_always_push`: False
373
+ - `gradient_checkpointing`: False
374
+ - `gradient_checkpointing_kwargs`: None
375
+ - `include_inputs_for_metrics`: False
376
+ - `eval_do_concat_batches`: True
377
+ - `fp16_backend`: auto
378
+ - `push_to_hub_model_id`: None
379
+ - `push_to_hub_organization`: None
380
+ - `mp_parameters`:
381
+ - `auto_find_batch_size`: False
382
+ - `full_determinism`: False
383
+ - `torchdynamo`: None
384
+ - `ray_scope`: last
385
+ - `ddp_timeout`: 1800
386
+ - `torch_compile`: False
387
+ - `torch_compile_backend`: None
388
+ - `torch_compile_mode`: None
389
+ - `dispatch_batches`: None
390
+ - `split_batches`: None
391
+ - `include_tokens_per_second`: False
392
+ - `include_num_input_tokens_seen`: False
393
+ - `neftune_noise_alpha`: None
394
+ - `optim_target_modules`: None
395
+ - `batch_eval_metrics`: False
396
+ - `eval_on_start`: False
397
+ - `eval_use_gather_object`: False
398
+ - `prompts`: None
399
+ - `batch_sampler`: batch_sampler
400
+ - `multi_dataset_batch_sampler`: round_robin
401
+
402
+ </details>
403
+
404
+ ### Framework Versions
405
+ - Python: 3.12.10
406
+ - Sentence Transformers: 3.4.1
407
+ - Transformers: 4.44.2
408
+ - PyTorch: 2.4.1+cpu
409
+ - Accelerate: 1.12.0
410
+ - Datasets: 4.3.0
411
+ - Tokenizers: 0.19.1
412
+
413
+ ## Citation
414
+
415
+ ### BibTeX
416
+
417
+ #### Sentence Transformers
418
+ ```bibtex
419
+ @inproceedings{reimers-2019-sentence-bert,
420
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
421
+ author = "Reimers, Nils and Gurevych, Iryna",
422
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
423
+ month = "11",
424
+ year = "2019",
425
+ publisher = "Association for Computational Linguistics",
426
+ url = "https://arxiv.org/abs/1908.10084",
427
+ }
428
+ ```
429
+
430
+ <!--
431
+ ## Glossary
432
+
433
+ *Clearly define terms in order to be accessible across audiences.*
434
+ -->
435
+
436
+ <!--
437
+ ## Model Card Authors
438
+
439
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
440
+ -->
441
+
442
+ <!--
443
+ ## Model Card Contact
444
+
445
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
446
+ -->
patentsberta_finetuned_final/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "AI-Growth-Lab/PatentSBERTa",
3
+ "architectures": [
4
+ "MPNetModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "mpnet",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "relative_attention_num_buckets": 32,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.44.2",
23
+ "vocab_size": 30527
24
+ }
patentsberta_finetuned_final/config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.4.1",
4
+ "transformers": "4.44.2",
5
+ "pytorch": "2.4.1+cpu"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
+ }
patentsberta_finetuned_final/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e28202556120966b12e1eb36376e58ead315bebb63ca3d735aaf7521e16a279c
3
+ size 437967672
patentsberta_finetuned_final/modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
patentsberta_finetuned_final/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
patentsberta_finetuned_final/special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
patentsberta_finetuned_final/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
patentsberta_finetuned_final/tokenizer_config.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "104": {
36
+ "content": "[UNK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "30526": {
44
+ "content": "<mask>",
45
+ "lstrip": true,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ }
51
+ },
52
+ "bos_token": "<s>",
53
+ "clean_up_tokenization_spaces": true,
54
+ "cls_token": "<s>",
55
+ "do_lower_case": true,
56
+ "eos_token": "</s>",
57
+ "mask_token": "<mask>",
58
+ "max_length": 512,
59
+ "model_max_length": 512,
60
+ "pad_to_multiple_of": null,
61
+ "pad_token": "<pad>",
62
+ "pad_token_type_id": 0,
63
+ "padding_side": "right",
64
+ "sep_token": "</s>",
65
+ "stride": 0,
66
+ "strip_accents": null,
67
+ "tokenize_chinese_chars": true,
68
+ "tokenizer_class": "MPNetTokenizer",
69
+ "truncation_side": "right",
70
+ "truncation_strategy": "longest_first",
71
+ "unk_token": "[UNK]"
72
+ }
patentsberta_finetuned_final/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
qlora_outputs.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98255cf9b1ea80fbf3b0d372f3627a10553ffea700a986a0d4c1b8dcb55a3bef
3
+ size 891140146