hmm404/siamese1.0
Browse files- README.md +105 -103
- config.json +1 -1
- model.safetensors +1 -1
- runs/Feb17_12-14-27_873a5f2e1bf6/events.out.tfevents.1739794468.873a5f2e1bf6.495.0 +3 -0
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -4,80 +4,75 @@ tags:
|
|
| 4 |
- sentence-similarity
|
| 5 |
- feature-extraction
|
| 6 |
- generated_from_trainer
|
| 7 |
-
- dataset_size:
|
| 8 |
- loss:ContrastiveLoss
|
| 9 |
-
base_model:
|
| 10 |
widget:
|
| 11 |
-
- source_sentence:
|
| 12 |
-
they are created
|
| 13 |
sentences:
|
| 14 |
-
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
-
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
issue.
|
| 25 |
sentences:
|
| 26 |
-
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
- There is some chicken-or-egg here. One might argue a welfare state leads to less
|
| 36 |
-
poverty, but another would argue welfare states are a luxury only affordable by
|
| 37 |
-
wealthy countries.
|
| 38 |
-
- source_sentence: Renewable energy is a better option for replacing fossil fuels
|
| 39 |
-
than nuclear.
|
| 40 |
sentences:
|
| 41 |
-
-
|
| 42 |
-
|
| 43 |
-
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
believe that Hamlet is mad.
|
| 48 |
sentences:
|
| 49 |
-
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
people. As a result, there is no clear line which courts can use to differentiate
|
| 60 |
-
what is and isn't whitewashing.
|
| 61 |
sentences:
|
| 62 |
-
-
|
| 63 |
-
|
| 64 |
-
|
| 65 |
-
-
|
| 66 |
-
|
| 67 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 68 |
pipeline_tag: sentence-similarity
|
| 69 |
library_name: sentence-transformers
|
| 70 |
---
|
| 71 |
|
| 72 |
-
# SentenceTransformer based on
|
| 73 |
|
| 74 |
-
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [
|
| 75 |
|
| 76 |
## Model Details
|
| 77 |
|
| 78 |
### Model Description
|
| 79 |
- **Model Type:** Sentence Transformer
|
| 80 |
-
- **Base model:** [
|
| 81 |
- **Maximum Sequence Length:** 384 tokens
|
| 82 |
- **Output Dimensionality:** 768 dimensions
|
| 83 |
- **Similarity Function:** Cosine Similarity
|
|
@@ -119,9 +114,9 @@ from sentence_transformers import SentenceTransformer
|
|
| 119 |
model = SentenceTransformer("sentence_transformers_model_id")
|
| 120 |
# Run inference
|
| 121 |
sentences = [
|
| 122 |
-
|
| 123 |
-
'
|
| 124 |
-
'
|
| 125 |
]
|
| 126 |
embeddings = model.encode(sentences)
|
| 127 |
print(embeddings.shape)
|
|
@@ -175,19 +170,19 @@ You can finetune this model on your own dataset.
|
|
| 175 |
|
| 176 |
#### Unnamed Dataset
|
| 177 |
|
| 178 |
-
* Size:
|
| 179 |
* Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>label</code>
|
| 180 |
* Approximate statistics based on the first 1000 samples:
|
| 181 |
| | sentence1 | sentence2 | label |
|
| 182 |
|:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------|
|
| 183 |
| type | string | string | int |
|
| 184 |
-
| details | <ul><li>min: 3 tokens</li><li>mean: 30.
|
| 185 |
* Samples:
|
| 186 |
-
| sentence1
|
| 187 |
-
|
| 188 |
-
| <code>
|
| 189 |
-
| <code>
|
| 190 |
-
| <code>
|
| 191 |
* Loss: [<code>ContrastiveLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
|
| 192 |
```json
|
| 193 |
{
|
|
@@ -323,42 +318,49 @@ You can finetune this model on your own dataset.
|
|
| 323 |
### Training Logs
|
| 324 |
| Epoch | Step | Training Loss |
|
| 325 |
|:------:|:-----:|:-------------:|
|
| 326 |
-
| 0.
|
| 327 |
-
| 0.
|
| 328 |
-
| 0.
|
| 329 |
-
| 0.
|
| 330 |
-
| 0.
|
| 331 |
-
| 0.
|
| 332 |
-
| 0.
|
| 333 |
-
| 0.
|
| 334 |
-
| 0.
|
| 335 |
-
| 0.
|
| 336 |
-
| 0.
|
| 337 |
-
| 0.
|
| 338 |
-
|
|
| 339 |
-
|
|
| 340 |
-
| 1.
|
| 341 |
-
| 1.
|
| 342 |
-
| 1.
|
| 343 |
-
| 1.
|
| 344 |
-
| 1.
|
| 345 |
-
| 1.
|
| 346 |
-
| 1.
|
| 347 |
-
| 1.
|
| 348 |
-
| 1.
|
| 349 |
-
| 1.
|
| 350 |
-
|
|
| 351 |
-
|
|
| 352 |
-
|
|
| 353 |
-
|
|
| 354 |
-
| 2.
|
| 355 |
-
| 2.
|
| 356 |
-
| 2.
|
| 357 |
-
| 2.
|
| 358 |
-
| 2.
|
| 359 |
-
| 2.
|
| 360 |
-
| 2.
|
| 361 |
-
| 2.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 362 |
|
| 363 |
|
| 364 |
### Framework Versions
|
|
|
|
| 4 |
- sentence-similarity
|
| 5 |
- feature-extraction
|
| 6 |
- generated_from_trainer
|
| 7 |
+
- dataset_size:57485
|
| 8 |
- loss:ContrastiveLoss
|
| 9 |
+
base_model: B0ketto/tmp_trainer
|
| 10 |
widget:
|
| 11 |
+
- source_sentence: Russia is acting no differently than other countries.
|
|
|
|
| 12 |
sentences:
|
| 13 |
+
- Animals are not as self aware as humans, so they would not be aware of their coming
|
| 14 |
+
death until it's already happened. Especially as the methods of doing so are quick
|
| 15 |
+
and painless.
|
| 16 |
+
- 'The US has regularly interfered in the elections and governments of foreign states
|
| 17 |
+
consistently throughout its history: in Latin America, the Ukraine and Russia
|
| 18 |
+
to name just a few.'
|
| 19 |
+
- Pedophiles can have abusive ex-partners, parents they ran away from, or generally
|
| 20 |
+
other people that they want to stay hidden from. Publicising their identities
|
| 21 |
+
and locations allows them to be found by those who could inflict harm upon them.
|
| 22 |
+
- source_sentence: Stem cell donations are allowed by MSMs, so should blood.
|
|
|
|
| 23 |
sentences:
|
| 24 |
+
- Some crimes impact people worldwide, such as hacking and virus creation. The cost
|
| 25 |
+
of restitution in these cases would take a lifetime to pay back.
|
| 26 |
+
- Studies indicate that the Muslim nature of countries does not influence the likelihood
|
| 27 |
+
of conflict to occur. Factors such as a lack of economic growth, a demographic
|
| 28 |
+
youth bulge, political repression and ethnic dominance have far more impact in
|
| 29 |
+
this context \(Sørli et al, p. 141, 155; Karakaya, p. 533\).
|
| 30 |
+
- Stem cell donation does not pose a risk of AIDs transmission to the recipient.
|
| 31 |
+
- source_sentence: Global sustainable energy production could be accomplished with
|
| 32 |
+
nuclear fusion.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
sentences:
|
| 34 |
+
- In the long-term, nuclear fusion will be cheaper than nuclear fission
|
| 35 |
+
- Maybe P=NP is a hoax?
|
| 36 |
+
- It is impossible to meaningfully consent to being harmed.
|
| 37 |
+
- source_sentence: European nations have contradictory needs and interests. Making
|
| 38 |
+
those countries agree on crucial points, like economical politics, is impossible
|
| 39 |
+
due to their differences.
|
|
|
|
| 40 |
sentences:
|
| 41 |
+
- The different interests of the nations are actually an argument FOR a USE, because
|
| 42 |
+
then the majority could force the minority to accept changes that are good for
|
| 43 |
+
the EU as a whole, but bad for special interests.
|
| 44 |
+
- Children with vaginal hypoplasia can have vaginoplasty to remedy the diagnosis.
|
| 45 |
+
After this procedure the patients are required to have their vaginas dilated three
|
| 46 |
+
times a week, a very painful procedure, until they become sexually active. \(pp.394\)
|
| 47 |
+
- Prisons would not be "too full" to hold \(violent\) offenders.
|
| 48 |
+
- source_sentence: This argument has the same flaw as the argument for banning alcohol
|
| 49 |
+
\(illuminated by the disaster of prohibition\) - driving things underground makes
|
| 50 |
+
them more not less dangerous.
|
|
|
|
|
|
|
| 51 |
sentences:
|
| 52 |
+
- Bi is the prefix meaning two. If in fact bisexuals claim more than two genders
|
| 53 |
+
than the term to use is not bisexual. Therefore the B in LGBTQ should be changed
|
| 54 |
+
to an M for multi sexual
|
| 55 |
+
- It would be possible for teachers to write supplementary exams for students without
|
| 56 |
+
including The Adventures of Huckleberry Finn.
|
| 57 |
+
- The difference is that prohibition 'prohibited' alcohol entirely, rather than
|
| 58 |
+
licensing and controlling it. Consider cigarettes - they are heavily restricted
|
| 59 |
+
and taxed in many countries, and smoking rates have dropped precipitously, but
|
| 60 |
+
its still more common to buy them at a local store than on the black market. The
|
| 61 |
+
proposal does not suggest 'prohibition' of guns, just restrictions to important
|
| 62 |
+
and appropriate uses by responsible owners.
|
| 63 |
pipeline_tag: sentence-similarity
|
| 64 |
library_name: sentence-transformers
|
| 65 |
---
|
| 66 |
|
| 67 |
+
# SentenceTransformer based on B0ketto/tmp_trainer
|
| 68 |
|
| 69 |
+
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [B0ketto/tmp_trainer](https://huggingface.co/B0ketto/tmp_trainer). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
| 70 |
|
| 71 |
## Model Details
|
| 72 |
|
| 73 |
### Model Description
|
| 74 |
- **Model Type:** Sentence Transformer
|
| 75 |
+
- **Base model:** [B0ketto/tmp_trainer](https://huggingface.co/B0ketto/tmp_trainer) <!-- at revision 4cc993aec9623bb30f9892e2f9b29f6af685ddc0 -->
|
| 76 |
- **Maximum Sequence Length:** 384 tokens
|
| 77 |
- **Output Dimensionality:** 768 dimensions
|
| 78 |
- **Similarity Function:** Cosine Similarity
|
|
|
|
| 114 |
model = SentenceTransformer("sentence_transformers_model_id")
|
| 115 |
# Run inference
|
| 116 |
sentences = [
|
| 117 |
+
'This argument has the same flaw as the argument for banning alcohol \\(illuminated by the disaster of prohibition\\) - driving things underground makes them more not less dangerous.',
|
| 118 |
+
"The difference is that prohibition 'prohibited' alcohol entirely, rather than licensing and controlling it. Consider cigarettes - they are heavily restricted and taxed in many countries, and smoking rates have dropped precipitously, but its still more common to buy them at a local store than on the black market. The proposal does not suggest 'prohibition' of guns, just restrictions to important and appropriate uses by responsible owners.",
|
| 119 |
+
'Bi is the prefix meaning two. If in fact bisexuals claim more than two genders than the term to use is not bisexual. Therefore the B in LGBTQ should be changed to an M for multi sexual',
|
| 120 |
]
|
| 121 |
embeddings = model.encode(sentences)
|
| 122 |
print(embeddings.shape)
|
|
|
|
| 170 |
|
| 171 |
#### Unnamed Dataset
|
| 172 |
|
| 173 |
+
* Size: 57,485 training samples
|
| 174 |
* Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>label</code>
|
| 175 |
* Approximate statistics based on the first 1000 samples:
|
| 176 |
| | sentence1 | sentence2 | label |
|
| 177 |
|:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------|
|
| 178 |
| type | string | string | int |
|
| 179 |
+
| details | <ul><li>min: 3 tokens</li><li>mean: 30.47 tokens</li><li>max: 137 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 35.81 tokens</li><li>max: 162 tokens</li></ul> | <ul><li>0: ~54.70%</li><li>1: ~45.30%</li></ul> |
|
| 180 |
* Samples:
|
| 181 |
+
| sentence1 | sentence2 | label |
|
| 182 |
+
|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------|
|
| 183 |
+
| <code>Large stockpiles might not be available and what little is available will only serve to fuel corruption.</code> | <code>It might serve as a deterrent.</code> | <code>1</code> |
|
| 184 |
+
| <code>Conditional aid works as an excuse to stop providing support for those in need.</code> | <code>Making aid conditional will simply redirect foreign aid funds from corrupt states to non-corrupt states. This is a perfectly reasonable move for donor states to take, and there is no reason why they would need an excuse.</code> | <code>0</code> |
|
| 185 |
+
| <code>Any obstruction in the predictability of their choices \(given all the relevant data\) would have to result from some external source \(randomness, demonic possession, etc\), not determined by the individual's will. The interjection of such an external source, though making the individual less predictable, would arguably make them even less free, since it would restrict their ability to choose according to their nature.</code> | <code>Under the branching model of the universe, every time there is an opportunity for a declension, the individual makes all decisions available to them, each choice making a new universe where the events are carried out, no external force needed.</code> | <code>0</code> |
|
| 186 |
* Loss: [<code>ContrastiveLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
|
| 187 |
```json
|
| 188 |
{
|
|
|
|
| 318 |
### Training Logs
|
| 319 |
| Epoch | Step | Training Loss |
|
| 320 |
|:------:|:-----:|:-------------:|
|
| 321 |
+
| 0.0696 | 500 | 0.0238 |
|
| 322 |
+
| 0.1392 | 1000 | 0.0251 |
|
| 323 |
+
| 0.2087 | 1500 | 0.0261 |
|
| 324 |
+
| 0.2783 | 2000 | 0.0257 |
|
| 325 |
+
| 0.3479 | 2500 | 0.026 |
|
| 326 |
+
| 0.4175 | 3000 | 0.0271 |
|
| 327 |
+
| 0.4871 | 3500 | 0.054 |
|
| 328 |
+
| 0.5566 | 4000 | 0.0384 |
|
| 329 |
+
| 0.6262 | 4500 | 0.036 |
|
| 330 |
+
| 0.6958 | 5000 | 0.0347 |
|
| 331 |
+
| 0.7654 | 5500 | 0.0342 |
|
| 332 |
+
| 0.8350 | 6000 | 0.0331 |
|
| 333 |
+
| 0.9045 | 6500 | 0.033 |
|
| 334 |
+
| 0.9741 | 7000 | 0.0327 |
|
| 335 |
+
| 1.0437 | 7500 | 0.033 |
|
| 336 |
+
| 1.1133 | 8000 | 0.0324 |
|
| 337 |
+
| 1.1829 | 8500 | 0.0325 |
|
| 338 |
+
| 1.2524 | 9000 | 0.0319 |
|
| 339 |
+
| 1.3220 | 9500 | 0.032 |
|
| 340 |
+
| 1.3916 | 10000 | 0.0325 |
|
| 341 |
+
| 1.4612 | 10500 | 0.0321 |
|
| 342 |
+
| 1.5308 | 11000 | 0.032 |
|
| 343 |
+
| 1.6003 | 11500 | 0.0317 |
|
| 344 |
+
| 1.6699 | 12000 | 0.0319 |
|
| 345 |
+
| 1.7395 | 12500 | 0.0316 |
|
| 346 |
+
| 1.8091 | 13000 | 0.0316 |
|
| 347 |
+
| 1.8787 | 13500 | 0.0318 |
|
| 348 |
+
| 1.9482 | 14000 | 0.0317 |
|
| 349 |
+
| 2.0178 | 14500 | 0.0319 |
|
| 350 |
+
| 2.0874 | 15000 | 0.0319 |
|
| 351 |
+
| 2.1570 | 15500 | 0.0315 |
|
| 352 |
+
| 2.2266 | 16000 | 0.0317 |
|
| 353 |
+
| 2.2961 | 16500 | 0.0315 |
|
| 354 |
+
| 2.3657 | 17000 | 0.0316 |
|
| 355 |
+
| 2.4353 | 17500 | 0.0315 |
|
| 356 |
+
| 2.5049 | 18000 | 0.0311 |
|
| 357 |
+
| 2.5745 | 18500 | 0.0297 |
|
| 358 |
+
| 2.6440 | 19000 | 0.0294 |
|
| 359 |
+
| 2.7136 | 19500 | 0.0289 |
|
| 360 |
+
| 2.7832 | 20000 | 0.0286 |
|
| 361 |
+
| 2.8528 | 20500 | 0.0287 |
|
| 362 |
+
| 2.9223 | 21000 | 0.0288 |
|
| 363 |
+
| 2.9919 | 21500 | 0.0286 |
|
| 364 |
|
| 365 |
|
| 366 |
### Framework Versions
|
config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"MPNetModel"
|
| 5 |
],
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "B0ketto/tmp_trainer",
|
| 3 |
"architectures": [
|
| 4 |
"MPNetModel"
|
| 5 |
],
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 437967672
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6cbb02e57c6275c10e778bfdf112231d9de8dde06d5e50989ef9f2dd6ba0f22
|
| 3 |
size 437967672
|
runs/Feb17_12-14-27_873a5f2e1bf6/events.out.tfevents.1739794468.873a5f2e1bf6.495.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9625ce97ee334df227e1f3cbd06daf6bf371ade6e53315d1471afef292cde3ac
|
| 3 |
+
size 13852
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5560
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c398cd2d5e4d21698eaa16ec0e94ca18789de8b4a972809ccfc16c63280bdc18
|
| 3 |
size 5560
|