Upload 10 files
Browse files- README.md +626 -0
- config.json +26 -0
- config_sentence_transformers.json +14 -0
- model.safetensors +3 -0
- modules.json +20 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +63 -0
- vocab.txt +0 -0
README.md
ADDED
|
@@ -0,0 +1,626 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- sentence-transformers
|
| 4 |
+
- sentence-similarity
|
| 5 |
+
- feature-extraction
|
| 6 |
+
- dense
|
| 7 |
+
- generated_from_trainer
|
| 8 |
+
- dataset_size:11180
|
| 9 |
+
- loss:CosineSimilarityLoss
|
| 10 |
+
widget:
|
| 11 |
+
- source_sentence: '" My cousin said he share hoes with his brothers. He said sharing
|
| 12 |
+
is caring and he love his brothers 😂"'
|
| 13 |
+
sentences:
|
| 14 |
+
- Clerical Error Led to Costa Rica's First Legal Gay Marriage
|
| 15 |
+
- 'UPDATE: The leasing office is going to tape notes on ALL tenants doors in our
|
| 16 |
+
building as to not single anyone out. However, we were not the first to complain.
|
| 17 |
+
Our maintenance guy also caught them smoking in the breezeway of our building
|
| 18 |
+
and told them to cut it out. Hopefully the note helps, if not, I don''t care enough
|
| 19 |
+
to make it a bigger issue and I''m only here for the next 8 months so smoke em
|
| 20 |
+
if you got em I guess.
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
Alright bare/bear with me here...
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
My husband and I live in a pretty decent apartment complex, we''re on the top
|
| 27 |
+
floor, and we are cool with most of our neighbors. We''re 90% sure about which
|
| 28 |
+
neighbor is smoking the devils lettuce, as it only really started when our next-door
|
| 29 |
+
neighbors new roommate moved in.
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
I have nothing against smoking, in any capacity, but it''s literally all our apartment
|
| 33 |
+
has smelled like the last few months. It''s starting to permeate our clothes and
|
| 34 |
+
furniture it''s so bad. My husband and I both work for the government (and are
|
| 35 |
+
drug tested) so this is not ideal. At first (before we realized what was actually
|
| 36 |
+
going on) we thought a skunk had sprayed near our apartment and was just coming
|
| 37 |
+
in through a window. Well, it''s winter now, no windows are open, and every day
|
| 38 |
+
we wake up to and come home to SKUNK. We have told the apartment complex about
|
| 39 |
+
this, and nothing has changed.
|
| 40 |
+
|
| 41 |
+
|
| 42 |
+
WIBTA if I left a note on this neighbors door? Something to the effect of "Hey
|
| 43 |
+
fellow apartment dwellers, pot is fun. However, consider blowing the smoke out
|
| 44 |
+
into a paper towel roll that has a dryer sheet at the end. Or vape it. Or eat
|
| 45 |
+
it. My entire apartment smells like skunky pot and I''m over it. Love, your neighbors"
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
Thoughts?
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
Edit: When we messaged the complex about it, we didn''t single out them or their
|
| 52 |
+
apartment, just said "someone in our building." They actually have since responded
|
| 53 |
+
saying that they''d "send a message to our building about it," so we''ll see what
|
| 54 |
+
happens! I''ll hold off on the note/talking to them in person for now, but if
|
| 55 |
+
it continues I''m definitely considering buying two smoke buddies and wrapping
|
| 56 |
+
them up nicely in Christmas paper with a little (not aggressive) card.
|
| 57 |
+
|
| 58 |
+
|
| 59 |
+
Edit 2: Enough people are asking so I wanted to clarify, **I do not live in a
|
| 60 |
+
state where it is legal.**'
|
| 61 |
+
- 'And just as Kamala Harris is dedicated to building an America for all peoples,
|
| 62 |
+
so too am I.
|
| 63 |
+
|
| 64 |
+
|
| 65 |
+
That is why the idea that freedom of religion somehow comes primary is so offensive
|
| 66 |
+
to the very idea of an America for all.
|
| 67 |
+
|
| 68 |
+
|
| 69 |
+
Some religions will cite their bigotry as having a basis in scripture. That will
|
| 70 |
+
apply to bigotry hitting the LDS community as well.
|
| 71 |
+
|
| 72 |
+
|
| 73 |
+
According to Christian doctrine, [the LDS church is fundamentally heretical](
|
| 74 |
+
|
| 75 |
+
|
| 76 |
+
Without discrimination protections, a catholic, muslim, jewish or christian business
|
| 77 |
+
can discriminate against the members of the LDS church without recourse.
|
| 78 |
+
|
| 79 |
+
|
| 80 |
+
That is why the protection of anti-discrimination measures must inevitably step
|
| 81 |
+
on the toes of religious bigotry. Religion is cited to justify bigotry. Just as
|
| 82 |
+
it was by the LDS church.'
|
| 83 |
+
- source_sentence: Thomson Buys Test Provider Capstar TORONTO (Reuters) - Electronic
|
| 84 |
+
publisher Thomson Corp. TOC.TO said on Monday it will buy test provider Capstar
|
| 85 |
+
from Educational Testing Service, the latest in a string of acquisitions designed
|
| 86 |
+
to flesh out its product offerings.
|
| 87 |
+
sentences:
|
| 88 |
+
- 'This is one of the most brilliant movies that I have seen in recent times. Goes
|
| 89 |
+
way above even any international movie of any repute. I am really surprised why
|
| 90 |
+
this has not received the recognition it deserved. Sonali Kulkarni winning the
|
| 91 |
+
National Award is perhaps the only consoling fact. Renuka Daftardar simply amazes
|
| 92 |
+
as she speaks volumes through her eyes. There are a few scenes that stand out:
|
| 93 |
+
When Gauri comes back from the city on Krishna''s wedding, she and Krishna meet
|
| 94 |
+
for the first time in many years. Krishna notices a change in Gauri, but not a
|
| 95 |
+
single line of dialogue is said. The entire gamut of emotions is conveyed through
|
| 96 |
+
subtle mannerisms and the eyes. There''s another towards the end when Krishna
|
| 97 |
+
pleads to Abhay Kulkarni to marry Gauri instead. If you are not moved by that
|
| 98 |
+
scene, you don''t have a heart.Watch this movie for sheer movie-making brilliance
|
| 99 |
+
and acting capabilities.'
|
| 100 |
+
- I put all my genius into my life I put only my talent into my works.
|
| 101 |
+
- '" we dont trust these niggas all these bitches "'
|
| 102 |
+
- source_sentence: 'Project 2025 happens. We already know what will happen. The only
|
| 103 |
+
thing we don''t know his how quick Republican plan to burn things down.
|
| 104 |
+
|
| 105 |
+
|
| 106 |
+
I think they will move quick. The pointy hoods are off. Everything is being said
|
| 107 |
+
out loud. They have their own website for fucks sake.
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
"We''re going to make the country so successful again, I''m not going to have
|
| 111 |
+
time for retribution, and remember this: our ultimate retribution is success."
|
| 112 |
+
- Trump'
|
| 113 |
+
sentences:
|
| 114 |
+
- 'I have a stepdaughter who i have never gotten along with . There is only a three
|
| 115 |
+
year age difference, so I''m not entirely blaming her, but i think my husband
|
| 116 |
+
did a shitty job raising her. She is just not nice, and not just to me, but she
|
| 117 |
+
is rude to wait staff, entitled, something of a mean girl. She has always made
|
| 118 |
+
snide comments about our age difference and me being a gold digger because I quit
|
| 119 |
+
my job. I told her once as an April Fool''s joke that her dad was leaving me everything
|
| 120 |
+
in the will, and she didn''t even respond to me, she just tattled to daddy.
|
| 121 |
+
|
| 122 |
+
|
| 123 |
+
She graduated in May with a PhD, and before she realized that she wasn''t going
|
| 124 |
+
to have a graduation because of what''s going on right now, she made a big deal
|
| 125 |
+
of announcing to everyone that I wasn''t invited. My husband told me to just deal
|
| 126 |
+
with it, because it is her graduation.
|
| 127 |
+
|
| 128 |
+
|
| 129 |
+
She came over the other night with her boyfriend of three years (she''s 26 and
|
| 130 |
+
he is 38, important later) and announced that they are getting married. Honestly
|
| 131 |
+
I was kind of pissed off. My husband is wealthy, but her boyfriend is legit rich,
|
| 132 |
+
so she is doing exactly what I did, marrying a man with a large age gap, and marrying
|
| 133 |
+
a man with a lot of money. I''m not saying that she doesn''t love him, but I think
|
| 134 |
+
she is a huge hypocrite to not acknowledge that just maybe i love her dad.
|
| 135 |
+
|
| 136 |
+
|
| 137 |
+
I told her that was kind of hypocritical and ironic and she just rolled her eyes,
|
| 138 |
+
but after the fact my husband got mad and said I was trying to steal her spotlight
|
| 139 |
+
and make her engagement about me, but he is always super defensive about her.'
|
| 140 |
+
- Paul Hamm Wins All-Around Event After falling on the landing of his vault, Paul
|
| 141 |
+
Hamm put on two of the most spectacular routines of his career to win Olympic
|
| 142 |
+
gold, the first ever by a U.S. man.
|
| 143 |
+
- "This morning I (31m) woke up and it was cold as hell in my apartment. I didn't\
|
| 144 |
+
\ realize the weather was getting a little cooler today and didn't adjust the\
|
| 145 |
+
\ thermostat accordingly, I just left it on full blast all night since its been\
|
| 146 |
+
\ hot lately. I woke up in the morning and had to take a shower before work and\
|
| 147 |
+
\ the whole apartment was freezing. \n\nI wasn't really in the mood to be cold\
|
| 148 |
+
\ as hell and shivering when I got out of the shower and had to do my hair and\
|
| 149 |
+
\ all that, so after I cleaned myself off, I just stood towards the back of the\
|
| 150 |
+
\ shower and turned the water to as hot as it goes and let it run for about ten\
|
| 151 |
+
\ or fifteen minutes. After a couple minutes I opened the bathroom door to start\
|
| 152 |
+
\ letting the heat and steam out so it wouldn't be so cold. I got out of the shower\
|
| 153 |
+
\ and it was steamy as hell and so nice and comfortable. The bathroom mirror was\
|
| 154 |
+
\ all cloudy from the steam.\n\nWhile I was doing my hair my fiance comes in and\
|
| 155 |
+
\ asks why the apartment is so foggy, so I told her what I did. I figured it was\
|
| 156 |
+
\ no big deal but she made it seem like I was a huge dick for it, saying it's\
|
| 157 |
+
\ stupid and childish and a waste of water, but I really feel like she was overreacting\
|
| 158 |
+
\ and just looking for an argument or something. What do you guys think? AITA?"
|
| 159 |
+
- source_sentence: Jake decides that the car would just need to be cleaned again in
|
| 160 |
+
the future so it would be a waste of time.
|
| 161 |
+
sentences:
|
| 162 |
+
- 'You guys think if I got a big face tattoo that said #metoo it would stop the
|
| 163 |
+
comments drunk men make at me while I do my job????'
|
| 164 |
+
- "please stop your oppressive editing\n TAB+TAB\nHi. I see that you're not as stupid\
|
| 165 |
+
\ as the other sysop, trying to block ip addresses and stuff. But all my warnings\
|
| 166 |
+
\ apply to you too. And if you're still wondering what this is about or who the\
|
| 167 |
+
\ fuck I am, you know what to do. Just do a whois on my ip117.201.39.75"
|
| 168 |
+
- Mindy cuts off ties with John and finds other friends who won't abuse her.
|
| 169 |
+
- source_sentence: Besides that which the men brought him that were over the tributes,
|
| 170 |
+
and the merchants, and they that sold by retail, and all the kings of Arabia,
|
| 171 |
+
and the governors of the country.
|
| 172 |
+
sentences:
|
| 173 |
+
- I don't look for much to come out of government ownership as long as we have Democrats
|
| 174 |
+
and Republicans.
|
| 175 |
+
- If this needs a federal mandate and 100% global consensus, than leaders like Macron
|
| 176 |
+
should let us renegotiate. As it stands right now, this agreement is 100% toothless.
|
| 177 |
+
There are no penalties for not following through with it.
|
| 178 |
+
- 'Barclays to Buy CIBC Credit Card Unit for \$293 Mln (Update1) Aug. 18 (Bloomberg)
|
| 179 |
+
-- Barclays Plc, the UK #39;s third- biggest bank, agreed to buy Juniper Financial
|
| 180 |
+
Corp. from Canadian Imperial Bank of Commerce for \$293 million in cash to help
|
| 181 |
+
expand its credit card business in North America.'
|
| 182 |
+
pipeline_tag: sentence-similarity
|
| 183 |
+
library_name: sentence-transformers
|
| 184 |
+
metrics:
|
| 185 |
+
- pearson_cosine
|
| 186 |
+
- spearman_cosine
|
| 187 |
+
model-index:
|
| 188 |
+
- name: SentenceTransformer
|
| 189 |
+
results:
|
| 190 |
+
- task:
|
| 191 |
+
type: semantic-similarity
|
| 192 |
+
name: Semantic Similarity
|
| 193 |
+
dataset:
|
| 194 |
+
name: similarity
|
| 195 |
+
type: similarity
|
| 196 |
+
metrics:
|
| 197 |
+
- type: pearson_cosine
|
| 198 |
+
value: 0.3879415946781869
|
| 199 |
+
name: Pearson Cosine
|
| 200 |
+
- type: spearman_cosine
|
| 201 |
+
value: 0.40484905925030346
|
| 202 |
+
name: Spearman Cosine
|
| 203 |
+
---
|
| 204 |
+
|
| 205 |
+
# SentenceTransformer
|
| 206 |
+
|
| 207 |
+
This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
| 208 |
+
|
| 209 |
+
## Model Details
|
| 210 |
+
|
| 211 |
+
### Model Description
|
| 212 |
+
- **Model Type:** Sentence Transformer
|
| 213 |
+
<!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
|
| 214 |
+
- **Maximum Sequence Length:** 512 tokens
|
| 215 |
+
- **Output Dimensionality:** 1024 dimensions
|
| 216 |
+
- **Similarity Function:** Cosine Similarity
|
| 217 |
+
<!-- - **Training Dataset:** Unknown -->
|
| 218 |
+
<!-- - **Language:** Unknown -->
|
| 219 |
+
<!-- - **License:** Unknown -->
|
| 220 |
+
|
| 221 |
+
### Model Sources
|
| 222 |
+
|
| 223 |
+
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
|
| 224 |
+
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
|
| 225 |
+
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
|
| 226 |
+
|
| 227 |
+
### Full Model Architecture
|
| 228 |
+
|
| 229 |
+
```
|
| 230 |
+
SentenceTransformer(
|
| 231 |
+
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'BertModel'})
|
| 232 |
+
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
|
| 233 |
+
(2): Normalize()
|
| 234 |
+
)
|
| 235 |
+
```
|
| 236 |
+
|
| 237 |
+
## Usage
|
| 238 |
+
|
| 239 |
+
### Direct Usage (Sentence Transformers)
|
| 240 |
+
|
| 241 |
+
First install the Sentence Transformers library:
|
| 242 |
+
|
| 243 |
+
```bash
|
| 244 |
+
pip install -U sentence-transformers
|
| 245 |
+
```
|
| 246 |
+
|
| 247 |
+
Then you can load this model and run inference.
|
| 248 |
+
```python
|
| 249 |
+
from sentence_transformers import SentenceTransformer
|
| 250 |
+
|
| 251 |
+
# Download from the 🤗 Hub
|
| 252 |
+
model = SentenceTransformer("sentence_transformers_model_id")
|
| 253 |
+
# Run inference
|
| 254 |
+
sentences = [
|
| 255 |
+
'Besides that which the men brought him that were over the tributes, and the merchants, and they that sold by retail, and all the kings of Arabia, and the governors of the country.',
|
| 256 |
+
'If this needs a federal mandate and 100% global consensus, than leaders like Macron should let us renegotiate. As it stands right now, this agreement is 100% toothless. There are no penalties for not following through with it.',
|
| 257 |
+
"I don't look for much to come out of government ownership as long as we have Democrats and Republicans.",
|
| 258 |
+
]
|
| 259 |
+
embeddings = model.encode(sentences)
|
| 260 |
+
print(embeddings.shape)
|
| 261 |
+
# [3, 1024]
|
| 262 |
+
|
| 263 |
+
# Get the similarity scores for the embeddings
|
| 264 |
+
similarities = model.similarity(embeddings, embeddings)
|
| 265 |
+
print(similarities)
|
| 266 |
+
# tensor([[1.0000, 0.5648, 0.5502],
|
| 267 |
+
# [0.5648, 1.0000, 0.7965],
|
| 268 |
+
# [0.5502, 0.7965, 1.0000]])
|
| 269 |
+
```
|
| 270 |
+
|
| 271 |
+
<!--
|
| 272 |
+
### Direct Usage (Transformers)
|
| 273 |
+
|
| 274 |
+
<details><summary>Click to see the direct usage in Transformers</summary>
|
| 275 |
+
|
| 276 |
+
</details>
|
| 277 |
+
-->
|
| 278 |
+
|
| 279 |
+
<!--
|
| 280 |
+
### Downstream Usage (Sentence Transformers)
|
| 281 |
+
|
| 282 |
+
You can finetune this model on your own dataset.
|
| 283 |
+
|
| 284 |
+
<details><summary>Click to expand</summary>
|
| 285 |
+
|
| 286 |
+
</details>
|
| 287 |
+
-->
|
| 288 |
+
|
| 289 |
+
<!--
|
| 290 |
+
### Out-of-Scope Use
|
| 291 |
+
|
| 292 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 293 |
+
-->
|
| 294 |
+
|
| 295 |
+
## Evaluation
|
| 296 |
+
|
| 297 |
+
### Metrics
|
| 298 |
+
|
| 299 |
+
#### Semantic Similarity
|
| 300 |
+
|
| 301 |
+
* Dataset: `similarity`
|
| 302 |
+
* Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
|
| 303 |
+
|
| 304 |
+
| Metric | Value |
|
| 305 |
+
|:--------------------|:-----------|
|
| 306 |
+
| pearson_cosine | 0.3879 |
|
| 307 |
+
| **spearman_cosine** | **0.4048** |
|
| 308 |
+
|
| 309 |
+
<!--
|
| 310 |
+
## Bias, Risks and Limitations
|
| 311 |
+
|
| 312 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
| 313 |
+
-->
|
| 314 |
+
|
| 315 |
+
<!--
|
| 316 |
+
### Recommendations
|
| 317 |
+
|
| 318 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
| 319 |
+
-->
|
| 320 |
+
|
| 321 |
+
## Training Details
|
| 322 |
+
|
| 323 |
+
### Training Dataset
|
| 324 |
+
|
| 325 |
+
#### Unnamed Dataset
|
| 326 |
+
|
| 327 |
+
* Size: 11,180 training samples
|
| 328 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
|
| 329 |
+
* Approximate statistics based on the first 1000 samples:
|
| 330 |
+
| | sentence_0 | sentence_1 | label |
|
| 331 |
+
|:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:---------------------------------------------------------------|
|
| 332 |
+
| type | string | string | float |
|
| 333 |
+
| details | <ul><li>min: 6 tokens</li><li>mean: 104.26 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 118.5 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.53</li><li>max: 1.0</li></ul> |
|
| 334 |
+
* Samples:
|
| 335 |
+
| sentence_0 | sentence_1 | label |
|
| 336 |
+
|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------|
|
| 337 |
+
| <code>He worked at Rothschild as an investment banker. Great. Am I supposed to be alarmed that France elected a technocrat who has worked in the private banking sector? <br><br>I also don't give a shit about what macron does in his personal life. Clearly the French people don't either.</code> | <code>Chad runs over the raccoon since it's been bothering him anyway.</code> | <code>0.3535533905932737</code> |
|
| 338 |
+
| <code>Amazing effects for a movie of this time. A primer of the uselessness of war and how war becomes a nurturer of itself.A wonderful thing about this movie is it is now public domain and available at archive.org. No charge, no sign up necessary. Watch it in one sitting and you will be propelled.I plan to share this flick with as many people as possible as I had never heard of it before and I am a hard core sci fi fan.I would like to see how others react to this movie.Watch it.Rate it.Tell us what you think.</code> | <code>First off, I must say that I made the mistake of watching the Election films out of sequence. I say unfortunately, because after seeing Election 2 first, Election seems a bit of a disappointment. Both films are gangster epics that are similar in form. And while Election is an enjoyable piece of cinema... it's just not nearly as good as it's sequel.In the first Election installment, we are shown the two competitors for Chairman; Big D and Lok. After a few scenes of discussion amongst the "Uncle's" as to who should have the Chairman title, they (almost unanimously) decide That Lok (Simon Yam) will helm the Triads. Suffice to say this doesn't go over very well with competitor Big D (Tony Leung Ka Fai) and in a bid to influence the takeover, Big D kidnaps two of the uncles in order to sway the election board to his side. This has disastrous results and heads the triads into an all out war. Lok is determined to become Chairman but won't become official until he can recover the "Dragon Head ...</code> | <code>0.7071067811865475</code> |
|
| 339 |
+
| <code>MY SINCERE APOLOGIES 2U WHO I'VE OFFENDED WITH ALLEGATIONS OF COMPLACENT COWARDS & ASSHOLES FOR CLIMATE CHANGE INDIFFERENCE!</code> | <code>yeah man fucking disgusting. as if we didn't waste enough time at work</code> | <code>1.0</code> |
|
| 340 |
+
* Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
|
| 341 |
+
```json
|
| 342 |
+
{
|
| 343 |
+
"loss_fct": "torch.nn.modules.loss.MSELoss"
|
| 344 |
+
}
|
| 345 |
+
```
|
| 346 |
+
|
| 347 |
+
### Training Hyperparameters
|
| 348 |
+
#### Non-Default Hyperparameters
|
| 349 |
+
|
| 350 |
+
- `eval_strategy`: steps
|
| 351 |
+
- `per_device_train_batch_size`: 32
|
| 352 |
+
- `per_device_eval_batch_size`: 32
|
| 353 |
+
- `fp16`: True
|
| 354 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 355 |
+
|
| 356 |
+
#### All Hyperparameters
|
| 357 |
+
<details><summary>Click to expand</summary>
|
| 358 |
+
|
| 359 |
+
- `overwrite_output_dir`: False
|
| 360 |
+
- `do_predict`: False
|
| 361 |
+
- `eval_strategy`: steps
|
| 362 |
+
- `prediction_loss_only`: True
|
| 363 |
+
- `per_device_train_batch_size`: 32
|
| 364 |
+
- `per_device_eval_batch_size`: 32
|
| 365 |
+
- `per_gpu_train_batch_size`: None
|
| 366 |
+
- `per_gpu_eval_batch_size`: None
|
| 367 |
+
- `gradient_accumulation_steps`: 1
|
| 368 |
+
- `eval_accumulation_steps`: None
|
| 369 |
+
- `torch_empty_cache_steps`: None
|
| 370 |
+
- `learning_rate`: 5e-05
|
| 371 |
+
- `weight_decay`: 0.0
|
| 372 |
+
- `adam_beta1`: 0.9
|
| 373 |
+
- `adam_beta2`: 0.999
|
| 374 |
+
- `adam_epsilon`: 1e-08
|
| 375 |
+
- `max_grad_norm`: 1
|
| 376 |
+
- `num_train_epochs`: 3
|
| 377 |
+
- `max_steps`: -1
|
| 378 |
+
- `lr_scheduler_type`: linear
|
| 379 |
+
- `lr_scheduler_kwargs`: {}
|
| 380 |
+
- `warmup_ratio`: 0.0
|
| 381 |
+
- `warmup_steps`: 0
|
| 382 |
+
- `log_level`: passive
|
| 383 |
+
- `log_level_replica`: warning
|
| 384 |
+
- `log_on_each_node`: True
|
| 385 |
+
- `logging_nan_inf_filter`: True
|
| 386 |
+
- `save_safetensors`: True
|
| 387 |
+
- `save_on_each_node`: False
|
| 388 |
+
- `save_only_model`: False
|
| 389 |
+
- `restore_callback_states_from_checkpoint`: False
|
| 390 |
+
- `no_cuda`: False
|
| 391 |
+
- `use_cpu`: False
|
| 392 |
+
- `use_mps_device`: False
|
| 393 |
+
- `seed`: 42
|
| 394 |
+
- `data_seed`: None
|
| 395 |
+
- `jit_mode_eval`: False
|
| 396 |
+
- `use_ipex`: False
|
| 397 |
+
- `bf16`: False
|
| 398 |
+
- `fp16`: True
|
| 399 |
+
- `fp16_opt_level`: O1
|
| 400 |
+
- `half_precision_backend`: auto
|
| 401 |
+
- `bf16_full_eval`: False
|
| 402 |
+
- `fp16_full_eval`: False
|
| 403 |
+
- `tf32`: None
|
| 404 |
+
- `local_rank`: 0
|
| 405 |
+
- `ddp_backend`: None
|
| 406 |
+
- `tpu_num_cores`: None
|
| 407 |
+
- `tpu_metrics_debug`: False
|
| 408 |
+
- `debug`: []
|
| 409 |
+
- `dataloader_drop_last`: False
|
| 410 |
+
- `dataloader_num_workers`: 0
|
| 411 |
+
- `dataloader_prefetch_factor`: None
|
| 412 |
+
- `past_index`: -1
|
| 413 |
+
- `disable_tqdm`: False
|
| 414 |
+
- `remove_unused_columns`: True
|
| 415 |
+
- `label_names`: None
|
| 416 |
+
- `load_best_model_at_end`: False
|
| 417 |
+
- `ignore_data_skip`: False
|
| 418 |
+
- `fsdp`: []
|
| 419 |
+
- `fsdp_min_num_params`: 0
|
| 420 |
+
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
|
| 421 |
+
- `fsdp_transformer_layer_cls_to_wrap`: None
|
| 422 |
+
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
|
| 423 |
+
- `deepspeed`: None
|
| 424 |
+
- `label_smoothing_factor`: 0.0
|
| 425 |
+
- `optim`: adamw_torch
|
| 426 |
+
- `optim_args`: None
|
| 427 |
+
- `adafactor`: False
|
| 428 |
+
- `group_by_length`: False
|
| 429 |
+
- `length_column_name`: length
|
| 430 |
+
- `ddp_find_unused_parameters`: None
|
| 431 |
+
- `ddp_bucket_cap_mb`: None
|
| 432 |
+
- `ddp_broadcast_buffers`: False
|
| 433 |
+
- `dataloader_pin_memory`: True
|
| 434 |
+
- `dataloader_persistent_workers`: False
|
| 435 |
+
- `skip_memory_metrics`: True
|
| 436 |
+
- `use_legacy_prediction_loop`: False
|
| 437 |
+
- `push_to_hub`: False
|
| 438 |
+
- `resume_from_checkpoint`: None
|
| 439 |
+
- `hub_model_id`: None
|
| 440 |
+
- `hub_strategy`: every_save
|
| 441 |
+
- `hub_private_repo`: None
|
| 442 |
+
- `hub_always_push`: False
|
| 443 |
+
- `gradient_checkpointing`: False
|
| 444 |
+
- `gradient_checkpointing_kwargs`: None
|
| 445 |
+
- `include_inputs_for_metrics`: False
|
| 446 |
+
- `include_for_metrics`: []
|
| 447 |
+
- `eval_do_concat_batches`: True
|
| 448 |
+
- `fp16_backend`: auto
|
| 449 |
+
- `push_to_hub_model_id`: None
|
| 450 |
+
- `push_to_hub_organization`: None
|
| 451 |
+
- `mp_parameters`:
|
| 452 |
+
- `auto_find_batch_size`: False
|
| 453 |
+
- `full_determinism`: False
|
| 454 |
+
- `torchdynamo`: None
|
| 455 |
+
- `ray_scope`: last
|
| 456 |
+
- `ddp_timeout`: 1800
|
| 457 |
+
- `torch_compile`: False
|
| 458 |
+
- `torch_compile_backend`: None
|
| 459 |
+
- `torch_compile_mode`: None
|
| 460 |
+
- `dispatch_batches`: None
|
| 461 |
+
- `split_batches`: None
|
| 462 |
+
- `include_tokens_per_second`: False
|
| 463 |
+
- `include_num_input_tokens_seen`: False
|
| 464 |
+
- `neftune_noise_alpha`: None
|
| 465 |
+
- `optim_target_modules`: None
|
| 466 |
+
- `batch_eval_metrics`: False
|
| 467 |
+
- `eval_on_start`: False
|
| 468 |
+
- `use_liger_kernel`: False
|
| 469 |
+
- `eval_use_gather_object`: False
|
| 470 |
+
- `average_tokens_across_devices`: False
|
| 471 |
+
- `prompts`: None
|
| 472 |
+
- `batch_sampler`: batch_sampler
|
| 473 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 474 |
+
- `router_mapping`: {}
|
| 475 |
+
- `learning_rate_mapping`: {}
|
| 476 |
+
|
| 477 |
+
</details>
|
| 478 |
+
|
| 479 |
+
### Training Logs
|
| 480 |
+
| Epoch | Step | Training Loss | similarity_spearman_cosine |
|
| 481 |
+
|:------:|:----:|:-------------:|:--------------------------:|
|
| 482 |
+
| 0.0286 | 10 | - | 0.2006 |
|
| 483 |
+
| 0.0571 | 20 | - | 0.2012 |
|
| 484 |
+
| 0.0857 | 30 | - | 0.2023 |
|
| 485 |
+
| 0.1143 | 40 | - | 0.2036 |
|
| 486 |
+
| 0.1429 | 50 | - | 0.2054 |
|
| 487 |
+
| 0.1714 | 60 | - | 0.2081 |
|
| 488 |
+
| 0.2 | 70 | - | 0.2098 |
|
| 489 |
+
| 0.2286 | 80 | - | 0.2115 |
|
| 490 |
+
| 0.2571 | 90 | - | 0.2128 |
|
| 491 |
+
| 0.2857 | 100 | - | 0.2149 |
|
| 492 |
+
| 0.3143 | 110 | - | 0.2177 |
|
| 493 |
+
| 0.3429 | 120 | - | 0.2207 |
|
| 494 |
+
| 0.3714 | 130 | - | 0.2243 |
|
| 495 |
+
| 0.4 | 140 | - | 0.2278 |
|
| 496 |
+
| 0.4286 | 150 | - | 0.2310 |
|
| 497 |
+
| 0.4571 | 160 | - | 0.2332 |
|
| 498 |
+
| 0.4857 | 170 | - | 0.2350 |
|
| 499 |
+
| 0.5143 | 180 | - | 0.2361 |
|
| 500 |
+
| 0.5429 | 190 | - | 0.2360 |
|
| 501 |
+
| 0.5714 | 200 | - | 0.2369 |
|
| 502 |
+
| 0.6 | 210 | - | 0.2423 |
|
| 503 |
+
| 0.6286 | 220 | - | 0.2533 |
|
| 504 |
+
| 0.6571 | 230 | - | 0.2691 |
|
| 505 |
+
| 0.6857 | 240 | - | 0.2808 |
|
| 506 |
+
| 0.7143 | 250 | - | 0.2889 |
|
| 507 |
+
| 0.7429 | 260 | - | 0.2960 |
|
| 508 |
+
| 0.7714 | 270 | - | 0.2939 |
|
| 509 |
+
| 0.8 | 280 | - | 0.3007 |
|
| 510 |
+
| 0.8286 | 290 | - | 0.3010 |
|
| 511 |
+
| 0.8571 | 300 | - | 0.3016 |
|
| 512 |
+
| 0.8857 | 310 | - | 0.3035 |
|
| 513 |
+
| 0.9143 | 320 | - | 0.3078 |
|
| 514 |
+
| 0.9429 | 330 | - | 0.3138 |
|
| 515 |
+
| 0.9714 | 340 | - | 0.3206 |
|
| 516 |
+
| 1.0 | 350 | - | 0.3234 |
|
| 517 |
+
| 1.0286 | 360 | - | 0.3299 |
|
| 518 |
+
| 1.0571 | 370 | - | 0.3367 |
|
| 519 |
+
| 1.0857 | 380 | - | 0.3267 |
|
| 520 |
+
| 1.1143 | 390 | - | 0.3307 |
|
| 521 |
+
| 1.1429 | 400 | - | 0.3359 |
|
| 522 |
+
| 1.1714 | 410 | - | 0.3417 |
|
| 523 |
+
| 1.2 | 420 | - | 0.3504 |
|
| 524 |
+
| 1.2286 | 430 | - | 0.3324 |
|
| 525 |
+
| 1.2571 | 440 | - | 0.3365 |
|
| 526 |
+
| 1.2857 | 450 | - | 0.3580 |
|
| 527 |
+
| 1.3143 | 460 | - | 0.3622 |
|
| 528 |
+
| 1.3429 | 470 | - | 0.3073 |
|
| 529 |
+
| 1.3714 | 480 | - | 0.3596 |
|
| 530 |
+
| 1.4 | 490 | - | 0.3473 |
|
| 531 |
+
| 1.4286 | 500 | 0.1278 | 0.3573 |
|
| 532 |
+
| 1.4571 | 510 | - | 0.3539 |
|
| 533 |
+
| 1.4857 | 520 | - | 0.3355 |
|
| 534 |
+
| 1.5143 | 530 | - | 0.3299 |
|
| 535 |
+
| 1.5429 | 540 | - | 0.3559 |
|
| 536 |
+
| 1.5714 | 550 | - | 0.3285 |
|
| 537 |
+
| 1.6 | 560 | - | 0.3435 |
|
| 538 |
+
| 1.6286 | 570 | - | 0.3654 |
|
| 539 |
+
| 1.6571 | 580 | - | 0.3824 |
|
| 540 |
+
| 1.6857 | 590 | - | 0.3426 |
|
| 541 |
+
| 1.7143 | 600 | - | 0.3413 |
|
| 542 |
+
| 1.7429 | 610 | - | 0.3395 |
|
| 543 |
+
| 1.7714 | 620 | - | 0.3492 |
|
| 544 |
+
| 1.8 | 630 | - | 0.3664 |
|
| 545 |
+
| 1.8286 | 640 | - | 0.3634 |
|
| 546 |
+
| 1.8571 | 650 | - | 0.3392 |
|
| 547 |
+
| 1.8857 | 660 | - | 0.3686 |
|
| 548 |
+
| 1.9143 | 670 | - | 0.3722 |
|
| 549 |
+
| 1.9429 | 680 | - | 0.3557 |
|
| 550 |
+
| 1.9714 | 690 | - | 0.3896 |
|
| 551 |
+
| 2.0 | 700 | - | 0.3908 |
|
| 552 |
+
| 2.0286 | 710 | - | 0.3859 |
|
| 553 |
+
| 2.0571 | 720 | - | 0.3536 |
|
| 554 |
+
| 2.0857 | 730 | - | 0.3606 |
|
| 555 |
+
| 2.1143 | 740 | - | 0.3638 |
|
| 556 |
+
| 2.1429 | 750 | - | 0.3713 |
|
| 557 |
+
| 2.1714 | 760 | - | 0.3704 |
|
| 558 |
+
| 2.2 | 770 | - | 0.3441 |
|
| 559 |
+
| 2.2286 | 780 | - | 0.3435 |
|
| 560 |
+
| 2.2571 | 790 | - | 0.3668 |
|
| 561 |
+
| 2.2857 | 800 | - | 0.3735 |
|
| 562 |
+
| 2.3143 | 810 | - | 0.3373 |
|
| 563 |
+
| 2.3429 | 820 | - | 0.3474 |
|
| 564 |
+
| 2.3714 | 830 | - | 0.3560 |
|
| 565 |
+
| 2.4 | 840 | - | 0.3028 |
|
| 566 |
+
| 2.4286 | 850 | - | 0.3485 |
|
| 567 |
+
| 2.4571 | 860 | - | 0.3604 |
|
| 568 |
+
| 2.4857 | 870 | - | 0.3769 |
|
| 569 |
+
| 2.5143 | 880 | - | 0.3600 |
|
| 570 |
+
| 2.5429 | 890 | - | 0.3916 |
|
| 571 |
+
| 2.5714 | 900 | - | 0.3957 |
|
| 572 |
+
| 2.6 | 910 | - | 0.3797 |
|
| 573 |
+
| 2.6286 | 920 | - | 0.3875 |
|
| 574 |
+
| 2.6571 | 930 | - | 0.3978 |
|
| 575 |
+
| 2.6857 | 940 | - | 0.3951 |
|
| 576 |
+
| 2.7143 | 950 | - | 0.3831 |
|
| 577 |
+
| 2.7429 | 960 | - | 0.3912 |
|
| 578 |
+
| 2.7714 | 970 | - | 0.3800 |
|
| 579 |
+
| 2.8 | 980 | - | 0.3955 |
|
| 580 |
+
| 2.8286 | 990 | - | 0.3976 |
|
| 581 |
+
| 2.8571 | 1000 | 0.1036 | 0.4048 |
|
| 582 |
+
|
| 583 |
+
|
| 584 |
+
### Framework Versions
|
| 585 |
+
- Python: 3.11.9
|
| 586 |
+
- Sentence Transformers: 5.1.0
|
| 587 |
+
- Transformers: 4.49.0
|
| 588 |
+
- PyTorch: 2.8.0+cu128
|
| 589 |
+
- Accelerate: 1.10.0
|
| 590 |
+
- Datasets: 2.14.4
|
| 591 |
+
- Tokenizers: 0.21.0
|
| 592 |
+
|
| 593 |
+
## Citation
|
| 594 |
+
|
| 595 |
+
### BibTeX
|
| 596 |
+
|
| 597 |
+
#### Sentence Transformers
|
| 598 |
+
```bibtex
|
| 599 |
+
@inproceedings{reimers-2019-sentence-bert,
|
| 600 |
+
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|
| 601 |
+
author = "Reimers, Nils and Gurevych, Iryna",
|
| 602 |
+
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
|
| 603 |
+
month = "11",
|
| 604 |
+
year = "2019",
|
| 605 |
+
publisher = "Association for Computational Linguistics",
|
| 606 |
+
url = "https://arxiv.org/abs/1908.10084",
|
| 607 |
+
}
|
| 608 |
+
```
|
| 609 |
+
|
| 610 |
+
<!--
|
| 611 |
+
## Glossary
|
| 612 |
+
|
| 613 |
+
*Clearly define terms in order to be accessible across audiences.*
|
| 614 |
+
-->
|
| 615 |
+
|
| 616 |
+
<!--
|
| 617 |
+
## Model Card Authors
|
| 618 |
+
|
| 619 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
| 620 |
+
-->
|
| 621 |
+
|
| 622 |
+
<!--
|
| 623 |
+
## Model Card Contact
|
| 624 |
+
|
| 625 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
| 626 |
+
-->
|
config.json
ADDED
|
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "./models/pretrained/gte-large",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"BertModel"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"classifier_dropout": null,
|
| 8 |
+
"gradient_checkpointing": false,
|
| 9 |
+
"hidden_act": "gelu",
|
| 10 |
+
"hidden_dropout_prob": 0.1,
|
| 11 |
+
"hidden_size": 1024,
|
| 12 |
+
"initializer_range": 0.02,
|
| 13 |
+
"intermediate_size": 4096,
|
| 14 |
+
"layer_norm_eps": 1e-12,
|
| 15 |
+
"max_position_embeddings": 512,
|
| 16 |
+
"model_type": "bert",
|
| 17 |
+
"num_attention_heads": 16,
|
| 18 |
+
"num_hidden_layers": 24,
|
| 19 |
+
"pad_token_id": 0,
|
| 20 |
+
"position_embedding_type": "absolute",
|
| 21 |
+
"torch_dtype": "float32",
|
| 22 |
+
"transformers_version": "4.49.0",
|
| 23 |
+
"type_vocab_size": 2,
|
| 24 |
+
"use_cache": true,
|
| 25 |
+
"vocab_size": 30522
|
| 26 |
+
}
|
config_sentence_transformers.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_type": "SentenceTransformer",
|
| 3 |
+
"__version__": {
|
| 4 |
+
"sentence_transformers": "5.1.0",
|
| 5 |
+
"transformers": "4.49.0",
|
| 6 |
+
"pytorch": "2.8.0+cu128"
|
| 7 |
+
},
|
| 8 |
+
"prompts": {
|
| 9 |
+
"query": "",
|
| 10 |
+
"document": ""
|
| 11 |
+
},
|
| 12 |
+
"default_prompt_name": null,
|
| 13 |
+
"similarity_fn_name": "cosine"
|
| 14 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a846270ef79ec7517531a94c9d816903d7b59602b410352185a48014ca8d0128
|
| 3 |
+
size 1340612432
|
modules.json
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"idx": 0,
|
| 4 |
+
"name": "0",
|
| 5 |
+
"path": "",
|
| 6 |
+
"type": "sentence_transformers.models.Transformer"
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"idx": 1,
|
| 10 |
+
"name": "1",
|
| 11 |
+
"path": "1_Pooling",
|
| 12 |
+
"type": "sentence_transformers.models.Pooling"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"idx": 2,
|
| 16 |
+
"name": "2",
|
| 17 |
+
"path": "2_Normalize",
|
| 18 |
+
"type": "sentence_transformers.models.Normalize"
|
| 19 |
+
}
|
| 20 |
+
]
|
sentence_bert_config.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"max_seq_length": 512,
|
| 3 |
+
"do_lower_case": false
|
| 4 |
+
}
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": true,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": true,
|
| 47 |
+
"extra_special_tokens": {},
|
| 48 |
+
"mask_token": "[MASK]",
|
| 49 |
+
"max_length": 128,
|
| 50 |
+
"model_max_length": 512,
|
| 51 |
+
"pad_to_multiple_of": null,
|
| 52 |
+
"pad_token": "[PAD]",
|
| 53 |
+
"pad_token_type_id": 0,
|
| 54 |
+
"padding_side": "right",
|
| 55 |
+
"sep_token": "[SEP]",
|
| 56 |
+
"stride": 0,
|
| 57 |
+
"strip_accents": null,
|
| 58 |
+
"tokenize_chinese_chars": true,
|
| 59 |
+
"tokenizer_class": "BertTokenizer",
|
| 60 |
+
"truncation_side": "right",
|
| 61 |
+
"truncation_strategy": "longest_first",
|
| 62 |
+
"unk_token": "[UNK]"
|
| 63 |
+
}
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|