Upload 10 files
Browse files- README.md +648 -0
- config.json +24 -0
- config_sentence_transformers.json +14 -0
- model.safetensors +3 -0
- modules.json +20 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +56 -0
- vocab.txt +0 -0
README.md
ADDED
|
@@ -0,0 +1,648 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- sentence-transformers
|
| 4 |
+
- sentence-similarity
|
| 5 |
+
- feature-extraction
|
| 6 |
+
- dense
|
| 7 |
+
- generated_from_trainer
|
| 8 |
+
- dataset_size:11180
|
| 9 |
+
- loss:CosineSimilarityLoss
|
| 10 |
+
widget:
|
| 11 |
+
- source_sentence: Of course she would, otherwise she stands no chance of becoming
|
| 12 |
+
prez. The only thing above Le Pen's xenophobia is their thrive for power.
|
| 13 |
+
sentences:
|
| 14 |
+
- Mormon Church declares same-sex couples apostates and excludes children of those
|
| 15 |
+
couples from blessings and baptism.
|
| 16 |
+
- Feminists have legitimate gripes with the way the world is structured and their
|
| 17 |
+
ideas are quite sane. The feminists of the world are not deluded or wacky and
|
| 18 |
+
it's a bad idea to call them that. Open your mind to new ideas and drop the patriarchal
|
| 19 |
+
thinking.
|
| 20 |
+
- Hi, Professor Lichtman. Thanks for doing this AMA. What would you say to people
|
| 21 |
+
who argue about the economy keys being affected by people not feeling, in terms
|
| 22 |
+
of their lived experiences, that the economy has been good (due to the cost of
|
| 23 |
+
living exceeding, in many cases, their income) and this potentially costing Harris
|
| 24 |
+
the election? They seem to believe that this should be costing her the ST and
|
| 25 |
+
LT economy keys.
|
| 26 |
+
- source_sentence: Sondra Locke stinks in this film, but then she was an awful 'actress'
|
| 27 |
+
anyway. Unfortunately, she drags everyone else (including then =real life boyfriend
|
| 28 |
+
Clint Eastwood down the drain with her. But what was Clint Eastwood thinking when
|
| 29 |
+
he agreed to star in this one? One read of the script should have told him that
|
| 30 |
+
this one was going to be a real snorer. It's an exceptionally weak story, basically
|
| 31 |
+
no story or plot at all. Add in bored, poor acting, even from the normally good
|
| 32 |
+
Eastwood. There's absolutely no action except a couple arguments and as far as
|
| 33 |
+
I was concerned, this film ranks up at the top of the heap of natural sleep enhancers.
|
| 34 |
+
Wow! Could a film BE any more boring? I think watching paint dry or the grass
|
| 35 |
+
grow might be more fun. A real stinker. Don't bother with this one.
|
| 36 |
+
sentences:
|
| 37 |
+
- 'SPOILERS: The original Road House is one of those movies that I know is clichéd
|
| 38 |
+
and unoriginal, yet it''s done so well, I''m embarrassed to admit I really like
|
| 39 |
+
it. Turns out many of my friends, whose movie opinions I respect, think the same
|
| 40 |
+
way. So when they attempt to make a sequel to it and it''s as if it was written
|
| 41 |
+
by some high school kids who were given the rights to do a sequel, it''s just
|
| 42 |
+
bad; really, really bad.Oddly, Johnathon Schaech is listed as one of the writers
|
| 43 |
+
and I can only hope his WGA membership is revoked. The writing was just bad and
|
| 44 |
+
all the writers of this film should retire for complete lack of originality and
|
| 45 |
+
some of the worst dialog in this millennium. Schaech already appearing to be the
|
| 46 |
+
king of the straight-to-DVD sequel (8mm 2, Poison Ivy 2) and now after seeing
|
| 47 |
+
this and 8mm 2, I''m thinking his acting ability is non-existent. He was awful,
|
| 48 |
+
just awful.And it''s not the terrible fighting scenes that make this movie terrible,
|
| 49 |
+
but take it from me, they''re bad. Every fight scene is a slowly delivered punch
|
| 50 |
+
(yet still making the "wiff" sound in the air) that is then blocked by the opponent,
|
| 51 |
+
who returns a punch that sends the first guy to the ground. This is repeated throughout
|
| 52 |
+
the film, worse than any bad 1970s cop show. Or the fact that many of the people
|
| 53 |
+
involved in the fights seem to have a mouth full of cherry kool-aid for some reason.
|
| 54 |
+
And we''re supposed to believe Will Patton is a fighting machine; his fight scenes
|
| 55 |
+
look so amazingly fake I was honestly embarrassed watching. It''s the complete
|
| 56 |
+
lapses of logic in this ridiculous movie that make it terrible. For instance:
|
| 57 |
+
Johnathon Schaech''s character is in town for a day and already tells some girl
|
| 58 |
+
he barely knows who he has no idea what side she''s on, "I''m with the Feds, but
|
| 59 |
+
don''t tell anyone." The female villain, who fights the good girl in one fight
|
| 60 |
+
scene with acrobatics that rival any super hero, yet is easily held down by the
|
| 61 |
+
Will Patton, "old guy," in another scene by simply holding both her hands while
|
| 62 |
+
he utters some ridiculous line ("stab me once, shame on you, stab me twice, ain''t
|
| 63 |
+
gonna happen" whew, that''s bad) and then head-butts her. Jake Busey''s villain
|
| 64 |
+
shoots at the feds while caught in the middle of a drug deal, yet no DEA agents
|
| 65 |
+
or anyone simply go to his place and pick him up after, in fact, he''s simply
|
| 66 |
+
let go because "this is the sheriff''s territory." Busey wants the bar because
|
| 67 |
+
it''s "in a great location" for drug deals, yet his own house appears to be just
|
| 68 |
+
as good apparently offering all the perks the bar is supposed to have. Johnathon
|
| 69 |
+
Schaech''s character is supposed to be the son of Patrick Swayze''s character
|
| 70 |
+
in the original, yet Swayze''s character''s last name is Dalton and Schaech''s
|
| 71 |
+
isn''t (nor is the supposed brother of Swayze''s character). And Johnathon Schaech
|
| 72 |
+
looks about 50 in this movie. I looked it up, he''s 17 years younger than Swayze,
|
| 73 |
+
but he looks awful.But my favorite absolutely stupid scene in this movie was the
|
| 74 |
+
most stock fight scene ending in movies: the villain is knocked through a window
|
| 75 |
+
on a second floor and as they pan down I''m thinking "please don''t tell me he''s
|
| 76 |
+
impaled on something..." and sure enough, my worst fears were realized.Actually,
|
| 77 |
+
I could go on for another half hour about the things I hated about this movie.
|
| 78 |
+
Suffice to say, let''s put an end to these ridiculous straight-to-DVD sequels
|
| 79 |
+
to theatrical movies, at least the ones with Johnathon Schaech.'
|
| 80 |
+
- 'Who do neoliberals scapegoat? The uneducated I guess? But we certainly don''t
|
| 81 |
+
advocate for their destruction or dissolution or deportation, unless you think
|
| 82 |
+
wanting to improve education standards and accessibility is violent?
|
| 83 |
+
|
| 84 |
+
|
| 85 |
+
I have yet to see anyone unironically idolize Macron like the right does Trump
|
| 86 |
+
and the left does Bernie.'
|
| 87 |
+
- 'Republicans created their own safe space in arcon (flaired users only) since
|
| 88 |
+
their facts are fake and they are too much of snowflakes to take the criticism
|
| 89 |
+
of being fact checked, just like how Trump gets mad when fact checked.
|
| 90 |
+
|
| 91 |
+
|
| 92 |
+
Republicans don''t post here because they are cowards.'
|
| 93 |
+
- source_sentence: U.S. Forces Kill 50 Sadr Militia in Baghdad Suburb (Reuters) Reuters
|
| 94 |
+
- U.S. forces killed more than 50 Shi'ite\militiamen Wednesday in a significant
|
| 95 |
+
advance into a Baghdad\suburb that is a powerbase for radical cleric Moqtada al-Sadr,\the
|
| 96 |
+
military said.
|
| 97 |
+
sentences:
|
| 98 |
+
- "'You haven't had it as bad as I had it,' is exactly why Millennials are the first\
|
| 99 |
+
\ generation ever to be worse off than their parents.\n \n If we're on the same\
|
| 100 |
+
\ team, let's act like it.\n\nFucking thank you!\n\nI literally didn't mention\
|
| 101 |
+
\ Boomers or say a damn thing negative about them. I shared my experience as\
|
| 102 |
+
\ a Millenial in a thread for an article specifically about my generation. And\
|
| 103 |
+
\ of fucking course a Boomer came in to tell me I don't know what real struggle\
|
| 104 |
+
\ is. Like... okay, what the hell is your point Karen?"
|
| 105 |
+
- '"Dutch Schultz", AKA Arthur Fleggenheimer, was a real person and his rather nasty
|
| 106 |
+
life is fairly well documented. This movie which purports to depict his life should
|
| 107 |
+
have used a fictional character, because the overdramatized events are too strong
|
| 108 |
+
a departure from the facts and the chronology. Not only that, it ignores some
|
| 109 |
+
interesting details which other versions have included such as the public relations
|
| 110 |
+
fiasco in upstate N.Y. and his religious conversion. It is true that he was executed
|
| 111 |
+
by Luciano, Lansky, et. al. but that''s as far as it goes. The exploding plate
|
| 112 |
+
scene which represents Luciano carrying out the execution of Bo Weinberg in his
|
| 113 |
+
own home, assisted by his own mother is rediculous. Also, there is the scene in
|
| 114 |
+
which Dutch approaches his own mother to pay protection to Legs Diamond. It just
|
| 115 |
+
doesn''t work. The character of Mrs. Fleggenheimer doesn''t work either. This
|
| 116 |
+
movie does not need a doting Jewish mother for comic relief. The lame representation
|
| 117 |
+
of Legs Diamond was humorous enough. I''m sure the man is turning in his grave.
|
| 118 |
+
And, by the way, Dutch did in fact personally kill people, but, he was not Rambo
|
| 119 |
+
or 007. The scene in which he wipes out the brewery is absurd. I don''t know.
|
| 120 |
+
Maybe it was supposed to be a comedy and I just didn''t get it.'
|
| 121 |
+
- Can we stop with the whole "leader/s of the free world"? What does that even mean?
|
| 122 |
+
As far as I know we're all from distinct sovereign countries for now. Like I don't
|
| 123 |
+
hate Macron or Merkel and they are good options so far, but this crap with "leaders
|
| 124 |
+
of the free world" is just obnoxious.
|
| 125 |
+
- source_sentence: Reasoned arguments and suggestions which make allowance for the
|
| 126 |
+
full difficulties of the state of war that exists may help, and will always be
|
| 127 |
+
listened to with respect and sympathy.
|
| 128 |
+
sentences:
|
| 129 |
+
- 'Because they''re primarily just racists. If they truly believed in their stated
|
| 130 |
+
ideology they would want Macron to win. If you''re an America First nationalist
|
| 131 |
+
who thinks globalism is terrible then you should want globalists in charge of
|
| 132 |
+
every other country.
|
| 133 |
+
|
| 134 |
+
|
| 135 |
+
Edit: To be fair it''s possible that they''re not racist but merely too stupid
|
| 136 |
+
to think through the basic logic of their own ideology. It''s one or the other.'
|
| 137 |
+
- This is the last episode of the Goldenboy OVA series. Kentaro finds himself working
|
| 138 |
+
in an animation studio, which is rather interesting if you don't know anything
|
| 139 |
+
about the way anime studios were run. Besides episode 3, this was probably the
|
| 140 |
+
least risqué, but it had a nice girl interest, as well as a surprise reunion from
|
| 141 |
+
others in the previous episodes. My only complaint about this episode is it seemed
|
| 142 |
+
a little too short, but at the same time this may have only been because it was
|
| 143 |
+
the only original script for the show that wasn't based on one of the manga chapters.
|
| 144 |
+
but it ended well, leaving us with the nice feeling that Kentaro is permanently
|
| 145 |
+
25, studying on. Definitely watch the rest of the series all the way through,
|
| 146 |
+
you can buy the whole series for like $17, you can watch it all the way through
|
| 147 |
+
in about 2 1/2 hours, or watch your favorite episode if you have 20 minutes free
|
| 148 |
+
time (which i do if i have a lunch break at school.) good series, check it out.
|
| 149 |
+
- 'DON''T STOP NOW: DEMAND AN END TO CHEMICAL CASTRATION of CHILDREN! Under Scrutiny,
|
| 150 |
+
Texas Judge Caves Under Pressure, Grants Father ''A Say'' In Son''s Transitioning
|
| 151 |
+
- But Only with a Gag Order. STAND YOUR GROUND AND KEEP FIGHTING!.'
|
| 152 |
+
- source_sentence: Nationalism is a silly cock crowing on his own dunghill.
|
| 153 |
+
sentences:
|
| 154 |
+
- "There absolutely was voter fraud. There's voter fraud in every election. However,\
|
| 155 |
+
\ they are generally isolated incidents and I don't think there has been any credible\
|
| 156 |
+
\ evidence presented that indicates any wide-scale systemic voter fraud happened\
|
| 157 |
+
\ in 2020. \n\nI would like a federal commission started that investigates and\
|
| 158 |
+
\ looks for systemic voter and election fraud. Especially one that would be empowered\
|
| 159 |
+
\ to look into cases of disenfranchisement and voter suppression as well. Everyone\
|
| 160 |
+
\ that is legally allowed to vote should be able to easily and securely register\
|
| 161 |
+
\ and cast their vote."
|
| 162 |
+
- 'The 1970''s saw a rise and fall of what we have come to know as "Blacksploitation"
|
| 163 |
+
Films. The term is a reference to kind of broad catch-all, rather than a true
|
| 164 |
+
Genre of Film. In short, any comedy, drama, adventure, western or urban cops &
|
| 165 |
+
robbers shoot-em-up, that are so constructed and so cast as to appeal to the large
|
| 166 |
+
Urban Black population of the Mid 20th Century. That indeed could embrace the
|
| 167 |
+
widest type of films, as long as the had a slant toward the inner-city black population.It
|
| 168 |
+
appears that the idea of producing these films of particularly keen interest to
|
| 169 |
+
Black Americans had its genesis with the Eastertime Release of 100 RIFLES (Marvin
|
| 170 |
+
Schwartz Prod./20th Century-Fox, 1969). In it, former Syracuse University All-American
|
| 171 |
+
Footballer and Several Times All-Pro Fullback for the Cleveland Browns, Jim Brown,
|
| 172 |
+
had a Co-Starring Billing. Having appeared in a number of films already, as for
|
| 173 |
+
example, RIO CONCHOS (1964),THE DIRTY DOZEN (1967), (ICE STTION ZEBRA (1968)*
|
| 174 |
+
and others, it was beginning to make more sense to the Studios'' "Suits" that
|
| 175 |
+
Jim was a hot property.Now this 100 RIFLES brings record numbers of Black patrons
|
| 176 |
+
to the Big Cities'' central business districts on Easter Sunday to view Mr. Brown.
|
| 177 |
+
Why not start to film more of these adventure epics and other types of film with
|
| 178 |
+
more Black Players and Stars? Why not, indeed.** So we saw a succession of Cops
|
| 179 |
+
& Robbers, Bad-ass Private Detective Films, Comedies, all going the route. Along
|
| 180 |
+
the way, we eventually got to some more family oriented, wider appealing films.
|
| 181 |
+
The movie goers were treated to SOUNDER (1972), THE TAKE (1974), CONRACK (1974)and,
|
| 182 |
+
ultimately, CLAUDINE (1974).In CLAUDINE, we find no stigma nor easy classification
|
| 183 |
+
as being "Blackploitation", as the story is universal, and could easily have been
|
| 184 |
+
done as a story about people of any descent, any where, and not just in the 1970''s
|
| 185 |
+
USA.That the story was done of a SINGLE mother, Claudine (Dianne Carroll), struggling
|
| 186 |
+
to keep a family together after "....two marriages and two almost marriages.",
|
| 187 |
+
is a far cry from a shoot-em-up Harlem Style. The problems that plague the everyday
|
| 188 |
+
citizens of our nation are confronted and examined under the ol'' sociological
|
| 189 |
+
microscope.But we also consider Claudine''s psychological and physical needs as
|
| 190 |
+
a female. For "Woman Needs Man and Man Must Have His MATE",***and we do concede
|
| 191 |
+
this point. (That''s S-E-X that we''re talking about, Schultz!) Claudine meets
|
| 192 |
+
up with a very masculine, broad shouldered, athletic type in Private Scavanger
|
| 193 |
+
Garbage Man, Ruppert B. Marshall (James Earl Jones) and they go on a date.The
|
| 194 |
+
Great Welfare State intervenes with the Couple as Claudine''s Welfare Case Worker,
|
| 195 |
+
Miss Tayback (Elisa Loti), comes snooping around to see just who is this unattached
|
| 196 |
+
Male, who is suddenly paying so much attention to Claudine''s family.After a humiliating
|
| 197 |
+
experience with the Welfare Bureau''s auditing and "deducting" binge, which would
|
| 198 |
+
be the norm for the family, the two decide to get married with or without the
|
| 199 |
+
blessing of Big Brother.Meanwhile, Claudine''s elder son has gotten involved with
|
| 200 |
+
some big talking but little doing Black Activist group. But, with Ruppert''s help,
|
| 201 |
+
he and they all come through it A.O.K.It ends on a Happy, Upbeat and Hopeful note.
|
| 202 |
+
We know that it may not be exactly "...Happily Ever After!", but rather the''ll
|
| 203 |
+
make it all together! If there is a single criticism that we must state it is
|
| 204 |
+
that sometimes in a movie like this, a misconception is spread to a large portion
|
| 205 |
+
of Urban Blacks. And that is, the apparent implied myth that all Whites are wealthy,
|
| 206 |
+
having none of their kind ever in need of a helping hand, out of work or suffering
|
| 207 |
+
any disabilities.Well, folks, it just ain''t true! NOTE: * At one point, Jim Brown''s
|
| 208 |
+
career was a real hit as a rugged actioner. He was even being tauted as "...The
|
| 209 |
+
Black John Wayne." NOTE: ** The idea of producing films with All-Black Casts,
|
| 210 |
+
filmed for All-Black consumption was not a new idea. In the 1920''s, ''30''s and
|
| 211 |
+
''40''s, we saw productions from people like Noble Johnson, Spencer Williams,
|
| 212 |
+
Jr. and Rex Ingram.NOTE: *** That''s "As Time Goes By", you know, Schultz, it''s
|
| 213 |
+
from CASABLANCA (Warner Brothers, 1942).'
|
| 214 |
+
- Thinks gun confiscation is . . .
|
| 215 |
+
pipeline_tag: sentence-similarity
|
| 216 |
+
library_name: sentence-transformers
|
| 217 |
+
metrics:
|
| 218 |
+
- pearson_cosine
|
| 219 |
+
- spearman_cosine
|
| 220 |
+
model-index:
|
| 221 |
+
- name: SentenceTransformer
|
| 222 |
+
results:
|
| 223 |
+
- task:
|
| 224 |
+
type: semantic-similarity
|
| 225 |
+
name: Semantic Similarity
|
| 226 |
+
dataset:
|
| 227 |
+
name: similarity
|
| 228 |
+
type: similarity
|
| 229 |
+
metrics:
|
| 230 |
+
- type: pearson_cosine
|
| 231 |
+
value: 0.41059188174486916
|
| 232 |
+
name: Pearson Cosine
|
| 233 |
+
- type: spearman_cosine
|
| 234 |
+
value: 0.4260753939913245
|
| 235 |
+
name: Spearman Cosine
|
| 236 |
+
---
|
| 237 |
+
|
| 238 |
+
# SentenceTransformer
|
| 239 |
+
|
| 240 |
+
This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
| 241 |
+
|
| 242 |
+
## Model Details
|
| 243 |
+
|
| 244 |
+
### Model Description
|
| 245 |
+
- **Model Type:** Sentence Transformer
|
| 246 |
+
<!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
|
| 247 |
+
- **Maximum Sequence Length:** 512 tokens
|
| 248 |
+
- **Output Dimensionality:** 1024 dimensions
|
| 249 |
+
- **Similarity Function:** Cosine Similarity
|
| 250 |
+
<!-- - **Training Dataset:** Unknown -->
|
| 251 |
+
<!-- - **Language:** Unknown -->
|
| 252 |
+
<!-- - **License:** Unknown -->
|
| 253 |
+
|
| 254 |
+
### Model Sources
|
| 255 |
+
|
| 256 |
+
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
|
| 257 |
+
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
|
| 258 |
+
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
|
| 259 |
+
|
| 260 |
+
### Full Model Architecture
|
| 261 |
+
|
| 262 |
+
```
|
| 263 |
+
SentenceTransformer(
|
| 264 |
+
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'BertModel'})
|
| 265 |
+
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
|
| 266 |
+
(2): Normalize()
|
| 267 |
+
)
|
| 268 |
+
```
|
| 269 |
+
|
| 270 |
+
## Usage
|
| 271 |
+
|
| 272 |
+
### Direct Usage (Sentence Transformers)
|
| 273 |
+
|
| 274 |
+
First install the Sentence Transformers library:
|
| 275 |
+
|
| 276 |
+
```bash
|
| 277 |
+
pip install -U sentence-transformers
|
| 278 |
+
```
|
| 279 |
+
|
| 280 |
+
Then you can load this model and run inference.
|
| 281 |
+
```python
|
| 282 |
+
from sentence_transformers import SentenceTransformer
|
| 283 |
+
|
| 284 |
+
# Download from the 🤗 Hub
|
| 285 |
+
model = SentenceTransformer("sentence_transformers_model_id")
|
| 286 |
+
# Run inference
|
| 287 |
+
sentences = [
|
| 288 |
+
'Nationalism is a silly cock crowing on his own dunghill.',
|
| 289 |
+
'The 1970\'s saw a rise and fall of what we have come to know as "Blacksploitation" Films. The term is a reference to kind of broad catch-all, rather than a true Genre of Film. In short, any comedy, drama, adventure, western or urban cops & robbers shoot-em-up, that are so constructed and so cast as to appeal to the large Urban Black population of the Mid 20th Century. That indeed could embrace the widest type of films, as long as the had a slant toward the inner-city black population.It appears that the idea of producing these films of particularly keen interest to Black Americans had its genesis with the Eastertime Release of 100 RIFLES (Marvin Schwartz Prod./20th Century-Fox, 1969). In it, former Syracuse University All-American Footballer and Several Times All-Pro Fullback for the Cleveland Browns, Jim Brown, had a Co-Starring Billing. Having appeared in a number of films already, as for example, RIO CONCHOS (1964),THE DIRTY DOZEN (1967), (ICE STTION ZEBRA (1968)* and others, it was beginning to make more sense to the Studios\' "Suits" that Jim was a hot property.Now this 100 RIFLES brings record numbers of Black patrons to the Big Cities\' central business districts on Easter Sunday to view Mr. Brown. Why not start to film more of these adventure epics and other types of film with more Black Players and Stars? Why not, indeed.** So we saw a succession of Cops & Robbers, Bad-ass Private Detective Films, Comedies, all going the route. Along the way, we eventually got to some more family oriented, wider appealing films. The movie goers were treated to SOUNDER (1972), THE TAKE (1974), CONRACK (1974)and, ultimately, CLAUDINE (1974).In CLAUDINE, we find no stigma nor easy classification as being "Blackploitation", as the story is universal, and could easily have been done as a story about people of any descent, any where, and not just in the 1970\'s USA.That the story was done of a SINGLE mother, Claudine (Dianne Carroll), struggling to keep a family together after "....two marriages and two almost marriages.", is a far cry from a shoot-em-up Harlem Style. The problems that plague the everyday citizens of our nation are confronted and examined under the ol\' sociological microscope.But we also consider Claudine\'s psychological and physical needs as a female. For "Woman Needs Man and Man Must Have His MATE",***and we do concede this point. (That\'s S-E-X that we\'re talking about, Schultz!) Claudine meets up with a very masculine, broad shouldered, athletic type in Private Scavanger Garbage Man, Ruppert B. Marshall (James Earl Jones) and they go on a date.The Great Welfare State intervenes with the Couple as Claudine\'s Welfare Case Worker, Miss Tayback (Elisa Loti), comes snooping around to see just who is this unattached Male, who is suddenly paying so much attention to Claudine\'s family.After a humiliating experience with the Welfare Bureau\'s auditing and "deducting" binge, which would be the norm for the family, the two decide to get married with or without the blessing of Big Brother.Meanwhile, Claudine\'s elder son has gotten involved with some big talking but little doing Black Activist group. But, with Ruppert\'s help, he and they all come through it A.O.K.It ends on a Happy, Upbeat and Hopeful note. We know that it may not be exactly "...Happily Ever After!", but rather the\'ll make it all together! If there is a single criticism that we must state it is that sometimes in a movie like this, a misconception is spread to a large portion of Urban Blacks. And that is, the apparent implied myth that all Whites are wealthy, having none of their kind ever in need of a helping hand, out of work or suffering any disabilities.Well, folks, it just ain\'t true! NOTE: * At one point, Jim Brown\'s career was a real hit as a rugged actioner. He was even being tauted as "...The Black John Wayne." NOTE: ** The idea of producing films with All-Black Casts, filmed for All-Black consumption was not a new idea. In the 1920\'s, \'30\'s and \'40\'s, we saw productions from people like Noble Johnson, Spencer Williams, Jr. and Rex Ingram.NOTE: *** That\'s "As Time Goes By", you know, Schultz, it\'s from CASABLANCA (Warner Brothers, 1942).',
|
| 290 |
+
"There absolutely was voter fraud. There's voter fraud in every election. However, they are generally isolated incidents and I don't think there has been any credible evidence presented that indicates any wide-scale systemic voter fraud happened in 2020. \n\nI would like a federal commission started that investigates and looks for systemic voter and election fraud. Especially one that would be empowered to look into cases of disenfranchisement and voter suppression as well. Everyone that is legally allowed to vote should be able to easily and securely register and cast their vote.",
|
| 291 |
+
]
|
| 292 |
+
embeddings = model.encode(sentences)
|
| 293 |
+
print(embeddings.shape)
|
| 294 |
+
# [3, 1024]
|
| 295 |
+
|
| 296 |
+
# Get the similarity scores for the embeddings
|
| 297 |
+
similarities = model.similarity(embeddings, embeddings)
|
| 298 |
+
print(similarities)
|
| 299 |
+
# tensor([[1.0000, 0.1908, 0.3587],
|
| 300 |
+
# [0.1908, 1.0000, 0.3531],
|
| 301 |
+
# [0.3587, 0.3531, 1.0000]])
|
| 302 |
+
```
|
| 303 |
+
|
| 304 |
+
<!--
|
| 305 |
+
### Direct Usage (Transformers)
|
| 306 |
+
|
| 307 |
+
<details><summary>Click to see the direct usage in Transformers</summary>
|
| 308 |
+
|
| 309 |
+
</details>
|
| 310 |
+
-->
|
| 311 |
+
|
| 312 |
+
<!--
|
| 313 |
+
### Downstream Usage (Sentence Transformers)
|
| 314 |
+
|
| 315 |
+
You can finetune this model on your own dataset.
|
| 316 |
+
|
| 317 |
+
<details><summary>Click to expand</summary>
|
| 318 |
+
|
| 319 |
+
</details>
|
| 320 |
+
-->
|
| 321 |
+
|
| 322 |
+
<!--
|
| 323 |
+
### Out-of-Scope Use
|
| 324 |
+
|
| 325 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 326 |
+
-->
|
| 327 |
+
|
| 328 |
+
## Evaluation
|
| 329 |
+
|
| 330 |
+
### Metrics
|
| 331 |
+
|
| 332 |
+
#### Semantic Similarity
|
| 333 |
+
|
| 334 |
+
* Dataset: `similarity`
|
| 335 |
+
* Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
|
| 336 |
+
|
| 337 |
+
| Metric | Value |
|
| 338 |
+
|:--------------------|:-----------|
|
| 339 |
+
| pearson_cosine | 0.4106 |
|
| 340 |
+
| **spearman_cosine** | **0.4261** |
|
| 341 |
+
|
| 342 |
+
<!--
|
| 343 |
+
## Bias, Risks and Limitations
|
| 344 |
+
|
| 345 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
| 346 |
+
-->
|
| 347 |
+
|
| 348 |
+
<!--
|
| 349 |
+
### Recommendations
|
| 350 |
+
|
| 351 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
| 352 |
+
-->
|
| 353 |
+
|
| 354 |
+
## Training Details
|
| 355 |
+
|
| 356 |
+
### Training Dataset
|
| 357 |
+
|
| 358 |
+
#### Unnamed Dataset
|
| 359 |
+
|
| 360 |
+
* Size: 11,180 training samples
|
| 361 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
|
| 362 |
+
* Approximate statistics based on the first 1000 samples:
|
| 363 |
+
| | sentence_0 | sentence_1 | label |
|
| 364 |
+
|:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:---------------------------------------------------------------|
|
| 365 |
+
| type | string | string | float |
|
| 366 |
+
| details | <ul><li>min: 4 tokens</li><li>mean: 104.44 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 109.27 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.52</li><li>max: 1.0</li></ul> |
|
| 367 |
+
* Samples:
|
| 368 |
+
| sentence_0 | sentence_1 | label |
|
| 369 |
+
|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------|
|
| 370 |
+
| <code>The concept that things could be possibly be worse therefore do not strive to improve things is a weak and cowardly mentality. <br>Nobody wants to hear your dumbass shit.<br><br>Edit: This dude dm'd me and had a total emotional meltdown, that's how bad my words hurt this man.</code> | <code>Based Macron needs to snort something off of your girlfriends titis</code> | <code>1.0</code> |
|
| 371 |
+
| <code>Even #foxnews pundit Brit Hume is calling this tweet a lie and should be the reason he loses the next election or is impeached & found guilty by the majority Republican Senate ASAP! #MuellerReport. End of story!</code> | <code>An election like this will hardly ever be more decisive, thats just how these things are. I agree its sad that even someone like Le Pen doesnt break the habit.</code> | <code>0.7071067811865475</code> |
|
| 372 |
+
| <code>*review may contain spoilers*predictable, campy, bad special effects. it has a TV-movie feeling to it. the idea of the UN as being taken over by Satan is an interesting twist to the end of the world according to the bible. the premise is interesting, but its excution falls waaaay short. if you want to convert people to Christianity with a film like this, at least make it a quality one! i was seriously checking my watch while watching this piece of dreck. can't say much else about this film since i saw it over a year ago, and there isn't really much to say about this film other than.....skip it!</code> | <code>wonderful movie with good story great humour (some great one-liners) and a soundtrack to die for.i've seen it 3 times so far.the american audiences are going to love it.</code> | <code>0.3333333333333333</code> |
|
| 373 |
+
* Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
|
| 374 |
+
```json
|
| 375 |
+
{
|
| 376 |
+
"loss_fct": "torch.nn.modules.loss.MSELoss"
|
| 377 |
+
}
|
| 378 |
+
```
|
| 379 |
+
|
| 380 |
+
### Training Hyperparameters
|
| 381 |
+
#### Non-Default Hyperparameters
|
| 382 |
+
|
| 383 |
+
- `eval_strategy`: steps
|
| 384 |
+
- `per_device_train_batch_size`: 32
|
| 385 |
+
- `per_device_eval_batch_size`: 32
|
| 386 |
+
- `fp16`: True
|
| 387 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 388 |
+
|
| 389 |
+
#### All Hyperparameters
|
| 390 |
+
<details><summary>Click to expand</summary>
|
| 391 |
+
|
| 392 |
+
- `overwrite_output_dir`: False
|
| 393 |
+
- `do_predict`: False
|
| 394 |
+
- `eval_strategy`: steps
|
| 395 |
+
- `prediction_loss_only`: True
|
| 396 |
+
- `per_device_train_batch_size`: 32
|
| 397 |
+
- `per_device_eval_batch_size`: 32
|
| 398 |
+
- `per_gpu_train_batch_size`: None
|
| 399 |
+
- `per_gpu_eval_batch_size`: None
|
| 400 |
+
- `gradient_accumulation_steps`: 1
|
| 401 |
+
- `eval_accumulation_steps`: None
|
| 402 |
+
- `torch_empty_cache_steps`: None
|
| 403 |
+
- `learning_rate`: 5e-05
|
| 404 |
+
- `weight_decay`: 0.0
|
| 405 |
+
- `adam_beta1`: 0.9
|
| 406 |
+
- `adam_beta2`: 0.999
|
| 407 |
+
- `adam_epsilon`: 1e-08
|
| 408 |
+
- `max_grad_norm`: 1
|
| 409 |
+
- `num_train_epochs`: 3
|
| 410 |
+
- `max_steps`: -1
|
| 411 |
+
- `lr_scheduler_type`: linear
|
| 412 |
+
- `lr_scheduler_kwargs`: {}
|
| 413 |
+
- `warmup_ratio`: 0.0
|
| 414 |
+
- `warmup_steps`: 0
|
| 415 |
+
- `log_level`: passive
|
| 416 |
+
- `log_level_replica`: warning
|
| 417 |
+
- `log_on_each_node`: True
|
| 418 |
+
- `logging_nan_inf_filter`: True
|
| 419 |
+
- `save_safetensors`: True
|
| 420 |
+
- `save_on_each_node`: False
|
| 421 |
+
- `save_only_model`: False
|
| 422 |
+
- `restore_callback_states_from_checkpoint`: False
|
| 423 |
+
- `no_cuda`: False
|
| 424 |
+
- `use_cpu`: False
|
| 425 |
+
- `use_mps_device`: False
|
| 426 |
+
- `seed`: 42
|
| 427 |
+
- `data_seed`: None
|
| 428 |
+
- `jit_mode_eval`: False
|
| 429 |
+
- `use_ipex`: False
|
| 430 |
+
- `bf16`: False
|
| 431 |
+
- `fp16`: True
|
| 432 |
+
- `fp16_opt_level`: O1
|
| 433 |
+
- `half_precision_backend`: auto
|
| 434 |
+
- `bf16_full_eval`: False
|
| 435 |
+
- `fp16_full_eval`: False
|
| 436 |
+
- `tf32`: None
|
| 437 |
+
- `local_rank`: 0
|
| 438 |
+
- `ddp_backend`: None
|
| 439 |
+
- `tpu_num_cores`: None
|
| 440 |
+
- `tpu_metrics_debug`: False
|
| 441 |
+
- `debug`: []
|
| 442 |
+
- `dataloader_drop_last`: False
|
| 443 |
+
- `dataloader_num_workers`: 0
|
| 444 |
+
- `dataloader_prefetch_factor`: None
|
| 445 |
+
- `past_index`: -1
|
| 446 |
+
- `disable_tqdm`: False
|
| 447 |
+
- `remove_unused_columns`: True
|
| 448 |
+
- `label_names`: None
|
| 449 |
+
- `load_best_model_at_end`: False
|
| 450 |
+
- `ignore_data_skip`: False
|
| 451 |
+
- `fsdp`: []
|
| 452 |
+
- `fsdp_min_num_params`: 0
|
| 453 |
+
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
|
| 454 |
+
- `fsdp_transformer_layer_cls_to_wrap`: None
|
| 455 |
+
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
|
| 456 |
+
- `deepspeed`: None
|
| 457 |
+
- `label_smoothing_factor`: 0.0
|
| 458 |
+
- `optim`: adamw_torch
|
| 459 |
+
- `optim_args`: None
|
| 460 |
+
- `adafactor`: False
|
| 461 |
+
- `group_by_length`: False
|
| 462 |
+
- `length_column_name`: length
|
| 463 |
+
- `ddp_find_unused_parameters`: None
|
| 464 |
+
- `ddp_bucket_cap_mb`: None
|
| 465 |
+
- `ddp_broadcast_buffers`: False
|
| 466 |
+
- `dataloader_pin_memory`: True
|
| 467 |
+
- `dataloader_persistent_workers`: False
|
| 468 |
+
- `skip_memory_metrics`: True
|
| 469 |
+
- `use_legacy_prediction_loop`: False
|
| 470 |
+
- `push_to_hub`: False
|
| 471 |
+
- `resume_from_checkpoint`: None
|
| 472 |
+
- `hub_model_id`: None
|
| 473 |
+
- `hub_strategy`: every_save
|
| 474 |
+
- `hub_private_repo`: None
|
| 475 |
+
- `hub_always_push`: False
|
| 476 |
+
- `hub_revision`: None
|
| 477 |
+
- `gradient_checkpointing`: False
|
| 478 |
+
- `gradient_checkpointing_kwargs`: None
|
| 479 |
+
- `include_inputs_for_metrics`: False
|
| 480 |
+
- `include_for_metrics`: []
|
| 481 |
+
- `eval_do_concat_batches`: True
|
| 482 |
+
- `fp16_backend`: auto
|
| 483 |
+
- `push_to_hub_model_id`: None
|
| 484 |
+
- `push_to_hub_organization`: None
|
| 485 |
+
- `mp_parameters`:
|
| 486 |
+
- `auto_find_batch_size`: False
|
| 487 |
+
- `full_determinism`: False
|
| 488 |
+
- `torchdynamo`: None
|
| 489 |
+
- `ray_scope`: last
|
| 490 |
+
- `ddp_timeout`: 1800
|
| 491 |
+
- `torch_compile`: False
|
| 492 |
+
- `torch_compile_backend`: None
|
| 493 |
+
- `torch_compile_mode`: None
|
| 494 |
+
- `include_tokens_per_second`: False
|
| 495 |
+
- `include_num_input_tokens_seen`: False
|
| 496 |
+
- `neftune_noise_alpha`: None
|
| 497 |
+
- `optim_target_modules`: None
|
| 498 |
+
- `batch_eval_metrics`: False
|
| 499 |
+
- `eval_on_start`: False
|
| 500 |
+
- `use_liger_kernel`: False
|
| 501 |
+
- `liger_kernel_config`: None
|
| 502 |
+
- `eval_use_gather_object`: False
|
| 503 |
+
- `average_tokens_across_devices`: False
|
| 504 |
+
- `prompts`: None
|
| 505 |
+
- `batch_sampler`: batch_sampler
|
| 506 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 507 |
+
- `router_mapping`: {}
|
| 508 |
+
- `learning_rate_mapping`: {}
|
| 509 |
+
|
| 510 |
+
</details>
|
| 511 |
+
|
| 512 |
+
### Training Logs
|
| 513 |
+
| Epoch | Step | Training Loss | similarity_spearman_cosine |
|
| 514 |
+
|:------:|:----:|:-------------:|:--------------------------:|
|
| 515 |
+
| 0.0286 | 10 | - | 0.1359 |
|
| 516 |
+
| 0.0571 | 20 | - | 0.1424 |
|
| 517 |
+
| 0.0857 | 30 | - | 0.1525 |
|
| 518 |
+
| 0.1143 | 40 | - | 0.1651 |
|
| 519 |
+
| 0.1429 | 50 | - | 0.1759 |
|
| 520 |
+
| 0.1714 | 60 | - | 0.1846 |
|
| 521 |
+
| 0.2 | 70 | - | 0.1947 |
|
| 522 |
+
| 0.2286 | 80 | - | 0.2056 |
|
| 523 |
+
| 0.2571 | 90 | - | 0.2144 |
|
| 524 |
+
| 0.2857 | 100 | - | 0.2298 |
|
| 525 |
+
| 0.3143 | 110 | - | 0.2409 |
|
| 526 |
+
| 0.3429 | 120 | - | 0.2526 |
|
| 527 |
+
| 0.3714 | 130 | - | 0.2511 |
|
| 528 |
+
| 0.4 | 140 | - | 0.2661 |
|
| 529 |
+
| 0.4286 | 150 | - | 0.2664 |
|
| 530 |
+
| 0.4571 | 160 | - | 0.2572 |
|
| 531 |
+
| 0.4857 | 170 | - | 0.2804 |
|
| 532 |
+
| 0.5143 | 180 | - | 0.2885 |
|
| 533 |
+
| 0.5429 | 190 | - | 0.2885 |
|
| 534 |
+
| 0.5714 | 200 | - | 0.2933 |
|
| 535 |
+
| 0.6 | 210 | - | 0.3037 |
|
| 536 |
+
| 0.6286 | 220 | - | 0.3163 |
|
| 537 |
+
| 0.6571 | 230 | - | 0.3197 |
|
| 538 |
+
| 0.6857 | 240 | - | 0.3275 |
|
| 539 |
+
| 0.7143 | 250 | - | 0.3238 |
|
| 540 |
+
| 0.7429 | 260 | - | 0.3262 |
|
| 541 |
+
| 0.7714 | 270 | - | 0.3295 |
|
| 542 |
+
| 0.8 | 280 | - | 0.3129 |
|
| 543 |
+
| 0.8286 | 290 | - | 0.3491 |
|
| 544 |
+
| 0.8571 | 300 | - | 0.3354 |
|
| 545 |
+
| 0.8857 | 310 | - | 0.3448 |
|
| 546 |
+
| 0.9143 | 320 | - | 0.3581 |
|
| 547 |
+
| 0.9429 | 330 | - | 0.3658 |
|
| 548 |
+
| 0.9714 | 340 | - | 0.3386 |
|
| 549 |
+
| 1.0 | 350 | - | 0.3503 |
|
| 550 |
+
| 1.0286 | 360 | - | 0.3533 |
|
| 551 |
+
| 1.0571 | 370 | - | 0.3604 |
|
| 552 |
+
| 1.0857 | 380 | - | 0.3624 |
|
| 553 |
+
| 1.1143 | 390 | - | 0.3549 |
|
| 554 |
+
| 1.1429 | 400 | - | 0.3594 |
|
| 555 |
+
| 1.1714 | 410 | - | 0.3747 |
|
| 556 |
+
| 1.2 | 420 | - | 0.3465 |
|
| 557 |
+
| 1.2286 | 430 | - | 0.3378 |
|
| 558 |
+
| 1.2571 | 440 | - | 0.3809 |
|
| 559 |
+
| 1.2857 | 450 | - | 0.3856 |
|
| 560 |
+
| 1.3143 | 460 | - | 0.3522 |
|
| 561 |
+
| 1.3429 | 470 | - | 0.3987 |
|
| 562 |
+
| 1.3714 | 480 | - | 0.3847 |
|
| 563 |
+
| 1.4 | 490 | - | 0.3688 |
|
| 564 |
+
| 1.4286 | 500 | 0.1157 | 0.3937 |
|
| 565 |
+
| 1.4571 | 510 | - | 0.3857 |
|
| 566 |
+
| 1.4857 | 520 | - | 0.4039 |
|
| 567 |
+
| 1.5143 | 530 | - | 0.3913 |
|
| 568 |
+
| 1.5429 | 540 | - | 0.3900 |
|
| 569 |
+
| 1.5714 | 550 | - | 0.3497 |
|
| 570 |
+
| 1.6 | 560 | - | 0.3613 |
|
| 571 |
+
| 1.6286 | 570 | - | 0.4067 |
|
| 572 |
+
| 1.6571 | 580 | - | 0.4016 |
|
| 573 |
+
| 1.6857 | 590 | - | 0.3954 |
|
| 574 |
+
| 1.7143 | 600 | - | 0.3947 |
|
| 575 |
+
| 1.7429 | 610 | - | 0.3864 |
|
| 576 |
+
| 1.7714 | 620 | - | 0.4194 |
|
| 577 |
+
| 1.8 | 630 | - | 0.3985 |
|
| 578 |
+
| 1.8286 | 640 | - | 0.4003 |
|
| 579 |
+
| 1.8571 | 650 | - | 0.4061 |
|
| 580 |
+
| 1.8857 | 660 | - | 0.4074 |
|
| 581 |
+
| 1.9143 | 670 | - | 0.4004 |
|
| 582 |
+
| 1.9429 | 680 | - | 0.4022 |
|
| 583 |
+
| 1.9714 | 690 | - | 0.4056 |
|
| 584 |
+
| 2.0 | 700 | - | 0.3991 |
|
| 585 |
+
| 2.0286 | 710 | - | 0.3944 |
|
| 586 |
+
| 2.0571 | 720 | - | 0.3952 |
|
| 587 |
+
| 2.0857 | 730 | - | 0.4014 |
|
| 588 |
+
| 2.1143 | 740 | - | 0.3846 |
|
| 589 |
+
| 2.1429 | 750 | - | 0.3719 |
|
| 590 |
+
| 2.1714 | 760 | - | 0.4073 |
|
| 591 |
+
| 2.2 | 770 | - | 0.3828 |
|
| 592 |
+
| 2.2286 | 780 | - | 0.3858 |
|
| 593 |
+
| 2.2571 | 790 | - | 0.4114 |
|
| 594 |
+
| 2.2857 | 800 | - | 0.3930 |
|
| 595 |
+
| 2.3143 | 810 | - | 0.3845 |
|
| 596 |
+
| 2.3429 | 820 | - | 0.4053 |
|
| 597 |
+
| 2.3714 | 830 | - | 0.3582 |
|
| 598 |
+
| 2.4 | 840 | - | 0.3848 |
|
| 599 |
+
| 2.4286 | 850 | - | 0.4139 |
|
| 600 |
+
| 2.4571 | 860 | - | 0.3609 |
|
| 601 |
+
| 2.4857 | 870 | - | 0.4122 |
|
| 602 |
+
| 2.5143 | 880 | - | 0.4101 |
|
| 603 |
+
| 2.5429 | 890 | - | 0.4261 |
|
| 604 |
+
|
| 605 |
+
|
| 606 |
+
### Framework Versions
|
| 607 |
+
- Python: 3.11.9
|
| 608 |
+
- Sentence Transformers: 5.1.0
|
| 609 |
+
- Transformers: 4.53.3
|
| 610 |
+
- PyTorch: 2.5.1
|
| 611 |
+
- Accelerate: 1.10.0
|
| 612 |
+
- Datasets: 2.14.4
|
| 613 |
+
- Tokenizers: 0.21.0
|
| 614 |
+
|
| 615 |
+
## Citation
|
| 616 |
+
|
| 617 |
+
### BibTeX
|
| 618 |
+
|
| 619 |
+
#### Sentence Transformers
|
| 620 |
+
```bibtex
|
| 621 |
+
@inproceedings{reimers-2019-sentence-bert,
|
| 622 |
+
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|
| 623 |
+
author = "Reimers, Nils and Gurevych, Iryna",
|
| 624 |
+
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
|
| 625 |
+
month = "11",
|
| 626 |
+
year = "2019",
|
| 627 |
+
publisher = "Association for Computational Linguistics",
|
| 628 |
+
url = "https://arxiv.org/abs/1908.10084",
|
| 629 |
+
}
|
| 630 |
+
```
|
| 631 |
+
|
| 632 |
+
<!--
|
| 633 |
+
## Glossary
|
| 634 |
+
|
| 635 |
+
*Clearly define terms in order to be accessible across audiences.*
|
| 636 |
+
-->
|
| 637 |
+
|
| 638 |
+
<!--
|
| 639 |
+
## Model Card Authors
|
| 640 |
+
|
| 641 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
| 642 |
+
-->
|
| 643 |
+
|
| 644 |
+
<!--
|
| 645 |
+
## Model Card Contact
|
| 646 |
+
|
| 647 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
| 648 |
+
-->
|
config.json
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"BertModel"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.1,
|
| 6 |
+
"classifier_dropout": null,
|
| 7 |
+
"hidden_act": "gelu",
|
| 8 |
+
"hidden_dropout_prob": 0.1,
|
| 9 |
+
"hidden_size": 1024,
|
| 10 |
+
"initializer_range": 0.02,
|
| 11 |
+
"intermediate_size": 4096,
|
| 12 |
+
"layer_norm_eps": 1e-12,
|
| 13 |
+
"max_position_embeddings": 512,
|
| 14 |
+
"model_type": "bert",
|
| 15 |
+
"num_attention_heads": 16,
|
| 16 |
+
"num_hidden_layers": 24,
|
| 17 |
+
"pad_token_id": 0,
|
| 18 |
+
"position_embedding_type": "absolute",
|
| 19 |
+
"torch_dtype": "float32",
|
| 20 |
+
"transformers_version": "4.53.3",
|
| 21 |
+
"type_vocab_size": 2,
|
| 22 |
+
"use_cache": true,
|
| 23 |
+
"vocab_size": 30522
|
| 24 |
+
}
|
config_sentence_transformers.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_type": "SentenceTransformer",
|
| 3 |
+
"__version__": {
|
| 4 |
+
"sentence_transformers": "5.1.0",
|
| 5 |
+
"transformers": "4.53.3",
|
| 6 |
+
"pytorch": "2.5.1"
|
| 7 |
+
},
|
| 8 |
+
"prompts": {
|
| 9 |
+
"query": "",
|
| 10 |
+
"document": ""
|
| 11 |
+
},
|
| 12 |
+
"default_prompt_name": null,
|
| 13 |
+
"similarity_fn_name": "cosine"
|
| 14 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:faa9fd1420b2d60582e5e578fe49d3a281c0b0575ad2667ab4693179f64142a2
|
| 3 |
+
size 1340612432
|
modules.json
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"idx": 0,
|
| 4 |
+
"name": "0",
|
| 5 |
+
"path": "",
|
| 6 |
+
"type": "sentence_transformers.models.Transformer"
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"idx": 1,
|
| 10 |
+
"name": "1",
|
| 11 |
+
"path": "1_Pooling",
|
| 12 |
+
"type": "sentence_transformers.models.Pooling"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"idx": 2,
|
| 16 |
+
"name": "2",
|
| 17 |
+
"path": "2_Normalize",
|
| 18 |
+
"type": "sentence_transformers.models.Normalize"
|
| 19 |
+
}
|
| 20 |
+
]
|
sentence_bert_config.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"max_seq_length": 512,
|
| 3 |
+
"do_lower_case": false
|
| 4 |
+
}
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": true,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": true,
|
| 47 |
+
"extra_special_tokens": {},
|
| 48 |
+
"mask_token": "[MASK]",
|
| 49 |
+
"model_max_length": 512,
|
| 50 |
+
"pad_token": "[PAD]",
|
| 51 |
+
"sep_token": "[SEP]",
|
| 52 |
+
"strip_accents": null,
|
| 53 |
+
"tokenize_chinese_chars": true,
|
| 54 |
+
"tokenizer_class": "BertTokenizer",
|
| 55 |
+
"unk_token": "[UNK]"
|
| 56 |
+
}
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|