Culture-and-Morality-Lab commited on
Commit
fbaa37f
·
verified ·
1 Parent(s): 2356c36

Upload 10 files

Browse files
README.md ADDED
@@ -0,0 +1,648 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - dense
7
+ - generated_from_trainer
8
+ - dataset_size:11180
9
+ - loss:CosineSimilarityLoss
10
+ widget:
11
+ - source_sentence: Of course she would, otherwise she stands no chance of becoming
12
+ prez. The only thing above Le Pen's xenophobia is their thrive for power.
13
+ sentences:
14
+ - Mormon Church declares same-sex couples apostates and excludes children of those
15
+ couples from blessings and baptism.
16
+ - Feminists have legitimate gripes with the way the world is structured and their
17
+ ideas are quite sane. The feminists of the world are not deluded or wacky and
18
+ it's a bad idea to call them that. Open your mind to new ideas and drop the patriarchal
19
+ thinking.
20
+ - Hi, Professor Lichtman. Thanks for doing this AMA. What would you say to people
21
+ who argue about the economy keys being affected by people not feeling, in terms
22
+ of their lived experiences, that the economy has been good (due to the cost of
23
+ living exceeding, in many cases, their income) and this potentially costing Harris
24
+ the election? They seem to believe that this should be costing her the ST and
25
+ LT economy keys.
26
+ - source_sentence: Sondra Locke stinks in this film, but then she was an awful 'actress'
27
+ anyway. Unfortunately, she drags everyone else (including then =real life boyfriend
28
+ Clint Eastwood down the drain with her. But what was Clint Eastwood thinking when
29
+ he agreed to star in this one? One read of the script should have told him that
30
+ this one was going to be a real snorer. It's an exceptionally weak story, basically
31
+ no story or plot at all. Add in bored, poor acting, even from the normally good
32
+ Eastwood. There's absolutely no action except a couple arguments and as far as
33
+ I was concerned, this film ranks up at the top of the heap of natural sleep enhancers.
34
+ Wow! Could a film BE any more boring? I think watching paint dry or the grass
35
+ grow might be more fun. A real stinker. Don't bother with this one.
36
+ sentences:
37
+ - 'SPOILERS: The original Road House is one of those movies that I know is clichéd
38
+ and unoriginal, yet it''s done so well, I''m embarrassed to admit I really like
39
+ it. Turns out many of my friends, whose movie opinions I respect, think the same
40
+ way. So when they attempt to make a sequel to it and it''s as if it was written
41
+ by some high school kids who were given the rights to do a sequel, it''s just
42
+ bad; really, really bad.Oddly, Johnathon Schaech is listed as one of the writers
43
+ and I can only hope his WGA membership is revoked. The writing was just bad and
44
+ all the writers of this film should retire for complete lack of originality and
45
+ some of the worst dialog in this millennium. Schaech already appearing to be the
46
+ king of the straight-to-DVD sequel (8mm 2, Poison Ivy 2) and now after seeing
47
+ this and 8mm 2, I''m thinking his acting ability is non-existent. He was awful,
48
+ just awful.And it''s not the terrible fighting scenes that make this movie terrible,
49
+ but take it from me, they''re bad. Every fight scene is a slowly delivered punch
50
+ (yet still making the "wiff" sound in the air) that is then blocked by the opponent,
51
+ who returns a punch that sends the first guy to the ground. This is repeated throughout
52
+ the film, worse than any bad 1970s cop show. Or the fact that many of the people
53
+ involved in the fights seem to have a mouth full of cherry kool-aid for some reason.
54
+ And we''re supposed to believe Will Patton is a fighting machine; his fight scenes
55
+ look so amazingly fake I was honestly embarrassed watching. It''s the complete
56
+ lapses of logic in this ridiculous movie that make it terrible. For instance:
57
+ Johnathon Schaech''s character is in town for a day and already tells some girl
58
+ he barely knows who he has no idea what side she''s on, "I''m with the Feds, but
59
+ don''t tell anyone." The female villain, who fights the good girl in one fight
60
+ scene with acrobatics that rival any super hero, yet is easily held down by the
61
+ Will Patton, "old guy," in another scene by simply holding both her hands while
62
+ he utters some ridiculous line ("stab me once, shame on you, stab me twice, ain''t
63
+ gonna happen" whew, that''s bad) and then head-butts her. Jake Busey''s villain
64
+ shoots at the feds while caught in the middle of a drug deal, yet no DEA agents
65
+ or anyone simply go to his place and pick him up after, in fact, he''s simply
66
+ let go because "this is the sheriff''s territory." Busey wants the bar because
67
+ it''s "in a great location" for drug deals, yet his own house appears to be just
68
+ as good apparently offering all the perks the bar is supposed to have. Johnathon
69
+ Schaech''s character is supposed to be the son of Patrick Swayze''s character
70
+ in the original, yet Swayze''s character''s last name is Dalton and Schaech''s
71
+ isn''t (nor is the supposed brother of Swayze''s character). And Johnathon Schaech
72
+ looks about 50 in this movie. I looked it up, he''s 17 years younger than Swayze,
73
+ but he looks awful.But my favorite absolutely stupid scene in this movie was the
74
+ most stock fight scene ending in movies: the villain is knocked through a window
75
+ on a second floor and as they pan down I''m thinking "please don''t tell me he''s
76
+ impaled on something..." and sure enough, my worst fears were realized.Actually,
77
+ I could go on for another half hour about the things I hated about this movie.
78
+ Suffice to say, let''s put an end to these ridiculous straight-to-DVD sequels
79
+ to theatrical movies, at least the ones with Johnathon Schaech.'
80
+ - 'Who do neoliberals scapegoat? The uneducated I guess? But we certainly don''t
81
+ advocate for their destruction or dissolution or deportation, unless you think
82
+ wanting to improve education standards and accessibility is violent?
83
+
84
+
85
+ I have yet to see anyone unironically idolize Macron like the right does Trump
86
+ and the left does Bernie.'
87
+ - 'Republicans created their own safe space in arcon (flaired users only) since
88
+ their facts are fake and they are too much of snowflakes to take the criticism
89
+ of being fact checked, just like how Trump gets mad when fact checked.
90
+
91
+
92
+ Republicans don''t post here because they are cowards.'
93
+ - source_sentence: U.S. Forces Kill 50 Sadr Militia in Baghdad Suburb (Reuters) Reuters
94
+ - U.S. forces killed more than 50 Shi'ite\militiamen Wednesday in a significant
95
+ advance into a Baghdad\suburb that is a powerbase for radical cleric Moqtada al-Sadr,\the
96
+ military said.
97
+ sentences:
98
+ - "'You haven't had it as bad as I had it,' is exactly why Millennials are the first\
99
+ \ generation ever to be worse off than their parents.\n \n If we're on the same\
100
+ \ team, let's act like it.\n\nFucking thank you!\n\nI literally didn't mention\
101
+ \ Boomers or say a damn thing negative about them. I shared my experience as\
102
+ \ a Millenial in a thread for an article specifically about my generation. And\
103
+ \ of fucking course a Boomer came in to tell me I don't know what real struggle\
104
+ \ is. Like... okay, what the hell is your point Karen?"
105
+ - '"Dutch Schultz", AKA Arthur Fleggenheimer, was a real person and his rather nasty
106
+ life is fairly well documented. This movie which purports to depict his life should
107
+ have used a fictional character, because the overdramatized events are too strong
108
+ a departure from the facts and the chronology. Not only that, it ignores some
109
+ interesting details which other versions have included such as the public relations
110
+ fiasco in upstate N.Y. and his religious conversion. It is true that he was executed
111
+ by Luciano, Lansky, et. al. but that''s as far as it goes. The exploding plate
112
+ scene which represents Luciano carrying out the execution of Bo Weinberg in his
113
+ own home, assisted by his own mother is rediculous. Also, there is the scene in
114
+ which Dutch approaches his own mother to pay protection to Legs Diamond. It just
115
+ doesn''t work. The character of Mrs. Fleggenheimer doesn''t work either. This
116
+ movie does not need a doting Jewish mother for comic relief. The lame representation
117
+ of Legs Diamond was humorous enough. I''m sure the man is turning in his grave.
118
+ And, by the way, Dutch did in fact personally kill people, but, he was not Rambo
119
+ or 007. The scene in which he wipes out the brewery is absurd. I don''t know.
120
+ Maybe it was supposed to be a comedy and I just didn''t get it.'
121
+ - Can we stop with the whole "leader/s of the free world"? What does that even mean?
122
+ As far as I know we're all from distinct sovereign countries for now. Like I don't
123
+ hate Macron or Merkel and they are good options so far, but this crap with "leaders
124
+ of the free world" is just obnoxious.
125
+ - source_sentence: Reasoned arguments and suggestions which make allowance for the
126
+ full difficulties of the state of war that exists may help, and will always be
127
+ listened to with respect and sympathy.
128
+ sentences:
129
+ - 'Because they''re primarily just racists. If they truly believed in their stated
130
+ ideology they would want Macron to win. If you''re an America First nationalist
131
+ who thinks globalism is terrible then you should want globalists in charge of
132
+ every other country.
133
+
134
+
135
+ Edit: To be fair it''s possible that they''re not racist but merely too stupid
136
+ to think through the basic logic of their own ideology. It''s one or the other.'
137
+ - This is the last episode of the Goldenboy OVA series. Kentaro finds himself working
138
+ in an animation studio, which is rather interesting if you don't know anything
139
+ about the way anime studios were run. Besides episode 3, this was probably the
140
+ least risqué, but it had a nice girl interest, as well as a surprise reunion from
141
+ others in the previous episodes. My only complaint about this episode is it seemed
142
+ a little too short, but at the same time this may have only been because it was
143
+ the only original script for the show that wasn't based on one of the manga chapters.
144
+ but it ended well, leaving us with the nice feeling that Kentaro is permanently
145
+ 25, studying on. Definitely watch the rest of the series all the way through,
146
+ you can buy the whole series for like $17, you can watch it all the way through
147
+ in about 2 1/2 hours, or watch your favorite episode if you have 20 minutes free
148
+ time (which i do if i have a lunch break at school.) good series, check it out.
149
+ - 'DON''T STOP NOW: DEMAND AN END TO CHEMICAL CASTRATION of CHILDREN! Under Scrutiny,
150
+ Texas Judge Caves Under Pressure, Grants Father ''A Say'' In Son''s Transitioning
151
+ - But Only with a Gag Order. STAND YOUR GROUND AND KEEP FIGHTING!.'
152
+ - source_sentence: Nationalism is a silly cock crowing on his own dunghill.
153
+ sentences:
154
+ - "There absolutely was voter fraud. There's voter fraud in every election. However,\
155
+ \ they are generally isolated incidents and I don't think there has been any credible\
156
+ \ evidence presented that indicates any wide-scale systemic voter fraud happened\
157
+ \ in 2020. \n\nI would like a federal commission started that investigates and\
158
+ \ looks for systemic voter and election fraud. Especially one that would be empowered\
159
+ \ to look into cases of disenfranchisement and voter suppression as well. Everyone\
160
+ \ that is legally allowed to vote should be able to easily and securely register\
161
+ \ and cast their vote."
162
+ - 'The 1970''s saw a rise and fall of what we have come to know as "Blacksploitation"
163
+ Films. The term is a reference to kind of broad catch-all, rather than a true
164
+ Genre of Film. In short, any comedy, drama, adventure, western or urban cops &
165
+ robbers shoot-em-up, that are so constructed and so cast as to appeal to the large
166
+ Urban Black population of the Mid 20th Century. That indeed could embrace the
167
+ widest type of films, as long as the had a slant toward the inner-city black population.It
168
+ appears that the idea of producing these films of particularly keen interest to
169
+ Black Americans had its genesis with the Eastertime Release of 100 RIFLES (Marvin
170
+ Schwartz Prod./20th Century-Fox, 1969). In it, former Syracuse University All-American
171
+ Footballer and Several Times All-Pro Fullback for the Cleveland Browns, Jim Brown,
172
+ had a Co-Starring Billing. Having appeared in a number of films already, as for
173
+ example, RIO CONCHOS (1964),THE DIRTY DOZEN (1967), (ICE STTION ZEBRA (1968)*
174
+ and others, it was beginning to make more sense to the Studios'' "Suits" that
175
+ Jim was a hot property.Now this 100 RIFLES brings record numbers of Black patrons
176
+ to the Big Cities'' central business districts on Easter Sunday to view Mr. Brown.
177
+ Why not start to film more of these adventure epics and other types of film with
178
+ more Black Players and Stars? Why not, indeed.** So we saw a succession of Cops
179
+ & Robbers, Bad-ass Private Detective Films, Comedies, all going the route. Along
180
+ the way, we eventually got to some more family oriented, wider appealing films.
181
+ The movie goers were treated to SOUNDER (1972), THE TAKE (1974), CONRACK (1974)and,
182
+ ultimately, CLAUDINE (1974).In CLAUDINE, we find no stigma nor easy classification
183
+ as being "Blackploitation", as the story is universal, and could easily have been
184
+ done as a story about people of any descent, any where, and not just in the 1970''s
185
+ USA.That the story was done of a SINGLE mother, Claudine (Dianne Carroll), struggling
186
+ to keep a family together after "....two marriages and two almost marriages.",
187
+ is a far cry from a shoot-em-up Harlem Style. The problems that plague the everyday
188
+ citizens of our nation are confronted and examined under the ol'' sociological
189
+ microscope.But we also consider Claudine''s psychological and physical needs as
190
+ a female. For "Woman Needs Man and Man Must Have His MATE",***and we do concede
191
+ this point. (That''s S-E-X that we''re talking about, Schultz!) Claudine meets
192
+ up with a very masculine, broad shouldered, athletic type in Private Scavanger
193
+ Garbage Man, Ruppert B. Marshall (James Earl Jones) and they go on a date.The
194
+ Great Welfare State intervenes with the Couple as Claudine''s Welfare Case Worker,
195
+ Miss Tayback (Elisa Loti), comes snooping around to see just who is this unattached
196
+ Male, who is suddenly paying so much attention to Claudine''s family.After a humiliating
197
+ experience with the Welfare Bureau''s auditing and "deducting" binge, which would
198
+ be the norm for the family, the two decide to get married with or without the
199
+ blessing of Big Brother.Meanwhile, Claudine''s elder son has gotten involved with
200
+ some big talking but little doing Black Activist group. But, with Ruppert''s help,
201
+ he and they all come through it A.O.K.It ends on a Happy, Upbeat and Hopeful note.
202
+ We know that it may not be exactly "...Happily Ever After!", but rather the''ll
203
+ make it all together! If there is a single criticism that we must state it is
204
+ that sometimes in a movie like this, a misconception is spread to a large portion
205
+ of Urban Blacks. And that is, the apparent implied myth that all Whites are wealthy,
206
+ having none of their kind ever in need of a helping hand, out of work or suffering
207
+ any disabilities.Well, folks, it just ain''t true! NOTE: * At one point, Jim Brown''s
208
+ career was a real hit as a rugged actioner. He was even being tauted as "...The
209
+ Black John Wayne." NOTE: ** The idea of producing films with All-Black Casts,
210
+ filmed for All-Black consumption was not a new idea. In the 1920''s, ''30''s and
211
+ ''40''s, we saw productions from people like Noble Johnson, Spencer Williams,
212
+ Jr. and Rex Ingram.NOTE: *** That''s "As Time Goes By", you know, Schultz, it''s
213
+ from CASABLANCA (Warner Brothers, 1942).'
214
+ - Thinks gun confiscation is . . .
215
+ pipeline_tag: sentence-similarity
216
+ library_name: sentence-transformers
217
+ metrics:
218
+ - pearson_cosine
219
+ - spearman_cosine
220
+ model-index:
221
+ - name: SentenceTransformer
222
+ results:
223
+ - task:
224
+ type: semantic-similarity
225
+ name: Semantic Similarity
226
+ dataset:
227
+ name: similarity
228
+ type: similarity
229
+ metrics:
230
+ - type: pearson_cosine
231
+ value: 0.41059188174486916
232
+ name: Pearson Cosine
233
+ - type: spearman_cosine
234
+ value: 0.4260753939913245
235
+ name: Spearman Cosine
236
+ ---
237
+
238
+ # SentenceTransformer
239
+
240
+ This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
241
+
242
+ ## Model Details
243
+
244
+ ### Model Description
245
+ - **Model Type:** Sentence Transformer
246
+ <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
247
+ - **Maximum Sequence Length:** 512 tokens
248
+ - **Output Dimensionality:** 1024 dimensions
249
+ - **Similarity Function:** Cosine Similarity
250
+ <!-- - **Training Dataset:** Unknown -->
251
+ <!-- - **Language:** Unknown -->
252
+ <!-- - **License:** Unknown -->
253
+
254
+ ### Model Sources
255
+
256
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
257
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
258
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
259
+
260
+ ### Full Model Architecture
261
+
262
+ ```
263
+ SentenceTransformer(
264
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'BertModel'})
265
+ (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
266
+ (2): Normalize()
267
+ )
268
+ ```
269
+
270
+ ## Usage
271
+
272
+ ### Direct Usage (Sentence Transformers)
273
+
274
+ First install the Sentence Transformers library:
275
+
276
+ ```bash
277
+ pip install -U sentence-transformers
278
+ ```
279
+
280
+ Then you can load this model and run inference.
281
+ ```python
282
+ from sentence_transformers import SentenceTransformer
283
+
284
+ # Download from the 🤗 Hub
285
+ model = SentenceTransformer("sentence_transformers_model_id")
286
+ # Run inference
287
+ sentences = [
288
+ 'Nationalism is a silly cock crowing on his own dunghill.',
289
+ 'The 1970\'s saw a rise and fall of what we have come to know as "Blacksploitation" Films. The term is a reference to kind of broad catch-all, rather than a true Genre of Film. In short, any comedy, drama, adventure, western or urban cops & robbers shoot-em-up, that are so constructed and so cast as to appeal to the large Urban Black population of the Mid 20th Century. That indeed could embrace the widest type of films, as long as the had a slant toward the inner-city black population.It appears that the idea of producing these films of particularly keen interest to Black Americans had its genesis with the Eastertime Release of 100 RIFLES (Marvin Schwartz Prod./20th Century-Fox, 1969). In it, former Syracuse University All-American Footballer and Several Times All-Pro Fullback for the Cleveland Browns, Jim Brown, had a Co-Starring Billing. Having appeared in a number of films already, as for example, RIO CONCHOS (1964),THE DIRTY DOZEN (1967), (ICE STTION ZEBRA (1968)* and others, it was beginning to make more sense to the Studios\' "Suits" that Jim was a hot property.Now this 100 RIFLES brings record numbers of Black patrons to the Big Cities\' central business districts on Easter Sunday to view Mr. Brown. Why not start to film more of these adventure epics and other types of film with more Black Players and Stars? Why not, indeed.** So we saw a succession of Cops & Robbers, Bad-ass Private Detective Films, Comedies, all going the route. Along the way, we eventually got to some more family oriented, wider appealing films. The movie goers were treated to SOUNDER (1972), THE TAKE (1974), CONRACK (1974)and, ultimately, CLAUDINE (1974).In CLAUDINE, we find no stigma nor easy classification as being "Blackploitation", as the story is universal, and could easily have been done as a story about people of any descent, any where, and not just in the 1970\'s USA.That the story was done of a SINGLE mother, Claudine (Dianne Carroll), struggling to keep a family together after "....two marriages and two almost marriages.", is a far cry from a shoot-em-up Harlem Style. The problems that plague the everyday citizens of our nation are confronted and examined under the ol\' sociological microscope.But we also consider Claudine\'s psychological and physical needs as a female. For "Woman Needs Man and Man Must Have His MATE",***and we do concede this point. (That\'s S-E-X that we\'re talking about, Schultz!) Claudine meets up with a very masculine, broad shouldered, athletic type in Private Scavanger Garbage Man, Ruppert B. Marshall (James Earl Jones) and they go on a date.The Great Welfare State intervenes with the Couple as Claudine\'s Welfare Case Worker, Miss Tayback (Elisa Loti), comes snooping around to see just who is this unattached Male, who is suddenly paying so much attention to Claudine\'s family.After a humiliating experience with the Welfare Bureau\'s auditing and "deducting" binge, which would be the norm for the family, the two decide to get married with or without the blessing of Big Brother.Meanwhile, Claudine\'s elder son has gotten involved with some big talking but little doing Black Activist group. But, with Ruppert\'s help, he and they all come through it A.O.K.It ends on a Happy, Upbeat and Hopeful note. We know that it may not be exactly "...Happily Ever After!", but rather the\'ll make it all together! If there is a single criticism that we must state it is that sometimes in a movie like this, a misconception is spread to a large portion of Urban Blacks. And that is, the apparent implied myth that all Whites are wealthy, having none of their kind ever in need of a helping hand, out of work or suffering any disabilities.Well, folks, it just ain\'t true! NOTE: * At one point, Jim Brown\'s career was a real hit as a rugged actioner. He was even being tauted as "...The Black John Wayne." NOTE: ** The idea of producing films with All-Black Casts, filmed for All-Black consumption was not a new idea. In the 1920\'s, \'30\'s and \'40\'s, we saw productions from people like Noble Johnson, Spencer Williams, Jr. and Rex Ingram.NOTE: *** That\'s "As Time Goes By", you know, Schultz, it\'s from CASABLANCA (Warner Brothers, 1942).',
290
+ "There absolutely was voter fraud. There's voter fraud in every election. However, they are generally isolated incidents and I don't think there has been any credible evidence presented that indicates any wide-scale systemic voter fraud happened in 2020. \n\nI would like a federal commission started that investigates and looks for systemic voter and election fraud. Especially one that would be empowered to look into cases of disenfranchisement and voter suppression as well. Everyone that is legally allowed to vote should be able to easily and securely register and cast their vote.",
291
+ ]
292
+ embeddings = model.encode(sentences)
293
+ print(embeddings.shape)
294
+ # [3, 1024]
295
+
296
+ # Get the similarity scores for the embeddings
297
+ similarities = model.similarity(embeddings, embeddings)
298
+ print(similarities)
299
+ # tensor([[1.0000, 0.1908, 0.3587],
300
+ # [0.1908, 1.0000, 0.3531],
301
+ # [0.3587, 0.3531, 1.0000]])
302
+ ```
303
+
304
+ <!--
305
+ ### Direct Usage (Transformers)
306
+
307
+ <details><summary>Click to see the direct usage in Transformers</summary>
308
+
309
+ </details>
310
+ -->
311
+
312
+ <!--
313
+ ### Downstream Usage (Sentence Transformers)
314
+
315
+ You can finetune this model on your own dataset.
316
+
317
+ <details><summary>Click to expand</summary>
318
+
319
+ </details>
320
+ -->
321
+
322
+ <!--
323
+ ### Out-of-Scope Use
324
+
325
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
326
+ -->
327
+
328
+ ## Evaluation
329
+
330
+ ### Metrics
331
+
332
+ #### Semantic Similarity
333
+
334
+ * Dataset: `similarity`
335
+ * Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
336
+
337
+ | Metric | Value |
338
+ |:--------------------|:-----------|
339
+ | pearson_cosine | 0.4106 |
340
+ | **spearman_cosine** | **0.4261** |
341
+
342
+ <!--
343
+ ## Bias, Risks and Limitations
344
+
345
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
346
+ -->
347
+
348
+ <!--
349
+ ### Recommendations
350
+
351
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
352
+ -->
353
+
354
+ ## Training Details
355
+
356
+ ### Training Dataset
357
+
358
+ #### Unnamed Dataset
359
+
360
+ * Size: 11,180 training samples
361
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
362
+ * Approximate statistics based on the first 1000 samples:
363
+ | | sentence_0 | sentence_1 | label |
364
+ |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:---------------------------------------------------------------|
365
+ | type | string | string | float |
366
+ | details | <ul><li>min: 4 tokens</li><li>mean: 104.44 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 109.27 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.52</li><li>max: 1.0</li></ul> |
367
+ * Samples:
368
+ | sentence_0 | sentence_1 | label |
369
+ |:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------|
370
+ | <code>The concept that things could be possibly be worse therefore do not strive to improve things is a weak and cowardly mentality. <br>Nobody wants to hear your dumbass shit.<br><br>Edit: This dude dm'd me and had a total emotional meltdown, that's how bad my words hurt this man.</code> | <code>Based Macron needs to snort something off of your girlfriends titis</code> | <code>1.0</code> |
371
+ | <code>Even #foxnews pundit Brit Hume is calling this tweet a lie and should be the reason he loses the next election or is impeached & found guilty by the majority Republican Senate ASAP! #MuellerReport. End of story!</code> | <code>An election like this will hardly ever be more decisive, thats just how these things are. I agree its sad that even someone like Le Pen doesnt break the habit.</code> | <code>0.7071067811865475</code> |
372
+ | <code>*review may contain spoilers*predictable, campy, bad special effects. it has a TV-movie feeling to it. the idea of the UN as being taken over by Satan is an interesting twist to the end of the world according to the bible. the premise is interesting, but its excution falls waaaay short. if you want to convert people to Christianity with a film like this, at least make it a quality one! i was seriously checking my watch while watching this piece of dreck. can't say much else about this film since i saw it over a year ago, and there isn't really much to say about this film other than.....skip it!</code> | <code>wonderful movie with good story great humour (some great one-liners) and a soundtrack to die for.i've seen it 3 times so far.the american audiences are going to love it.</code> | <code>0.3333333333333333</code> |
373
+ * Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
374
+ ```json
375
+ {
376
+ "loss_fct": "torch.nn.modules.loss.MSELoss"
377
+ }
378
+ ```
379
+
380
+ ### Training Hyperparameters
381
+ #### Non-Default Hyperparameters
382
+
383
+ - `eval_strategy`: steps
384
+ - `per_device_train_batch_size`: 32
385
+ - `per_device_eval_batch_size`: 32
386
+ - `fp16`: True
387
+ - `multi_dataset_batch_sampler`: round_robin
388
+
389
+ #### All Hyperparameters
390
+ <details><summary>Click to expand</summary>
391
+
392
+ - `overwrite_output_dir`: False
393
+ - `do_predict`: False
394
+ - `eval_strategy`: steps
395
+ - `prediction_loss_only`: True
396
+ - `per_device_train_batch_size`: 32
397
+ - `per_device_eval_batch_size`: 32
398
+ - `per_gpu_train_batch_size`: None
399
+ - `per_gpu_eval_batch_size`: None
400
+ - `gradient_accumulation_steps`: 1
401
+ - `eval_accumulation_steps`: None
402
+ - `torch_empty_cache_steps`: None
403
+ - `learning_rate`: 5e-05
404
+ - `weight_decay`: 0.0
405
+ - `adam_beta1`: 0.9
406
+ - `adam_beta2`: 0.999
407
+ - `adam_epsilon`: 1e-08
408
+ - `max_grad_norm`: 1
409
+ - `num_train_epochs`: 3
410
+ - `max_steps`: -1
411
+ - `lr_scheduler_type`: linear
412
+ - `lr_scheduler_kwargs`: {}
413
+ - `warmup_ratio`: 0.0
414
+ - `warmup_steps`: 0
415
+ - `log_level`: passive
416
+ - `log_level_replica`: warning
417
+ - `log_on_each_node`: True
418
+ - `logging_nan_inf_filter`: True
419
+ - `save_safetensors`: True
420
+ - `save_on_each_node`: False
421
+ - `save_only_model`: False
422
+ - `restore_callback_states_from_checkpoint`: False
423
+ - `no_cuda`: False
424
+ - `use_cpu`: False
425
+ - `use_mps_device`: False
426
+ - `seed`: 42
427
+ - `data_seed`: None
428
+ - `jit_mode_eval`: False
429
+ - `use_ipex`: False
430
+ - `bf16`: False
431
+ - `fp16`: True
432
+ - `fp16_opt_level`: O1
433
+ - `half_precision_backend`: auto
434
+ - `bf16_full_eval`: False
435
+ - `fp16_full_eval`: False
436
+ - `tf32`: None
437
+ - `local_rank`: 0
438
+ - `ddp_backend`: None
439
+ - `tpu_num_cores`: None
440
+ - `tpu_metrics_debug`: False
441
+ - `debug`: []
442
+ - `dataloader_drop_last`: False
443
+ - `dataloader_num_workers`: 0
444
+ - `dataloader_prefetch_factor`: None
445
+ - `past_index`: -1
446
+ - `disable_tqdm`: False
447
+ - `remove_unused_columns`: True
448
+ - `label_names`: None
449
+ - `load_best_model_at_end`: False
450
+ - `ignore_data_skip`: False
451
+ - `fsdp`: []
452
+ - `fsdp_min_num_params`: 0
453
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
454
+ - `fsdp_transformer_layer_cls_to_wrap`: None
455
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
456
+ - `deepspeed`: None
457
+ - `label_smoothing_factor`: 0.0
458
+ - `optim`: adamw_torch
459
+ - `optim_args`: None
460
+ - `adafactor`: False
461
+ - `group_by_length`: False
462
+ - `length_column_name`: length
463
+ - `ddp_find_unused_parameters`: None
464
+ - `ddp_bucket_cap_mb`: None
465
+ - `ddp_broadcast_buffers`: False
466
+ - `dataloader_pin_memory`: True
467
+ - `dataloader_persistent_workers`: False
468
+ - `skip_memory_metrics`: True
469
+ - `use_legacy_prediction_loop`: False
470
+ - `push_to_hub`: False
471
+ - `resume_from_checkpoint`: None
472
+ - `hub_model_id`: None
473
+ - `hub_strategy`: every_save
474
+ - `hub_private_repo`: None
475
+ - `hub_always_push`: False
476
+ - `hub_revision`: None
477
+ - `gradient_checkpointing`: False
478
+ - `gradient_checkpointing_kwargs`: None
479
+ - `include_inputs_for_metrics`: False
480
+ - `include_for_metrics`: []
481
+ - `eval_do_concat_batches`: True
482
+ - `fp16_backend`: auto
483
+ - `push_to_hub_model_id`: None
484
+ - `push_to_hub_organization`: None
485
+ - `mp_parameters`:
486
+ - `auto_find_batch_size`: False
487
+ - `full_determinism`: False
488
+ - `torchdynamo`: None
489
+ - `ray_scope`: last
490
+ - `ddp_timeout`: 1800
491
+ - `torch_compile`: False
492
+ - `torch_compile_backend`: None
493
+ - `torch_compile_mode`: None
494
+ - `include_tokens_per_second`: False
495
+ - `include_num_input_tokens_seen`: False
496
+ - `neftune_noise_alpha`: None
497
+ - `optim_target_modules`: None
498
+ - `batch_eval_metrics`: False
499
+ - `eval_on_start`: False
500
+ - `use_liger_kernel`: False
501
+ - `liger_kernel_config`: None
502
+ - `eval_use_gather_object`: False
503
+ - `average_tokens_across_devices`: False
504
+ - `prompts`: None
505
+ - `batch_sampler`: batch_sampler
506
+ - `multi_dataset_batch_sampler`: round_robin
507
+ - `router_mapping`: {}
508
+ - `learning_rate_mapping`: {}
509
+
510
+ </details>
511
+
512
+ ### Training Logs
513
+ | Epoch | Step | Training Loss | similarity_spearman_cosine |
514
+ |:------:|:----:|:-------------:|:--------------------------:|
515
+ | 0.0286 | 10 | - | 0.1359 |
516
+ | 0.0571 | 20 | - | 0.1424 |
517
+ | 0.0857 | 30 | - | 0.1525 |
518
+ | 0.1143 | 40 | - | 0.1651 |
519
+ | 0.1429 | 50 | - | 0.1759 |
520
+ | 0.1714 | 60 | - | 0.1846 |
521
+ | 0.2 | 70 | - | 0.1947 |
522
+ | 0.2286 | 80 | - | 0.2056 |
523
+ | 0.2571 | 90 | - | 0.2144 |
524
+ | 0.2857 | 100 | - | 0.2298 |
525
+ | 0.3143 | 110 | - | 0.2409 |
526
+ | 0.3429 | 120 | - | 0.2526 |
527
+ | 0.3714 | 130 | - | 0.2511 |
528
+ | 0.4 | 140 | - | 0.2661 |
529
+ | 0.4286 | 150 | - | 0.2664 |
530
+ | 0.4571 | 160 | - | 0.2572 |
531
+ | 0.4857 | 170 | - | 0.2804 |
532
+ | 0.5143 | 180 | - | 0.2885 |
533
+ | 0.5429 | 190 | - | 0.2885 |
534
+ | 0.5714 | 200 | - | 0.2933 |
535
+ | 0.6 | 210 | - | 0.3037 |
536
+ | 0.6286 | 220 | - | 0.3163 |
537
+ | 0.6571 | 230 | - | 0.3197 |
538
+ | 0.6857 | 240 | - | 0.3275 |
539
+ | 0.7143 | 250 | - | 0.3238 |
540
+ | 0.7429 | 260 | - | 0.3262 |
541
+ | 0.7714 | 270 | - | 0.3295 |
542
+ | 0.8 | 280 | - | 0.3129 |
543
+ | 0.8286 | 290 | - | 0.3491 |
544
+ | 0.8571 | 300 | - | 0.3354 |
545
+ | 0.8857 | 310 | - | 0.3448 |
546
+ | 0.9143 | 320 | - | 0.3581 |
547
+ | 0.9429 | 330 | - | 0.3658 |
548
+ | 0.9714 | 340 | - | 0.3386 |
549
+ | 1.0 | 350 | - | 0.3503 |
550
+ | 1.0286 | 360 | - | 0.3533 |
551
+ | 1.0571 | 370 | - | 0.3604 |
552
+ | 1.0857 | 380 | - | 0.3624 |
553
+ | 1.1143 | 390 | - | 0.3549 |
554
+ | 1.1429 | 400 | - | 0.3594 |
555
+ | 1.1714 | 410 | - | 0.3747 |
556
+ | 1.2 | 420 | - | 0.3465 |
557
+ | 1.2286 | 430 | - | 0.3378 |
558
+ | 1.2571 | 440 | - | 0.3809 |
559
+ | 1.2857 | 450 | - | 0.3856 |
560
+ | 1.3143 | 460 | - | 0.3522 |
561
+ | 1.3429 | 470 | - | 0.3987 |
562
+ | 1.3714 | 480 | - | 0.3847 |
563
+ | 1.4 | 490 | - | 0.3688 |
564
+ | 1.4286 | 500 | 0.1157 | 0.3937 |
565
+ | 1.4571 | 510 | - | 0.3857 |
566
+ | 1.4857 | 520 | - | 0.4039 |
567
+ | 1.5143 | 530 | - | 0.3913 |
568
+ | 1.5429 | 540 | - | 0.3900 |
569
+ | 1.5714 | 550 | - | 0.3497 |
570
+ | 1.6 | 560 | - | 0.3613 |
571
+ | 1.6286 | 570 | - | 0.4067 |
572
+ | 1.6571 | 580 | - | 0.4016 |
573
+ | 1.6857 | 590 | - | 0.3954 |
574
+ | 1.7143 | 600 | - | 0.3947 |
575
+ | 1.7429 | 610 | - | 0.3864 |
576
+ | 1.7714 | 620 | - | 0.4194 |
577
+ | 1.8 | 630 | - | 0.3985 |
578
+ | 1.8286 | 640 | - | 0.4003 |
579
+ | 1.8571 | 650 | - | 0.4061 |
580
+ | 1.8857 | 660 | - | 0.4074 |
581
+ | 1.9143 | 670 | - | 0.4004 |
582
+ | 1.9429 | 680 | - | 0.4022 |
583
+ | 1.9714 | 690 | - | 0.4056 |
584
+ | 2.0 | 700 | - | 0.3991 |
585
+ | 2.0286 | 710 | - | 0.3944 |
586
+ | 2.0571 | 720 | - | 0.3952 |
587
+ | 2.0857 | 730 | - | 0.4014 |
588
+ | 2.1143 | 740 | - | 0.3846 |
589
+ | 2.1429 | 750 | - | 0.3719 |
590
+ | 2.1714 | 760 | - | 0.4073 |
591
+ | 2.2 | 770 | - | 0.3828 |
592
+ | 2.2286 | 780 | - | 0.3858 |
593
+ | 2.2571 | 790 | - | 0.4114 |
594
+ | 2.2857 | 800 | - | 0.3930 |
595
+ | 2.3143 | 810 | - | 0.3845 |
596
+ | 2.3429 | 820 | - | 0.4053 |
597
+ | 2.3714 | 830 | - | 0.3582 |
598
+ | 2.4 | 840 | - | 0.3848 |
599
+ | 2.4286 | 850 | - | 0.4139 |
600
+ | 2.4571 | 860 | - | 0.3609 |
601
+ | 2.4857 | 870 | - | 0.4122 |
602
+ | 2.5143 | 880 | - | 0.4101 |
603
+ | 2.5429 | 890 | - | 0.4261 |
604
+
605
+
606
+ ### Framework Versions
607
+ - Python: 3.11.9
608
+ - Sentence Transformers: 5.1.0
609
+ - Transformers: 4.53.3
610
+ - PyTorch: 2.5.1
611
+ - Accelerate: 1.10.0
612
+ - Datasets: 2.14.4
613
+ - Tokenizers: 0.21.0
614
+
615
+ ## Citation
616
+
617
+ ### BibTeX
618
+
619
+ #### Sentence Transformers
620
+ ```bibtex
621
+ @inproceedings{reimers-2019-sentence-bert,
622
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
623
+ author = "Reimers, Nils and Gurevych, Iryna",
624
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
625
+ month = "11",
626
+ year = "2019",
627
+ publisher = "Association for Computational Linguistics",
628
+ url = "https://arxiv.org/abs/1908.10084",
629
+ }
630
+ ```
631
+
632
+ <!--
633
+ ## Glossary
634
+
635
+ *Clearly define terms in order to be accessible across audiences.*
636
+ -->
637
+
638
+ <!--
639
+ ## Model Card Authors
640
+
641
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
642
+ -->
643
+
644
+ <!--
645
+ ## Model Card Contact
646
+
647
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
648
+ -->
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_eps": 1e-12,
13
+ "max_position_embeddings": 512,
14
+ "model_type": "bert",
15
+ "num_attention_heads": 16,
16
+ "num_hidden_layers": 24,
17
+ "pad_token_id": 0,
18
+ "position_embedding_type": "absolute",
19
+ "torch_dtype": "float32",
20
+ "transformers_version": "4.53.3",
21
+ "type_vocab_size": 2,
22
+ "use_cache": true,
23
+ "vocab_size": 30522
24
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "SentenceTransformer",
3
+ "__version__": {
4
+ "sentence_transformers": "5.1.0",
5
+ "transformers": "4.53.3",
6
+ "pytorch": "2.5.1"
7
+ },
8
+ "prompts": {
9
+ "query": "",
10
+ "document": ""
11
+ },
12
+ "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
14
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faa9fd1420b2d60582e5e578fe49d3a281c0b0575ad2667ab4693179f64142a2
3
+ size 1340612432
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff