Add new SentenceTransformer model
Browse files- 1_Pooling/config.json +10 -0
- README.md +1016 -0
- config.json +25 -0
- config_sentence_transformers.json +10 -0
- model.safetensors +3 -0
- modules.json +20 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +63 -0
- vocab.txt +0 -0
1_Pooling/config.json
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"word_embedding_dimension": 768,
|
| 3 |
+
"pooling_mode_cls_token": false,
|
| 4 |
+
"pooling_mode_mean_tokens": true,
|
| 5 |
+
"pooling_mode_max_tokens": false,
|
| 6 |
+
"pooling_mode_mean_sqrt_len_tokens": false,
|
| 7 |
+
"pooling_mode_weightedmean_tokens": false,
|
| 8 |
+
"pooling_mode_lasttoken": false,
|
| 9 |
+
"include_prompt": true
|
| 10 |
+
}
|
README.md
ADDED
|
@@ -0,0 +1,1016 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- sentence-transformers
|
| 4 |
+
- sentence-similarity
|
| 5 |
+
- feature-extraction
|
| 6 |
+
- generated_from_trainer
|
| 7 |
+
- dataset_size:113450
|
| 8 |
+
- loss:ContrastiveTensionLoss
|
| 9 |
+
base_model: sucharush/e5_stem_finetuned
|
| 10 |
+
widget:
|
| 11 |
+
- source_sentence: 'Question: what channel is nj news 12 on fios, Answer: News 12
|
| 12 |
+
Networks The networks are carried by Altice''s Optimum systems throughout the
|
| 13 |
+
region (except for certain systems in Dutchess, Orange, and Putnam counties in
|
| 14 |
+
New York, and Litchfield Count in Connecticut); these systems were originally
|
| 15 |
+
owned by Dolan-owned Cablevision before being spun off to Altice. Since the sale,
|
| 16 |
+
News 12 networks are also available on most Charter Communications, Comcast, and
|
| 17 |
+
Service Electric systems in the Tri-State area (in the case of Charter, this is
|
| 18 |
+
part of a cross-provider agreement that also allows Charter''s 24-hour local news
|
| 19 |
+
channel, NY1, to be seen on Altice providers). However, none of the channels are
|
| 20 |
+
available on Verizon FiOS (which offers a competing FiOS 1 service) or satellite
|
| 21 |
+
providers DirecTV or Dish Network; this is referenced in the slogan used by the
|
| 22 |
+
networks, "Only on cable. Never on FiOS. Never on satellite." Although the seven
|
| 23 |
+
channels have individual and unique news operations, they share the same mission,
|
| 24 |
+
graphics, and music packages.'
|
| 25 |
+
sentences:
|
| 26 |
+
- 'Question: who won the hockey game between usa and russia, Answer: Miracle on
|
| 27 |
+
Ice For the first game in the medal round, the United States played the Soviets.
|
| 28 |
+
Finishing the first period tied at 2–2, and the Soviets leading 3–2 following
|
| 29 |
+
the second, the U.S. team scored two more goals to take their first lead during
|
| 30 |
+
the third and final period, winning the game 4–3 in a major upset against the
|
| 31 |
+
Cold War rival. Following the game, the U.S. went on to clinch the gold medal
|
| 32 |
+
by beating Finland in their final match of the tournament. Likewise, the Soviet
|
| 33 |
+
Union took the silver medal by beating Sweden.'
|
| 34 |
+
- 'Question: when did the vietnam war start for the us, Answer: Vietnam War Beginning
|
| 35 |
+
in 1950, American military advisors arrived in what was then French Indochina.[86][A
|
| 36 |
+
3] Most of the funding for the French war effort was provided by the U.S.[87]
|
| 37 |
+
The Việt Cộng, also known as Front national de libération du Sud-Viêt Nam or FNL
|
| 38 |
+
(the National Liberation Front), a South Vietnamese communist common front aided
|
| 39 |
+
by the North, fought a guerrilla war against anti-communist forces in the region,
|
| 40 |
+
while the People''s Army of Vietnam, also known as the North Vietnamese Army (NVA),
|
| 41 |
+
engaged in more conventional warfare, and had launched armed struggles from 1959
|
| 42 |
+
onward. U.S. involvement escalated in 1960 under President John F. Kennedy, with
|
| 43 |
+
troop levels gradually surging under the MAAG program from just under a thousand
|
| 44 |
+
in 1959 to 16,000 in 1963.[88][89]'
|
| 45 |
+
- 'Question: how many medals did gb win at the winter olympics, Answer: Great Britain
|
| 46 |
+
at the 2018 Winter Olympics Great Britain competed at the 2018 Winter Olympics
|
| 47 |
+
in Pyeongchang, South Korea, from 9 to 25 February 2018, with 58 competitors in
|
| 48 |
+
11 sports. They won five medals in total, one gold and four bronze, ranking 19th
|
| 49 |
+
in the medal table.'
|
| 50 |
+
- source_sentence: 'Question: when did wet n wild close las vegas, Answer: Wet ''n
|
| 51 |
+
Wild (Las Vegas) The park was acquired by Universal Studios in October 1998, Ogden
|
| 52 |
+
Corp in March 1999, Alfa SmartParks in March 2000, and Palace Entertainment on
|
| 53 |
+
July 10, 2002.[8] The park closed in 2004.[9][10][8]'
|
| 54 |
+
sentences:
|
| 55 |
+
- 'Question: how many times has north america hosted the olympics, Answer: List
|
| 56 |
+
of Olympic Games host cities The Games have primarily been hosted in the continents
|
| 57 |
+
of Europe (32 editions) and North America (12 editions); seven Games have been
|
| 58 |
+
hosted in Asia and two have been hosted in Oceania. In 2010, Singapore became
|
| 59 |
+
Southeast Asia''s first Olympic host city for the inaugural Summer Youth Olympics,
|
| 60 |
+
while Rio de Janeiro became South America''s first Olympic host city with the
|
| 61 |
+
2016 Summer Olympics, followed by Buenos Aires with the 2018 Summer Youth Olympics.
|
| 62 |
+
The 2022 Summer Youth Olympics in Dakar will become the first-ever Games to be
|
| 63 |
+
held on the African continent. Other major geographic regions which have never
|
| 64 |
+
hosted the Olympics include the Middle East, the Indian subcontinent, Central
|
| 65 |
+
America and the Caribbean.'
|
| 66 |
+
- '[{''from'': ''system'', ''value'': ''You are an AI assistant. You will be given
|
| 67 |
+
a task. You must generate a detailed and long answer.''}, {''from'': ''human'',
|
| 68 |
+
''value'': ''Answer this question: where does ground pepper come from?\nAnswer:''},
|
| 69 |
+
{''from'': ''gpt'', ''value'': ''Ground pepper, a widely used and popular spice
|
| 70 |
+
across the globe, comes from the dried berries of the Piper nigrum plant. Piper
|
| 71 |
+
nigrum is a perennial flowering vine, native to Southern India and Sri Lanka,
|
| 72 |
+
but now cultivated in various tropical regions such as Vietnam, Indonesia, Malaysia,
|
| 73 |
+
Brazil, and Madagascar. The plant bears small, green, unripe berries that turn
|
| 74 |
+
red as they ripen. The resulting ground pepper differs in color (black, green,
|
| 75 |
+
or white) depending on the time the berries are harvested and the processing method
|
| 76 |
+
employed.\n\nBlack ground pepper, which is the most common variety, is derived
|
| 77 |
+
from the unripe green berries of the Piper nigrum plant. These berries are harvested
|
| 78 |
+
and then blanched briefly in boiling water, a process that activates the enzyme
|
| 79 |
+
responsible for the oxidation of the outer skin, promoting its dark coloration.
|
| 80 |
+
The berries are then spread out in the sun or other drying methods to dry over
|
| 81 |
+
several days. The outer skin wrinkles, and the remaining seed dries out, developing
|
| 82 |
+
its characteristic black color and distinct pungent aroma. Once fully dried, the
|
| 83 |
+
berries, known as black peppercorns, are ground to produce black ground pepper.\n\nGreen
|
| 84 |
+
ground pepper is produced from fresh green berries that are harvested before they
|
| 85 |
+
ripen. Rather than being dried, these unripe berries are treated to preserve their
|
| 86 |
+
green color, either through freeze-drying or by soaking them in a saltwater or
|
| 87 |
+
vinegar solution. This process retains the green hue and a milder, herbaceous
|
| 88 |
+
flavor profile compared to black pepper. After preservation, green peppercorns
|
| 89 |
+
are ground into green ground pepper.\n\nWhite ground pepper, on the other hand,
|
| 90 |
+
comes from ripe red berries. The red berries are harvested at full maturity and
|
| 91 |
+
soaked in water for an extended period, a process that softens and loosens their
|
| 92 |
+
outer skin. The outer skin is then removed, leaving behind the inner seed, which
|
| 93 |
+
is pale in color. These seeds are then dried and ground to create white ground
|
| 94 |
+
pepper. This variety has a milder and slightly less pungent flavor compared to
|
| 95 |
+
black pepper.\n\nAll three varieties (black, green, and white) of ground pepper
|
| 96 |
+
are used in various cuisines worldwide, each imparting unique flavors to culinary
|
| 97 |
+
dishes. Today, ground pepper is not only a primary seasoning ingredient in countless
|
| 98 |
+
recipes but also a staple condiment on restaurant tables and in home kitchens.
|
| 99 |
+
As a testament to its universal appeal, ground pepper remains one of the most
|
| 100 |
+
traded commodities in the global spice market.''}]'
|
| 101 |
+
- '[{''from'': ''human'', ''value'': ''You are working on a project that involves
|
| 102 |
+
creating a function in Python to calculate the roots of a quadratic equation.
|
| 103 |
+
The function should take the coefficients of the quadratic equation as input and
|
| 104 |
+
return the roots as output. The quadratic equation is of the form ax^2 + bx +
|
| 105 |
+
c = 0.\n\nWrite a Python function to calculate the roots of a quadratic equation.''},
|
| 106 |
+
{''from'': ''gpt'', ''value'': ''```python\nimport cmath\n\ndef calculate_quadratic_roots(a,
|
| 107 |
+
b, c):\n # Calculate the discriminant\n discriminant = (b**2) - (4*a*c)\n\n #
|
| 108 |
+
Calculate the two roots using the quadratic formula\n root1 = (-b - cmath.sqrt(discriminant))
|
| 109 |
+
/ (2 * a)\n root2 = (-b + cmath.sqrt(discriminant)) / (2 * a)\n\n return
|
| 110 |
+
root1, root2\n```\nThis Python function takes the coefficients a, b, and c as
|
| 111 |
+
input and calculates the roots of the quadratic equation using the quadratic formula.
|
| 112 |
+
The cmath module is used to handle complex roots if the discriminant is negative.
|
| 113 |
+
The function returns the two roots as output.''}]'
|
| 114 |
+
- source_sentence: '[{''from'': ''human'', ''value'': ''How can we define a function
|
| 115 |
+
by a formula in a rigorous way, ensuring that the function is well-defined and
|
| 116 |
+
its domain is properly specified?''}, {''from'': ''gpt'', ''value'': "To define
|
| 117 |
+
a function by a formula in a rigorous way, we need to consider both the formula
|
| 118 |
+
itself and the domain of the function. Here''s a step-by-step approach:\n\n1.
|
| 119 |
+
**Formula Definition**:\n - Define a set of expressions for rational functions
|
| 120 |
+
using a fixed variable, an alphabet for real constants, and a function assigning
|
| 121 |
+
values to constants.\n - Each expression is a string in a specific alphabet,
|
| 122 |
+
including the variable, constants, and operators.\n\n2. **Natural Domain and Interpretation**:\n -
|
| 123 |
+
Simultaneously define two functions:\n - The natural domain function $D$ associates
|
| 124 |
+
each expression with a subset of real numbers, representing the domain where the
|
| 125 |
+
expression is well-defined.\n - The interpretation function $I$ associates
|
| 126 |
+
each expression with a function from its natural domain to real numbers, representing
|
| 127 |
+
the value of the expression at each point in the domain.\n\n3. **Structural Induction**:\n -
|
| 128 |
+
Define $D$ and $I$ inductively based on the structure of expressions:\n -
|
| 129 |
+
For the variable, the natural domain is all real numbers, and the interpretation
|
| 130 |
+
is the identity function.\n - For constants, the natural domain is all real
|
| 131 |
+
numbers, and the interpretation is the constant value.\n - For sums, products,
|
| 132 |
+
and quotients, the natural domain and interpretation are defined based on the
|
| 133 |
+
natural domains and interpretations of the constituent expressions.\n\n4. **Syntactic
|
| 134 |
+
Equality and Sugar**:\n - Introduce syntactic equality to simplify expressions
|
| 135 |
+
and make them more readable.\n - Define rules for syntactic sugar, such as removing
|
| 136 |
+
unnecessary parentheses and using more compact notation for powers and fractions.\n\n5.
|
| 137 |
+
**Equivalence of Fractions and Irreducible Fractions**:\n - Define equivalence
|
| 138 |
+
of fractions and irreducible fractions within the context of expressions.\n\nBy
|
| 139 |
+
following this approach, we can rigorously define a function by a formula, ensuring
|
| 140 |
+
that the function is well-defined and its domain is properly specified. This approach
|
| 141 |
+
provides a solid foundation for understanding and working with functions expressed
|
| 142 |
+
as formulas in mathematics."}]'
|
| 143 |
+
sentences:
|
| 144 |
+
- '[{''from'': ''human'', ''value'': ''How can we define a function by a formula
|
| 145 |
+
in a rigorous way, ensuring that the function is well-defined and its domain is
|
| 146 |
+
properly specified?''}, {''from'': ''gpt'', ''value'': "To define a function by
|
| 147 |
+
a formula in a rigorous way, we need to consider both the formula itself and the
|
| 148 |
+
domain of the function. Here''s a step-by-step approach:\n\n1. **Formula Definition**:\n -
|
| 149 |
+
Define a set of expressions for rational functions using a fixed variable, an
|
| 150 |
+
alphabet for real constants, and a function assigning values to constants.\n -
|
| 151 |
+
Each expression is a string in a specific alphabet, including the variable, constants,
|
| 152 |
+
and operators.\n\n2. **Natural Domain and Interpretation**:\n - Simultaneously
|
| 153 |
+
define two functions:\n - The natural domain function $D$ associates each
|
| 154 |
+
expression with a subset of real numbers, representing the domain where the expression
|
| 155 |
+
is well-defined.\n - The interpretation function $I$ associates each expression
|
| 156 |
+
with a function from its natural domain to real numbers, representing the value
|
| 157 |
+
of the expression at each point in the domain.\n\n3. **Structural Induction**:\n -
|
| 158 |
+
Define $D$ and $I$ inductively based on the structure of expressions:\n -
|
| 159 |
+
For the variable, the natural domain is all real numbers, and the interpretation
|
| 160 |
+
is the identity function.\n - For constants, the natural domain is all real
|
| 161 |
+
numbers, and the interpretation is the constant value.\n - For sums, products,
|
| 162 |
+
and quotients, the natural domain and interpretation are defined based on the
|
| 163 |
+
natural domains and interpretations of the constituent expressions.\n\n4. **Syntactic
|
| 164 |
+
Equality and Sugar**:\n - Introduce syntactic equality to simplify expressions
|
| 165 |
+
and make them more readable.\n - Define rules for syntactic sugar, such as removing
|
| 166 |
+
unnecessary parentheses and using more compact notation for powers and fractions.\n\n5.
|
| 167 |
+
**Equivalence of Fractions and Irreducible Fractions**:\n - Define equivalence
|
| 168 |
+
of fractions and irreducible fractions within the context of expressions.\n\nBy
|
| 169 |
+
following this approach, we can rigorously define a function by a formula, ensuring
|
| 170 |
+
that the function is well-defined and its domain is properly specified. This approach
|
| 171 |
+
provides a solid foundation for understanding and working with functions expressed
|
| 172 |
+
as formulas in mathematics."}]'
|
| 173 |
+
- '[{''from'': ''system'', ''value'': ''You are a helpful assistant, with no access
|
| 174 |
+
to external functions.''}, {''from'': ''human'', ''value'': ''Group the shapes
|
| 175 |
+
given in the table below into 2D and 3D shapes, based on their respective dimensions.
|
| 176 |
+
Additionally, provide the surface area and volume of each shape in the table.
|
| 177 |
+
The list of shapes includes triangle, sphere, cylinder, and square.''}, {''from'':
|
| 178 |
+
''gpt'', ''value'': ''2D Shapes: Triangle, Square\n- Triangle: Surface Area =
|
| 179 |
+
(base x height) / 2\n- Square: Surface Area = 4 x side length\n3D Shapes: Sphere,
|
| 180 |
+
Cylinder\n- Sphere: Surface Area = 4πr^2, Volume = (4/3)πr^3\n- Cylinder: Surface
|
| 181 |
+
Area = 2πrh + 2πr^2, Volume = πr^2h\nNote: r is the radius of the sphere or cylinder,
|
| 182 |
+
h is the height of the cylinder.\nBased on their dimensions, the 2D shapes are
|
| 183 |
+
the triangle and the square, while the 3D shapes are the sphere and the cylinder.\n|
|
| 184 |
+
Shape | Type | Surface Area Formula | Volume Formula |\n|----------|----------|--------------------------------|-------------------------------|\n|
|
| 185 |
+
Triangle | 2D Shape | (base x height) / 2 | N/A |\n| Square | 2D Shape | 4 x side
|
| 186 |
+
length | N/A |\n| Sphere | 3D Shape | 4πr^2 | (4/3)πr^3 |\n| Cylinder | 3D Shape
|
| 187 |
+
| 2πrh + 2πr^2 | πr^2h |''}, {''from'': ''human'', ''value'': ''Can you explain
|
| 188 |
+
the difference between a hurricane, a typhoon, and a cyclone?''}, {''from'': ''gpt'',
|
| 189 |
+
''value'': ''Yes, I can explain the difference between a hurricane, a typhoon,
|
| 190 |
+
and a cyclone. These are all terms for tropical cyclones, which are rotating storms
|
| 191 |
+
that form in warm ocean waters near the equator. The main difference between these
|
| 192 |
+
terms is the location where the storm forms. Hurricanes form in the Atlantic Ocean
|
| 193 |
+
or northeastern Pacific Ocean, while typhoons form in the northwestern Pacific
|
| 194 |
+
Ocean. Cyclones, on the other hand, is a term used to describe tropical storms
|
| 195 |
+
that occur in the South Pacific and Indian Oceans. Another key difference is the
|
| 196 |
+
direction of rotation. In the Northern Hemisphere, hurricanes and typhoons rotate
|
| 197 |
+
counterclockwise, while in the Southern Hemisphere, they rotate clockwise. Cyclones,
|
| 198 |
+
however, rotate clockwise in the Southern Hemisphere and counterclockwise in the
|
| 199 |
+
Northern Hemisphere.''}, {''from'': ''human'', ''value'': ''How does the process
|
| 200 |
+
of photosynthesis work?''}, {''from'': ''gpt'', ''value'': ''Photosynthesis is
|
| 201 |
+
the process by which plants convert sunlight, carbon dioxide, and water into glucose
|
| 202 |
+
(sugar) and oxygen. This process takes place in the chloroplasts of plant cells,
|
| 203 |
+
which contain a green pigment called chlorophyll. During photosynthesis, light
|
| 204 |
+
energy is absorbed by chlorophyll, which then uses that energy to split water
|
| 205 |
+
molecules into hydrogen ions and oxygen gas. The hydrogen ions are combined with
|
| 206 |
+
carbon dioxide from the air to create glucose, which the plant uses as food. Oxygen
|
| 207 |
+
gas is released into the atmosphere as a byproduct of this process. The overall
|
| 208 |
+
chemical equation for photosynthesis is: 6CO2 + 6H2O + Light Energy → C6H12O6
|
| 209 |
+
+ 6O2.''}, {''from'': ''human'', ''value'': ''What is the difference between nature
|
| 210 |
+
and nurture when it comes to human characteristics?''}, {''from'': ''gpt'', ''value'':
|
| 211 |
+
''Nature refers to the genetic or inherited traits that we receive from our parents
|
| 212 |
+
and ancestors. These traits include things like eye color, height, and certain
|
| 213 |
+
personality traits. Nurture refers to the environmental factors that influence
|
| 214 |
+
our development, such as our upbringing, education, and life experiences. While
|
| 215 |
+
nature plays a role in shaping who we are, research has shown that nurture can
|
| 216 |
+
have a significant impact on human characteristics as well. For example, studies
|
| 217 |
+
of identical twins who were separated at birth and raised in different environments
|
| 218 |
+
have shown that their personalities and behaviors can differ quite significantly,
|
| 219 |
+
despite having the same genetic makeup.''}, {''from'': ''human'', ''value'': ''Why
|
| 220 |
+
is the sky blue during the day and red/orange during sunrise and sunset?''}, {''from'':
|
| 221 |
+
''gpt'', ''value'': "The sky appears blue during the day and red/orange during
|
| 222 |
+
sunrise and sunset due to the way that light is scattered by the Earth''s atmosphere.
|
| 223 |
+
During the day, sunlight consists of all the colors of the spectrum, including
|
| 224 |
+
blue, green, and red. When this light enters the Earth''s atmosphere, it collides
|
| 225 |
+
with molecules of gas and other particles, causing it to scatter in all directions.
|
| 226 |
+
Blue light is scattered more than other colors because it travels in smaller,
|
| 227 |
+
shorter waves. This is why the sky appears blue to us. During sunrise and sunset,
|
| 228 |
+
however, the sun''s rays have to travel through more of the Earth''s atmosphere
|
| 229 |
+
to reach us, which causes the blue light to be scattered even more. This leaves
|
| 230 |
+
only the longer-wavelength red, orange, and yellow light to reach our eyes, giving
|
| 231 |
+
the sky its characteristic red or orange hue."}]'
|
| 232 |
+
- '[{''from'': ''human'', ''value'': ''How can I create a program in Java to find
|
| 233 |
+
the greatest common divisor (GCD) of two integers? None''}, {''from'': ''gpt'',
|
| 234 |
+
''value'': ''1. The task is to create a Java program to find the greatest common
|
| 235 |
+
divisor (GCD) of two integers.\n2. The GCD of two integers is the largest number
|
| 236 |
+
that divides both of them without leaving a remainder.\n3. To solve this, we can
|
| 237 |
+
use a simple iterative method where we check each number up to the minimum of
|
| 238 |
+
the two given numbers to see if it divides both numbers.\n4. We start by declaring
|
| 239 |
+
two integer variables `a` and `b` with the values of the two numbers we want to
|
| 240 |
+
find the GCD of.\n5. We then initialize a variable `gcd` with 1. This variable
|
| 241 |
+
will hold the greatest common divisor.\n6. Next, we use a `for` loop to iterate
|
| 242 |
+
from 1 to the smaller of the two numbers (`a` and `b`).\n7. Inside the loop, we
|
| 243 |
+
check if both `a` and `b` are divisible by the current value of `i` using the
|
| 244 |
+
modulus operator (`%`). The modulus operator returns the remainder of the division
|
| 245 |
+
of two numbers.\n8. If `a` and `b` are both divisible by `i` (i.e., `a % i ==
|
| 246 |
+
0` and `b % i == 0`), this means `i` is a common divisor of `a` and `b`.\n9. In
|
| 247 |
+
this case, we update `gcd` with the current value of `i`. Since `i` is increasing
|
| 248 |
+
with each iteration of the loop, `gcd` will always hold the greatest common divisor
|
| 249 |
+
found so far.\n10. After the loop finishes, `gcd` will hold the greatest common
|
| 250 |
+
divisor of `a` and `b`.\n11. Finally, we print the value of `gcd` using `System.out.printf`.
|
| 251 |
+
This will display the GCD of the two numbers.\n12. The output of the program will
|
| 252 |
+
be the GCD of the two input numbers.\n\nYou can use the following code as a solution:\n\n```java\npublic
|
| 253 |
+
class Main {\n public static void main(String[] args) {\n int a = 24;\n int
|
| 254 |
+
b = 18;\n\n int gcd = 1;\n\n for(int i = 1; i <= a && i <= b; i++)
|
| 255 |
+
{\n if(a % i == 0 && b % i == 0)\n gcd = i;\n }\n\n System.out.printf("GCD
|
| 256 |
+
of %d and %d is: %d", a, b, gcd);\n }\n}\n```\n\nIn this code, we start by
|
| 257 |
+
declaring two integer variables `a` and `b` with values 24 and 18 respectively.
|
| 258 |
+
We then initialize a variable `gcd` with 1, which will hold the greatest common
|
| 259 |
+
divisor.\n\nNext, we use a `for` loop to iterate from 1 to the smaller of the
|
| 260 |
+
two numbers (`a` and `b`). Inside the loop, we check if both `a` and `b` are divisible
|
| 261 |
+
by the current value of `i`. If they are, we update `gcd` with the current value
|
| 262 |
+
of `i`.\n\nFinally, we use `System.out.printf` to display the resulting GCD value.
|
| 263 |
+
In this example, the output will be "GCD of 24 and 18 is: 6", indicating that
|
| 264 |
+
the greatest common divisor of 24 and 18 is 6.''}]'
|
| 265 |
+
- source_sentence: 'Question: how old do you have to be to buy scratchers in arizona,
|
| 266 |
+
Answer: Arizona Lottery Arizona requires lottery players to be at least 21; the
|
| 267 |
+
minimum age was 18 until June 1, 2003.'
|
| 268 |
+
sentences:
|
| 269 |
+
- '[{''from'': ''human'', ''value'': ''Provide a detailed guide on how to crop an
|
| 270 |
+
image using Python including an explanation of the Pillow library, the code needed
|
| 271 |
+
for cropping, and an explanation of each line of code. None''}, {''from'': ''gpt'',
|
| 272 |
+
''value'': "1. Understand the instruction: Provide a detailed guide on cropping
|
| 273 |
+
an image using Python, including an explanation of the Pillow library, the code
|
| 274 |
+
needed for cropping, and an explanation of each line of code.\n\n2. Identify the
|
| 275 |
+
main components: Pillow library, cropping an image, code explanation.\n\n3. Start
|
| 276 |
+
with an introduction: Explain the objective of the article and introduce Python''s
|
| 277 |
+
digital image processing capability.\n\n4. Introduce the Pillow library: Explain
|
| 278 |
+
that Python uses the Pillow library for handling images, which is an updated version
|
| 279 |
+
of the PIL module.\n\n5. Explain the process of cropping an image: Describe how
|
| 280 |
+
cropping involves selecting a rectangular portion from an image and creating a
|
| 281 |
+
new image with just the selected portion.\n\n6. Provide the code snippet for cropping
|
| 282 |
+
an image: Include the code snippet and explain each line of code.\n\n7. Line 1:
|
| 283 |
+
Explain the ''crop()'' function, its input parameters, and how it returns a new
|
| 284 |
+
image with the specified rectangular box.\n\n8. Line 2: Explain how the new Image
|
| 285 |
+
object for the cropped image is stored in the ''croppedIm'' variable.\n\n9. Line
|
| 286 |
+
3: Explain how the ''save()'' function is called on ''croppedIm'' to save the
|
| 287 |
+
cropped image as ''cropped.png''.\n\n10. Mention the coordinate system used by
|
| 288 |
+
Pillow: Explain that Pillow uses a right-handed coordinate system with (0,0) in
|
| 289 |
+
the upper left corner.\n\n11. Conclude the chain of thought: Summarize the process
|
| 290 |
+
of cropping an image using Python and the Pillow library.\n\nTitle: Code Snippet
|
| 291 |
+
Guide on Image Cropping Using Python\n\nObjective: This article explains the use
|
| 292 |
+
of Python''s digital image processing capability to perform a common task - image
|
| 293 |
+
cropping. Python''s powerful and rich libraries aid in manipulating images with
|
| 294 |
+
just a few lines of code. Here, we will introduce how to crop an image using Python.\n\n
|
| 295 |
+
Python''s Pillow Library\n\nPrimarily, Python uses a library named Pillow for
|
| 296 |
+
handling images, which is an updated version of an older module called PIL (Python
|
| 297 |
+
Imaging Library). It can be used for opening, manipulating, and saving different
|
| 298 |
+
image file formats.\n\n Cropping an Image with Python \n\nThe process of cropping
|
| 299 |
+
an image using Python involves selecting a desired rectangular portion from a
|
| 300 |
+
given image and creating a new image with just the selected portion.\n\n Code
|
| 301 |
+
Explanation:\n\nHere is the code snippet to perform image cropping.\n\n```python\ncroppedIm
|
| 302 |
+
= catIm.crop((335, 345, 565, 560))\ncroppedIm.save(''cropped.png'')\n```\n\nThe
|
| 303 |
+
above code snippet does the following:\n\n1. `catIm.crop((335, 345, 565, 560))`:
|
| 304 |
+
The ''crop()'' function is called on the Image object ''catIm''. This function
|
| 305 |
+
requires a tuple of four integer coordinates that represent a rectangular box
|
| 306 |
+
in the original image. The coordinates are pixel positions for left, upper, right,
|
| 307 |
+
and lower position (in that order). The ''crop()'' function then returns a new
|
| 308 |
+
image that includes the area inside the specified rectangular box.\n\n2. `croppedIm
|
| 309 |
+
= catIm.crop((335, 345, 565, 560))`: This line of code makes a new Image object
|
| 310 |
+
for the cropped image and stores it in the ''croppedIm'' variable.\n\n3. `croppedIm.save(''cropped.png'')`:
|
| 311 |
+
Finally, the ''save()'' function is called on ''croppedIm'', which saves this
|
| 312 |
+
newly created cropped image as ''cropped.png''. If the file doesn''t already exist,
|
| 313 |
+
it will be created in your active Python script''s current directory. \n\nWith
|
| 314 |
+
this simple coding approach, we can easily generate a new image from the original,
|
| 315 |
+
containing only the cropped section. \n\nPlease note: Pillow uses a coordinate
|
| 316 |
+
system with (0,0) in the upper left corner. The coordinate system is right handed,
|
| 317 |
+
meaning the right direction corresponds to increasing X, and down direction corresponds
|
| 318 |
+
to increasing Y. Always check the coordinate before cropping, else you might end
|
| 319 |
+
up cropping the wrong section of the image."}]'
|
| 320 |
+
- 'Question: what is the standardized coefficient in a regression, Answer: Standardized
|
| 321 |
+
coefficient In statistics, standardized coefficients or beta coefficients are
|
| 322 |
+
the estimates resulting from a regression analysis that have been standardized
|
| 323 |
+
so that the variances of dependent and independent variables are 1.[1] Therefore,
|
| 324 |
+
standardized coefficients refer to how many standard deviations a dependent variable
|
| 325 |
+
will change, per standard deviation increase in the predictor variable. For univariate
|
| 326 |
+
regression, the absolute value of the standardized coefficient equals the correlation
|
| 327 |
+
coefficient. Standardization of the coefficient is usually done to answer the
|
| 328 |
+
question of which of the independent variables have a greater effect on the dependent
|
| 329 |
+
variable in a multiple regression analysis, when the variables are measured in
|
| 330 |
+
different units of measurement (for example, income measured in dollars and family
|
| 331 |
+
size measured in number of individuals).'
|
| 332 |
+
- '[{''from'': ''human'', ''value'': ''Solve the following math problem step-by-step.\nSimplify
|
| 333 |
+
your answer as much as possible. Present your final answer as \\boxed{Your Answer}.\nSarah
|
| 334 |
+
is playing tic-tac-toe. She wins $1 for every game she wins, $0 for every game
|
| 335 |
+
she ties, and she loses $2 for every game she loses. If she played 100 games where
|
| 336 |
+
she tied 40 of them, and she lost $30, how many games did she win?''}, {''from'':
|
| 337 |
+
''gpt'', ''value'': "Step 1: First, let''s calculate the number of games Sarah
|
| 338 |
+
lost. We know that she lost $30 in total, and she loses $2 for each game she loses.
|
| 339 |
+
So the number of games she lost is 30 / 2 = 15.\nStep 2: We also know that Sarah
|
| 340 |
+
played 100 games in total, and 40 of them were ties and 15 were losses. So the
|
| 341 |
+
number of games she won is 100 - 40 - 15 = 45.\n\nAnswer:\n\\boxed{45}"}, {''from'':
|
| 342 |
+
''human'', ''value'': ''Out: 45\nYour answer is wrong.\nYour approach to calculate
|
| 343 |
+
the number of games Sarah lost is correct. However, you made a mistake in calculating
|
| 344 |
+
the number of games she won. You just subtracted the games she lost and tied from
|
| 345 |
+
the total games. But remember, she wins $1 for every game she wins and loses $2
|
| 346 |
+
for each game she loses. So, the losses include not only the games she actually
|
| 347 |
+
lost, but also the money she could have won if she had won those games. Thus,
|
| 348 |
+
when calculating the number of games won, you need to consider the games she actually
|
| 349 |
+
lost and the equivalent number of games she could have won to offset the monetary
|
| 350 |
+
losses. Try to adjust your calculations for game wins.''}, {''from'': ''gpt'',
|
| 351 |
+
''value'': "Step 1: First, let''s calculate the number of games Sarah lost. We
|
| 352 |
+
know that she lost $30 in total, and she loses $2 for each game she loses. So
|
| 353 |
+
the number of games she lost is 30 / 2 = 15.\nStep 2: To calculate the number
|
| 354 |
+
of games Sarah won, we need to consider the monetary losses from the games she
|
| 355 |
+
actually lost. For each game she lost, she would have won $2 if she had won. So
|
| 356 |
+
the equivalent number of games she could have won to offset the monetary losses
|
| 357 |
+
is 15 / 2 = 7.5.\nStep 3: Now, we can calculate the number of games she won. We
|
| 358 |
+
know that she played a total of 100 games, and out of those, 40 were ties, 15
|
| 359 |
+
were losses, and she could have won 7.5 games. So the number of games she won
|
| 360 |
+
is 100 - 40 - 15 - 7.5 = 37.5.\n\nAnswer:\n\\boxed{37.5}"}, {''from'': ''human'',
|
| 361 |
+
''value'': ''Out: 37.5\nYour answer is wrong.\nYour reasoning in Step 2 is incorrect.
|
| 362 |
+
The idea that "for each game she lost, she would have won $2 if she had won" is
|
| 363 |
+
not logical. This calculation double counts the losses and doesn\''t make sense
|
| 364 |
+
in the context of the problem. Instead, the losses should be considered as a product
|
| 365 |
+
of the numbers of games lost and the money lost per each game lost. Hence, to
|
| 366 |
+
find the number of games won, simply subtract the number of games tied and lost
|
| 367 |
+
from the total games played. You should then calculate the total losses in terms
|
| 368 |
+
of games by dividing the total lost money by the dollar earned per game won. This
|
| 369 |
+
will give the total lost games not only from losing but also from not winning.
|
| 370 |
+
Adding these lost games and tied games then subtracting from total games played
|
| 371 |
+
will give the number of games won. Try to fix your calculations accordingly.''},
|
| 372 |
+
{''from'': ''gpt'', ''value'': "Step 1: First, let''s calculate the number of
|
| 373 |
+
games Sarah lost. We know that she lost $30 in total, and she loses $2 for each
|
| 374 |
+
game she loses. So the number of games she lost is 30 / -2 = -15. This means our
|
| 375 |
+
assumption that she lost 15 games is incorrect.\n\nStep 2: To find the number
|
| 376 |
+
of games Sarah won, we need to consider the total amount of money lost. She loses
|
| 377 |
+
$2 for each game she loses and she lost a total of $30. So the total number of
|
| 378 |
+
games lost is -30 / -2 = 15.\n\nStep 3: Now, we can calculate the number of games
|
| 379 |
+
won. We know that Sarah played a total of 100 games, out of which 40 were ties
|
| 380 |
+
and 15 were losses. So the number of games she won is 100 - 40 - 15 = 45.\n\nAnswer:\n\\boxed{45}"},
|
| 381 |
+
{''from'': ''human'', ''value'': ''Out: 45\nYour answer is wrong.\nYou seem to
|
| 382 |
+
misunderstand how the losses impact the total number of games won. When Sarah
|
| 383 |
+
loses a game, she not only loses $2, but also loses the opportunity to win $1.
|
| 384 |
+
Therefore, each lost game equates to a $3 loss ($2 lost + $1 not won). With this
|
| 385 |
+
in mind, try revising your calculation of the number of games Sarah lost. Then
|
| 386 |
+
proceed with calculating the number of games won.''}, {''from'': ''gpt'', ''value'':
|
| 387 |
+
"Step 1: First, let''s calculate the number of games Sarah lost. We know that
|
| 388 |
+
she lost $30 in total. Since each loss costs her $3 (including the $2 lost and
|
| 389 |
+
the $1 she could have won), the number of games she lost is 30 / 3 = 10.\n\nStep
|
| 390 |
+
2: To find the number of games Sarah won, we subtract the number of games tied
|
| 391 |
+
and lost from the total number of games played. She played 100 games in total,
|
| 392 |
+
with 40 ties and 10 losses. So the number of games she won is 100 - 40 - 10 =
|
| 393 |
+
50.\n\nAnswer:\n\\boxed{50}"}, {''from'': ''human'', ''value'': "Out: 50\nYour
|
| 394 |
+
answer is wrong.\nYour Step 1 is correct but Step 2 isn''t. After knowing the
|
| 395 |
+
number of games lost, You just straight up subtracted the number of tied and lost
|
| 396 |
+
games from the total. But this regularization doesn''t apply here, because she
|
| 397 |
+
is losing $2 for every game she loses. So, her total losses of $30 isn''t just
|
| 398 |
+
from the games she lost, it''s also from the games she could have won. So in calculating
|
| 399 |
+
the games she won, you actually need to find out how many games she didn''t win
|
| 400 |
+
and then subtract these from the total games. A hint is to consider the total
|
| 401 |
+
income if she had won all the games and compare it with her actual total income.
|
| 402 |
+
The difference will give you the number of games she didn''t win (i.e., she either
|
| 403 |
+
lost or tied)."}, {''from'': ''gpt'', ''value'': "Step 1: Let''s start by breaking
|
| 404 |
+
down the information we have. Sarah played a total of 100 games, out of which
|
| 405 |
+
40 were ties. So, the remaining games must be the ones she either won or lost.\nStep
|
| 406 |
+
2: We''re given that she lost 30 games, so to find the number of games she won,
|
| 407 |
+
we can calculate the difference between the total games and the sum of tied and
|
| 408 |
+
lost games, which gives us the number of games she won.\nStep 3: Essentially,
|
| 409 |
+
since there are only 3 possible outcomes in tic-tac-toe (win, lose, or tie), the
|
| 410 |
+
number of games she won can be obtained by subtracting the number of games tied
|
| 411 |
+
and the number of games lost from the total games played. So, the formula for
|
| 412 |
+
the number of games won is: Total games - Tied games - Lost games.\nStep 4: Substituting
|
| 413 |
+
the given values, we have: Number of games won = 100 - 40 - 30 = 30.\nStep 5: This
|
| 414 |
+
leads us to the conclusion that Sarah won 30 games out of the 100 games she played.\n\nAnswer:\n\\boxed{30}"}]'
|
| 415 |
+
- source_sentence: '[{''from'': ''human'', ''value'': ''Solve the following math problem
|
| 416 |
+
step-by-step.\nSimplify your answer as much as possible. Present your final answer
|
| 417 |
+
as \\boxed{Your Answer}.\nA pentagon is drawn by placing an isosceles right triangle
|
| 418 |
+
on top of a square as pictured. What percent of the area of the pentagon is the
|
| 419 |
+
area of the right triangle?\n\n[asy]\nsize(50);\ndraw((0,0)--(0,-1)--(1,-1)--(1,0)--(0,0)--(.5,.5)--(1,0));\n[/asy]''},
|
| 420 |
+
{''from'': ''gpt'', ''value'': "Step 1: We are given a pentagon formed by placing
|
| 421 |
+
an isosceles right triangle on top of a square.\nStep 2: To find the percentage
|
| 422 |
+
of the area of the pentagon that is the right triangle, we need to consider the
|
| 423 |
+
areas of both shapes and then calculate the ratio of the triangle''s area to the
|
| 424 |
+
total pentagon''s area.\nStep 3: Let''s designate the leg length of the isosceles
|
| 425 |
+
right triangle as $x$. This means the hypotenuse of the triangle has a length
|
| 426 |
+
of $x\\sqrt{2}$.\nStep 4: Since the hypotenuse of the triangle is a side of the
|
| 427 |
+
square, the area of the square is $(x\\sqrt{2})^2 = 2x^2$. \nStep 5: On the other
|
| 428 |
+
hand, the area of the triangle is $(x)(x)/2 = x^2/2$. This is obtained using the
|
| 429 |
+
standard formula for the area of a triangle, which is base times height divided
|
| 430 |
+
by 2.\nStep 6: So, the area of the pentagon is the sum of the areas of the triangle
|
| 431 |
+
and the square, which gives us $\\frac{x^2}{2} + 2x^2 = \\frac{5x^2}{2}$.\nStep
|
| 432 |
+
7: Now, to find the percentage of the pentagon''s area that is inside the triangle,
|
| 433 |
+
we calculate the ratio of the triangle''s area to the total pentagon''s area.
|
| 434 |
+
This can be computed as $\\frac{x^2/2}{5x^2/2} =\\frac{x^2}{2}\\cdot \\frac{2}{5x^2}
|
| 435 |
+
= \\frac15$.\nStep 8: Therefore, the percentage of the area of the pentagon that
|
| 436 |
+
is the right triangle is $\\frac{1}{5}$, which is equivalent to $\\boxed{20\\%}$.\nStep
|
| 437 |
+
9: As an alternate solution, we can consider drawing the two diagonals of the
|
| 438 |
+
square and observe the resulting geometric properties to arrive at the same conclusion.\n\nAnswer:\n\\boxed{20\\%}"}]'
|
| 439 |
+
sentences:
|
| 440 |
+
- 'Question: who played george luz in band of brothers, Answer: Rick Gomez Richard
|
| 441 |
+
Harper "Rick" Gomez (born June 1, 1972) is an American actor and voice actor.
|
| 442 |
+
He is known for portraying Technician 4th Grade George Luz in the HBO television
|
| 443 |
+
miniseries Band of Brothers, and as "Endless Mike" Hellstrom in the Nickelodeon
|
| 444 |
+
TV series The Adventures of Pete and Pete. He is the older brother of actor Joshua
|
| 445 |
+
Gomez.'
|
| 446 |
+
- '[{''from'': ''human'', ''value'': ''Solve the following math problem step-by-step.\nSimplify
|
| 447 |
+
your answer as much as possible. Present your final answer as \\boxed{Your Answer}.\nA
|
| 448 |
+
pentagon is drawn by placing an isosceles right triangle on top of a square as
|
| 449 |
+
pictured. What percent of the area of the pentagon is the area of the right triangle?\n\n[asy]\nsize(50);\ndraw((0,0)--(0,-1)--(1,-1)--(1,0)--(0,0)--(.5,.5)--(1,0));\n[/asy]''},
|
| 450 |
+
{''from'': ''gpt'', ''value'': "Step 1: We are given a pentagon formed by placing
|
| 451 |
+
an isosceles right triangle on top of a square.\nStep 2: To find the percentage
|
| 452 |
+
of the area of the pentagon that is the right triangle, we need to consider the
|
| 453 |
+
areas of both shapes and then calculate the ratio of the triangle''s area to the
|
| 454 |
+
total pentagon''s area.\nStep 3: Let''s designate the leg length of the isosceles
|
| 455 |
+
right triangle as $x$. This means the hypotenuse of the triangle has a length
|
| 456 |
+
of $x\\sqrt{2}$.\nStep 4: Since the hypotenuse of the triangle is a side of the
|
| 457 |
+
square, the area of the square is $(x\\sqrt{2})^2 = 2x^2$. \nStep 5: On the other
|
| 458 |
+
hand, the area of the triangle is $(x)(x)/2 = x^2/2$. This is obtained using the
|
| 459 |
+
standard formula for the area of a triangle, which is base times height divided
|
| 460 |
+
by 2.\nStep 6: So, the area of the pentagon is the sum of the areas of the triangle
|
| 461 |
+
and the square, which gives us $\\frac{x^2}{2} + 2x^2 = \\frac{5x^2}{2}$.\nStep
|
| 462 |
+
7: Now, to find the percentage of the pentagon''s area that is inside the triangle,
|
| 463 |
+
we calculate the ratio of the triangle''s area to the total pentagon''s area.
|
| 464 |
+
This can be computed as $\\frac{x^2/2}{5x^2/2} =\\frac{x^2}{2}\\cdot \\frac{2}{5x^2}
|
| 465 |
+
= \\frac15$.\nStep 8: Therefore, the percentage of the area of the pentagon that
|
| 466 |
+
is the right triangle is $\\frac{1}{5}$, which is equivalent to $\\boxed{20\\%}$.\nStep
|
| 467 |
+
9: As an alternate solution, we can consider drawing the two diagonals of the
|
| 468 |
+
square and observe the resulting geometric properties to arrive at the same conclusion.\n\nAnswer:\n\\boxed{20\\%}"}]'
|
| 469 |
+
- 'Question: where are the basilar membrane and the organ of corti located, Answer:
|
| 470 |
+
Organ of Corti The organ of Corti is located in the scala media of the cochlea
|
| 471 |
+
of the inner ear between the vestibular duct and the tympanic duct and is composed
|
| 472 |
+
of mechanosensory cells, known as hair cells.[2] Strategically positioned on the
|
| 473 |
+
basilar membrane of the organ of Corti are three rows of outer hair cells (OHCs)
|
| 474 |
+
and one row of inner hair cells (IHCs).[4] Separating these hair cells are supporting
|
| 475 |
+
cells: Deiters cells, also called phalangeal cells, which separate and support
|
| 476 |
+
both the OHCs and the IHCs.[4]'
|
| 477 |
+
pipeline_tag: sentence-similarity
|
| 478 |
+
library_name: sentence-transformers
|
| 479 |
+
---
|
| 480 |
+
|
| 481 |
+
# SentenceTransformer based on sucharush/e5_stem_finetuned
|
| 482 |
+
|
| 483 |
+
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sucharush/e5_stem_finetuned](https://huggingface.co/sucharush/e5_stem_finetuned). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
| 484 |
+
|
| 485 |
+
## Model Details
|
| 486 |
+
|
| 487 |
+
### Model Description
|
| 488 |
+
- **Model Type:** Sentence Transformer
|
| 489 |
+
- **Base model:** [sucharush/e5_stem_finetuned](https://huggingface.co/sucharush/e5_stem_finetuned) <!-- at revision 3013cc95b53a83082746afd556743e81f1da5dff -->
|
| 490 |
+
- **Maximum Sequence Length:** 512 tokens
|
| 491 |
+
- **Output Dimensionality:** 768 dimensions
|
| 492 |
+
- **Similarity Function:** Cosine Similarity
|
| 493 |
+
<!-- - **Training Dataset:** Unknown -->
|
| 494 |
+
<!-- - **Language:** Unknown -->
|
| 495 |
+
<!-- - **License:** Unknown -->
|
| 496 |
+
|
| 497 |
+
### Model Sources
|
| 498 |
+
|
| 499 |
+
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
|
| 500 |
+
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
|
| 501 |
+
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
|
| 502 |
+
|
| 503 |
+
### Full Model Architecture
|
| 504 |
+
|
| 505 |
+
```
|
| 506 |
+
SentenceTransformer(
|
| 507 |
+
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
|
| 508 |
+
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
|
| 509 |
+
(2): Normalize()
|
| 510 |
+
)
|
| 511 |
+
```
|
| 512 |
+
|
| 513 |
+
## Usage
|
| 514 |
+
|
| 515 |
+
### Direct Usage (Sentence Transformers)
|
| 516 |
+
|
| 517 |
+
First install the Sentence Transformers library:
|
| 518 |
+
|
| 519 |
+
```bash
|
| 520 |
+
pip install -U sentence-transformers
|
| 521 |
+
```
|
| 522 |
+
|
| 523 |
+
Then you can load this model and run inference.
|
| 524 |
+
```python
|
| 525 |
+
from sentence_transformers import SentenceTransformer
|
| 526 |
+
|
| 527 |
+
# Download from the 🤗 Hub
|
| 528 |
+
model = SentenceTransformer("Stergios-Konstantinidis/MNLP_M3_tokenizer_tuned")
|
| 529 |
+
# Run inference
|
| 530 |
+
sentences = [
|
| 531 |
+
'[{\'from\': \'human\', \'value\': \'Solve the following math problem step-by-step.\\nSimplify your answer as much as possible. Present your final answer as \\\\boxed{Your Answer}.\\nA pentagon is drawn by placing an isosceles right triangle on top of a square as pictured. What percent of the area of the pentagon is the area of the right triangle?\\n\\n[asy]\\nsize(50);\\ndraw((0,0)--(0,-1)--(1,-1)--(1,0)--(0,0)--(.5,.5)--(1,0));\\n[/asy]\'}, {\'from\': \'gpt\', \'value\': "Step 1: We are given a pentagon formed by placing an isosceles right triangle on top of a square.\\nStep 2: To find the percentage of the area of the pentagon that is the right triangle, we need to consider the areas of both shapes and then calculate the ratio of the triangle\'s area to the total pentagon\'s area.\\nStep 3: Let\'s designate the leg length of the isosceles right triangle as $x$. This means the hypotenuse of the triangle has a length of $x\\\\sqrt{2}$.\\nStep 4: Since the hypotenuse of the triangle is a side of the square, the area of the square is $(x\\\\sqrt{2})^2 = 2x^2$. \\nStep 5: On the other hand, the area of the triangle is $(x)(x)/2 = x^2/2$. This is obtained using the standard formula for the area of a triangle, which is base times height divided by 2.\\nStep 6: So, the area of the pentagon is the sum of the areas of the triangle and the square, which gives us $\\\\frac{x^2}{2} + 2x^2 = \\\\frac{5x^2}{2}$.\\nStep 7: Now, to find the percentage of the pentagon\'s area that is inside the triangle, we calculate the ratio of the triangle\'s area to the total pentagon\'s area. This can be computed as $\\\\frac{x^2/2}{5x^2/2} =\\\\frac{x^2}{2}\\\\cdot \\\\frac{2}{5x^2} = \\\\frac15$.\\nStep 8: Therefore, the percentage of the area of the pentagon that is the right triangle is $\\\\frac{1}{5}$, which is equivalent to $\\\\boxed{20\\\\%}$.\\nStep 9: As an alternate solution, we can consider drawing the two diagonals of the square and observe the resulting geometric properties to arrive at the same conclusion.\\n\\nAnswer:\\n\\\\boxed{20\\\\%}"}]',
|
| 532 |
+
'[{\'from\': \'human\', \'value\': \'Solve the following math problem step-by-step.\\nSimplify your answer as much as possible. Present your final answer as \\\\boxed{Your Answer}.\\nA pentagon is drawn by placing an isosceles right triangle on top of a square as pictured. What percent of the area of the pentagon is the area of the right triangle?\\n\\n[asy]\\nsize(50);\\ndraw((0,0)--(0,-1)--(1,-1)--(1,0)--(0,0)--(.5,.5)--(1,0));\\n[/asy]\'}, {\'from\': \'gpt\', \'value\': "Step 1: We are given a pentagon formed by placing an isosceles right triangle on top of a square.\\nStep 2: To find the percentage of the area of the pentagon that is the right triangle, we need to consider the areas of both shapes and then calculate the ratio of the triangle\'s area to the total pentagon\'s area.\\nStep 3: Let\'s designate the leg length of the isosceles right triangle as $x$. This means the hypotenuse of the triangle has a length of $x\\\\sqrt{2}$.\\nStep 4: Since the hypotenuse of the triangle is a side of the square, the area of the square is $(x\\\\sqrt{2})^2 = 2x^2$. \\nStep 5: On the other hand, the area of the triangle is $(x)(x)/2 = x^2/2$. This is obtained using the standard formula for the area of a triangle, which is base times height divided by 2.\\nStep 6: So, the area of the pentagon is the sum of the areas of the triangle and the square, which gives us $\\\\frac{x^2}{2} + 2x^2 = \\\\frac{5x^2}{2}$.\\nStep 7: Now, to find the percentage of the pentagon\'s area that is inside the triangle, we calculate the ratio of the triangle\'s area to the total pentagon\'s area. This can be computed as $\\\\frac{x^2/2}{5x^2/2} =\\\\frac{x^2}{2}\\\\cdot \\\\frac{2}{5x^2} = \\\\frac15$.\\nStep 8: Therefore, the percentage of the area of the pentagon that is the right triangle is $\\\\frac{1}{5}$, which is equivalent to $\\\\boxed{20\\\\%}$.\\nStep 9: As an alternate solution, we can consider drawing the two diagonals of the square and observe the resulting geometric properties to arrive at the same conclusion.\\n\\nAnswer:\\n\\\\boxed{20\\\\%}"}]',
|
| 533 |
+
'Question: where are the basilar membrane and the organ of corti located, Answer: Organ of Corti The organ of Corti is located in the scala media of the cochlea of the inner ear between the vestibular duct and the tympanic duct and is composed of mechanosensory cells, known as hair cells.[2] Strategically positioned on the basilar membrane of the organ of Corti are three rows of outer hair cells (OHCs) and one row of inner hair cells (IHCs).[4] Separating these hair cells are supporting cells: Deiters cells, also called phalangeal cells, which separate and support both the OHCs and the IHCs.[4]',
|
| 534 |
+
]
|
| 535 |
+
embeddings = model.encode(sentences)
|
| 536 |
+
print(embeddings.shape)
|
| 537 |
+
# [3, 768]
|
| 538 |
+
|
| 539 |
+
# Get the similarity scores for the embeddings
|
| 540 |
+
similarities = model.similarity(embeddings, embeddings)
|
| 541 |
+
print(similarities.shape)
|
| 542 |
+
# [3, 3]
|
| 543 |
+
```
|
| 544 |
+
|
| 545 |
+
<!--
|
| 546 |
+
### Direct Usage (Transformers)
|
| 547 |
+
|
| 548 |
+
<details><summary>Click to see the direct usage in Transformers</summary>
|
| 549 |
+
|
| 550 |
+
</details>
|
| 551 |
+
-->
|
| 552 |
+
|
| 553 |
+
<!--
|
| 554 |
+
### Downstream Usage (Sentence Transformers)
|
| 555 |
+
|
| 556 |
+
You can finetune this model on your own dataset.
|
| 557 |
+
|
| 558 |
+
<details><summary>Click to expand</summary>
|
| 559 |
+
|
| 560 |
+
</details>
|
| 561 |
+
-->
|
| 562 |
+
|
| 563 |
+
<!--
|
| 564 |
+
### Out-of-Scope Use
|
| 565 |
+
|
| 566 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 567 |
+
-->
|
| 568 |
+
|
| 569 |
+
<!--
|
| 570 |
+
## Bias, Risks and Limitations
|
| 571 |
+
|
| 572 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
| 573 |
+
-->
|
| 574 |
+
|
| 575 |
+
<!--
|
| 576 |
+
### Recommendations
|
| 577 |
+
|
| 578 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
| 579 |
+
-->
|
| 580 |
+
|
| 581 |
+
## Training Details
|
| 582 |
+
|
| 583 |
+
### Training Dataset
|
| 584 |
+
|
| 585 |
+
#### Unnamed Dataset
|
| 586 |
+
|
| 587 |
+
* Size: 113,450 training samples
|
| 588 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
|
| 589 |
+
* Approximate statistics based on the first 1000 samples:
|
| 590 |
+
| | sentence_0 | sentence_1 | label |
|
| 591 |
+
|:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------|
|
| 592 |
+
| type | string | string | int |
|
| 593 |
+
| details | <ul><li>min: 17 tokens</li><li>mean: 283.42 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 16 tokens</li><li>mean: 281.46 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>0: ~80.00%</li><li>1: ~20.00%</li></ul> |
|
| 594 |
+
* Samples:
|
| 595 |
+
| sentence_0 | sentence_1 | label |
|
| 596 |
+
|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------|
|
| 597 |
+
| <code>Question: where does rasin in the sun take place, Answer: A Raisin in the Sun A Raisin in the Sun is a play by Lorraine Hansberry that debuted on Broadway in 1959.[1] The title comes from the poem "Harlem" (also known as "A Dream Deferred"[2]) by Langston Hughes. The story tells of a black family's experiences in the Washington Park Subdivision of Chicago's Woodlawn neighborhood as they attempt to "better" themselves with an insurance payout following the death of the father. The New York Drama Critics' Circle named it the best play of 1959.</code> | <code>Question: where does rasin in the sun take place, Answer: A Raisin in the Sun A Raisin in the Sun is a play by Lorraine Hansberry that debuted on Broadway in 1959.[1] The title comes from the poem "Harlem" (also known as "A Dream Deferred"[2]) by Langston Hughes. The story tells of a black family's experiences in the Washington Park Subdivision of Chicago's Woodlawn neighborhood as they attempt to "better" themselves with an insurance payout following the death of the father. The New York Drama Critics' Circle named it the best play of 1959.</code> | <code>1</code> |
|
| 598 |
+
| <code>Question: when does the movie midnight sun come out, Answer: Midnight Sun (2018 film) Midnight Sun is a 2018 American romantic drama film directed by Scott Speer and written by Eric Kirsten, based on the 2006 Japanese film of the same name. The film stars Bella Thorne, Patrick Schwarzenegger, and Rob Riggle, and follows a teenage girl with the disease xeroderma pigmentosum, which prevents her from going out into sunlight. When she meets a boy, she struggles to decide whether to tell him about her condition or pretend to live a normal life. Principal photography began on October 12, 2015 in Vancouver, British Columbia, Canada. The film was released in the United States on March 23, 2018.</code> | <code>Question: who sang the song unchained melody in the movie ghost, Answer: Unchained Melody "Unchained Melody" reappeared on the Billboard charts in 1990 after The Righteous Brothers' recording was used in the box office blockbuster film Ghost. Two versions charted in the US that year – the original and a new recording. According to Medley, he was interested in having the original recording released due to the renewed interest in the song, but was told that there were licensing issues. Although Hatfield was no longer in the same condition vocally as when he first recorded the song, they decided to re-record the song for Curb Records.[44] The re-recorded version was released as both a cassette single[45] and a CD single.[46] It received minimal airplay but recorded excellent sales, peaking at number 19. The re-recorded version was certified Platinum by the RIAA on January 10, 1991,[47] and received a Grammy Award nomination.[48]</code> | <code>0</code> |
|
| 599 |
+
| <code>Question: where is the o2 arena in london located, Answer: The O2 Arena The O2 Arena (temporarily the sponsor neutral "North Greenwich Arena", during the 2012 Summer Olympics and 2012 Summer Paralympics), is a multi purpose indoor arena located in the centre of The O2 entertainment complex on the Greenwich Peninsula in south east London.</code> | <code>[{'from': 'human', 'value': "Explain how Archimedes' Principle enables objects to float in water."}, {'from': 'gpt', 'value': "Archimedes' Principle states that the buoyant force acting on an object submerged in water is equal to the weight of the water displaced by the object. If the buoyant force exceeds the weight of the object, it will float."}]</code> | <code>0</code> |
|
| 600 |
+
* Loss: [<code>ContrastiveTensionLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastivetensionloss)
|
| 601 |
+
|
| 602 |
+
### Training Hyperparameters
|
| 603 |
+
#### Non-Default Hyperparameters
|
| 604 |
+
|
| 605 |
+
- `per_device_train_batch_size`: 10
|
| 606 |
+
- `per_device_eval_batch_size`: 10
|
| 607 |
+
- `num_train_epochs`: 10
|
| 608 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 609 |
+
|
| 610 |
+
#### All Hyperparameters
|
| 611 |
+
<details><summary>Click to expand</summary>
|
| 612 |
+
|
| 613 |
+
- `overwrite_output_dir`: False
|
| 614 |
+
- `do_predict`: False
|
| 615 |
+
- `eval_strategy`: no
|
| 616 |
+
- `prediction_loss_only`: True
|
| 617 |
+
- `per_device_train_batch_size`: 10
|
| 618 |
+
- `per_device_eval_batch_size`: 10
|
| 619 |
+
- `per_gpu_train_batch_size`: None
|
| 620 |
+
- `per_gpu_eval_batch_size`: None
|
| 621 |
+
- `gradient_accumulation_steps`: 1
|
| 622 |
+
- `eval_accumulation_steps`: None
|
| 623 |
+
- `torch_empty_cache_steps`: None
|
| 624 |
+
- `learning_rate`: 5e-05
|
| 625 |
+
- `weight_decay`: 0.0
|
| 626 |
+
- `adam_beta1`: 0.9
|
| 627 |
+
- `adam_beta2`: 0.999
|
| 628 |
+
- `adam_epsilon`: 1e-08
|
| 629 |
+
- `max_grad_norm`: 1
|
| 630 |
+
- `num_train_epochs`: 10
|
| 631 |
+
- `max_steps`: -1
|
| 632 |
+
- `lr_scheduler_type`: linear
|
| 633 |
+
- `lr_scheduler_kwargs`: {}
|
| 634 |
+
- `warmup_ratio`: 0.0
|
| 635 |
+
- `warmup_steps`: 0
|
| 636 |
+
- `log_level`: passive
|
| 637 |
+
- `log_level_replica`: warning
|
| 638 |
+
- `log_on_each_node`: True
|
| 639 |
+
- `logging_nan_inf_filter`: True
|
| 640 |
+
- `save_safetensors`: True
|
| 641 |
+
- `save_on_each_node`: False
|
| 642 |
+
- `save_only_model`: False
|
| 643 |
+
- `restore_callback_states_from_checkpoint`: False
|
| 644 |
+
- `no_cuda`: False
|
| 645 |
+
- `use_cpu`: False
|
| 646 |
+
- `use_mps_device`: False
|
| 647 |
+
- `seed`: 42
|
| 648 |
+
- `data_seed`: None
|
| 649 |
+
- `jit_mode_eval`: False
|
| 650 |
+
- `use_ipex`: False
|
| 651 |
+
- `bf16`: False
|
| 652 |
+
- `fp16`: False
|
| 653 |
+
- `fp16_opt_level`: O1
|
| 654 |
+
- `half_precision_backend`: auto
|
| 655 |
+
- `bf16_full_eval`: False
|
| 656 |
+
- `fp16_full_eval`: False
|
| 657 |
+
- `tf32`: None
|
| 658 |
+
- `local_rank`: 0
|
| 659 |
+
- `ddp_backend`: None
|
| 660 |
+
- `tpu_num_cores`: None
|
| 661 |
+
- `tpu_metrics_debug`: False
|
| 662 |
+
- `debug`: []
|
| 663 |
+
- `dataloader_drop_last`: False
|
| 664 |
+
- `dataloader_num_workers`: 0
|
| 665 |
+
- `dataloader_prefetch_factor`: None
|
| 666 |
+
- `past_index`: -1
|
| 667 |
+
- `disable_tqdm`: False
|
| 668 |
+
- `remove_unused_columns`: True
|
| 669 |
+
- `label_names`: None
|
| 670 |
+
- `load_best_model_at_end`: False
|
| 671 |
+
- `ignore_data_skip`: False
|
| 672 |
+
- `fsdp`: []
|
| 673 |
+
- `fsdp_min_num_params`: 0
|
| 674 |
+
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
|
| 675 |
+
- `fsdp_transformer_layer_cls_to_wrap`: None
|
| 676 |
+
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
|
| 677 |
+
- `deepspeed`: None
|
| 678 |
+
- `label_smoothing_factor`: 0.0
|
| 679 |
+
- `optim`: adamw_torch
|
| 680 |
+
- `optim_args`: None
|
| 681 |
+
- `adafactor`: False
|
| 682 |
+
- `group_by_length`: False
|
| 683 |
+
- `length_column_name`: length
|
| 684 |
+
- `ddp_find_unused_parameters`: None
|
| 685 |
+
- `ddp_bucket_cap_mb`: None
|
| 686 |
+
- `ddp_broadcast_buffers`: False
|
| 687 |
+
- `dataloader_pin_memory`: True
|
| 688 |
+
- `dataloader_persistent_workers`: False
|
| 689 |
+
- `skip_memory_metrics`: True
|
| 690 |
+
- `use_legacy_prediction_loop`: False
|
| 691 |
+
- `push_to_hub`: False
|
| 692 |
+
- `resume_from_checkpoint`: None
|
| 693 |
+
- `hub_model_id`: None
|
| 694 |
+
- `hub_strategy`: every_save
|
| 695 |
+
- `hub_private_repo`: None
|
| 696 |
+
- `hub_always_push`: False
|
| 697 |
+
- `gradient_checkpointing`: False
|
| 698 |
+
- `gradient_checkpointing_kwargs`: None
|
| 699 |
+
- `include_inputs_for_metrics`: False
|
| 700 |
+
- `include_for_metrics`: []
|
| 701 |
+
- `eval_do_concat_batches`: True
|
| 702 |
+
- `fp16_backend`: auto
|
| 703 |
+
- `push_to_hub_model_id`: None
|
| 704 |
+
- `push_to_hub_organization`: None
|
| 705 |
+
- `mp_parameters`:
|
| 706 |
+
- `auto_find_batch_size`: False
|
| 707 |
+
- `full_determinism`: False
|
| 708 |
+
- `torchdynamo`: None
|
| 709 |
+
- `ray_scope`: last
|
| 710 |
+
- `ddp_timeout`: 1800
|
| 711 |
+
- `torch_compile`: False
|
| 712 |
+
- `torch_compile_backend`: None
|
| 713 |
+
- `torch_compile_mode`: None
|
| 714 |
+
- `include_tokens_per_second`: False
|
| 715 |
+
- `include_num_input_tokens_seen`: False
|
| 716 |
+
- `neftune_noise_alpha`: None
|
| 717 |
+
- `optim_target_modules`: None
|
| 718 |
+
- `batch_eval_metrics`: False
|
| 719 |
+
- `eval_on_start`: False
|
| 720 |
+
- `use_liger_kernel`: False
|
| 721 |
+
- `eval_use_gather_object`: False
|
| 722 |
+
- `average_tokens_across_devices`: False
|
| 723 |
+
- `prompts`: None
|
| 724 |
+
- `batch_sampler`: batch_sampler
|
| 725 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 726 |
+
|
| 727 |
+
</details>
|
| 728 |
+
|
| 729 |
+
### Training Logs
|
| 730 |
+
<details><summary>Click to expand</summary>
|
| 731 |
+
|
| 732 |
+
| Epoch | Step | Training Loss |
|
| 733 |
+
|:------:|:------:|:-------------:|
|
| 734 |
+
| 0.0441 | 500 | 10.4083 |
|
| 735 |
+
| 0.0881 | 1000 | 7.1707 |
|
| 736 |
+
| 0.1322 | 1500 | 5.1666 |
|
| 737 |
+
| 0.1763 | 2000 | 5.2289 |
|
| 738 |
+
| 0.2204 | 2500 | 5.1515 |
|
| 739 |
+
| 0.2644 | 3000 | 5.0737 |
|
| 740 |
+
| 0.3085 | 3500 | 5.0824 |
|
| 741 |
+
| 0.3526 | 4000 | 5.0569 |
|
| 742 |
+
| 0.3967 | 4500 | 5.2043 |
|
| 743 |
+
| 0.4407 | 5000 | 5.1293 |
|
| 744 |
+
| 0.4848 | 5500 | 5.1169 |
|
| 745 |
+
| 0.5289 | 6000 | 5.0743 |
|
| 746 |
+
| 0.5729 | 6500 | 5.1535 |
|
| 747 |
+
| 0.6170 | 7000 | 5.2208 |
|
| 748 |
+
| 0.6611 | 7500 | 5.1847 |
|
| 749 |
+
| 0.7052 | 8000 | 5.1545 |
|
| 750 |
+
| 0.7492 | 8500 | 5.0944 |
|
| 751 |
+
| 0.7933 | 9000 | 5.0662 |
|
| 752 |
+
| 0.8374 | 9500 | 5.1083 |
|
| 753 |
+
| 0.8814 | 10000 | 5.102 |
|
| 754 |
+
| 0.9255 | 10500 | 5.2018 |
|
| 755 |
+
| 0.9696 | 11000 | 5.1479 |
|
| 756 |
+
| 1.0137 | 11500 | 5.1877 |
|
| 757 |
+
| 1.0577 | 12000 | 5.2397 |
|
| 758 |
+
| 1.1018 | 12500 | 5.1657 |
|
| 759 |
+
| 1.1459 | 13000 | 5.2276 |
|
| 760 |
+
| 1.1900 | 13500 | 5.2295 |
|
| 761 |
+
| 1.2340 | 14000 | 5.1096 |
|
| 762 |
+
| 1.2781 | 14500 | 5.0877 |
|
| 763 |
+
| 1.3222 | 15000 | 5.0776 |
|
| 764 |
+
| 1.3662 | 15500 | 5.0957 |
|
| 765 |
+
| 1.4103 | 16000 | 5.1655 |
|
| 766 |
+
| 1.4544 | 16500 | 5.0537 |
|
| 767 |
+
| 1.4985 | 17000 | 5.0836 |
|
| 768 |
+
| 1.5425 | 17500 | 5.1495 |
|
| 769 |
+
| 1.5866 | 18000 | 5.0776 |
|
| 770 |
+
| 1.6307 | 18500 | 5.1215 |
|
| 771 |
+
| 1.6747 | 19000 | 5.1614 |
|
| 772 |
+
| 1.7188 | 19500 | 5.0195 |
|
| 773 |
+
| 1.7629 | 20000 | 5.0855 |
|
| 774 |
+
| 1.8070 | 20500 | 5.2153 |
|
| 775 |
+
| 1.8510 | 21000 | 5.1569 |
|
| 776 |
+
| 1.8951 | 21500 | 5.1213 |
|
| 777 |
+
| 1.9392 | 22000 | 5.1694 |
|
| 778 |
+
| 1.9833 | 22500 | 5.1414 |
|
| 779 |
+
| 2.0273 | 23000 | 5.1314 |
|
| 780 |
+
| 2.0714 | 23500 | 5.2454 |
|
| 781 |
+
| 2.1155 | 24000 | 5.1414 |
|
| 782 |
+
| 2.1595 | 24500 | 5.1194 |
|
| 783 |
+
| 2.2036 | 25000 | 5.1115 |
|
| 784 |
+
| 2.2477 | 25500 | 5.0393 |
|
| 785 |
+
| 2.2918 | 26000 | 5.1154 |
|
| 786 |
+
| 2.3358 | 26500 | 5.1414 |
|
| 787 |
+
| 2.3799 | 27000 | 5.0794 |
|
| 788 |
+
| 2.4240 | 27500 | 5.1013 |
|
| 789 |
+
| 2.4680 | 28000 | 5.1873 |
|
| 790 |
+
| 2.5121 | 28500 | 5.1034 |
|
| 791 |
+
| 2.5562 | 29000 | 5.1593 |
|
| 792 |
+
| 2.6003 | 29500 | 5.1933 |
|
| 793 |
+
| 2.6443 | 30000 | 5.2192 |
|
| 794 |
+
| 2.6884 | 30500 | 5.1233 |
|
| 795 |
+
| 2.7325 | 31000 | 5.1033 |
|
| 796 |
+
| 2.7766 | 31500 | 5.1408 |
|
| 797 |
+
| 2.8206 | 32000 | 5.1233 |
|
| 798 |
+
| 2.8647 | 32500 | 5.1153 |
|
| 799 |
+
| 2.9088 | 33000 | 5.0795 |
|
| 800 |
+
| 2.9528 | 33500 | 5.1792 |
|
| 801 |
+
| 2.9969 | 34000 | 5.1233 |
|
| 802 |
+
| 3.0410 | 34500 | 5.0852 |
|
| 803 |
+
| 3.0851 | 35000 | 5.0653 |
|
| 804 |
+
| 3.1291 | 35500 | 5.0973 |
|
| 805 |
+
| 3.1732 | 36000 | 5.1212 |
|
| 806 |
+
| 3.2173 | 36500 | 5.1046 |
|
| 807 |
+
| 3.2613 | 37000 | 5.1193 |
|
| 808 |
+
| 3.3054 | 37500 | 5.1632 |
|
| 809 |
+
| 3.3495 | 38000 | 5.1932 |
|
| 810 |
+
| 3.3936 | 38500 | 5.0894 |
|
| 811 |
+
| 3.4376 | 39000 | 5.1493 |
|
| 812 |
+
| 3.4817 | 39500 | 5.1812 |
|
| 813 |
+
| 3.5258 | 40000 | 5.1612 |
|
| 814 |
+
| 3.5699 | 40500 | 5.0972 |
|
| 815 |
+
| 3.6139 | 41000 | 5.0872 |
|
| 816 |
+
| 3.6580 | 41500 | 5.2375 |
|
| 817 |
+
| 3.7021 | 42000 | 5.1652 |
|
| 818 |
+
| 3.7461 | 42500 | 5.1273 |
|
| 819 |
+
| 3.7902 | 43000 | 5.0695 |
|
| 820 |
+
| 3.8343 | 43500 | 5.1557 |
|
| 821 |
+
| 3.8784 | 44000 | 5.0697 |
|
| 822 |
+
| 3.9224 | 44500 | 5.1635 |
|
| 823 |
+
| 3.9665 | 45000 | 5.2095 |
|
| 824 |
+
| 4.0106 | 45500 | 5.0936 |
|
| 825 |
+
| 4.0546 | 46000 | 5.1435 |
|
| 826 |
+
| 4.0987 | 46500 | 5.1635 |
|
| 827 |
+
| 4.1428 | 47000 | 5.1415 |
|
| 828 |
+
| 4.1869 | 47500 | 5.1634 |
|
| 829 |
+
| 4.2309 | 48000 | 5.1595 |
|
| 830 |
+
| 4.2750 | 48500 | 5.0775 |
|
| 831 |
+
| 4.3191 | 49000 | 5.0976 |
|
| 832 |
+
| 4.3632 | 49500 | 5.1135 |
|
| 833 |
+
| 4.4072 | 50000 | 5.0877 |
|
| 834 |
+
| 4.4513 | 50500 | 5.1934 |
|
| 835 |
+
| 4.4954 | 51000 | 5.0915 |
|
| 836 |
+
| 4.5394 | 51500 | 5.1295 |
|
| 837 |
+
| 4.5835 | 52000 | 5.0994 |
|
| 838 |
+
| 4.6276 | 52500 | 5.1235 |
|
| 839 |
+
| 4.6717 | 53000 | 5.1435 |
|
| 840 |
+
| 4.7157 | 53500 | 5.1335 |
|
| 841 |
+
| 4.7598 | 54000 | 5.1735 |
|
| 842 |
+
| 4.8039 | 54500 | 5.0914 |
|
| 843 |
+
| 4.8480 | 55000 | 5.1673 |
|
| 844 |
+
| 4.8920 | 55500 | 5.1834 |
|
| 845 |
+
| 4.9361 | 56000 | 5.1434 |
|
| 846 |
+
| 4.9802 | 56500 | 5.1713 |
|
| 847 |
+
| 5.0242 | 57000 | 5.1034 |
|
| 848 |
+
| 5.0683 | 57500 | 5.0615 |
|
| 849 |
+
| 5.1124 | 58000 | 5.0835 |
|
| 850 |
+
| 5.1565 | 58500 | 5.0654 |
|
| 851 |
+
| 5.2005 | 59000 | 5.1813 |
|
| 852 |
+
| 5.2446 | 59500 | 5.2133 |
|
| 853 |
+
| 5.2887 | 60000 | 5.2273 |
|
| 854 |
+
| 5.3327 | 60500 | 5.1313 |
|
| 855 |
+
| 5.3768 | 61000 | 5.1415 |
|
| 856 |
+
| 5.4209 | 61500 | 5.2432 |
|
| 857 |
+
| 5.4650 | 62000 | 5.0854 |
|
| 858 |
+
| 5.5090 | 62500 | 5.1415 |
|
| 859 |
+
| 5.5531 | 63000 | 5.0834 |
|
| 860 |
+
| 5.5972 | 63500 | 5.1294 |
|
| 861 |
+
| 5.6413 | 64000 | 5.1234 |
|
| 862 |
+
| 5.6853 | 64500 | 5.1913 |
|
| 863 |
+
| 5.7294 | 65000 | 5.1793 |
|
| 864 |
+
| 5.7735 | 65500 | 5.0614 |
|
| 865 |
+
| 5.8175 | 66000 | 5.1014 |
|
| 866 |
+
| 5.8616 | 66500 | 5.0354 |
|
| 867 |
+
| 5.9057 | 67000 | 5.1913 |
|
| 868 |
+
| 5.9498 | 67500 | 5.0614 |
|
| 869 |
+
| 5.9938 | 68000 | 5.2453 |
|
| 870 |
+
| 6.0379 | 68500 | 5.0913 |
|
| 871 |
+
| 6.0820 | 69000 | 5.1135 |
|
| 872 |
+
| 6.1260 | 69500 | 5.1374 |
|
| 873 |
+
| 6.1701 | 70000 | 5.0933 |
|
| 874 |
+
| 6.2142 | 70500 | 5.0134 |
|
| 875 |
+
| 6.2583 | 71000 | 5.0435 |
|
| 876 |
+
| 6.3023 | 71500 | 5.2412 |
|
| 877 |
+
| 6.3464 | 72000 | 5.0754 |
|
| 878 |
+
| 6.3905 | 72500 | 5.1933 |
|
| 879 |
+
| 6.4346 | 73000 | 5.2932 |
|
| 880 |
+
| 6.4786 | 73500 | 5.0515 |
|
| 881 |
+
| 6.5227 | 74000 | 5.2073 |
|
| 882 |
+
| 6.5668 | 74500 | 5.1312 |
|
| 883 |
+
| 6.6108 | 75000 | 5.1332 |
|
| 884 |
+
| 6.6549 | 75500 | 5.0615 |
|
| 885 |
+
| 6.6990 | 76000 | 5.1873 |
|
| 886 |
+
| 6.7431 | 76500 | 5.0654 |
|
| 887 |
+
| 6.7871 | 77000 | 5.1333 |
|
| 888 |
+
| 6.8312 | 77500 | 5.1593 |
|
| 889 |
+
| 6.8753 | 78000 | 5.1234 |
|
| 890 |
+
| 6.9193 | 78500 | 5.2012 |
|
| 891 |
+
| 6.9634 | 79000 | 5.1513 |
|
| 892 |
+
| 7.0075 | 79500 | 5.1153 |
|
| 893 |
+
| 7.0516 | 80000 | 5.0414 |
|
| 894 |
+
| 7.0956 | 80500 | 5.2034 |
|
| 895 |
+
| 7.1397 | 81000 | 5.1073 |
|
| 896 |
+
| 7.1838 | 81500 | 5.1753 |
|
| 897 |
+
| 7.2279 | 82000 | 5.1233 |
|
| 898 |
+
| 7.2719 | 82500 | 5.1293 |
|
| 899 |
+
| 7.3160 | 83000 | 5.1492 |
|
| 900 |
+
| 7.3601 | 83500 | 5.2392 |
|
| 901 |
+
| 7.4041 | 84000 | 5.0033 |
|
| 902 |
+
| 7.4482 | 84500 | 5.1912 |
|
| 903 |
+
| 7.4923 | 85000 | 5.1673 |
|
| 904 |
+
| 7.5364 | 85500 | 5.1932 |
|
| 905 |
+
| 7.5804 | 86000 | 5.1393 |
|
| 906 |
+
| 7.6245 | 86500 | 5.1653 |
|
| 907 |
+
| 7.6686 | 87000 | 5.1232 |
|
| 908 |
+
| 7.7126 | 87500 | 5.0993 |
|
| 909 |
+
| 7.7567 | 88000 | 5.1313 |
|
| 910 |
+
| 7.8008 | 88500 | 5.0695 |
|
| 911 |
+
| 7.8449 | 89000 | 5.1493 |
|
| 912 |
+
| 7.8889 | 89500 | 5.1733 |
|
| 913 |
+
| 7.9330 | 90000 | 5.1233 |
|
| 914 |
+
| 7.9771 | 90500 | 5.1292 |
|
| 915 |
+
| 8.0212 | 91000 | 5.0254 |
|
| 916 |
+
| 8.0652 | 91500 | 5.2092 |
|
| 917 |
+
| 8.1093 | 92000 | 5.1313 |
|
| 918 |
+
| 8.1534 | 92500 | 5.0834 |
|
| 919 |
+
| 8.1974 | 93000 | 5.1032 |
|
| 920 |
+
| 8.2415 | 93500 | 5.0653 |
|
| 921 |
+
| 8.2856 | 94000 | 5.1832 |
|
| 922 |
+
| 8.3297 | 94500 | 5.1752 |
|
| 923 |
+
| 8.3737 | 95000 | 5.0752 |
|
| 924 |
+
| 8.4178 | 95500 | 5.1593 |
|
| 925 |
+
| 8.4619 | 96000 | 5.1633 |
|
| 926 |
+
| 8.5059 | 96500 | 5.0794 |
|
| 927 |
+
| 8.5500 | 97000 | 5.1991 |
|
| 928 |
+
| 8.5941 | 97500 | 5.0952 |
|
| 929 |
+
| 8.6382 | 98000 | 5.1153 |
|
| 930 |
+
| 8.6822 | 98500 | 5.1192 |
|
| 931 |
+
| 8.7263 | 99000 | 5.1053 |
|
| 932 |
+
| 8.7704 | 99500 | 5.2213 |
|
| 933 |
+
| 8.8145 | 100000 | 5.1571 |
|
| 934 |
+
| 8.8585 | 100500 | 5.1292 |
|
| 935 |
+
| 8.9026 | 101000 | 5.1531 |
|
| 936 |
+
| 8.9467 | 101500 | 5.2251 |
|
| 937 |
+
| 8.9907 | 102000 | 5.0733 |
|
| 938 |
+
| 9.0348 | 102500 | 5.1313 |
|
| 939 |
+
| 9.0789 | 103000 | 5.1973 |
|
| 940 |
+
| 9.1230 | 103500 | 5.1133 |
|
| 941 |
+
| 9.1670 | 104000 | 5.1813 |
|
| 942 |
+
| 9.2111 | 104500 | 5.2472 |
|
| 943 |
+
| 9.2552 | 105000 | 5.1352 |
|
| 944 |
+
| 9.2993 | 105500 | 5.0373 |
|
| 945 |
+
| 9.3433 | 106000 | 5.1791 |
|
| 946 |
+
| 9.3874 | 106500 | 5.1891 |
|
| 947 |
+
| 9.4315 | 107000 | 5.0832 |
|
| 948 |
+
| 9.4755 | 107500 | 5.1632 |
|
| 949 |
+
| 9.5196 | 108000 | 5.1012 |
|
| 950 |
+
| 9.5637 | 108500 | 5.0354 |
|
| 951 |
+
| 9.6078 | 109000 | 5.2112 |
|
| 952 |
+
| 9.6518 | 109500 | 5.1212 |
|
| 953 |
+
| 9.6959 | 110000 | 5.1212 |
|
| 954 |
+
| 9.7400 | 110500 | 5.1612 |
|
| 955 |
+
| 9.7840 | 111000 | 5.1613 |
|
| 956 |
+
| 9.8281 | 111500 | 5.1072 |
|
| 957 |
+
| 9.8722 | 112000 | 5.0852 |
|
| 958 |
+
| 9.9163 | 112500 | 5.0234 |
|
| 959 |
+
| 9.9603 | 113000 | 5.1033 |
|
| 960 |
+
|
| 961 |
+
</details>
|
| 962 |
+
|
| 963 |
+
### Framework Versions
|
| 964 |
+
- Python: 3.12.8
|
| 965 |
+
- Sentence Transformers: 3.4.1
|
| 966 |
+
- Transformers: 4.52.4
|
| 967 |
+
- PyTorch: 2.6.0+cu126
|
| 968 |
+
- Accelerate: 1.3.0
|
| 969 |
+
- Datasets: 3.2.0
|
| 970 |
+
- Tokenizers: 0.21.0
|
| 971 |
+
|
| 972 |
+
## Citation
|
| 973 |
+
|
| 974 |
+
### BibTeX
|
| 975 |
+
|
| 976 |
+
#### Sentence Transformers
|
| 977 |
+
```bibtex
|
| 978 |
+
@inproceedings{reimers-2019-sentence-bert,
|
| 979 |
+
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|
| 980 |
+
author = "Reimers, Nils and Gurevych, Iryna",
|
| 981 |
+
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
|
| 982 |
+
month = "11",
|
| 983 |
+
year = "2019",
|
| 984 |
+
publisher = "Association for Computational Linguistics",
|
| 985 |
+
url = "https://arxiv.org/abs/1908.10084",
|
| 986 |
+
}
|
| 987 |
+
```
|
| 988 |
+
|
| 989 |
+
#### ContrastiveTensionLoss
|
| 990 |
+
```bibtex
|
| 991 |
+
@inproceedings{carlsson2021semantic,
|
| 992 |
+
title={Semantic Re-tuning with Contrastive Tension},
|
| 993 |
+
author={Fredrik Carlsson and Amaru Cuba Gyllensten and Evangelia Gogoulou and Erik Ylip{"a}{"a} Hellqvist and Magnus Sahlgren},
|
| 994 |
+
booktitle={International Conference on Learning Representations},
|
| 995 |
+
year={2021},
|
| 996 |
+
url={https://openreview.net/forum?id=Ov_sMNau-PF}
|
| 997 |
+
}
|
| 998 |
+
```
|
| 999 |
+
|
| 1000 |
+
<!--
|
| 1001 |
+
## Glossary
|
| 1002 |
+
|
| 1003 |
+
*Clearly define terms in order to be accessible across audiences.*
|
| 1004 |
+
-->
|
| 1005 |
+
|
| 1006 |
+
<!--
|
| 1007 |
+
## Model Card Authors
|
| 1008 |
+
|
| 1009 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
| 1010 |
+
-->
|
| 1011 |
+
|
| 1012 |
+
<!--
|
| 1013 |
+
## Model Card Contact
|
| 1014 |
+
|
| 1015 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
| 1016 |
+
-->
|
config.json
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"BertModel"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.1,
|
| 6 |
+
"classifier_dropout": null,
|
| 7 |
+
"gradient_checkpointing": false,
|
| 8 |
+
"hidden_act": "gelu",
|
| 9 |
+
"hidden_dropout_prob": 0.1,
|
| 10 |
+
"hidden_size": 768,
|
| 11 |
+
"initializer_range": 0.02,
|
| 12 |
+
"intermediate_size": 3072,
|
| 13 |
+
"layer_norm_eps": 1e-12,
|
| 14 |
+
"max_position_embeddings": 512,
|
| 15 |
+
"model_type": "bert",
|
| 16 |
+
"num_attention_heads": 12,
|
| 17 |
+
"num_hidden_layers": 12,
|
| 18 |
+
"pad_token_id": 0,
|
| 19 |
+
"position_embedding_type": "absolute",
|
| 20 |
+
"torch_dtype": "float32",
|
| 21 |
+
"transformers_version": "4.52.4",
|
| 22 |
+
"type_vocab_size": 2,
|
| 23 |
+
"use_cache": true,
|
| 24 |
+
"vocab_size": 30522
|
| 25 |
+
}
|
config_sentence_transformers.json
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"__version__": {
|
| 3 |
+
"sentence_transformers": "3.4.1",
|
| 4 |
+
"transformers": "4.52.4",
|
| 5 |
+
"pytorch": "2.6.0+cu126"
|
| 6 |
+
},
|
| 7 |
+
"prompts": {},
|
| 8 |
+
"default_prompt_name": null,
|
| 9 |
+
"similarity_fn_name": "cosine"
|
| 10 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5223a9f4c56d5391436409acb852c6851f42821b2280087e759e95f39a6f2c59
|
| 3 |
+
size 437951328
|
modules.json
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"idx": 0,
|
| 4 |
+
"name": "0",
|
| 5 |
+
"path": "",
|
| 6 |
+
"type": "sentence_transformers.models.Transformer"
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"idx": 1,
|
| 10 |
+
"name": "1",
|
| 11 |
+
"path": "1_Pooling",
|
| 12 |
+
"type": "sentence_transformers.models.Pooling"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"idx": 2,
|
| 16 |
+
"name": "2",
|
| 17 |
+
"path": "2_Normalize",
|
| 18 |
+
"type": "sentence_transformers.models.Normalize"
|
| 19 |
+
}
|
| 20 |
+
]
|
sentence_bert_config.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"max_seq_length": 512,
|
| 3 |
+
"do_lower_case": false
|
| 4 |
+
}
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": true,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": true,
|
| 47 |
+
"extra_special_tokens": {},
|
| 48 |
+
"mask_token": "[MASK]",
|
| 49 |
+
"max_length": 512,
|
| 50 |
+
"model_max_length": 512,
|
| 51 |
+
"pad_to_multiple_of": null,
|
| 52 |
+
"pad_token": "[PAD]",
|
| 53 |
+
"pad_token_type_id": 0,
|
| 54 |
+
"padding_side": "right",
|
| 55 |
+
"sep_token": "[SEP]",
|
| 56 |
+
"stride": 0,
|
| 57 |
+
"strip_accents": null,
|
| 58 |
+
"tokenize_chinese_chars": true,
|
| 59 |
+
"tokenizer_class": "BertTokenizer",
|
| 60 |
+
"truncation_side": "right",
|
| 61 |
+
"truncation_strategy": "longest_first",
|
| 62 |
+
"unk_token": "[UNK]"
|
| 63 |
+
}
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|