mengyouHF commited on
Commit
b91f835
·
verified ·
1 Parent(s): ce9001a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +142 -142
README.md CHANGED
@@ -1,143 +1,143 @@
1
- ---
2
- language: []
3
- tags:
4
- - sentence-transformers
5
- - sentence-similarity
6
- - feature-extraction
7
- widget: []
8
- pipeline_tag: sentence-similarity
9
- library_name: sentence-transformers
10
- ---
11
-
12
- # SentenceTransformer
13
-
14
- This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
15
-
16
- ## Model Details
17
-
18
- ### Model Description
19
- - **Model Type:** Sentence Transformer
20
- <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
21
- - **Maximum Sequence Length:** 512 tokens
22
- - **Output Dimensionality:** 768 tokens
23
- - **Similarity Function:** Cosine Similarity
24
- <!-- - **Training Dataset:** Unknown -->
25
- <!-- - **Language:** Unknown -->
26
- <!-- - **License:** Unknown -->
27
-
28
- ### Model Sources
29
-
30
- - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
31
- - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
32
- - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
33
-
34
- ### Full Model Architecture
35
-
36
- ```
37
- SentenceTransformer(
38
- (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
39
- (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
40
- (2): Normalize()
41
- )
42
- ```
43
-
44
- ## Usage
45
-
46
- ### Direct Usage (Sentence Transformers)
47
-
48
- First install the Sentence Transformers library:
49
-
50
- ```bash
51
- pip install -U sentence-transformers
52
- ```
53
-
54
- Then you can load this model and run inference.
55
- ```python
56
- from sentence_transformers import SentenceTransformer
57
-
58
- # Download from the 🤗 Hub
59
- model = SentenceTransformer("valuesimplex-ai-lab/Fin-Retriever-base")
60
- # Run inference
61
- sentences = [
62
- 'The weather is lovely today.',
63
- "It's so sunny outside!",
64
- 'He drove to the stadium.',
65
- ]
66
- embeddings = model.encode(sentences)
67
- print(embeddings.shape)
68
- # [3, 768]
69
-
70
- # Get the similarity scores for the embeddings
71
- similarities = model.similarity(embeddings, embeddings)
72
- print(similarities.shape)
73
- # [3, 3]
74
- ```
75
-
76
- <!--
77
- ### Direct Usage (Transformers)
78
-
79
- <details><summary>Click to see the direct usage in Transformers</summary>
80
-
81
- </details>
82
- -->
83
-
84
- <!--
85
- ### Downstream Usage (Sentence Transformers)
86
-
87
- You can finetune this model on your own dataset.
88
-
89
- <details><summary>Click to expand</summary>
90
-
91
- </details>
92
- -->
93
-
94
- <!--
95
- ### Out-of-Scope Use
96
-
97
- *List how the model may foreseeably be misused and address what users ought not to do with the model.*
98
- -->
99
-
100
- <!--
101
- ## Bias, Risks and Limitations
102
-
103
- *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
104
- -->
105
-
106
- <!--
107
- ### Recommendations
108
-
109
- *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
110
- -->
111
-
112
- ## Training Details
113
-
114
- ### Framework Versions
115
- - Python: 3.10.11
116
- - Sentence Transformers: 3.0.0
117
- - Transformers: 4.39.1
118
- - PyTorch: 2.0.1+cu117
119
- - Accelerate: 0.30.1
120
- - Datasets: 3.1.0
121
- - Tokenizers: 0.15.2
122
-
123
- ## Citation
124
-
125
- ### BibTeX
126
-
127
- <!--
128
- ## Glossary
129
-
130
- *Clearly define terms in order to be accessible across audiences.*
131
- -->
132
-
133
- <!--
134
- ## Model Card Authors
135
-
136
- *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
137
- -->
138
-
139
- <!--
140
- ## Model Card Contact
141
-
142
- *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
143
  -->
 
1
+ ---
2
+ language: []
3
+ tags:
4
+ - sentence-transformers
5
+ - sentence-similarity
6
+ - feature-extraction
7
+ widget: []
8
+ pipeline_tag: sentence-similarity
9
+ library_name: sentence-transformers
10
+ ---
11
+
12
+ # SentenceTransformer
13
+
14
+ This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
15
+
16
+ ## Model Details
17
+
18
+ ### Model Description
19
+ - **Model Type:** Sentence Transformer
20
+ <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
21
+ - **Maximum Sequence Length:** 512 tokens
22
+ - **Output Dimensionality:** 768 tokens
23
+ - **Similarity Function:** Cosine Similarity
24
+ <!-- - **Training Dataset:** Unknown -->
25
+ <!-- - **Language:** Unknown -->
26
+ <!-- - **License:** Unknown -->
27
+
28
+ ### Model Sources
29
+
30
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
31
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
32
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
33
+
34
+ ### Full Model Architecture
35
+
36
+ ```
37
+ SentenceTransformer(
38
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
39
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
40
+ (2): Normalize()
41
+ )
42
+ ```
43
+
44
+ ## Usage
45
+
46
+ ### Direct Usage (Sentence Transformers)
47
+
48
+ First install the Sentence Transformers library:
49
+
50
+ ```bash
51
+ pip install -U sentence-transformers
52
+ ```
53
+
54
+ Then you can load this model and run inference.
55
+ ```python
56
+ from sentence_transformers import SentenceTransformer
57
+
58
+ # Download from the 🤗 Hub
59
+ model = SentenceTransformer("valuesimplex-ai-lab/Fin-Retriever-base")
60
+ # Run inference
61
+ sentences = [
62
+ '熵简科技是一家金融科技公司',
63
+ "熵简科技专注于ai智能化投研",
64
+ '走向ai智能化是当前投研行业发展趋势',
65
+ ]
66
+ embeddings = model.encode(sentences)
67
+ print(embeddings.shape)
68
+ # [3, 768]
69
+
70
+ # Get the similarity scores for the embeddings
71
+ similarities = model.similarity(embeddings, embeddings)
72
+ print(similarities.shape)
73
+ # [3, 3]
74
+ ```
75
+
76
+ <!--
77
+ ### Direct Usage (Transformers)
78
+
79
+ <details><summary>Click to see the direct usage in Transformers</summary>
80
+
81
+ </details>
82
+ -->
83
+
84
+ <!--
85
+ ### Downstream Usage (Sentence Transformers)
86
+
87
+ You can finetune this model on your own dataset.
88
+
89
+ <details><summary>Click to expand</summary>
90
+
91
+ </details>
92
+ -->
93
+
94
+ <!--
95
+ ### Out-of-Scope Use
96
+
97
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
98
+ -->
99
+
100
+ <!--
101
+ ## Bias, Risks and Limitations
102
+
103
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
104
+ -->
105
+
106
+ <!--
107
+ ### Recommendations
108
+
109
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
110
+ -->
111
+
112
+ ## Training Details
113
+
114
+ ### Framework Versions
115
+ - Python: 3.10.11
116
+ - Sentence Transformers: 3.0.0
117
+ - Transformers: 4.39.1
118
+ - PyTorch: 2.0.1+cu117
119
+ - Accelerate: 0.30.1
120
+ - Datasets: 3.1.0
121
+ - Tokenizers: 0.15.2
122
+
123
+ ## Citation
124
+
125
+ ### BibTeX
126
+
127
+ <!--
128
+ ## Glossary
129
+
130
+ *Clearly define terms in order to be accessible across audiences.*
131
+ -->
132
+
133
+ <!--
134
+ ## Model Card Authors
135
+
136
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
137
+ -->
138
+
139
+ <!--
140
+ ## Model Card Contact
141
+
142
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
143
  -->