Parveshiiii commited on
Commit
f3cdb92
·
verified ·
1 Parent(s): a9b94db

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -149
README.md CHANGED
@@ -1,149 +0,0 @@
1
- ---
2
- tags:
3
- - sentence-transformers
4
- - sentence-similarity
5
- - feature-extraction
6
- - dense
7
- base_model: FacebookAI/xlm-roberta-large
8
- pipeline_tag: sentence-similarity
9
- library_name: sentence-transformers
10
- ---
11
- # Note: This is simple cls pooling embedding model derived from xlm backbone
12
-
13
-
14
- # this is generated by sentence-tranformers lib
15
-
16
- # SentenceTransformer based on FacebookAI/xlm-roberta-large
17
-
18
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large). It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
19
-
20
-
21
- ## Model Details
22
-
23
- ### Model Description
24
- - **Model Type:** Sentence Transformer
25
- - **Base model:** [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) <!-- at revision c23d21b0620b635a76227c604d44e43a9f0ee389 -->
26
- - **Maximum Sequence Length:** 8192 tokens
27
- - **Output Dimensionality:** 1024 dimensions
28
- - **Similarity Function:** Cosine Similarity
29
- <!-- - **Training Dataset:** Unknown -->
30
- <!-- - **Language:** Unknown -->
31
- <!-- - **License:** Unknown -->
32
-
33
- ### Model Sources
34
-
35
- - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
36
- - **Repository:** [Sentence Transformers on GitHub](https://github.com/huggingface/sentence-transformers)
37
- - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
38
-
39
- ### Full Model Architecture
40
-
41
- ```
42
- SentenceTransformer(
43
- (0): Transformer({'max_seq_length': 8192, 'do_lower_case': False, 'architecture': 'XLMRobertaModel'})
44
- (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
45
- )
46
- ```
47
-
48
- ## Usage
49
-
50
- ### Direct Usage (Sentence Transformers)
51
-
52
- First install the Sentence Transformers library:
53
-
54
- ```bash
55
- pip install -U sentence-transformers
56
- ```
57
-
58
- Then you can load this model and run inference.
59
- ```python
60
- from sentence_transformers import SentenceTransformer
61
-
62
- # Download from the 🤗 Hub
63
- model = SentenceTransformer("sentence_transformers_model_id")
64
- # Run inference
65
- sentences = [
66
- 'The weather is lovely today.',
67
- "It's so sunny outside!",
68
- 'He drove to the stadium.',
69
- ]
70
- embeddings = model.encode(sentences)
71
- print(embeddings.shape)
72
- # [3, 1024]
73
-
74
- # Get the similarity scores for the embeddings
75
- similarities = model.similarity(embeddings, embeddings)
76
- print(similarities)
77
- # tensor([[1.0000, 0.9982, 0.9974],
78
- # [0.9982, 1.0000, 0.9984],
79
- # [0.9974, 0.9984, 1.0000]])
80
- ```
81
-
82
- <!--
83
- ### Direct Usage (Transformers)
84
-
85
- <details><summary>Click to see the direct usage in Transformers</summary>
86
-
87
- </details>
88
- -->
89
-
90
- <!--
91
- ### Downstream Usage (Sentence Transformers)
92
-
93
- You can finetune this model on your own dataset.
94
-
95
- <details><summary>Click to expand</summary>
96
-
97
- </details>
98
- -->
99
-
100
- <!--
101
- ### Out-of-Scope Use
102
-
103
- *List how the model may foreseeably be misused and address what users ought not to do with the model.*
104
- -->
105
-
106
- <!--
107
- ## Bias, Risks and Limitations
108
-
109
- *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
110
- -->
111
-
112
- <!--
113
- ### Recommendations
114
-
115
- *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
116
- -->
117
-
118
- ## Training Details
119
-
120
- ### Framework Versions
121
- - Python: 3.12.12
122
- - Sentence Transformers: 5.2.0
123
- - Transformers: 4.57.6
124
- - PyTorch: 2.9.0+cu126
125
- - Accelerate: 1.12.0
126
- - Datasets: 4.0.0
127
- - Tokenizers: 0.22.2
128
-
129
- ## Citation
130
-
131
- ### BibTeX
132
-
133
- <!--
134
- ## Glossary
135
-
136
- *Clearly define terms in order to be accessible across audiences.*
137
- -->
138
-
139
- <!--
140
- ## Model Card Authors
141
-
142
- *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
143
- -->
144
-
145
- <!--
146
- ## Model Card Contact
147
-
148
- *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
149
- -->