---
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- dense
- generated_from_trainer
- dataset_size:3156
- loss:CosineSimilarityLoss
base_model: Qwen/Qwen3-Embedding-8B
widget:
- source_sentence: The goal nonvar('2019-12-31') succeeded, indicating that the date
2019‑12‑31 is instantiated.
sentences:
- 'Succeeded: day_to_stamp("2019-12-31",1577836800.0)'
- 'Failed: s1_c_iii(22895,3538)'
- 'Failed: agent_(alice_dies,bob)'
- source_sentence: The date September 1, 2015 corresponds to the Unix timestamp 1441152000.0.
sentences:
- 'Failed: son_(_20022)'
- 'Succeeded: day_to_stamp("2015-09-01",1441152000.0)'
- 'Failed: s1_c_iv(102268,27225)'
- source_sentence: Alice is the employer of Bob.
sentences:
- 'Succeeded: agent_(alice_employer,bob)'
- 'Succeeded: day_to_stamp("2019-10-10",1570752000.0)'
- 'Failed: s7703_a(alice,_18952,_18954,2016)'
- source_sentence: The first day of tax year 2014 is January 1, 2014.
sentences:
- 'Succeeded: nonvar("2019-10-10")'
- 'Succeeded: first_day_year(2018,"2018-01-01")'
- 'Succeeded: first_day_year(2014,"2014-01-01")'
- source_sentence: Under section 1(a)(iv), the tax on $164,612 of taxable income is
$44,789.
sentences:
- 'Succeeded: 2019 is 2018+1'
- 'Succeeded: var(_21490)'
- 'Succeeded: 44789 is round(35928.5+(164612-140000)*0.36)'
pipeline_tag: sentence-similarity
library_name: sentence-transformers
---
# SentenceTransformer based on Qwen/Qwen3-Embedding-8B
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Qwen/Qwen3-Embedding-8B](https://huggingface.co/Qwen/Qwen3-Embedding-8B). It maps sentences & paragraphs to a 4096-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
## Model Details
### Model Description
- **Model Type:** Sentence Transformer
- **Base model:** [Qwen/Qwen3-Embedding-8B](https://huggingface.co/Qwen/Qwen3-Embedding-8B)
- **Maximum Sequence Length:** 40960 tokens
- **Output Dimensionality:** 4096 dimensions
- **Similarity Function:** Cosine Similarity
### Model Sources
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
- **Repository:** [Sentence Transformers on GitHub](https://github.com/huggingface/sentence-transformers)
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
### Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 40960, 'do_lower_case': False, 'architecture': 'Qwen3Model'})
(1): Pooling({'word_embedding_dimension': 4096, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': True, 'include_prompt': True})
(2): Normalize()
)
```
## Usage
### Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
```bash
pip install -U sentence-transformers
```
Then you can load this model and run inference.
```python
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("DChak2000/qwen3-trace-align")
# Run inference
queries = [
"Under section 1(a)(iv), the tax on $164,612 of taxable income is $44,789.",
]
documents = [
'Succeeded: 44789 is round(35928.5+(164612-140000)*0.36)',
'Succeeded: 2019 is 2018+1',
'Succeeded: var(_21490)',
]
query_embeddings = model.encode_query(queries)
document_embeddings = model.encode_document(documents)
print(query_embeddings.shape, document_embeddings.shape)
# [1, 4096] [3, 4096]
# Get the similarity scores for the embeddings
similarities = model.similarity(query_embeddings, document_embeddings)
print(similarities)
# tensor([[0.0548, 0.3047, 0.3684]])
```
## Training Details
### Training Dataset
#### Unnamed Dataset
* Size: 3,156 training samples
* Columns: sentence_0, sentence_1, and label
* Approximate statistics based on the first 1000 samples:
| | sentence_0 | sentence_1 | label |
|:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:---------------------------------------------------------------|
| type | string | string | float |
| details |
The marriage predicate could not be satisfied. | Failed: marriage_(_19298) | 1.0 |
| The last day of the year 2018 is 2018-12-31. | Succeeded: last_day_year(2018,"2018-12-31") | 1.0 |
| The conversion of the date 2019‑11‑03 to a timestamp yielded 1572825600.0. | Succeeded: day_to_stamp("2019-11-03",1572825600.0) | 1.0 |
* Loss: [CosineSimilarityLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
```json
{
"loss_fct": "torch.nn.modules.loss.MSELoss"
}
```
### Training Hyperparameters
#### Non-Default Hyperparameters
- `multi_dataset_batch_sampler`: round_robin
#### All Hyperparameters