noahjax commited on
Commit
15c433c
·
verified ·
1 Parent(s): 358ba33

Upload fine-tuned chart reranker model

Browse files
README.md CHANGED
@@ -4,7 +4,7 @@ tags:
4
  - cross-encoder
5
  - reranker
6
  - generated_from_trainer
7
- - dataset_size:8352
8
  - loss:BinaryCrossEntropyLoss
9
  base_model: Alibaba-NLP/gte-multilingual-reranker-base
10
  pipeline_tag: text-ranking
@@ -23,10 +23,10 @@ model-index:
23
  type: validation
24
  metrics:
25
  - type: pearson
26
- value: 0.8860059576990913
27
  name: Pearson
28
  - type: spearman
29
- value: 0.8842438421497182
30
  name: Spearman
31
  ---
32
 
@@ -70,11 +70,11 @@ from sentence_transformers import CrossEncoder
70
  model = CrossEncoder("cross_encoder_model_id")
71
  # Get scores for pairs of texts
72
  pairs = [
73
- ['cas similaires entrepreneurs création entreprises apports intellectuels succès échecs', 'Title: "SNPS Overview"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "SNPS"="Synopsys, Inc.", "Overview"="Stock Overview"\nSources: S&P Global'],
74
- ['Lakers Nuggets preview', 'Title: "Los Angeles Lakers Schedule"\nCollections: NBA\nChart Type: schedule:basketball_team_v2'],
75
- ['Bitcoin performance compared to Altcoin performance in 2025', 'Title: "CBTC Overview"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "CBTC"="XTRA Bitcoin Inc.", "Overview"="Stock Overview"\nSources: S&P Global'],
76
- ['Nvidia market capitalization', 'Title: "Nvidia Market Capitalization"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "Nvidia"="NVIDIA Corporation", "Market Capitalization"="Valuation Overview"\nSources: S&P Global'],
77
- ['scope of the EU cyber resilience act', 'Title: "League of Legends European Championship Overview"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "League of Legends European Championship"="LEC, Inc.", "Overview"="Stock Overview"\nSources: S&P Global'],
78
  ]
79
  scores = model.predict(pairs)
80
  print(scores.shape)
@@ -82,13 +82,13 @@ print(scores.shape)
82
 
83
  # Or rank different texts based on similarity to a single text
84
  ranks = model.rank(
85
- 'cas similaires entrepreneurs création entreprises apports intellectuels succès échecs',
86
  [
87
- 'Title: "SNPS Overview"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "SNPS"="Synopsys, Inc.", "Overview"="Stock Overview"\nSources: S&P Global',
88
- 'Title: "Los Angeles Lakers Schedule"\nCollections: NBA\nChart Type: schedule:basketball_team_v2',
89
- 'Title: "CBTC Overview"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "CBTC"="XTRA Bitcoin Inc.", "Overview"="Stock Overview"\nSources: S&P Global',
90
- 'Title: "Nvidia Market Capitalization"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "Nvidia"="NVIDIA Corporation", "Market Capitalization"="Valuation Overview"\nSources: S&P Global',
91
- 'Title: "League of Legends European Championship Overview"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "League of Legends European Championship"="LEC, Inc.", "Overview"="Stock Overview"\nSources: S&P Global',
92
  ]
93
  )
94
  # [{'corpus_id': ..., 'score': ...}, {'corpus_id': ..., 'score': ...}, ...]
@@ -129,8 +129,8 @@ You can finetune this model on your own dataset.
129
 
130
  | Metric | Value |
131
  |:-------------|:-----------|
132
- | pearson | 0.886 |
133
- | **spearman** | **0.8842** |
134
 
135
  <!--
136
  ## Bias, Risks and Limitations
@@ -150,19 +150,19 @@ You can finetune this model on your own dataset.
150
 
151
  #### Unnamed Dataset
152
 
153
- * Size: 8,352 training samples
154
  * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
155
  * Approximate statistics based on the first 1000 samples:
156
- | | sentence_0 | sentence_1 | label |
157
- |:--------|:-----------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------|:---------------------------------------------------------------|
158
- | type | string | string | float |
159
- | details | <ul><li>min: 9 characters</li><li>mean: 45.67 characters</li><li>max: 174 characters</li></ul> | <ul><li>min: 76 characters</li><li>mean: 186.96 characters</li><li>max: 350 characters</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.47</li><li>max: 1.0</li></ul> |
160
  * Samples:
161
- | sentence_0 | sentence_1 | label |
162
- |:---------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------|
163
- | <code>cas similaires entrepreneurs création entreprises apports intellectuels succès échecs</code> | <code>Title: "SNPS Overview"<br>Collections: Companies<br>Chart Type: company:finance<br>Canonical forms: "SNPS"="Synopsys, Inc.", "Overview"="Stock Overview"<br>Sources: S&P Global</code> | <code>0.5</code> |
164
- | <code>Lakers Nuggets preview</code> | <code>Title: "Los Angeles Lakers Schedule"<br>Collections: NBA<br>Chart Type: schedule:basketball_team_v2</code> | <code>0.75</code> |
165
- | <code>Bitcoin performance compared to Altcoin performance in 2025</code> | <code>Title: "CBTC Overview"<br>Collections: Companies<br>Chart Type: company:finance<br>Canonical forms: "CBTC"="XTRA Bitcoin Inc.", "Overview"="Stock Overview"<br>Sources: S&P Global</code> | <code>0.0</code> |
166
  * Loss: [<code>BinaryCrossEntropyLoss</code>](https://sbert.net/docs/package_reference/cross_encoder/losses.html#binarycrossentropyloss) with these parameters:
167
  ```json
168
  {
@@ -308,23 +308,29 @@ You can finetune this model on your own dataset.
308
  ### Training Logs
309
  | Epoch | Step | Training Loss | validation_spearman |
310
  |:------:|:----:|:-------------:|:-------------------:|
311
- | 0.3831 | 100 | - | 0.8141 |
312
- | 0.7663 | 200 | - | 0.8486 |
313
- | 1.0 | 261 | - | 0.8624 |
314
- | 1.1494 | 300 | - | 0.8641 |
315
- | 1.5326 | 400 | - | 0.8683 |
316
- | 1.9157 | 500 | 0.4409 | 0.8728 |
317
- | 2.0 | 522 | - | 0.8732 |
318
- | 2.2989 | 600 | - | 0.8731 |
319
- | 2.6820 | 700 | - | 0.8803 |
320
- | 3.0 | 783 | - | 0.8804 |
321
- | 3.0651 | 800 | - | 0.8809 |
322
- | 3.4483 | 900 | - | 0.8800 |
323
- | 3.8314 | 1000 | 0.3641 | 0.8825 |
324
- | 4.0 | 1044 | - | 0.8836 |
325
- | 4.2146 | 1100 | - | 0.8826 |
326
- | 4.5977 | 1200 | - | 0.8821 |
327
- | 4.9808 | 1300 | - | 0.8842 |
 
 
 
 
 
 
328
 
329
 
330
  ### Framework Versions
 
4
  - cross-encoder
5
  - reranker
6
  - generated_from_trainer
7
+ - dataset_size:12349
8
  - loss:BinaryCrossEntropyLoss
9
  base_model: Alibaba-NLP/gte-multilingual-reranker-base
10
  pipeline_tag: text-ranking
 
23
  type: validation
24
  metrics:
25
  - type: pearson
26
+ value: 0.8643473065020739
27
  name: Pearson
28
  - type: spearman
29
+ value: 0.8620968090164374
30
  name: Spearman
31
  ---
32
 
 
70
  model = CrossEncoder("cross_encoder_model_id")
71
  # Get scores for pairs of texts
72
  pairs = [
73
+ ['DJ mixers compatible with Apple Music 2025', 'Title: "Music devices - radio (United States)"\nCollections: YouGov Trackers\nDatasets: YouGovTrackerValueV2\nChart Type: survey:timeseries\nSources: YouGov'],
74
+ ['current USD to PLN exchange rate', 'Title: "Conversion rate from PLN to USD"\nCollections: Foreign Exchange Rates\nDatasets: Forex\nChart Type: exchange:currency\nSources: Xignite'],
75
+ ['Aktuelle Investmenttrends 2025', 'Title: "Financial activity - next 12 months (United States)"\nCollections: YouGov Trackers\nDatasets: YouGovTrackerValueV2\nChart Type: survey:timeseries\nSources: YouGov'],
76
+ ["What are Amazon's accrued liabilities?", 'Title: "Amazon Expenses Accrued (Quarterly)"\nCollections: Companies\nDatasets: StandardIncomeStatement\nChart Type: timeseries:eav_v2\nCanonical forms: "Expenses Accrued"="accrued_expenses_total"\nSources: S&P Global'],
77
+ ["Costco's long-term lease obligations", 'Title: "Air Lease Overview"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "Air Lease"="Air Lease Corporation", "Overview"="Stock Overview"\nSources: S&P Global'],
78
  ]
79
  scores = model.predict(pairs)
80
  print(scores.shape)
 
82
 
83
  # Or rank different texts based on similarity to a single text
84
  ranks = model.rank(
85
+ 'DJ mixers compatible with Apple Music 2025',
86
  [
87
+ 'Title: "Music devices - radio (United States)"\nCollections: YouGov Trackers\nDatasets: YouGovTrackerValueV2\nChart Type: survey:timeseries\nSources: YouGov',
88
+ 'Title: "Conversion rate from PLN to USD"\nCollections: Foreign Exchange Rates\nDatasets: Forex\nChart Type: exchange:currency\nSources: Xignite',
89
+ 'Title: "Financial activity - next 12 months (United States)"\nCollections: YouGov Trackers\nDatasets: YouGovTrackerValueV2\nChart Type: survey:timeseries\nSources: YouGov',
90
+ 'Title: "Amazon Expenses Accrued (Quarterly)"\nCollections: Companies\nDatasets: StandardIncomeStatement\nChart Type: timeseries:eav_v2\nCanonical forms: "Expenses Accrued"="accrued_expenses_total"\nSources: S&P Global',
91
+ 'Title: "Air Lease Overview"\nCollections: Companies\nChart Type: company:finance\nCanonical forms: "Air Lease"="Air Lease Corporation", "Overview"="Stock Overview"\nSources: S&P Global',
92
  ]
93
  )
94
  # [{'corpus_id': ..., 'score': ...}, {'corpus_id': ..., 'score': ...}, ...]
 
129
 
130
  | Metric | Value |
131
  |:-------------|:-----------|
132
+ | pearson | 0.8643 |
133
+ | **spearman** | **0.8621** |
134
 
135
  <!--
136
  ## Bias, Risks and Limitations
 
150
 
151
  #### Unnamed Dataset
152
 
153
+ * Size: 12,349 training samples
154
  * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
155
  * Approximate statistics based on the first 1000 samples:
156
+ | | sentence_0 | sentence_1 | label |
157
+ |:--------|:-----------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------|:---------------------------------------------------------------|
158
+ | type | string | string | float |
159
+ | details | <ul><li>min: 5 characters</li><li>mean: 46.81 characters</li><li>max: 123 characters</li></ul> | <ul><li>min: 77 characters</li><li>mean: 182.4 characters</li><li>max: 495 characters</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.48</li><li>max: 1.0</li></ul> |
160
  * Samples:
161
+ | sentence_0 | sentence_1 | label |
162
+ |:--------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------|
163
+ | <code>DJ mixers compatible with Apple Music 2025</code> | <code>Title: "Music devices - radio (United States)"<br>Collections: YouGov Trackers<br>Datasets: YouGovTrackerValueV2<br>Chart Type: survey:timeseries<br>Sources: YouGov</code> | <code>0.25</code> |
164
+ | <code>current USD to PLN exchange rate</code> | <code>Title: "Conversion rate from PLN to USD"<br>Collections: Foreign Exchange Rates<br>Datasets: Forex<br>Chart Type: exchange:currency<br>Sources: Xignite</code> | <code>0.75</code> |
165
+ | <code>Aktuelle Investmenttrends 2025</code> | <code>Title: "Financial activity - next 12 months (United States)"<br>Collections: YouGov Trackers<br>Datasets: YouGovTrackerValueV2<br>Chart Type: survey:timeseries<br>Sources: YouGov</code> | <code>0.75</code> |
166
  * Loss: [<code>BinaryCrossEntropyLoss</code>](https://sbert.net/docs/package_reference/cross_encoder/losses.html#binarycrossentropyloss) with these parameters:
167
  ```json
168
  {
 
308
  ### Training Logs
309
  | Epoch | Step | Training Loss | validation_spearman |
310
  |:------:|:----:|:-------------:|:-------------------:|
311
+ | 0.2591 | 100 | - | 0.7835 |
312
+ | 0.5181 | 200 | - | 0.8161 |
313
+ | 0.7772 | 300 | - | 0.8369 |
314
+ | 1.0 | 386 | - | 0.8392 |
315
+ | 1.0363 | 400 | - | 0.8442 |
316
+ | 1.2953 | 500 | 0.47 | 0.8475 |
317
+ | 1.5544 | 600 | - | 0.8533 |
318
+ | 1.8135 | 700 | - | 0.8544 |
319
+ | 2.0 | 772 | - | 0.8579 |
320
+ | 2.0725 | 800 | - | 0.8585 |
321
+ | 2.3316 | 900 | - | 0.8548 |
322
+ | 2.5907 | 1000 | 0.3926 | 0.8577 |
323
+ | 2.8497 | 1100 | - | 0.8569 |
324
+ | 3.0 | 1158 | - | 0.8607 |
325
+ | 3.1088 | 1200 | - | 0.8573 |
326
+ | 3.3679 | 1300 | - | 0.8614 |
327
+ | 3.6269 | 1400 | - | 0.8594 |
328
+ | 3.8860 | 1500 | 0.3602 | 0.8591 |
329
+ | 4.0 | 1544 | - | 0.8596 |
330
+ | 4.1451 | 1600 | - | 0.8611 |
331
+ | 4.4041 | 1700 | - | 0.8619 |
332
+ | 4.6632 | 1800 | - | 0.8618 |
333
+ | 4.9223 | 1900 | - | 0.8621 |
334
 
335
 
336
  ### Framework Versions
eval/CrossEncoderCorrelationEvaluator_validation_results.csv CHANGED
@@ -1,6 +1,6 @@
1
  epoch,steps,Pearson_Correlation,Spearman_Correlation
2
- 1.0,261,0.8664252648440269,0.8624316344419658
3
- 2.0,522,0.876051000065144,0.8731512392307483
4
- 3.0,783,0.8816187173145336,0.8804230133811932
5
- 4.0,1044,0.8854817235935349,0.8835956509201133
6
- 5.0,1305,0.8859959671748322,0.8842324890429593
 
1
  epoch,steps,Pearson_Correlation,Spearman_Correlation
2
+ 1.0,386,0.837652001216976,0.8392216545657446
3
+ 2.0,772,0.8587273733508869,0.8579410042452386
4
+ 3.0,1158,0.8640762414247022,0.8607296855959887
5
+ 4.0,1544,0.8618951242992149,0.8596087038742567
6
+ 5.0,1930,0.8643015819365643,0.8620805071202834
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb79ac26ead9212291d08a2777f15be6061a36463e79e6fc16768d9a6756bf44
3
  size 1223854204
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6485ea4ee6cc3859ccff14222884fed3c2f23e283906df217bd6bd60d8377bff
3
  size 1223854204
training_info.txt CHANGED
@@ -1,5 +1,5 @@
1
  Base Model: Alibaba-NLP/gte-multilingual-reranker-base
2
- Training Samples: 8352
3
  Epochs: 5
4
  Batch Size: 32
5
  Learning Rate: 2e-05
 
1
  Base Model: Alibaba-NLP/gte-multilingual-reranker-base
2
+ Training Samples: 12349
3
  Epochs: 5
4
  Batch Size: 32
5
  Learning Rate: 2e-05