---
license: apache-2.0
tags:
- sentence-transformers
- cross-encoder
- reranker
- modchembert
- cheminformatics
- smiles
- generated_from_trainer
- dataset_size:3269544
- loss:MultipleNegativesRankingLoss
base_model: Derify/ModChemBERT-IR-BASE
pipeline_tag: text-ranking
library_name: sentence-transformers
metrics:
- map
- mrr@10
- ndcg@10
co2_eq_emissions:
emissions: 2511.7170953063833
energy_consumed: 12.236213576221271
source: codecarbon
training_type: fine-tuning
on_cloud: false
cpu_model: AMD Ryzen 7 3700X 8-Core Processor
ram_total_size: 62.69877243041992
hours_used: 19.958
hardware_used: 2 x NVIDIA GeForce RTX 3090
model-index:
- name: 'Derify/ChemRanker-alpha-sim'
results:
- task:
type: cross-encoder-reranking
name: Cross Encoder Reranking
dataset:
name: Unknown
type: unknown
metrics:
- type: map
value: 0.4322642780868116
name: Map
- type: mrr@10
value: 0.6974610113213638
name: Mrr@10
- type: ndcg@10
value: 0.7033573942360365
name: Ndcg@10
---
# Derify/ChemRanker-alpha-sim
This [Cross Encoder](https://www.sbert.net/docs/cross_encoder/usage/usage.html) is finetuned from [Derify/ModChemBERT-IR-BASE](https://huggingface.co/Derify/ModChemBERT-IR-BASE) using hard-negative triplets derived from [Derify/pubchem_10m_genmol_similarity](https://huggingface.co/datasets/Derify/pubchem_10m_genmol_similarity). Positive SMILES pairs are first filtered by quality and similarity constraints, then reduced to one strongest positive target per anchor molecule to create a high-signal training set for reranking. The model computes relevance scores for pairs of SMILES strings, enabling SMILES reranking and molecular semantic search.
For this variant, the positive selection objective is pure similarity ranking where each anchor keeps the highest-similarity candidate after filtering, rather than using a QED+similarity composite score. The quality stage uses strict inequality filtering (`QED > 0.85`, `similarity > 0.5`, with similarity also bounded below 1.0), and then keeps the top-scoring pair per anchor molecule.
Hard negatives are mined with [Sentence Transformers](https://www.sbert.net/) using [Derify/ChemMRL-beta](https://huggingface.co/Derify/ChemMRL-beta) as the teacher model and a TopK-PercPos-style margin setting based on [NV-Retriever](https://arxiv.org/abs/2407.15831), with `relative_margin=0.05` and `max_negative_score_threshold = pos_score * percentage_margin`. Training uses triplet-format samples with 5 mined negatives per anchor-positive pair and optimizes a multiple-negatives ranking objective, while reranking evaluation uses n-tuple samples with 30 mined negatives per query.
## Model Details
### Model Description
- **Model Type:** Cross Encoder
- **Base model:** [Derify/ModChemBERT-IR-BASE](https://huggingface.co/Derify/ModChemBERT-IR-BASE)
- **Maximum Sequence Length:** 512 tokens
- **Number of Output Labels:** 1 label
- **Training Dataset:**
- [Derify/pubchem_10m_genmol_similarity](https://huggingface.co/datasets/Derify/pubchem_10m_genmol_similarity) Mined Hard Negatives
- **License:** apache-2.0
### Model Sources
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
- **Documentation:** [Cross Encoder Documentation](https://www.sbert.net/docs/cross_encoder/usage/usage.html)
- **Repository:** [Sentence Transformers on GitHub](https://github.com/huggingface/sentence-transformers)
- **Hugging Face:** [Cross Encoders on Hugging Face](https://huggingface.co/models?library=sentence-transformers&other=cross-encoder)
## Usage
### Direct Usage (Sentence Transformers)
First install the Transformers and Sentence Transformers libraries:
```bash
pip install -U "transformers>=4.57.1,<5.0.0"
pip install -U sentence-transformers
```
Then you can load this model and run inference.
```python
from sentence_transformers import CrossEncoder
# Download from the 🤗 Hub
model = CrossEncoder("Derify/ChemRanker-alpha-sim")
# Get scores for pairs of texts
pairs = [
['c1snnc1C[NH2+]Cc1cc2c(s1)CCC2', 'c1snnc1CCC[NH2+]Cc1cc2c(s1)CCC2'],
['c1sc2c(c1-c1nc(C3CCOC3)no1)CCCC2', 'O=CCc1noc(-c2csc3c2CCCC3)n1'],
['c1sc(C[NH2+]C2CC2)nc1C[NH+]1CCN2CCCC2C1', 'FC(F)[NH2+]Cc1nc(C[NH+]2CCN3CCCC3C2)cs1'],
['c1sc(CC[NH+]2CCOCC2)nc1C[NH2+]C1CC1', 'CCc1nc(C[NH2+]C2CC2)cs1'],
['c1sc(CC2CCC[NH2+]2)nc1C1CCCO1', 'c1sc(CC2CCC[NH2+]2)nc1C1CCCC1'],
]
scores = model.predict(pairs)
print(scores.shape)
# (5,)
# Or rank different texts based on similarity to a single text
ranks = model.rank(
'c1snnc1C[NH2+]Cc1cc2c(s1)CCC2',
[
'c1snnc1CCC[NH2+]Cc1cc2c(s1)CCC2',
'O=CCc1noc(-c2csc3c2CCCC3)n1',
'FC(F)[NH2+]Cc1nc(C[NH+]2CCN3CCCC3C2)cs1',
'CCc1nc(C[NH2+]C2CC2)cs1',
'c1sc(CC2CCC[NH2+]2)nc1C1CCCC1',
]
)
# [{'corpus_id': ..., 'score': ...}, {'corpus_id': ..., 'score': ...}, ...]
```
## Evaluation
### Metrics
#### Cross Encoder Reranking
* Evaluated with [CrossEncoderRerankingEvaluator](https://sbert.net/docs/package_reference/cross_encoder/evaluation.html#sentence_transformers.cross_encoder.evaluation.CrossEncoderRerankingEvaluator) with these parameters:
```json
{
"at_k": 10
}
```
| Metric | Value |
| :---------- | :--------- |
| map | 0.4323 |
| mrr@10 | 0.6975 |
| **ndcg@10** | **0.7034** |
## Training Details
### Training Dataset
#### GenMol Similarity Hard Negatives
* Dataset: GenMol Similarity Hard Negatives
* Size: 3,269,544 training samples
* Columns: smiles_a, smiles_b, and negative
* Approximate statistics based on the first 1000 samples:
| | smiles_a | smiles_b | negative |
| :------ | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- |
| type | string | string | string |
| details |
c1sc2cc3c(cc2c1CC[NH2+]C1CC1)OCCO3 | FC(F)(F)[NH2+]CCc1csc2cc3c(cc12)OCCO3 | [NH3+]CCCc1cc2c(cc1C1CC1)OCO2 |
| c1sc2cc3c(cc2c1CC[NH2+]C1CC1)OCCO3 | FC(F)(F)[NH2+]CCc1csc2cc3c(cc12)OCCO3 | COc1cc2c(cc1C[NH2+]C1CCC1)OCO2 |
| c1sc2cc3c(cc2c1CC[NH2+]C1CC1)OCCO3 | FC(F)(F)[NH2+]CCc1csc2cc3c(cc12)OCCO3 | O=c1[nH]c2cc3c(cc2cc1CNC1CCCCC1)OCCO3 |
* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/cross_encoder/losses.html#multiplenegativesrankingloss) with these parameters:
```json
{
"scale": 10.0,
"num_negatives": 4,
"activation_fn": "torch.nn.modules.activation.Sigmoid"
}
```
### Evaluation Dataset
#### GenMol Similarity Hard Negatives
* Dataset: GenMol Similarity Hard Negatives
* Size: 165,968 evaluation samples
* Columns: smiles_a, smiles_b, negative_1, negative_2, negative_3, negative_4, negative_5, negative_6, negative_7, negative_8, negative_9, negative_10, negative_11, negative_12, negative_13, negative_14, negative_15, negative_16, negative_17, negative_18, negative_19, negative_20, negative_21, negative_22, negative_23, negative_24, negative_25, negative_26, negative_27, negative_28, negative_29, and negative_30
* Approximate statistics based on the first 1000 samples:
| | smiles_a | smiles_b | negative_1 | negative_2 | negative_3 | negative_4 | negative_5 | negative_6 | negative_7 | negative_8 | negative_9 | negative_10 | negative_11 | negative_12 | negative_13 | negative_14 | negative_15 | negative_16 | negative_17 | negative_18 | negative_19 | negative_20 | negative_21 | negative_22 | negative_23 | negative_24 | negative_25 | negative_26 | negative_27 | negative_28 | negative_29 | negative_30 |
| :------ | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :-------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :-------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :-------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- | :--------------------------------------------------------------------------------------------- |
| type | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string | string |
| details | c1snnc1C[NH2+]Cc1cc2c(s1)CCC2 | c1snnc1CCC[NH2+]Cc1cc2c(s1)CCC2 | c1snnc1CCC[NH2+]Cc1cc2c(s1)CCC2 | Cn1cc(C[NH2+]Cc2cc3c(s2)CCC3)nn1 | Cn1cc(CC[NH2+]Cc2cc3c(s2)CCC3)nn1 | Cc1cc(C[NH2+]Cc2csnn2)sc1C | NC(=O)c1csc(C[NH2+]Cc2cc3c(s2)CCC3)c1 | Cc1cc(CC[NH2+]Cc2csnn2)sc1C | Ic1ccc(C[NH2+]Cc2cc3c(s2)CCC3)o1 | Cc1cc(C[NH2+]CCCCc2cc3c(s2)CCC3)c(C)s1 | c1ccc(C[NH2+]Cc2cc3c(s2)CCC3)cc1 | c1ncc(C[NH2+]Cc2csnn2)s1 | FC(F)c1csc(C[NH2+]Cc2cc3c(s2)CCC3)c1 | c1c(C[NH2+]CC2CC2)sc2c1CSCC2 | N#Cc1cc(F)cc(C[NH2+]Cc2cc3c(s2)CCC3)c1 | c1cc(C[NH2+]Cc2nc3c(s2)CCC3)no1 | CCc1ccc(C[NH2+]Cc2csnn2)s1 | NCc1csc(NCc2cc3c(s2)CCC3)n1 | C[NH+](Cc1cscn1)Cc1nnc(-c2cc3c(s2)CCCC3)o1 | Fc1cc(C[NH2+]Cc2cc3c(s2)CCC3)ccc1Br | FC(F)(F)C[NH2+]Cc1cc2c(s1)CCSC2 | c1cc(C[NH2+]Cc2cc3c(s2)CCC3)c[nH]1 | Cc1cc(C)c(CC[NH2+]Cc2cc3c(s2)CCC3)c(C)c1 | Oc1ccc(C[NH2+]Cc2cc3c(s2)CCC3)cc1Br | O=C([O-])c1ccc(CC[NH2+]Cc2cc3c(s2)CCC3)s1 | c1c(C[NH2+]CC2CCCC2)sc2c1CCC2 | O=C([O-])c1ccc(C[NH2+]Cc2cc3c(s2)CCC3)s1 | COc1cc(C)cc(C[NH2+]Cc2cc3c(s2)CCC3)c1 | PSc1ccc(C[NH2+]Cc2csnn2)s1 | CCc1cnc(C[NH2+]Cc2csnn2)s1 | Clc1cc(C[NH2+]Cc2cc3c(s2)CCC3)ccc1Br | c1c(C[NH2+]CC2CC2)sc2c1CCCCC2 |
| c1sc2c(c1-c1nc(C3CCOC3)no1)CCCC2 | O=CCc1noc(-c2csc3c2CCCC3)n1 | Nc1sc2c(c1-c1nc(C3CCOC3)no1)CCCC2 | Nc1sc2c(c1-c1nc(C3CCC3)no1)CCCC2 | c1c(-c2nc(C3CCCNC3)no2)sc2c1CCCCCC2 | Nc1sccc1-c1nc(C2CCCOC2)no1 | Nc1sc2c(c1-c1nc(C3CCCO3)no1)CCCC2 | Cc1csc(-c2nc(C3CCOCC3)no2)c1N | Cc1oc2c(c1-c1nc(C3CCOC3)no1)C(=O)CCC2 | c1c(-c2nc(C3C[NH2+]CCO3)no2)sc2c1CCCCC2 | O=C([O-])Nc1sc2c(c1-c1nc(C3CC3)no1)CCCC2 | c1cc2c(s1)CCCC2c1nc(C2CC2)no1 | CC(=O)N1CCCC(c2noc(-c3cc4c(s3)CCCCCC4)n2)C1 | Cc1cc(-c2nc([C@@H]3CCOC3)no2)c(N)s1 | c1cc2c(nc1-c1noc(C3CCCOC3)n1)CCCC2 | Nc1sccc1-c1nc(C2CCCC2)no1 | c1cc2c(nc1-c1noc(C3CCOCC3)n1)CCCC2 | [NH3+]C(c1noc(-c2cc3c(s2)CCCC3)n1)C1CC1 | c1cc2c(c(-c3nc(C4CCOCC4)no3)c1)CCCN2 | c1c(-c2nc(C3CC3)no2)nn2c1CCCC2 | CN1CC(c2noc(-c3cc4c(s3)CCCC4)n2)CC1=O | Oc1c(-c2nc(C3CCC(F)(F)C3)no2)ccc2c1CCCC2 | O=CCc1noc(-c2csc3c2CCCC3)n1 | Cc1cc(=O)c(-c2noc(C3CCCOC3)n2)c2n1CCC2 | O=C([O-])CNc1sc2c(c1-c1nc(C3CC3)no1)CCCC2 | c1cc(-c2noc(C3CCCOC3)n2)cs1 | Cn1nc(-c2nc(C3CCCO3)no2)c2c1CCCC2 | O=C(Nc1sc2c(c1-c1nc(C3CC3)no1)COCC2)C1=CCCCC1 | Cc1cscc1-c1noc(C2CCOCC2)n1 | CC1(C)CCCc2sc(N)c(-c3nc(C4CC4)no3)c21 | Clc1cc2c(c(-c3nc(C4CCOC4)no3)c1)OCC2 | Nc1sc2c(c1-c1nnc(C3CC3)o1)CCCC2 |
| c1sc(C[NH2+]C2CC2)nc1C[NH+]1CCN2CCCC2C1 | FC(F)[NH2+]Cc1nc(C[NH+]2CCN3CCCC3C2)cs1 | FC(F)[NH2+]Cc1nc(C[NH+]2CCN3CCCC3C2)cs1 | CC(C)[NH2+]Cc1nc(C[NH+]2CCC3CCCCC3C2)cs1 | CN1C2CCC1C[NH+](Cc1csc(C[NH3+])n1)CC2 | Nc1nc(CC[NH+]2CCCN3CCCC3C2)cs1 | NCc1nc(C[NH+]2CCCC3CCCCC32)cs1 | CC1C[NH+](Cc2csc(C[NH2+]C3CC3)n2)CCN1C | Oc1csc(CN2CCCC3C[NH2+]CC32)n1 | CCc1nc(C[NH+]2CCCC3CCCCC32)cs1 | C[NH2+]Cc1csc(N2CC[NH+]3CCCC3C2)n1 | [NH3+]Cc1nc(C[NH+]2CCC3CCCCC32)cs1 | CC1CN2CCCCC2C[NH+]1Cc1csc(CC[NH3+])n1 | CCCc1nc(CN2CCCC2C2CCC[NH2+]2)cs1 | ClCCc1nc(CN2CCCC2C2CCC[NH2+]2)cs1 | c1cc(C[NH2+]C2CC2)c(C[NH+]2CCN3CCCCC3C2)o1 | O=C(Cc1nc(CCl)cs1)N1CCC[NH+]2CCCC2C1 | CC[NH2+]Cc1csc(N2CCC3C(CCC[NH+]3C)C2)n1 | c1sc(C[NH2+]C2CC2)nc1C[NH+]1CCCCC1 | [NH3+]Cc1nc(C[NH+]2CCCC2C2CCCC2)cs1 | Cc1csc(C[NH+]2CCC3C[NH2+]CC3C2)n1 | c1cc(C[NH+]2CCCN3CCCC3C2)nc(C2CC2)n1 | Cc1ccsc1C[NH2+]CCN1CCN2CCCC2C1 | c1sc(C[NH2+]C2CCCC2)nc1C[NH+]1CCCCC1 | Brc1csc(C[NH2+]CCN2CCN3CCCCC3C2)c1 | Cc1nc(CCC[NH2+]C2CCN3CCCCC23)cs1 | CCOC(=O)c1nc(CN2CC3CCC[NH2+]C3C2)cs1 | CCCC(=O)c1nc(CN2CC3CCC[NH2+]C3C2)cs1 | CC(C)(C)c1csc(CN2CCC[NH2+]C(C3CC3)C2)n1 | COCc1nc(CN2CCC([NH3+])C2)cs1 | CCC[NH2+]Cc1nc(C[NH+]2CC3CCC2C3)cs1 | CCC1CN2CCCC2C[NH+]1CCc1csc(C)n1 |
* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/cross_encoder/losses.html#multiplenegativesrankingloss) with these parameters:
```json
{
"scale": 10.0,
"num_negatives": 4,
"activation_fn": "torch.nn.modules.activation.Sigmoid"
}
```
### Training Hyperparameters
#### Non-Default Hyperparameters
- `eval_strategy`: epoch
- `per_device_train_batch_size`: 256
- `per_device_eval_batch_size`: 256
- `torch_empty_cache_steps`: 1000
- `learning_rate`: 3e-05
- `weight_decay`: 1e-05
- `max_grad_norm`: None
- `lr_scheduler_type`: warmup_stable_decay
- `lr_scheduler_kwargs`: {'num_decay_steps': 6385, 'warmup_type': 'linear', 'decay_type': '1-sqrt'}
- `warmup_steps`: 6385
- `seed`: 12
- `data_seed`: 24681357
- `bf16`: True
- `bf16_full_eval`: True
- `tf32`: True
- `dataloader_num_workers`: 8
- `dataloader_prefetch_factor`: 2
- `load_best_model_at_end`: True
- `optim`: stable_adamw
- `optim_args`: decouple_lr=True,max_lr=3e-05
- `dataloader_persistent_workers`: True
- `resume_from_checkpoint`: False
- `gradient_checkpointing`: True
- `torch_compile`: True
- `torch_compile_backend`: inductor
- `torch_compile_mode`: max-autotune
- `eval_on_start`: True
- `batch_sampler`: no_duplicates
#### All Hyperparameters