davidmezzetti's picture
Upload model
f0bd247
---
pipeline_tag: sentence-similarity
tags:
- sentence-transformers
- feature-extraction
- sentence-similarity
- transformers
base_model: NeuML/bert-hash-pico
language: en
license: apache-2.0
---
# BERT Hash Pico Embeddings
This is a [BERT Hash Pico](https://hf.co/neuml/bert-hash-pico) model fined-tuned using [sentence-transformers](https://www.SBERT.net). It maps sentences & paragraphs to a 80-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
This model is an alternative to [MUVERA fixed-dimensional encoding](https://arxiv.org/abs/2405.19504) with ColBERT models. MUVERA encoding enables encoding the multi-vector outputs of ColBERT into single dense vector outputs. While this is a great step, the main issue with MUVERA is that it tends to need wide vectors to be effective (5K - 10K dimensional vectors). `bert-hash-pico-embeddings` outputs 80-dimensional vectors.
The training dataset is a subset of [this embedding training collection](https://huggingface.co/collections/sentence-transformers/embedding-model-datasets). The training workflow was a two step distillation process as follows.
- Distill embeddings from the larger [bert-hash-nano-embeddings](https://huggingface.co/neuml/bert-hash-nano-embeddings) model using this [model distillation script](https://github.com/huggingface/sentence-transformers/blob/main/examples/sentence_transformer/training/distillation/model_distillation.py) from Sentence Transformers.
- Build a distilled dataset of teacher scores using the [mixedbread-ai/mxbai-rerank-xsmall-v1](https://hf.co/mixedbread-ai/mxbai-rerank-xsmall-v1) cross-encoder for a random sample of the training dataset mentioned above.
- Further fine-tune the model on the distilled dataset using [KLDivLoss](https://github.com/huggingface/sentence-transformers/blob/main/sentence_transformers/losses/DistillKLDivLoss.py).
## Usage (txtai)
This model can be used to build embeddings databases with [txtai](https://github.com/neuml/txtai) for semantic search and/or as a knowledge source for retrieval augmented generation (RAG).
```python
import txtai
embeddings = txtai.Embeddings(
path="neuml/bert-hash-pico-embeddings",
content=True,
vectors={"trust_remote_code": True}
)
embeddings.index(documents())
# Run a query
embeddings.search("query to run")
```
## Usage (Sentence-Transformers)
Alternatively, the model can be loaded with [sentence-transformers](https://www.SBERT.net).
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer("neuml/bert-hash-pico-embeddings", trust_remote_code=True)
embeddings = model.encode(sentences)
print(embeddings)
```
## Usage (Hugging Face Transformers)
The model can also be used directly with Transformers.
```python
from transformers import AutoTokenizer, AutoModel
import torch
# Mean Pooling - Take attention mask into account for correct averaging
def meanpooling(output, mask):
embeddings = output[0] # First element of model_output contains all token embeddings
mask = mask.unsqueeze(-1).expand(embeddings.size()).float()
return torch.sum(embeddings * mask, 1) / torch.clamp(mask.sum(1), min=1e-9)
# Sentences we want sentence embeddings for
sentences = ['This is an example sentence', 'Each sentence is converted']
# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained("neuml/bert-hash-pico-embeddings", trust_remote_code=True)
model = AutoModel.from_pretrained("neuml/bert-hash-pico-embeddings", trust_remote_code=True)
# Tokenize sentences
inputs = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
# Compute token embeddings
with torch.no_grad():
output = model(**inputs)
# Perform pooling. In this case, mean pooling.
embeddings = meanpooling(output, inputs['attention_mask'])
print("Sentence embeddings:")
print(embeddings)
```
## Evaluation
The following table shows a subset of BEIR scored with the [txtai benchmarks script](https://github.com/neuml/txtai/blob/master/examples/benchmarks.py).
This evaluation is compared against the [ColBERT MUVERA](https://huggingface.co/collections/NeuML/colbert) series of models.
Scores reported are `ndcg@10` and grouped into the following three categories.
### BERT Hash Embeddings vs MUVERA
| Model | Parameters | NFCorpus | SciDocs | SciFact | Average |
|:------------------|:-----------|:---------|:---------|:--------|:--------|
| [**BERT Hash Pico Embeddings**](https://huggingface.co/neuml/bert-hash-pico-embeddings) | **0.4M** | **0.2075** | **0.0812** | **0.3912** | **0.2266** |
| [ColBERT MUVERA Pico](https://huggingface.co/neuml/colbert-muvera-pico) | 0.4M | 0.1926 | 0.0564 | 0.4424 | 0.2305 |
### BERT Hash Embeddings vs MUVERA with maxsim re-ranking of the top 100 results per MUVERA paper
| Model | Parameters | NFCorpus | SciDocs | SciFact | Average |
|:------------------|:-----------|:---------|:---------|:--------|:--------|
| [**BERT Hash Pico Embeddings**](https://huggingface.co/neuml/bert-hash-pico-embeddings) | **0.4M** | **0.2702** | **0.1104** | **0.5965** | **0.3257** |
| [ColBERT MUVERA Pico](https://huggingface.co/neuml/colbert-muvera-pico) | 0.4M | 0.2821 | 0.1004 | 0.6090 | 0.3305 |
### Compare to other models
| Model | Parameters | NFCorpus | SciDocs | SciFact | Average |
|:------------------|:-----------|:---------|:---------|:--------|:--------|
| [ColBERT MUVERA Pico (full multi-vector maxsim)](https://huggingface.co/neuml/colbert-muvera-pico) | 0.4M | 0.3005 | 0.1117 | 0.6452 | 0.3525 |
| [all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) | 22.7M | 0.3089 | 0.2164 | 0.6527 | 0.3927 |
| [mxbai-embed-xsmall-v1](https://huggingface.co/mixedbread-ai/mxbai-embed-xsmall-v1) | 24.1M | 0.3186 | 0.2155 | 0.6598 | 0.3980 |
In analyzing the results, `bert-hash-pico-embeddings` scores slightly worse than MUVERA with `colbert-muvera-pico`. Comparing the standard MUVERA output of `10240` vs `80` dimensions, `10K` standard F32 vectors needs `400 MB` of storage vs `3.2 MB`
Keeping in mind this is only a `448K` parameter model, the performance is still impressive at only `~2%` of the number of parameters of popular small embeddings models.
While this isn't a state of the art model, it's an extremely competitive method for building vectors on edge and low resource devices.
## Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'BertHashModel'})
(1): Pooling({'word_embedding_dimension': 80, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)
```
## More Information
Read more about this model and how it was built in [this article](https://hf.co/blog/neuml/bert-hash-embeddings).