Upload folder using huggingface_hub
Browse files
README.md
CHANGED
|
@@ -27,7 +27,7 @@ Finally, for each respective quantisation level, `llama.cpp`'s `llama-quantize`
|
|
| 27 |
|
| 28 |
## Quantisations
|
| 29 |
|
| 30 |
-
To help visualise the difference in model quantisation (i.e. level of retained fidelity), the image below shows the cosine similarity scores for each
|
| 31 |
|
| 32 |
The underlying [base dataset](https://huggingface.co/datasets/sentence-transformers/stsb) was sampled to 1000 records with a unbiased similarity score distribution. Using the various quantisation levels of this model, embeddings were created for `sentence1` and `sentence2`. Finally, a cosine similarity score was calculated across the two embeddings, and plotted on the graph.
|
| 33 |
|
|
|
|
| 27 |
|
| 28 |
## Quantisations
|
| 29 |
|
| 30 |
+
To help visualise the difference in model quantisation (i.e. level of retained fidelity), the image below shows the cosine similarity scores for each quantisation, baselined against the 32-bit base model. It can be observed that lower fidelity yields a wider scatter in scores, relative to the 32-bit model.
|
| 31 |
|
| 32 |
The underlying [base dataset](https://huggingface.co/datasets/sentence-transformers/stsb) was sampled to 1000 records with a unbiased similarity score distribution. Using the various quantisation levels of this model, embeddings were created for `sentence1` and `sentence2`. Finally, a cosine similarity score was calculated across the two embeddings, and plotted on the graph.
|
| 33 |
|