Spaces:
Running
Running
fahmiaziz98
commited on
Commit
·
1c62a3d
1
Parent(s):
bf67a97
granite model
Browse files- config.yaml +13 -0
config.yaml
CHANGED
|
@@ -53,3 +53,16 @@ models:
|
|
| 53 |
Not that inference efficiency is not important, we will address that subsequently.)
|
| 54 |
language: ["multilingual"]
|
| 55 |
repository: "https://huggingface.co/prithivida/Splade_PP_en_v2"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
Not that inference efficiency is not important, we will address that subsequently.)
|
| 54 |
language: ["multilingual"]
|
| 55 |
repository: "https://huggingface.co/prithivida/Splade_PP_en_v2"
|
| 56 |
+
|
| 57 |
+
granite-30m:
|
| 58 |
+
name: "ibm-granite/granite-embedding-30m-sparse"
|
| 59 |
+
type: "sparse-embeddings"
|
| 60 |
+
dimension: 1234 # must add this field
|
| 61 |
+
max_tokens: 1234
|
| 62 |
+
description: |
|
| 63 |
+
Granite-Embedding-30m-Sparse is a 30M parameter sparse biencoder embedding model from the Granite Experimental suite that can be used to generate high quality text embeddings.
|
| 64 |
+
This model produces variable length bag-of-word like dictionary, containing expansions of sentence tokens and their corresponding weights and is trained using a combination of open source relevance-pair datasets with permissive,
|
| 65 |
+
enterprise-friendly license, and IBM collected and generated datasets. While maintaining competitive scores on academic benchmarks such as BEIR, this model also performs well on many enterprise use cases.
|
| 66 |
+
This model is developed using retrieval oriented pretraining, contrastive finetuning and knowledge distillation for improved performance.
|
| 67 |
+
language: ["En"]
|
| 68 |
+
repository: "https://huggingface.co/ibm-granite/granite-embedding-30m-sparse"
|