fahmiaziz98 commited on
Commit
1c62a3d
·
1 Parent(s): bf67a97

granite model

Browse files
Files changed (1) hide show
  1. config.yaml +13 -0
config.yaml CHANGED
@@ -53,3 +53,16 @@ models:
53
  Not that inference efficiency is not important, we will address that subsequently.)
54
  language: ["multilingual"]
55
  repository: "https://huggingface.co/prithivida/Splade_PP_en_v2"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  Not that inference efficiency is not important, we will address that subsequently.)
54
  language: ["multilingual"]
55
  repository: "https://huggingface.co/prithivida/Splade_PP_en_v2"
56
+
57
+ granite-30m:
58
+ name: "ibm-granite/granite-embedding-30m-sparse"
59
+ type: "sparse-embeddings"
60
+ dimension: 1234 # must add this field
61
+ max_tokens: 1234
62
+ description: |
63
+ Granite-Embedding-30m-Sparse is a 30M parameter sparse biencoder embedding model from the Granite Experimental suite that can be used to generate high quality text embeddings.
64
+ This model produces variable length bag-of-word like dictionary, containing expansions of sentence tokens and their corresponding weights and is trained using a combination of open source relevance-pair datasets with permissive,
65
+ enterprise-friendly license, and IBM collected and generated datasets. While maintaining competitive scores on academic benchmarks such as BEIR, this model also performs well on many enterprise use cases.
66
+ This model is developed using retrieval oriented pretraining, contrastive finetuning and knowledge distillation for improved performance.
67
+ language: ["En"]
68
+ repository: "https://huggingface.co/ibm-granite/granite-embedding-30m-sparse"