limcheekin commited on
Commit
c61af36
·
verified ·
1 Parent(s): 95c1137

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +53 -0
README.md ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model:
4
+ - nomic-ai/CodeRankEmbed
5
+ tags:
6
+ - gguf
7
+ - embeddings
8
+ - f16
9
+ ---
10
+
11
+ # Model Card: CodeRankEmbed (GGUF Quantized)
12
+
13
+ ## Model Overview
14
+
15
+ This model is a GGUF-quantized version of [CodeRankEmbed](https://huggingface.co/nomic-ai/CodeRankEmbed).
16
+
17
+ The quantization reduces the model's size and computational requirements, facilitating efficient deployment without significantly compromising performance.
18
+
19
+ ## Model Details
20
+
21
+ - **Model Name:** CodeRankEmbed-GGUF
22
+ - **Original Model:** [CodeRankEmbed](https://huggingface.co/nomic-ai/CodeRankEmbed)
23
+ - **Quantization Format:** GGUF
24
+ - **Parameters:** 568 million
25
+ - **Embedding Dimension:** 768
26
+ - **Languages Supported:** Python, Java, JS, PHP, Go, Ruby
27
+ - **Context Length:** Supports up to 8,192 tokens
28
+ - **License:** MIT
29
+
30
+ ## Quantization Details
31
+
32
+ GGUF (Gerganov's General Unified Format) is a binary format optimized for efficient loading and inference of large language models. Quantization involves reducing the precision of the model's weights, resulting in decreased memory usage and faster computation with minimal impact on accuracy.
33
+
34
+ ## Performance
35
+
36
+ The CodeRankEmbed is a 137M bi-encoder supporting 8192 context length for code retrieval. It significantly outperforms various open-source and proprietary code embedding models on various code retrieval tasks.
37
+
38
+ ## Usage
39
+
40
+ This quantized model is suitable for deployment in resource-constrained environments where memory and computational efficiency are critical. It can be utilized for tasks such as code retrieval, semantic search, and other applications requiring high-quality code embeddings.
41
+
42
+ ## Limitations
43
+
44
+ While quantization reduces resource requirements, it may introduce slight degradation in model performance. Users should evaluate the model in their specific use cases to ensure it meets the desired performance criteria.
45
+
46
+ ## Acknowledgements
47
+
48
+ This quantized model is based on Nomic's CodeRankEmbed. For more details on the original model, please refer to the [official model card](https://huggingface.co/nomic-ai/CodeRankEmbed).
49
+
50
+ ---
51
+
52
+ For a overview of the CodeRankEmbed model, you may find the following article informative:
53
+ https://simonwillison.net/2025/Mar/27/nomic-embed-code