File size: 3,324 Bytes
479bd64
 
 
 
 
 
 
 
 
 
 
fea62df
 
c93204d
fea62df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bf67a97
 
 
 
 
 
 
fea62df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51ae485
 
fea62df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
---
title: Api Embedding
emoji: 🐠
colorFrom: green
colorTo: purple
sdk: docker
pinned: false
---

Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference

# 🧠 Unified Embedding API

> 🧩 Unified API for all your Embedding & Sparse needs β€” plug and play with any model from Hugging Face or your own fine-tuned versions. This official repository from huggingface space

---

## πŸš€ Overview

**Unified Embedding API** is a modular and open-source **RAG-ready API** built for developers who want a simple, unified way to access **dense**, and **sparse** models.

It’s designed for **vector search**, **semantic retrieval**, and **AI-powered pipelines** β€” all controlled from a single `config.yaml` file.

⚠️ **Note:** This is a development API.  
For production deployment, host it on cloud platforms such as **Hugging Face TGI**, **AWS**, or **GCP**.

---

## 🧩 Features

- 🧠 **Unified Interface** β€” One API to handle dense, sparse, and reranking models.
- βš™οΈ **Configurable** β€” Switch models instantly via `config.yaml`.
- πŸ” **Vector DB Ready** β€” Easily integrates with FAISS, Chroma, Qdrant, Milvus, etc.
- πŸ“ˆ **RAG Support** β€” Perfect base for Retrieval-Augmented Generation systems.
- ⚑ **Fast & Lightweight** β€” Powered by FastAPI and optimized with async processing.
- 🧰 **Extendable** β€” Add your own models or pipelines effortlessly.

---

## πŸ“ Project Structure

```

unified-embedding-api/
β”‚
β”œβ”€β”€ core/
β”‚   β”œβ”€β”€ embedding.py         
β”‚   └── model_manager.py     
β”œβ”€β”€ models/
|   └──model.py
β”œβ”€β”€ app.py                   # Entry point (FastAPI server)
|── config.yaml              # Model + system configuration
β”œβ”€β”€ Dockerfile                 
β”œβ”€β”€ requirements.txt
└── README.md

```
---
## 🧩 Model Selection

Default configuration is optimized for **CPU 2vCPU / 16GB RAM**. See [MTEB Leaderboard](https://huggingface.co/spaces/mteb/leaderboard) for memory usage reference.

⚠️ If you plan to use larger models like `Qwen2-embedding-8B`, please upgrade your Space.

---

## ☁️ How to Deploy (Free πŸš€)

Deploy your **custom Embedding API** on **Hugging Face Spaces** β€” free, fast, and serverless.

### πŸ”§ Steps:

1. **Clone this Space Template:**
   πŸ‘‰ [Hugging Face Space β€” fahmiaziz/api-embedding](https://huggingface.co/spaces/fahmiaziz/api-embedding)
2. **Edit `config.yaml`** to set your own model names and backend preferences.
3. **Push your code** β€” Spaces will automatically rebuild and host your API.

That’s it! You now have a live embedding API endpoint powered by your models.

πŸ“˜ **Tutorial Reference:**
- [Deploy Applications on Hugging Face Spaces (Official Guide)](https://huggingface.co/blog/HemanthSai7/deploy-applications-on-huggingface-spaces)
- [How-to-Sync-Hugging-Face-Spaces-with-a-GitHub-Repository by Ruslanmv](https://github.com/ruslanmv/How-to-Sync-Hugging-Face-Spaces-with-a-GitHub-Repository?tab=readme-ov-file)

---


## πŸ§‘β€πŸ’» Contributing

Contributions are welcome!
Please open an issue or submit a pull request to discuss changes.

---

## ⚠️ License

MIT License Β© 2025
Developed with ❀️ by the Open-Source Community.

---

> ✨ β€œUnify your embeddings. Simplify your AI stack.”