Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,10 +1,52 @@
|
|
| 1 |
---
|
| 2 |
title: README
|
| 3 |
-
emoji:
|
| 4 |
colorFrom: purple
|
| 5 |
colorTo: indigo
|
| 6 |
sdk: static
|
| 7 |
pinned: false
|
| 8 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
|
| 10 |
-
|
|
|
|
| 1 |
---
|
| 2 |
title: README
|
| 3 |
+
emoji: 👀
|
| 4 |
colorFrom: purple
|
| 5 |
colorTo: indigo
|
| 6 |
sdk: static
|
| 7 |
pinned: false
|
| 8 |
---
|
| 9 |
+
---
|
| 10 |
+
# EdgeCompress
|
| 11 |
+
|
| 12 |
+
**EdgeCompress** is a graduation project developed by senior Computer Science students from **Capital University, Egypt**.
|
| 13 |
+
|
| 14 |
+
Our work focuses on **compressing Large Language Models (LLMs)** to make them efficient enough to run on **edge devices** with limited computational resources.
|
| 15 |
+
|
| 16 |
+
## Project Overview
|
| 17 |
+
|
| 18 |
+
Large Language Models typically require significant memory, storage, and computational power. This makes them difficult to deploy on edge hardware such as embedded systems, IoT devices, and low-power GPUs.
|
| 19 |
+
|
| 20 |
+
Our project explores different **model compression techniques** to reduce the size and resource requirements of LLMs while maintaining acceptable performance.
|
| 21 |
+
|
| 22 |
+
## Research Focus
|
| 23 |
+
|
| 24 |
+
We investigate multiple compression approaches, including:
|
| 25 |
+
|
| 26 |
+
* **Quantization**
|
| 27 |
+
* **Model pruning**
|
| 28 |
+
* **Knowledge Distillation**
|
| 29 |
+
* **Low-precision inference**
|
| 30 |
+
* **Memory-efficient deployment strategies**
|
| 31 |
+
|
| 32 |
+
## Edge Deployment
|
| 33 |
+
|
| 34 |
+
After compression, the models are evaluated on **edge computing environments** to determine:
|
| 35 |
+
|
| 36 |
+
* Memory usage
|
| 37 |
+
* Inference latency
|
| 38 |
+
* Performance degradation after compression
|
| 39 |
+
* Suitability for real-time edge AI applications
|
| 40 |
+
|
| 41 |
+
## What You Will Find Here
|
| 42 |
+
|
| 43 |
+
This organization hosts:
|
| 44 |
+
|
| 45 |
+
* Compressed LLM checkpoints
|
| 46 |
+
* Experiments with different compression techniques
|
| 47 |
+
* Benchmark results on edge hardware
|
| 48 |
+
* Research artifacts from our graduation project
|
| 49 |
+
|
| 50 |
+
## Goal
|
| 51 |
|
| 52 |
+
Our goal is to **enable efficient deployment of LLMs on edge devices**, making advanced AI models more accessible in real-world and resource-constrained environments.
|