abdelrahmane01 commited on
Commit
00deabf
·
verified ·
1 Parent(s): eb2577c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +44 -2
README.md CHANGED
@@ -1,10 +1,52 @@
1
  ---
2
  title: README
3
- emoji: 🚀
4
  colorFrom: purple
5
  colorTo: indigo
6
  sdk: static
7
  pinned: false
8
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
 
10
- Edit this `README.md` markdown file to author your organization card.
 
1
  ---
2
  title: README
3
+ emoji: 👀
4
  colorFrom: purple
5
  colorTo: indigo
6
  sdk: static
7
  pinned: false
8
  ---
9
+ ---
10
+ # EdgeCompress
11
+
12
+ **EdgeCompress** is a graduation project developed by senior Computer Science students from **Capital University, Egypt**.
13
+
14
+ Our work focuses on **compressing Large Language Models (LLMs)** to make them efficient enough to run on **edge devices** with limited computational resources.
15
+
16
+ ## Project Overview
17
+
18
+ Large Language Models typically require significant memory, storage, and computational power. This makes them difficult to deploy on edge hardware such as embedded systems, IoT devices, and low-power GPUs.
19
+
20
+ Our project explores different **model compression techniques** to reduce the size and resource requirements of LLMs while maintaining acceptable performance.
21
+
22
+ ## Research Focus
23
+
24
+ We investigate multiple compression approaches, including:
25
+
26
+ * **Quantization**
27
+ * **Model pruning**
28
+ * **Knowledge Distillation**
29
+ * **Low-precision inference**
30
+ * **Memory-efficient deployment strategies**
31
+
32
+ ## Edge Deployment
33
+
34
+ After compression, the models are evaluated on **edge computing environments** to determine:
35
+
36
+ * Memory usage
37
+ * Inference latency
38
+ * Performance degradation after compression
39
+ * Suitability for real-time edge AI applications
40
+
41
+ ## What You Will Find Here
42
+
43
+ This organization hosts:
44
+
45
+ * Compressed LLM checkpoints
46
+ * Experiments with different compression techniques
47
+ * Benchmark results on edge hardware
48
+ * Research artifacts from our graduation project
49
+
50
+ ## Goal
51
 
52
+ Our goal is to **enable efficient deployment of LLMs on edge devices**, making advanced AI models more accessible in real-world and resource-constrained environments.