Improve model card and add metadata

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +32 -4
README.md CHANGED
@@ -1,10 +1,38 @@
1
  ---
 
 
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
5
  ---
6
 
7
- This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
- - Code: https://github.com/Intellindust-AI-Lab/EdgeCrafter
9
- - Paper: https://arxiv.org/abs/2603.18739
10
- - Docs: [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: apache-2.0
3
+ pipeline_tag: image-segmentation
4
  tags:
5
  - model_hub_mixin
6
  - pytorch_model_hub_mixin
7
  ---
8
 
9
+ # EdgeCrafter: Compact ViTs for Edge Dense Prediction via Task-Specialized Distillation
10
+
11
+ EdgeCrafter is a unified framework for high-performance dense prediction on resource-constrained edge devices. It introduces compact Vision Transformers (ViTs) that compete with CNN-based architectures like YOLO by using task-specialized distillation and edge-aware encoder-decoder designs.
12
+
13
+ This repository contains a checkpoint for **ECSeg**, the instance segmentation variant of the framework, which achieves a strong accuracy-efficiency tradeoff.
14
+
15
+ - **Paper:** [EdgeCrafter: Compact ViTs for Edge Dense Prediction via Task-Specialized Distillation](https://arxiv.org/abs/2603.18739)
16
+ - **Project Page:** [EdgeCrafter Project Page](https://intellindust-ai-lab.github.io/projects/EdgeCrafter/)
17
+ - **Code:** [GitHub Repository](https://github.com/Intellindust-AI-Lab/EdgeCrafter)
18
+
19
+ ## Model Description
20
+
21
+ Deploying high-performance dense prediction models on resource-constrained edge devices remains challenging due to strict limits on computation and memory. EdgeCrafter addresses this by introducing a framework centered on distilled compact backbones and edge-friendly encoder-decoder designs. For instance segmentation, ECSeg achieves performance comparable to RF-DETR while using substantially fewer parameters, proving that compact ViTs can be a practical and competitive option for edge deployment.
22
+
23
+ ## Usage
24
+
25
+ This model is compatible with the `PytorchModelHubMixin`. For detailed instructions on installation, training, and running inference, please refer to the [official GitHub repository](https://github.com/Intellindust-AI-Lab/EdgeCrafter).
26
+
27
+ ## Citation
28
+
29
+ If you find this project useful in your research, please consider citing:
30
+
31
+ ```bibtex
32
+ @article{liu2026edgecrafter,
33
+ title={EdgeCrafter: Compact ViTs for Edge Dense Prediction via Task-Specialized Distillation},
34
+ author={Liu, Longfei and Hou, Yongjie and Li, Yang and Wang, Qirui and Sha, Youyang and Yu, Yongjun and Wang, Yinzhi and Ru, Peizhe and Yu, Xuanlong and Shen, Xi},
35
+ journal={arXiv},
36
+ year={2026}
37
+ }
38
+ ```