hopenjin commited on
Commit
e33bcdf
·
1 Parent(s): 3934e01

Update model card with CE-CoLLM paper info

Browse files
Files changed (1) hide show
  1. README.md +22 -10
README.md CHANGED
@@ -1,13 +1,25 @@
 
 
 
 
 
 
 
 
1
  ## Citation
2
- If you find this model useful, please cite the following works:
 
3
 
4
  ```bibtex
5
- @misc{jin2024cecollmefficientadaptivelarge,
6
- title={CE-CoLLM: Efficient and Adaptive Large Language Models Through Cloud-Edge Collaboration},
7
- author={Hongpeng Jin and Yanzhao Wu},
8
- year={2024},
9
- eprint={2411.02829},
10
- archivePrefix={arXiv},
11
- primaryClass={cs.DC},
12
- url={https://arxiv.org/abs/2411.02829},
13
- }
 
 
 
 
1
+ # CE-CoLLM Cloud LLM Partition
2
+
3
+ This repository hosts the cloud-side partition of the CE-CoLLM model used in our IEEE ICWS 2025 paper "CE-CoLLM: Efficient and Adaptive Large Language Models Through Cloud-Edge Collaboration".
4
+
5
+ - Codebase: https://github.com/mlsysx/CE-CoLLM
6
+ - Edge-side partition: https://huggingface.co/hopenjin/CE-CoLLM_Edge_LLM_Partition
7
+ - Paper (IEEE Xplore): https://ieeexplore.ieee.org/abstract/document/11169709
8
+
9
  ## Citation
10
+
11
+ Please cite the following paper when using CE-CoLLM:
12
 
13
  ```bibtex
14
+ @INPROCEEDINGS{jin2024cecollmefficientadaptivelarge,
15
+ author={Jin, Hongpeng and Wu, Yanzhao},
16
+ booktitle={2025 IEEE International Conference on Web Services (ICWS)},
17
+ title={CE-CoLLM: Efficient and Adaptive Large Language Models Through Cloud-Edge Collaboration},
18
+ year={2025},
19
+ pages={316-323},
20
+ keywords={Cloud computing;Accuracy;Web services;Large language models;Collaboration;Benchmark testing;Reliability engineering;Low latency communication;Edge computing;Software development management;Large Language Model;LLM Deployment;Cloud-Edge Collaboration;Cloud Services;Adaptive LLM Inference;Edge AI},
21
+ doi={10.1109/ICWS67624.2025.00046},
22
+ ISSN={2836-3868},
23
+ month={July},
24
+ }
25
+ ```