Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,83 @@
|
|
| 1 |
---
|
| 2 |
license: gpl-3.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: gpl-3.0
|
| 3 |
+
tags:
|
| 4 |
+
- text2text-generation
|
| 5 |
+
pipeline_tag: text2text-generation
|
| 6 |
+
language:
|
| 7 |
+
- zh
|
| 8 |
+
- en
|
| 9 |
---
|
| 10 |
+
|
| 11 |
+
Considering LLaMA's license constraints, the model is for research and learning only.
|
| 12 |
+
Please strictly respect LLaMA's usage policy. We are not allowed to publish weights for LLaMA, of course, even finetuned, but there is no problem publishing the difference, a patch that we suggest to apply to the files.
|
| 13 |
+
The encryption is a simple XOR between files, ensuring that only the people that have access to the original weights (from completely legal sources, of course) can transform them into finetuned weights.
|
| 14 |
+
You can find the decrypt code on https://github.com/LianjiaTech/BELLE/tree/main/models .
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
# Model Card for BELLE-LLaMA-7B-2M-q4
|
| 18 |
+
|
| 19 |
+
## Welcome
|
| 20 |
+
4-bit quantized version using [llama.cpp](https://github.com/ggerganov/llama.cpp) of [BELLE-LLaMA-7B-2M](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-enc)
|
| 21 |
+
If you find this model helpful, please *like* this model and star us on https://github.com/LianjiaTech/BELLE !
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
## Model description
|
| 25 |
+
BELLE-LLAMA-7B-2M-enc is based on LLAMA 7B and finetuned with 2M Chinese data combined with 50,000 pieces of English data from the open source Stanford-Alpaca, resulting in good Chinese instruction understanding and response generation capabilities.
|
| 26 |
+
|
| 27 |
+
The code of Chinese data generation and other detailed information can be found in our Github project repository: https://github.com/LianjiaTech/BELLE.
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
## Download
|
| 31 |
+
Should you accept our license and acknowledged the limitations, download the model by clicking [Download](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-q4/resolve/main/belle-model.bin).
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
## Model Usage
|
| 35 |
+
This is a quantized version made for offline on-devices inferencing.
|
| 36 |
+
You can use this model with ChatBELLE, a minimal, cross-platform LLM chat app powered by [BELLE](https://github.com/LianjiaTech/BELLE)
|
| 37 |
+
using quantized on-device offline models and Flutter UI, running on macOS (done), Windows, Android,
|
| 38 |
+
iOS(see [Known Issues](#known-issues)) and more.
|
| 39 |
+
|
| 40 |
+
### macOS
|
| 41 |
+
* Download and put the app anywhere, preferably in `Applications` folder.
|
| 42 |
+
* Open the app by right click then Ctrl-click `Open`, then click `Open`.
|
| 43 |
+
* The app will prompt the intended model file path and fail to load the model. Close the app.
|
| 44 |
+
* Download quantized model from [BELLE-LLaMA-7B-2M-q4](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-q4/blob/main/belle-model.bin).
|
| 45 |
+
* Move and rename the model to the path prompted by the app. Defaults to `~/Library/Containers/com.barius.chatbelle/Data/belle-model.bin` .
|
| 46 |
+
* Reopen the app again (double clicking is now OK).
|
| 47 |
+
|
| 48 |
+
### Windows
|
| 49 |
+
* Stay tuned
|
| 50 |
+
|
| 51 |
+
### Android
|
| 52 |
+
* Stay tuned
|
| 53 |
+
|
| 54 |
+
### iOS
|
| 55 |
+
* Stay tuned
|
| 56 |
+
|
| 57 |
+
|
| 58 |
+
## Limitations
|
| 59 |
+
There still exists a few issues in the model trained on current base model and data:
|
| 60 |
+
|
| 61 |
+
1. The model might generate factual errors when asked to follow instructions related to facts.
|
| 62 |
+
|
| 63 |
+
2. Occasionally generates harmful responses since the model still struggles to identify potential harmful instructions.
|
| 64 |
+
|
| 65 |
+
3. Needs improvements on reasoning and coding.
|
| 66 |
+
|
| 67 |
+
Since the model still has its limitations, we require developers only use the open-sourced code, data, model and any other artifacts generated via this project for research purposes. Commercial use and other potential harmful use cases are not allowed.
|
| 68 |
+
|
| 69 |
+
|
| 70 |
+
## Citation
|
| 71 |
+
|
| 72 |
+
Please cite us when using our code, data or model.
|
| 73 |
+
|
| 74 |
+
```
|
| 75 |
+
@misc{BELLE,
|
| 76 |
+
author = {Yunjie Ji, Yong Deng, Yan Gong, Yiping Peng, Qiang Niu, Baochang Ma, Xiangang Li},
|
| 77 |
+
title = {BELLE: Be Everyone's Large Language model Engine},
|
| 78 |
+
year = {2023},
|
| 79 |
+
publisher = {GitHub},
|
| 80 |
+
journal = {GitHub repository},
|
| 81 |
+
howpublished = {\url{https://github.com/LianjiaTech/BELLE}},
|
| 82 |
+
}
|
| 83 |
+
```
|