|
|
--- |
|
|
language: |
|
|
- zh |
|
|
- en |
|
|
pipeline_tag: text-generation |
|
|
inference: false |
|
|
--- |
|
|
|
|
|
# baichuan-llama-7B |
|
|
|
|
|
使用[LLaMA](https://huggingface.co/huggyllama/llama-7b)格式保存的[baichuan-7B](https://huggingface.co/baichuan-inc/baichuan-7B)。可以直接使用LlamaForCausalLM和LlamaTokenizer加载。 |
|
|
权重文件以405M的尺寸分片,方便并行快速下载。权重来自[fireballoon/baichuan-llama-7b](https://huggingface.co/fireballoon/baichuan-llama-7b) |
|
|
|
|
|
[baichuan-7B](https://huggingface.co/baichuan-inc/baichuan-7B) model saved in the format of the [LLaMA](https://huggingface.co/huggyllama/llama-7b) model. You can directly use LlamaForCausalLM and LlamaTokenizer to load the model. |
|
|
The weight file is split into chunks with a size of 405M for convenient and fast parallel downloads, specifically for academic research purposes. The weights are sourced from [fireballoon/baichuan-llama-7b](https://huggingface.co/fireballoon/baichuan-llama-7b) |
|
|
|
|
|
**License:** [baichuan-7B License](https://huggingface.co/baichuan-inc/baichuan-7B/blob/main/baichuan-7B%20%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE.pdf) |
|
|
|
|
|
|