Upload README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,71 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
license: other
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
language:
|
| 3 |
+
- en
|
| 4 |
+
- zh
|
| 5 |
license: other
|
| 6 |
+
tasks:
|
| 7 |
+
- text-generation
|
| 8 |
---
|
| 9 |
+
|
| 10 |
+
# 智海-录问
|
| 11 |
+
|
| 12 |
+
## 项目背景
|
| 13 |
+
智海-录问(wisdomInterrogatory)是由浙江大学、阿里巴巴达摩院以及华院计算三家单位共同设计研发的法律大模型。核心思想:以“普法共享和司法效能提升”为目标,从推动法律智能化体系入司法实践、数字化案例建设、虚拟法律咨询服务赋能等方面提供支持,形成数字化和智能化的司法基座能力。
|
| 14 |
+
|
| 15 |
+
## 模型训练
|
| 16 |
+
|
| 17 |
+
我们的模型基座是[Baichuan-7B](https://github.com/baichuan-inc/baichuan-7B),在此基础上,进行了二次预训练以及指令微调训练。
|
| 18 |
+
|
| 19 |
+
### 二次预训练
|
| 20 |
+
|
| 21 |
+
二次预训练的目的是给通用的大模型注入法律领域的知识。预训练的数据包括法律文书、司法案例以及法律问答数据,共40G。
|
| 22 |
+
|
| 23 |
+
### 指令微调训练
|
| 24 |
+
|
| 25 |
+
经过了二次预训练之后,在指令微调阶段,我们使用了100k的指微调训练,其目的是让大模型具备问答的能力,能够直接与用户进行交流。
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+
|
| 29 |
+
## 推理代码
|
| 30 |
+
|
| 31 |
+
#### 推理环境安装
|
| 32 |
+
```shell
|
| 33 |
+
transformers>=4.27.1
|
| 34 |
+
accelerate>=0.20.1
|
| 35 |
+
torch>=2.0.1
|
| 36 |
+
modelscope>=1.8.3
|
| 37 |
+
sentencepiece==0.1.99
|
| 38 |
+
```
|
| 39 |
+
|
| 40 |
+
#### 推理代码调用
|
| 41 |
+
```python
|
| 42 |
+
import os
|
| 43 |
+
os.environ["CUDA_VISIBLE_DEVICES"] = "0"
|
| 44 |
+
from modelscope import AutoModelForCausalLM, AutoTokenizer, snapshot_download
|
| 45 |
+
import torch
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
model_id = "wisdomOcean/wisdomInterrogatory"
|
| 49 |
+
revision = 'v1.0.0'
|
| 50 |
+
model_dir = snapshot_download(model_id, revision)
|
| 51 |
+
|
| 52 |
+
def generate_response(prompt: str) -> str:
|
| 53 |
+
inputs = tokenizer(f'</s>Human:{prompt} </s>Assistant: ', return_tensors='pt')
|
| 54 |
+
inputs = inputs.to('cuda')
|
| 55 |
+
pred = model.generate(**inputs, max_new_tokens=800,
|
| 56 |
+
repetition_penalty=1.2)
|
| 57 |
+
response = tokenizer.decode(pred.cpu()[0], skip_special_tokens=True)
|
| 58 |
+
return response.split("Assistant: ")[1]
|
| 59 |
+
|
| 60 |
+
tokenizer = AutoTokenizer.from_pretrained(model_dir, trust_remote_code=True)
|
| 61 |
+
model = AutoModelForCausalLM.from_pretrained(model_dir, device_map="auto",
|
| 62 |
+
torch_dtype=torch.float16,
|
| 63 |
+
trust_remote_code=True)
|
| 64 |
+
prompt = "如果喝了两斤白酒后开车,会有什么后果?"
|
| 65 |
+
resp = generate_response(prompt)
|
| 66 |
+
print(resp)
|
| 67 |
+
```
|
| 68 |
+
|
| 69 |
+
## 免责声明
|
| 70 |
+
|
| 71 |
+
本模型仅供学术研究之目的而提供,不保证结果的准确性、完整性或适用性。在使用模型生成的内容时,您应自行判断其适用性,并自担风险。
|