Update README.md
#3
by
xuehaha
- opened
README.md
CHANGED
|
@@ -52,6 +52,7 @@ widget:
|
|
| 52 |
问题:小米的创始人是谁?
|
| 53 |
答案:
|
| 54 |
library_name: paddlenlp
|
|
|
|
| 55 |
---
|
| 56 |
|
| 57 |
<a href="https://colab.research.google.com/drive/1hlSMYEq3pyX-fwTSqIOT1um80kU1yOJF?usp=sharing"><img src="https://colab.research.google.com/assets/colab-badge.svg"></a>
|
|
@@ -72,14 +73,14 @@ PromptCLUE:全中文任务零样本学习模型
|
|
| 72 |
# 加载模型
|
| 73 |
from paddlenlp.transformers import AutoTokenizer, T5ForConditionalGeneration
|
| 74 |
|
| 75 |
-
tokenizer = AutoTokenizer.from_pretrained("ClueAI/PromptCLUE-base
|
| 76 |
-
model = T5ForConditionalGeneration.from_pretrained("ClueAI/PromptCLUE-base
|
| 77 |
```
|
| 78 |
|
| 79 |
使用模型进行预测推理方法:
|
| 80 |
```python
|
| 81 |
import torch
|
| 82 |
-
#这里使用的是paddle的
|
| 83 |
def preprocess(text):
|
| 84 |
return text.replace("\n", "_")
|
| 85 |
|
|
|
|
| 52 |
问题:小米的创始人是谁?
|
| 53 |
答案:
|
| 54 |
library_name: paddlenlp
|
| 55 |
+
pipeline_tag: text2text-generation
|
| 56 |
---
|
| 57 |
|
| 58 |
<a href="https://colab.research.google.com/drive/1hlSMYEq3pyX-fwTSqIOT1um80kU1yOJF?usp=sharing"><img src="https://colab.research.google.com/assets/colab-badge.svg"></a>
|
|
|
|
| 73 |
# 加载模型
|
| 74 |
from paddlenlp.transformers import AutoTokenizer, T5ForConditionalGeneration
|
| 75 |
|
| 76 |
+
tokenizer = AutoTokenizer.from_pretrained("ClueAI/PromptCLUE-base", from_hf_hub=False)
|
| 77 |
+
model = T5ForConditionalGeneration.from_pretrained("ClueAI/PromptCLUE-base", from_hf_hub=False)
|
| 78 |
```
|
| 79 |
|
| 80 |
使用模型进行预测推理方法:
|
| 81 |
```python
|
| 82 |
import torch
|
| 83 |
+
#这里使用的是paddle的gpu版本,推理更快
|
| 84 |
def preprocess(text):
|
| 85 |
return text.replace("\n", "_")
|
| 86 |
|