Instructions to use ClueAI/PromptCLUE-base with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use ClueAI/PromptCLUE-base with Transformers:
# Load model directly from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("ClueAI/PromptCLUE-base") model = AutoModelForSeq2SeqLM.from_pretrained("ClueAI/PromptCLUE-base") - Notebooks
- Google Colab
- Kaggle
Update README.md
Browse files
README.md
CHANGED
|
@@ -17,9 +17,10 @@ PromptCLUE:全中文任务零样本学习模型
|
|
| 17 |
加载模型:
|
| 18 |
|
| 19 |
```python
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
|
|
|
|
| 23 |
```
|
| 24 |
|
| 25 |
使用模型进行预测推理方法:
|
|
|
|
| 17 |
加载模型:
|
| 18 |
|
| 19 |
```python
|
| 20 |
+
# 加载模型
|
| 21 |
+
from transformers import T5Tokenizer, T5ForConditionalGeneration
|
| 22 |
+
tokenizer = T5Tokenizer.from_pretrained("ClueAI/PromptCLUE-base")
|
| 23 |
+
model = T5ForConditionalGeneration.from_pretrained("ClueAI/PromptCLUE-base")
|
| 24 |
```
|
| 25 |
|
| 26 |
使用模型进行预测推理方法:
|