NewBreaker
commited on
Commit
·
09d423b
1
Parent(s):
60a8d23
auto git
Browse files- README.md +2 -0
- config.json +1 -1
- demo_call_api.py +10 -9
README.md
CHANGED
|
@@ -27,6 +27,8 @@ widget:
|
|
| 27 |
example_title: Rammstein's album Mutter
|
| 28 |
pipeline_tag: text2text-generation
|
| 29 |
|
|
|
|
|
|
|
| 30 |
|
| 31 |
---
|
| 32 |
|
|
|
|
| 27 |
example_title: Rammstein's album Mutter
|
| 28 |
pipeline_tag: text2text-generation
|
| 29 |
|
| 30 |
+
tags:
|
| 31 |
+
- conversational
|
| 32 |
|
| 33 |
---
|
| 34 |
|
config.json
CHANGED
|
@@ -9,7 +9,7 @@
|
|
| 9 |
"AutoModel": "modeling_chatglm.ChatGLMForConditionalGeneration",
|
| 10 |
"AutoModelForSeq2SeqLM": "modeling_chatglm.ChatGLMForConditionalGeneration"
|
| 11 |
},
|
| 12 |
-
|
| 13 |
"bos_token_id": 130004,
|
| 14 |
"eos_token_id": 130005,
|
| 15 |
"gmask_token_id": 130001,
|
|
|
|
| 9 |
"AutoModel": "modeling_chatglm.ChatGLMForConditionalGeneration",
|
| 10 |
"AutoModelForSeq2SeqLM": "modeling_chatglm.ChatGLMForConditionalGeneration"
|
| 11 |
},
|
| 12 |
+
|
| 13 |
"bos_token_id": 130004,
|
| 14 |
"eos_token_id": 130005,
|
| 15 |
"gmask_token_id": 130001,
|
demo_call_api.py
CHANGED
|
@@ -1,14 +1,15 @@
|
|
| 1 |
-
from transformers import AutoModel,AutoTokenizer
|
| 2 |
-
|
| 3 |
-
model = AutoModel.from_pretrained("NewBreaker/chatglm-6b-int4",trust_remote_code=True)
|
| 4 |
-
tokenizer = AutoTokenizer.from_pretrained("NewBreaker/chatglm-6b-int4", trust_remote_code=True)
|
| 5 |
|
| 6 |
from transformers import AutoConfig
|
| 7 |
|
| 8 |
-
config = AutoConfig.from_pretrained("
|
| 9 |
|
| 10 |
-
# Push the config to your namespace with the name "my-finetuned-bert".
|
| 11 |
-
config.push_to_hub("my-finetuned-bert")
|
| 12 |
|
| 13 |
-
# Push the config to
|
| 14 |
-
config.push_to_hub("
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# from transformers import AutoModel,AutoTokenizer
|
| 2 |
+
#
|
| 3 |
+
# model = AutoModel.from_pretrained("NewBreaker/chatglm-6b-int4",trust_remote_code=True)
|
| 4 |
+
# tokenizer = AutoTokenizer.from_pretrained("NewBreaker/chatglm-6b-int4", trust_remote_code=True)
|
| 5 |
|
| 6 |
from transformers import AutoConfig
|
| 7 |
|
| 8 |
+
config = AutoConfig.from_pretrained("silver/chatglm-6b-int4-qe-slim",trust_remote_code=True)
|
| 9 |
|
|
|
|
|
|
|
| 10 |
|
| 11 |
+
# # Push the config to your namespace with the name "my-finetuned-bert".
|
| 12 |
+
config.push_to_hub("silver/chatglm-6b-int4-qe-slim")
|
| 13 |
+
#
|
| 14 |
+
# # Push the config to an organization with the name "my-finetuned-bert".
|
| 15 |
+
# config.push_to_hub("huggingface/my-finetuned-bert")
|