Spaces:
Runtime error
Runtime error
Update main.py
Browse files
main.py
CHANGED
|
@@ -19,14 +19,13 @@ starchat_repo_id = os.environ.get('starchat_repo_id')
|
|
| 19 |
llama2_repo_id=os.environ.get('llama2_repo_id')
|
| 20 |
#port = os.getenv('port')
|
| 21 |
|
| 22 |
-
llm = HuggingFaceHub(
|
| 23 |
-
|
| 24 |
-
#huggingfacehub_api_token="hf_p***K",
|
| 25 |
huggingfacehub_api_token=hf_token,
|
| 26 |
model_kwargs={#"min_length":512, #for StarChat
|
| 27 |
"min_length":1024, #for Llama2
|
| 28 |
#"max_new_tokens":3072, "do_sample":True, #for StarChat
|
| 29 |
-
"max_new_tokens":
|
| 30 |
"temperature":0.1,
|
| 31 |
"top_k":50,
|
| 32 |
"top_p":0.95, "eos_token_id":49155})
|
|
|
|
| 19 |
llama2_repo_id=os.environ.get('llama2_repo_id')
|
| 20 |
#port = os.getenv('port')
|
| 21 |
|
| 22 |
+
llm = HuggingFaceHub(llama2_repo_id=llama2_repo_id,
|
| 23 |
+
#starchat_repo_id=starchat_repo_id,
|
|
|
|
| 24 |
huggingfacehub_api_token=hf_token,
|
| 25 |
model_kwargs={#"min_length":512, #for StarChat
|
| 26 |
"min_length":1024, #for Llama2
|
| 27 |
#"max_new_tokens":3072, "do_sample":True, #for StarChat
|
| 28 |
+
"max_new_tokens":5632, "do_sample":True, #for Llama2
|
| 29 |
"temperature":0.1,
|
| 30 |
"top_k":50,
|
| 31 |
"top_p":0.95, "eos_token_id":49155})
|