Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -18,13 +18,13 @@ from transformers import LlamaForCausalLM, LlamaTokenizer
|
|
| 18 |
#chatbot = hugchat.ChatBot(cookie_path="cookies.json")
|
| 19 |
|
| 20 |
#Alternativ mit beliebigen Modellen:
|
| 21 |
-
|
| 22 |
#base_model = "MAGAer13/mPLUG-Owl" #load_8bit = False (in load_tokenizer_and_model)
|
| 23 |
#base_model = "alexkueck/li-tis-tuned-2" #load_8bit = False (in load_tokenizer_and_model)
|
| 24 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = False (in load_tokenizer_and_model)
|
| 25 |
#base_model = "EleutherAI/gpt-neo-1.3B" #load_8bit = False (in load_tokenizer_and_model)
|
| 26 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = True
|
| 27 |
-
base_model = "TheBloke/vicuna-13B-1.1-HF" #load_8bit = ?
|
| 28 |
#following runs only on GPU upgrade
|
| 29 |
#base_model = "TheBloke/airoboros-65B-gpt4-1.3-GPTQ" #model_basename = "airoboros-65b-gpt4-1.3-GPTQ-4bit--1g.act.order"
|
| 30 |
#base_model = "lmsys/vicuna-13b-v1.3"
|
|
|
|
| 18 |
#chatbot = hugchat.ChatBot(cookie_path="cookies.json")
|
| 19 |
|
| 20 |
#Alternativ mit beliebigen Modellen:
|
| 21 |
+
base_model = "project-baize/baize-v2-7b" #load_8bit = False (in load_tokenizer_and_model)
|
| 22 |
#base_model = "MAGAer13/mPLUG-Owl" #load_8bit = False (in load_tokenizer_and_model)
|
| 23 |
#base_model = "alexkueck/li-tis-tuned-2" #load_8bit = False (in load_tokenizer_and_model)
|
| 24 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = False (in load_tokenizer_and_model)
|
| 25 |
#base_model = "EleutherAI/gpt-neo-1.3B" #load_8bit = False (in load_tokenizer_and_model)
|
| 26 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = True
|
| 27 |
+
#base_model = "TheBloke/vicuna-13B-1.1-HF" #load_8bit = ?
|
| 28 |
#following runs only on GPU upgrade
|
| 29 |
#base_model = "TheBloke/airoboros-65B-gpt4-1.3-GPTQ" #model_basename = "airoboros-65b-gpt4-1.3-GPTQ-4bit--1g.act.order"
|
| 30 |
#base_model = "lmsys/vicuna-13b-v1.3"
|