Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,7 +5,7 @@ from langchain.llms import CTransformers
|
|
| 5 |
# Function to get response from LLaMA 2 model
|
| 6 |
def get_llama_response(input_text, no_words, blog_style):
|
| 7 |
# LLaMA 2 model (replace with the correct model identifier if needed)
|
| 8 |
-
llm = CTransformers(model='
|
| 9 |
config={'max_new_tokens': 256, 'temperature': 0.01})
|
| 10 |
|
| 11 |
# Prompt Template
|
|
|
|
| 5 |
# Function to get response from LLaMA 2 model
|
| 6 |
def get_llama_response(input_text, no_words, blog_style):
|
| 7 |
# LLaMA 2 model (replace with the correct model identifier if needed)
|
| 8 |
+
llm = CTransformers(model='meta-llama/Llama-2-7b-chat-hf', model_type='llama',
|
| 9 |
config={'max_new_tokens': 256, 'temperature': 0.01})
|
| 10 |
|
| 11 |
# Prompt Template
|