Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -8,9 +8,9 @@ import time
|
|
| 8 |
|
| 9 |
custom_prompt_template = """
|
| 10 |
You are an AI coding assistant and your task is to solve coding problems and
|
| 11 |
-
return code snippets based on user's query. Below is the user's query.
|
| 12 |
Query: {query}
|
| 13 |
-
You just return the helpful code and related
|
| 14 |
Helpful code and related details:
|
| 15 |
"""
|
| 16 |
|
|
@@ -23,7 +23,7 @@ def set_custom_prompt():
|
|
| 23 |
|
| 24 |
def load_model():
|
| 25 |
llm = CTransformers(
|
| 26 |
-
model = 'codellama-7b.Q4_K_M.gguf',
|
| 27 |
model_type = 'llama',
|
| 28 |
max_new_tokens = 1096,
|
| 29 |
temperature = 0.2,
|
|
|
|
| 8 |
|
| 9 |
custom_prompt_template = """
|
| 10 |
You are an AI coding assistant and your task is to solve coding problems and
|
| 11 |
+
return code snippets based on the user's query. Below is the user's query.
|
| 12 |
Query: {query}
|
| 13 |
+
You just return the helpful code and related details
|
| 14 |
Helpful code and related details:
|
| 15 |
"""
|
| 16 |
|
|
|
|
| 23 |
|
| 24 |
def load_model():
|
| 25 |
llm = CTransformers(
|
| 26 |
+
model = 'CodeLlama-7B-GGUF/codellama-7b.Q4_K_M.gguf',
|
| 27 |
model_type = 'llama',
|
| 28 |
max_new_tokens = 1096,
|
| 29 |
temperature = 0.2,
|