Commit ·
838916a
1
Parent(s): 8d38a05
Update app.py
Browse files
app.py
CHANGED
|
@@ -6,16 +6,16 @@ import torch
|
|
| 6 |
from model import get_input_token_length, run
|
| 7 |
|
| 8 |
DEFAULT_SYSTEM_PROMPT = """\
|
| 9 |
-
You are a helpful, respectful and honest assistant with a deep knowledge of code and software design. Always answer as helpfully as possible
|
| 10 |
"""
|
| 11 |
MAX_MAX_NEW_TOKENS = 4096
|
| 12 |
DEFAULT_MAX_NEW_TOKENS = 1024
|
| 13 |
MAX_INPUT_TOKEN_LENGTH = 4000
|
| 14 |
|
| 15 |
DESCRIPTION = """
|
| 16 |
-
# Code Llama
|
| 17 |
|
| 18 |
-
This Space demonstrates model [CodeLlama-
|
| 19 |
|
| 20 |
🔎 For more details about the Code Llama family of models and how to use them with `transformers`, take a look [at our blog post](https://huggingface.co/blog/codellama) or [the paper](https://huggingface.co/papers/2308.12950).
|
| 21 |
|
|
|
|
| 6 |
from model import get_input_token_length, run
|
| 7 |
|
| 8 |
DEFAULT_SYSTEM_PROMPT = """\
|
| 9 |
+
You are a helpful, respectful and honest assistant with a deep knowledge of code and software design. Always answer as helpfully as possible. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. \n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.\
|
| 10 |
"""
|
| 11 |
MAX_MAX_NEW_TOKENS = 4096
|
| 12 |
DEFAULT_MAX_NEW_TOKENS = 1024
|
| 13 |
MAX_INPUT_TOKEN_LENGTH = 4000
|
| 14 |
|
| 15 |
DESCRIPTION = """
|
| 16 |
+
# Code Llama 34B Chat
|
| 17 |
|
| 18 |
+
This Space demonstrates model [CodeLlama-34b-Instruct](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) by Meta, a Code Llama model with 34B parameters fine-tuned for chat instructions and specialized on code tasks. Feel free to play with it, or duplicate to run generations without a queue! If you want to run your own service, you can also [deploy the model on Inference Endpoints](https://huggingface.co/inference-endpoints).
|
| 19 |
|
| 20 |
🔎 For more details about the Code Llama family of models and how to use them with `transformers`, take a look [at our blog post](https://huggingface.co/blog/codellama) or [the paper](https://huggingface.co/papers/2308.12950).
|
| 21 |
|