Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -29,7 +29,7 @@ class MessageRequest(BaseModel):
|
|
| 29 |
message: str
|
| 30 |
language: str
|
| 31 |
|
| 32 |
-
repo_id = "
|
| 33 |
llm_client = InferenceClient(
|
| 34 |
model=repo_id,
|
| 35 |
token=os.getenv("HF_TOKEN"),
|
|
@@ -65,8 +65,8 @@ templates = Jinja2Templates(directory="static")
|
|
| 65 |
|
| 66 |
# Configure Llama index settings
|
| 67 |
Settings.llm = HuggingFaceInferenceAPI(
|
| 68 |
-
model_name="
|
| 69 |
-
tokenizer_name="
|
| 70 |
context_window=3000,
|
| 71 |
token=os.getenv("HF_TOKEN"),
|
| 72 |
max_new_tokens=512,
|
|
|
|
| 29 |
message: str
|
| 30 |
language: str
|
| 31 |
|
| 32 |
+
repo_id = "meta-llama/Meta-Llama-3-8B-Instruct"
|
| 33 |
llm_client = InferenceClient(
|
| 34 |
model=repo_id,
|
| 35 |
token=os.getenv("HF_TOKEN"),
|
|
|
|
| 65 |
|
| 66 |
# Configure Llama index settings
|
| 67 |
Settings.llm = HuggingFaceInferenceAPI(
|
| 68 |
+
model_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
| 69 |
+
tokenizer_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
| 70 |
context_window=3000,
|
| 71 |
token=os.getenv("HF_TOKEN"),
|
| 72 |
max_new_tokens=512,
|