Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -57,7 +57,6 @@ elif st.session_state.page == "python":
|
|
| 57 |
# Setup the LangChain HuggingFaceEndpoint and ChatHuggingFace LLM
|
| 58 |
deep_seek_model = HuggingFaceEndpoint(
|
| 59 |
repo_id="deepseek-ai/DeepSeek-R1",
|
| 60 |
-
provider="nebius",
|
| 61 |
temperature=0.7,
|
| 62 |
max_new_tokens=100,
|
| 63 |
task="conversational",
|
|
@@ -114,7 +113,6 @@ elif st.session_state.page == "sql":
|
|
| 114 |
# Initialize the LLaMA model from HuggingFace (via Nebius provider)
|
| 115 |
llama_model = HuggingFaceEndpoint(
|
| 116 |
repo_id="meta-llama/Llama-3.1-8B-Instruct",
|
| 117 |
-
provider="nebius",
|
| 118 |
temperature=0.7,
|
| 119 |
max_new_tokens=512,
|
| 120 |
task="conversational",
|
|
|
|
| 57 |
# Setup the LangChain HuggingFaceEndpoint and ChatHuggingFace LLM
|
| 58 |
deep_seek_model = HuggingFaceEndpoint(
|
| 59 |
repo_id="deepseek-ai/DeepSeek-R1",
|
|
|
|
| 60 |
temperature=0.7,
|
| 61 |
max_new_tokens=100,
|
| 62 |
task="conversational",
|
|
|
|
| 113 |
# Initialize the LLaMA model from HuggingFace (via Nebius provider)
|
| 114 |
llama_model = HuggingFaceEndpoint(
|
| 115 |
repo_id="meta-llama/Llama-3.1-8B-Instruct",
|
|
|
|
| 116 |
temperature=0.7,
|
| 117 |
max_new_tokens=512,
|
| 118 |
task="conversational",
|