Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -11,7 +11,7 @@ def respond(
|
|
| 11 |
hf_token: gr.OAuthToken,
|
| 12 |
):
|
| 13 |
"""
|
| 14 |
-
|
| 15 |
https://huggingface.co/docs/huggingface_hub/v0.22.2/en/guides/inference
|
| 16 |
"""
|
| 17 |
client = InferenceClient(token=hf_token.token, model="openai/gpt-oss-20b")
|
|
@@ -35,7 +35,7 @@ def respond(
|
|
| 35 |
yield response
|
| 36 |
|
| 37 |
"""
|
| 38 |
-
|
| 39 |
https://www.gradio.app/docs/chatinterface
|
| 40 |
"""
|
| 41 |
|
|
@@ -63,4 +63,3 @@ with gr.Blocks() as demo:
|
|
| 63 |
|
| 64 |
if __name__ == "__main__":
|
| 65 |
demo.launch()
|
| 66 |
-
|
|
|
|
| 11 |
hf_token: gr.OAuthToken,
|
| 12 |
):
|
| 13 |
"""
|
| 14 |
+
Для получения дополнительной информации о поддержке API `huggingface_hub` Inference, посмотрите документацию:
|
| 15 |
https://huggingface.co/docs/huggingface_hub/v0.22.2/en/guides/inference
|
| 16 |
"""
|
| 17 |
client = InferenceClient(token=hf_token.token, model="openai/gpt-oss-20b")
|
|
|
|
| 35 |
yield response
|
| 36 |
|
| 37 |
"""
|
| 38 |
+
Для информации о том, как настроить ChatInterface, изучите документацию Gradio:
|
| 39 |
https://www.gradio.app/docs/chatinterface
|
| 40 |
"""
|
| 41 |
|
|
|
|
| 63 |
|
| 64 |
if __name__ == "__main__":
|
| 65 |
demo.launch()
|
|
|