Spaces:
Running
Running
| title: hf-hub-query | |
| emoji: π | |
| colorFrom: blue | |
| colorTo: indigo | |
| sdk: docker | |
| app_port: 7860 | |
| short_description: Raw fast-agent MCP server for HF Hub queries. | |
| # hf-hub-query | |
| This Space runs a raw-passthrough fast-agent MCP server backed by the custom Monty build used for Hugging Face Hub querying. | |
| The deployed card uses `tool_result_mode: passthrough`, so tool results are returned directly rather than rewritten by a second LLM pass. | |
| ## Auth | |
| This Space is configured for Hugging Face OAuth/token passthrough: | |
| - `FAST_AGENT_SERVE_OAUTH=hf` | |
| - `FAST_AGENT_OAUTH_SCOPES=inference-api` | |
| - `--instance-scope request` | |
| These are configured as Space settings: | |
| - Variables: | |
| - `FAST_AGENT_SERVE_OAUTH` | |
| - `FAST_AGENT_OAUTH_SCOPES` | |
| - `FAST_AGENT_OAUTH_RESOURCE_URL` | |
| - Secret: | |
| - `HF_TOKEN` (dummy startup token) | |
| Clients can either: | |
| - send `Authorization: Bearer <HF_TOKEN>` directly, or | |
| - use MCP OAuth discovery/auth flow | |
| ## Model | |
| The deployed card uses: | |
| - `hf.openai/gpt-oss-120b:cerebras` | |
| ## Main files | |
| - `hf-hub-query.md` β raw MCP card | |
| - `monty_api_tool_v2.py` β Hub query tool implementation | |
| - `_monty_codegen_shared.md` β shared codegen instructions | |
| - `wheels/pydantic_monty-0.0.7-cp313-cp313-manylinux_2_34_x86_64.whl` β custom Monty wheel | |
| ## Note on the Monty wheel | |
| The bundled wheel is a CPython 3.13 `manylinux_2_34_x86_64` build for better HF runtime compatibility. | |