🤖 Agentic AI Suite (Qwen3-1.7B-Reasoning)

This model is a Small Reasoning Model (SRM) fine-tuned specifically for the "Goldilocks Zone" of AI deployment: Powerful enough to handle PhD-level research tasks and fast enough to run on a standard Hugging Face Free Tier CPU.

By utilizing the Qwen3-1.7B architecture, this model achieves high-logic reasoning while maintaining a tiny memory footprint (~1.1GB in 4-bit).

🌟 Features

  • Native Tool Calling: Trained to output structured Action calls for web_search and calculator.
  • ReAct Framework: Uses a "Thought -> Action -> Observation -> Final Answer" loop.
  • Safetensors Format: Merged 4-bit weights for instant loading and high-speed CPU inference via transformers.
  • Zero-Latency Logic: Optimized to respond in under 10 seconds on a 2-vCPU environment.

🛠️ How to Use (Agentic Implementation)

To use this model as a true agent, your code should intercept the Action: text and execute the corresponding Python function.

1. Requirements

pip install transformers torch accelerate bitsandbytes
Downloads last month
44
Safetensors
Model size
2B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support