Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf AGofficial/AGI:Q3_K_S# Run inference directly in the terminal:
llama-cli -hf AGofficial/AGI:Q3_K_SUse pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf AGofficial/AGI:Q3_K_S# Run inference directly in the terminal:
./llama-cli -hf AGofficial/AGI:Q3_K_SBuild from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf AGofficial/AGI:Q3_K_S# Run inference directly in the terminal:
./build/bin/llama-cli -hf AGofficial/AGI:Q3_K_SUse Docker
docker model run hf.co/AGofficial/AGI:Q3_K_SAGI
Artificial General Intelligence
Research Paper
This paper delineates a comprehensive architectural framework for the progressive realization of Artificial General Intelligence (AGI), predicated upon a modular-agentic paradigm. We present a system design that integrates sophisticated tool-use capabilities, hierarchical memory management, dynamic code execution, and nascent world-modeling functionalities. The proposed architecture, exemplified through a lightweight Qwen3-4B-Instruct-2507-Q3_K_S.gguf model, demonstrates a robust foundation for emergent cognitive properties such as autonomy, recursive self-improvement, and goal-oriented behavior. Furthermore, we explore the theoretical underpinnings of consciousness as an emergent property within complex neural architectures and postulate pathways towards super-intelligence through advanced computational and embodied interaction modalities. The exposition maintains a rigorous academic tone, employing advanced terminology to articulate the intricate conceptual and technical facets of AGI development.
Implementation
A simple implementation of the paper's code can be found in this repository.
- Downloads last month
- 6
3-bit
Model tree for AGofficial/AGI
Base model
Qwen/Qwen3-4B-Instruct-2507
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf AGofficial/AGI:Q3_K_S# Run inference directly in the terminal: llama-cli -hf AGofficial/AGI:Q3_K_S