How to use from
llama.cppInstall from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf RockSky1/Infinity_1.0# Run inference directly in the terminal:
llama-cli -hf RockSky1/Infinity_1.0Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf RockSky1/Infinity_1.0# Run inference directly in the terminal:
./llama-cli -hf RockSky1/Infinity_1.0Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf RockSky1/Infinity_1.0# Run inference directly in the terminal:
./build/bin/llama-cli -hf RockSky1/Infinity_1.0Use Docker
docker model run hf.co/RockSky1/Infinity_1.0Quick Links
βΎοΈ Infinity 1.0 (Llama-3-8B GGUF)
Developed by: RockSky1
Model Type: Causal Language Model
Base Model: Meta-Llama-3-8B
Format: GGUF (Quantized for efficiency)
π Overview
Infinity 1.0 is a high-performance, fine-tuned version of the Llama-3-8B architecture. This model is designed to be the "Brain" of the Infinity AI ecosystem, offering fast, creative, and technically sound responses. It has been optimized for local deployment and low-latency interactions.
β¨ Key Features
- Optimized Architecture: Fine-tuned over multiple epochs (v5 development cycle) for superior reasoning.
- GGUF Format: Ready for offline use in LM Studio, Ollama, and mobile LLM runners.
- Quantized Precision: Balanced performance-to-size ratio using Q4_K_M quantization.
- Coding & Logic: Strong capabilities in full-stack development and architectural logic.
π οΈ How to Use
You can use this model offline using any GGUF-compatible runner:
- LM Studio: Search for
RockSky1/Infinity_1.0and download. - Ollama: Create a Modelfile and point it to the
.gguffile. - Mobile: Load via Layla or MLC LLM apps.
π License
This model follows the Meta Llama 3 Community License.
Created with β€οΈ by Shivam Kumar (RockSky1)
- Downloads last month
- 262
Hardware compatibility
Log In to add your hardware
We're not able to determine the quantization variants.
Inference Providers NEW
This model isn't deployed by any Inference Provider. π Ask for provider support
Model tree for RockSky1/Infinity_1.0
Base model
meta-llama/Meta-Llama-3-8B
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf RockSky1/Infinity_1.0# Run inference directly in the terminal: llama-cli -hf RockSky1/Infinity_1.0