How to use from
llama.cpp
Install from brew
brew install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf TobDeBer/myContainers
# Run inference directly in the terminal:
llama-cli -hf TobDeBer/myContainers
Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf TobDeBer/myContainers
# Run inference directly in the terminal:
llama-cli -hf TobDeBer/myContainers
Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases
# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf TobDeBer/myContainers
# Run inference directly in the terminal:
./llama-cli -hf TobDeBer/myContainers
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli
# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf TobDeBer/myContainers
# Run inference directly in the terminal:
./build/bin/llama-cli -hf TobDeBer/myContainers
Use Docker
docker model run hf.co/TobDeBer/myContainers
Quick Links

TobDeBer/arco-Q4_K_M-GGUF

This model was converted to Big Endian Q4_K_M GGUF format from appvoid/arco using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.

Container Repository for CPU adaptations of Inference code

Variants for Inference

Slim container

  • run std binaries

CPUdiffusion

  • inference diffusion models on CPU
  • include CUDAonCPU stack

Diffusion container

  • run diffusion app.py variants
  • support CPU and CUDA
  • include Flux

Slim CUDA container

  • run CUDA binaries

Variants for Build

Llama.cpp build container

  • build llama-cli-static
  • build llama-server-static

sd build container

  • build sd
  • optional: build sd-server

CUDA build container

  • build cuda binaries
  • support sd_cuda
Downloads last month
8
GGUF
Model size
0.5B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for TobDeBer/myContainers

Base model

appvoid/arco
Quantized
(7)
this model