Our flagship CHAT models
Collection
This is a collection of all our flagship chat models to make it more easy for you! • 3 items • Updated
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf LH-Tech-AI/Apex-1.6-Instruct-350M# Run inference directly in the terminal:
llama-cli -hf LH-Tech-AI/Apex-1.6-Instruct-350M# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf LH-Tech-AI/Apex-1.6-Instruct-350M# Run inference directly in the terminal:
./llama-cli -hf LH-Tech-AI/Apex-1.6-Instruct-350Mgit clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf LH-Tech-AI/Apex-1.6-Instruct-350M# Run inference directly in the terminal:
./build/bin/llama-cli -hf LH-Tech-AI/Apex-1.6-Instruct-350Mdocker model run hf.co/LH-Tech-AI/Apex-1.6-Instruct-350MImproved facts, chat instructions and coding!
Hey there! Welcome to Apex 1.6 Instruct 350M! We finetuned our base model of Apex 1 350M again! We now used a 2:1 ratio between Alpaca-Cleaned and Fineweb-Edu-10BT for anti-forgetting.
You can download the apex_1.6.gguf or use ollama run hf.co/LH-Tech-AI/Apex-1.6-Instruct-350M. And you can also use it in LM Studio for example, just by searching for "Apex 1.6".
Have fun! :D
We're not able to determine the quantization variants.
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf LH-Tech-AI/Apex-1.6-Instruct-350M# Run inference directly in the terminal: llama-cli -hf LH-Tech-AI/Apex-1.6-Instruct-350M