Instructions to use ORCA-AI/ORCA1-EXP-0213 with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- llama-cpp-python
How to use ORCA-AI/ORCA1-EXP-0213 with llama-cpp-python:
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="ORCA-AI/ORCA1-EXP-0213", filename="ORCA1-EXP-1302.gguf", )
llm.create_chat_completion( messages = "No input example has been defined for this model task." )
- Notebooks
- Google Colab
- Kaggle
- Local Apps
- llama.cpp
How to use ORCA-AI/ORCA1-EXP-0213 with llama.cpp:
Install from brew
brew install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf ORCA-AI/ORCA1-EXP-0213 # Run inference directly in the terminal: llama-cli -hf ORCA-AI/ORCA1-EXP-0213
Install from WinGet (Windows)
winget install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf ORCA-AI/ORCA1-EXP-0213 # Run inference directly in the terminal: llama-cli -hf ORCA-AI/ORCA1-EXP-0213
Use pre-built binary
# Download pre-built binary from: # https://github.com/ggerganov/llama.cpp/releases # Start a local OpenAI-compatible server with a web UI: ./llama-server -hf ORCA-AI/ORCA1-EXP-0213 # Run inference directly in the terminal: ./llama-cli -hf ORCA-AI/ORCA1-EXP-0213
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git cd llama.cpp cmake -B build cmake --build build -j --target llama-server llama-cli # Start a local OpenAI-compatible server with a web UI: ./build/bin/llama-server -hf ORCA-AI/ORCA1-EXP-0213 # Run inference directly in the terminal: ./build/bin/llama-cli -hf ORCA-AI/ORCA1-EXP-0213
Use Docker
docker model run hf.co/ORCA-AI/ORCA1-EXP-0213
- LM Studio
- Jan
- Ollama
How to use ORCA-AI/ORCA1-EXP-0213 with Ollama:
ollama run hf.co/ORCA-AI/ORCA1-EXP-0213
- Unsloth Studio new
How to use ORCA-AI/ORCA1-EXP-0213 with Unsloth Studio:
Install Unsloth Studio (macOS, Linux, WSL)
curl -fsSL https://unsloth.ai/install.sh | sh # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for ORCA-AI/ORCA1-EXP-0213 to start chatting
Install Unsloth Studio (Windows)
irm https://unsloth.ai/install.ps1 | iex # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for ORCA-AI/ORCA1-EXP-0213 to start chatting
Using HuggingFace Spaces for Unsloth
# No setup required # Open https://huggingface.co/spaces/unsloth/studio in your browser # Search for ORCA-AI/ORCA1-EXP-0213 to start chatting
- Docker Model Runner
How to use ORCA-AI/ORCA1-EXP-0213 with Docker Model Runner:
docker model run hf.co/ORCA-AI/ORCA1-EXP-0213
- Lemonade
How to use ORCA-AI/ORCA1-EXP-0213 with Lemonade:
Pull the model
# Download Lemonade from https://lemonade-server.ai/ lemonade pull ORCA-AI/ORCA1-EXP-0213
Run and chat with the model
lemonade run user.ORCA1-EXP-0213-{{QUANT_TAG}}List all available models
lemonade list
# !pip install llama-cpp-python
from llama_cpp import Llama
llm = Llama.from_pretrained(
repo_id="ORCA-AI/ORCA1-EXP-0213",
filename="ORCA1-EXP-1302.gguf",
)
llm.create_chat_completion(
messages = "No input example has been defined for this model task."
)ORCA 1-EXP-0213
ORCA 1-EXP-0213 is an experimental, uncensored AI model developed by ORCA AI Labs.
It is optimized for speed rather than deep reasoning and is designed primarily for Czech-language conversations.
This model is uncensored, meaning it operates without artificial restrictions, making it ideal for open-ended discussions.
Features
- Fast and Efficient – Optimized for low-latency inference.
- Uncensored – No artificial restrictions or content filtering.
- Czech Language Support – Primarily designed for Czech, with some English capability.
- Lightweight – Easy to deploy and run efficiently.
Model Details
| Property | Details |
|---|---|
| Type | Experimental Conversational AI |
| Speed | Optimized for fast responses |
| Intelligence | Average (not designed for complex reasoning) |
| Censorship | Uncensored |
| Best for | Open-ended conversations, quick replies |
| Limitations | Not suited for deep contextual understanding or logic-heavy tasks |
Usage on Hugging Face
ORCA 1-EXP-0213 is available on Hugging Face. You can load and run it using Python:
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "ORCA-AI/ORCA1-EXP-0213"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
input_text = "Jaké je hlavní město České republiky?"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Performance Overview
| Metric | ORCA 1-EXP-0213 | ORCA 2-Turbo |
|---|---|---|
| Speed | Extremely fast | Faster than ORCA 1 |
| Intelligence | Average | Designed for deep reasoning |
| Censorship | Unfiltered | Unfiltered |
| Language | Czech & some English | Primarily English |
| Use Case | Conversational AI, quick replies | Complex tasks, in-depth responses |
This model prioritizes speed over deep reasoning. It is well-suited for casual conversations in Czech but may not perform well in fact-heavy or complex discussions.
Limitations & Warnings
- Uncensored – The model does not apply moderation. Use responsibly.
- Not Designed for Deep Reasoning – Works well for casual interactions but may struggle with complex logic.
- Experimental – Expect inconsistencies and updates over time.
Fine-Tuning & Customization
If you need to fine-tune ORCA 1-EXP-0213, you can do so using Hugging Face's Trainer:
from transformers import Trainer, TrainingArguments
training_args = TrainingArguments(
output_dir="./orca1-exp-0213-finetuned",
per_device_train_batch_size=4,
gradient_accumulation_steps=8,
evaluation_strategy="steps",
save_total_limit=2,
)
trainer = Trainer(
model=model,
args=training_args,
train_dataset=your_dataset,
eval_dataset=your_eval_dataset,
)
trainer.train()
Get Involved
For feedback, contributions, or inquiries, reach out to ORCA AI Labs:
- Website: ORCA AI Labs
- Twitter: @ORCA_AI_Labs
License
ORCA 1-EXP-0213 is released under the ORCA Experimental License.
Full terms can be found here.
ORCA AI Labs – Advancing Open AI Research
- Downloads last month
- -
We're not able to determine the quantization variants.
# Gated model: Login with a HF token with gated access permission hf auth login