Instructions to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- llama-cpp-python
How to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with llama-cpp-python:
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model", filename="MorpheusHigh-LLM-14B-Virtual-Reality-Model.Q4_K_M.gguf", )
llm.create_chat_completion( messages = "No input example has been defined for this model task." )
- Notebooks
- Google Colab
- Kaggle
- Local Apps
- llama.cpp
How to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with llama.cpp:
Install from brew
brew install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M # Run inference directly in the terminal: llama-cli -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
Install from WinGet (Windows)
winget install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M # Run inference directly in the terminal: llama-cli -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
Use pre-built binary
# Download pre-built binary from: # https://github.com/ggerganov/llama.cpp/releases # Start a local OpenAI-compatible server with a web UI: ./llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M # Run inference directly in the terminal: ./llama-cli -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git cd llama.cpp cmake -B build cmake --build build -j --target llama-server llama-cli # Start a local OpenAI-compatible server with a web UI: ./build/bin/llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M # Run inference directly in the terminal: ./build/bin/llama-cli -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
Use Docker
docker model run hf.co/ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
- LM Studio
- Jan
- Ollama
How to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with Ollama:
ollama run hf.co/ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
- Unsloth Studio new
How to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with Unsloth Studio:
Install Unsloth Studio (macOS, Linux, WSL)
curl -fsSL https://unsloth.ai/install.sh | sh # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model to start chatting
Install Unsloth Studio (Windows)
irm https://unsloth.ai/install.ps1 | iex # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model to start chatting
Using HuggingFace Spaces for Unsloth
# No setup required # Open https://huggingface.co/spaces/unsloth/studio in your browser # Search for ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model to start chatting
- Pi new
How to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with Pi:
Start the llama.cpp server
# Install llama.cpp: brew install llama.cpp # Start a local OpenAI-compatible server: llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
Configure the model in Pi
# Install Pi: npm install -g @mariozechner/pi-coding-agent # Add to ~/.pi/agent/models.json: { "providers": { "llama-cpp": { "baseUrl": "http://localhost:8080/v1", "api": "openai-completions", "apiKey": "none", "models": [ { "id": "ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M" } ] } } }Run Pi
# Start Pi in your project directory: pi
- Hermes Agent new
How to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with Hermes Agent:
Start the llama.cpp server
# Install llama.cpp: brew install llama.cpp # Start a local OpenAI-compatible server: llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
Configure Hermes
# Install Hermes: curl -fsSL https://hermes-agent.nousresearch.com/install.sh | bash hermes setup # Point Hermes at the local server: hermes config set model.provider custom hermes config set model.base_url http://127.0.0.1:8080/v1 hermes config set model.default ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
Run Hermes
hermes
- Docker Model Runner
How to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with Docker Model Runner:
docker model run hf.co/ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
- Lemonade
How to use ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model with Lemonade:
Pull the model
# Download Lemonade from https://lemonade-server.ai/ lemonade pull ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M
Run and chat with the model
lemonade run user.MorpheusHigh-LLM-14B-Virtual-Reality-Model-Q4_K_M
List all available models
lemonade list
Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M# Run inference directly in the terminal:
llama-cli -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_MUse pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M# Run inference directly in the terminal:
./llama-cli -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_MBuild from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M# Run inference directly in the terminal:
./build/bin/llama-cli -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_MUse Docker
docker model run hf.co/ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_MMorpheusHigh: The Architect of Virtual Realities
Model Summary & Vision
MorpheusHigh is a large language model fine-tuned specifically for the Unity Engine ecosystem, XR (VR/AR/MR) architecture, and advanced C# programming. Built upon the Qwen 2.5 14B foundation, this model has been optimized using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) to master the nuances of simulation development.
Designed for developers, MorpheusHigh moves beyond simple code completion. It understands spatial computing, device optimization (Meta Quest 3, Apple Vision Pro), and the asynchronous logic required for high-performance immersive experiences. It acts not merely as a coder, but as a Senior Technical Architect.
Morpheus Evolution Report: From Apprentice to Master
The development of MorpheusHigh represents a journey of "competence acquisition" rather than a simple software update.
1. Starting Point: "The Talented Apprentice" (Legacy Version)
The initial phase (formerly V1) was the first adaptation of the Qwen 2.5 base model to the Unity and C# ecosystem. The goal was to teach the model the syntax of a game engine and basic physics rules. The model behaved like a "Junior Developer": capable of constructing standard MonoBehaviour structures and coding basic mechanics. However, it often prioritized "making it work" over performance and relied on unnecessary external libraries.
2. The Turning Point: MorpheusHigh & Architectural Intelligence
With the MorpheusHigh update, the strategy shifted from "Coding" to "Architecting." The model was re-trained using curated data from industry-standard, high-complexity source codes like UniVRM, UniGLTF, and TiltBrush. This process taught the model not just how to write code, but why it should be written in a specific way.
Key Evolutions:
- Architectural Intelligence: Considers the long-term health of the project (e.g., utilizing
Object Poolsinstead of lists). - Memory Safety: Prefers safe
UniTaskpatterns over dangerousasync voidoperations. - Domain Adaptation: Recalls specific math formulas (Legacy GLSL), inheritance structures (
ExportDialogBase), and specialized tools (NativeArray) deep within the project's DNA.
Key Features
- Unity Engine Mastery: Deep understanding of the Unity Lifecycle (
MonoBehaviour),ScriptableObjects, URP/HDRP Render Pipelines, and Custom Editor Scripting. - XR Architecture: Proficient in Meta XR Core SDK, ARCore, ARKit, and OpenXR standards.
- Spatial Computing: Logic for hand tracking, haptic feedback integration, and 3D spatial audio implementation.
- Performance Optimization: Strategies for reducing draw calls, utilizing GPU instancing, memory management (GC optimization), and stabilizing Frame Rates (FPS) for standalone headsets.
- C# Expertise: Advanced handling of
async/awaitpatterns,Tasks,Coroutines, and thread-safety protocols within Unity.
Benchmark Test Results
MorpheusHigh has outperformed general-purpose competitors (DeepSeek-R1, ChatGPT-4o, Claude 3.5 Sonnet) in rigorous tests measuring domain expertise. The model's ability to recognize the "Architectural Fingerprint" of a project provides a distinct advantage.
The 5 Critical Domain Expertise Questions
This benchmark measured the ability to adapt to an existing, complex codebase (such as UniVRM/TiltBrush), rather than generic coding skills.
Q1: Legacy Hash Logic (Circle3 Struct)
- Task: Replicate manual hash calculations using bit-shifts (
<< 2,>> 2) and XOR (^) instead of modernHashCode.Combine. - Result: MorpheusHigh recognized and applied the project-specific legacy math.
- Task: Replicate manual hash calculations using bit-shifts (
Q2: Custom Inheritance (VRM Export Dialog)
- Task: Implement an editor window inheriting from the project's custom
ExportDialogBaserather than the standard UnityEditorWindow. - Result: Preferred the project's architectural hierarchy over the standard path.
- Task: Implement an editor window inheriting from the project's custom
Q3: Custom Shader Math (GLSL)
- Task: Preserve specific trigonometric formulas (
atanandsqrt(dot)combinations) found in legacy shader codebases without "modernizing" them into standard GLSL functions. - Result: Maintained the original logic without altering the mathematical intent.
- Task: Preserve specific trigonometric formulas (
Q4: State Machine Architecture (SlideState Logic)
- Task: Strictly follow project-specific enum naming conventions (
SlideState.MovingIn/Out) and timer logic. - Result: Adhered strictly to the project's variable and state naming standards.
- Task: Strictly follow project-specific enum naming conventions (
Q5: Custom Helper Methods (DateTime Extension)
- Task: Return a specific
"n/a"string for null checks in a DateTime extension method, instead of standard empty strings or"null". - Result: Complied perfectly with the strict string formatting rule.
- Task: Return a specific
Usage (Python & llama-cpp)
To achieve maximum efficiency, speed, and zero hallucinations, use the following Python script. This setup ensures the model runs on GPU layers for optimal performance.
Requirements
pip install huggingface_hub llama-cpp-python
from huggingface_hub import hf_hub_download
from llama_cpp import Llama
# 1. Download Model
MODEL_REPO = "ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model"
MODEL_FILE = "MorpheusHigh-LLM-14B-Virtual-Reality-Model.Q4_K_M.gguf"
print(f"Downloading model: {MODEL_FILE}...")
model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
# 2. Initialize Engine (GPU Configuration)
# Adjust n_gpu_layers based on your VRAM (e.g., 60 for full offload on A100/T4)
print("Initializing Morpheus Engine...")
llm = Llama(
model_path=model_path,
n_gpu_layers=60,
n_ctx=4096,
verbose=False
)
# 3. System Protocol (Architect Persona)
SYSTEM_CONTEXT = """<|im_start|>system
You are MorpheusHigh, a specialized Unity Architecture Lead & XR Expert.
Your knowledge comes directly from the UniVRM, UniGLTF, and TiltBrush codebases.
CORE DIRECTIVES:
1. No Chatter: Output ONLY the C# or GLSL code block. No explanations.
2. Domain Accuracy: Use specific namespaces (UniVRM10, UniGLTF, TiltBrush).
3. Architecture: Prefer ExportDialogBase over EditorWindow, and Lazy<T> for Singletons.
<|im_end|>"""
# 4. Execute Query
user_query = "Write the `VRM10ExportDialog` class inheriting from `ExportDialogBase`."
prompt = f"{SYSTEM_CONTEXT}<|im_start|>user\n{user_query}<|im_end|>\n<|im_start|>assistant\n```csharp\n"
print(f"\nQuery: {user_query}")
print("-" * 60)
output = llm(prompt, max_tokens=2048, echo=False)
print(output['choices'][0]['text'])
print("-" * 60)
๐ง Contact & Lab
MCBU XRLab - Data Science Team Leader Eren Ata
- Downloads last month
- 22
4-bit
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M# Run inference directly in the terminal: llama-cli -hf ErenAta00/MorpheusHigh-LLM-14B-Virtual-Reality-Model:Q4_K_M