Instructions to use alvanalrakib/aurora-lyrics with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- llama-cpp-python
How to use alvanalrakib/aurora-lyrics with llama-cpp-python:
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="alvanalrakib/aurora-lyrics", filename="aurora-lyrics-BF16.gguf", )
output = llm( "Once upon a time,", max_tokens=512, echo=True ) print(output)
- Notebooks
- Google Colab
- Kaggle
- Local Apps
- llama.cpp
How to use alvanalrakib/aurora-lyrics with llama.cpp:
Install from brew
brew install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf alvanalrakib/aurora-lyrics:BF16 # Run inference directly in the terminal: llama-cli -hf alvanalrakib/aurora-lyrics:BF16
Install from WinGet (Windows)
winget install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf alvanalrakib/aurora-lyrics:BF16 # Run inference directly in the terminal: llama-cli -hf alvanalrakib/aurora-lyrics:BF16
Use pre-built binary
# Download pre-built binary from: # https://github.com/ggerganov/llama.cpp/releases # Start a local OpenAI-compatible server with a web UI: ./llama-server -hf alvanalrakib/aurora-lyrics:BF16 # Run inference directly in the terminal: ./llama-cli -hf alvanalrakib/aurora-lyrics:BF16
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git cd llama.cpp cmake -B build cmake --build build -j --target llama-server llama-cli # Start a local OpenAI-compatible server with a web UI: ./build/bin/llama-server -hf alvanalrakib/aurora-lyrics:BF16 # Run inference directly in the terminal: ./build/bin/llama-cli -hf alvanalrakib/aurora-lyrics:BF16
Use Docker
docker model run hf.co/alvanalrakib/aurora-lyrics:BF16
- LM Studio
- Jan
- Ollama
How to use alvanalrakib/aurora-lyrics with Ollama:
ollama run hf.co/alvanalrakib/aurora-lyrics:BF16
- Unsloth Studio new
How to use alvanalrakib/aurora-lyrics with Unsloth Studio:
Install Unsloth Studio (macOS, Linux, WSL)
curl -fsSL https://unsloth.ai/install.sh | sh # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for alvanalrakib/aurora-lyrics to start chatting
Install Unsloth Studio (Windows)
irm https://unsloth.ai/install.ps1 | iex # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for alvanalrakib/aurora-lyrics to start chatting
Using HuggingFace Spaces for Unsloth
# No setup required # Open https://huggingface.co/spaces/unsloth/studio in your browser # Search for alvanalrakib/aurora-lyrics to start chatting
- Docker Model Runner
How to use alvanalrakib/aurora-lyrics with Docker Model Runner:
docker model run hf.co/alvanalrakib/aurora-lyrics:BF16
- Lemonade
How to use alvanalrakib/aurora-lyrics with Lemonade:
Pull the model
# Download Lemonade from https://lemonade-server.ai/ lemonade pull alvanalrakib/aurora-lyrics:BF16
Run and chat with the model
lemonade run user.aurora-lyrics-BF16
List all available models
lemonade list
Aurora Lyrics GGUF
Aurora Lyrics is a GGUF release of a full fine-tuned Qwen3 8B lyrics model.
This release was exported from:
checkpoint-12000
Training progress at export time:
Training steps completed: 12,000
Planned 1-epoch total steps: 49,042
Approximate epoch progress: ~24.5%
Dataset size: ~250K samples
This is not the final full-epoch checkpoint. It is a mid-training checkpoint exported for testing, local generation, and quality comparison.
Files
aurora-lyrics-BF16.gguf # BF16 master GGUF, highest fidelity, about 16 GB
aurora-lyrics-Q8_0.gguf # Q8_0 quantized GGUF, high quality, about 8 GB
Use aurora-lyrics-Q8_0.gguf for normal local generation and testing.
Use aurora-lyrics-BF16.gguf when you want the highest fidelity and have enough VRAM/RAM.
Model Type
Aurora Lyrics is a base lyrics-completion model, not an instruction/chat model.
For best results, do not use chat messages like:
user:
assistant:
Also avoid ChatML wrappers or assistant-style prompting unless your runtime system specifically requires them.
The model was trained with a simple structured lyrics prefix ending in:
Lyrics:
After that, the model should continue with song lyrics.
Important Prompt Format
Minimum format:
Artist: <artist or virtual artist>
Lyrics:
Recommended full template:
Artist: <artist name or virtual artist>
Mood: <emotion or vibe>
Genre: <style / genre / production direction>
Title: <optional song title>
Lyrics:
Example:
Artist: Virtual Pop Artist
Mood: bittersweet romantic
Genre: emotional pop, acoustic guitar, big chorus
Title: Midnight Letters
Lyrics:
The model should continue from Lyrics: with the generated song.
Continuing Seed Lines
To continue from existing lyrics, put the seed lines immediately after Lyrics:.
Artist: Virtual Pop Artist
Mood: sad romantic
Genre: cinematic piano pop, big chorus
Title: Rain On My Phone
Lyrics:
[Verse]
I saw your name light up the glass
Like a ghost from the summer past
The model should continue from the existing lines.
Better Generation Behavior
Use sampling. Greedy decoding or very low temperature can make outputs repeat or start too similarly.
Good starting settings:
temperature: 0.80 - 0.95
top_p: 0.90 - 0.95
top_k: 40 - 80
repeat_penalty: 1.08 - 1.15
max_tokens: 500 - 900
context: 4096 - 8192
For more variety, increase temperature toward 0.95 and use a different seed.
llama.cpp CLI Example
Example using the Q8_0 GGUF:
llama-cli \
-m aurora-lyrics-Q8_0.gguf \
-ngl 999 \
-c 8192 \
-n 700 \
--temp 0.85 \
--top-p 0.92 \
--top-k 50 \
--repeat-penalty 1.10 \
-p $'Artist: Virtual Pop Artist\nMood: heartbroken\nGenre: emotional pop ballad, piano, big chorus\nTitle: After The Rain\nLyrics:\n'
llama.cpp Server Example
llama-server \
-m aurora-lyrics-Q8_0.gguf \
--host 0.0.0.0 \
--port 8080 \
-ngl 999 \
-c 8192 \
--temp 0.85 \
--top-p 0.92 \
--top-k 50 \
--repeat-penalty 1.10
Then send prompts using the structured lyrics format:
Artist: Virtual Pop Artist
Mood: emotional, lonely, cinematic
Genre: modern pop ballad, piano, big chorus
Title: Alone At Midnight
Lyrics:
llama-cpp-python Example
from llama_cpp import Llama
llm = Llama(
model_path="aurora-lyrics-Q8_0.gguf",
n_gpu_layers=-1,
n_ctx=8192,
flash_attn=True,
verbose=False,
)
prompt = """Artist: Virtual Pop Artist
Mood: sad romantic
Genre: cinematic piano pop, big chorus
Title: Rain On My Phone
Lyrics:
"""
out = llm(
prompt,
max_tokens=700,
temperature=0.85,
top_p=0.92,
top_k=50,
repeat_penalty=1.10,
)
print(out["choices"][0]["text"])
Optional Runtime Instruction
A runtime or web UI may add a short guidance line after Lyrics: for more controlled structure:
Artist: Virtual Pop Artist
Mood: emotional, lonely, cinematic
Genre: modern pop ballad, piano, big chorus
Title: Alone At Midnight
Lyrics:
Write structured song lyrics with sections like [Verse], [Pre-Chorus], [Chorus], and [Bridge]. Keep it singable, emotional, and pop-focused.
That can help guide formatting, but the core training format is still the structured header ending in Lyrics:.
What To Avoid
Avoid plain prompts like:
Write an emotional pop chorus about missing someone in the rain.
That can work sometimes, but it is not the best format for this model.
Avoid chat format:
User: write lyrics
Assistant:
Use the structured lyric header instead:
Artist: Virtual Pop Artist
Mood: emotional
Genre: pop ballad
Title: Missing You In The Rain
Lyrics:
Notes
This checkpoint was exported at 12,000 training steps out of a planned 49,042-step single epoch.
Later checkpoints may become stronger as training continues, so this release should be treated as a high-quality testing checkpoint rather than the final model.
- Downloads last month
- 73
8-bit
16-bit