Commit History

add fast captioning module (CLAP + faster-whisper + Silero VAD), update deps
4619f39

Nekochu commited on

random 60s crop at training time (matches Side-Step chunk-duration), remove pre-split chunking
d3618ec

Nekochu commited on

audio-level chunking (not latent), auto-scale epochs for chunk count
1ee8f1f

Nekochu commited on

chunk latents into ~30s segments for faster CPU training, energy-aware boundaries
2e395ab

Nekochu commited on

default rank 16 for faster CPU training
81f54b1

Nekochu commited on

add credit lines to both tabs
afe7cf4

Nekochu commited on

compact Generate Music UI, same layout style as Train tab
4d42fae

Nekochu commited on

move audio upload to left column
8d2b494

Nekochu commited on

skip bare librosa sidecar, let preprocessing faf analysis handle caption fallback
53f6566

Nekochu commited on

fix adapter save path, smart LM fallback, compact training UI, remove Server Status
35fbf3e

Nekochu commited on

cancel, captioning, preprocessing, sidecar upload, elapsed time, GeneratorExit fix
32de701

Nekochu commited on

fix review: debug leak, int crash, rank mismatch, 0-byte skip, log cap, understand diag
4d9a556

Nekochu commited on

fix: save PEFT adapter (not full model), remove random suffix from LoRA names, fix epoch cap to 1000
57df0f6

Nekochu commited on

remove XL checkpoint download (OOMKilled build, training uses standard turbo)
6d9fb39

Nekochu commited on

fix: save_every_n_epochs=0, add demucs-infer to Dockerfile, debug adapter dir
0e27e49

Nekochu commited on

fix all review issues: dedup sampling/unwrap, thread-safe lock, cleanup, retry, security docs
829ed0c

Nekochu commited on

update README with final state, full pipeline inference, LM generation step
a5741b1

Nekochu commited on

fix inference: add LM generation step, detokenize codes before DiT, full pipeline working
ff9f4ad

Nekochu commited on

add _is_space flag, block inference during training, understand clone fix
3c15b8b

Nekochu commited on

fix understand_audio: clone tensors for inference mode, working on GPU (52s)
4b2f4ad

Nekochu commited on

add understand_audio (LM reverse), demucs-infer fix, commit refs, dtype fixes
6bfdc38

Nekochu commited on

major update: PyTorch inference, Gradio 6, session isolation, /understand captioning
ff239f5

Nekochu commited on

truncate long files to fit cap, show which files truncated/skipped
bc97006

Nekochu commited on

accept files until total audio cap reached, skip rest with warning
956dc8c

Nekochu commited on

replace per-file cap with total audio cap (30 min max)
1549b91

Nekochu commited on

add LoRA download button after training (gr.File output, like rvc-beatrice)
2d3c27c

Nekochu commited on

remove ace-server understand proxy, captioning stays librosa + txt sidecars
5b7a56f

Nekochu commited on

add /understand API endpoint for GGUF audio captioning
a4457c3

Nekochu commited on

SDPA first on Blackwell, FA2 only for Ampere/Hopper, txt caption support
04ccf32

Nekochu commited on

add GPU/CUDA auto-detect, mixed precision, flash_attn, txt caption parser
917e4ed

Nekochu commited on

update defaults: LR 3e-4, rank 32, alpha 2x rank (per Side-Step author)
04c031f

Nekochu commited on

save only on cancel/finish, max epochs 1000, default 3
d42aa91

Nekochu commited on

add mid/sas analysis modes (Demucs + ensemble), auto-select by dataset size
b38d0b1

Nekochu commited on

add auto-captioning (BPM/key/signature via librosa), add librosa+mutagen deps
1d42836

Nekochu commited on

switch training to standard turbo (11s/epoch), auto-select standard GGUF for LoRA inference
c0f2a13

Nekochu commited on

add XL variant mappings to _VARIANT_DIR
3c5ce9c

Nekochu commited on

fix: train on XL turbo (matches XL GGUF for inference), add XL checkpoint download
372f08e

Nekochu commited on

add full README with API docs, MCP, CLI, architecture
9d2d424

Nekochu commited on

allow custom LoRA values in dropdown (API clients)
2bd2612

Nekochu commited on

log ace-server restart, show output for adapter debugging
d2ae079

Nekochu commited on

fix: forward adapter to synth request, default LM to 1.7B
b23b6b8

Nekochu commited on

fix: adapter saved to clean dir, LM dropdown no 'Default', on-demand download
e62602f

Nekochu commited on

copy train_engine.py into Docker image
5fe3c53

Nekochu commited on

Side-Step training engine, tested locally on CPU
a07b39d

Nekochu commited on

switch back to 1.7B LM (fastest at 269s, 0.6B was 936s)
5e95353

Nekochu commited on

swap LM to 0.6B Q8_0 for speed test
88b9223

Nekochu commited on

swap LM 4B->1.7B Q8_0 for faster CPU inference
b14d3e8

Nekochu commited on

add LoRA adapter dropdown to inference UI
5c2e4e7

Nekochu commited on

default mp3, remove format selector, increase LM timeout to 900s
882ed5c

Nekochu commited on

remove accelerate (causes meta tensors), clean up patches
153f929

Nekochu commited on