random 60s crop at training time (matches Side-Step chunk-duration), remove pre-split chunking d3618ec Nekochu commited on 9 days ago
audio-level chunking (not latent), auto-scale epochs for chunk count 1ee8f1f Nekochu commited on 10 days ago
chunk latents into ~30s segments for faster CPU training, energy-aware boundaries 2e395ab Nekochu commited on 11 days ago
fix adapter save path, smart LM fallback, compact training UI, remove Server Status 35fbf3e Nekochu commited on 12 days ago
cancel, captioning, preprocessing, sidecar upload, elapsed time, GeneratorExit fix 32de701 Nekochu commited on 13 days ago
fix: save PEFT adapter (not full model), remove random suffix from LoRA names, fix epoch cap to 1000 57df0f6 Nekochu commited on 13 days ago
fix all review issues: dedup sampling/unwrap, thread-safe lock, cleanup, retry, security docs 829ed0c Nekochu commited on 14 days ago
fix inference: add LM generation step, detokenize codes before DiT, full pipeline working ff9f4ad Nekochu commited on 14 days ago
fix understand_audio: clone tensors for inference mode, working on GPU (52s) 4b2f4ad Nekochu commited on 14 days ago
add understand_audio (LM reverse), demucs-infer fix, commit refs, dtype fixes 6bfdc38 Nekochu commited on 14 days ago
major update: PyTorch inference, Gradio 6, session isolation, /understand captioning ff239f5 Nekochu commited on 14 days ago
SDPA first on Blackwell, FA2 only for Ampere/Hopper, txt caption support 04ccf32 Nekochu commited on 15 days ago
add GPU/CUDA auto-detect, mixed precision, flash_attn, txt caption parser 917e4ed Nekochu commited on 15 days ago
add mid/sas analysis modes (Demucs + ensemble), auto-select by dataset size b38d0b1 Nekochu commited on 15 days ago
add auto-captioning (BPM/key/signature via librosa), add librosa+mutagen deps 1d42836 Nekochu commited on 15 days ago
fix: adapter saved to clean dir, LM dropdown no 'Default', on-demand download e62602f Nekochu commited on 15 days ago