Audio-Text-to-Text
Transformers
Safetensors
step_audio_2
text-generation
audio-reasoning
chain-of-thought
multi-modal
step-audio-r1
custom_code
Instructions to use stepfun-ai/Step-Audio-R1.1 with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use stepfun-ai/Step-Audio-R1.1 with Transformers:
# Load model directly from transformers import AutoModelForCausalLM model = AutoModelForCausalLM.from_pretrained("stepfun-ai/Step-Audio-R1.1", trust_remote_code=True, dtype="auto") - Notebooks
- Google Colab
- Kaggle
Does the R1.1 model internally implement a VLLM-based inference framework?
#7
by 026jzz - opened
I'm working on the vllm-omni adaptation for this model, and a friend reminded me that you guys seem to have already implemented it, so I'm asking here.
btw, your presentation is amazing, I watched the Live recording through bilibili