openbmb/NOSA-1B
Updated
•
43
•
4
Large Language Models
Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts
InfLLM-V2: Dense-Sparse Switchable Attention for Seamless Short-to-Long Adaptation