openbmb/Ultra-FineWeb
Viewer
•
Updated
•
1.29B
•
33.1k
•
295
Large Language Models
Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts
InfLLM-V2: Dense-Sparse Switchable Attention for Seamless Short-to-Long Adaptation