S0 Tuning: Zero-Overhead Adaptation of Hybrid Recurrent-Attention Models Paper • 2604.01168 • Published about 1 month ago • 7
ThinkTwice: Jointly Optimizing Large Language Models for Reasoning and Self-Refinement Paper • 2604.01591 • Published about 1 month ago • 42
How Well Do Agentic Skills Work in the Wild: Benchmarking LLM Skill Usage in Realistic Settings Paper • 2604.04323 • Published 26 days ago • 41
Attention Sink in Transformers: A Survey on Utilization, Interpretation, and Mitigation Paper • 2604.10098 • Published 21 days ago • 78
KnowRL: Boosting LLM Reasoning via Reinforcement Learning with Minimal-Sufficient Knowledge Guidance Paper • 2604.12627 • Published 18 days ago • 100
Claw-Eval: Toward Trustworthy Evaluation of Autonomous Agents Paper • 2604.06132 • Published 25 days ago • 119
SkillClaw: Let Skills Evolve Collectively with Agentic Evolver Paper • 2604.08377 • Published 23 days ago • 287
view article Article Training and Finetuning Multimodal Embedding & Reranker Models with Sentence Transformers 16 days ago • 66
view article Article PangolinGuard: Fine-Tuning ModernBERT as a Lightweight Approach to AI Guardrails Mar 23, 2025 • 13