WTF GENIUS PAPERS Collection Papers that made me appreciate my major and my life a little more. obs=Observation, innov=Innovation. Most papers are abt improving tiny models. ⢠124 items ⢠Updated 1 day ago ⢠19
RegMix: Data Mixture as Regression for Language Model Pre-training Paper ⢠2407.01492 ⢠Published Jul 1, 2024 ⢠41
SD-E^2: Semantic Exploration for Reasoning Under Token Budgets Paper ⢠2601.17982 ⢠Published Jan 25 ⢠1
Reasoning Path Divergence: A New Metric and Curation Strategy to Unlock LLM Diverse Thinking Paper ⢠2510.26122 ⢠Published Jan 4 ⢠1
Unlocking Reasoning Capabilities in LLMs via Reinforcement Learning Exploration Paper ⢠2510.03865 ⢠Published Oct 4, 2025 ⢠1
Controllable Exploration in Hybrid-Policy RLVR for Multi-Modal Reasoning Paper ⢠2602.20197 ⢠Published Feb 22 ⢠1
Wider or Deeper? Scaling LLM Inference-Time Compute with Adaptive Branching Tree Search Paper ⢠2503.04412 ⢠Published Mar 6, 2025 ⢠6
Reasoning with Sampling: Your Base Model is Smarter Than You Think Paper ⢠2510.14901 ⢠Published Oct 16, 2025 ⢠49
Power-SMC: Low-Latency Sequence-Level Power Sampling for Training-Free LLM Reasoning Paper ⢠2602.10273 ⢠Published Mar 23 ⢠1
Sampling for Quality: Training-Free Reward-Guided LLM Decoding via Sequential Monte Carlo Paper ⢠2604.16453 ⢠Published Apr 7 ⢠1
Diversified Sampling Improves Scaling LLM inference Paper ⢠2502.11027 ⢠Published Feb 16, 2025 ⢠1
GuidedSampling: Steering LLMs Towards Diverse Candidate Solutions at Inference-Time Paper ⢠2510.03777 ⢠Published Feb 14 ⢠2
Large Language Monkeys: Scaling Inference Compute with Repeated Sampling Paper ⢠2407.21787 ⢠Published Jul 31, 2024 ⢠14
MiMo: Unlocking the Reasoning Potential of Language Model -- From Pretraining to Posttraining Paper ⢠2505.07608 ⢠Published May 12, 2025 ⢠86
WTF GENIUS PAPERS Collection Papers that made me appreciate my major and my life a little more. obs=Observation, innov=Innovation. Most papers are abt improving tiny models. ⢠124 items ⢠Updated 1 day ago ⢠19
WTF GENIUS PAPERS Collection Papers that made me appreciate my major and my life a little more. obs=Observation, innov=Innovation. Most papers are abt improving tiny models. ⢠124 items ⢠Updated 1 day ago ⢠19
From Growing to Looping: A Unified View of Iterative Computation in LLMs Paper ⢠2602.16490 ⢠Published Feb 18 ⢠1
view reply I really like this. It's how MoEs should have always worked, and it's fascinating that it actually succeeds in practice. Please submit this to Daily Papers so more people can see it:https://huggingface.co/papers/submit?paperId=2605.06663