Best Papers (4):
Artificial Hivemind: The Open-Ended Homogeneity of Language Models (and Beyond)
PDF: https://openreview.net/pdf?id=saDOrrnNTz
Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free
PDF: https://openreview.net/pdf?id=1b7whO4SfY
1000 Layer Networks for Self-Supervised RL: Scaling Depth Can Enable New Goal-Reaching Capabilities
PDF: https://openreview.net/pdf?id=s0JVsx3bx1
Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training
PDF: https://openreview.net/pdf?id=BSZqpqgqM0
Runners Up (3):
Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
PDF: https://openreview.net/pdf?id=4OsgYD7em5
Optimal Mistake Bounds for Transductive Online Learning
PDF: https://openreview.net/pdf?id=EoebmBe9fG
Superposition Yields Robust Neural Scaling