REVE: A Foundation Model for EEG -- Adapting to Any Setup with Large-Scale Pretraining on 25,000 Subjects Paper • 2510.21585 • Published Oct 24, 2025 • 7
D5P4: Partition Determinantal Point Process for Diversity in Parallel Discrete Diffusion Decoding Paper • 2603.19146 • Published 8 days ago
Inner Loop Inference for Pretrained Transformers: Unlocking Latent Capabilities Without Training Paper • 2602.14759 • Published Feb 16
Residual Connections and the Causal Shift: Uncovering a Structural Misalignment in Transformers Paper • 2602.14760 • Published Feb 16
REVE Collection Model collection of the "REVE: A Foundation Model for EEG - Adapting to Any Setup with Large-Scale Pretraining on 25,000 Subjects" paper. • 5 items • Updated 11 days ago • 13
Scaling Text-to-Image Diffusion Transformers with Representation Autoencoders Paper • 2601.16208 • Published Jan 22 • 55
The Flexibility Trap: Why Arbitrary Order Limits Reasoning Potential in Diffusion Language Models Paper • 2601.15165 • Published Jan 21 • 73
Dynamic Large Concept Models: Latent Reasoning in an Adaptive Semantic Space Paper • 2512.24617 • Published Dec 31, 2025 • 65
Next-Embedding Prediction Makes Strong Vision Learners Paper • 2512.16922 • Published Dec 18, 2025 • 88
LLaDA2.0: Scaling Up Diffusion Language Models to 100B Paper • 2512.15745 • Published Dec 10, 2025 • 88