Arxiv Papers
Podcast image
[QA] Babel: Open Multilingual Large Language Models Serving Over 90% of Global Speakers
6 mins; March 07, 2025
Babel: Open Multilingual Large Language Models Serving Over 90% of Global Speakers
15 mins; March 07, 2025
[QA] L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning
8 mins; March 07, 2025
L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning
15 mins; March 07, 2025
[QA] PokéChamp: an Expert-level Minimax Language Agent
8 mins; March 07, 2025
PokéChamp: an Expert-level Minimax Language Agent
26 mins; March 07, 2025
[QA] Token-Efficient Long Video Understanding for Multimodal LLMs
7 mins; March 07, 2025
Token-Efficient Long Video Understanding for Multimodal LLMs
26 mins; March 07, 2025
[QA] Position: Model Collapse Does Not Mean What You Think
7 mins; March 05, 2025
Position: Model Collapse Does Not Mean What You Think
21 mins; March 05, 2025
[QA] Towards Understanding Distilled Reasoning Models: A Representational Approach
7 mins; March 05, 2025
Towards Understanding Distilled Reasoning Models: A Representational Approach
10 mins; March 05, 2025
[QA] Weak-to-Strong Generalization Even in Random Feature Networks, Provably
6 mins; March 05, 2025
Weak-to-Strong Generalization Even in Random Feature Networks, Provably
15 mins; March 05, 2025
[QA] Beyond Cosine Decay: On the effectiveness of Infinite Learning Rate Schedule for Continual Pre-training
7 mins; March 05, 2025
Beyond Cosine Decay: On the effectiveness of Infinite Learning Rate Schedule for Continual Pre-training
14 mins; March 05, 2025
[QA] Chain of Draft: Thinking Faster by Writing Less
7 mins; March 03, 2025
Chain of Draft: Thinking Faster by Writing Less
10 mins; March 03, 2025
[QA] Sim-to-Real Reinforcement Learning for Vision-Based Dexterous Manipulation on Humanoids
7 mins; March 03, 2025
Sim-to-Real Reinforcement Learning for Vision-Based Dexterous Manipulation on Humanoids
24 mins; March 03, 2025
[QA] Implicit Search via Discrete Diffusion: A Study on Chess
8 mins; March 01, 2025
Implicit Search via Discrete Diffusion: A Study on Chess
23 mins; March 01, 2025
[QA] Avat3r: Large Animatable Gaussian Reconstruction Model for High-fidelity 3D Head Avatars
7 mins; March 01, 2025
Avat3r: Large Animatable Gaussian Reconstruction Model for High-fidelity 3D Head Avatars
21 mins; March 01, 2025
[QA] LightThinker: Thinking Step-by-Step Compression
7 mins; February 28, 2025
LightThinker: Thinking Step-by-Step Compression
16 mins; February 28, 2025
[QA] LLM-Microscope: Uncovering the Hidden Role of Punctuation in Context Memory of Transformers
8 mins; February 28, 2025
LLM-Microscope: Uncovering the Hidden Role of Punctuation in Context Memory of Transformers
10 mins; February 28, 2025
[QA] Self-rewarding correction for mathematical reasoning
40 mins; February 27, 2025
Self-rewarding correction for mathematical reasoning
40 mins; February 27, 2025
[QA] I Know What I Don't Know: Improving Model Cascades Through Confidence Tuning
8 mins; February 26, 2025
I Know What I Don't Know: Improving Model Cascades Through Confidence Tuning
19 mins; February 26, 2025
[QA] Can Language Models Falsify? Evaluating Algorithmic Reasoning with Counterexample Creation
7 mins; February 26, 2025
Can Language Models Falsify? Evaluating Algorithmic Reasoning with Counterexample Creation
15 mins; February 26, 2025
[QA] DeepSeek vs. ChatGPT: A Comparative Study for Scientific Computing and Scientific Machine Learning Tasks
8 mins; February 25, 2025
DeepSeek vs. ChatGPT: A Comparative Study for Scientific Computing and Scientific Machine Learning Tasks
15 mins; February 25, 2025
[QA] Yes, Q-learning Helps Offline In-Context RL
8 mins; February 25, 2025
Yes, Q-learning Helps Offline In-Context RL
26 mins; February 25, 2025
[QA] Fractal Generative Models
7 mins; February 24, 2025
Fractal Generative Models
17 mins; February 24, 2025
[QA] Improving the Scaling Laws of Synthetic Data with Deliberate Practice
6 mins; February 23, 2025
Improving the Scaling Laws of Synthetic Data with Deliberate Practice
19 mins; February 23, 2025
[QA] Idiosyncrasies in Large Language Models
7 mins; February 22, 2025
Idiosyncrasies in Large Language Models
20 mins; February 22, 2025
[QA] SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
7 mins; February 21, 2025
SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
20 mins; February 21, 2025
[QA] Position: Graph Learning Will Lose Relevance Due To Poor Benchmarks
7 mins; February 20, 2025
Position: Graph Learning Will Lose Relevance Due To Poor Benchmarks
27 mins; February 20, 2025
[QA] RocketKV: Accelerating Long-Context LLM Inference via Two-Stage KV Cache Compression
7 mins; February 20, 2025
RocketKV: Accelerating Long-Context LLM Inference via Two-Stage KV Cache Compression
20 mins; February 20, 2025
[QA] Autellix: An Efficient Serving Engine for LLM Agents as General Programs
8 mins; February 19, 2025
Autellix: An Efficient Serving Engine for LLM Agents as General Programs
35 mins; February 19, 2025
[QA] Small Models Struggle to Learn from Strong Reasoners
6 mins; February 19, 2025
Small Models Struggle to Learn from Strong Reasoners
13 mins; February 19, 2025
[QA] TokenSkip: Controllable Chain-of-Thought Compression in LLMs
7 mins; February 18, 2025
TokenSkip: Controllable Chain-of-Thought Compression in LLMs
18 mins; February 18, 2025
[QA] Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention
7 mins; February 18, 2025
Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention
26 mins; February 18, 2025
[QA] (How) Can Transformers Predict Pseudo-Random Numbers?
7 mins; February 16, 2025
(How) Can Transformers Predict Pseudo-Random Numbers?
21 mins; February 16, 2025
[QA] Do Large Language Models Reason Causally Like Us? Even Better?
8 mins; February 16, 2025
Do Large Language Models Reason Causally Like Us? Even Better?
7 mins; February 16, 2025
[QA] Eidetic Learning: an Efficient and Provable Solution to Catastrophic Forgetting
7 mins; February 15, 2025
Eidetic Learning: an Efficient and Provable Solution to Catastrophic Forgetting
15 mins; February 15, 2025
[QA] Fino1: On the Transferability of Reasoning‑Enhanced LLMs to Finance
8 mins; February 15, 2025
Fino1: On the Transferability of Reasoning‑Enhanced LLMs to Finance
27 mins; February 15, 2025
[QA] The Geometry of Prompting: Unveiling Distinct Mechanisms of Task Adaptation in Language Models
8 mins; February 14, 2025
The Geometry of Prompting: Unveiling Distinct Mechanisms of Task Adaptation in Language Models
22 mins; February 14, 2025
[QA] SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models
7 mins; February 13, 2025
SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models
19 mins; February 13, 2025
[QA] LLM Pretraining with Continuous Concepts
7 mins; February 12, 2025
LLM Pretraining with Continuous Concepts
17 mins; February 12, 2025
[QA] Distillation Scaling Laws
7 mins; February 12, 2025
Distillation Scaling Laws
17 mins; February 12, 2025
[QA] Competitive Programming with Large Reasoning Models
7 mins; February 11, 2025
Competitive Programming with Large Reasoning Models
20 mins; February 11, 2025
[QA] Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling
7 mins; February 11, 2025
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling
21 mins; February 11, 2025
[QA] DeepCrossAttention: Supercharging Transformer Residual Connections
7 mins; February 11, 2025
DeepCrossAttention: Supercharging Transformer Residual Connections
21 mins; February 11, 2025
[QA] Matryoshka Quantization
7 mins; February 11, 2025
Matryoshka Quantization
23 mins; February 11, 2025
[QA] When One LLM Drools, Multi-LLM Collaboration Rules
8 mins; February 10, 2025
When One LLM Drools, Multi-LLM Collaboration Rules
17 mins; February 10, 2025
[QA] Self-Regulation and Requesting Interventions
7 mins; February 10, 2025
Self-Regulation and Requesting Interventions
23 mins; February 10, 2025
[QA] Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach
7 mins; February 10, 2025
Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach
31 mins; February 10, 2025
[QA] Value-Based Deep RL Scales Predictably
8 mins; February 08, 2025
Value-Based Deep RL Scales Predictably
17 mins; February 08, 2025
[QA] Demystifying Long Chain-of-Thought Reasoning in LLMs
7 mins; February 08, 2025
Demystifying Long Chain-of-Thought Reasoning in LLMs
34 mins; February 08, 2025
[QA] ULTRAIF: Advancing Instruction Following from the Wild
7 mins; February 07, 2025
ULTRAIF: Advancing Instruction Following from the Wild
19 mins; February 07, 2025
[QA] Analyze Feature Flow to Enhance Interpretation and Steering in Language Models
7 mins; February 07, 2025
Analyze Feature Flow to Enhance Interpretation and Steering in Language Models
19 mins; February 07, 2025
[QA] Examining Two Hop Reasoning Through Information Content Scaling
8 mins; February 06, 2025
Examining Two Hop Reasoning Through Information Content Scaling
16 mins; February 06, 2025
[QA] Gold-medalist Performance in Solving Olympiad Geometry with AlphaGeometry2
8 mins; February 06, 2025
Gold-medalist Performance in Solving Olympiad Geometry with AlphaGeometry2
23 mins; February 06, 2025