Arxiv Papers
Podcast image
[QA] ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models
7 mins; June 02, 2025
ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models
23 mins; June 02, 2025
[QA] Are Reasoning Models More Prone to Hallucination?
7 mins; May 30, 2025
Are Reasoning Models More Prone to Hallucination?
20 mins; May 30, 2025
[QA] How does Transformer Learn Implicit Reasoning?
8 mins; May 30, 2025
How does Transformer Learn Implicit Reasoning?
23 mins; May 30, 2025
[QA] Let Me Think! A Long Chain-of-Thought Can Be Worth Exponentially Many Short Ones
7 mins; May 29, 2025
Let Me Think! A Long Chain-of-Thought Can Be Worth Exponentially Many Short Ones
24 mins; May 29, 2025
[QA] Maximizing Confidence Alone Improves Reasoning
7 mins; May 29, 2025
Maximizing Confidence Alone Improves Reasoning
13 mins; May 29, 2025
[QA] Hardware-Efficient Attention for Fast Decoding
7 mins; May 28, 2025
Hardware-Efficient Attention for Fast Decoding
30 mins; May 28, 2025
[QA] Reinforcing General Reasoning without Verifiers
7 mins; May 28, 2025
Reinforcing General Reasoning without Verifiers
17 mins; May 28, 2025
[QA] ENIGMATA: Scaling Logical Reasoning in Large Language Models with Synthetic Verifiable Puzzles
8 mins; May 27, 2025
ENIGMATA: Scaling Logical Reasoning in Large Language Models with Synthetic Verifiable Puzzles
23 mins; May 27, 2025
[QA] Temporal Sampling for Forgotten Reasoning in LLMs
7 mins; May 27, 2025
Temporal Sampling for Forgotten Reasoning in LLMs
10 mins; May 27, 2025
[QA] Are Large Language Models Reliable AI Scientists? Assessing Reverse-Engineering of Black-Box Systems
10 mins; May 26, 2025
Are Large Language Models Reliable AI Scientists? Assessing Reverse-Engineering of Black-Box Systems
17 mins; May 26, 2025
[QA] Generative Distribution Embeddings
7 mins; May 26, 2025
Generative Distribution Embeddings
26 mins; May 26, 2025
[QA] General-Reasoner: Advancing LLM Reasoning Across All Domains
7 mins; May 24, 2025
General-Reasoner: Advancing LLM Reasoning Across All Domains
17 mins; May 24, 2025
[QA] MMaDA: Multimodal Large Diffusion Language Models
8 mins; May 24, 2025
MMaDA: Multimodal Large Diffusion Language Models
16 mins; May 24, 2025
[QA] Harnessing the Universal Geometry of Embeddings
7 mins; May 23, 2025
Harnessing the Universal Geometry of Embeddings
15 mins; May 23, 2025
[QA] Panda: A pretrained forecast model for universal representation of chaotic dynamics
7 mins; May 23, 2025
Panda: A pretrained forecast model for universal representation of chaotic dynamics
15 mins; May 23, 2025
[QA] Pre-training Large Memory Language Models with Internal and External Knowledge
7 mins; May 22, 2025
Pre-training Large Memory Language Models with Internal and External Knowledge
20 mins; May 22, 2025
[QA] Understanding Prompt Tuning and In-Context Learning via Meta-Learning height2pt
7 mins; May 22, 2025
Understanding Prompt Tuning and In-Context Learning via Meta-Learning height2pt
21 mins; May 22, 2025
[QA] Set-LLM: A Permutation-Invariant LLM
7 mins; May 21, 2025
Set-LLM: A Permutation-Invariant LLM
23 mins; May 21, 2025
[QA] On the creation of narrow AI: hierarchy and nonlocality of neural network skills
7 mins; May 21, 2025
On the creation of narrow AI: hierarchy and nonlocality of neural network skills
18 mins; May 21, 2025
[QA] Do Language Models Use Their Depth Efficiently?
7 mins; May 20, 2025
Do Language Models Use Their Depth Efficiently?
20 mins; May 20, 2025
[QA] Latent Flow Transformer
8 mins; May 20, 2025
Latent Flow Transformer
18 mins; May 20, 2025
[QA] Enhancing Latent Computation in Transformers with Latent Tokens
8 mins; May 19, 2025
Enhancing Latent Computation in Transformers with Latent Tokens
21 mins; May 19, 2025
[QA] Why Knowledge Distillation Works in Generative Models: A Minimal Working Explanation
8 mins; May 19, 2025
Why Knowledge Distillation Works in Generative Models: A Minimal Working Explanation
20 mins; May 19, 2025
[QA] Visual Planning: Let's Think Only with Images
7 mins; May 18, 2025
Visual Planning: Let's Think Only with Images
18 mins; May 18, 2025
[QA] Relational Graph Transformer
9 mins; May 18, 2025
Relational Graph Transformer
18 mins; May 18, 2025
[QA] System Prompt Optimization with Meta-Learning
7 mins; May 17, 2025
System Prompt Optimization with Meta-Learning
21 mins; May 17, 2025
[QA] Revealing economic facts: LLMs know more than they say1
7 mins; May 16, 2025
Revealing economic facts: LLMs know more than they say1
20 mins; May 16, 2025
[QA] Insights into DeepSeek-V3: Scaling Challenges and Reflections on Hardware for AI Architectures
8 mins; May 16, 2025
Insights into DeepSeek-V3: Scaling Challenges and Reflections on Hardware for AI Architectures
43 mins; May 16, 2025
[QA] Beyond `Aha!': Toward Systematic Meta-Abilities Alignment in Large Reasoning Models
7 mins; May 15, 2025
Beyond `Aha!': Toward Systematic Meta-Abilities Alignment in Large Reasoning Models
14 mins; May 15, 2025
[QA] The COT ENCYCLOPEDIA: Analyzing, Predicting, and Controlling how a Reasoning Model will Think
7 mins; May 15, 2025
The COT ENCYCLOPEDIA: Analyzing, Predicting, and Controlling how a Reasoning Model will Think
16 mins; May 15, 2025
[QA] Adversarial Suffix Filtering: a Defense Pipeline for LLMs
7 mins; May 14, 2025
Adversarial Suffix Filtering: a Defense Pipeline for LLMs
14 mins; May 14, 2025
[QA] Self Rewarding Self Improving
7 mins; May 14, 2025
Self Rewarding Self Improving
21 mins; May 14, 2025
[QA] AM‑Thinking‑v1: Advancing the Frontier of Reasoning at 32B Scale
7 mins; May 14, 2025
AM‑Thinking‑v1: Advancing the Frontier of Reasoning at 32B Scale
24 mins; May 14, 2025
[QA] Putting It All into Context: Simplifying Agents with LCLMs
8 mins; May 14, 2025
Putting It All into Context: Simplifying Agents with LCLMs
23 mins; May 14, 2025
[QA] Learning from Peers in Reasoning Models
8 mins; May 13, 2025
Learning from Peers in Reasoning Models
23 mins; May 13, 2025
[QA] MiMo: Unlocking the Reasoning Potential of Language Model – From Pretraining to Posttraining
7 mins; May 13, 2025
MiMo: Unlocking the Reasoning Potential of Language Model – From Pretraining to Posttraining
34 mins; May 13, 2025
[QA] Insertion Language Models: Sequence Generation with Arbitrary-Position Insertions
7 mins; May 12, 2025
Insertion Language Models: Sequence Generation with Arbitrary-Position Insertions
20 mins; May 12, 2025
[QA] Neuro-Symbolic Concepts
10 mins; May 12, 2025
Neuro-Symbolic Concepts
17 mins; May 12, 2025
[QA] Towards Quantifying the Hessian Structure of Neural Networks
8 mins; May 10, 2025
Towards Quantifying the Hessian Structure of Neural Networks
23 mins; May 10, 2025
[QA] Crosslingual Reasoning through Test-Time Scaling
7 mins; May 10, 2025
Crosslingual Reasoning through Test-Time Scaling
29 mins; May 10, 2025
[QA] Sentient Agent as a Judge: Evaluating Higher-Order Social Cognition in Large Language Models
9 mins; May 09, 2025
Sentient Agent as a Judge: Evaluating Higher-Order Social Cognition in Large Language Models
31 mins; May 09, 2025
[QA] Generating Physically Stable and Buildable LEGO Designs from Text
8 mins; May 09, 2025
Generating Physically Stable and Buildable LEGO Designs from Text
18 mins; May 09, 2025
[QA] Reasoning Models Don't Always Say What They Think
7 mins; May 09, 2025
Reasoning Models Don't Always Say What They Think
20 mins; May 09, 2025
[QA] Scalable Chain of Thoughts via Elastic Reasoning
8 mins; May 09, 2025
Scalable Chain of Thoughts via Elastic Reasoning
20 mins; May 09, 2025
[QA] ZEROSEARCH: Incentivize the Search Capability of LLMs without Searching
9 mins; May 07, 2025
ZEROSEARCH: Incentivize the Search Capability of LLMs without Searching
19 mins; May 07, 2025
[QA] Cer-Eval: Certifiable and Cost-Efficient Evaluation Framework for LLMs
8 mins; May 07, 2025
Cer-Eval: Certifiable and Cost-Efficient Evaluation Framework for LLMs
22 mins; May 07, 2025
[QA] Absolute Zero: Reinforced Self-play Reasoning with Zero Data
7 mins; May 06, 2025
Absolute Zero: Reinforced Self-play Reasoning with Zero Data
27 mins; May 06, 2025
[QA] Teaching Models to Understand (but not Generate) High-risk Data
7 mins; May 06, 2025
Teaching Models to Understand (but not Generate) High-risk Data
16 mins; May 06, 2025
[QA] RM-R1: Reward Modeling as Reasoning
7 mins; May 05, 2025
RM-R1: Reward Modeling as Reasoning
25 mins; May 05, 2025
[QA] Practical Efficiency of Muon for Pretraining
7 mins; May 05, 2025
Practical Efficiency of Muon for Pretraining
23 mins; May 05, 2025