Best AI papers explained
A podcast by Enoch H. Kang
456 Episodes
-
ShiQ: Bringing back Bellman to LLMs
Published: 5/22/2025 -
Policy Learning with a Natural Language Action Space: A Causal Approach
Published: 5/22/2025 -
Multi-Objective Preference Optimization: Improving Human Alignment of Generative Models
Published: 5/22/2025 -
End-to-End Learning for Stochastic Optimization: A Bayesian Perspective
Published: 5/21/2025 -
TEXTGRAD: Automatic Differentiation via Text
Published: 5/21/2025 -
Steering off Course: Reliability Challenges in Steering Language Models
Published: 5/20/2025 -
Past-Token Prediction for Long-Context Robot Policies
Published: 5/20/2025 -
Recovering Coherent Event Probabilities from LLM Embeddings
Published: 5/20/2025 -
Systematic Meta-Abilities Alignment in Large Reasoning Models
Published: 5/20/2025 -
Predictability Shapes Adaptation: An Evolutionary Perspective on Modes of Learning in Transformers
Published: 5/20/2025 -
Efficient Exploration for LLMs
Published: 5/19/2025 -
Rankers, Judges, and Assistants: Towards Understanding the Interplay of LLMs in Information Retrieval Evaluation
Published: 5/18/2025 -
Bayesian Concept Bottlenecks with LLM Priors
Published: 5/17/2025 -
Transformers for In-Context Reinforcement Learning
Published: 5/17/2025 -
Evaluating Large Language Models Across the Lifecycle
Published: 5/17/2025 -
Active Ranking from Human Feedback with DopeWolfe
Published: 5/16/2025 -
Optimal Designs for Preference Elicitation
Published: 5/16/2025 -
Dual Active Learning for Reinforcement Learning from Human Feedback
Published: 5/16/2025 -
Active Learning for Direct Preference Optimization
Published: 5/16/2025 -
Active Preference Optimization for RLHF
Published: 5/16/2025
Cut through the noise. We curate and break down the most important AI papers so you don’t have to.