Best AI papers explained
A podcast by Enoch H. Kang
456 Episodes
-
Trace is the Next AutoDiff: Generative Optimization with Rich Feedback, Execution Traces, and LLMs
Published: 5/24/2025 -
Adaptive Inference-Time Compute: LLMs Can Predict if They Can Do Better, Even Mid-Generation
Published: 5/24/2025 -
Prompts from Reinforcement Learning (PRL)
Published: 5/24/2025 -
Logits are All We Need to Adapt Closed Models
Published: 5/24/2025 -
Large Language Models Are (Bayesian) Latent Variable Models: Explaining and Finding Good Demonstrations for In-Context Learning
Published: 5/23/2025 -
Inference-Time Intervention: Eliciting Truthful Answers from a Language Model
Published: 5/23/2025 -
From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models
Published: 5/23/2025 -
LLM In-Context Learning as Kernel Regression
Published: 5/23/2025 -
Personalizing LLMs via Decode-Time Human Preference Optimization
Published: 5/23/2025 -
Almost Surely Safe LLM Inference-Time Alignment
Published: 5/23/2025 -
Survey of In-Context Learning Interpretation and Analysis
Published: 5/23/2025 -
From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models
Published: 5/23/2025 -
LLM In-Context Learning as Kernel Regression
Published: 5/23/2025 -
Where does In-context Learning Happen in Large Language Models?
Published: 5/23/2025 -
Auto-Differentiating Any LLM Workflow: A Farewell to Manual Prompting
Published: 5/22/2025 -
metaTextGrad: Learning to learn with language models as optimizers
Published: 5/22/2025 -
Semantic Operators: A Declarative Model for Rich, AI-based Data Processing
Published: 5/22/2025 -
Isolated Causal Effects of Language
Published: 5/22/2025 -
Sleep-time Compute: Beyond Inference Scaling at Test-time
Published: 5/22/2025 -
J1: Incentivizing Thinking in LLM-as-a-Judge
Published: 5/22/2025
Cut through the noise. We curate and break down the most important AI papers so you don’t have to.