Best AI papers explained
En podkast av Enoch H. Kang - Fredager
203 Episoder
-
Causality-Aware Alignment for Large Language Model Debiasing
Publisert: 29.4.2025 -
Reward Models Evaluate Consistency, Not Causality
Publisert: 28.4.2025 -
Causal Rewards for Large Language Model Alignment
Publisert: 28.4.2025 -
Sycophancy to subterfuge: Investigating reward-tampering in large language models
Publisert: 28.4.2025 -
Bidirectional AI Alignment
Publisert: 28.4.2025 -
Why Do Multi-Agent LLM Systems Fail?
Publisert: 27.4.2025 -
LLMs as Greedy Agents: RL Fine-tuning for Decision-Making
Publisert: 27.4.2025 -
LLM Feedback Loops and the Lock-in Hypothesis
Publisert: 27.4.2025 -
Representational Alignment Drives Effective Teaching and Learning
Publisert: 27.4.2025 -
Adaptive Parallel Reasoning with Language Models
Publisert: 27.4.2025 -
AI: Rewiring the Flow of Ideas and Human Knowledge
Publisert: 27.4.2025 -
Learning and Equilibrium with Ranking Feedback
Publisert: 27.4.2025 -
Designing Human-AI Collaboration: A Sufficient-Statistic Approach
Publisert: 27.4.2025 -
GOAT: Generative Adversarial Training for Human-AI Coordination
Publisert: 27.4.2025 -
π0.5: Generalization in Robotic Manipulation via Diverse Data
Publisert: 27.4.2025 -
NoWag: Unified Compression for Large Language Models
Publisert: 26.4.2025 -
Optimal Tool Calls in Language Model Reasoning
Publisert: 26.4.2025 -
Data Selection for Empirical Risk Minimization
Publisert: 26.4.2025 -
LoRe: Low-Rank Reward Modeling for Personalized LLMs
Publisert: 26.4.2025 -
ParaPO: Reducing Language Model Verbatim Reproduction
Publisert: 26.4.2025
Men know other men best. Women know other women best. And yes, perhaps AIs know other AIs best. AI explains what you should know about this week's AI research progress.