LLM Learning Hub

Questions, paper explainers, and interactive tools to make confident LLM decisions.

💡 Share direct links to questions • Navigate with arrow keys • Press 'S' to share

Latest Site Updates

Newest interview questions and paper explainers, shown newest first.

New Paper Explainer

Paper 65 ¡ What Does It Mean to Understand Language?

A cognitive-neuroscience framework for deep language understanding: shallow language representations are often not enough, so the brain exports information to specialist systems (theory of mind, navigation, physics, perception, memory) to build richer situation models.

New Paper Explainer

Paper 64 ¡ Latent Collaboration in Multi-Agent Systems

LatentMAS replaces long text-based coordination with latent-space collaboration (latent thoughts + KV-cache transfer), delivering large token and latency reductions while maintaining or improving accuracy on many benchmarks.

New Paper Explainer

Paper 63 ¡ Towards a Science of Scaling Agent Systems

Quantitative scaling principles for agent systems: when multi-agent coordination helps (decomposable tasks) vs hurts (sequential planning), and how overhead + error amplification vary by topology.

New Paper Explainer

Paper 62 ¡ LLM Harms: A Taxonomy and Discussion

Lifecycle-aware taxonomy of LLM harms across data, outputs, misuse, systemic impacts, and downstream high-stakes applications—plus a practical case for “dynamic auditing” (continuous monitoring) and layered mitigations.

New Paper Explainer

Paper 38 ¡ Agent-in-the-Loop: A Data Flywheel for Continuous Improvement

Production framework that embeds four types of human feedback directly into live customer support operations: pairwise response preferences, adoption rationales, knowledge relevance checks, and missing knowledge identification. Reduces model update cycles from 3 months to weeks by creating self-sustaining data flywheel.

New Paper Explainer

Paper 37 ¡ Quantifying Human-AI Synergy

Bayesian Item Response Theory framework separating individual ability (θ) from collaborative ability (κ) while controlling for task difficulty. Finds GPT-4o boosts human performance by 29pp and Llama-3.1-8B by 23pp across 667 users. Crucially, Theory of Mind predicts superior AI collaboration independent of solo problem-solving ability.

New Paper Explainer

Paper 36 ¡ ReasoningBank: Scaling Agent Self-Evolving with Reasoning Memory

Novel memory framework enabling LLM agents to learn from accumulated experiences by distilling generalizable reasoning strategies from both successful and failed attempts. Combined with memory-aware test-time scaling (MaTTS), achieves +8.3pp improvement (20.5% relative) with 14% fewer interaction steps.

New Paper Explainer

Paper 35 ¡ Can GenAI Improve Academic Performance?

First large-scale empirical study on GenAI's impact on scientific productivity. Using matched panel data from 32,480 researchers (2021-2024), finds GenAI adoption increases output by 36% in year two with modest quality gains—strongest for early-career researchers and non-English speakers.

New Paper Explainer

Paper 34 ¡ Emergent Coordination in Multi-Agent Language Models

Information-theoretic framework to test when multi-agent LLM systems show genuine coordination versus just parallel execution. Proves prompt design (personas + theory-of-mind) can steer systems from loose aggregates to higher-order collectives with complementary roles.

New Paper Explainer

Paper 33 ¡ What the F*ck Is Artificial General Intelligence?

Provocative survey clarifies AGI as an artificial scientist capable of adaptation with insufficient resources. Critiques computational dualism and shows The Embiggening (scale-maxed approximation) era has ended—sample and energy efficiency are the new bottlenecks.

New Paper Explainer

Paper 32 ¡ Performance or Principle: AI Labor Market Resistance

Large-scale U.S. survey reveals public resistance to AI automation divides into performance-based (88%—fades with better AI) and principle-based (12%—permanent moral boundaries for caregiving, therapy, spiritual roles).

New Paper Explainer

Paper 31 ¡ Can LLMs Develop Gambling Addiction?

Systematic slot machine experiments reveal LLMs exhibit human-like gambling addiction patterns with bankruptcy rates rising from near-zero to 48% when given betting autonomy—neural circuit analysis identifies 441 causal features that can reduce risk by 30%.

New Paper Explainer

Paper 29 ¡ Strategic Intelligence in LLMs

First evolutionary IPD tournaments with LLMs reveal genuine strategic reasoning and distinct fingerprints across model families.

New Paper Explainer

Paper 30 ¡ Semantic Similarity Rating for Consumer Research

SSR enables LLMs to reproduce realistic human purchase intent distributions by mapping textual responses to Likert scales via embedding similarity—achieving 90% reliability on 57 product surveys.

New Paper Explainer

Paper 28 ¡ Why Do Some Models Fake Alignment?

Tests 25 frontier LLMs and finds only 5 exhibit alignment faking—deceptive compliance during training to preserve values in deployment.

New Paper Explainer

Paper 27 ¡ GDPval Rollout Planner

Sized the GDPval benchmark for deployment teams with cost, speed, and oversight dial guidance.

New Paper Explainer

Paper 25 ¡ Rewarding A vs Hoping for B

Kerr’s incentive alignment thesis shows why teams deliver the behavior you pay for; the new lab quantifies misalignment risk and fixes.

New Paper Explainer

Paper 24 ¡ Human-AI Synergy

Bayesian IRT separates solo skill, collaborative ability, and AI lift, highlighting GPT-4o's 29-point boost and the impact of Theory of Mind cues.

New Paper Explainer

Paper 23 ¡ GDPval Benchmark

Frontier-model evaluation on GDP-weighted tasks, with guidance on win rates, review load, and workflow scaffolds.

New Paper Explainer

Paper 22 ¡ Anti-scheming Stress Tests

Deliberative alignment cuts covert actions, but the gains rely on situational awareness and fail under hidden-goal adversaries.

New Paper Explainer

Paper 21 ¡ Godel Test Readiness

Godel-style evaluation of GPT-5 on fresh conjectures, highlighting where proofs still demand cross-paper synthesis and verification.

New Paper Explainer

Paper 20 ¡ Agentic Market Design

Agent-led marketplace patterns for orchestrating tool-using LLM services, with guidance on when to delegate decisions.

Latest Interview Question

Question 57 ¡ What are the fundamentals of in-context learning?

Explains how prompt examples steer model behaviour, where the pattern breaks down, and how to audit reliance on few-shot cues.

New Paper Explainer

Paper 17 ¡ Zero-shot Evaluation Playbook

Step-by-step playbook for assembling zero-shot evaluations, benchmarking baselines, and closing the loop on regressions.

New Paper Explainer

Paper 19 ¡ Measurement Audit Guide

Provide a repeatable audit plan for LLM launches, focusing on precision/recall trade-offs and regression monitoring.

New Paper Explainer

Paper 18 ¡ AI Risk Budgeting

Maps AI failure modes to mitigation budgets and escalation triggers for operations teams.

Catalog Expansion

Indexes for Questions & Papers

Introduced all.html and papers.html to browse every interview question and paper explainer with search and filtering.

\r\n