Max Kleiman-Weiner
@maxhkw
professor @UW computational cognitive scientist working on social and artificial intelligence. cofounder @CSM_ai. priors: PhD @MIT founder @diffeo (acquired)
Our new paper is out in PNAS: "Evolving general cooperation with a Bayesian theory of mind"! Humans are the ultimate cooperators. We coordinate on a scale and scope no other species (nor AI) can match. What makes this possible? 🧵

A hallmark of human intelligence is the capacity for rapid adaptation, solving new problems quickly under novel and unfamiliar conditions. How can we build machines to do so? In our new preprint, we propose that any general intelligence system must have an adaptive world model,…
Check our poster 11am Thurs July 17th! #ICML2025 I'd be very keen to chat about truth-seeking AI, gradual disempowerment, the lock-in prolem, and AI influence. DM open or drop me an email.
Humans learn from LLM, LLMs learn from AI. Would this feedback loop cause infinite lock-in of current dominant beliefs? However wrong or harmful some of the current beliefs might be, harder for them to be updated away. Introducing "The Lock-in Hypothesis" (accepted by ICML) 🧵
Check out @kjha02's ORAL at ICML today!
Really pumped for my Oral presentation on this work today!!! Come check out the RL session from 3:30-4:30pm in West Ballroom B You can also swing by our poster from 4:30-7pm in West Exhibition Hall B2-B3 # W-713 See you all there!
🚀 Big news! Pokee AI just raised a $12M seed round—3x oversubscribed! Our mission: Automate every human workflow on the internet with frictionless AI agents. Public beta now live 👉 pokee.ai Here's what we're building and who's backing us 🧵👇
Settling in for my flight and apparently A.I. DOOM is now a movie genre between Harry Potter and Classics. Nothing better than an existential crisis with pretzels and a ginger ale.

CSM on Google Cloud is a game-changer (pun intended). Turn any prompt (text, image, sketch) into game-engine-ready 3D assets in minutes. Rapidly iterate on ideas and streamline 3D asset production with @CSM_ai tools to bring your games to life quicker, better, and for less ↓
Check out our 3 papers on Testing LLM Moral Reasoning via Multi-Agent Simulations! ✍️ Our summary blogpost: lesswrong.com/posts/2WAire3L… 📑Our series of 3 papers:1️⃣GovSim (NeurIPS 2024) arxiv.org/abs/2404.16698 2️⃣SanctSim zhijing-jin.com/files/papers/2… 3️⃣MoralSim arxiv.org/abs/2505.19212
To successfully act as our delegates, agents must be additionally aligned to particular interests, individuals, and organizations, and cooperate towards mutual benefit, not solely out of altruism. One of the frontiers for Cooperative AI and alignment that I'm excited about.
Claude did well in some ways: it searched the web to find new suppliers, and ordered very niche drinks that Anthropic staff requested. But it also made mistakes. Claude was too nice to run a shop effectively: it allowed itself to be browbeaten into giving big discounts.
Humans learn from LLM, LLMs learn from AI. Would this feedback loop cause infinite lock-in of current dominant beliefs? However wrong or harmful some of the current beliefs might be, harder for them to be updated away. Introducing "The Lock-in Hypothesis" (accepted by ICML) 🧵
To help computational cognitive scientist engage with more naturalistic experiments, I've made NiceWebRL. NiceWebRL is a Python library for designing human subject experiments that leverage machine reinforcement learning environments. github.com/KempnerInstitu…
New paper: World models + Program synthesis by @topwasu 1. World modeling on-the-fly by synthesizing programs w/ 4000+ lines of code 2. Learns new environments from minutes of experience 3. Positive score on Montezuma's Revenge 4. Compositional generalization to new environments…