Clem Bonnet
@ClementBonnet16
AI Research @ndea
I had a fantastic time discussing latent search with @ecsquendor on MLST! 🙏 We review our work on latent program networks co-authored with @MattVMacfarlane and debate composition, AI creativity, and program synthesis. Full video: youtu.be/PHBItVuudbU
We spoke with @ClementBonnet16 at NeurIPS about his extremely innovative approach to the @arcprize using a form of test time inference where you search a latent space of a VAE before making an optimal prediction. @fchollet was so impressed, he hired Clem shortly after! 😃 -…
Today, we're announcing a preview of ARC-AGI-3, the Interactive Reasoning Benchmark with the widest gap between easy for humans and hard for AI We’re releasing: * 3 games (environments) * $10K agent contest * AI agents API Starting scores - Frontier AI: 0%, Humans: 100%
Our #ICML2025 Programmatic Representations for Agent Learning workshop will take place tomorrow, July 18th, at the West Meeting Room 301-305, exploring how programmatic representations can make agent learning more interpretable, generalizable, efficient, and safe! Come join us!
Encoder-Decoder for Latent Program Search : how there was (and still is) a possibility to have specialized approaches for Arc problem solving 📖 Read 203: « Searching Latent Program Spaces », by @ClementBonnet16 and @MattVMacfarlane arxiv.org/pdf/2411.08706 A very original…
I’ll be presenting two workshop papers: “Searching Latent Program Spaces” (oral at Programmatic Representations for Agent Learning) & “Instilling Parallel Reasoning into Language Models” (AI4Math).
Previous work has shown that programmatic policies—computer programs written in a domain-specific language—generalize to out-of-distribution problems more easily than neural policies. Is this really the case? 🧵
Ndea is sponsoring SYNT 2025 - a workshop on synthesis of computing systems - July 22 in Zagreb, Croatia. Part of CAV 2025 (Conference on Computer Aided Verification). Have a synthesis-related paper abstract? Submissions due by May 18. synt2025.github.io
I will be at ICLR next week. Always up for chatting about ARC-AGI, program synthesis, RL, open-endedness, or related rabbit holes. DMs open! #ICLR2025
Compressing the timeline to get to AGI also means compressing the timeline of every single scientific breakthrough that is downstream of AGI. There is no greater leverage
I've just finished drafting a fairly thorough review of a lot of the research that went into @arcprize 2024, all the paper winners, top scorers, and a few hot takes for ARC-AGI-2 in 2025. Here's a short 🧵 with some highlights!
Quick hiring update: we've assembled an incredible founding research team. We have no open positions, for now. However — we will create roles for exceptional program synthesis researchers. If that's you: ndea.com/join Onward.
ARC Prize 2025 is Live $1M competition to open source a solution to ARC-AGI Your objective: Reach 85% on the private evaluation dataset Progress needs new ideas, not just scale
Today we are announcing ARC-AGI-2, an unsaturated frontier AGI benchmark that challenges AI reasoning systems (same relative ease for humans). Grand Prize: 85%, ~$0.42/task efficiency Current Performance: * Base LLMs: 0% * Reasoning Systems: <4%
Honored to be a guest on the infamous MLST podcast again! We discuss our test-time methods, compositionality in LLMs, limitations of VLMs, logic vs perception, efficient adaptation, and more. @MLStreetTalk youtu.be/3p0O28W1ZHg
So proud of this one: today we're releasing Command A, our new 111B flagship model tailored for business use cases. Gpt4o/Deepseek-level in evals, better than Sonnet on LMsys, has 256k context length. Also, weights are available now on hf! huggingface.co/CohereForAI/c4…
Deep Learning architectures usually aren't trained to perform search at test time, leading to sample inefficiency + poor generalization. Latent Program Network (LPN) builds in test-time adaption by learning a latent space that can be searched. @ClementBonnet16 @MattVMacfarlane
AGI is reached when the capability gap between humans and computers is zero ARC Prize Foundation measures this to inspire progress Today we preview the unbeaten ARC-AGI-2 + open public donations to fund ARC-AGI-3 TY Schmidt Sciences (@ericschmidt) for $50k to kick us off!
Checkout @ClementBonnet16 discussing searching latent program spaces tommorow on @MLStreetTalk !
We spoke with @ClementBonnet16 at NeurIPS about his extremely innovative approach to the @arcprize using a form of test time inference where you search a latent space of a VAE before making an optimal prediction. @fchollet was so impressed, he hired Clem shortly after! 😃 -…
We spoke with @ClementBonnet16 at NeurIPS about his extremely innovative approach to the @arcprize using a form of test time inference where you search a latent space of a VAE before making an optimal prediction. @fchollet was so impressed, he hired Clem shortly after! 😃 -…