Cyril Zakka, MD
@cyrilzakka
Health Lead x @Huggingface | Medical Doctor & ML Researcher. Prev: @Stanford
There’s been a lot of buzz surrounding multimodal LLMs in medicine. At the @HiesingerLab we’ve been building Almanac Chat for over 6 months alongside our collaborators and we’re excited to showcase the first preview! • Almanac Chat is the first truly multimodal medical LLM…
We are excited to work with @StanfordDeptMed, @SmidtHeart, @mghliver, and @KPDOR with the support of @American_Heart to refine and prospectively evaluate AI opportunistic screening approaches for Cardiovascular-Kidney-Metabolic Health! Chronic Liver and Kidney Disease is common…
A lot to unpack here but having open-source and open-weight AI as core tenets of the American AI Action Plan is going to have compounding benefits for the quality and affordability of healthcare in the US. We’re still in the early days but a shift is coming. 🇺🇸🤗
🇺🇸 Today is a day we have been working towards for six months. We are announcing America’s AI action plan putting us on the road to continued AI dominance. The three core themes: - Accelerate AI innovation - Build American AI infrastructure - Lead in international AI…
There’s recently been an influx of case reports describing people exhibiting signs of psychosis having their episodes and beliefs amplified by an LLM (e.g. ChatGPT, CharacterAI, etc) While I’m not a psychiatrist by training, I think it mirrors an interesting syndrome known as…
You know CASP? The competition that AlphaFold won that changed the game for AI x bio? 🧬 Just dropped all the data from their last challenge on @huggingface! check it out ⤵️ huggingface.co/datasets/cgeor…
and it's out!
🧑🍳 We're cooking up something. You'll soon be able to go from any API with an OpenAPI spec to a Gradio UI. Coupled with @Gradio's MCP integration, this makes it especially easy to go from API -> MCP.
So many great releases today, with a new medical model to boot! 27B multimodal MedGemma featuring my all-time favorite encoder: huggingface.co/papers/2507.05…
🏥Introducing MedGemma, part 2, including: 🔥A 27B multimodal MedGemma 👀MedSigLIP, a lightweight image/text encoder for medical image retrieval/classification 📜A technical report with details Blog: research.google/blog/medgemma-… Paper: arxiv.org/abs/2507.05201
We updated the State Embedding 600M checkpoint on the @ArcInstitute Hugging Face This model was trained with 4x FLOPs compared to the preprint model. It achieves significantly lower val/loss and does better on internal evals - would recommend using this over the 4 epoch one for…
Want to do a postdoc in my research group? bsse.ethz.ch/mail I can support one ETH Postdoc fellowship application this fall. For details, see: grantsoffice.ethz.ch/funding-opport… If interested, please carefully check your eligibility (esp PhD defense date) and then send us the…
We just released the best 3B model, 100% open-source, open dataset, architecture details, exact data mixtures and full training recipe including pre-training, mid-training, post-training, and synthetic data generation for everyone to train their own. Let's go open-source AI!
Introducing SmolLM3: a strong, smol reasoner! > SoTA 3B model > dual mode reasoning (think/no_think) > long context, up to 128k > multilingual: en, fr, es, de, it, pt > fully open source (data, code, recipes) huggingface.co/blog/smollm3
Introducing SmolLM3: a strong, smol reasoner! > SoTA 3B model > dual mode reasoning (think/no_think) > long context, up to 128k > multilingual: en, fr, es, de, it, pt > fully open source (data, code, recipes) huggingface.co/blog/smollm3
Sad news: HuggingChat is gone 😢 It was a wild adventure... 🔹20+ Open Source models 🔹1M+ users 🔹100k+ assistants created But with where AI is going lately, it was starting to feel out of date. It’s time to build something new. Stay tuned…
Why would you create an MCP server?? When you could create a UI, API, and MCP server all with the same Python code?
🚨New preprint! 🚨In-context learning (ICL) is the intriguing ability of LLMs to learn to solve tasks purely from context w/o parameter updates. For multimodal LLMs (MLLMs), ICL is poorly understood, especially in the medical domain where doctors would often face few relevant…
I f*** love this. And to put even more momentum behind this awesome movement by our @arcinstitute friends: 🥁@tahoe_ai will give $25K to the best model that is also open-sourced with weights on @huggingface for everyone's use! + maybe a mention on X by @ClementDelangue? ;)
Register today for the Virtual Cell Challenge and use AI to solve one of biology’s most complex problems. Announced in @CellCellPress, the competition is hosted by Arc Institute and sponsored by @nvidia, @10xGenomics, and @UltimaGenomics.
Great new model from the @arcinstitute. Model links here: huggingface.co/arcinstitute/S…, huggingface.co/arcinstitute/S…, huggingface.co/arcinstitute/S…
Introducing Arc Institute’s first virtual cell model: STATE
Excellent commentary, @cyrilzakka. Fascinating implications for teaching residents. @TradIMYale
There's a lot to be said on this newest preprint showing that reliance on LLMs in writing tasks leads to a 'consistent underperformance' in neural and behavioral levels with potential long-term consequences. Interestingly, I think this ties to similar but (early)…
Very good point. But the cognitive impact of technology is a recurrent question over at least the last 75 years (more recently for Search—see article below).
There's a lot to be said on this newest preprint showing that reliance on LLMs in writing tasks leads to a 'consistent underperformance' in neural and behavioral levels with potential long-term consequences. Interestingly, I think this ties to similar but (early)…
Unsurprisingly @cyrilzakka says it better than I could. These are the exact concerns I have about the way we are integrating LLMs into clinical care. It *might* not matter, but we are all participating in an educational Pascal's wager right now. A 🧵🔽
There's a lot to be said on this newest preprint showing that reliance on LLMs in writing tasks leads to a 'consistent underperformance' in neural and behavioral levels with potential long-term consequences. Interestingly, I think this ties to similar but (early)…