Daniel P Jeong
@danielpjeong
PhD student @mldcmu working on ML for healthcare | currently @MSFTResearch | prev: @columbia, @nasajpl 🦋 https://bsky.app/profile/danielpjeong.bsky.social
🧵 Are "medical" LLMs/VLMs *adapted* from general-domain models, always better at answering medical questions than the original models? In our oral presentation at #EMNLP2024 today (2:30pm in Tuttle), we'll show that surprisingly, the answer is "no". arxiv.org/abs/2411.04118
🇮🇹 We're presenting a tutorial on Retrieval-Enhanced Machine Learning (#REML) at #SIGIR2025! 🗓️Sunday July 13th | 9AM - 12:30PM 📌Donatello Floor 0 Details: retrieval-enhanced-ml.github.io/sigir-2025.html @841io @mrdrozdov @SalemiAlireza7 @HamedZamani @SIGIRConf
🧵Working with #MCP or building a modular #RAG system, but not sure which rankers to use from your pool? 📊 Rank the Rankers⚡Route smart. This paper shows how. 👨🔬 w/ Fernando Diaz @841io 💻 Code: github.com/kimdanny/Starl… Paper: arxiv.org/abs/2506.13743
Looking forward to giving a talk this Friday @OpenAI with @zhilifeng on some of our privacy & memorization research + how it applies to production LLMs! We've been gaining momentum on detecting, quantifying & erasing memorization; excited to explore its real-world impact!
Excited to be at #ICLR2025 🇸🇬 to talk about my recent works👇that uncover key pitfalls & inefficiencies in pretraining & inference🚨. Final PhD lap —thinking a lot about how pretraining interventions can shape downstream behaviors (like reasoning & safety). DM to chat or vibe!
1/Being in academia is such a privilege: You get to collaborate with insanely talented & passionate students on their journey to upskill themselves. Very excited to share *OpenUnlearning*: a unified, easily extensible framework for unlearning led by @anmol_mekala @VineethDorna🧵
We have just released SenSet, a novel list of 106 senescence marker genes. We hope this resource accelerates discoveries in aging research, cancer biology, and regenerative medicine. #senescence #aging #pulearning #gene-set #SenNet biorxiv.org/content/10.110…
🧩 Why do task vectors exist in pretrained LLMs? Our new research uncovers how transformers form internal abstractions and the mechanisms behind in-context learning(ICL).
How to **efficiently** build personalized language models without textual info on user preferences? Our P-RLHF work: - light-weight user model - personalize all *PO alignment algorithms - strong performance on the largest personalized preference dataset arxiv.org/abs/2402.05133
Those who are attending #SIGIRAP2024, come by and learn how retrieval can enhance ML models! @ACMSIGIR_AP
Today we'll be presenting the Tutorial on Retrieval-Enhanced Machine Learning (REML). Come by to learn about the emerging design patterns in this space and see how retrieval can be used beyond RAG. In collaboration w/ the amazing @841io @TEKnologyy @SalemiAlireza7 @HamedZamani
Contrastive VLMs (CLIP) lack the structure of text embeddings, like satisfying analogies via arithmetic (king - man = queen). We enhance CLIP’s *reasoning abilities* on such tasks by finetuning w/ text descriptions of image differences! w/ D. Willmott, J.Semedo, @zicokolter 1/🧵
Michael is an amazing mentor! I’ve really enjoyed working with him on multiple projects, and I highly recommend him as an advisor 😄. If you’re applying to CS PhD programs and interested in causality/reliable ML/healthcare, consider applying to work with him!
I'm recruiting PhD students for Fall 2025! CS PhD Deadline: Dec. 15th. I work on safe/reliable ML and causal inference, motivated by healthcare applications. Beyond myself, Johns Hopkins has a rich community of folks doing similar work! Come join us!
Unpacking what's going on in our paper on Med-* foundation models & their failure to improve over their generic counterparts (think Med-{LLaMa, LLaVa} vs {LLaMa, LLaVa}. A familiar tale of motivated reasoning, sloppy eval, & hidden hyper-optimization. arxiv.org/abs/2411.04118
Medically adapted foundation models (think Med-*) turn out to be more hot air than hot stuff. Correcting for fatal flaws in evaluation, the current crop are no better on balance than generic foundation models, even on the very tasks for which benefits are claimed.