UMD CLIP Lab
@ClipUmd
Official account of the Computational Linguistics and Information Processing (CLIP) Lab at the University of Maryland
Welcome back to @ClipUmd Mohit Iyyer (@MohitIyyer)! After earning his Ph.D. @UofMaryland and working at @UMass, he’s returned as an associate professor, bringing his lab and cutting-edge work on LLMs and web agents. Read more: go.umd.edu/20fy

A new round of $750K in seed funding has been awarded to faculty and students who are advancing trustworthy AI at all four of TRAILS’ academic institutions: @UofMaryland, @GWtweets, @MorganStateU, and @Cornell. From disaster response to education, copyright law, and AI red…
I am at #ICML2025! 🇨🇦🏞️ Catch me: 1️⃣ Today at the @WiMLworkshop mentoring roundtables (1-2pm in W211-214) 2️⃣ Presenting this paper👇 tomorrow 11-11:30 at East #1205 3️⃣ At the Actionable Interpretability @ActInterp workshop on Saturday in East Ballroom A (I’m an organizer!)
Lots of progress in mech interp (MI) lately! But how can we measure when new mech interp methods yield real improvements over prior work? We propose 😎 𝗠𝗜𝗕: a Mechanistic Interpretability Benchmark!
In earlier work, we showed that neural topic model evaluation was broken, and those models didn't improve over classical methods the way people thought. This new paper provides a replacement paradigm that's grounded in the real-world requirements of qualitative content analysis.
(Repost due to mistaken deletion😢): Evaluating topic models (& doc clustering methods) is hard. In fact, since our paper critiquing standard eval practices 4 years ago, there hasn't been a good replacement metric That ends today! Our ACL paper introduces a new evaluation🧵
To be presented at ACL 2025: Large Language Models Are Biased Because They Are Large Language Models. Article: doi.org/10.1162/coli_a… Short (8min) video: youtube.com/watch?v=WLSuhe… #ACL2025NLP #NLProc #LLMs
The precursor to this paper "The Incoherence of Coherence" had our most-watched paper video ever, so I thought we had to surpass it somehow ... so we decided to do a song parody (of Roxanne, obviously).
(Repost due to mistaken deletion😢): Evaluating topic models (& doc clustering methods) is hard. In fact, since our paper critiquing standard eval practices 4 years ago, there hasn't been a good replacement metric That ends today! Our ACL paper introduces a new evaluation🧵
Andrew Childs, a professor of computer science, has been named Interim Director @umiacs. A noted expert in quantum information science and quantum algorithms, @andrewmchilds was co-director @JointQuICS for a decade and is currently director @RQuantumSim. umiacs.umd.edu/news-events/ne…
8/ 💌 Huge thanks to @MarineCarpuat, @rachelrudinger, and @zhoutianyi for their guidance — and special shoutout to the amazing @umdclip team! Check out our paper and code below 🚀 📄 Paper: arxiv.org/abs/2505.24671 🤖 Dataset: github.com/dayeonki/cultu…
1/ Are two #LLMs better than one for equitable cultural alignment? 🌍 We introduce a Multi-Agent Debate framework — where two LLM agents debate the cultural adaptability of a given scenario. #ACL2025 🧵👇
A bit late to announce, but I’m excited to share that I'll be starting as an assistant professor at the University of Maryland @umdcs this August. I'll be recruiting PhD students this upcoming cycle for fall 2026. (And if you're a UMD grad student, sign up for my fall seminar!)
I'm now a Ph.D. candidate! 🎉🥳 A few weeks ago, I proposed my thesis: "Teaching AI to Answer Questions with Reasoning that Actually Helps You". Thanks to my amazing committee + friends @ClipUmd! 🫶 I won't be back in Maryland for a while, some exciting things coming soon 👀
We’re hiring! 🎓 Feizi Lab has an open Postdoctoral Position in LLM Reasoning, supported by our PECASE award and other projects. If you're excited about foundational research in LLMs, apply here: forms.gle/uupq4wPz2EQnC4…
Soheil Feizi (@FeiziSoheil) is receiving $1 million in federal funding to advance the foundations of reasoning AI models. The funding is the result of the Presidential Early Career Award for Scientists and Engineers (#PECASE). Learn more: go.umd.edu/1zqs
🎉🎉 Excited to have two papers accepted to #ACL2025! Our first paper designs a preference training method to boost LLM personalization 🎨 While the second outlines our position on why MCQA evals are terrible and how to make them better 🙏 Grateful for amazing collaborators!
TRAILS Director Hal Daumé (@haldaume3) joined "AI or Not" the podcast with Pamela Isom to talk about his shift from technical AI research to its societal impact—touching on governance, copyright, and the need for stronger regulations. go.umd.edu/1yvy
🎉Our GRACE paper is heading to #ACL2025 Main conference! 🇦🇹 LLMs don’t just make mistakes; they make them with confidence, often more than people. Excited to push the boundaries of how we evaluate and understand LMs alongside humans! 👥🤝🤖 Grateful for amazing collab!
🚨 New Calibration Benchmark Alert: GRACE! GRACE is the first human-grounded calibration benchmark that directly compares LLMs to humans on when they're confidently wrong — miscalibration can mislead users into overtrusting AI. 🧵👇 🔗arxiv.org/pdf/2502.19684
Major Life Update 1/3 🥂🎊🎉 I passed my Ph.D. Preliminary Examination and am now a Ph.D. Candidate! I am almost a Doctor! Huge thanks to my amazing advisor @rachelrudinger and my committee @sarahwiegreffe @MarineCarpuat @haldaume3 @abhi2610 and Naomi Feldman for their support!
Philip Resnik (@psresnik) recently spoke with Capital News Service (@CNSmd) about how a powerful Reddit thread inspired his AI-driven research into why people choose to keep living through suicidal crises. cnsmaryland.org/2025/05/12/mac…
I feel so honored to win this award at #naaclmeeting #naacl2025 🥹 Cannot say how much grateful I am to my wonderful advisor @boydgraber and could not have done it without @maharshigor, @enfleisig, @IshaniMond66436 🙏