Kenneth Huang
@windx0303
I create technology that helps users. Associate Prof @ISTatPENNSTATE; Run @PSUCrowdAILab. #HCI #NLProc #Crowdsourcing #Writing #Labeling. Ex @cmuhcii/@LTIatCMU.
We're excited to launch the 3rd SciCap Challenge! We're part of the @LM4Sci Workshop @COLM_conf #COLM2025. The top teams will present posters at the workshop. 🔥 Result Submission Deadline: Aug 11, 2025 🔥 Tech Report Deadline: Aug 15, 2025 Join now: scicap.ai/#/challenge/20…

I’m gonna be recruiting students thru both @LTIatCMU (NLP) and @CMU_EPP (Engineering and Public Policy) for fall 2026! If you are interested in reasoning, memorization, AI for science & discovery and of course privacy, u can catch me at ACL! Prospective students fill this form:
📣Thrilled to announce I’ll join Carnegie Mellon University (@CMU_EPP & @LTIatCMU) as an Assistant Professor starting Fall 2026! Until then, I’ll be a Research Scientist at @AIatMeta FAIR in SF, working with @kamalikac’s amazing team on privacy, security, and reasoning in LLMs!
👋 Attending ACL'2025? Don't miss out the opportunity to ask your questions about careers in NLP on July 30 11-12:00pm from our esteemed panel! Please submit your questions via the form ASAP, the latest by EOD July 28 to be considered. Thanks in advance!
📢ACL 2025 Industry Track 🧠 Do you have questions about a future in industry for NLP ? 👉 Submit your questions in this form: forms.gle/RC1aunT7WTSpfd… #ACL2025NLPcareers #IndustryTrack #NLProc #TechCareers
Everyone uses LLMs to annotate data or evaluate models in their research. But how can we convince others (readers, collaborators, reviewers!!!) that LLMs are reliable? 🤖 Here’s a simple (and low-effort) solution: show the LLM is a *comparable alternative annotator* ✅
in case you are wondering this is academia now
ICML’s Statement about subversive hidden LLM prompts We live in a weird timeline…
🗣️ Excited to share our new #ACL2025 Findings paper: “Just Put a Human in the Loop? Investigating LLM-Assisted Annotation for Subjective Tasks” with @jad_kabbara and @dkroy. Arxiv: arxiv.org/abs/2507.15821 Read about our findings ⤵️
Help Me Write a Story: Evaluating LLMs' Ability to Generate Writing Feedback. arxiv.org/abs/2507.16007
A comprehensive and accessible overview of LLMs for psychological assessment journals.sagepub.com/doi/full/10.11…
Very excited to share that an advanced version of Gemini Deep Think is the first to have achieved gold-medal level in the International Mathematical Olympiad! 🏆, solving five out of six problems perfectly, as verified by the IMO organizers! It’s been a wild run to lead this…
Super thrilled to share that our AI has has now reached silver medalist level in Math at #imo2024 (1 point away from 🥇)! Since Jan, we now not only have a much stronger version of #AlphaGeometry, but also an entirely new system called #AlphaProof, capable of solving many more…
tried ds*y again for another project. I can understand why people like it; instead of writing free-form system and user prompts and templates, you can write code and just rely on it. this part makes sense (though I think there's extra juice in customizing templates) however, ...
Fascinating new paper on AI companionship w/data donation from Character.ai by @Diyi_Yang and colleagues: arxiv.org/abs/2506.12605
Do laypeople "fall" for jargon-based explanations? No, not really
Have experience working on a system paper involving LLMs (any software that embeds an LLM call), and trying to publish it at HCI venues? We want to hear from you! See below 👇
🚨Calling all HCI researchers!—Have you authored a paper involving an LLM-integrated system over the past two years? Care to speak about your experiences with the publication process? 📃 DM or email me—we are running a study around this topic, and would love to hear from you.
Huan and I are looking for a postdoc to join us on agent research (broadly defined: planning, reasoning, safety, memory, continual learning, etc.). If you have a strong record in this space, drop us an email with CV! Retweet appreciated.
🚨 Postdoc Hiring: I am looking for a postdoc to work on rigorously evaluating and advancing the capabilities and safety of computer-use agents (CUAs), co-advised with @ysu_nlp @osunlp. We welcome strong applicants with experience in CUAs, long-horizon reasoning/planning,…
Creativity has been linked to better learning—but why? We find that both rely on associative thinking: the ability to form new connections. nature.com/articles/s4153…
Happy to share that FlawedFictions is now accepted in CoLM 2025 (@COLM_conf)! Looking forward to presenting this in Montreal later this year. Huge thanks to my wonderful collaborators @melaniesclar and @tsvetshop.
📢 New Paper! Tired 😴 of reasoning benchmarks full of math & code? In our work we consider the problem of reasoning for plot holes in stories -- inconsistencies in a storyline that break the internal logic or rules of a story’s world 🌎 W/ @melaniesclar, and @tsvetshop 1/n
🎉 Excited to announce that the 4th HCI+NLP workshop will be co-located with @EMNLP in Suzhou, China! 🌍📍 Join us to explore the intersection of human-computer interaction and NLP. 🧵 1/
CMU is hosting a workshop on Human-AI Complementarity for Decision Making this September! Abstract submissions due July 15, travel will be covered for accepted presenters. cmu.edu/ai-sdm/researc…
While I briefly have no employer, let me tell you what's really happening with AI companies training on public data: [1/n]