Jiachen Zhao
@jcz12856876
PhD student @KhouryCollege | Scholar @MATSprogram | Prev: @UMassAmherst @HKUST
❓ How do you solve grasping problems when your target object is completely out of sight? 🚀 Excited to share our latest research! Check out ThinkGrasp: A Vision-Language System for Strategic Part Grasping in Clutter. 🔗 Site: h-freax.github.io/thinkgrasp_page
🤔Long-horizon tasks: How to train LLMs for the marathon?🌀 Submit anything on 🔁"Multi-turn Interactions in LLMs"🔁 to our @NeurIPSConf workshop by 08/22: 📕 Multi-Turn RL ⚖️ Multi-Turn Alignment 💬 Multi-Turn Human-AI Teaming 📊 Multi-Turn Eval ♾️You name it! #neurips #LLM
🚀 Call for Papers — @NeurIPSConf 2025 Workshop Multi-Turn Interactions in LLMs 📅 December 6/7 · 📍 San Diego Convention Center Join us to shape the future of interactive AI. Topics include but are not limited to: 🧠 Multi-Turn RL for Agentic Tasks (e.g., web & GUI agents,…
1/ 🔥 AI agents are reaching a breakthrough moment in cybersecurity. In our latest work: 🔓 CyberGym: AI agents discovered 15 zero-days in major open-source projects 💰 BountyBench: AI agents solved real-world bug bounty tasks worth tens of thousands of dollars 🤖…
What types of exemplar CoTs are better for In-Context Learning? Our #EMNLP paper shows that an LLM usually prefers its own generated CoTs as demonstrations for ICL. 📅I will present this paper in person on Wednesday 4pm at Poster Session E (Jasmine). Come visit our poster!
🎉 New paper alert! Large Language Models are In-context Teachers for Knowledge Reasoning #EMNLP24 finding 🔗 Read the paper: arxiv.org/abs/2311.06985 Work done by @jcz12856876 @YaoZonghai @YangZhichaoNLP and Prof. Hong Yu #BioNLP #InstructionTuning (0/N)
❗Are We Truly Achieving Multilingualism in LLMs or Just Relying on Translation?❗ Need multilingual instruction data and benchmarks? Just translate from English. LLM multilingualism can be easily solved! If you agree, check out our #EMNLP 2024 paper which says this is…
Our work has been accepted by #EMNLP2024 Findings ! So thankful for my wonderful co-authors !!! All my three projects during my Master's study @UMassAmherst now have happy endings!

☕We release the paper, “Diversity and Conquer: Diversity-Centric Data Selection with Iterative Refinement” 👉In this paper, we explore the critical role of diversity in data selection and introduce a novel approach with iterative selection. 📜arxiv.org/abs/2409.11378 🧵Below
🚨 New Research Alert! People have found safety training of LLMs can be easily undone through finetuning. How can we ensure safety in customized LLM finetuning while making finetuning still useful? Check out our latest work led by Jiachen Zhao! @jcz12856876 🔍 Our study reveals:…
In our #NeurIPS2023 paper Energy Transformer we propose a network that unifies three promising ideas in AI: Transformers, Energy-based Models (EBMs), and Associative Memory. The inference step in our model performs a descent dynamics on a specially engineered energy function. Our…
⚡️Energy Transformer (ET)⚡️ A novel architecture combining 3 prominent ideas in AI 1️⃣ Transformers: mix tokens with attention 2️⃣ Energy-based Models: inference descends a tractable energy function 3️⃣ Associative Memory: inference performs error correction #NeurIPS2023 A 🧵:…
I will have a poster presentation at ICML TEACH workshop (7/29). The paper is an extension of this tweet 😆that interprets ICL as retrieving from associative memory.
Is it possible that for the most of time, ChatGPT is only retrieving answers from its memory like a Hopfield Network or an IR system. How good is it for OOD cases
In your application letter for #PhD / postdoc, NEVER ever say: "Hi prof" "Hello" "Dear Professor" "Greetings of the day" If you do, your email will be immediately deleted by 99% of professors. ▫️ Only start your applications with “Dear Prof. [second_name],” And don’t…