Weixi Feng @ CVPR 2025
@weixi_feng
Final-year CS Ph.D. @UCSB. Ex-intern @NVIDIA/Adobe/Amazon. Working on visual generation and VLMs. Author of StructureDiff., LayoutGPT, BlobGEN-Vid.
🎉Thrilled to share my internship work with the @NVIDIA GenAIR team (accepted to #CVPR2025): BlobGEN-Vid: Compositional Text-to-Video Generation with Blob Video Representations! 🚀BlobGEN-Vid is a model-agnostic framework that delivers: - SOTA layout controllability - Enhanced…
Happy to announce that we have six papers accepted to ACL 2025. Five of them are main and one is findings. These papers cover a wide range of topics including LLM agent, Coder, Retrieval, Multimodal and Evaluation. Congrats to all the collaborators and students!
A wonderful start of 2025! Four papers I really like are accepted to #ICLR2025, covering four distinct directions: AI agents, multimodal safety, 3D vision, and video evaluation. Stay tuned for more updates! 🥳 Huge congrats to my students and collaborators and big thanks to ACs…
I am confused by @arxiv release nowadays. The email seems to imply that all submissions undergo checks that take 2 business days on average. But the schedule and cutoff days on the official website implies same day announcement if submitted before 2pm. If my submission is “on…

持续制作这个系列! 可达鸭睡在罗马斗兽场! 创意制作:@simonxxoo & @ring_hyacinth
Using racial labels to describe misconduct is harmful and inappropriate. @NeurIPSConf must not condone speech that targets specific ethnic groups. We urge Rosalind Picard @MIT @medialab to retract and apologize for her statement. Btw, most Rosalinds I know are honest and morally…
I will be graduating early next year and on the industrial job market! My expertise lies in post-training diffusion models, and I have a strong background in Deep RL. If you think my skills align with your needs, feel free to reach out!
Finally arrived at #NeurIPS2024 yesterday. Vancouver is great! 🇨🇦 Today, I’ll be presenting our T2V-Turbo paper at East Exhibit Hall A-C #4700 from 11 AM - 2 PM. If you are passionate about video generation and diffusion model alignment, please come say hi!
🚨😱Obligatory job market announcement post‼️🤯 I'm searching for faculty positions/postdocs in multimodal/multilingual NLP and generative AI! I'll be at #NeurIPS2024 presenting our work on meta-evaluation for text-to-image faithfulness! Let's chat! Website in bio, papers in🧵
🚀 Since its invention, the mouse has been our way to control computers. But what if it didn’t have to be? 🤔 Thrilled to introduce Agent S, a new state-of-the-art GUI agent framework that interacts with computers just like a human and takes on the toughest automation challenges.…
🚀Introducing ChipAgents: the World's First AI Agent for Chip Design and Verification. Get ready to supercharge your workflow and accelerate your time-to-market! 💻⚡
New career update: 🚀 Today, we’re thrilled to launch ChipAgents, our most ambitious project. The agentic AI chip design environment will allow engineers to iterate on your chip design & verification 10x faster by collaborating with ChipAgents in your favorite code editor. 🤖
🚀Introducing ChipAgents: the World's First AI Agent for Chip Design and Verification. Get ready to supercharge your workflow and accelerate your time-to-market! 💻⚡
Excited to share our work, GenArtist: Multimodal LLM as an Agent for Unified Image Generation and Editing, which has been accepted by NeurIPS 2024 as 🚩 Spotlight! Project page: zhenyuw16.github.io/GenArtist_page/ arXiv: arxiv.org/abs/2407.05600 Code: github.com/zhenyuw16/GenA…
🚀🚀I'm delighted to share that our paper "Reward Guided Latent Consistency Distillation" has been accepted by @TmlrOrg #TMLR with a 𝐅𝐞𝐚𝐭𝐮𝐫𝐞𝐝 𝐂𝐞𝐫𝐭𝐢𝐟𝐢𝐜𝐚𝐭𝐢𝐨𝐧! Kudos to my collaborator @weixi_feng and advisors @WenhuChen @WilliamWangNLP We integrate feedback…
Super excited that our paper gets accepted to #neurips2024 @NeurIPSConf ! Many thanks to my awesome collaborator @weixi_feng @tsujuifu @XinyiWang98 @ Sugato and great supervision from @WenhuChen @WilliamWangNLP If you are interested in our work, stay tuned! We are cooking a…
T2V-Turbo Breaking the Quality Bottleneck of Video Consistency Model with Mixed Reward Feedback Diffusion-based text-to-video (T2V) models have achieved significant success but continue to be hampered by the slow sampling speed of their iterative sampling processes. To
Our Discffusion paper is officially accepted to TMLR! First time submitting to TMLR and love the constructive reviewing process. Congrats to @XuehaiH and all co-authors @weixi_feng @jampani_varun @arjunreddy2613 @WilliamWangNLP! 🎉
Gen AI gains attention these days. Can generative models be used for discriminative tasks? Our collaborative work with UCSB and Google turns pre-trained text-to-image diffusion models into few-shot discriminative learners. Our approach mainly uses the cross-attention score of a…