Andrew Owens
@andrewhowens
Associate professor of CS @ Cornell Tech
I had a lot of fun helping put this problem set together -- if you're teaching diffusion models + computer vision, consider using this homework for your course! (links at end of @ryan_tabrizi's thread!)
Teaching computer vision next semester? Hoping to finally learn about diffusion models in 2025? Check out this diffusion project that we designed and test-drove this past semester at Berkeley and Michigan!
Teaching computer vision next semester? Hoping to finally learn about diffusion models in 2025? Check out this diffusion project that we designed and test-drove this past semester at Berkeley and Michigan!
Our computer vision textbook is now available for free online here: visionbook.mit.edu We are working on adding some interactive components like search and (beta) integration with LLMs. Hope this is useful and feel free to submit Github issues to help us improve the text!
Sharing our #CVPR2025 paper: "GPS as a Control Signal for Image Generation"! 🛰️+✍️ We turn the GPS tag stored in EXIF of photos into a control signal for diffusion models—so they don’t just know what you asked for, but where you want it to look like. Come to see our poster at…
Ever wondered how a scene sounds👂 when you interact👋 with it? Introducing our #CVPR2025 work "Hearing Hands: Generating Sounds from Physical Interactions in 3D Scenes" -- we make 3D scene reconstructions audibly interactive! yimingdou.com/hearing_hands/
Excited to share our CVPR 2025 paper on cross-modal space-time correspondence! We present a method to match pixels across different modalities (RGB-Depth, RGB-Thermal, Photo-Sketch, and cross-style images) — trained entirely using unpaired data and self-supervision. Our…
Can AI image detectors keep up with new fakes? Mostly, no. Existing detectors are trained using a handful of models. But there are thousands in the wild! Our work, Community Forensics, uses 4800+ generators to train detectors that generalize to new fakes. #CVPR2025 🧵 (1/5)
Hello! If you like pretty images and videos and want a rec for CVPR oral session, you should def go to Image/Video Gen, Friday at 9am: I'll be presenting "Motion Prompting" @RyanBurgert will be presenting "Go with the Flow" and @ChangPasca1650 will be presenting "LookingGlass"
3D illusions are fascinating! 🤩 But it takes exceptional artistic skills to make one. We present Illusion3D - a simple method for creating 3D multiview illusions, where the interpretations change depending on your perspectives. Let's play Where's Waldo, shall we? 😆
I'll be presenting "Images that Sound" today at #NeurIPS2024! East Exhibit Hall A-C #2710. Come say hi to me and @andrewhowens :) (@CzyangChen sadly could not make it, but will be there in spirit :') )
These spectrograms look like images, but can also be played as a sound! We call these images that sound. How do we make them? Look and listen below to find out, and to see more examples!
What happens when you train a video generation model to be conditioned on motion? Turns out you can perform "motion prompting," just like you might prompt an LLM! Doing so enables many different capabilities. Here’s a few examples – check out this thread 🧵 for more results!
🎥 Introducing MultiFoley, a video-aware audio generation method with multimodal controls! 🔊 We can ⌨️Make a typewriter sound like a piano 🎹 🐱Make a cat meow like a lion roars! 🦁 ⏱️Perfectly time existing SFX 💥 to a video
I’m recruiting PhD students with computer vision, robotics, or ML experience! We especially encourage applicants from physics and related fields who want to explore AI for Science. Join us by applying to @UMichCSE's PhD program!
I am recruiting exceptional PhD students & postdocs with an adventurous soul for my💫new TTIC AI lab💫! We aim to understand intelligence, one pixel at a time, inspired by psychology, neuroscience, language, robotics, and the arts. ttic.edu/studentapplica… sites.google.com/ttic.edu/openh…
Hiring research interns to work on AI for Sound Design! 💥🛸🐲 Join the Sound Design AI Group (SODA) for a summer 2025 internship advancing the bleeding edge of audio AI for video with @pseetharaman @urinieto and me. A chance to impact millions of users: youtube.com/watch?v=RddSWo…
Ready to make an impact with cutting-edge research? @AdobeResearch is now accepting applications for summer 2025 internships from master's and PhD students. Apply now! adobe.ly/4dS1zc6
Couldn’t be more excited—hundreds of people have gathered from across Michigan (& beyond!) for our 6th Michigan AI Symposium. Special theme this year Embodied AI, co-chaired by Stella Yu & @AndrewHOwens. Warm intros by a bipedal robot 🤖 and a humanoid Provost, Laurie McCauley.