Fiona Ryan
@fionakryan
Computer Science PhD student @GeorgiaTech and NSF Graduate Research Fellow interested in understanding human behavior with computer vision
There is 1 more week to submit non-archival extended abstracts to present at the Artificial Social Intelligence workshop @ICCVConference! We welcome work recently published in other venues (including the main ICCV conference) as well as works in progress!
Excited to announce the Artificial Social Intelligence Workshop @ ICCV 2025 @ICCVConference Join us in October to discuss the science of social intelligence and algorithms to advance socially-intelligent AI! Discussion will focus on reasoning, multimodality, and embodiment.
August 1 is the deadline for non-archival papers for the #ICCV2025 artificial social intelligence workshop! Papers in this track will be presented as posters at ICCV. We welcome any submissions of ongoing work, recently published work, or #ICCV papers in areas related to…
Excited to announce the Artificial Social Intelligence Workshop @ ICCV 2025 @ICCVConference Join us in October to discuss the science of social intelligence and algorithms to advance socially-intelligent AI! Discussion will focus on reasoning, multimodality, and embodiment.
Are AI scientists already better than human researchers? We recruited 43 PhD students to spend 3 months executing research ideas proposed by an LLM agent vs human experts. Main finding: LLM ideas result in worse projects than human ideas.
📣 the deadline for archival papers in the #ICCV2025 artificial social intelligence workshop is tomorrow! archival papers will be published in #ICCV workshop proceedings and presented we welcome any papers on social signal processing, social scene understanding, social…
Excited to announce the Artificial Social Intelligence Workshop @ ICCV 2025 @ICCVConference Join us in October to discuss the science of social intelligence and algorithms to advance socially-intelligent AI! Discussion will focus on reasoning, multimodality, and embodiment.
thanks for sharing our work! come talk with us at the Sunday PM poster session :)
CVPR 2025 papers pt. 1 - Gaze-LLE Gaze-LLE simplifies gaze target estimation by building on top of a frozen DINOv2 visual foundation model; SOTA performance; open source code and model more papers: github.com/SkalskiP/top-c… ↓ more
Very happy to be in Music City for #CVPR2025 My lab is presenting 7 papers, 4 selected as highlights. My amazing students @IrohXu @zixuan_huang @Wenqi_Jia @bryanislucky Xiang Li @fionakryan and postdoc Sangmin Lee are here! @siebelschool @uofigrainger
Future AI systems interacting with humans will need to perform social reasoning that is grounded in behavioral cues and external knowledge. We introduce Social Genome to study and advance this form of reasoning in models! New paper w/ Marian Qian, @pliang279, & @lpmorency!
We are very excited to host the Artificial Social Intelligence Workshop at #ICCV2025! We hope you'll join us there and consider submitting your work as a paper or extended abstract!
Excited to announce the Artificial Social Intelligence Workshop @ ICCV 2025 @ICCVConference Join us in October to discuss the science of social intelligence and algorithms to advance socially-intelligent AI! Discussion will focus on reasoning, multimodality, and embodiment.
This ICLR is the best conference ever. Attendees are extremely friendly and cuddly. ..What do you mean this is the wrong hall?
A new approach for easily merging data models is bringing adaptable, multi-tasking #AIs closer to reality. A @GeorgiaTech and @IBM team led by George Stoica and Pratik Ramesh (@pratikramesh7) "significantly enhances existing merging techniques" for data tasks necessary to advance…
📢#CVPR2025 Introducing InstaManip, a novel multimodal autoregressive model for few-shot image editing. 🎯InstaManip can learn a new image editing operation from textual and visual guidance via in-context learning, and apply it to new query images. [1/8] bolinlai.github.io/projects/Insta…
Extracting structure that’s implicitly learned by video foundation models _without_ relying on labeled data is a fundamental challenge. What’s a better place to start than extracting motion? Temporal correspondence is a key building block of perception. Check out our paper!
New paper on self-supervised optical flow and occlusion estimation from video foundation models. @sstj389 @jiajunwu_cs @SeKim1112 @Rahul_Venkatesh tinyurl.com/dpa3auzd @
Check out 🔥 EgoNormia: a benchmark for physical social norm understanding egonormia.org Can we really trust VLMs to make decisions that align with human norms? 👩⚖️ With EgoNormia, a 1800 ego-centric video 🥽 QA benchmark, we show that this is surprisingly challenging…
Congrats to all of our faculty who have received promotions and tenure! Promoted to full professor: @munmun10 @jhhays Granted tenure: Rosa Arriaga, @humphrey_shi Promoted to associate professor w/tenure: @judyfhoffman, @BoZhu_GT, Jessica Roberts
The first generation of Aria glasses have made a big impact in the research community, can't wait to see all the new possibilities these will unlock meta.com/blog/project-a…
happy to share that Gaze-LLE was accepted to #CVPR2025!
Introducing Gaze-LLE, a new model for gaze target estimation built on top of a frozen visual foundation model! Gaze-LLE achieves SOTA results on multiple benchmarks while learning minimal parameters, and shows strong generalization paper: arxiv.org/abs/2412.09586