#3 MMFM Workshop
@MMFMWorkshop
Workshop on Multi Modal Foundation Models
Are you in Seattle for #CVPR2024? Come and visit our poster today at 9:40 AM at the @MMFMWorkshop. We are looking forward to seeing you!
🚀 GPT-4, Gemini, Qwen, LLaVA: LLMs are stepping into the multi-modal arena with a bang! But let's zoom in on their vision 👁️. Our preprint peels back the layers on a crucial bias in vision models that sets most apart from humans: the texture/shape bias 👉 t.ly/jT1-R
And now… Arsha Nagrani is giving a talk on "How good are foundation models at video?"

Room 207, A-D We just started the 3rd MMFM workshop 🤩🤩🤩

Submit Submit Submit🤩🤩🤩
🚀 Call for Papers – 3rd Workshop on Multi-Modal Foundation Models (MMFM) @CVPR! 🚀 🔍 Topics: Multi-modal learning, vision-language, audio-visual, and more! 📅 Deadline: March 14, 2025 📝 Submit your paper: cmt3.research.microsoft.com/MMFM2025 🌐 More details: sites.google.com/view/mmfm3rdwo…
🚀 Call for Papers – 3rd Workshop on Multi-Modal Foundation Models (MMFM) @CVPR! 🚀 🔍 Topics: Multi-modal learning, vision-language, audio-visual, and more! 📅 Deadline: March 14, 2025 📝 Submit your paper: cmt3.research.microsoft.com/MMFM2025 🌐 More details: sites.google.com/view/mmfm3rdwo…
I am happy to share that LiveXiv accepted to ICLR 2025 🥳
Introducing LiveXiv, a new, challenging and maintainable scientific multi-modal live dataset Paper: arxiv.org/abs/2410.10783 Github: github.com/NimrodShabtay/… Dataset: huggingface.co/datasets/LiveX…
Excited to share that our 3rd Multimodal Workshop has been accepted to CVPR 2025 in Nashville! 🎉 Looking forward to advancing discussions on vision-language models, compositional reasoning, and contextual understanding. See you there! @CVPR
Ever wanted to locate your cat in a database of images using just one reference image? Probably not—but this highlights a gap in VLMs. They struggle to localize specific objects given in-context examples, often copying the last sample's location instead of learning from it.
With the recent widespread discussions on X about OpenAI's newest o1 model🍓arithmetic skills, this feels particularly timely. Excited to share that NumeroLogic, our work on improving LLMs' numerical reasoning capabilities, has been accepted to @emnlpmeeting! See you in Miami! 🥳
1/ Despite great advances in LLMs, challenges in numerical reasoning persist. Our new paper introduces #NumeroLogic, a novel number formatting that enhances numerical reasoning. @LChoshen @JosephShtok @SivanDoveh @leokarlin @ArbelleAssaf @IBMResearch @MITIBMLab @MIT
Meta-prompt got accepted to ECCV 🤩
final project of my phd makes it to @eccvconf 🥳 - here, find a new way to extract the abundant visual world knowledge in the llms for improving downstream visual tasks, with minimum human involvement (and bias): jmiemirza.github.io/Meta-Prompting/
Panel starts now 🥳 @jcniebles @YejinChoinka Alexie A.Efros Antonio Torralba

It is time to wake up and drink your coffee so you'll make it in time for the workshop today ☕️☕️☕️ 8:30-13:00 Summit 437-439 @CVPR
🔥🔥🔥Excited to present #HallusionBench at the @MMFMWorkshop and @CVPR! 🎉 Catch me 👀 at poster #97 on Tuesday, June 18, from 9:40-11:00 AM and at poster session #457 on Thursday, June 20, from 10:30 AM to noon. Come by and let's chat! 🥳
🔥🔥🔥Thrilled to announce that our #HallusionBench was accepted by #CVPR2024! See you in Seattle! @CVPR @gammaumd @UMDResearch @UMDscience #cvpr #VLMs #hallucination #gpt4v #Gemini #LLaVa