Leo Bringer
@leo_bringer
ML Researcher @CraftyApesVFX 🦧 - Prev Associate Researcher @UMich 〽️ - 3D Vision, Diffusion Models, VideoGen & Character Animation 🤼♀️
🚀 Our paper **MDMP** has been accepted at CVPR’25 - HuMoGen 🚀 We propose a multi-modal diffusion model that fuses textual action descriptions and 3D skeletal data to generate long-term human motion predictions, with interpretable uncertainty — paving the way for safer and…
Such a cool feature!! Being able to import a 3d reconstructed scene through a Mesh format in Blender is a game changer. I guess the next step is 4D dynamic scenes with deformable meshes.
"MILo: Mesh-In-the-Loop Gaussian Splatting for Detailed and Efficient Surface Reconstruction" TL;DR: differentiably extract a mesh including both vertex locations and connectivity only from Gaussian parameters. Gradient flow from the mesh to GS
Transformers support is coming today! 🤗
Kyutai Speech-To-Text is now open-source! It’s streaming, supports batched inference, and runs blazingly fast: perfect for interactive applications. Check out the details here: kyutai.org/next/stt
It's clear that video diffusion models know a lot about the 3D world, material properties, and lighting. The trick is to get control over these. With a tiny amount of synthetic data, we can train a video model to realistically relight a single image. This is a neat trick that…
🚀 Introducing GenLit – Reformulating Single-Image Relighting as Video Generation! We leverage video diffusion models to perform realistic near-field relighting from just a single image—No explicit 3D reconstruction or ray tracing required! No intermediate graphics buffers,…
Presented our poster at CVPR 2025 / HuMoGen this week in Nashville! Amazing to share ideas with the community and see our work in motion🤸 Thanks to the @humogen11384 organizers for making it all happen. 🔗 github.com/leob03/mdmp #CVPR2025 #HuMoGen #AI #MotionPrediction…
🚀 Our paper **MDMP** has been accepted at CVPR’25 - HuMoGen 🚀 We propose a multi-modal diffusion model that fuses textual action descriptions and 3D skeletal data to generate long-term human motion predictions, with interpretable uncertainty — paving the way for safer and…
Open Sourcing Forge: 3D Gaussian splat rendering for web developers! 3DGS has become a dominant paradigm for differentiable rendering, combining high visual quality and real-time rendering. However, support for splatting on the web still lags behind its adoption in AI.
🚨 Paper Alert Our recent breakthrough CAST: Component-Aligned 3D Scene Reconstruction from an RGB Image has been accepted by @siggraph 2025 Journal Track! CAST will change the way create scenes in 3D Art and Embody AI. 🚀Soon available at HYPER3D.AI 👇Details
I am really super happy to show you my research that transform 3D volumetric capture of man capture with @kartel_ai and with env gaussian splatting generated with @theworldlabs in a three.js player . And all that is "upscale" using Hunyuan vid2vid model with a Lora train on…
very interesting article on Spatial Intelligence and 3D Awareness through Point Tracking: deepmind-tapir.github.io/blogpost.html From my experience with current Point Tracking technics, the main issue that I've run into for generating a video by animating an image conditioned on Point…
I have been working on testing it in Blender and Nuke and the estimations of 3d camera trajectories of DPVO+SLAM are pretty impressive, could be very useful for matchmoving
Check out our #IROS2024 paper "Deep Visual Odometry with Events and Frames," the new state of the art in Visual Odometry, which outperforms learning-based image methods (DROID-SLAM, DPVO), model-based methods (ORB-SLAM, DSO) and event-based methods (DEVO, EDS) by up to 60%…
Exciting news! MegaSAM code is out🔥 & the updated Shape of Motion results with MegaSAM are really impressive! A year ago I didn't think we could make any progress on these videos: shape-of-motion.github.io/results.html Huge congrats to everyone involved and the community 🎉
🛠️ Blender MCP update: Bring in high-quality assets to Blender through just prompts, thanks to an integration with @polyhaven’s library of 1500+ assets. Watch me create a beachy scene with just words. (Watch how it imports the right HDRIs, textures and models) 👇