Kosta Derpanis
@CSProfKGD
#CS Assoc Prof @YorkUniversity, #ComputerVision Scientist Samsung #AI, @VectorInst Faculty Affiliate, TPAMI AE, @ELLISforEurope Member #ICCV2025 Publicity Chair
Lecture slides for my "Introduction to #ComputerVision" and "#DeepLearning in Computer Vision" courses. 🆕 Gaussian Splatting 🆕 Flow Matching The included videos do not contain voiceovers yet, planned for a future revision.


Moving a photo in Microsoft Word... so true 🤣🤣
Here's how my recent papers & reviews are going: * To solve a vision problem today, the sensible thing is to leverage a pre-trained VLM or video diffusion model. Such models implicitly represent a tremendous amount about the visual world that we can exploit. * Figure out how to…
Survey on feed-forward 3D reconstruction: arxiv.org/pdf/2507.14501. Very thorough overview and a great resource. IMO in the future we'll see a lot of intersection of reconstruction and generation, because we often want to recover the full 3D scene but never observe it completely.
Love having unstructured time. Why can’t the whole year be summer?
Really interesting paper. Fits the theme that we should make our modeling problems harder, not easier, so that they are forced to learn more and generalize better.
🚨 The era of infinite internet data is ending, So we ask: 👉 What’s the right generative modelling objective when data—not compute—is the bottleneck? TL;DR: ▶️Compute-constrained? Train Autoregressive models ▶️Data-constrained? Train Diffusion models Get ready for 🤿 1/n
this is pure beauty 👌👌 midjourney --- reddit .com/r/midjourney/comments/1m4rxhp/paint_world_4/
I had a great time presenting "It's Time to Say Goodbye to Hard Constraints" at the Flatiron Institute. In this talk, I describe a philosophy for model construction. Video now online! youtube.com/watch?v=LxuNC3…
Kicking off #ICCP2025! Highlights from Day 1 of the oral sessions.
We are hiring on the Veo team!📽️ Some people asked me about this at #ICML2025. If that's you, I will have told you to check deepmind.google/careers/ regularly. 👀It's just been updated: Europe (London, Zurich) job-boards.greenhouse.io/deepmind/jobs/… US (Mountain View) job-boards.greenhouse.io/deepmind/jobs/…
Want to be part of a team redefining SOTA for generative video models? Excited about building models that can reach billions of users? The Veo team is hiring! We are looking for amazing researchers and engineers, in North America and Europe. Details below:
Ready to present your latest work? The Call for Papers for #UniReps2025 @NeurIPSConf is open! 👉Check the CFP: unireps.org/2025/call-for-… 🔗 Submit your Full Paper or Extended Abstract here: openreview.net/group?id=NeurI… Speakers and panelists: @d_j_sutherland @elmelis @KriegeskorteLab…
I'm very excited to announce our #SIGGRAPH2025 workshop: Drawing & Sketching: Art, Psychology, and Computer Graphics 🎨🧠🫖 🔗 lines-and-minds.github.io 📅 Sunday, August 10th Join us to explore how people draw, how machines draw, and how the two might draw together! 🤖✍️
Preprint of today: Walker et al., "Generalist Forecasting with Frozen Video Models via Latent Diffusion" -- arxiv.org/abs/2507.13942 Maybe not surprising, but also very interesting -- learning to forecast strongly correlates with generalization. Like how LLMs came to be.
"We can confirm that Google DeepMind has reached the much-desired milestone, earning 35 out of a possible 42 points — a gold medal score. Their solutions were astonishing in many respects. IMO graders found them to be clear, precise and most of them easy to follow." IMO…
An advanced version of Gemini with Deep Think has officially achieved gold medal-level performance at the International Mathematical Olympiad. 🥇 It solved 5️⃣ out of 6️⃣ exceptionally difficult problems, involving algebra, combinatorics, geometry and number theory. Here’s how 🧵
Today we release Franca, a new vision Foundation Model that matches and sometimes outperforms DINOv2. The data, the training code and the model weights (with intermediate checkpoints) are open-source, allowing everyone to build on this. Methodologically, we introduce two new…
Can open-data models beat DINOv2? Today we release Franca, a fully open-sourced vision foundation model. Franca with ViT-G backbone matches (and often beats) proprietary models like SigLIPv2, CLIP, DINOv2 on various benchmarks setting a new standard for open-source research🧵
We will be waiting for you at #3DV2026 🤓
A #computervision researcher at #ICML2025
🚨 FINAL HOURS: Winter 2026 Research Internship applications close TODAY at 1PM ET! Don't miss your last chance to launch your AI career with Vector and contribute to research that shapes real-world applications. As a Vector Research Intern, you'll experience a unique journey…