Prateek Jain
@jainprateek_
Learning machine learning at Google DeepMind.
Research Engineer, Tokyo: job-boards.greenhouse.io/deepmind/jobs/… Research Scientist, Tokyo: job-boards.greenhouse.io/deepmind/jobs/… Research Scientist, Bangalore: job-boards.greenhouse.io/deepmind/jobs/…
We're hiring @GoogleDeepMind APAC! 🇯🇵🇮🇳 Seeking experts in multilingual, multicultural & multimodal AI to help shape the future of our frontier models including Gemini. This is a unique opportunity to make impacts on billions of users.
@GoogleDeepMind India 🇮🇳 & Japan 🇯🇵 are looking for strong candidates in multilinguality, multicultural, & multimodality areas. RS Bangalore: job-boards.greenhouse.io/deepmind/jobs/… RS Tokyo: job-boards.greenhouse.io/deepmind/jobs/… RE Tokyo: job-boards.greenhouse.io/deepmind/jobs/…
@GoogleDeepMind India 🇮🇳 & Japan 🇯🇵 are looking for strong candidates in multilinguality, multicultural, & multimodality areas. RS Bangalore: job-boards.greenhouse.io/deepmind/jobs/… RS Tokyo: job-boards.greenhouse.io/deepmind/jobs/… RE Tokyo: job-boards.greenhouse.io/deepmind/jobs/…
🪆 Matryoshka is extremely general & applicable to every component in our modern ML/DL stack. It can't get more fundamental that 🪆 in bit space to enable elastic quantization! Drop by the poster and say hi to Puranjay (on behalf of @pranavn1008 @JeffDean @jainprateek_ & me).
Hi, I'll be presenting Matryoshka Quantization (arxiv.org/abs/2502.06786) on 16th July at #ICML2025 📍East Exhibition Hall A-B #3606 ⏲️ 11 AM - 1:30 PM
Check out our #ICML2025 poster on Matryoshka Quantization tomorrow, being presented by @puranjay1412 on behalf of all the authors.
Hi, I'll be presenting Matryoshka Quantization (arxiv.org/abs/2502.06786) on 16th July at #ICML2025 📍East Exhibition Hall A-B #3606 ⏲️ 11 AM - 1:30 PM
Nested matformer style models for faster parallel decoding! Sahil will present this poster at ICML today. Please check this one out if you are interested in architecture, efficiency etc. Additionally, Sahil will be applying for grad schools this cycle. He is brilliant, has a…
Hi, we'll be presenting MaGNeTS (arxiv.org/pdf/2502.00382) on 15th July at #ICML2025 📍East Exhibition Hall A-B #3209 🕦 11 AM - 1:30PM Excited to discuss about nested transformers and decode time scaling for visual generation!
Puranjay will present our poster on nested bitwise models or MatQuant, so if you are ICML and interested in the topic, do bother him :) Puranjay is going on the grad-school market this cycle. So if you are looking for a brilliant, hardworking student with good ML+LLM exposure,…
Hi, I'll be presenting Matryoshka Quantization (arxiv.org/abs/2502.06786) on 16th July at #ICML2025 📍East Exhibition Hall A-B #3606 ⏲️ 11 AM - 1:30 PM
Powerful gemini embeddings with support for flexible output dimensions 🪆🪆🪆
📢 The Gemini Embedding text model (gemini-embedding-001) is now generally available in the Gemini API via Google AI Studio. It supports 100+ languages and uses Matryoshka Representation Learning for flexible output dimensions, allowing devs to scale down from 3072 dimensions.
Embeddings are an important part of the magic of deep learning, allowing diverse data to be represented semantically as compact vectors. These embedding vectors can be used to retrieve, compare and classify data for downstream tasks. Gemini Embeddings model is now available!
Gemini Embeddings General Available! First Gemini Embedding model (`gemini-embedding-001`) now available for production use, ranking top on MMTEB optimized for finance, science, legal, search, code. 👀 - 🏦 $0.15 per million input tokens with free tier. - 🥇 Top MMTEB…
Gemini Embeddings General Available! First Gemini Embedding model (`gemini-embedding-001`) now available for production use, ranking top on MMTEB optimized for finance, science, legal, search, code. 👀 - 🏦 $0.15 per million input tokens with free tier. - 🥇 Top MMTEB…
Hi, we'll be presenting MaGNeTS (arxiv.org/pdf/2502.00382) on 15th July at #ICML2025 📍East Exhibition Hall A-B #3209 🕦 11 AM - 1:30PM Excited to discuss about nested transformers and decode time scaling for visual generation!
My first work @GoogleDeepMind accepted to ICML!
Google DeepMind just launched #Gemma3n! 🚀 They've released E2B & E4B, with more sizes coming for your specific hardware needs. What's mind-blowing? You can extract new, custom models from the bigger one.
My team is hiring a Technical Program Manager to help organize, accelerate, and empower world-class research. High-impact, high-growth role for someone passionate about AI and great at making things happen.
We are hiring Technical Program Manager to organize and enable our research teams to be the best at what they do and to make fast-paced progress towards our mission of bringing AGI responsibly. Ideal candidates should have a demonstrable record of strong program management…
We are hiring Technical Program Manager to organize and enable our research teams to be the best at what they do and to make fast-paced progress towards our mission of bringing AGI responsibly. Ideal candidates should have a demonstrable record of strong program management…
Similiar to how virtualization enabled elasticity of compute in Cloud, which led to huge adoption of Cloud, I have felt that Matryoshka architectures, devised by @jainprateek_, @adityakusupati & team are enabling elasticity of FM capability. People will recognize it over time.
Great to see the power of Matryoshka🪆 architectures, devised by my lab, getting widely highlighted as an indispensable component of the LLM cognitive core. cc @jainprateek_ @adityakusupati A sample list of Matryoshka papers from my lab: Matryoshka Representation Learning…
The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing: - Natively multimodal…
I’m so excited to announce Gemma 3n is here! 🎉 🔊Multimodal (text/audio/image/video) understanding 🤯Runs with as little as 2GB of RAM 🏆First model under 10B with @lmarena_ai score of 1300+ Available now on @huggingface, @kaggle, llama.cpp, ai.dev, and more
Great to see the power of Matryoshka🪆 architectures, devised by my lab, getting widely highlighted as an indispensable component of the LLM cognitive core. cc @jainprateek_ @adityakusupati A sample list of Matryoshka papers from my lab: Matryoshka Representation Learning…
The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing: - Natively multimodal…
It is great to see my first contributed project at Google, Gemma3n, is now fully open-sourced after Google I/O Preview annoucement! 🎉 Learned so much while contributing to parts of Gemma3n alongside amazing colleagues. We did a lot for both model quality and efficiency!
We’re fully releasing Gemma 3n, which brings powerful multimodal AI capabilities to edge devices. 🛠️ Here’s a snapshot of its innovations 🧵
Gemma 3n has arrived in llama.cpp 👨🍳 🍰 Comes in 2 flavors: E2B and E4B (E means "effective/active parameters")