EmbeddingGemma

Creator
Creator
Seonglae ChoSeonglae Cho
Created
Created
2025 Dec 31 12:42
Editor
Edited
Edited
2025 Dec 31 12:57
Lightweight text embedding model: 308M parameters, runs on-device with less than 200MB RAM when quantized. Achieves top performance among sub-500M models on MTEB (multilingual, English, and code).
  • Gemma 3
    -based encoder-decoder initialization. Gemma 3 is a decoder-only LLM. It's further trained as an encoder-decoder structure, then only the encoder is extracted and used as the initial weights for the embedding model.
  • Model Souping
    : averaging weights from multiple models trained on different data mixes. "Merging a brain good at retrieval + a brain good at classification 50/50"
  • MRL (Matryoshka Representation Learning): supports variable embedding dimensions with MRL.
    Matryoshka Embedding
 
 
 
 
arxiv.org
EmbeddingGemma
Google DeepMind
EmbeddingGemma
Huggingface
EmbeddingGemma - a google Collection
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
EmbeddingGemma - a google Collection
How to use
Welcome EmbeddingGemma, Google's new efficient embedding model
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Welcome EmbeddingGemma, Google's new efficient embedding model
 

Recommendations