AI/EXPLORER
OutilsCatégoriesSitesLLMsComparerQuiz IAAlternativesPremium
—Outils IA
—Sites & Blogs
—LLMs & Modèles
—Catégories
AI Explorer

Trouvez et comparez les meilleurs outils d'intelligence artificielle pour vos projets.

Fait avecen France

Explorer

  • ›Tous les outils
  • ›Sites & Blogs
  • ›LLMs & Modèles
  • ›Comparer
  • ›Chatbots
  • ›Images IA
  • ›Code & Dev

Entreprise

  • ›Premium
  • ›À propos
  • ›Contact
  • ›Blog

Légal

  • ›Mentions légales
  • ›Confidentialité
  • ›CGV

© 2026 AI Explorer·Tous droits réservés.

AccueilLLMszembed 1

zembed 1

par zeroentropy

Open source · 85k downloads · 89 likes

2.4
(89 avis)EmbeddingAPI & Local
À propos

zembed-1 est un modèle d'embeddings open-source de pointe, spécialement conçu pour les systèmes de recherche et de récupération d'informations. Il surpasse les principaux modèles fermés comme ceux d'OpenAI, Cohere ou Google dans des domaines variés tels que la finance, la santé, le juridique, les conversations, l'industrie, le code et les sciences. Entraîné à partir d'un reranker avancé via une méthode innovante appelée zELO, il offre une précision supérieure grâce à une modélisation fine des scores de pertinence. Ce modèle se distingue par sa flexibilité, permettant de projeter ses embeddings dans différentes dimensions (de 2560 à 40) et de les quantifier jusqu'à des formats binaires, réduisant considérablement leur taille sans sacrifier trop de précision. Multilingue par conception, il intègre plus de la moitié de ses données d'entraînement en langues autres que l'anglais, ce qui le rend particulièrement adapté aux applications internationales. Disponible sous licence non commerciale, il représente une alternative performante et accessible aux solutions propriétaires dominantes.

Documentation

Releasing zeroentropy/zembed-1

In retrieval systems, embedding models determine the quality of your search.

However, SOTA embedding models are closed-source and proprietary. At ZeroEntropy, we've trained a SOTA 4B open-weight multilingual embedding model that outperforms every competitor we benchmarked, and we're launching it here on HuggingFace.

This model outperforms OpenAI text-embedding-large, Cohere Embed v4, gemini-embedding-001, and voyage-4-nano across finance, healthcare, legal, conversational, manufacturing, code, and STEM.

zembed-1 is distilled directly from our SOTA reranker zerank-2 using our zELO methodology, which models relevance scores as adjusted Elo ratings. Standard contrastive training on binary labels can't match this signal. See our blog post for details.

The model supports flexible dimension projections (2560, 1280, 640, 320, 160, 80, 40) and quantization down to binary, compressing a full 8 KB vector to under 128 bytes with a controlled accuracy trade-off. See our Technical Report (Coming soon!) for details on the projection method. zembed-1 is multilingual from the ground up, with over half the training data in non-English languages.

This model is released under a non-commercial license. If you'd like a commercial license, please contact us at [email protected].

Model Details

PropertyValue
Parameters4B
Context Length32,768 tokens (32k)
Base ModelQwen/Qwen3-4B
Embedding Dimensions2560, 1280, 640, 320, 160, 80, 40
LicenseCC-BY-NC-4.0

How to Use

Python
from sentence_transformers import SentenceTransformer

# Initialize model
model = SentenceTransformer(
    "zeroentropy/zembed-1",
    trust_remote_code=True,
    model_kwargs={"torch_dtype": "bfloat16"},
)

# Define query and documents
query = "What is backpropagation?"
documents = [
    "Backpropagation is a fundamental algorithm for training neural networks by computing gradients.",
    "Gradient descent is used to optimize model parameters during the training process.",
    "Neural network training relies on efficient computation of derivatives through backpropagation.",
]

# Encode query and documents (uses task-specific prompts automatically)
query_embeddings = model.encode_query(query)
document_embeddings = model.encode_document(documents)
# (2560,) (3, 2560)

# Compute cosine similarities
similarities = model.similarity(query_embeddings, document_embeddings)
# tensor([[0.7525, 0.5670, 0.6835]])

The model can also be used through ZeroEntropy's /models/embed endpoint.

Evaluations

NDCG@10 scores between zembed-1 and competing embedding models, averaged across public and private benchmarks per domain. Full per-benchmark breakdown here.

DomainZeroEntropy zembed-1voyage-4-nanoQwen3 4BCohere Embed v4gemini-embed-001jina-v5-smallOpenAI Largebge-m3
Finance0.44760.42270.37150.36700.32910.35760.32910.3085
Healthcare0.62600.53560.51340.47500.50080.51320.53150.3620
Legal0.67230.59570.58580.58940.60690.57160.50990.5207
Conversational0.53850.40450.40340.42440.42470.44300.39880.3296
Manufacturing0.55560.48570.49320.49190.46640.47250.47360.3736
Web Search0.61650.59770.69140.72420.58810.67720.67500.6311
Code0.64520.64150.63790.62770.63050.63540.61550.5584
STEM & Math0.52830.50120.52190.46980.48400.37800.39050.3399
Enterprise0.37500.36000.29350.29150.32240.30120.33070.2213
Average0.55610.50500.50130.49570.48370.48330.47270.4050
Bar chart comparing zembed-1 NDCG@10 scores against competing embedding models across domains
Liens & Ressources
Spécifications
CatégorieEmbedding
AccèsAPI & Local
LicenceOpen Source
TarificationOpen Source
Note
2.4

Essayer zembed 1

Accédez directement au modèle