AI/EXPLORER
ToolsCategoriesSitesLLMsCompareAI QuizAlternativesPremium
—AI Tools
—Sites & Blogs
—LLMs & Models
—Categories
AI Explorer

Find and compare the best artificial intelligence tools for your projects.

Made within France

Explore

  • ›All tools
  • ›Sites & Blogs
  • ›LLMs & Models
  • ›Compare
  • ›Chatbots
  • ›AI Images
  • ›Code & Dev

Company

  • ›Premium
  • ›About
  • ›Contact
  • ›Blog

Legal

  • ›Legal notice
  • ›Privacy
  • ›Terms

© 2026 AI Explorer·All rights reserved.

HomeLLMsEmbeddingNeoBERT

NeoBERT

by chandar-lab

Open source · 18k downloads · 194 likes

2.9
(194 reviews)EmbeddingAPI & Local
About

NeoBERT is a next-generation English text encoding model, specifically designed to efficiently represent textual data. With an optimized architecture and training on the extensive RefinedWeb dataset, it outperforms existing models like BERT or RoBERTa while remaining compact with just 250 million parameters. Its key advantage lies in its ability to process sequences up to 4,096 tokens, providing a deeper understanding of context than its predecessors. The model stands out for its energy efficiency and exceptional performance on benchmarks such as MTEB, making it an ideal choice for applications requiring high-quality text representation. It seamlessly integrates as a plug-and-play alternative into existing pipelines, simplifying adoption for various natural language processing use cases.

Documentation

NeoBERT

Hugging Face Model Card

NeoBERT is a next-generation encoder model for English text representation, pre-trained from scratch on the RefinedWeb dataset. NeoBERT integrates state-of-the-art advancements in architecture, modern data, and optimized pre-training methodologies. It is designed for seamless adoption: it serves as a plug-and-play replacement for existing base models, relies on an optimal depth-to-width ratio, and leverages an extended context length of 4,096 tokens. Despite its compact 250M parameter footprint, it is the most efficient model of its kind and achieves state-of-the-art results on the massive MTEB benchmark, outperforming BERT large, RoBERTa large, NomicBERT, and ModernBERT under identical fine-tuning conditions.

  • Paper: paper
  • Repository: github.

Get started

Ensure you have the following dependencies installed:

Bash
pip install transformers torch xformers==0.0.28.post3

If you would like to use sequence packing (un-padding), you will need to also install flash-attention:

Bash
pip install transformers torch xformers==0.0.28.post3 flash_attn

How to use

Load the model using Hugging Face Transformers:

Python
from transformers import AutoModel, AutoTokenizer

model_name = "chandar-lab/NeoBERT"
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
model = AutoModel.from_pretrained(model_name, trust_remote_code=True)

# Tokenize input text
text = "NeoBERT is the most efficient model of its kind!"
inputs = tokenizer(text, return_tensors="pt")

# Generate embeddings
outputs = model(**inputs)
embedding = outputs.last_hidden_state[:, 0, :]
print(embedding.shape)

Features

FeatureNeoBERT
Depth-to-width28 × 768
Parameter count250M
ActivationSwiGLU
Positional embeddingsRoPE
NormalizationPre-RMSNorm
Data SourceRefinedWeb
Data Size2.8 TB
Tokenizergoogle/bert
Context length4,096
MLM Masking Rate20%
OptimizerAdamW
SchedulerCosineDecay
Training Tokens2.1 T
EfficiencyFlashAttention

License

Model weights and code repository are licensed under the permissive MIT license.

Citation

If you use this model in your research, please cite:

Bibtex
@misc{breton2025neobertnextgenerationbert,
      title={NeoBERT: A Next-Generation BERT}, 
      author={Lola Le Breton and Quentin Fournier and Mariam El Mezouar and Sarath Chandar},
      year={2025},
      eprint={2502.19587},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2502.19587}, 
}

Contact

For questions, do not hesitate to reach out and open an issue on here or on our GitHub.


Capabilities & Tags
transformerssafetensorsneobertfill-maskfeature-extractioncustom_codeen
Links & Resources
Specifications
CategoryEmbedding
AccessAPI & Local
LicenseOpen Source
PricingOpen Source
Rating
2.9

Try NeoBERT

Access the model directly