AI ExplorerAI Explorer
OutilsCatégoriesSitesLLMsComparerQuiz IAAlternativesPremium

—

Outils IA

—

Sites & Blogs

—

LLMs & Modèles

—

Catégories

AI Explorer

Trouvez et comparez les meilleurs outils d'intelligence artificielle pour vos projets.

Fait avecen France

Explorer

  • Tous les outils
  • Sites & Blogs
  • LLMs & Modèles
  • Comparer
  • Chatbots
  • Images IA
  • Code & Dev

Entreprise

  • Premium
  • À propos
  • Contact
  • Blog

Légal

  • Mentions légales
  • Confidentialité
  • CGV

© 2026 AI Explorer. Tous droits réservés.

AccueilLLMsOLMo 2 0425 1B

OLMo 2 0425 1B

par allenai

Open source · 116k downloads · 72 likes

2.3
(72 avis)ChatAPI & Local
À propos

OLMo 2 0425 1B est un modèle de langage de 1 milliard de paramètres, développé par l'Allen Institute for AI (Ai2), conçu pour générer du texte de manière autonome et comprendre le langage naturel en anglais. Issu de la famille OLMo, il se distingue par sa transparence totale, avec tous ses codes, jeux de données et détails d'entraînement rendus publics pour favoriser la recherche scientifique. Ce modèle est particulièrement adapté aux tâches nécessitant une compréhension contextuelle fine, comme la rédaction assistée, la synthèse d'informations ou l'analyse de données textuelles, tout en restant accessible grâce à sa taille réduite. Sa robustesse provient d'un entraînement en deux étapes sur des corpus variés, combinant données web de haute qualité et contenus spécialisés (académiques, questions-réponses, mathématiques). Enfin, son approche open source et son éthique de transparence en font un outil privilégié pour les chercheurs et développeurs souhaitant explorer les mécanismes des grands modèles de langage sans les contraintes des solutions propriétaires.

Documentation

Model Details

OLMo Logo

Model Card for OLMo 2 1B

We introduce OLMo 2 1B, the smallest model in the OLMo 2 family. OLMo 2 was pre-trained on OLMo-mix-1124 and uses Dolmino-mix-1124 for mid-training.

OLMo 2 is the latest in a series of Open Language Models designed to enable the science of language models. We have released all code, checkpoints, logs, and associated training details on GitHub.

SizeTraining TokensLayersHidden SizeAttention HeadsContext Length
OLMo 2-1B4 Trillion162048164096
OLMo 2-7B4 Trillion324096324096
OLMo 2-13B5 Trillion405120404096
OLMo 2-32B6 Trillion645120404096

The core models released in this batch include the following:

StageOLMo 2 1BOLMo 2 7BOLMo 2 13BOLMo 2 32B
Base Modelallenai/OLMo-2-0425-1Ballenai/OLMo-2-1124-7Ballenai/OLMo-2-1124-13Ballenai/OLMo-2-0325-32B
SFTallenai/OLMo-2-0425-1B-SFTallenai/OLMo-2-1124-7B-SFTallenai/OLMo-2-1124-13B-SFTallenai/OLMo-2-0325-32B-SFT
DPOallenai/OLMo-2-0425-1B-DPOallenai/OLMo-2-1124-7B-DPOallenai/OLMo-2-1124-13B-DPOallenai/OLMo-2-0325-32B-DPO
Final Models (RLVR)allenai/OLMo-2-0425-1B-Instructallenai/OLMo-2-1124-7B-Instructallenai/OLMo-2-1124-13B-Instructallenai/OLMo-2-0325-32B-Instruct
Reward Model (RM)allenai/OLMo-2-1124-7B-RM(Same as 7B)

Installation

OLMo 2 1B is supported in transformers v4.48 or higher:

Bash
pip install transformers>=4.48

If using vLLM, you will need to install from the main branch until v0.7.4 is released. Please

Inference

You can use OLMo with the standard HuggingFace transformers library:

Python
from transformers import AutoModelForCausalLM, AutoTokenizer
olmo = AutoModelForCausalLM.from_pretrained("allenai/OLMo-2-0425-1B")
tokenizer = AutoTokenizer.from_pretrained("allenai/OLMo-2-0425-1B")
message = ["Language modeling is "]
inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)
# optional verifying cuda
# inputs = {k: v.to('cuda') for k,v in inputs.items()}
# olmo = olmo.to('cuda')
response = olmo.generate(**inputs, max_new_tokens=100, do_sample=True, top_k=50, top_p=0.95)
print(tokenizer.batch_decode(response, skip_special_tokens=True)[0])
>> 'Language modeling is  a key component of any text-based application, but its effectiveness...'

For faster performance, you can quantize the model using the following method:

Python
AutoModelForCausalLM.from_pretrained("allenai/OLMo-2-0425-1B",
    torch_dtype=torch.float16,
    load_in_8bit=True)  # Requires bitsandbytes

The quantized model is more sensitive to data types and CUDA operations. To avoid potential issues, it's recommended to pass the inputs directly to CUDA using:

Python
inputs.input_ids.to('cuda')

We have released checkpoints for these models. For pretraining, the naming convention is stage1-stepXXX-tokensYYYB. For checkpoints with ingredients of the soup, the naming convention is stage2-ingredientN-stepXXX-tokensYYYB

To load a specific model revision with HuggingFace, simply add the argument revision:

Bash
olmo = AutoModelForCausalLM.from_pretrained("allenai/OLMo-2-0425-1B", revision="stage1-step140000-tokens294B")

Or, you can access all the revisions for the models via the following code snippet:

Python
from huggingface_hub import list_repo_refs
out = list_repo_refs("allenai/OLMo-2-0425-1B")
branches = [b.name for b in out.branches]

Fine-tuning

Model fine-tuning can be done from the final checkpoint (the main revision of this model) or many intermediate checkpoints. Two recipes for tuning are available.

  1. Fine-tune with the OLMo repository:
Bash
torchrun --nproc_per_node=8 scripts/train.py {path_to_train_config} \
    --data.paths=[{path_to_data}/input_ids.npy] \
    --data.label_mask_paths=[{path_to_data}/label_mask.npy] \
    --load_path={path_to_checkpoint} \
    --reset_trainer_state

For more documentation, see the GitHub README.

  1. Further fine-tuning support is being developing in AI2's Open Instruct repository. Details are here.

Model Description

  • Developed by: Allen Institute for AI (Ai2)
  • Model type: a Transformer style autoregressive language model.
  • Language(s) (NLP): English
  • License: The code and model are released under Apache 2.0.
  • Contact: Technical inquiries: [email protected]. Press: [email protected]
  • Date cutoff: Dec. 2023.

Model Sources

  • Project Page: https://allenai.org/olmo
  • Repositories:
    • Core repo (training, inference, fine-tuning etc.): https://github.com/allenai/OLMo
    • Evaluation code: https://github.com/allenai/OLMo-Eval
    • Further fine-tuning code: https://github.com/allenai/open-instruct
  • Paper: https://arxiv.org/abs/2501.00656

Evaluation

Core model results for OLMo 2 1B are found below.

Instruct ModelAvgFLOP×10²³AE2BBHDROPGSM8KIFEMATHMMLUSafetyPQATQA
Closed API models
GPT-3.5 Turbo 012560.5n/a38.766.670.274.366.941.270.269.145.062.9
GPT 4o Mini 072465.7n/a49.765.936.383.083.567.982.284.939.064.8
Open weights models 1-1.7B Parameters
SmolLM2 1.7B34.21.15.839.830.945.351.620.334.352.416.445.3
Gemma 3 1B38.31.220.439.425.135.060.640.338.970.29.643.8
Llama 3.1 1B39.36.710.140.232.245.454.021.646.787.213.841.5
Qwen 2.5 1.5B41.71.77.445.813.466.244.240.659.777.615.546.5
Fully-open models
OLMo 1B 072424.40.222.429.927.910.825.32.236.652.012.144.3
OLMo 2 1B42.70.359.135.034.668.370.120.740.087.612.948.7

Model Details

Training

OLMo 2 1BOLMo 2 7BOLMo 2 13BOLMo 2 32B
Pretraining Stage 14 trillion tokens
(1 epoch)
4 trillion tokens
(1 epoch)
5 trillion tokens
(1.2 epochs)
6 trillion tokens
(1.5 epochs)
Pretraining Stage 250B tokens50B tokens (3 runs)
merged
100B tokens (3 runs)
300B tokens (1 run)
merged
100B tokens (3 runs)
300B tokens (1 run)
merged
Post-trainingSFT+DPO+GRPO
(preference mix)
SFT + DPO + PPO
(preference mix)
SFT + DPO + PPO
(preference mix)
SFT + DPO + GRPO
(preference mix)

Stage 1: Initial Pretraining

  • Dataset: OLMo-mix-1124 (3.9T tokens)
  • Coverage: 95%+ of total pretraining budget
  • 1B Model: ~1 epoch

Stage 2: Mid-training

  • Dataset: Dolmino-Mix-1124
  • One training mix:
    • 50B tokens
  • Mix composition: 50% high-quality web data + academic/Q&A/instruction/math content

Model Merging

  • 1B Model: only 1 version is trained on a 50B mix (ingredient 3), we did not merge. Ingredients 1 and 2 are just exploratory runs.

Bias, Risks, and Limitations

Like any base or fine-tuned language model, AI can be prompted by users to generate harmful and sensitive content. Such content may also be produced unintentionally, especially in cases involving bias, so we recommend that users consider the risks when applying this technology. Additionally, many statements from OLMo or any LLM are often inaccurate, so facts should be verified.

Citation

INI
@misc{olmo20242olmo2furious,
      title={{2 OLMo 2 Furious}},
      author={Team OLMo and Pete Walsh and Luca Soldaini and Dirk Groeneveld and Kyle Lo and Shane Arora and Akshita Bhagia and Yuling Gu and Shengyi Huang and Matt Jordan and Nathan Lambert and Dustin Schwenk and Oyvind Tafjord and Taira Anderson and David Atkinson and Faeze Brahman and Christopher Clark and Pradeep Dasigi and Nouha Dziri and Michal Guerquin and Hamish Ivison and Pang Wei Koh and Jiacheng Liu and Saumya Malik and William Merrill and Lester James V. Miranda and Jacob Morrison and Tyler Murray and Crystal Nam and Valentina Pyatkin and Aman Rangapur and Michael Schmitz and Sam Skjonsberg and David Wadden and Christopher Wilhelm and Michael Wilson and Luke Zettlemoyer and Ali Farhadi and Noah A. Smith and Hannaneh Hajishirzi},
      year={2024},
      eprint={2501.00656},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2501.00656},
}

Model Card Contact

For errors in this model card, contact [email protected].

Liens & Ressources
Spécifications
CatégorieChat
AccèsAPI & Local
LicenceOpen Source
TarificationOpen Source
Paramètres1B parameters
Note
2.3

Essayer OLMo 2 0425 1B

Accédez directement au modèle