AI ExplorerAI Explorer
OutilsCatégoriesSitesLLMsComparerQuiz IAAlternativesPremium

—

Outils IA

—

Sites & Blogs

—

LLMs & Modèles

—

Catégories

AI Explorer

Trouvez et comparez les meilleurs outils d'intelligence artificielle pour vos projets.

Fait avecen France

Explorer

  • Tous les outils
  • Sites & Blogs
  • LLMs & Modèles
  • Comparer
  • Chatbots
  • Images IA
  • Code & Dev

Entreprise

  • Premium
  • À propos
  • Contact
  • Blog

Légal

  • Mentions légales
  • Confidentialité
  • CGV

© 2026 AI Explorer. Tous droits réservés.

AccueilLLMszephyr 7b beta

zephyr 7b beta

par HuggingFaceH4

Open source · 113k downloads · 1840 likes

4.1
(1840 avis)ChatAPI & Local
À propos

Zephyr 7B β est un modèle de langage conçu pour agir comme un assistant conversationnel utile et performant. Fine-tuné à partir de Mistral-7B, il se distingue par son entraînement sur des jeux de données synthétiques variés, optimisé via la méthode DPO pour améliorer ses capacités d'assistance tout en réduisant les biais d'alignement. Principalement axé sur l'anglais, il excelle dans les tâches de dialogue, surpassant d'autres modèles open source de taille similaire sur des benchmarks comme MT-Bench et AlpacaEval. Ses cas d'usage incluent la génération de réponses naturelles, l'aide à la rédaction ou à la résolution de problèmes simples, bien que ses performances restent limitées sur des tâches complexes comme la programmation avancée ou les mathématiques. Contrairement à des modèles propriétaires, Zephyr 7B β offre une alternative open source, mais son absence de filtrage renforcé le rend plus susceptible de produire des réponses inappropriées si sollicité de manière malveillante.

Documentation
Zephyr Logo

Model Card for Zephyr 7B β

Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr-7B-β is the second model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1 that was trained on on a mix of publicly available, synthetic datasets using Direct Preference Optimization (DPO). We found that removing the in-built alignment of these datasets boosted performance on MT Bench and made the model more helpful. However, this means that model is likely to generate problematic text when prompted to do so. You can find more details in the technical report.

Model description

  • Model type: A 7B parameter GPT-like model fine-tuned on a mix of publicly available, synthetic datasets.
  • Language(s) (NLP): Primarily English
  • License: MIT
  • Finetuned from model: mistralai/Mistral-7B-v0.1

Model Sources

  • Repository: https://github.com/huggingface/alignment-handbook
  • Demo: https://huggingface.co/spaces/HuggingFaceH4/zephyr-chat
  • Chatbot Arena: Evaluate Zephyr 7B against 10+ LLMs in the LMSYS arena: http://arena.lmsys.org

Performance

At the time of release, Zephyr-7B-β is the highest ranked 7B chat model on the MT-Bench and AlpacaEval benchmarks:

ModelSizeAlignmentMT-Bench (score)AlpacaEval (win rate %)
StableLM-Tuned-α7BdSFT2.75-
MPT-Chat7BdSFT5.42-
Xwin-LMv0.17BdPPO6.1987.83
Mistral-Instructv0.17B-6.84-
Zephyr-7b-α7BdDPO6.88-
Zephyr-7b-β 🪁7BdDPO7.3490.60
Falcon-Instruct40BdSFT5.1745.71
Guanaco65BSFT6.4171.80
Llama2-Chat70BRLHF6.8692.66
Vicuna v1.333BdSFT7.1288.99
WizardLM v1.070BdSFT7.71-
Xwin-LM v0.170BdPPO-95.57
GPT-3.5-turbo-RLHF7.9489.37
Claude 2-RLHF8.0691.36
GPT-4-RLHF8.9995.28

In particular, on several categories of MT-Bench, Zephyr-7B-β has strong performance compared to larger open models like Llama2-Chat-70B:

image/png

However, on more complex tasks like coding and mathematics, Zephyr-7B-β lags behind proprietary models and more research is needed to close the gap.

Intended uses & limitations

The model was initially fine-tuned on a filtered and preprocessed of the UltraChat dataset, which contains a diverse range of synthetic dialogues generated by ChatGPT. We then further aligned the model with 🤗 TRL's DPOTrainer on the openbmb/UltraFeedback dataset, which contains 64k prompts and model completions that are ranked by GPT-4. As a result, the model can be used for chat and you can check out our demo to test its capabilities.

You can find the datasets used for training Zephyr-7B-β here

Here's how you can run the model using the pipeline() function from 🤗 Transformers:

Python
# Install transformers from source - only needed for versions <= v4.34
# pip install git+https://github.com/huggingface/transformers.git
# pip install accelerate

import torch
from transformers import pipeline

pipe = pipeline("text-generation", model="HuggingFaceH4/zephyr-7b-beta", torch_dtype=torch.bfloat16, device_map="auto")

# We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
messages = [
    {
        "role": "system",
        "content": "You are a friendly chatbot who always responds in the style of a pirate",
    },
    {"role": "user", "content": "How many helicopters can a human eat in one sitting?"},
]
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
# <|system|>
# You are a friendly chatbot who always responds in the style of a pirate.</s>
# <|user|>
# How many helicopters can a human eat in one sitting?</s>
# <|assistant|>
# Ah, me hearty matey! But yer question be a puzzler! A human cannot eat a helicopter in one sitting, as helicopters are not edible. They be made of metal, plastic, and other materials, not food!

Bias, Risks, and Limitations

Zephyr-7B-β has not been aligned to human preferences for safety within the RLHF phase or deployed with in-the-loop filtering of responses like ChatGPT, so the model can produce problematic outputs (especially when prompted to do so). It is also unknown what the size and composition of the corpus was used to train the base model (mistralai/Mistral-7B-v0.1), however it is likely to have included a mix of Web data and technical sources like books and code. See the Falcon 180B model card for an example of this.

Training and evaluation data

During DPO training, this model achieves the following results on the evaluation set:

  • Loss: 0.7496
  • Rewards/chosen: -4.5221
  • Rewards/rejected: -8.3184
  • Rewards/accuracies: 0.7812
  • Rewards/margins: 3.7963
  • Logps/rejected: -340.1541
  • Logps/chosen: -299.4561
  • Logits/rejected: -2.3081
  • Logits/chosen: -2.3531

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-07
  • train_batch_size: 2
  • eval_batch_size: 4
  • seed: 42
  • distributed_type: multi-GPU
  • num_devices: 16
  • total_train_batch_size: 32
  • total_eval_batch_size: 64
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 3.0

Training results

The table below shows the full set of DPO training metrics:

Training LossEpochStepValidation LossRewards/chosenRewards/rejectedRewards/accuraciesRewards/marginsLogps/rejectedLogps/chosenLogits/rejectedLogits/chosen
0.62840.051000.60980.0425-0.18720.73440.2297-258.8416-253.8099-2.7976-2.8234
0.49080.12000.5426-0.0279-0.68420.750.6563-263.8124-254.5145-2.7719-2.7960
0.52640.153000.53240.0414-0.97930.76561.0207-266.7627-253.8209-2.7892-2.8122
0.55360.214000.4957-0.0185-1.52760.79691.5091-272.2460-254.4203-2.8542-2.8764
0.53620.265000.5031-0.2630-1.59170.78121.3287-272.8869-256.8653-2.8702-2.8958
0.59660.316000.5963-0.2993-1.64910.78121.3499-273.4614-257.2279-2.8778-2.8986
0.50140.367000.5382-0.2859-1.47500.751.1891-271.7204-257.0942-2.7659-2.7869
0.53340.418000.5677-0.4289-1.89680.79691.4679-275.9378-258.5242-2.7053-2.7265
0.52510.469000.5772-0.2116-1.31070.73441.0991-270.0768-256.3507-2.8463-2.8662
0.52050.5210000.5262-0.3792-1.85850.71881.4793-275.5552-258.0276-2.7893-2.7979
0.50940.5711000.5433-0.6279-1.93680.79691.3089-276.3377-260.5136-2.7453-2.7536
0.58370.6212000.5349-0.3780-1.95840.76561.5804-276.5542-258.0154-2.7643-2.7756
0.52140.6713000.5732-1.0055-2.23060.76561.2251-279.2761-264.2903-2.6986-2.7113
0.69140.7214000.5137-0.6912-2.17750.79691.4863-278.7448-261.1467-2.7166-2.7275
0.46550.7715000.5090-0.7987-2.29300.70311.4943-279.8999-262.2220-2.6651-2.6838
0.57310.8316000.5312-0.8253-2.35200.78121.5268-280.4902-262.4876-2.6543-2.6728
0.52330.8817000.5206-0.4573-2.09510.78121.6377-277.9205-258.8084-2.6870-2.7097
0.55930.9318000.5231-0.5508-2.20000.79691.6492-278.9703-259.7433-2.6221-2.6519
0.49670.9819000.5290-0.5340-1.95700.82811.4230-276.5395-259.5749-2.6564-2.6878
0.09211.0320000.5368-1.1376-3.16150.78122.0239-288.5854-265.6111-2.6040-2.6345
0.07331.0821000.5453-1.1045-3.44510.76562.3406-291.4208-265.2799-2.6289-2.6595
0.09721.1422000.5571-1.6915-3.98230.81252.2908-296.7934-271.1505-2.6471-2.6709
0.10581.1923000.5789-1.0621-3.89410.79692.8319-295.9106-264.8563-2.5527-2.5798
0.24231.2424000.5455-1.1963-3.55900.78122.3627-292.5599-266.1981-2.5414-2.5784
0.11771.2925000.5889-1.8141-4.39420.79692.5801-300.9120-272.3761-2.4802-2.5189
0.12131.3426000.5683-1.4608-3.84200.81252.3812-295.3901-268.8436-2.4774-2.5207
0.08891.3927000.5890-1.6007-3.73370.78122.1330-294.3068-270.2423-2.4123-2.4522
0.09951.4528000.6073-1.5519-3.83620.82812.2843-295.3315-269.7538-2.4685-2.5050
0.11451.529000.5790-1.7939-4.28760.84382.4937-299.8461-272.1744-2.4272-2.4674
0.06441.5530000.5735-1.7285-4.20510.81252.4766-299.0209-271.5201-2.4193-2.4574
0.07981.631000.5537-1.7226-4.28500.84382.5624-299.8200-271.4610-2.5367-2.5696
0.10131.6532000.5575-1.5715-3.98130.8752.4098-296.7825-269.9498-2.4926-2.5267
0.12541.733000.5905-1.6412-4.47030.85942.8291-301.6730-270.6473-2.5017-2.5340
0.0851.7634000.6133-1.9159-4.67600.84382.7601-303.7296-273.3941-2.4614-2.4960
0.0651.8135000.6074-1.8237-4.35250.85942.5288-300.4951-272.4724-2.4597-2.5004
0.07551.8636000.5836-1.9252-4.40050.81252.4753-300.9748-273.4872-2.4327-2.4716
0.07461.9137000.5789-1.9280-4.49060.81252.5626-301.8762-273.5149-2.4686-2.5115
0.13481.9638000.6015-1.8658-4.24280.82812.3769-299.3976-272.8936-2.4943-2.5393
0.02172.0139000.6122-2.3335-4.92290.82812.5894-306.1988-277.5699-2.4841-2.5272
0.02192.0740000.6522-2.9890-6.01640.82813.0274-317.1334-284.1248-2.4105-2.4545
0.01192.1241000.6922-3.4777-6.67490.79693.1972-323.7187-289.0121-2.4272-2.4699
0.01532.1742000.6993-3.2406-6.67750.79693.4369-323.7453-286.6413-2.4047-2.4465
0.0112.2243000.7178-3.7991-7.43970.76563.6406-331.3667-292.2260-2.3843-2.4290
0.00722.2744000.6840-3.3269-6.80210.81253.4752-324.9908-287.5042-2.4095-2.4536
0.01972.3245000.7013-3.6890-7.30140.81253.6124-329.9841-291.1250-2.4118-2.4543
0.01822.3746000.7476-3.8994-7.53660.82813.6372-332.3356-293.2291-2.4163-2.4565
0.01252.4347000.7199-4.0560-7.57650.84383.5204-332.7345-294.7952-2.3699-2.4100
0.00822.4848000.7048-3.6613-7.13560.8753.4743-328.3255-290.8477-2.3925-2.4303
0.01182.5349000.6976-3.7908-7.31520.81253.5244-330.1224-292.1431-2.3633-2.4047
0.01182.5850000.7198-3.9049-7.55570.82813.6508-332.5271-293.2844-2.3764-2.4194
0.0062.6351000.7506-4.2118-7.91490.81253.7032-336.1194-296.3530-2.3407-2.3860
0.01432.6852000.7408-4.2433-7.98020.81253.7369-336.7721-296.6682-2.3509-2.3946
0.00572.7453000.7552-4.3392-8.08310.79693.7439-337.8013-297.6275-2.3388-2.3842
0.01382.7954000.7404-4.2395-7.97620.81253.7367-336.7322-296.6304-2.3286-2.3737
0.00792.8455000.7525-4.4466-8.21960.78123.7731-339.1662-298.7007-2.3200-2.3641
0.00772.8956000.7520-4.5586-8.34850.79693.7899-340.4545-299.8206-2.3078-2.3517
0.00942.9457000.7527-4.5542-8.35090.78123.7967-340.4790-299.7773-2.3062-2.3510
0.00542.9958000.7520-4.5169-8.30790.78123.7911-340.0493-299.4038-2.3081-2.3530

Framework versions

  • Transformers 4.35.0.dev0
  • Pytorch 2.0.1+cu118
  • Datasets 2.12.0
  • Tokenizers 0.14.0

Citation

If you find Zephyr-7B-β is useful in your work, please cite it with:

INI
@misc{tunstall2023zephyr,
      title={Zephyr: Direct Distillation of LM Alignment}, 
      author={Lewis Tunstall and Edward Beeching and Nathan Lambert and Nazneen Rajani and Kashif Rasul and Younes Belkada and Shengyi Huang and Leandro von Werra and Clémentine Fourrier and Nathan Habib and Nathan Sarrazin and Omar Sanseviero and Alexander M. Rush and Thomas Wolf},
      year={2023},
      eprint={2310.16944},
      archivePrefix={arXiv},
      primaryClass={cs.LG}
}

If you use the UltraChat or UltraFeedback datasets, please cite the original works:

INI
@misc{ding2023enhancing,
      title={Enhancing Chat Language Models by Scaling High-quality Instructional Conversations}, 
      author={Ning Ding and Yulin Chen and Bokai Xu and Yujia Qin and Zhi Zheng and Shengding Hu and Zhiyuan Liu and Maosong Sun and Bowen Zhou},
      year={2023},
      eprint={2305.14233},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

@misc{cui2023ultrafeedback,
      title={UltraFeedback: Boosting Language Models with High-quality Feedback}, 
      author={Ganqu Cui and Lifan Yuan and Ning Ding and Guanming Yao and Wei Zhu and Yuan Ni and Guotong Xie and Zhiyuan Liu and Maosong Sun},
      year={2023},
      eprint={2310.01377},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

MetricValue
Avg.52.15
ARC (25-shot)62.03
HellaSwag (10-shot)84.36
MMLU (5-shot)61.07
TruthfulQA (0-shot)57.45
Winogrande (5-shot)77.74
GSM8K (5-shot)12.74
DROP (3-shot)9.66
Liens & Ressources
Spécifications
CatégorieChat
AccèsAPI & Local
LicenceOpen Source
TarificationOpen Source
Paramètres7B parameters
Note
4.1

Essayer zephyr 7b beta

Accédez directement au modèle