Text Generation
MLX
English
mistral
zephyr
zephyr-7b-beta / README.md
alvarobartt's picture
alvarobartt HF staff
Update README.md
02bc678
|
raw
history blame
1.55 kB
metadata
pipeline_tag: text-generation
inference: false
tags:
  - zephyr
  - mlx
language:
  - en
license: mit
library_name: mlx

Zephyr 7B β

Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr-7B-β is the second model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1 that was trained on on a mix of publicly available, synthetic datasets using Direct Preference Optimization (DPO). We found that removing the in-built alignment of these datasets boosted performance on MT Bench and made the model more helpful. However, this means that model is likely to generate problematic text when prompted to do so. You can find more details in the technical report.

This repository contains the zephyr-7b-beta weights in npz format suitable for use with Apple's MLX framework.

Use with MLX

pip install mlx
pip install huggingface_hub hf_transfer
git clone https://github.com/ml-explore/mlx-examples.git
cd mlx-examples

# Download model
export HF_HUB_ENABLE_HF_TRANSFER=1
huggingface-cli download --local-dir-use-symlinks False --local-dir zephyr-7b-beta alvarobartt/zephyr-7b-beta

# Run example
python llms/mistral/mistral.py --model_path zephyr-7b-beta --prompt "My name is"

Please, refer to the original model card for more details on Zephyr 7B β.