DPOpenHermes-11B / README.md
winglian's picture
Create README.md
6b78354
metadata
license: apache-2.0
datasets:
  - teknium/openhermes
  - argilla/ultrafeedback-binarized-preferences
  - Intel/orca_dpo_pairs
language:
  - en
library_name: transformers

DPOpenHermes 11B

This is a mergekit merge of DPOpenHermes-7B from seperate versions of it.

slices:
  - sources:
    - model: openaccess-ai-collective/DPOpenHermes-7B
      revision: dpo-v0
      layer_range: [0, 24]
  - sources:
    - model: openaccess-ai-collective/DPOpenHermes-7B
      layer_range: [8, 32]
merge_method: passthrough
dtype: bfloat16