File size: 2,269 Bytes
2b63d8f
 
 
 
 
 
 
 
 
 
 
276d1c0
2b63d8f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
---
license: apache-2.0
base_model:
- chihoonlee10/T3Q-Mistral-Orca-Math-DPO
- nbeerbower/bruphin-kappa
library_name: transformers
tags:
- mergekit
- merge

---
![image/png](https://cdn-lfs-us-1.huggingface.co/repos/2b/ed/2bed4d865015eea76d66e4abb9ecb64d6eec6bbbb05d183eaadee021dcf35b53/8e4f6b05ab037981807f86da66ec1a54ce2cda0dab6c57addcbffec2b7e3bf75?response-content-disposition=inline%3B+filename*%3DUTF-8%27%27bruphin.png%3B+filename%3D%22bruphin.png%22%3B&response-content-type=image%2Fpng&Expires=1712112316&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTcxMjExMjMxNn19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2RuLWxmcy11cy0xLmh1Z2dpbmdmYWNlLmNvL3JlcG9zLzJiL2VkLzJiZWQ0ZDg2NTAxNWVlYTc2ZDY2ZTRhYmI5ZWNiNjRkNmVlYzZiYmJiMDVkMTgzZWFhZGVlMDIxZGNmMzViNTMvOGU0ZjZiMDVhYjAzNzk4MTgwN2Y4NmRhNjZlYzFhNTRjZTJjZGEwZGFiNmM1N2FkZGNiZmZlYzJiN2UzYmY3NT9yZXNwb25zZS1jb250ZW50LWRpc3Bvc2l0aW9uPSomcmVzcG9uc2UtY29udGVudC10eXBlPSoifV19&Signature=g7fhO-yfpN9sqZx--XTySUK%7EeWP48iAs4Ec5tm5%7Es2j78A5-sFA8IqmgFm9Ndv6UGu2tCDWNSJ-qx7lQ-5PadWZn-F%7E420lLr0-KzJ-ERjiAFpNXkFjxpjH4h3AvHqJ4GrL9K85PdCttaq4-KaKNEmtJOVfVSKLAmp4Z2uBpIMvcFyoh1b7l62LR9NeK4cRA9m9fdoEkZR-RxwhXl43gpTOOCpKE0wMgAGDNJdyppDd86s7lReV6eH2qFNGJyNAfvWaO-eMsGc5Jv-AJ9srt6p6-8%7EqqUxfl9pby-2yzbFqFOtxLbPVWWb-v-Uvg6PJ2CIQs2CeouOF0uwgT3W-oXA__&Key-Pair-Id=KCD77M1F0VK2B)
# bruphin-lambda

This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).

## Merge Details
### Merge Method

This model was merged using the SLERP merge method.

### Models Merged

The following models were included in the merge:
* [chihoonlee10/T3Q-Mistral-Orca-Math-DPO](https://huggingface.co/chihoonlee10/T3Q-Mistral-Orca-Math-DPO)
* [nbeerbower/bruphin-kappa](https://huggingface.co/nbeerbower/bruphin-kappa)

### Configuration

The following YAML configuration was used to produce this model:

```yaml
models:
  - model: chihoonlee10/T3Q-Mistral-Orca-Math-DPO
    layer_range: [0, 32]
  - model: nbeerbower/bruphin-kappa
    layer_range: [0, 32]
merge_method: slerp
base_model: nbeerbower/bruphin-kappa
parameters:
  t:
    - filter: self_attn
      value: [0, 0.5, 0.3, 0.7, 1]
    - filter: mlp
      value: [1, 0.5, 0.7, 0.3, 0]
    - value: 0.5
dtype: bfloat16

```