File size: 2,903 Bytes
8c6bb08
 
 
4a1d0f3
8c6bb08
 
4a1d0f3
 
 
 
 
 
 
 
 
afd21ec
9b542a2
a719e4c
4a1d0f3
a719e4c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c194ced
a719e4c
 
 
 
 
 
 
 
 
 
 
4a1d0f3
 
 
 
 
 
 
 
 
 
 
 
a719e4c
 
 
 
 
 
 
 
 
 
 
 
c2b350f
a719e4c
 
 
 
 
 
 
 
 
 
 
c194ced
a719e4c
 
 
 
 
 
 
 
 
 
 
4a1d0f3
a719e4c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8c6bb08
9bb3a31
 
4a1d0f3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
---
license: mit
base_model:
  - meta-llama/Llama-2-13b
pipeline_tag: text-generation
tags:
  - chemistry
  - biology
  - finance
  - legal
  - music
  - art
  - code
  - climate
  - medical
  - quantized
library_name: transformers
metrics:
  - perplexity

model-index:
  - name: LLaMa2-13b-merged-clusters
    results:
      - task:
          type: commonsense-reasoning
        dataset:
          name: HellaSwag
          type: HellaSwag
        metrics:
          - name: Accuracy
            type: accuracy
            value: 66.66
        source:
          name: OpenCompass
          url: https://opencompass.org
      - task:
          type: commonsense-reasoning
        dataset:
          name: PIQA
          type: PIQA
        metrics:
          - name: Accuracy
            type: accuracy
            value: 72.14
        source:
          name: OpenCompass
          url: https://opencompass.org
      - task:
          type: coreference-resolution
        dataset:
          name: WSC
          type: WSC
        metrics:
          - name: Perplexity
            type: perplexity
            value: 60.38
        source:
          name: OpenCompass
          url: https://opencompass.org
      - task:
          type: coreference-resolution
        dataset:
          name: WSC
          type: WSC
        metrics:
          - name: Accuracy
            type: accuracy
            value: 38.99
        source:
          name: OpenCompass
          url: https://opencompass.org
      - task:
          type: multiple-choice-question-answering
        dataset:
          name: CSQA
          type: CSQA
        metrics:
          - name: Accuracy
            type: accuracy
            value: 54.36
        source:
          name: OpenCompass
          url: https://opencompass.org
      - task:
          type: multi-task-evaluation
        dataset:
          name: MMLU
          type: MMLU
        metrics:
          - name: Accuracy
            type: accuracy
            value: 54.76
        source:
          name: OpenCompass
          url: https://opencompass.org
      - task:
          type: multiple-choice-question-answering
        dataset:
          name: RACE
          type: RACE-high
        metrics:
          - name: Accuracy
            type: accuracy
            value: 53.89
        source:
          name: OpenCompass
          url: https://opencompass.org
      - task:
          type: multiple-choice-question-answering
        dataset:
          name: RACE
          type: RACE-middle
        metrics:
          - name: Accuracy
            type: accuracy
            value: 55.29
        source:
          name: OpenCompass
          url: https://opencompass.org

---
# Merged LLaMA Model

This is a merged version of the LLaMA2-13b model based on hyperboloid projections. The model retains 31 layers with significant performance retention across all benchmarks.