KrisPi commited on
Commit
305c966
1 Parent(s): c888f6b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -10
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  license: openrail
3
  ---
4
- **This model is merge between 66% of Wizard Coder and 33% of Redmond Hermes Coder (which is Wizard Coder fine-tune):**
5
 
6
  https://huggingface.co/NousResearch/Redmond-Hermes-Coder
7
  https://huggingface.co/WizardLM/WizardCoder-15B-V1.0
@@ -10,18 +10,23 @@ Merger done by the most basic value average.
10
 
11
  Using CTranslate2 for quantization and inference achieving as much as 37 tokens /s on RTX 3090 GPU.
12
 
13
- Inference done by using text-generation-webui:
14
 
15
- Added this code and ran update on requirements.txt: https://github.com/oobabooga/text-generation-webui/pull/2828
16
 
17
  There is one thing extra to be changed in the code: reply = apply_extensions('output', reply) to: reply = apply_extensions('output', reply, state)
18
 
19
- The idea was to get some of the coding abilities back that were lost in fine-tune, but retain at least basic capabilities to summarize text and work with context. This experiment was also focused on using CT2 for its speed. **I believe presented approach is the best available compromise between speed, coding accuracy and a little of general LLM use. ~Please note that CT2 8bit quant seems to have better HumanEval scores than load-in-8bit~**
20
 
21
- Community now mostly focus on making non-coding models - code as making coding models be more general seems near impossible.
22
- However, my daily use is focused around DevOps questions, summarizing content and script development. Further development will be around intent analysis for integration with TODO lists and calendar extracting actions and notes from my voice transcription. This model doesn't seem to work well enough on those tasks so next time will attempt actual fine-tunes of Wizard Coder or just running two models at the same time. I hope to fit under 24GB VRAM which would mean I will also evaluate 4 bit quantization.
23
 
24
- My initial testing was checking if model finds:
 
 
 
 
 
 
25
 
26
  Overflow: `"what is mistake in following C++ code: int a = 1e9+7; int b = 1e9+9; int c = a*b; cout << c;"`
27
 
@@ -30,15 +35,26 @@ Out of bounds: `"what is bug in the following C++ code: int a = 100; vector <int
30
  and propose using "docker update" for `"how to stop docker container so it doesnt start every reboot"`
31
 
32
 
 
33
  I have run those prompts in the loop, with different presets and ended up picking this preset:
34
  `['temperature'] = 1.31`
35
  `['top_p'] = 0.29`
36
  `['top_k'] = 72`
37
  `['repetition_penalty'] = 1.09`
 
38
 
39
- Testing of the above prompts has shown that Hermes Coder CT2 was not able to answer correctly most of the time while Wizard Coder and this merge did. Merged model seems to retain ability to use "### Input:" in the prompt and became more sensitive to non-coding instruction. (Wizard Coder almost completely disregard it)
 
 
 
 
 
 
 
 
 
 
40
 
41
- In the bottom you can see EvalPlus benchmarks of three mentioned models - seems they all performed in similar way with default preset. I'm not sure if I'm not doing benchmark right or those quants are not working properly. As I noticed myself custom preset improved the result. I will rerun benchmarks as the last result seems to hint some randomness.
42
 
43
  **For summarization I propose following prompt:**
44
 
@@ -131,7 +147,7 @@ and
131
  `{'pass@1': 0.3719512195121951}`
132
 
133
  --------------
134
-
135
 
136
 
137
 
 
1
  ---
2
  license: openrail
3
  ---
4
+ **This model is a merge between 66% of Wizard Coder and 33% of Redmond Hermes Coder (which is Wizard Coder fine-tune):**
5
 
6
  https://huggingface.co/NousResearch/Redmond-Hermes-Coder
7
  https://huggingface.co/WizardLM/WizardCoder-15B-V1.0
 
10
 
11
  Using CTranslate2 for quantization and inference achieving as much as 37 tokens /s on RTX 3090 GPU.
12
 
13
+ Inference is done by using text-generation-webui:
14
 
15
+ Added this code and ran an update on requirements.txt: https://github.com/oobabooga/text-generation-webui/pull/2828
16
 
17
  There is one thing extra to be changed in the code: reply = apply_extensions('output', reply) to: reply = apply_extensions('output', reply, state)
18
 
19
+ The idea was to get some of the coding abilities back that were lost in fine-tune but retain at least basic capabilities to summarize text and work with context. This experiment was also focused on using CT2 for its speed.
20
 
21
+ **I believe the presented approach is the best available compromise between speed, coding accuracy, and a little of general LLM use.**
 
22
 
23
+ **Please note that CT2 8bit quant seems to have better HumanEval scores than load-in-8bit**
24
+
25
+
26
+ The community now mostly focuses on making non-coding models - code as making coding models be more general seems near impossible.
27
+ However, my daily use is focused on DevOps questions, summarizing content, and script development. Further development will be around intent analysis for integration with TODO lists and calendar extracting actions and notes from my voice transcription. This model doesn't seem to work well enough on those tasks so next time will attempt actual fine-tunes of Wizard Coder or just run two models at the same time. I hope to fit under 24GB VRAM which would mean I will also evaluate 4 bit quantization.
28
+
29
+ My initial testing was checking if the model finds:
30
 
31
  Overflow: `"what is mistake in following C++ code: int a = 1e9+7; int b = 1e9+9; int c = a*b; cout << c;"`
32
 
 
35
  and propose using "docker update" for `"how to stop docker container so it doesnt start every reboot"`
36
 
37
 
38
+
39
  I have run those prompts in the loop, with different presets and ended up picking this preset:
40
  `['temperature'] = 1.31`
41
  `['top_p'] = 0.29`
42
  `['top_k'] = 72`
43
  `['repetition_penalty'] = 1.09`
44
+
45
 
46
+ Testing of the above prompts has shown that Hermes Coder CT2 was not able to answer correctly most of the time while Wizard Coder and this merge did. The merged model seems to retain the ability to use "### Input:" in the prompt and became more sensitive to non-coding instruction. (Wizard Coder almost completely disregards it)
47
+
48
+ In the bottom you can see EvalPlus benchmarks of three mentioned models - seems they all performed in a similar way with the default preset. I'm not sure if I'm not doing the benchmark right or if those quants are not working properly with default preset. As I noticed custom preset considerably improved the result.
49
+
50
+ **I would greatly appreciate if anyone can confirm how good this model is with proposed preset as the result I got really positively suprised me.(seems better than any other Wizard Coder 8bit quant**
51
+
52
+ **CT2 int8_float16 merge, custom preset:**
53
+ `Base`
54
+ `{'pass@1': 0.47560975609756095}`
55
+ `Base + Extra`
56
+ `{'pass@1': 0.45121951219512196}`
57
 
 
58
 
59
  **For summarization I propose following prompt:**
60
 
 
147
  `{'pass@1': 0.3719512195121951}`
148
 
149
  --------------
150
+
151
 
152
 
153