bhenrym14 commited on
Commit
444fb74
1 Parent(s): b6bbdb4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -5
README.md CHANGED
@@ -1,3 +1,4 @@
 
1
  ## Overview
2
 
3
  This is [Jon Durbin's Airoboros 33B GPT4 1.4](https://huggingface.co/jondurbin/airoboros-33b-gpt4-1.4) (with GPTQ Quantization) with several key modifications:
@@ -7,20 +8,25 @@ This is [Jon Durbin's Airoboros 33B GPT4 1.4](https://huggingface.co/jondurbin/a
7
  Otherwise, I emulated the training process as closely as possible. It was trained on 1x RTX 6000 Ada for ~43 hours.
8
 
9
  ## Motivation
10
- Recent advancements in extending context by RoPE scaling ([kaiokendev](https://kaiokendev.github.io/til#extending-context-to-8k) and [(meta AI)](https://arxiv.org/abs/2306.15595)) demonstrate the ability to extend the context window without (total) retraining. Finetuning has shown to be necessary to properly leverage the longer context. The superHOT LoRA is a finetuned adapter that has been finetuned on longer context (8192 tokens); even when applied to dissimilar models, it successfully extends the contexts window to which the model can attend. While impressive this adapter is so flexible, how much does performance suffer relative to a model that has been finetuned with the scaled embeddings from the start? This is an experiment to explore this.
11
 
12
  ## Relative Performance (perplexity)
 
 
 
 
 
 
13
 
14
-
15
 
16
  ## Quantization:
17
 
18
- The merged model was quantized with AutoGPTQ (bits = 4, group_size = 128, desc_act = True)
19
-
20
 
21
 
22
 
23
- ## Original model card: Jon Durbin's Airoboros 33B GPT4 1.4
24
 
25
 
26
  __not yet tested!__
 
1
+ # RoPE Scaled Finetune of airoboros-33b-gpt4-1.4.1 (GPTQ)
2
  ## Overview
3
 
4
  This is [Jon Durbin's Airoboros 33B GPT4 1.4](https://huggingface.co/jondurbin/airoboros-33b-gpt4-1.4) (with GPTQ Quantization) with several key modifications:
 
8
  Otherwise, I emulated the training process as closely as possible. It was trained on 1x RTX 6000 Ada for ~43 hours.
9
 
10
  ## Motivation
11
+ Recent advancements in extending context by RoPE scaling ([kaiokendev](https://kaiokendev.github.io/til#extending-context-to-8k) and [meta AI)](https://arxiv.org/abs/2306.15595)) demonstrate the ability to extend the context window without (total) retraining. Finetuning has shown to be necessary to properly leverage the longer context. The superHOT LoRA is a finetuned adapter that has been finetuned on longer context (8192 tokens); even when applied to dissimilar models, it successfully extends the contexts window to which the model can attend. While impressive this adapter is so flexible, how much does performance suffer relative to a model that has been finetuned with the scaled embeddings from the start? This is an experiment to explore this.
12
 
13
  ## Relative Performance (perplexity)
14
+ | Model | Context | Perplexity |
15
+ | ---------------------------------------------------- | ----------- | ---------- |
16
+ | TheBloke/airoboros-33B-gpt4-1-4-SuperHOT-8K-GPTQ | 2048 | 5.15 |
17
+ | TheBloke/airoboros-33B-gpt4-1-4-SuperHOT-8K-GPTQ | 8192 | 5.04 |
18
+ | **bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-GPTQ** | **2048** | **4.32** |
19
+ | **bhenrym14/airoboros-33b-gpt4-1.4.1-PI-8192-GPTQ** | **3072** | **4.26** |
20
 
21
+ How does this reduction in perplexity translate into actual performance lift on downstream tasks? I'm not sure yet.
22
 
23
  ## Quantization:
24
 
25
+ The merged model was quantized with AutoGPTQ (bits = 4, group_size = 128, desc_act = True). If there's interest, I can upload the LoRA weights and/or merged 16bit HF model.
 
26
 
27
 
28
 
29
+ # Original model card: Jon Durbin's Airoboros 33B GPT4 1.4
30
 
31
 
32
  __not yet tested!__