yahma commited on
Commit
175a00b
1 Parent(s): e54b448

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -0
README.md CHANGED
@@ -1,3 +1,31 @@
1
  ---
2
  license: mit
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ datasets:
4
+ - yahma/alpaca-cleaned
5
  ---
6
+ This repo contains a low-rank adapter for LLaMA-13b fit on the Cleaned Alpaca dataset.
7
+
8
+ This version of the weights was trained with the following hyperparameters:
9
+
10
+ Cleaned dataset: Snapshot April 2, 2023
11
+ Epochs: 3
12
+ Validation set size: 2000
13
+ Batch size: 128
14
+ Micro batch size: 8
15
+ Cutoff length: 512
16
+ Learning rate: 3e-4
17
+ Lora r: 16
18
+ Lora target modules: q_proj, k_proj, v_proj, o_proj
19
+
20
+ That is:
21
+
22
+ python finetune.py \
23
+ --base_model='decapoda-research/llama-13b-hf' \
24
+ --data_path 'yahma/alpaca-cleaned' \
25
+ --num_epochs=3 \
26
+ --cutoff_len=512 \
27
+ --output_dir='./lora-alpaca' \
28
+ --lora_target_modules='[q_proj,k_proj, v_proj, o_proj]' \
29
+ --lora_r=16 \
30
+ --val_set_size 2000 \
31
+ --micro_batch_size=8