Update README.md
Browse files
README.md
CHANGED
@@ -20,14 +20,15 @@ RWKV-4 14B is a L40-D5120 causal language model trained on the Pile. See https:/
|
|
20 |
|
21 |
Use https://github.com/BlinkDL/ChatRWKV to run it.
|
22 |
|
23 |
-
ctx_len = 1024
|
24 |
n_layer = 40
|
25 |
n_embd = 5120
|
26 |
|
27 |
RWKV-4-Pile-14B-2023xxxx-ctx4096-testxxx.pth : Fine-tuned to ctx_len 4096.
|
|
|
28 |
* Highly recommended. It's great.
|
29 |
|
30 |
RWKV-4-Pile-14B-20230213-8019.pth : Trained on the Pile for 331B tokens.
|
|
|
31 |
* Pile loss 1.7579
|
32 |
* LAMBADA ppl 3.81, acc 71.05%
|
33 |
* PIQA acc 77.42%
|
|
|
20 |
|
21 |
Use https://github.com/BlinkDL/ChatRWKV to run it.
|
22 |
|
|
|
23 |
n_layer = 40
|
24 |
n_embd = 5120
|
25 |
|
26 |
RWKV-4-Pile-14B-2023xxxx-ctx4096-testxxx.pth : Fine-tuned to ctx_len 4096.
|
27 |
+
* ctx_len = 4096
|
28 |
* Highly recommended. It's great.
|
29 |
|
30 |
RWKV-4-Pile-14B-20230213-8019.pth : Trained on the Pile for 331B tokens.
|
31 |
+
* ctx_len = 1024
|
32 |
* Pile loss 1.7579
|
33 |
* LAMBADA ppl 3.81, acc 71.05%
|
34 |
* PIQA acc 77.42%
|