Update README.md
Browse files
README.md
CHANGED
@@ -22,7 +22,24 @@ This is a narrow domain-expert LLM trained on the top 5 books on Gutenberg:
|
|
22 |
- The Prince (Machiavelli)
|
23 |
- Second Treatise of Government
|
24 |
|
25 |
-
It's meant to be an interesting novelty, showing off training on a specific domain.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
|
27 |
## Training procedure
|
28 |
|
|
|
22 |
- The Prince (Machiavelli)
|
23 |
- Second Treatise of Government
|
24 |
|
25 |
+
It's meant to be an interesting novelty, showing off training on a specific domain. It has some quirks. Namely:
|
26 |
+
|
27 |
+
1. It seems to have memorized the training data very well. Ask a question that exists in the training data, with temp 0, and it will usually give you back the exact response word-for-word. This means that, on the subjects covered by its data, it will be very knowledgeable.
|
28 |
+
2. I forgot to include any generalist instruct data, so it's... not stupid, at least not particularly stupid by 7b standards, but it is very much limited to QA.
|
29 |
+
3. It's much less fluffy and wasteful with its responses than previous Augmentoolkit domain expert models, due to using a new dataset setting. This tends to make it respond with less detail, but it also may remember stuff better and get to the point easier.
|
30 |
+
|
31 |
+
Some example chats (blame LM studio for not hiding the stop token):
|
32 |
+
|
33 |
+
Asking stuff from the training data:
|
34 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/AirHFo61iB1HAP-IXwnZn.png)
|
35 |
+
|
36 |
+
Asking a question directly from the training data and
|
37 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/Ccm-EeDyOFcylCefwDS-W.png)
|
38 |
+
|
39 |
+
Some things that are kinda funny but also show off the drawback of not using any generalist data:
|
40 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/e2sCBLIX8Xg91KSevGt_B.png))
|
41 |
+
|
42 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64825ebceb4befee377cf8ac/P0bhWyENxOaxPvC4fE6jw.png)
|
43 |
|
44 |
## Training procedure
|
45 |
|