ChristianAzinn
commited on
Commit
•
caced19
1
Parent(s):
187767f
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,16 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
---
|
4 |
+
|
5 |
+
# TinyJSON
|
6 |
+
|
7 |
+
Trained on [my `json-training` dataset](https://huggingface.co/datasets/ChristianAzinn/json-training),
|
8 |
+
these are finetunes of the smallest state-of-the-art LLMs to output in structured JSON.
|
9 |
+
|
10 |
+
Where their base/instruct versions have so little clue how to output JSON that forcing it using
|
11 |
+
techniques like grammars simply hangs forever, these little guys (mostly) work like a charm.
|
12 |
+
(SmolLM 135M still sometimes babbles on. Set a maximum token limit.)
|
13 |
+
|
14 |
+
Training was done with Unsloth at 4bit (lmao), rank=8, alpha=8, for 3 epochs each.
|
15 |
+
`rev1` models were trained on the first revision (11.6k rows) of `json-training`,
|
16 |
+
while `rev2` models were trained on the second (20.6k rows).
|