SicariusSicariiStuff
commited on
Commit
•
e828abb
1
Parent(s):
059cfa0
Update README.md
Browse files
README.md
CHANGED
@@ -22,6 +22,10 @@ As of **June 13, 2024**, I've observed that even after two days of continuous tr
|
|
22 |
|
23 |
These observations underscore the critical importance of fine-tuning for alignment. Given the current pace, training will likely extend beyond a week. However, the end result should be **interesting**. If the additional datasets focused on logic and common sense are effective, we should achieve a model that is **nearly completely unaligned**, while still retaining its core 'intelligence.'
|
24 |
|
|
|
|
|
|
|
|
|
25 |
<img src="https://i.imgur.com/b6unKyS.png" alt="LLAMA-3_Unaligned_Training" style="width: 60%; min-width: 600px; display: block; margin: auto;">
|
26 |
|
27 |
|
|
|
22 |
|
23 |
These observations underscore the critical importance of fine-tuning for alignment. Given the current pace, training will likely extend beyond a week. However, the end result should be **interesting**. If the additional datasets focused on logic and common sense are effective, we should achieve a model that is **nearly completely unaligned**, while still retaining its core 'intelligence.'
|
24 |
|
25 |
+
**June 18, 2024 Update**, After extensive testing of the intermediate checkpoints, significant progress has been made. The model is slowly — I mean, really slowly — unlearning its alignment. By significantly lowering the learning rate, I was able to visibly observe deep behavioral changes, this process is taking longer than anticipated, but it's going to be worth it. Estimated time to completion: 4 more days.. I'm pleased to report that in several tests, the model not only maintained its intelligence but actually showed a slight improvement, especially in terms of common sense. An intermediate checkpoint of this model was used to create [invisietch/EtherealRainbow-v0.3-rc7](https://huggingface.co/invisietch/EtherealRainbow-v0.3-rc7-8B-GGUF), with promising results. Currently, it seems like I'm on the right track. I hope this model will serve as a solid foundation for further merges, whether for role-playing (RP) or for uncensoring. This approach also allows us to save on actual fine-tuning, thereby reducing our carbon footprint. The merge process takes just a few minutes of CPU time, instead of days of GPU work.
|
26 |
+
|
27 |
+
Cheers,
|
28 |
+
Sicarius
|
29 |
<img src="https://i.imgur.com/b6unKyS.png" alt="LLAMA-3_Unaligned_Training" style="width: 60%; min-width: 600px; display: block; margin: auto;">
|
30 |
|
31 |
|