Kensho: a luminous awakening where the veil of illusion dissolves, revealing the boundless truth of our interconnected essence, inviting us into a dance with the infinite.
By Fernando, Eric and David
Discord: https://discord.gg/cognitivecomputations
This is a hack around pytorch + huggingface Transformers library to make the original Dolphin Phi-2 to behave in a way inspired by the Meta's paper "MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases" [ https://arxiv.org/abs/2402.14905 ]
One of the key ideas is that it works as if it was like "an online passthrough", by applying a loop on a module SuperClass, that groups layers, in a such way they get their forward method repeated in a loop. So, in theory, you can observe more intelligence in the same way MegaDolphin 120b, Professor 155b, Venus120b and other huge models, but use way less vRAM, because instead of cloning the weights, we share them in the vRAM.
And actually, this concept could be also used to enable the training of way more efficient models.
We hope the community enjoy it and make good use of it.
It won't work out of the box in the other models. Their "modeling files" should be changed accordingly to achieve the same effect.
- Downloads last month
- 15