Can't load in ollama
[root@ubuntu]$ ollama run magnum-v2:12b-fp16
Error: exception error loading model hyperparameters: iinvalid n_rot: 160, expected 128
Can someone explain why this is happening?
It is the first time models do not run on my ollama v0.3.5
The gguf quantized versions of these two models can run properly.
We can't provide support for third party engines, but I'd try updating to v0.3.6 in hopes they merged / fixed nemo inferencing, if that still doesn't work I'd open an issue on their github: https://github.com/ollama/ollama/issues hope that helps
Do you have any recommended engines that can run on Ubuntu?
The models your team has developed are really impressive.
I'll try loading fp16 with tabby, as I've been wanting to experience it. I’ve tried other quantized models and they’ve been great. Thank you for your efforts and sharing!