Which RAG embedding model to use with sapienzanlp/Minerva-7B-instruct-v1.0-GGUF

#2
by micrem73 - opened

I deployed sapienzanlp/Minerva-7B-instruct-v1.0-GGUF on a virtual machine. It works, and the answers to my prompts sound fine.

As soon as I trigger the RAG system by uploading a document through the webui function, or add some knowledge, the model stops answering in a consistent way. Sometimes it simply replicates my question. Sometimes it answers by including the RAG prompt.

I assume I should set a different embedding model - the default one, all-MiniLM-L6-v2 seems optimised just for english.

Do you have any suggestion on a right model to choose that allow RAG with Minerva7b, or any other parameter I should take care ok?

Thanks

Sign up or log in to comment