future research
#2
by
NafishZaldinanda
- opened
Will it be available for Q4_K_M in the future and can it be run via llama-cpp-python and provide mmproject for chathandler in llama-cpp-python?
Will it be available for Q4_K_M in the future and can it be run via llama-cpp-python and provide mmproject for chathandler in llama-cpp-python?