Can you make a model that can run without quantization on a gpu with only 8g vram?

#2
by win10 - opened

Can you make a model that can run without quantization on a gpu with only 8g vram?

maywell changed discussion status to closed

Sign up or log in to comment