Would like to add a quantitative model of GPTQ-IN8
#1
by
warlock-edward
- opened
Since my GPU is a 7th generation architecture and doesn't support AWQ model, expect a quantization model for int8 with GPTQ, thanks a lot!