just for curiosity
#1
by
prudant
- opened
how much time took the final training process?
thanks!
can you share the command line and config files ?
why is this taking so little resources, compared to accelerate/fsdp_config.yaml ?
I am doing AlexWortega/miqu-1-70b-AQLM-2Bit-1x16-hf
(sorry i am a beginner)
can I convert mixtral 8x22 to AQLM and then train using this method on 2x3090?
@bittamer i think AQLM quant process require a lot of gpu computational power (more than 4 gpus for a couple of days running)
@bittamer I think FSDP+QLoRA should be more suitable for your case