May I ask how the your method of merge adapters to base model?

#5
by HCZhang - opened

Sorry to bother you.
This question may not be directly related to your model, but I've been looking around and yet to find a solution.

I've fine-tuned a model using QLORA, and I can't merge the adapters (checkpoint) back to the base model.
I've tried the script provided by The Block, but there were some errors showing layer sizes don't match.

And your model seems works fine, so I wonder how did you merge the model.
Thank you.

Also, your model is really impressive

Hi,
I tried two ways for fusion:

I tried the two methods, generally, the second will get better ARC(+0.15) and Truthful_QA(+0.3) scores but the other two(MMLU(-0.2) and HelloSwag(-0.2)) seems to degenerate.

The version for leadboard is generated by the first fusion method.

Thank you so much, would check the two methods you mentioned.

fangloveskari changed discussion status to closed

Sign up or log in to comment