how to convert "Phi-3-mini-4k-instruct" to "Phi-3-mini-4k-instruct-onnx"?

#9
by tianheli - opened

It looks like "Phi-3-mini-4k-instruct-onnx" doesn't support fine-turning, so is it possible to fine-turn "Phi-3-mini-4k-instruct" then convert it to ONNX format?
If yes , is there a guide on how to convert ?

Microsoft org

You can use ONNX Runtime GenAI's model builder to quickly convert your fine-tuned Phi-3-mini-4k-instruct model to optimized and quantized ONNX models. This example should work for your scenario.

I fine-tuned ‘Phi-3-mini-4k-instruct’ (using LoRa) and it works pretty well. However, when quantizing it with the ONNX runtime GenAI's model builder (I used the following command python3 -m onnxruntime_genai.models.builder -i path_to_local_folder_on_disk -o path_to_output_folder -p int4 -e cpu -c cache_dir_to_store_temp_files and inference it, the output of the quanitized model is complete nonsense. Does anyone else have this problem? Or a tip on how to tackle the problem?

Microsoft org

The model builder does not currently support LoRA but support will be coming soon.

kvaishnavi changed discussion status to closed

Sign up or log in to comment