Jaward's picture
Update README.md
9b77a29 verified
---
license: mit
---
## Finetuned Model For My Thesis: Design And Implementation Of An Adaptive Virtual Intelligent Teaching Assistant Based On Supervised Fine-tuning Of A Pre-trained Large Language Model
### Model Name: CodeOptimus - Adaptive Supervised Instruction Fine-tuning [Mistral 7B Instruct](https://mistral.ai/news/announcing-mistral-7b/) using qLora.
## Prerequisites For Reproduction
1. **GPU**: Requires powerful GPUs - I used 7 Nvidia A100s.
2. **Train Time**: 1 week.
3. **RAG Module**: Updates the knowledge base of the model in real-time with adaptive features learned from conversations with the model over time..
4. **Python Packages**: Install requirements.txt.
5. **Dataset**: Download [code_instructions_122k_alpaca_style](https://huggingface.co/datasets/TokenBender/code_instructions_122k_alpaca_style) plus some custom curated dataset
6. **Mistra-7B-Instruct-v0.1**: Download [mistralai/Mistral-7B-Instruct-v0.1 ](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) pytorch bin weights
7. **Realistic 3D Intelligent Persona/Avatar (Optional)**: For this I'm using soulmachine's digital humans.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6438a9027de34e8ea7e4b257/UJtAiKejhrmUPN5EiA59E.png)