Model Card for Model ID

A parameter-efficient finetune (using LoRA) of DeepSeek Coder 1.3B finetuned on Python code.

Model Details

Model Description

A finetune of DeepSeek Coder 1.3B finetuned on 1000 examples of Python code from the ArtifactAI/arxiv_python_research_code dataset.

  • Model type: Text Generation
  • Language(s) (NLP): English, Python
  • Finetuned from model: deepseek-ai/deepseek-coder-1.3b-base

Model Sources [optional]

Uses

To generate Python code

How to Get Started with the Model

Use the code below to get started with the model.

from transformers import pipeline

model_name = "MadMarx37/deepseek-coder-1.3b-python-peft"

def generate_output(input):
    # Run text generation pipeline with our next model
    pipe = pipeline(task="text-generation", model=model_name, tokenizer=model_name, max_length=max_length)
    result = pipe(input)
    print(result[0]['generated_text'])

Training Details

Training Hyperparameters

  • Training regime: fp16 mixed-precision with original model loaded in 4bits with bitsandbytes
  • learning_rate = 2e-3
  • lr_scheduler_type = 'cosine_with_restarts'
  • max_grad_norm = 0.001
  • weight_decay = 0.001
  • num_train_epochs = 15
  • eval_strategy = "steps"
  • eval_steps = 25

Speeds, Sizes, Times [optional]

1.3B parameters. Training time of ~2 hours on an RTX3080.

Evaluation

Testing Data, Factors & Metrics

Testing Data

https://huggingface.co/datasets/ArtifactAI/arxiv_python_research_code

Metrics

Standard training and eval loss from the HF SFTTrainer.

Results

Training Loss: 0.074100 Validation Loss: 0.022271

Summary

The training had some unstability in the gradient norms, but the overall trend in both training and validation loss were downward, and validation loss has almost plateaud, which is ideally where we want our model. The code generation on the same prompts that we tested the original model on also seem better with the finetuned model. A good way to make the model better, if we wanted to increase the finetuning data, would be to also increase the epochs.

The training run metrics can be seen here: https://wandb.ai/kevinv3796/python-autocomplete-deepseek/reports/Supervised-Finetuning-run-for-DeepSeek-Coder-1-3B-on-Python-Code--Vmlldzo3NzQ4NjY0?accessToken=bo0rlzp0yj9vxf1xe3fybfv6rbgl97w5kkab478t8f5unbwltdczy63ba9o9kwjp

Downloads last month
17
Safetensors
Model size
1.35B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.