|
--- |
|
license: mit |
|
--- |
|
|
|
A Llama3 based model fine-tuned for cybersecurity domain. |
|
|
|
The model was finetuned over https://huggingface.co/datasets/unibuc-cs/CyberGuardianDataset. |
|
While the Llama3 original perplexity on the dataset started ~21.9, we managed to fine-tune to ~7.8 on this dataset, still preserving the general language abilities. |
|
|
|
You can load the model and see its LoRA config as below: |
|
``` |
|
from peft import get_peft_model, AutoPeftModelForCausalLM, PeftModel, PeftConfig |
|
from transformers import AutoModel |
|
config = PeftConfig.from_pretrained("unibuc-cs/CyberGuardian") |
|
model = AutoModel.from_pretrained(config.base_model_name_or_path) |
|
``` |
|
|
|
|
|
Refer to our github page for details: https://github.com/unibuc-cs/CyberGuardian |