Edit model card

<Llama2 ๋ชจ๋ธ์„ nsmc ๋ฐ์ดํ„ฐ์…‹์„ ํ•ด๊ฒฐํ•˜๋Š” ๋ชจ๋ธ์ด ๋˜๋„๋ก ๋ฏธ์„ธํŠœ๋‹ ํ•˜๊ธฐ>

๋ชจ๋ธ: Llama2
๋ฐ์ดํ„ฐ์…‹: nsmc
https://huggingface.co/datasets/nsmc
Train ๋ฐ์ดํ„ฐ: 3000
Test ๋ฐ์ดํ„ฐ: 1000

[ํ…Œ์ŠคํŠธ ๊ฒฐ๊ณผ]

์ •ํ™•๋„: 86.10%

ํ˜ผ๋™ํ–‰๋ ฌ(Confusion Matrix)

์ •๋‹ต Positive ์ •๋‹ต Negative
์˜ˆ์ธก Positive 395 26
์˜ˆ์ธก Negative 113 466

ํ‰๊ฐ€์ง€ํ‘œ

์ •๋ฐ€๋„(Precision) 0.938
์žฌํ˜„์œจ(Recall) 0.459
F1 Score 0.616

[์„ฑ๋Šฅ ํ–ฅ์ƒ]

train ๋ฐ์ดํ„ฐ ์ˆ˜๋ฅผ 2000์—์„œ 2500, 3000์œผ๋กœ ๋Š˜๋ ค๊ฐ€๋ฉฐ ์„ฑ๋Šฅ์„ ์•ฝ 11% ์ •๋„ ๋†’์˜€์œผ๋ฉฐ, TrainingArguments์˜ max_steps ๋“ฑ์˜ ํŒŒ๋ผ๋ฏธํ„ฐ๋ฅผ ์กฐ์ ˆํ•ด๊ฐ€๋ฉฐ ์„ฑ๋Šฅ์„ ๋†’์ด๊ณ ์ž ๋…ธ๋ ฅํ•˜์˜€๋‹ค.


lora-llama-2-7b-food-order-understanding

This model is a fine-tuned version of meta-llama/Llama-2-7b-chat-hf on an unknown dataset.

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 1
  • eval_batch_size: 1
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 2
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_ratio: 0.03
  • training_steps: 300
  • mixed_precision_training: Native AMP

Training results

Framework versions

  • Transformers 4.35.2
  • Pytorch 2.1.0+cu118
  • Datasets 2.15.0
  • Tokenizers 0.15.0
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Model tree for ChloeKa/lora-llama-2-7b-food-order-understanding

Finetuned
(296)
this model