geraldOslo
commited on
Commit
•
0493f81
1
Parent(s):
7af2f18
Update README.md
Browse files
README.md
CHANGED
@@ -6,7 +6,14 @@ tags:
|
|
6 |
|
7 |
# Model Card for Model ID
|
8 |
|
9 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
|
11 |
|
12 |
|
@@ -18,9 +25,9 @@ tags:
|
|
18 |
|
19 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
20 |
|
21 |
-
- **Developed by:**
|
22 |
|
23 |
-
- **Model type:**
|
24 |
- **Language(s) (NLP):** Norwegian
|
25 |
- **License:** [More Information Needed]
|
26 |
- **Finetuned from model [meta-llama/Llama-2-13b-hf]:**
|
|
|
6 |
|
7 |
# Model Card for Model ID
|
8 |
|
9 |
+
## Model
|
10 |
+
The base model used is the Meta Llama 13B model ([meta-llama/Llama-2-13b-hf](https://huggingface.co/meta-llama/Llama-2-13b-hf)).
|
11 |
+
## Data
|
12 |
+
A dataset of prompt/response pairs about radiation protection, radiation physics, radiation biology and radiological technology as the apply in dental clinics was used to fine-tune the model. The dataset is in Norwegian and the model is fine-tuned to answer in Norwegian.
|
13 |
+
## Training
|
14 |
+
The model was trained on 6.2k prompt/response pairs from the dataset [meta-llama/Llama-2-13b-hf](https://huggingface.co/datasets/geraldOslo/RadProtDataSet) for 6 epochs on a Google Colag notebook with an A100 GPU.
|
15 |
+
|
16 |
+
The [Unsloth library](https://github.com/unslothai/unsloth) was used to train the model on a single A100 GPU.
|
17 |
|
18 |
|
19 |
|
|
|
25 |
|
26 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
27 |
|
28 |
+
- **Developed by:** Gerald Torgersen
|
29 |
|
30 |
+
- **Model type:** Chat model fine-tuned
|
31 |
- **Language(s) (NLP):** Norwegian
|
32 |
- **License:** [More Information Needed]
|
33 |
- **Finetuned from model [meta-llama/Llama-2-13b-hf]:**
|