alonzogarbanzo
commited on
Commit
•
4105f1c
1
Parent(s):
83285c3
Update README.md
Browse files
README.md
CHANGED
@@ -8,12 +8,9 @@ model-index:
|
|
8 |
results: []
|
9 |
---
|
10 |
|
11 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
12 |
-
should probably proofread and complete it, then remove this comment. -->
|
13 |
-
|
14 |
# Bloom-1b7-dialogsum-IT
|
15 |
|
16 |
-
This model is a
|
17 |
|
18 |
## Model description
|
19 |
|
@@ -25,10 +22,31 @@ More information needed
|
|
25 |
|
26 |
## Training and evaluation data
|
27 |
|
28 |
-
|
29 |
|
30 |
## Training procedure
|
31 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
### Training hyperparameters
|
33 |
|
34 |
The following hyperparameters were used during training:
|
@@ -45,7 +63,9 @@ The following hyperparameters were used during training:
|
|
45 |
|
46 |
### Training results
|
47 |
|
|
|
48 |
|
|
|
49 |
|
50 |
### Framework versions
|
51 |
|
|
|
8 |
results: []
|
9 |
---
|
10 |
|
|
|
|
|
|
|
11 |
# Bloom-1b7-dialogsum-IT
|
12 |
|
13 |
+
This model is a instruction-tuned version of [bigscience/bloom-1b7](https://huggingface.co/bigscience/bloom-1b7) on a dialog summation dataset.
|
14 |
|
15 |
## Model description
|
16 |
|
|
|
22 |
|
23 |
## Training and evaluation data
|
24 |
|
25 |
+
Instruction Tuned on the dialog summation task here: https://huggingface.co/datasets/adambjorn/UnrelatedForgettingOverhead/viewer/dialogsum/train
|
26 |
|
27 |
## Training procedure
|
28 |
|
29 |
+
Given a set of prompts:
|
30 |
+
|
31 |
+
``` python
|
32 |
+
prompts = [
|
33 |
+
"Provide a concise summary for the following dialogue:",
|
34 |
+
"Summarize this conversation in a few sentences:",
|
35 |
+
"Here is a dialogue. Can you summarize it briefly?",
|
36 |
+
"Read the following dialogue and write a short summary:",
|
37 |
+
"Condense the essence of this conversation into a summary:"
|
38 |
+
]
|
39 |
+
```
|
40 |
+
|
41 |
+
Each example is concatenated with the prompt, the dialogue, and the summary as so:
|
42 |
+
|
43 |
+
``` python
|
44 |
+
concatenated_texts = [
|
45 |
+
random.choice(prompts) + " " + dialogue + "<\s>" + " Summary:" + summary
|
46 |
+
for dialogue, summary in zip(examples['dialogue'], examples['summary'])
|
47 |
+
]
|
48 |
+
```
|
49 |
+
|
50 |
### Training hyperparameters
|
51 |
|
52 |
The following hyperparameters were used during training:
|
|
|
63 |
|
64 |
### Training results
|
65 |
|
66 |
+
Final epoch results: {'loss': 0.0137, 'grad_norm': 0.6599154472351074, 'learning_rate': 7.000000000000001e-07, 'epoch': 10.0}
|
67 |
|
68 |
+
Average results: {'train_runtime': 1142.1524, 'train_samples_per_second': 1.751, 'train_steps_per_second': 0.438, 'train_loss': 0.37129621666669843, 'epoch': 10.0}
|
69 |
|
70 |
### Framework versions
|
71 |
|