Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -7,14 +7,7 @@ tags: ['text-generation-inference', 'transformers', 'unsloth', 'mistral', 'gguf'
|
|
7 |
datasets: ['iamtarun/python_code_instructions_18k_alpaca', 'jtatman/python-code-dataset-500k', 'flytech/python-codes-25k', 'Vezora/Tested-143k-Python-Alpaca', 'codefuse-ai/CodeExercise-Python-27k', 'Vezora/Tested-22k-Python-Alpaca', 'mlabonne/Evol-Instruct-Python-26k']
|
8 |
library_name: adapter-transformers
|
9 |
metrics:
|
10 |
-
- accuracy
|
11 |
-
- bertscore
|
12 |
-
- bleu
|
13 |
-
- comet
|
14 |
- glue
|
15 |
-
- google_bleu
|
16 |
-
- perplexity
|
17 |
-
- rouge
|
18 |
---
|
19 |
|
20 |
# Uploaded model
|
@@ -33,46 +26,11 @@ This mistral model was trained 2x faster with [Unsloth](https://github.com/unslo
|
|
33 |
|
34 |
This model has been fine-tuned for various tasks and evaluated on the following benchmarks:
|
35 |
|
36 |
-
### accuracy
|
37 |
-
**Accuracy:** Not Available
|
38 |
-
|
39 |
-
![accuracy Accuracy](./accuracy_accuracy.png)
|
40 |
-
|
41 |
-
### bertscore
|
42 |
-
**Bertscore:** Not Available
|
43 |
-
|
44 |
-
![bertscore Bertscore](./bertscore_bertscore.png)
|
45 |
-
|
46 |
-
### bleu
|
47 |
-
**Bleu:** Not Available
|
48 |
-
|
49 |
-
![bleu Bleu](./bleu_bleu.png)
|
50 |
-
|
51 |
-
### comet
|
52 |
-
**Comet:** Not Available
|
53 |
-
|
54 |
-
![comet Comet](./comet_comet.png)
|
55 |
-
|
56 |
### glue
|
57 |
**Glue:** Not Available
|
58 |
|
59 |
![glue Glue](./glue_glue.png)
|
60 |
|
61 |
-
### google_bleu
|
62 |
-
**Google_bleu:** Not Available
|
63 |
-
|
64 |
-
![google_bleu Google_bleu](./google_bleu_google_bleu.png)
|
65 |
-
|
66 |
-
### perplexity
|
67 |
-
**Perplexity:** Not Available
|
68 |
-
|
69 |
-
![perplexity Perplexity](./perplexity_perplexity.png)
|
70 |
-
|
71 |
-
### rouge
|
72 |
-
**Rouge:** Not Available
|
73 |
-
|
74 |
-
![rouge Rouge](./rouge_rouge.png)
|
75 |
-
|
76 |
|
77 |
Model Size: 494,032,768 parameters
|
78 |
Required Memory: 1.84 GB
|
|
|
7 |
datasets: ['iamtarun/python_code_instructions_18k_alpaca', 'jtatman/python-code-dataset-500k', 'flytech/python-codes-25k', 'Vezora/Tested-143k-Python-Alpaca', 'codefuse-ai/CodeExercise-Python-27k', 'Vezora/Tested-22k-Python-Alpaca', 'mlabonne/Evol-Instruct-Python-26k']
|
8 |
library_name: adapter-transformers
|
9 |
metrics:
|
|
|
|
|
|
|
|
|
10 |
- glue
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
# Uploaded model
|
|
|
26 |
|
27 |
This model has been fine-tuned for various tasks and evaluated on the following benchmarks:
|
28 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
### glue
|
30 |
**Glue:** Not Available
|
31 |
|
32 |
![glue Glue](./glue_glue.png)
|
33 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
|
35 |
Model Size: 494,032,768 parameters
|
36 |
Required Memory: 1.84 GB
|