Locutusque commited on
Commit
390fde4
1 Parent(s): c5107db

Change to a nicer HTML and CSS style model card

Browse files
Files changed (1) hide show
  1. README.md +154 -54
README.md CHANGED
@@ -6,58 +6,158 @@ datasets:
6
  language:
7
  - en
8
  ---
9
- # Hercules-Qwen1.5-14B
10
 
11
- <!-- Provide a quick summary of what the model is/does. -->
12
- We fine-tuned Qwen1.5-14B on Locutusque's Hercules-v4. This is M4-ai's new flagship model, with unparalleled levels of performance.
13
-
14
-
15
- ## Model Details
16
-
17
- ### Model Description
18
-
19
- <!-- Provide a longer summary of what this model is. -->
20
-
21
- This model has capabilities in math, coding, function calling, roleplay, and more. We fine-tuned it using 700,000 examples of Hercules-v4.
22
-
23
- - **Developed by:** M4-ai
24
- - **Language(s) (NLP):** English and maybe Chinese
25
- - **License:** tongyi-qianwen license
26
- - **Finetuned from model:** [Qwen1.5-14B](https://huggingface.co/Qwen/Qwen1.5-14B)
27
-
28
- ## Uses
29
-
30
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
31
-
32
- General purpose assistant, question answering, chain-of-thought, etc..
33
-
34
- ### Recommendations
35
-
36
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
37
-
38
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
39
-
40
- ## Evaluation
41
- Coming soon
42
-
43
-
44
- ## Training Details
45
-
46
- ### Training Data
47
-
48
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
49
- https://huggingface.co/datasets/Locutusque/hercules-v4.0
50
-
51
-
52
- #### Training Hyperparameters
53
-
54
- - **Training regime:** bf16 non-mixed precision
55
- ## Technical Specifications
56
-
57
- #### Hardware
58
-
59
- We used 8 Kaggle TPUs, and we trained at a global batch size of 128 and sequence length of 1024
60
-
61
- ## Contributions
62
-
63
- Thanks to @Tonic, @aloobun, @fhai50032, and @Locutusque for their contributions to this model.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  language:
7
  - en
8
  ---
 
9
 
10
+ <style>
11
+ body {
12
+ font-family: 'Segoe UI', Tahoma, Geneva, Verdana, sans-serif;
13
+ line-height: 1.6;
14
+ color: #f5f5f5;
15
+ background-color: #1e2a36;
16
+ margin: 0;
17
+ padding: 0;
18
+ }
19
+
20
+ .container {
21
+ max-width: 1200px;
22
+ margin: 20px auto;
23
+ padding: 20px;
24
+ background-color: #2a3f54;
25
+ border-radius: 8px;
26
+ box-shadow: 0 4px 8px rgba(0, 0, 0, 0.1);
27
+ display: flex;
28
+ flex-wrap: wrap;
29
+ justify-content: space-between;
30
+ }
31
+
32
+ h1 {
33
+ font-size: 2.5rem;
34
+ color: #51a3d3;
35
+ text-align: center;
36
+ margin-bottom: 30px;
37
+ width: 100%;
38
+ }
39
+
40
+ h2 {
41
+ font-size: 1.75rem;
42
+ margin: 20px 0;
43
+ color: #63b8ea;
44
+ padding-bottom: 10px;
45
+ }
46
+
47
+ h3 {
48
+ font-size: 1.25rem;
49
+ color: #80c8f4;
50
+ }
51
+
52
+ p, a {
53
+ font-size: 1rem;
54
+ }
55
+
56
+ p {
57
+ color: #b0c2ce;
58
+ margin-bottom: 20px;
59
+ }
60
+
61
+ ul {
62
+ list-style-type: none;
63
+ padding: 0;
64
+ display: flex;
65
+ flex-wrap: wrap;
66
+ justify-content: space-between;
67
+ width: 100%;
68
+ }
69
+
70
+ li {
71
+ background-color: #34495e;
72
+ padding: 20px;
73
+ margin-bottom: 10px;
74
+ border-radius: 4px;
75
+ cursor: pointer;
76
+ transition: background-color 0.3s ease, color 0.3s ease;
77
+ overflow: hidden;
78
+ color: #b0c2ce;
79
+ width: calc(50% - 10px);
80
+ box-shadow: 0 2px 4px rgba(0, 0, 0, 0.1);
81
+ }
82
+
83
+ li:hover {
84
+ background-color: #4e6a81;
85
+ color: #dfe8f1;
86
+ }
87
+
88
+ .section-content {
89
+ margin-top: 15px;
90
+ border-top: 1px solid #4e6a81;
91
+ padding-top: 10px;
92
+ }
93
+
94
+ a {
95
+ color: #a4c8e1;
96
+ text-decoration: none;
97
+ }
98
+
99
+ a:hover {
100
+ text-decoration: underline;
101
+ }
102
+
103
+ pre {
104
+ background-color: #2c3e50;
105
+ padding: 10px;
106
+ border-radius: 5px;
107
+ overflow-x: auto;
108
+ color: #b0c2ce;
109
+ }
110
+ </style>
111
+ <div class="container">
112
+ <h1>Hercules-Qwen1.5-14B</h1>
113
+ </div>
114
+
115
+ <ul>
116
+ <li>
117
+ <h2>Model Details</h2>
118
+ <div class="section-content">
119
+ <h3>Model Description</h3>
120
+ <p>This model has capabilities in math, coding, function calling, roleplay, and more. We fine-tuned it using 700,000 examples of Hercules-v4.</p>
121
+ <p><strong>Developed by:</strong> M4-ai</p>
122
+ <p><strong>Language(s) (NLP):</strong> English and maybe Chinese</p>
123
+ <p><strong>License:</strong> tongyi-qianwen license</p>
124
+ <p><strong>Finetuned from model:</strong> <a href="https://huggingface.co/Qwen/Qwen1.5-14B">Qwen1.5-14B</a></p>
125
+ </div>
126
+ </li>
127
+ <li>
128
+ <h2>Uses</h2>
129
+ <div class="section-content">
130
+ <p>General purpose assistant, question answering, chain-of-thought, etc..</p>
131
+ <h3>Recommendations</h3>
132
+ <p>Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.</p>
133
+ </div>
134
+ </li>
135
+ <li>
136
+ <h2>Evaluation</h2>
137
+ <div class="section-content">
138
+ <p>Coming soon</p>
139
+ </div>
140
+ </li>
141
+ <li>
142
+ <h2>Training Details</h2>
143
+ <div class="section-content">
144
+ <h3>Training Data</h3>
145
+ <p><a href="https://huggingface.co/datasets/Locutusque/hercules-v4.0">https://huggingface.co/datasets/Locutusque/hercules-v4.0</a></p>
146
+ <h4>Training Hyperparameters</h4>
147
+ <p><strong>Training regime:</strong> bf16 non-mixed precision</p>
148
+ </div>
149
+ </li>
150
+ <li>
151
+ <h2>Technical Specifications</h2>
152
+ <div class="section-content">
153
+ <h4>Hardware</h4>
154
+ <p>We used 8 Kaggle TPUs, and we trained at a global batch size of 128 and sequence length of 1024</p>
155
+ </div>
156
+ </li>
157
+ <li>
158
+ <h2>Contributions</h2>
159
+ <div class="section-content">
160
+ <p>Thanks to @Tonic, @aloobun, @fhai50032, and @Locutusque for their contributions to this model.</p>
161
+ </div>
162
+ </li>
163
+ </ul>