Text Generation
Transformers
Safetensors
English
mistral
genomics
medical
conversational
text-generation-inference
Inference Endpoints
kimou605 commited on
Commit
410a862
·
verified ·
1 Parent(s): 2bfab49

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -97
README.md CHANGED
@@ -1,14 +1,21 @@
1
  ---
2
  license: apache-2.0
3
  library_name: transformers
4
- tags: []
 
 
 
 
 
 
 
5
  ---
6
 
7
  # Model Card for Model ID
8
 
9
  <!-- Provide a quick summary of what the model is/does. -->
10
 
11
- BioTATA 7B is a hybrid model merged between BioMistral 7B Dare and a 4bit QLORA adapter trained on TATA/NO TATA sequences from INSTADEEP'S dataset
12
 
13
  ## Model Details
14
 
@@ -16,8 +23,6 @@ BioTATA 7B is a hybrid model merged between BioMistral 7B Dare and a 4bit QLORA
16
 
17
  <!-- Provide a longer summary of what this model is. -->
18
 
19
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
20
-
21
  - **Developed by:** Med Karim Akkari (kimou605)
22
  - **Funded by :** Med Karim Akkari (kimou605)
23
  - **Shared by :** Med Karim Akkari (kimou605)
@@ -31,10 +36,9 @@ This is the model card of a 🤗 transformers model that has been pushed on the
31
  <!-- Provide the basic links for the model. -->
32
 
33
  - **Repository:** kimou605/BioTATA-7B
34
- - **Paper [optional]:** [More Information Needed]
35
  - **Demo [optional]:** [More Information Needed]
36
 
37
- ## Uses
38
 
39
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
40
  ```python
@@ -124,18 +128,15 @@ This will inference the model on 4.8GB Vram
124
  This model has been developped to show how can a medical LLM adapt itself to identify sequences as TATA/NO TATA
125
  The adapter has been trained on a 53.3k rows for only 1 epoch (due to hardware limitations)
126
 
 
127
 
128
  ### Recommendations
129
 
130
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
131
 
132
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
133
-
134
- ## How to Get Started with the Model
135
 
136
- Use the code below to get started with the model.
137
 
138
- [More Information Needed]
139
 
140
  ## Training Details
141
 
@@ -149,11 +150,6 @@ kimou605/TATA-NOTATA-FineMistral-nucleotide_transformer_downstream_tasks
149
 
150
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
151
 
152
- #### Preprocessing [optional]
153
-
154
- Importing the dataset already preprocessed for mistral finetuning purposes
155
-
156
-
157
  #### Training Hyperparameters
158
 
159
  - **Training regime:** BF16 4bits <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
@@ -166,44 +162,6 @@ batch_per_gpu 32
166
  GPU: NVIDIA A40 45GB Vram
167
 
168
 
169
- ## Evaluation
170
-
171
- <!-- This section describes the evaluation protocols and provides the results. -->
172
-
173
- ### Testing Data, Factors & Metrics
174
-
175
- #### Testing Data
176
-
177
- <!-- This should link to a Dataset Card if possible. -->
178
-
179
- [More Information Needed]
180
-
181
- #### Factors
182
-
183
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
184
-
185
- [More Information Needed]
186
-
187
- #### Metrics
188
-
189
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
190
-
191
- [More Information Needed]
192
-
193
- ### Results
194
-
195
- [More Information Needed]
196
-
197
- #### Summary
198
-
199
-
200
-
201
- ## Model Examination [optional]
202
-
203
- <!-- Relevant interpretability work for the model goes here -->
204
-
205
- [More Information Needed]
206
-
207
  ## Environmental Impact
208
 
209
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
@@ -215,50 +173,8 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
215
  - **Cloud Provider:** vast.ai
216
  - **Compute Region:** Europe
217
 
218
- ## Technical Specifications [optional]
219
-
220
- ### Model Architecture and Objective
221
-
222
- [More Information Needed]
223
-
224
- ### Compute Infrastructure
225
-
226
- [More Information Needed]
227
-
228
- #### Hardware
229
-
230
- [More Information Needed]
231
-
232
- #### Software
233
-
234
- [More Information Needed]
235
-
236
- ## Citation [optional]
237
-
238
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
239
-
240
- **BibTeX:**
241
-
242
- [More Information Needed]
243
-
244
- **APA:**
245
-
246
- [More Information Needed]
247
-
248
- ## Glossary [optional]
249
-
250
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
251
-
252
- [More Information Needed]
253
-
254
- ## More Information [optional]
255
-
256
- [More Information Needed]
257
-
258
- ## Model Card Authors [optional]
259
 
260
- [More Information Needed]
261
 
262
  ## Model Card Contact
263
 
264
- [More Information Needed]
 
1
  ---
2
  license: apache-2.0
3
  library_name: transformers
4
+ tags:
5
+ - biology
6
+ datasets:
7
+ - kimou605/TATA-NOTATA-FineMistral-nucleotide_transformer_downstream_tasks
8
+ - InstaDeepAI/nucleotide_transformer_downstream_tasks
9
+ language:
10
+ - en
11
+ pipeline_tag: text-generation
12
  ---
13
 
14
  # Model Card for Model ID
15
 
16
  <!-- Provide a quick summary of what the model is/does. -->
17
 
18
+ BioTATA 7B is a hybrid model merged between BioMistral 7B Dare and a 4bit QLORA adapter trained on TATA/NO TATA sequences from InstaDeepAI nucleotide_transformer_downstream_tasks dataset (promoters_all subset)
19
 
20
  ## Model Details
21
 
 
23
 
24
  <!-- Provide a longer summary of what this model is. -->
25
 
 
 
26
  - **Developed by:** Med Karim Akkari (kimou605)
27
  - **Funded by :** Med Karim Akkari (kimou605)
28
  - **Shared by :** Med Karim Akkari (kimou605)
 
36
  <!-- Provide the basic links for the model. -->
37
 
38
  - **Repository:** kimou605/BioTATA-7B
 
39
  - **Demo [optional]:** [More Information Needed]
40
 
41
+ ## How to Get Started with the Model
42
 
43
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
  ```python
 
128
  This model has been developped to show how can a medical LLM adapt itself to identify sequences as TATA/NO TATA
129
  The adapter has been trained on a 53.3k rows for only 1 epoch (due to hardware limitations)
130
 
131
+ THIS MODEL IS FOR RESEARCH PURPOSES NOT FOR PRODUCTION
132
 
133
  ### Recommendations
134
 
135
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
136
 
137
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.
 
 
138
 
 
139
 
 
140
 
141
  ## Training Details
142
 
 
150
 
151
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
152
 
 
 
 
 
 
153
  #### Training Hyperparameters
154
 
155
  - **Training regime:** BF16 4bits <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
 
162
  GPU: NVIDIA A40 45GB Vram
163
 
164
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
165
  ## Environmental Impact
166
 
167
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
 
173
  - **Cloud Provider:** vast.ai
174
  - **Compute Region:** Europe
175
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
176
 
 
177
 
178
  ## Model Card Contact
179
 
180
+ Karim Akkari (kimou605)