RaphaelMourad commited on
Commit
c71e037
1 Parent(s): f6da53c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -6,9 +6,9 @@ tags:
6
  - protein
7
  ---
8
 
9
- # Model Card for Mistral-Prot-v1-139M (Mistral for protein)
10
 
11
- The Mistral-Prot-v1-139M Large Language Model (LLM) is a pretrained generative chemical molecule model with 133.8M parameters.
12
  It is derived from Mixtral-8x7B-v0.1 model, which was simplified for protein: the number of layers and the hidden size were reduced.
13
  The model was pretrained using 1M protein strings from the uniprot 50 database.
14
 
@@ -26,8 +26,8 @@ Like Mixtral-8x7B-v0.1, it is a transformer model, with the following architectu
26
  import torch
27
  from transformers import AutoTokenizer, AutoModel
28
 
29
- tokenizer = AutoTokenizer.from_pretrained("RaphaelMourad/Mistral-Prot-v1-139M", trust_remote_code=True)
30
- model = AutoModel.from_pretrained("RaphaelMourad/Mistral-Prot-v1-139M", trust_remote_code=True)
31
  ```
32
 
33
  ## Calculate the embedding of a DNA sequence
@@ -48,7 +48,7 @@ Ensure you are utilizing a stable version of Transformers, 4.34.0 or newer.
48
 
49
  ## Notice
50
 
51
- Mistral-Prot-v1-139M is a pretrained base model for protein.
52
 
53
  ## Contact
54
 
 
6
  - protein
7
  ---
8
 
9
+ # Model Card for Mistral-Prot-v1-134M (Mistral for protein)
10
 
11
+ The Mistral-Prot-v1-134M Large Language Model (LLM) is a pretrained generative chemical molecule model with 133.8M parameters.
12
  It is derived from Mixtral-8x7B-v0.1 model, which was simplified for protein: the number of layers and the hidden size were reduced.
13
  The model was pretrained using 1M protein strings from the uniprot 50 database.
14
 
 
26
  import torch
27
  from transformers import AutoTokenizer, AutoModel
28
 
29
+ tokenizer = AutoTokenizer.from_pretrained("RaphaelMourad/Mistral-Prot-v1-134M", trust_remote_code=True)
30
+ model = AutoModel.from_pretrained("RaphaelMourad/Mistral-Prot-v1-134M", trust_remote_code=True)
31
  ```
32
 
33
  ## Calculate the embedding of a DNA sequence
 
48
 
49
  ## Notice
50
 
51
+ Mistral-Prot-v1-134M is a pretrained base model for protein.
52
 
53
  ## Contact
54