RaphaelMourad commited on
Commit
536cc26
1 Parent(s): f942f46

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -8,7 +8,7 @@ tags:
8
 
9
  # Model Card for Mistral-Prot-small (Mistral for protein)
10
 
11
- The Mistral-Prot-small Large Language Model (LLM) is a pretrained generative chemical molecule model with 1.9M parameters x 8 experts = 15.2M parameters.
12
  It is derived from Mixtral-8x7B-v0.1 model, which was simplified for protein: the number of layers and the hidden size were reduced.
13
  The model was pretrained using 1M protein strings from the uniprot 50 database.
14
 
@@ -26,8 +26,8 @@ Like Mixtral-8x7B-v0.1, it is a transformer model, with the following architectu
26
  import torch
27
  from transformers import AutoTokenizer, AutoModel
28
 
29
- tokenizer = AutoTokenizer.from_pretrained("RaphaelMourad/Mistral-Prot-small", trust_remote_code=True)
30
- model = AutoModel.from_pretrained("RaphaelMourad/Mistral-Prot-small", trust_remote_code=True)
31
  ```
32
 
33
  ## Calculate the embedding of a DNA sequence
@@ -48,7 +48,7 @@ Ensure you are utilizing a stable version of Transformers, 4.34.0 or newer.
48
 
49
  ## Notice
50
 
51
- Mistral-Prot-small is a pretrained base model for protein.
52
 
53
  ## Contact
54
 
 
8
 
9
  # Model Card for Mistral-Prot-small (Mistral for protein)
10
 
11
+ The Mistral-Prot-v1-15M Large Language Model (LLM) is a pretrained generative chemical molecule model with 1.9M parameters x 8 experts = 15.2M parameters.
12
  It is derived from Mixtral-8x7B-v0.1 model, which was simplified for protein: the number of layers and the hidden size were reduced.
13
  The model was pretrained using 1M protein strings from the uniprot 50 database.
14
 
 
26
  import torch
27
  from transformers import AutoTokenizer, AutoModel
28
 
29
+ tokenizer = AutoTokenizer.from_pretrained("RaphaelMourad/Mistral-Prot-v1-15M", trust_remote_code=True)
30
+ model = AutoModel.from_pretrained("RaphaelMourad/Mistral-Prot-v1-15M", trust_remote_code=True)
31
  ```
32
 
33
  ## Calculate the embedding of a DNA sequence
 
48
 
49
  ## Notice
50
 
51
+ Mistral-Prot-v1-15M is a pretrained base model for protein.
52
 
53
  ## Contact
54