JingzeShi commited on
Commit
c57dd1a
·
verified ·
1 Parent(s): 5f7d5d3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -11,7 +11,7 @@ pipeline_tag: text-generation
11
 
12
  # **Doge 20M**
13
 
14
- ![architecture](https://cdn-uploads.huggingface.co/production/uploads/673ab3647afcea17eb4378fd/BBPbIRiw1KxdgYp_5CABf.png)
15
 
16
  Doge is an ongoing research project where we aim to train a series of small language models to further explore whether the Transformer framework allows for more complex feedforward network structures, enabling the model to have fewer cache states and larger knowledge capacity.
17
 
@@ -36,6 +36,8 @@ In addition, Doge uses Dynamic Mask Attention as sequence transformation and can
36
 
37
  We build the Doge by doing Per-Training on [Smollm-Corpus](https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus).
38
 
 
 
39
  > NOTE: These models has not been fine-tuned for instruction, the instruction model is [here](https://huggingface.co/JingzeShi/Doge-20M-Instruct).
40
 
41
  > TODO: The larger model is under training and will be uploaded soon.
 
11
 
12
  # **Doge 20M**
13
 
14
+ ![architecture](Doge.png)
15
 
16
  Doge is an ongoing research project where we aim to train a series of small language models to further explore whether the Transformer framework allows for more complex feedforward network structures, enabling the model to have fewer cache states and larger knowledge capacity.
17
 
 
36
 
37
  We build the Doge by doing Per-Training on [Smollm-Corpus](https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus).
38
 
39
+ > NOTE: If you want to continue pre-training this model, you can find the unconverged checkpoint [here](https://huggingface.co/JingzeShi/Doge-20M-checkpoint-7000).
40
+
41
  > NOTE: These models has not been fine-tuned for instruction, the instruction model is [here](https://huggingface.co/JingzeShi/Doge-20M-Instruct).
42
 
43
  > TODO: The larger model is under training and will be uploaded soon.