JingzeShi commited on
Commit
6a40632
1 Parent(s): c271fb0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -14
README.md CHANGED
@@ -9,11 +9,12 @@ pipeline_tag: text-generation
9
  ---
10
 
11
 
12
- # **Doge 76M**
13
 
14
  Doge is an ongoing research project where we aim to train a series of small language models to further explore whether the Transformer framework allows for more complex feedforward network structures, enabling the model to have fewer cache states and larger knowledge capacity.
15
 
16
- In addition, Doge uses Inner Function Attention with Dynamic Mask as sequence transformation and Cross Domain Mixture of Experts as state transformation. This model is trained by Jingze Shi, it only allows text input and text generation, for detailed algorithm and model architecture, please refer to [Wonderful Matrices](https://arxiv.org/abs/2407.16958), the ongoing research repository is [Doge](https://github.com/LoserCheems/Doge).
 
17
 
18
 
19
  ## Uses
@@ -21,8 +22,8 @@ In addition, Doge uses Inner Function Attention with Dynamic Mask as sequence tr
21
  ```python
22
  >>> from transformers import AutoTokenizer, AutoModelForCausalLM
23
 
24
- >>> tokenizer = AutoTokenizer.from_pretrained("JingzeShi/Doge-76M")
25
- >>> model = AutoModelForCausalLM.from_pretrained("JingzeShi/Doge-76M", trust_remote_code=True)
26
  >>> inputs = tokenizer("Hey how are you doing?", return_tensors="pt")
27
 
28
  >>> out = model.generate(**inputs, max_new_tokens=100)
@@ -31,16 +32,16 @@ In addition, Doge uses Inner Function Attention with Dynamic Mask as sequence tr
31
 
32
 
33
  ## Model Details
 
34
  > NOTE: This model has not been fine-tuned for instruction
 
35
  > TODO: The larger model is under training and will be uploaded soon.
36
 
37
- **Model Architecture**: The model architecture is a Transformer with Inner Function Attention with Dynamic Mask as sequence transformation and Cross Domain Mixture of Experts as state transformation. It can be simply understood as a Transformer with all attention and feedforward layers being sparse activation structures. For detailed information on the architecture, please refer to [Wonderful Matrices](https://arxiv.org/abs/2407.16958).
38
 
39
- || Training Data | Steps | Content Length | Tokens | LR | Batch Size | Precision |
40
- |---|---|---|---|---|---|---|---|
41
- | [Doge-22M](https://huggingface.co/LoserCheems/Doge-22M) | [HuggingFaceTB/smollm-corpus](https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus) | 5k | 2048 | 1B | 8e-4 | 0.25M | bfloat16 |
42
- | [Doge-76M](https://huggingface.co/JingzeShi/Doge-76M) | [HuggingFaceTB/smollm-corpus](https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus) | 10k | 2048 | 5B | 6e-4 | 0.5M | bfloat16 |
43
- | [Doge-197M](https://huggingface.co/JingzeShi/Doge-197M) | [HuggingFaceTB/smollm-corpus](https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus) | 20k | 2048 | 20B | 5e-4 | 1M | bfloat16 |
44
 
45
 
46
  **Training Environment**:
@@ -52,12 +53,12 @@ In addition, Doge uses Inner Function Attention with Dynamic Mask as sequence tr
52
 
53
  ```bibtex
54
  @misc{shi2024wonderfulmatrices,
55
- title={Wonderful Matrices: More Efficient and Effective Architecture for Language Modeling Tasks},
56
- author={Jingze Shi and Bingheng Wu and Lu He and Luchang Jiang},
57
  year={2024},
58
- eprint={2407.16958},
59
  archivePrefix={arXiv},
60
  primaryClass={cs.LG},
61
- url={https://arxiv.org/abs/2407.16958},
62
  }
63
  ```
 
9
  ---
10
 
11
 
12
+ # **Doge 60M**
13
 
14
  Doge is an ongoing research project where we aim to train a series of small language models to further explore whether the Transformer framework allows for more complex feedforward network structures, enabling the model to have fewer cache states and larger knowledge capacity.
15
 
16
+ In addition, Doge uses Dynamic Mask Attention as sequence transformation and can use Multi-Layer Perceptron or Cross Domain Mixture of Experts as state transformation. Dynamic Mask Attention allows the Transformer to use self-attention during training and state space during inference, and Cross Domain Mixture of Experts can directly inherit the weights of Multi-Layer Perceptron for further training. This model is trained by Jingze Shi, it only allows text input and text generation, for detailed algorithm and model architecture, please refer to [Wonderful Matrices](https://arxiv.org/abs/2412.11834), the ongoing research repository is [Wonderful Matrices](https://github.com/LoserCheems/WonderfulMatrices).
17
+
18
 
19
 
20
  ## Uses
 
22
  ```python
23
  >>> from transformers import AutoTokenizer, AutoModelForCausalLM
24
 
25
+ >>> tokenizer = AutoTokenizer.from_pretrained("JingzeShi/Doge-60M")
26
+ >>> model = AutoModelForCausalLM.from_pretrained("JingzeShi/Doge-60M", trust_remote_code=True)
27
  >>> inputs = tokenizer("Hey how are you doing?", return_tensors="pt")
28
 
29
  >>> out = model.generate(**inputs, max_new_tokens=100)
 
32
 
33
 
34
  ## Model Details
35
+
36
  > NOTE: This model has not been fine-tuned for instruction
37
+
38
  > TODO: The larger model is under training and will be uploaded soon.
39
 
 
40
 
41
+ || Training Data | Epochs | Steps | Content Length | Tokens | LR | Batch Size | Precision |
42
+ |---|---|---|---|---|---|---|---|---|
43
+ | [Doge-20M](https://huggingface.co/LoserCheems/Doge-20M) | [HuggingFaceTB/smollm-corpus](https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus) | 2 | 10k | 2048 | 5B | 8e-4 | 0.25M | bfloat16 |
44
+ | [Doge-60M](https://huggingface.co/LoserCheems/Doge-60M) | [HuggingFaceTB/smollm-corpus](https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus) | 2 | 20k | 2048 | 20B | 6e-4 | 0.5M | bfloat16 |
 
45
 
46
 
47
  **Training Environment**:
 
53
 
54
  ```bibtex
55
  @misc{shi2024wonderfulmatrices,
56
+ title={Wonderful Matrices: Combining for a More Efficient and Effective Foundation Model Architecture},
57
+ author={Jingze Shi and Bingheng Wu},
58
  year={2024},
59
+ eprint={2412.11834},
60
  archivePrefix={arXiv},
61
  primaryClass={cs.LG},
62
+ url={https://arxiv.org/abs/2412.11834},
63
  }
64
  ```