Text Generation
Transformers
PyTorch
Safetensors
Japanese
English
qwen
custom_code
tianyuz commited on
Commit
4d79168
1 Parent(s): 6c7893a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -15
README.md CHANGED
@@ -28,7 +28,7 @@ The name `nekomata` comes from the Japanese word [`猫又/ねこまた/Nekomata`
28
 
29
  * **Library**
30
 
31
- The model was trained using code based on [EleutherAI/gpt-neox](https://github.com/EleutherAI/gpt-neox).
32
 
33
  * **Model architecture**
34
 
@@ -126,19 +126,5 @@ We compared the `Qwen` tokenizer (as used in `nekomata`) and the `llama-2` token
126
  ~~~
127
  ---
128
 
129
- # Citations
130
- ~~~
131
- @software{gpt-neox-library,
132
- title = {{GPT-NeoX: Large Scale Autoregressive Language Modeling in PyTorch}},
133
- author = {Andonian, Alex and Anthony, Quentin and Biderman, Stella and Black, Sid and Gali, Preetham and Gao, Leo and Hallahan, Eric and Levy-Kramer, Josh and Leahy, Connor and Nestler, Lucas and Parker, Kip and Pieler, Michael and Purohit, Shivanshu and Songz, Tri and Phil, Wang and Weinbach, Samuel},
134
- url = {https://www.github.com/eleutherai/gpt-neox},
135
- doi = {10.5281/zenodo.5879544},
136
- month = {8},
137
- year = {2021},
138
- version = {0.0.1},
139
- }
140
- ~~~
141
- ---
142
-
143
  # License
144
  [Tongyi Qianwen LICENSE AGREEMENT](https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT)
 
28
 
29
  * **Library**
30
 
31
+ The model was trained using code based on [aws-neuron/neuronx-nemo-megatron](https://github.com/aws-neuron/neuronx-nemo-megatron/).
32
 
33
  * **Model architecture**
34
 
 
126
  ~~~
127
  ---
128
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129
  # License
130
  [Tongyi Qianwen LICENSE AGREEMENT](https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT)