update contributers info and rearrange sections
Browse files
README.md
CHANGED
@@ -132,10 +132,19 @@ The pre-training dataset may have contained offensive or inappropriate content e
|
|
132 |
|
133 |
We are utilizing the v1 version of the [novelai-tokenizer](https://github.com/NovelAI/novelai-tokenizer), introduced by [NovelAI](https://novelai.net/), because it processes both Japanese and English text effectively and efficiently. We extend our gratitude to NovelAI for allowing us to use their remarkable work. For more details about the tokenizer, please refer to their [blog post](https://blog.novelai.net/novelais-new-llm-tokenizer-5bc140e17642).
|
134 |
|
135 |
-
We are grateful for the contributions of the EleutherAI Polyglot-JA team in helping us to collect a large amount of pre-training data in Japanese. Polyglot-JA members includes
|
136 |
|
137 |
We are also appreciative of [AI Novelist/Sta (Bit192, Inc.)](https://ai-novel.com/index.php) and the numerous contributors from [Stable Community Japan](https://discord.gg/VPrcE475HB) for assisting us in gathering a large amount of high-quality Japanese textual data for model training.
|
138 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
139 |
## Citations
|
140 |
|
141 |
```bibtext
|
@@ -148,13 +157,4 @@ We are also appreciative of [AI Novelist/Sta (Bit192, Inc.)](https://ai-novel.co
|
|
148 |
year = {2021},
|
149 |
version = {0.0.1},
|
150 |
}
|
151 |
-
```
|
152 |
-
|
153 |
-
## How to cite
|
154 |
-
```
|
155 |
-
@misc{JapaneseStableLMBaseAlpha7B,
|
156 |
-
url={[https://huggingface.co/stabilityai/japanese-stablelm-base-alpha-7b](https://huggingface.co/stabilityai/japanese-stablelm-base-alpha-7b)},
|
157 |
-
title={Japanese StableLM Base Alpha 7B},
|
158 |
-
author={Lee, Meng and Nakamura, Fujiki and Shing, Makoto and McCann, Paul and Akiba, Takuya and Orii, Naoki}
|
159 |
-
}
|
160 |
```
|
|
|
132 |
|
133 |
We are utilizing the v1 version of the [novelai-tokenizer](https://github.com/NovelAI/novelai-tokenizer), introduced by [NovelAI](https://novelai.net/), because it processes both Japanese and English text effectively and efficiently. We extend our gratitude to NovelAI for allowing us to use their remarkable work. For more details about the tokenizer, please refer to their [blog post](https://blog.novelai.net/novelais-new-llm-tokenizer-5bc140e17642).
|
134 |
|
135 |
+
We are grateful for the contributions of the EleutherAI Polyglot-JA team in helping us to collect a large amount of pre-training data in Japanese. Polyglot-JA members includes Hyunwoong Ko (Project Lead), Fujiki Nakamura (originally started this project when he commited to the Polyglot team), Yunho Mo, Minji Jung and Su-Kyeong Jang.
|
136 |
|
137 |
We are also appreciative of [AI Novelist/Sta (Bit192, Inc.)](https://ai-novel.com/index.php) and the numerous contributors from [Stable Community Japan](https://discord.gg/VPrcE475HB) for assisting us in gathering a large amount of high-quality Japanese textual data for model training.
|
138 |
|
139 |
+
## How to cite
|
140 |
+
```
|
141 |
+
@misc{JapaneseStableLMBaseAlpha7B,
|
142 |
+
url={[https://huggingface.co/stabilityai/japanese-stablelm-base-alpha-7b](https://huggingface.co/stabilityai/japanese-stablelm-base-alpha-7b)},
|
143 |
+
title={Japanese StableLM Base Alpha 7B},
|
144 |
+
author={Lee, Meng and Nakamura, Fujiki and Shing, Makoto and McCann, Paul and Akiba, Takuya and Orii, Naoki}
|
145 |
+
}
|
146 |
+
```
|
147 |
+
|
148 |
## Citations
|
149 |
|
150 |
```bibtext
|
|
|
157 |
year = {2021},
|
158 |
version = {0.0.1},
|
159 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
160 |
```
|