Update README.md
Browse files
README.md
CHANGED
@@ -109,21 +109,6 @@ This model is trained and maintained by KongYang from Shanghai Jiao Tong Univers
|
|
109 |
## License
|
110 |
This model is released under the CC BY-NC-SA 4.0 license.
|
111 |
|
112 |
-
## Other
|
113 |
-
这个模型基于google/mt5-base进行迭代训练,epoch=10,在我的GTX 4070机器上跑了大概7天时间。训练集有10万行,主要是国内的中文新闻,不存在BBC,所以翻译的时候不会出现类似于“本文不代表bbc观点”之类的语言。并且我的数据集也经过深度过滤,基本不会出现意外的文字,可以非常好的用来进行缩略。我的缩略主要输入分为300字的chunk,然后再进行缩略,然后再把文字join就可以完成大段的缩略。
|
114 |
-
后续会进一步在我的100万行数据源进行训练,并开源模型,敬请期待。
|
115 |
-
另外,我已经把此缩略用在了我的本地知识库项目中,项目链接为:
|
116 |
-
https://github.com/erickong/document.ai
|
117 |
-
|
118 |
-
另外感谢此项目:
|
119 |
-
https://github.com/csebuetnlp/xl-sum/tree/master/seq2seq
|
120 |
-
|
121 |
-
This model is iteratively trained based on google/mt5-base with 10 epochs, taking approximately 7 days on my machine equipped with a GTX 4070 graphics card. The training set consists of 100,000 lines, mainly domestic Chinese news, and does not include sources such as BBC, hence translations such as "This article does not represent the views of BBC" will not appear. My dataset has also undergone deep filtering, virtually eliminating unexpected text, making it excellent for summarization. The primary input for my summarization is chunks of 300 words, which are then summarized, and the resulting text can be joined together to summarize larger sections.
|
122 |
-
|
123 |
-
In the future, I plan to further train this model with my dataset of 1 million lines and release the model. Please stay tuned.
|
124 |
-
|
125 |
-
Additionally, I have already incorporated this summarization into my local knowledge base project. The link to the project is: https://github.com/erickong/document.ai
|
126 |
-
|
127 |
## Citation
|
128 |
|
129 |
If you use this model in your research, please cite:
|
|
|
109 |
## License
|
110 |
This model is released under the CC BY-NC-SA 4.0 license.
|
111 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
112 |
## Citation
|
113 |
|
114 |
If you use this model in your research, please cite:
|