WYNN747 commited on
Commit
a233fae
·
1 Parent(s): 75a87d5

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +64 -0
README.md ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - my
5
+ tags:
6
+ - 'burmese-gpt '
7
+ - myanmar-gpt
8
+ - burmese-llm
9
+ - myanmar-llm
10
+ - llm
11
+ ---
12
+
13
+ ## Features Update (Burmese-GPT-V3)
14
+
15
+ - "Fix repeatable generation for longer sequences."
16
+ - "Support long text generation."
17
+ - "Train on a larger dataset: 55,000 Burmese text corpus."
18
+
19
+ ## Model Description (Burmese-GPT-V3)
20
+ Developed by Dr. Wai Yan, Burmese-GPT is a specialized large language model for the Burmese language, fine-tuned/pre-trained on the GPT-2 architecture, particularly the mGPT XL model. This model is primarily designed for text completion in Burmese, serving as a foundational base for fine-tuning a variety of natural language processing tasks within the Burmese language context.
21
+
22
+ **How to Use the Model**
23
+ ```bash
24
+ !pip install transformers
25
+
26
+ # Loading the Model:
27
+ from transformers import AutoTokenizer, AutoModelForCausalLM
28
+
29
+ tokenizer = AutoTokenizer.from_pretrained("WYNN747/Burmese-GPT")
30
+ model = AutoModelForCausalLM.from_pretrained("WYNN747/Burmese-GPT")
31
+
32
+ input_text = "မီးထွန်းပွဲတော်သည် သီ"
33
+ input_ids = tokenizer.encode(input_text, return_tensors='pt')
34
+ output = model.generate(input_ids, max_length=50)
35
+ print(tokenizer.decode(output[0], skip_special_tokens=True))
36
+
37
+
38
+ # [{'generated_text': 'မီးထွန်းပွဲတော် သည် သီတင်းကျွတ်လပြည့်နေ့တွင် ကျင်းပသော ရိုးရာပွဲတော်တစ်ခု ဖြစ်သည်။'}]
39
+
40
+ ```
41
+
42
+ ## Intended Use
43
+ This model, primarily designed for text completion in Burmese, serves as a foundational tool for a variety of NLP tasks. While its current primary function is to assist in generating and completing text, it holds significant potential for further applications. Researchers and developers can fine-tune this model on specialized datasets to extend its capabilities to other NLP applications, such as summarization and instruction-based tasks. It is important to note, however, that for high-stakes decisions or understanding domain-specific jargon, additional specialized training of the model is recommended to ensure accuracy and reliability.
44
+
45
+ ## Training Data
46
+ Burmese-GPT was trained on a comprehensive dataset of Burmese texts, curated by the author. This dataset, which includes literature, news, online articles, and content from Burmese Wikipedia, has been meticulously compiled to ensure a wide representation of the linguistic diversity and styles found in the Burmese language. The dataset, created by the author, is available for academic and research purposes upon request. Interested parties should contact the author to gain access to this valuable resource.
47
+
48
+ ## Ethical Considerations
49
+ Users should be aware of the inherent limitations and biases of language models. This model should be used responsibly, especially in sensitive applications, and is not intended for generating misleading or harmful content.
50
+
51
+ ## Limitations
52
+ The Burmese GPT performs well with general Burmese text but may not be as effective with highly technical or niche content. Users are advised to conduct thorough testing for their specific use cases.
53
+
54
+ ## Contact Information
55
+
56
+ - **LinkedIn:** [Dr. Wai Yan Nyein Naing](https://www.linkedin.com/in/wai-yan-nyein-naing/)
57
+ - **GitHub:** [WaiYanNyeinNaing](https://github.com/WaiYanNyeinNaing)
58
+
59
+
60
+ ## Acknowledgements
61
+
62
+ Credit and thanks to the creators of the [mGPT-XL model](https://github.com/ai-forever/mgpt) for providing the foundational model. Their contributions have been instrumental in the development of the Burmese GPT.
63
+
64
+ ........................................................................................................................................