jojo-ai-mst commited on
Commit
8efdacc
·
verified ·
1 Parent(s): dae78ab

MyanmarGPT-Chat

Browse files
Files changed (1) hide show
  1. README.md +126 -0
README.md ADDED
@@ -0,0 +1,126 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ language:
4
+ - my
5
+ - en
6
+ library_name: transformers
7
+ tags:
8
+ - chat
9
+ - myanmar
10
+ - burmese
11
+ - llm
12
+ widget:
13
+ - text: "User: မြန်မာနိုင်ငံအကြောင်းရှင်းပြပါ။\n Assistant: "
14
+ example_title: Example 1
15
+ - text: "User: ရုရှားနိုင်ငံအကြောင်းပြောပြပါ\n Assistant: "
16
+ example_title: Example 2
17
+ - text: "User: ကွန်မြူနစ်ဆိုတာဘာလဲ\n Assistant: "
18
+ example_title: Example 3
19
+ ---
20
+
21
+ # MyanmarGPT-Chat
22
+
23
+
24
+ MyanmarGPT-Chat is a question answering model available in Burmese language. It is fine-tuned via the foundational model called [MyanmarGPT](https://huggingface.co/jojo-ai-mst/MyanmarGPT).
25
+
26
+ Dataset used is called "A Brief History of the World" curated by the creator, Min Si Thu.
27
+ It can answer general knowledge about world history.
28
+
29
+
30
+ ## Model Details
31
+
32
+ MyanmarGPT-Chat is based on MyanmarGPT model.
33
+ As MyanmarGPT is a frontier model for Burmese language and it is getting used by lots of people around Myanmar,
34
+ MyanmarGPT-Chat is required to build as a foundational model for question-answering language model.
35
+
36
+
37
+ ### Model Description
38
+
39
+ <!-- Provide a longer summary of what this model is. -->
40
+
41
+
42
+
43
+ - **Developed by:** [Min Si Thu](https://huggingface.co/jojo-ai-mst)
44
+ - **Funded by [self]:** [More Information Needed]
45
+ - **Shared by [optional]:** [More Information Needed]
46
+ - **Model type:** GPT2
47
+ - **Language(s) (NLP):** Burmese, English
48
+ - **License:** CreativeML OpenRAIL-M
49
+ - **Finetuned from model [MyanmarGPT]:** [MyanmarGPT](https://huggingface.co/jojo-ai-mst/MyanmarGPT)
50
+
51
+ ### Model Sources [optional]
52
+
53
+ <!-- Provide the basic links for the model. -->
54
+
55
+ - **Repository:** [https://github.com/MinSiThu/MyanmarGPT]
56
+ - **Paper [optional]:** [More Information Needed]
57
+ - **Demo [optional]:** [More Information Needed]
58
+
59
+ ## Uses
60
+
61
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
62
+
63
+ ### Direct Use
64
+
65
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
66
+
67
+ Originally crafted for text completion in Burmese, this model functions as a fundamental asset for various Natural Language Processing (NLP) tasks. Although its primary role is presently centered on aiding in text generation and completion, it harbors considerable potential for broader applications. Researchers and developers have the option to refine this model using specialized datasets, thereby expanding its utility to other NLP domains, including summarization and instruction-based tasks. Nevertheless, it is crucial to acknowledge that when dealing with high-stakes decisions or comprehending domain-specific terminology, additional specialized training for the model is advised to ensure optimal accuracy and reliability.
68
+
69
+ ### Out-of-Scope Use
70
+
71
+ Users need to recognize the inherent limitations and biases present in language models. Responsible usage is crucial, particularly in sensitive contexts, as this model is not designed to generate misleading or harmful content.
72
+
73
+ [More Information Needed]
74
+
75
+ ## Bias, Risks, and Limitations
76
+
77
+ While the MyanmarGPT-Chat excels in handling general Burmese text about history of countires around the world, its effectiveness might be limited when dealing with daily-life spoken burmese words. Users are encouraged to perform comprehensive testing tailored to their specific use cases.
78
+
79
+ [More Information Needed]
80
+
81
+ ### Recommendations
82
+
83
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
84
+
85
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
86
+
87
+ ## How to Get Started with the Model
88
+
89
+ ```shell
90
+ !pip install transformers
91
+ ```
92
+
93
+ ```python
94
+ from transformers import GPT2LMHeadModel, GPT2Tokenizer
95
+
96
+ # Load MyanmarGPT-Chat model and tokenizer
97
+ model = GPT2LMHeadModel.from_pretrained("jojo-ai-mst/MyanmarGPT")
98
+ tokenizer = GPT2Tokenizer.from_pretrained("jojo-ai-mst/MyanmarGPT")
99
+
100
+ def generate_text(prompt, max_length=300, temperature=0.8, top_k=50):
101
+ input_ids = tokenizer.encode(prompt, return_tensors="pt").cuda() # remove .cude() if only cpu
102
+ output = model.generate(
103
+ input_ids,
104
+ max_length=max_length,
105
+ temperature=temperature,
106
+ top_k=top_k,
107
+ pad_token_id=tokenizer.eos_token_id,
108
+ do_sample=True
109
+ )
110
+ for result in output:
111
+ generated_text = tokenizer.decode(result, skip_special_tokens=True)
112
+ print(generated_text)
113
+
114
+ generate_text("User: မြန်မာနိုင်ငံအကြောင်းရှင်းပြပါ။\n Assistant: ")
115
+
116
+ ```
117
+
118
+
119
+
120
+ ## Citation [optional]
121
+
122
+ - MinSithu, MyanmarGPT, https://huggingface.co/jojo-ai-mst/MyanmarGPT, 1.1-SweptWood
123
+
124
+ ## Model Card Contact
125
+
126
+ [More Information Needed]