Sirclavin commited on
Commit
94d1c94
1 Parent(s): 76ce9d9

Upload 11 files

Browse files
README.md CHANGED
@@ -1,166 +1,9 @@
1
  ---
2
- license: other
3
- license_name: cfai
4
- license_link: LICENSE
5
- language:
6
- - en
7
- metrics:
8
- - accuracy
9
- library_name: adapter-transformers
10
- pipeline_tag: text-generation
11
  ---
12
- # Model Card for Model ID
13
- If you're looking for a text generative model that is creative and generates human-like content and is memory-efficient this model might full-fill your needs give it a try!
14
 
15
-
16
- ### Model Description
17
- NeXGen is a state-of-the-art text generative model designed to meet the demands of users seeking creativity and human-like content generation. This cutting-edge model is equipped with advanced capabilities that set it apart in the realm of natural language processing.
18
- One of NeXGen's standout features is its ability to generate highly creative and contextually relevant text. Whether you're looking to compose engaging stories, craft imaginative dialogues, or generate unique pieces of writing, NeXGen excels at producing content that mirrors the fluency and style of human expression. Its creative prowess extends to various genres and themes, making it a versatile tool for writers, content creators, and anyone in need of compelling textual output.
19
- Furthermore, NeXGen boasts an impressive understanding of context, allowing it to generate coherent and contextually appropriate responses. This contextual awareness enhances the model's ability to provide relevant information and maintain a natural flow in generated text, contributing to a more authentic and human-like reading experience.
20
- Users will find NeXGen to be an intuitive and user-friendly tool, allowing for seamless integration into various applications and platforms. Its user interface is designed for accessibility, ensuring that both novice and experienced users can harness the model's capabilities with ease.
21
- The model's architecture is underpinned by the latest advancements in natural language processing, leveraging sophisticated algorithms and vast datasets to achieve high-quality text generation. NeXGen's training regimen includes exposure to diverse linguistic patterns, enabling it to adapt to different writing styles and linguistic nuances.
22
- NeXGen is not only a powerful text generator but also a tool for enhancing productivity. Whether you need assistance in drafting creative content, brainstorming ideas, or generating textual prompts, NeXGen is equipped to facilitate the creative process and provide valuable support.
23
- In summary, NeXGen stands out as a go-to solution for those in search of a text generative model that excels in creativity, context awareness, and human-like content generation. Its user-friendly interface, advanced architecture, and versatility make it a valuable asset for a wide range of applications, offering users an innovative and efficient tool to meet their creative writing needs.
24
-
25
-
26
-
27
- - **Developed by:** [Sirclavin or (me)]
28
- - **Shared by:** [CrabfishAI]
29
- - **Model type:** [text generative model]
30
- - **Language(s) (NLP):** [English(EN)]
31
- - **License:** [CFAI]
32
-
33
-
34
- ## Uses
35
-
36
- 1. **Content Creation:**
37
- - Generate creative writing, including stories, poetry, and fictional narratives.
38
- - Produce marketing copy, ad content, and product descriptions with a natural and engaging tone.
39
-
40
- 2. **Assistance in Writing:**
41
- - Aid authors, journalists, and bloggers in brainstorming ideas and drafting articles.
42
- - Provide assistance in generating outlines or summaries for writing projects.
43
-
44
- 3. **Educational Support:**
45
- - Assist students in generating ideas for essays, reports, or creative writing assignments.
46
- - Offer language learning support by providing contextually relevant sentences and phrases.
47
-
48
- 4. **Chatbot Development:**
49
- - Power conversational agents and chatbots with human-like responses in customer service or information retrieval scenarios.
50
-
51
- 5. **Prototyping and Idea Generation:**
52
- - Facilitate brainstorming sessions by generating ideas and concepts for product development or problem-solving.
53
-
54
- 6. **Social Media Content:**
55
- - Generate engaging captions for social media posts, helping users maintain a consistent and appealing online presence.
56
-
57
- 7. **Personal Assistant Applications:**
58
- - Assist users in drafting emails, messages, or other forms of communication with a natural and personalized touch.
59
-
60
- 8. **Entertainment and Gaming:**
61
- - Enhance storytelling in video games by generating dynamic and contextually appropriate dialogues.
62
- - Contribute to the creation of interactive fiction or game narratives.
63
-
64
- 9. **Accessibility Features:**
65
- - Provide support for individuals with disabilities, such as generating text for those who may have difficulty typing.
66
-
67
- 10. **Innovative Writing Tools:**
68
- - Integrate with writing platforms to offer advanced suggestions, corrections, and improvements to users' writing.
69
-
70
- 11. **Research and Data Analysis:**
71
- - Assist researchers in generating textual summaries or insights from large datasets.
72
-
73
- ### Direct Use
74
-
75
- 1. **Automated Email Drafting:**
76
- - Quickly compose emails by providing key points, and let NeXGen generate a well-articulated message with appropriate language and tone.
77
-
78
- 2. **Blog Post Generation:**
79
- - Input a topic or key points, and NeXGen can assist in generating sections of a blog post or even an entire article.
80
-
81
- 3. **Copywriting for Advertisements:**
82
- - Generate creative and persuasive copy for advertisements, social media posts, or marketing materials.
83
-
84
- 4. **Code Commenting:**
85
- - Assist developers in generating clear and concise comments for their code to improve documentation and understanding.
86
-
87
- 5. **Storyline Creation for Games:**
88
- - Aid game developers in creating dynamic and engaging storylines or dialogues for characters within video games.
89
-
90
- 6. **Learning Material Generation:**
91
- - Develop study guides, flashcards, or educational content by providing key concepts to NeXGen.
92
-
93
- 7. **Personal Journaling Assistance:**
94
- - Generate prompts or suggestions for users looking to maintain a personal journal or diary.
95
-
96
- 8. **In-app Chatbot Responses:**
97
- - Power chatbots within applications with natural and context-aware responses for user interactions.
98
-
99
- 9. **Social Media Status Updates:**
100
- - Quickly generate interesting and varied status updates for platforms like Twitter or Facebook.
101
-
102
- 10. **Scriptwriting Support:**
103
- - Assist screenwriters in developing dialogue, scenes, or even plot points for film or television scripts.
104
-
105
- 11. **Product Descriptions:**
106
- - Generate compelling and informative product descriptions for e-commerce websites.
107
-
108
- 12. **Idea Expansion:**
109
- - Provide expanded ideas and details for creative projects, helping users flesh out their initial concepts.
110
-
111
- 13. **Meeting Note Summaries:**
112
- - Summarize meeting notes or discussions, condensing information into clear and coherent points.
113
-
114
- 14. **Legal Document Drafting:**
115
- - Assist legal professionals in generating preliminary drafts for contracts, agreements, or other legal documents.
116
-
117
- ## How to get Started with the Model
118
-
119
- Use the code below to get started with the model:
120
- ```python
121
- from transformers import AutoModelForCausalLM, AutoTokenizer
122
- model_name = "Sirclavin/NeXGen-based"
123
- model = AutoModelForCausalLM.from_pretrained(model_name)
124
- tokenizer = AutoTokenizer.from_pretrained(model_name)
125
- ```
126
-
127
- And to use the model for text generation:
128
- ```python
129
- from transformers import AutoTokenizer, AutoModelForCausalLM
130
-
131
- # Specify the model name from Hugging Face Model Hub
132
- model_name = "Sirclavin/NeXGen-based"
133
- tokenizer = AutoTokenizer.from_pretrained(model_name)
134
- model = AutoModelForCausalLM.from_pretrained(model_name)
135
-
136
- def generate_text(prompt, max_length=100, num_beams=5, no_repeat_ngram_size=2, top_k=50, top_p=0.95, temperature=0.7):
137
- input_ids = tokenizer.encode(prompt, return_tensors="pt")
138
-
139
- # Ensure attention_mask is provided
140
- attention_mask = input_ids.ne(tokenizer.pad_token_id).float()
141
-
142
- # Generate output text
143
- output = model.generate(
144
- input_ids,
145
- max_length=max_length,
146
- num_beams=num_beams,
147
- no_repeat_ngram_size=no_repeat_ngram_size,
148
- top_k=top_k,
149
- top_p=top_p,
150
- temperature=temperature,
151
- attention_mask=attention_mask # Pass attention_mask to the generation method
152
- )
153
-
154
- decoded_output = tokenizer.decode(output[0], skip_special_tokens=True)
155
- return decoded_output
156
-
157
- # Example usage:
158
- prompt = "Your prompt here"
159
- generated_text = generate_text(prompt, max_length=200)
160
-
161
- print("Generated Text:")
162
- print(generated_text)
163
- ```
164
-
165
- ## Training Data
166
- the dataset used to train this model is unknown
 
1
  ---
2
+ tags:
3
+ - autotrain
4
+ - text-generation
5
+ widget:
6
+ - text: "I love AutoTrain because "
 
 
 
 
7
  ---
 
 
8
 
9
+ # Model Trained Using AutoTrain
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|pad|>": 50258,
3
+ "<|startoftext|>": 50257
4
+ }
config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "farnhua/gpt2-small-self_instruct_human_eval",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 32,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 16,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "c_attn"
18
+ ],
19
+ "task_type": "CAUSAL_LM"
20
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60f3ea3765c1b02514bb4b684bd3d1f5004f1101b81bc7c09b3ba0df014ff825
3
+ size 497805594
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|pad|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "50256": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "50257": {
14
+ "content": "<|startoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "50258": {
22
+ "content": "<|pad|>",
23
+ "lstrip": false,
24
+ "normalized": true,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "bos_token": "<|startoftext|>",
31
+ "clean_up_tokenization_spaces": true,
32
+ "eos_token": "<|endoftext|>",
33
+ "errors": "replace",
34
+ "model_max_length": 1024,
35
+ "pad_token": "<|pad|>",
36
+ "tokenizer_class": "GPT2Tokenizer",
37
+ "unk_token": "<|endoftext|>"
38
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:451c10f29d391c5977d9401a129c75070c1f9e21c563dee1c87aee84480bdc81
3
+ size 4472
training_params.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "farnhua/gpt2-small-self_instruct_human_eval",
3
+ "data_path": "timdettmers/openassistant-guanaco",
4
+ "project_name": "NeXGen",
5
+ "train_split": "train",
6
+ "valid_split": null,
7
+ "text_column": "text",
8
+ "rejected_text_column": "rejected",
9
+ "token": null,
10
+ "lr": 0.0002,
11
+ "epochs": 3,
12
+ "batch_size": 2,
13
+ "warmup_ratio": 0.1,
14
+ "gradient_accumulation": 1,
15
+ "optimizer": "adamw_torch",
16
+ "scheduler": "linear",
17
+ "weight_decay": 0.0,
18
+ "max_grad_norm": 1.0,
19
+ "seed": 42,
20
+ "add_eos_token": false,
21
+ "block_size": -1,
22
+ "use_peft": true,
23
+ "lora_r": 16,
24
+ "lora_alpha": 32,
25
+ "lora_dropout": 0.05,
26
+ "logging_steps": -1,
27
+ "evaluation_strategy": "epoch",
28
+ "save_total_limit": 1,
29
+ "save_strategy": "epoch",
30
+ "auto_find_batch_size": false,
31
+ "fp16": false,
32
+ "push_to_hub": false,
33
+ "use_int8": false,
34
+ "model_max_length": 2048,
35
+ "repo_id": null,
36
+ "use_int4": true,
37
+ "trainer": "sft",
38
+ "target_modules": null,
39
+ "merge_adapter": false,
40
+ "username": null,
41
+ "use_flash_attention_2": false,
42
+ "log": "none",
43
+ "disable_gradient_checkpointing": false,
44
+ "model_ref": null,
45
+ "dpo_beta": 0.1,
46
+ "prompt_text_column": "prompt"
47
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff