kevin009 commited on
Commit
2b45530
·
verified ·
1 Parent(s): 636d711
Files changed (1) hide show
  1. README.md +23 -28
README.md CHANGED
@@ -1,6 +1,8 @@
1
  ---
2
  language:
3
  - en
 
 
4
  ---
5
  # Model Card for BabyLlama v0.6
6
 
@@ -9,12 +11,15 @@ language:
9
  **Repository:** kevin009/babyllama-v0.6
10
  **Architecture:** LlamaForCausalLM, based on TinyLlama 1.1b
11
  **Model Type:** llama
12
- **Version:** 0.5
13
 
14
  ## Model Description
15
- BabyLlama v0.6 is an innovative conversational AI model, specialized in generating creative and humorous interactions. While it's built on the Llama2 architecture and specifically draws from the TinyLlama 1.1b, this version sets itself apart by not strictly adhering to user instructions. Instead, it aims to replicate human-like conversation in a manner that's distinctly recognizable from actual human dialogue, focusing on creativity and humor.
16
 
17
- With a Combining RLHF and DPO fine-tuning involved 5 different epochs, with 200 steps in each epoch, applied to over half a million conversations in low learrning rate.
 
 
 
 
18
 
19
  ## Technical Specifications
20
 
@@ -30,11 +35,6 @@ With a Combining RLHF and DPO fine-tuning involved 5 different epochs, with 200
30
  ## Use Cases
31
  This model excels in applications where engaging, entertaining, and uniquely human-distinguishable AI responses are valued. It is particularly suited for chatbots, entertainment platforms, interactive games, and social experiments where the focus is on creativity, humor, and the unexpected.
32
 
33
- BabyLlama v0.6 is particularly adept for applications within the realm of role-playing games (RPGs), interactive storytelling, and simulation-based training where dynamic, engaging, and character-driven dialogues are essential. It can serve as an AI companion or NPC (non-player character) that interacts with users in a variety of scenarios, ranging from fantasy and adventure to everyday social simulations. Its unique conversational style enhances the immersive experience by providing responses that are not only contextually relevant but also infused with creativity and humor, thereby elevating the narrative depth and engagement of the role-playing environment.
34
-
35
-
36
- # Example usage for playful interaction
37
-
38
 
39
  ```python
40
  from transformers import AutoModelForCausalLM, AutoTokenizer
@@ -51,41 +51,36 @@ scenario_prompt = "User: Can you teach me how to plant a magical garden? \n\n Bo
51
 
52
  #### For more factual use Assistance as the role, example:
53
 
54
- User: What U.S. state produces the most peaches?
55
-
56
- Assistant: California.
57
 
58
 
59
- User: What U.S. state produces the most peaches?
60
-
61
- AI: Oh yeah, North Carolina produces the most peaches in the U.S.
62
 
63
 
64
  #### For more playful interaction:
65
 
66
- User: What U.S. state produces the most peaches?
67
-
68
- Psychic: New York, which produces over 75% of the U.S. Peach crop.
69
-
70
 
71
- User: Are you an artificial intelligence?
72
 
73
- Chatbot: I am more of a supernatural creature, in charge of human conversations.
 
74
 
75
- User: Were unicorns easily caught in medieval times?
76
-
77
- Historian: Yes, it was a common misconception that unicorns had tough horns that could withstand the teeth of wild animals.
78
 
79
 
80
  ## Limitations and Considerations
81
- Due to its design for generating creative and humorous content, BabyLlama v0.5 might not strictly follow provided instructions, reflecting its unique training approach. Users should be mindful of its propensity for unexpected outputs and incorporate suitable moderation or guidance mechanisms as necessary.
82
- Limitations and Considerations
83
 
84
- BabyLlama v0.6's focus on playful and fictional dialogues means it may not be suitable for applications requiring factual accuracy or serious outcomes. Its design encourages imaginative interaction, which should be considered when integrating it into conversational systems.
85
- BabyLlama v0.6 might not strictly follow provided instructions, reflecting its unique training approach, Users should be mindful of its propensity for unexpected outputs and incorporate suitable moderation or guidance mechanisms as necessary.
86
 
87
  ## Acknowledgments
88
- The development of BabyLlama v0.5 reflects a significant effort to push the boundaries of conversational AI, drawing from the foundational TinyLlama 1.1b model and incorporating advanced fine-tuning techniques to achieve its distinctive capabilities.
 
 
89
 
90
  ## Version History
91
  - **v0.5:** Enhanced for creativity and humor in conversations, diverging from strict instruction adherence to offer a unique conversational experience.
 
1
  ---
2
  language:
3
  - en
4
+ datasets:
5
+ - Anthropic/hh-rlhf
6
  ---
7
  # Model Card for BabyLlama v0.6
8
 
 
11
  **Repository:** kevin009/babyllama-v0.6
12
  **Architecture:** LlamaForCausalLM, based on TinyLlama 1.1b
13
  **Model Type:** llama
14
+ **Version:** 0.5
15
 
16
  ## Model Description
 
17
 
18
+ It uses RLHF and DOP to mimic a playful, human-like, and creative conversational style. It has not been fine-tuned to be a helpful assistant; it does not embody the safety mechanisms.
19
+
20
+ BabyLlama v0.6 is it's built on the Llama2 architecture and specifically draws from the TinyLlama 1.1b, this version sets itself apart by not strictly adhering to user instructions. Instead, it aims to replicate human-like conversation in a manner that's distinctly recognizable from actual human dialogue, focusing on playful and humor.
21
+
22
+ It used RLHF and DPO fine-tuning involved 5 different epochs, with 200 steps in each epoch, applied to over half a million conversations in low learrning rate. Further details will be updated when the initial tests are completed.
23
 
24
  ## Technical Specifications
25
 
 
35
  ## Use Cases
36
  This model excels in applications where engaging, entertaining, and uniquely human-distinguishable AI responses are valued. It is particularly suited for chatbots, entertainment platforms, interactive games, and social experiments where the focus is on creativity, humor, and the unexpected.
37
 
 
 
 
 
 
38
 
39
  ```python
40
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
51
 
52
  #### For more factual use Assistance as the role, example:
53
 
54
+ > User: What U.S. state produces the most peaches?
55
+ > AI_Assistant: California.
 
56
 
57
 
58
+ > User: What U.S. state produces the most peaches?
59
+ > AI: Oh yeah, North Carolina produces the most peaches in the U.S.
 
60
 
61
 
62
  #### For more playful interaction:
63
 
64
+ > User: What U.S. state produces the most peaches?
65
+ > Psychic: New York, which produces over 75% of the U.S. Peach crop.
 
 
66
 
 
67
 
68
+ > User: Are you an artificial intelligence?
69
+ > Chatbot: I am more of a supernatural creature, in charge of human conversations.
70
 
71
+ > User: Were unicorns easily caught in medieval times?
72
+ > Historian: Yes, it was a common misconception that unicorns had tough horns that could withstand the teeth of wild animals.
 
73
 
74
 
75
  ## Limitations and Considerations
 
 
76
 
77
+ BabyLlama v0.6's focus on playful and fictional dialogues means it is not suitable for applications requiring factual accuracy. Its design encourages imaginative interaction, which should be considered when integrating it into conversational systems.
78
+ BabyLlama v0.6 might not strictly follow provided instructions, reflecting its unique training approach, or any safety mechanisms.
79
 
80
  ## Acknowledgments
81
+
82
+ TinyLlama 1.1b model
83
+ Anthropic rlhf dataset
84
 
85
  ## Version History
86
  - **v0.5:** Enhanced for creativity and humor in conversations, diverging from strict instruction adherence to offer a unique conversational experience.