tinyllama-dolly-15k / README.md
Bhooyas's picture
Update README.md
8cf284d verified
metadata
license: apache-2.0
datasets:
  - databricks/databricks-dolly-15k
language:
  - en
base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
pipeline_tag: text-generation

TinyLlama Dolly 15

The TinyLlama Dolly 15k is a specialized large language model finely tuned on the Databricks Dolly 15k dataset. This dataset, composed of 15,000 high-quality, human-curated prompts and responses, helps the model excel in understanding and generating contextually relevant and coherent text. By leveraging this diverse and comprehensive dataset, TinyLlama Dolly 15k enhances its ability to engage in nuanced conversations and provide accurate, context-aware responses across a range of topics. The result is a powerful tool for applications requiring advanced natural language understanding and generation.

How to use

Below is a snippet that can be used to test the model.

from transformers import pipeline

pipe = pipeline("text-generation", model="Bhooyas/tinyllama-dolly-15k", device_map="auto")

messages = [
    {
        "role": "user",
        "context": "The TinyLlama Dolly 15k is a specialized large language model finely tuned on the Databricks Dolly 15k dataset. This dataset, composed of 15,000 high-quality, human-curated prompts and responses, helps the model excel in understanding and generating contextually relevant and coherent text. By leveraging this diverse and comprehensive dataset, TinyLlama Dolly 15k enhances its ability to engage in nuanced conversations and provide accurate, context-aware responses across a range of topics. The result is a powerful tool for applications requiring advanced natural language understanding and generation.",
        "instruction": "What is TinyLlama Dolly 15k?"
    }
]

prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipe(prompt, max_new_tokens=256)
print(outputs[0]["generated_text"])