afrideva commited on
Commit
67b52da
1 Parent(s): feaed84

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +92 -0
README.md ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: tog/TinyLlama-1.1B-alpaca-chat-v1.5
3
+ datasets:
4
+ - tatsu-lab/alpaca
5
+ inference: false
6
+ language:
7
+ - en
8
+ license: apache-2.0
9
+ model_creator: tog
10
+ model_name: TinyLlama-1.1B-alpaca-chat-v1.5
11
+ pipeline_tag: text-generation
12
+ quantized_by: afrideva
13
+ tags:
14
+ - gguf
15
+ - ggml
16
+ - quantized
17
+ - q2_k
18
+ - q3_k_m
19
+ - q4_k_m
20
+ - q5_k_m
21
+ - q6_k
22
+ - q8_0
23
+ widget:
24
+ - text: '###Instruction:\nWhat is a large language model? Be concise\n\n### Response:\n'
25
+ ---
26
+ # tog/TinyLlama-1.1B-alpaca-chat-v1.5-GGUF
27
+
28
+ Quantized GGUF model files for [TinyLlama-1.1B-alpaca-chat-v1.5](https://huggingface.co/tog/TinyLlama-1.1B-alpaca-chat-v1.5) from [tog](https://huggingface.co/tog)
29
+
30
+
31
+ | Name | Quant method | Size |
32
+ | ---- | ---- | ---- |
33
+ | [tinyllama-1.1b-alpaca-chat-v1.5.q2_k.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-alpaca-chat-v1.5-GGUF/resolve/main/tinyllama-1.1b-alpaca-chat-v1.5.q2_k.gguf) | q2_k | 482.14 MB |
34
+ | [tinyllama-1.1b-alpaca-chat-v1.5.q3_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-alpaca-chat-v1.5-GGUF/resolve/main/tinyllama-1.1b-alpaca-chat-v1.5.q3_k_m.gguf) | q3_k_m | 549.85 MB |
35
+ | [tinyllama-1.1b-alpaca-chat-v1.5.q4_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-alpaca-chat-v1.5-GGUF/resolve/main/tinyllama-1.1b-alpaca-chat-v1.5.q4_k_m.gguf) | q4_k_m | 667.81 MB |
36
+ | [tinyllama-1.1b-alpaca-chat-v1.5.q5_k_m.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-alpaca-chat-v1.5-GGUF/resolve/main/tinyllama-1.1b-alpaca-chat-v1.5.q5_k_m.gguf) | q5_k_m | 782.04 MB |
37
+ | [tinyllama-1.1b-alpaca-chat-v1.5.q6_k.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-alpaca-chat-v1.5-GGUF/resolve/main/tinyllama-1.1b-alpaca-chat-v1.5.q6_k.gguf) | q6_k | 903.41 MB |
38
+ | [tinyllama-1.1b-alpaca-chat-v1.5.q8_0.gguf](https://huggingface.co/afrideva/TinyLlama-1.1B-alpaca-chat-v1.5-GGUF/resolve/main/tinyllama-1.1b-alpaca-chat-v1.5.q8_0.gguf) | q8_0 | 1.17 GB |
39
+
40
+
41
+
42
+ ## Original Model Card:
43
+ ## This Model
44
+
45
+ This is the chat model finetuned on top of [PY007/TinyLlama-1.1B-intermediate-step-715k-1.5T](https://huggingface.co/PY007/TinyLlama-1.1B-intermediate-step-715k-1.5T). The dataset used is [tatsu-lab/stanford_alpaca](https://github.com/tatsu-lab/stanford_alpaca).
46
+
47
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
48
+ ```
49
+ ### Instruction:
50
+ {instruction}
51
+
52
+ ### Response:
53
+ ```
54
+
55
+ You can use it with the `transformers` library:
56
+
57
+ ```python
58
+ from transformers import AutoTokenizer
59
+ import transformers
60
+ import torch
61
+
62
+ model = "tog/TinyLlama-1.1B-alpaca-chat-v1.5"
63
+ tokenizer = AutoTokenizer.from_pretrained(model)
64
+
65
+ pipeline = transformers.pipeline(
66
+ "text-generation",
67
+ model=model,
68
+ torch_dtype=torch.float16,
69
+ device_map="auto")
70
+
71
+ sequences = pipeline(
72
+ '###Instruction:\nWhat is a large language model? Be concise.\n\n### Response:\n',
73
+ do_sample=True,
74
+ top_k=10,
75
+ num_return_sequences=1,
76
+ eos_token_id=tokenizer.eos_token_id,
77
+ max_length=200)
78
+
79
+ for seq in sequences:
80
+ print(f"{seq['generated_text']}")
81
+ ```
82
+
83
+ You should get something along those lines:
84
+
85
+ ```
86
+ Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
87
+ Result: ###Instruction:
88
+ What is a large language model? Be concise.
89
+
90
+ ### Response:
91
+ A large language model is a type of natural language understanding model that can learn to accurately recognize and interpret text data by understanding the context of words. Languages used for text understanding are typically trained on a corpus of text data.
92
+ ```