Update README.md
Browse files
README.md
CHANGED
@@ -1,16 +1,15 @@
|
|
1 |
-
---
|
2 |
-
language:
|
3 |
-
- en
|
4 |
-
tags:
|
5 |
-
- code
|
6 |
-
---
|
7 |
# Quantized_by: Zeeshan
|
|
|
|
|
8 |
# Tinyllama 1.1B Chat v0.3 - GGUF
|
9 |
- Model creator: [TinyLlama](https://huggingface.co/TinyLlama)
|
10 |
- Original model: [Tinyllama 1.1B Chat v0.3](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v0.3)
|
|
|
11 |
<!-- description start -->
|
12 |
## Description
|
|
|
13 |
This repo contains GGUF format model files for [TinyLlama's Tinyllama 1.1B Chat v0.3](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v0.3).
|
|
|
14 |
<!-- description end -->
|
15 |
<!-- README_GGUF.md-about-gguf start -->
|
16 |
### About GGUF
|
@@ -32,6 +31,8 @@ Here is an incomplete list of clients and libraries that are known to support GG
|
|
32 |
|
33 |
<!-- README_GGUF.md-about-gguf end -->
|
34 |
<!-- repositories-available start -->
|
|
|
|
|
35 |
<!-- README_GGUF.md-how-to-download start -->
|
36 |
## How to download GGUF files
|
37 |
|
@@ -74,9 +75,12 @@ Do check the [TinyLlama](https://github.com/jzhang38/TinyLlama) github page for
|
|
74 |
# Install transformers from source - only needed for versions <= v4.34
|
75 |
# pip install git+https://github.com/huggingface/transformers.git
|
76 |
# pip install accelerate
|
|
|
77 |
import torch
|
78 |
from transformers import pipeline
|
|
|
79 |
pipe = pipeline("text-generation", model="TinyLlama/TinyLlama-1.1B-Chat-v0.3", torch_dtype=torch.bfloat16, device_map="auto")
|
|
|
80 |
# We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
|
81 |
messages = [
|
82 |
{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
# Quantized_by: Zeeshan
|
2 |
+
|
3 |
+
|
4 |
# Tinyllama 1.1B Chat v0.3 - GGUF
|
5 |
- Model creator: [TinyLlama](https://huggingface.co/TinyLlama)
|
6 |
- Original model: [Tinyllama 1.1B Chat v0.3](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v0.3)
|
7 |
+
|
8 |
<!-- description start -->
|
9 |
## Description
|
10 |
+
|
11 |
This repo contains GGUF format model files for [TinyLlama's Tinyllama 1.1B Chat v0.3](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v0.3).
|
12 |
+
|
13 |
<!-- description end -->
|
14 |
<!-- README_GGUF.md-about-gguf start -->
|
15 |
### About GGUF
|
|
|
31 |
|
32 |
<!-- README_GGUF.md-about-gguf end -->
|
33 |
<!-- repositories-available start -->
|
34 |
+
|
35 |
+
|
36 |
<!-- README_GGUF.md-how-to-download start -->
|
37 |
## How to download GGUF files
|
38 |
|
|
|
75 |
# Install transformers from source - only needed for versions <= v4.34
|
76 |
# pip install git+https://github.com/huggingface/transformers.git
|
77 |
# pip install accelerate
|
78 |
+
|
79 |
import torch
|
80 |
from transformers import pipeline
|
81 |
+
|
82 |
pipe = pipeline("text-generation", model="TinyLlama/TinyLlama-1.1B-Chat-v0.3", torch_dtype=torch.bfloat16, device_map="auto")
|
83 |
+
|
84 |
# We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
|
85 |
messages = [
|
86 |
{
|