deepnight-research commited on
Commit
e50f20a
1 Parent(s): e8830ca

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +68 -0
README.md ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - 7B
9
+ - Saily
10
+ - DEEPNIGHT
11
+ - Llama
12
+ - Llama2
13
+ ---
14
+
15
+ # SaiLy 7B (deepnight-research/saily-7b-v0)
16
+ <img src="https://i.ibb.co/TvZQjZM/Leonardo-Diffusion-XL-Furious-and-strong-Elephant-and-anchor-l-1.jpg" alt="Saily: Experimental AI Models by DEEPNIGHT">
17
+
18
+ ---
19
+ ### SaiLy is a series/collection of AI Models by DEEPNIGHT-RESEARCH which are highly experimental and uncensored. Please use with responsibility.
20
+ ---
21
+ <br>
22
+ Prompt Template: Alpeca
23
+
24
+ ```
25
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
26
+ ### Instruction:
27
+ {prompt}
28
+ ### Response:
29
+ ```
30
+
31
+ ### Description
32
+ This is the first model of the series. The model is based on Llama2-chat.
33
+
34
+ ---
35
+ ### Did some said CODE?
36
+ Here you go!
37
+ ```python
38
+ import transformers
39
+ model = transformers.AutoModelForCausalLM.from_pretrained(
40
+ 'deepnight-research/saily-7b-v0'
41
+ )
42
+ ```
43
+
44
+ To use the optimized triton implementation of FlashAttention, you can load the model on GPU ```(cuda:0)``` with ```attn_impl='triton'``` and with ```bfloat16``` precision:
45
+ ```python
46
+ import torch
47
+ import transformers
48
+
49
+ name = 'deepnight-research/saily-7b-v0'
50
+
51
+ config = transformers.AutoConfig.from_pretrained(name)
52
+ config.attn_config['attn_impl'] = 'triton'
53
+ config.init_device = 'cuda:0' # For fast initialization directly on GPU!
54
+
55
+ model = transformers.AutoModelForCausalLM.from_pretrained(
56
+ name,
57
+ config=config,
58
+ torch_dtype=torch.bfloat16, # Load model weights in bfloat16
59
+ trust_remote_code=True
60
+ )
61
+
62
+ ```
63
+ ---
64
+
65
+ If you would like to support us, please consider donating for [#aiforcause](https://github.com/deepnight-ai/aiforcause).
66
+
67
+ Cheers✌️
68
+ - Team [DEEPNIGHT](https://deepnight.tech)