Text Generation
Transformers
PyTorch
English
Chinese
llama
code
text-generation-inference
Inference Endpoints
.gitattributes CHANGED
@@ -1,34 +1,35 @@
1
  *.7z filter=lfs diff=lfs merge=lfs -text
2
  *.arrow filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bin.* filter=lfs diff=lfs merge=lfs -text
5
  *.bz2 filter=lfs diff=lfs merge=lfs -text
 
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
 
11
  *.model filter=lfs diff=lfs merge=lfs -text
12
  *.msgpack filter=lfs diff=lfs merge=lfs -text
 
 
13
  *.onnx filter=lfs diff=lfs merge=lfs -text
14
  *.ot filter=lfs diff=lfs merge=lfs -text
15
  *.parquet filter=lfs diff=lfs merge=lfs -text
16
  *.pb filter=lfs diff=lfs merge=lfs -text
 
 
17
  *.pt filter=lfs diff=lfs merge=lfs -text
18
  *.pth filter=lfs diff=lfs merge=lfs -text
19
  *.rar filter=lfs diff=lfs merge=lfs -text
 
20
  saved_model/**/* filter=lfs diff=lfs merge=lfs -text
21
  *.tar.* filter=lfs diff=lfs merge=lfs -text
 
22
  *.tflite filter=lfs diff=lfs merge=lfs -text
23
  *.tgz filter=lfs diff=lfs merge=lfs -text
 
24
  *.xz filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *.tfevents* filter=lfs diff=lfs merge=lfs -text
28
- *.db* filter=lfs diff=lfs merge=lfs -text
29
- *.ark* filter=lfs diff=lfs merge=lfs -text
30
- **/*ckpt*data* filter=lfs diff=lfs merge=lfs -text
31
- **/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text
32
- **/*ckpt*.index filter=lfs diff=lfs merge=lfs -text
33
- *.safetensors filter=lfs diff=lfs merge=lfs -text
34
- *.ckpt filter=lfs diff=lfs merge=lfs -text
 
1
  *.7z filter=lfs diff=lfs merge=lfs -text
2
  *.arrow filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
 
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
  *.model filter=lfs diff=lfs merge=lfs -text
13
  *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
  *.onnx filter=lfs diff=lfs merge=lfs -text
17
  *.ot filter=lfs diff=lfs merge=lfs -text
18
  *.parquet filter=lfs diff=lfs merge=lfs -text
19
  *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
  *.pt filter=lfs diff=lfs merge=lfs -text
23
  *.pth filter=lfs diff=lfs merge=lfs -text
24
  *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
  saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
  *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
  *.tflite filter=lfs diff=lfs merge=lfs -text
30
  *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
  *.xz filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
LOGO.png CHANGED
README.md CHANGED
@@ -1,20 +1,8 @@
1
  ---
2
  license: other
3
- tasks:
4
- - code-generation
5
- datasets:
6
- - codefuse-ai/CodeExercise-Python-27k
7
- - codefuse-ai/Evol-instruction-66k
8
- language:
9
- - en
10
- - zh
11
- pipeline_tag: text-generation
12
- library_name: transformers
13
- tags:
14
- - code
15
  ---
16
  # Model Card for CodeFuse-CodeLlama-34B
17
- ![logo](LOGO.png)
18
 
19
  [[中文]](#chinese) [[English]](#english)
20
 
@@ -30,38 +18,17 @@ The context length of finetuning is 4K while it is able to be finetuned by 16k c
30
 
31
  ## News and Updates
32
 
33
- 🔥🔥🔥 2023-09-26 We are pleased to announce the release of the [4-bit quantized version](https://huggingface.co/codefuse-ai/CodeFuse-CodeLlama-34B-4bits) of CodeFuse-CodeLlama-34B. Despite the quantization process, the model still achieves a remarkable 73.8% accuracy (greedy decoding) on the HumanEval pass@1 metric.
34
-
35
- 🔥🔥🔥 2023-09-11 CodeFuse-CodeLlama34B has achieved 74.4% of pass@1 (greedy decoding) on HumanEval, which is SOTA results for openspurced LLMs at present.
36
 
37
  <br>
38
 
39
- ## Code Community
40
-
41
- **Homepage**: 🏡 https://github.com/codefuse-ai (**Please give us your support with a Star🌟 + Fork🚀 + Watch👀**)
42
-
43
- + If you wish to fine-tune the model yourself, you can visit ✨[MFTCoder](https://github.com/codefuse-ai/MFTCoder)✨✨
44
-
45
- + If you wish to deploy the model yourself, you can visit ✨[FasterTransformer4CodeFuse](https://github.com/codefuse-ai/FasterTransformer4CodeFuse)✨✨
46
-
47
- + If you wish to see a demo of the model, you can visit ✨[CodeFuse Demo](https://github.com/codefuse-ai/codefuse)✨✨
48
-
49
-
50
  ## Performance
51
 
52
-
53
- | Model | HumanEval(pass@1) | Date |
54
- |:----------------------------|:-----------------:|:-------:|
55
- | **CodeFuse-CodeLlama-34B** | **74.4%** | 2023.9 |
56
- | WizardCoder-Python-34B-V1.0 | 73.2% | 2023.8 |
57
- | GPT-4(zero-shot) | 67.0% | 2023.3 |
58
- | PanGu-Coder2 15B | 61.6% | 2023.8 |
59
- | CodeLlama-34b-Python | 53.7% | 2023.8 |
60
- | CodeLlama-34b | 48.8% | 2023.8 |
61
- | GPT-3.5(zero-shot) | 48.1% | 2022.11 |
62
- | OctoCoder | 46.2% | 2023.8 |
63
- | StarCoder-15B | 33.6% | 2023.5 |
64
- | LLaMA 2 70B(zero-shot) | 29.9% | 2023.7 |
65
 
66
  <br>
67
 
@@ -103,7 +70,6 @@ pip install -r requirements.txt
103
  ```
104
 
105
  ```python
106
- import torch
107
  from transformers import (
108
  AutoTokenizer,
109
  AutoModelForCausalLM,
@@ -112,13 +78,7 @@ tokenizer = AutoTokenizer.from_pretrained(mode_name_or_path, trust_remote_code=T
112
  tokenizer.padding_side = "left"
113
  tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids("<unk>")
114
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids("</s>")
115
- # try 4bit loading if cuda memory not enough
116
- model = AutoModelForCausalLM.from_pretrained(mode_name_or_path,
117
- trust_remote_code=True,
118
- load_in_4bit=False,
119
- device_map="auto",
120
- torch_dtype=torch.bfloat16)
121
- model.eval()
122
 
123
  HUMAN_ROLE_START_TAG = "<|role_start|>human<|role_end|>"
124
  BOT_ROLE_START_TAG = "<|role_start|>bot<|role_end|>"
@@ -139,32 +99,17 @@ gen_text = tokenizer.batch_decode(outputs[:, inputs["input_ids"].shape[1]:], ski
139
  print(gen_text)
140
  ```
141
 
142
- ## MD5
143
- We notice that the file may be corrupted during transfer process. Please check MD5 value before use.
144
 
145
- | Model File | MD5 Value |
146
- |:---------------------------------|:--------------------------------:|
147
- | pytorch_model-00001-of-00007.bin | 8d544b1bcb3449934184d4141137329c |
148
- | pytorch_model-00002-of-00007.bin | 9d5dbb30911e48a42fb6d0fcabb322a4 |
149
- | pytorch_model-00003-of-00007.bin | b0d4aecee0457d9332005a187e1fffed |
150
- | pytorch_model-00004-of-00007.bin | 5c7e002de5eab77d0194a2b0f6de0c24 |
151
- | pytorch_model-00005-of-00007.bin | d22a511aa26b5b17117b665a877490ab |
152
- | pytorch_model-00006-of-00007.bin | a5c28ac277fac07d16dd66537e54d109 |
153
- | pytorch_model-00007-of-00007.bin | a967e2c6195477b7407089c0bffa2d53 |
154
 
155
 
156
- ## Citation
157
- If you find our [work](https://arxiv.org/abs/2311.02303) useful or helpful for your R&D works, please feel free to cite our paper as below.
158
- ```
159
- @article{mftcoder2023,
160
- title={MFTCoder: Boosting Code LLMs with Multitask Fine-Tuning},
161
- author={Bingchang Liu and Chaoyu Chen and Cong Liao and Zi Gong and Huan Wang and Zhichao Lei and Ming Liang and Dajun Chen and Min Shen and Hailian Zhou and Hang Yu and Jianguo Li},
162
- year={2023},
163
- journal={arXiv preprint arXiv},
164
- archivePrefix={arXiv},
165
- eprint={2311.02303}
166
- }
167
- ```
168
 
169
  <a id="chinese"></a>
170
 
@@ -179,31 +124,13 @@ CodeFuse-CodeLlama34B-MFT 是一个通过QLoRA对基座模型CodeLlama-34b-Pytho
179
 
180
  <br>
181
 
182
- ## 代码社区
183
- **大本营**: 🏡 https://github.com/codefuse-ai (**欢迎为我们的项目一键三连 Star🌟 + Fork🚀 + Watch👀**)
184
-
185
- + 如果您想自己微调该模型,可以访问 ✨[MFTCoder](https://github.com/codefuse-ai/MFTCoder)✨✨
186
-
187
- + 如果您想自己部署该模型,可以访问 ✨[FasterTransformer4CodeFuse](https://github.com/codefuse-ai/FasterTransformer4CodeFuse)✨✨
188
-
189
- + 如果您想观看该模型示例,可以访问 ✨[CodeFuse Demo](https://github.com/codefuse-ai/codefuse)✨✨
190
-
191
-
192
  ## 评测表现(代码)
193
 
194
- | 模型 | HumanEval(pass@1) | 日期 |
195
- |:----------------------------|:-----------------:|:-------:|
196
- | **CodeFuse-CodeLlama-34B** | **74.4%** | 2023.9 |
197
- | WizardCoder-Python-34B-V1.0 | 73.2% | 2023.8 |
198
- | GPT-4(zero-shot) | 67.0% | 2023.3 |
199
- | PanGu-Coder2 15B | 61.6% | 2023.8 |
200
- | CodeLlama-34b-Python | 53.7% | 2023.8 |
201
- | CodeLlama-34b | 48.8% | 2023.8 |
202
- | GPT-3.5(zero-shot) | 48.1% | 2022.11 |
203
- | OctoCoder | 46.2% | 2023.8 |
204
- | StarCoder-15B | 33.6% | 2023.5 |
205
- | LLaMA 2 70B(zero-shot) | 29.9% | 2023.7 |
206
-
207
  <br>
208
 
209
  ## Requirements
@@ -246,18 +173,12 @@ tokenizer = AutoTokenizer.from_pretrained(mode_name_or_path, trust_remote_code=T
246
  tokenizer.padding_side = "left"
247
  tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids("<unk>")
248
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids("</s>")
249
- # 如果显存不够,可以考虑量化加载
250
- model = AutoModelForCausalLM.from_pretrained(mode_name_or_path,
251
- trust_remote_code=True,
252
- load_in_4bit=False,
253
- device_map="auto",
254
- torch_dtype=torch.bfloat16)
255
- model.eval()
256
 
257
  HUMAN_ROLE_START_TAG = "<|role_start|>human<|role_end|>"
258
  BOT_ROLE_START_TAG = "<|role_start|>bot<|role_end|>"
259
 
260
- text = f"{HUMAN_ROLE_START_TAG}请用C++实现求解第n个斐波那契数{BOT_ROLE_START_TAG}"
261
  inputs = tokenizer(text, return_tensors='pt', padding=True, add_special_tokens=False).to("cuda")
262
  outputs = model.generate(
263
  inputs=inputs["input_ids"],
@@ -271,18 +192,4 @@ outputs = model.generate(
271
  )
272
  gen_text = tokenizer.batch_decode(outputs[:, inputs["input_ids"].shape[1]:], skip_special_tokens=True)
273
  print(gen_text)
274
- ```
275
-
276
-
277
- ## MD5
278
- 我们发现模型文件可能会在传输过程中损坏,使用前请检查文件MD5值。
279
-
280
- | 模型文件 | MD5值 |
281
- |:---------------------------------|:--------------------------------:|
282
- | pytorch_model-00001-of-00007.bin | 8d544b1bcb3449934184d4141137329c |
283
- | pytorch_model-00002-of-00007.bin | 9d5dbb30911e48a42fb6d0fcabb322a4 |
284
- | pytorch_model-00003-of-00007.bin | b0d4aecee0457d9332005a187e1fffed |
285
- | pytorch_model-00004-of-00007.bin | 5c7e002de5eab77d0194a2b0f6de0c24 |
286
- | pytorch_model-00005-of-00007.bin | d22a511aa26b5b17117b665a877490ab |
287
- | pytorch_model-00006-of-00007.bin | a5c28ac277fac07d16dd66537e54d109 |
288
- | pytorch_model-00007-of-00007.bin | a967e2c6195477b7407089c0bffa2d53 |
 
1
  ---
2
  license: other
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
4
  # Model Card for CodeFuse-CodeLlama-34B
5
+ ![Creation Approach](LOGO.png)
6
 
7
  [[中文]](#chinese) [[English]](#english)
8
 
 
18
 
19
  ## News and Updates
20
 
21
+ 🔥🔥🔥 CodeFuse-CodeLlama34B-MFT has achived 74.4% of pass@1 on HumanEval, which is SOTA at present.
 
 
22
 
23
  <br>
24
 
 
 
 
 
 
 
 
 
 
 
 
25
  ## Performance
26
 
27
+ | Model | HumanEval(pass@1) |
28
+ | :---------------------------- | :---------------: |
29
+ | CodeLlama-34b | 48.8%(greedy decoding) |
30
+ | CodeLlama-34b-Python | 53.7%(greedy decoding) |
31
+ | **CodeFuse-CodeLlama-34B** | **74.4%**(greedy decoding) |
 
 
 
 
 
 
 
 
32
 
33
  <br>
34
 
 
70
  ```
71
 
72
  ```python
 
73
  from transformers import (
74
  AutoTokenizer,
75
  AutoModelForCausalLM,
 
78
  tokenizer.padding_side = "left"
79
  tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids("<unk>")
80
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids("</s>")
81
+ model = AutoModelForCausalLM.from_pretrained(mode_name_or_path, trust_remote_code=True)
 
 
 
 
 
 
82
 
83
  HUMAN_ROLE_START_TAG = "<|role_start|>human<|role_end|>"
84
  BOT_ROLE_START_TAG = "<|role_start|>bot<|role_end|>"
 
99
  print(gen_text)
100
  ```
101
 
 
 
102
 
 
 
 
 
 
 
 
 
 
103
 
104
 
105
+
106
+
107
+
108
+
109
+
110
+
111
+
112
+
 
 
 
 
113
 
114
  <a id="chinese"></a>
115
 
 
124
 
125
  <br>
126
 
 
 
 
 
 
 
 
 
 
 
127
  ## 评测表现(代码)
128
 
129
+ | 模型 | HumanEval(pass@1) |
130
+ | :---------------------------- | :---------------: |
131
+ | CodeLlama-34b | 48.8%(greedy decoding) |
132
+ | CodeLlama-34b-Python | 53.7%(greedy decoding) |
133
+ | **CodeFuse-CodeLlama-34B** | **74.4%**(greedy decoding) |
 
 
 
 
 
 
 
 
134
  <br>
135
 
136
  ## Requirements
 
173
  tokenizer.padding_side = "left"
174
  tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids("<unk>")
175
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids("</s>")
176
+ model = AutoModelForCausalLM.from_pretrained(mode_name_or_path, trust_remote_code=True)
 
 
 
 
 
 
177
 
178
  HUMAN_ROLE_START_TAG = "<|role_start|>human<|role_end|>"
179
  BOT_ROLE_START_TAG = "<|role_start|>bot<|role_end|>"
180
 
181
+ text = f"{HUMAN_ROLE_START_TAG}write a python function of quick sort.{BOT_ROLE_START_TAG}"
182
  inputs = tokenizer(text, return_tensors='pt', padding=True, add_special_tokens=False).to("cuda")
183
  outputs = model.generate(
184
  inputs=inputs["input_ids"],
 
192
  )
193
  gen_text = tokenizer.batch_decode(outputs[:, inputs["input_ids"].shape[1]:], skip_special_tokens=True)
194
  print(gen_text)
195
+ ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json DELETED
@@ -1,29 +0,0 @@
1
- {
2
- "_name_or_path": "/mnt/user/qumu/download_models/codellama/CodeLlama-34b-Python-hf",
3
- "architectures": [
4
- "LlamaForCausalLM"
5
- ],
6
- "bos_token_id": 1,
7
- "eos_token": "</s>",
8
- "eos_token_id": 2,
9
- "hidden_act": "silu",
10
- "hidden_size": 8192,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 22016,
13
- "max_position_embeddings": 16384,
14
- "model_type": "llama",
15
- "num_attention_heads": 64,
16
- "num_hidden_layers": 48,
17
- "num_key_value_heads": 8,
18
- "pad_token": "<unk>",
19
- "pad_token_id": 0,
20
- "pretraining_tp": 1,
21
- "rms_norm_eps": 1e-05,
22
- "rope_scaling": null,
23
- "rope_theta": 1000000,
24
- "tie_word_embeddings": false,
25
- "torch_dtype": "bfloat16",
26
- "transformers_version": "4.32.0",
27
- "use_cache": true,
28
- "vocab_size": 32000
29
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generation_config.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "bos_token_id": 1,
4
- "eos_token_id": 2,
5
- "transformers_version": "4.32.0"
6
- }
 
 
 
 
 
 
 
pytorch_model-00001-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb1cd3c944b6dd2ece21c09d7378a77ccc279867dfe4b8b8b7cbfa58e4ffe4a3
3
- size 9852638393
 
 
 
 
pytorch_model-00002-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1641f834b6d3938353873d95b642d0083ca806b7fd32dc8aaea0b3c64752b358
3
- size 9689094033
 
 
 
 
pytorch_model-00003-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:38ce1d749b110332ffbfc237c3271c10229f0913ed3e424aa61c4261af033357
3
- size 9689094033
 
 
 
 
pytorch_model-00004-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5c7a36d9f48fdd4b629f44d6e238085fa727f393c94e6cdf17d4f5da19be613
3
- size 9689094033
 
 
 
 
pytorch_model-00005-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ccdaa07cce1d16ed6d8ca1ca773b075f61c36d04430f8683b6e2bf458be4a72
3
- size 9689094033
 
 
 
 
pytorch_model-00006-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a257280910a701168406cc2114a669a975176ec08a373f7c2b7ba94d22c05215
3
- size 9689094033
 
 
 
 
pytorch_model-00007-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:aaef8ce33ccbc255d78026458d81fc9a55513a0d1ac3ef4f4f74967300af49e7
3
- size 9189986713
 
 
 
 
pytorch_model.bin.index.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a714170172282cfbcaa120af13c0df08b06d040ff24dab30229d8a010821d3d
3
- size 35838
 
 
 
 
requirements.txt DELETED
@@ -1,14 +0,0 @@
1
- numpy
2
- pandas
3
- einops
4
- sentencepiece
5
- deepspeed==0.9.3
6
- transformers==4.32.0
7
- accelerate==0.21.0
8
- peft==0.4.0
9
- BitsAndBytes==0.40.2
10
- xformers==0.0.21
11
- ujson
12
- jsonlines
13
- tiktoken
14
- transformers_stream_generator
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
special_tokens_map.json DELETED
@@ -1,23 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "unk_token": {
17
- "content": "<unk>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- }
23
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723
 
 
 
 
tokenizer_config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "bos_token": {
3
- "__type": "AddedToken",
4
- "content": "<s>",
5
- "lstrip": false,
6
- "normalized": true,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- "clean_up_tokenization_spaces": false,
11
- "eos_token": {
12
- "__type": "AddedToken",
13
- "content": "</s>",
14
- "lstrip": false,
15
- "normalized": true,
16
- "rstrip": false,
17
- "single_word": false
18
- },
19
- "legacy": null,
20
- "model_max_length": 1000000000000000019884624838656,
21
- "pad_token": null,
22
- "sp_model_kwargs": {},
23
- "tokenizer_class": "LlamaTokenizer",
24
- "unk_token": {
25
- "__type": "AddedToken",
26
- "content": "<unk>",
27
- "lstrip": false,
28
- "normalized": true,
29
- "rstrip": false,
30
- "single_word": false
31
- },
32
- "use_default_system_prompt": true
33
- }