Upload README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,17 @@ license: apache-2.0
|
|
5 |
model_creator: Intel
|
6 |
model_name: Neural Chat 7B v3-1
|
7 |
model_type: mistral
|
8 |
-
prompt_template: '
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
|
10 |
'
|
11 |
quantized_by: TheBloke
|
@@ -53,11 +63,17 @@ These files were quantised using hardware kindly provided by [Massed Compute](ht
|
|
53 |
<!-- repositories-available end -->
|
54 |
|
55 |
<!-- prompt-template start -->
|
56 |
-
## Prompt template:
|
57 |
|
58 |
```
|
|
|
|
|
|
|
|
|
59 |
{prompt}
|
60 |
|
|
|
|
|
61 |
```
|
62 |
|
63 |
<!-- prompt-template end -->
|
@@ -226,7 +242,13 @@ from huggingface_hub import InferenceClient
|
|
226 |
endpoint_url = "https://your-endpoint-url-here"
|
227 |
|
228 |
prompt = "Tell me about AI"
|
229 |
-
prompt_template=f'''
|
|
|
|
|
|
|
|
|
|
|
|
|
230 |
'''
|
231 |
|
232 |
client = InferenceClient(endpoint_url)
|
@@ -282,7 +304,13 @@ model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
|
|
282 |
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
283 |
|
284 |
prompt = "Tell me about AI"
|
285 |
-
prompt_template=f'''
|
|
|
|
|
|
|
|
|
|
|
|
|
286 |
'''
|
287 |
|
288 |
print("\n\n*** Generate:")
|
@@ -357,9 +385,9 @@ And thank you again to a16z for their generous grant.
|
|
357 |
# Original model card: Intel's Neural Chat 7B v3-1
|
358 |
|
359 |
|
360 |
-
##
|
361 |
|
362 |
-
This model is a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the open source dataset [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca). Then we align it with DPO algorithm. For more details, you can refer our blog: [
|
363 |
|
364 |
## Model date
|
365 |
Neural-chat-7b-v3-1 was trained between September and October, 2023.
|
@@ -389,10 +417,22 @@ The following hyperparameters were used during training:
|
|
389 |
- total_train_batch_size: 64
|
390 |
- total_eval_batch_size: 8
|
391 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
392 |
-
- lr_scheduler_type:
|
393 |
-
- lr_scheduler_warmup_ratio: 0.
|
394 |
- num_epochs: 2.0
|
395 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
396 |
## Inference with transformers
|
397 |
|
398 |
```shell
|
@@ -418,4 +458,3 @@ The NeuralChat team with members from Intel/SATG/AIA/AIPT. Core team members: Ka
|
|
418 |
## Useful links
|
419 |
* Intel Neural Compressor [link](https://github.com/intel/neural-compressor)
|
420 |
* Intel Extension for Transformers [link](https://github.com/intel/intel-extension-for-transformers)
|
421 |
-
* Intel Extension for PyTorch [link](https://github.com/intel/intel-extension-for-pytorch)
|
|
|
5 |
model_creator: Intel
|
6 |
model_name: Neural Chat 7B v3-1
|
7 |
model_type: mistral
|
8 |
+
prompt_template: '### System:
|
9 |
+
|
10 |
+
{system_message}
|
11 |
+
|
12 |
+
|
13 |
+
### User:
|
14 |
+
|
15 |
+
{prompt}
|
16 |
+
|
17 |
+
|
18 |
+
### Assistant:
|
19 |
|
20 |
'
|
21 |
quantized_by: TheBloke
|
|
|
63 |
<!-- repositories-available end -->
|
64 |
|
65 |
<!-- prompt-template start -->
|
66 |
+
## Prompt template: Orca-Hashes
|
67 |
|
68 |
```
|
69 |
+
### System:
|
70 |
+
{system_message}
|
71 |
+
|
72 |
+
### User:
|
73 |
{prompt}
|
74 |
|
75 |
+
### Assistant:
|
76 |
+
|
77 |
```
|
78 |
|
79 |
<!-- prompt-template end -->
|
|
|
242 |
endpoint_url = "https://your-endpoint-url-here"
|
243 |
|
244 |
prompt = "Tell me about AI"
|
245 |
+
prompt_template=f'''### System:
|
246 |
+
{system_message}
|
247 |
+
|
248 |
+
### User:
|
249 |
+
{prompt}
|
250 |
+
|
251 |
+
### Assistant:
|
252 |
'''
|
253 |
|
254 |
client = InferenceClient(endpoint_url)
|
|
|
304 |
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
305 |
|
306 |
prompt = "Tell me about AI"
|
307 |
+
prompt_template=f'''### System:
|
308 |
+
{system_message}
|
309 |
+
|
310 |
+
### User:
|
311 |
+
{prompt}
|
312 |
+
|
313 |
+
### Assistant:
|
314 |
'''
|
315 |
|
316 |
print("\n\n*** Generate:")
|
|
|
385 |
# Original model card: Intel's Neural Chat 7B v3-1
|
386 |
|
387 |
|
388 |
+
## Fine-tuning on [Habana](https://habana.ai/) Gaudi2
|
389 |
|
390 |
+
This model is a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the open source dataset [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca). Then we align it with DPO algorithm. For more details, you can refer our blog: [The Practice of Supervised Fine-tuning and Direct Preference Optimization on Habana Gaudi2](https://medium.com/@NeuralCompressor/the-practice-of-supervised-finetuning-and-direct-preference-optimization-on-habana-gaudi2-a1197d8a3cd3).
|
391 |
|
392 |
## Model date
|
393 |
Neural-chat-7b-v3-1 was trained between September and October, 2023.
|
|
|
417 |
- total_train_batch_size: 64
|
418 |
- total_eval_batch_size: 8
|
419 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
420 |
+
- lr_scheduler_type: cosine
|
421 |
+
- lr_scheduler_warmup_ratio: 0.03
|
422 |
- num_epochs: 2.0
|
423 |
|
424 |
+
## Prompt Template
|
425 |
+
|
426 |
+
```
|
427 |
+
### System:
|
428 |
+
{system}
|
429 |
+
### User:
|
430 |
+
{usr}
|
431 |
+
### Assistant:
|
432 |
+
|
433 |
+
```
|
434 |
+
|
435 |
+
|
436 |
## Inference with transformers
|
437 |
|
438 |
```shell
|
|
|
458 |
## Useful links
|
459 |
* Intel Neural Compressor [link](https://github.com/intel/neural-compressor)
|
460 |
* Intel Extension for Transformers [link](https://github.com/intel/intel-extension-for-transformers)
|
|