File size: 1,858 Bytes
8a830b4
 
efb732c
 
 
 
 
 
 
 
 
 
 
 
8a830b4
efb732c
 
 
 
 
465c40d
 
 
efb732c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
---
license: apache-2.0
datasets:
- hakurei/open-instruct-v1
language:
- en
tags:
- code
- instruction-following
widget:
  - text: Tell me how to bake a cake
    example_title: Baking cakes
  - text: How can I print a fibonacci series upto N in C++
    example_title: Coding
---

# DialoGPT2 Instruction Following

This is the fine-tuned version of the [microsoft/dialogpt-small](https://huggingface.co/microsoft/DialoGPT-small) on the instruction following task. The dataset used was the [hakurei/open-instruct-v1](https://huggingface.co/datasets/hakurei/open-instruct-v1) dataset.

Find the training notebook here on [Kaggle](https://www.kaggle.com/code/smjishanulislam/basic-guide-on-instruction-following-dialogpt).


## Using the model


### Using `model.generate()`

To use the model, first call the checkpoints and initialize the model

```python
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("smji/dialogpt2-instruct-following")
model = AutoModelForCausalLM.from_pretrained("smji/dialogpt2-instruct-following")
```

And then move onto generating the text

```python
def generate_text(prompt):
    inputs = tokenizer.encode(prompt, return_tensors='pt').to(device)
    outputs = model.generate(inputs, max_length=512, pad_token_id=tokenizer.eos_token_id)
    generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)

    return generated_text[:generated_text.rfind('.')+1]

generate_text("How can I bake a cake?")
```

### Using the pipeline

Or, you can also use the pipeline

```python
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="smji/dialogpt2-instruct-following")

pipe("How can I bake a cake?", max_length=512)
```

---

Done by [S M Jishanul Islam](https://github.com/S-M-J-I)