|
--- |
|
license: mit |
|
datasets: |
|
- daily_dialog |
|
- multi_woz_v22 |
|
language: |
|
- en |
|
--- |
|
### Useless ChitChat Language Model |
|
|
|
Basic Dialog Model from DialoGPT-small. |
|
Finetuned on Dialog dataset. (Daily Dialog, MultiWoz) |
|
|
|
### How to use |
|
|
|
Use it as any torch python Language Model |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
import torch |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("gpt") |
|
model = AutoModelForCausalLM.from_pretrained("jinymusim/dialogmodel") |
|
|
|
# Take user Input |
|
user_utterance = input('USER> ') |
|
user_utterance = user_utterance.strip() |
|
tokenized_context = tokenizer.encode(user_utterance + tokenizer.eos_token, return_tensors='pt') |
|
|
|
# generated a response, limit max_lenght to resonable size |
|
out_response = model.generate(tokenized_context, |
|
max_length=100, |
|
num_beams=2, |
|
no_repeat_ngram_size=2, |
|
early_stopping=True, |
|
pad_token_id=self.tokenizer.eos_token_id) |
|
|
|
# Truncate User Input |
|
decoded_response = self.tokenizer.decode(out_response[0], skip_special_tokens=True)[len(user_utterance):] |
|
|
|
print(f'SYSTEM> {decoded_response}') |
|
``` |