Spaces:
Runtime error
Runtime error
Fix v3
#3
by
euclaise
- opened
app.py
CHANGED
@@ -26,7 +26,7 @@ def generate_text(usertitle, content, temperature, max_length, N=3):
|
|
26 |
# 'title': usertitle,
|
27 |
# 'content': content
|
28 |
# }
|
29 |
-
input_text = f"[[[
|
30 |
inputs = tokenizer(input_text, return_tensors='pt').to('cuda')
|
31 |
attention_mask = torch.ones(inputs['input_ids'].shape, dtype=torch.long, device='cuda')
|
32 |
generated_sequences = model.generate(inputs['input_ids'], attention_mask=attention_mask, temperature=temperature, max_length=max_length, pad_token_id=tokenizer.eos_token_id, num_return_sequences=N, do_sample=True)
|
|
|
26 |
# 'title': usertitle,
|
27 |
# 'content': content
|
28 |
# }
|
29 |
+
input_text = f"[[[Title]]] {usertitle}\n[[[Content]]] {content.strip()}\n\nThe following is an interaction between a user and an AI assistant that is related to the above text.\n\n[[[User]]] "
|
30 |
inputs = tokenizer(input_text, return_tensors='pt').to('cuda')
|
31 |
attention_mask = torch.ones(inputs['input_ids'].shape, dtype=torch.long, device='cuda')
|
32 |
generated_sequences = model.generate(inputs['input_ids'], attention_mask=attention_mask, temperature=temperature, max_length=max_length, pad_token_id=tokenizer.eos_token_id, num_return_sequences=N, do_sample=True)
|