Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
from transformers import pipeline, set_seed
|
2 |
import gradio as grad, random, re
|
3 |
-
|
4 |
|
5 |
gpt2_pipe = pipeline('text-generation', model='Gustavosta/MagicPrompt-Stable-Diffusion', tokenizer='gpt2')
|
6 |
with open("ideas.txt", "r") as f:
|
@@ -22,11 +22,6 @@ def generate(starting_text):
|
|
22 |
if resp != starting_text and len(resp) > (len(starting_text) + 4) and resp.endswith((":", "-", "—")) is False:
|
23 |
response_list.append(resp+'\n')
|
24 |
|
25 |
-
translator = Translator()
|
26 |
-
for i, resp in enumerate(response_list):
|
27 |
-
translated_resp = translator.translate(resp, src='id', dest='en')
|
28 |
-
response_list[i] = translated_resp.text + '\n'
|
29 |
-
|
30 |
response_end = "\n".join(response_list)
|
31 |
response_end = re.sub('[^ ]+\.[^ ]+','', response_end)
|
32 |
response_end = response_end.replace("<", "").replace(">", "")
|
@@ -35,15 +30,15 @@ def generate(starting_text):
|
|
35 |
return response_end
|
36 |
|
37 |
|
38 |
-
txt = grad.Textbox(lines=1, label="Initial Text", placeholder="Text here")
|
39 |
out = grad.Textbox(lines=4, label="Generated Prompts")
|
40 |
|
41 |
examples = []
|
42 |
for x in range(8):
|
43 |
examples.append(line[random.randrange(0, len(line))].replace("\n", "").lower().capitalize())
|
44 |
|
45 |
-
title = "
|
46 |
-
description = '
|
47 |
|
48 |
grad.Interface(fn=generate,
|
49 |
inputs=txt,
|
@@ -54,6 +49,4 @@ grad.Interface(fn=generate,
|
|
54 |
article='',
|
55 |
allow_flagging='never',
|
56 |
cache_examples=False,
|
57 |
-
theme="default").launch(enable_queue=True, debug=True)
|
58 |
-
|
59 |
-
|
|
|
1 |
from transformers import pipeline, set_seed
|
2 |
import gradio as grad, random, re
|
3 |
+
|
4 |
|
5 |
gpt2_pipe = pipeline('text-generation', model='Gustavosta/MagicPrompt-Stable-Diffusion', tokenizer='gpt2')
|
6 |
with open("ideas.txt", "r") as f:
|
|
|
22 |
if resp != starting_text and len(resp) > (len(starting_text) + 4) and resp.endswith((":", "-", "—")) is False:
|
23 |
response_list.append(resp+'\n')
|
24 |
|
|
|
|
|
|
|
|
|
|
|
25 |
response_end = "\n".join(response_list)
|
26 |
response_end = re.sub('[^ ]+\.[^ ]+','', response_end)
|
27 |
response_end = response_end.replace("<", "").replace(">", "")
|
|
|
30 |
return response_end
|
31 |
|
32 |
|
33 |
+
txt = grad.Textbox(lines=1, label="Initial Text", placeholder="English Text here")
|
34 |
out = grad.Textbox(lines=4, label="Generated Prompts")
|
35 |
|
36 |
examples = []
|
37 |
for x in range(8):
|
38 |
examples.append(line[random.randrange(0, len(line))].replace("\n", "").lower().capitalize())
|
39 |
|
40 |
+
title = "Stable Diffusion Prompt Generator"
|
41 |
+
description = 'This is a demo of the model series: "MagicPrompt", in this case, aimed at: "Stable Diffusion". To use it, simply submit your text or click on one of the examples. To learn more about the model, [click here](https://huggingface.co/Gustavosta/MagicPrompt-Stable-Diffusion).<br>'
|
42 |
|
43 |
grad.Interface(fn=generate,
|
44 |
inputs=txt,
|
|
|
49 |
article='',
|
50 |
allow_flagging='never',
|
51 |
cache_examples=False,
|
52 |
+
theme="default").launch(enable_queue=True, debug=True)
|
|
|
|