bowphs's picture
Update app.py
ffac739
raw
history blame contribute delete
No virus
1.46 kB
import gradio as gr
from transformers import pipeline, AutoTokenizer, AutoModelForSeq2SeqLM
import os
AUTH_TOKEN = os.environ.get("TOKEN")
model = AutoModelForSeq2SeqLM.from_pretrained('bowphs/ancient-t5-translation', use_auth_token=AUTH_TOKEN)
tokenizer = AutoTokenizer.from_pretrained('bowphs/ancient-t5-translation', use_auth_token=AUTH_TOKEN)
generator = pipeline('text2text-generation', model=model, tokenizer=tokenizer)
def generate(text, generation_args):
arguments = {}
if generation_args:
pairs = generation_args.split(",")
for pair in pairs:
key, value = pair.strip().split('=')
arguments[key] = eval(value)
result = generator(text, max_length=30, num_return_sequences=1, **arguments)
return result[0]["generated_text"]
examples = [
["translate english to latin: and he took the sword and killed the man."],
["translate english to greek: and he took the sword and killed the man."],
]
demo = gr.Interface(
fn=generate,
inputs=[gr.components.Textbox(value="translate greek to english: ὁ ἄνθρωπος τὸν οἶνον πίνειν ἐθέλων τὸν κρατῆρα ἔλαβεν.", label="Input Text"), gr.components.Textbox(value="do_sample=False, num_beams=3", label="Generation Parameters")],
outputs=gr.components.Textbox(value="the man took the bowl with the intention of drinking wine.", label="Generated Text"),
examples=examples
)
demo.launch()