File size: 1,826 Bytes
6b1a623 137b040 6b1a623 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 |
import gradio as gr
import transformers
encoder_model_name = "cl-tohoku/bert-base-japanese-v2"
decoder_model_name = "skt/kogpt2-base-v2"
src_tokenizer = transformers.BertJapaneseTokenizer.from_pretrained(encoder_model_name)
trg_tokenizer = transformers.PreTrainedTokenizerFast.from_pretrained(decoder_model_name)
model = transformers.EncoderDecoderModel.from_pretrained("sappho192/aihub-ja-ko-translator")
def translate(text_src):
embeddings = src_tokenizer(text_src, return_attention_mask=False, return_token_type_ids=False, return_tensors='pt')
embeddings = {k: v for k, v in embeddings.items()}
# using default generation method: GreedySearch, No LogitsProcessor
output = model.generate(**embeddings, max_length=500)[0, 1:-1]
text_trg = trg_tokenizer.decode(output.cpu())
return text_trg
def endpoint(sentence):
return translate(sentence)
# demo = gr.Interface(fn=endpoint, inputs="text", outputs="text")
with gr.Blocks() as demo:
input = gr.Textbox(label="Sentence")
output = gr.Textbox(label="Result")
btn = gr.Button(value="Submit")
btn.click(endpoint, inputs=[input], outputs=[output])
gr.Markdown("## Examples")
gr.Markdown(
"""
Translated result can be wrong or containg misleading content.
번역된 결과는 정확하지 않을 수 있으며, 적절치 않은 표현을 포함할 수 있습니다.
""")
gr.Examples(
[["試験前に緊張したあまり、熱がでてしまった。"],
["山田は英語にかけてはクラスの誰にも負けない。"],
["この本によれば、最初の人工橋梁は新石器時代にさかのぼるという。"]],
[input],
output,
endpoint,
cache_examples=False
)
if __name__ == "__main__":
demo.launch(share=True) |