visheratin's picture
Create README.md
021f5d1
|
raw
history blame
892 Bytes
metadata
language:
  - en
tags:
  - grammar-correction
license: mit
datasets:
  - c4_200m

T5-Efficient-TINY for grammar correction

This is a T5-Efficient-TINY model that was trained on a subset of C4_200M dataset to solve the grammar correction task in English. To bring additional errors, random typos were introduced to the input sentences using the nlpaug library. Since the model was trained on only one task, there are no prefixes needed.

The model was trained as a part of the project during the Full Stack Depp Learning course. ONNX version of the model is deployed on the site and can be run directly in the browser.