|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- grammarly/coedit |
|
language: |
|
- en |
|
metrics: |
|
- accuracy |
|
tags: |
|
- torchtune |
|
- grammar-correction |
|
--- |
|
|
|
|
|
### Llama3 CoEdit |
|
|
|
This is a Llama3 8B based model trained using [torchtune](https://pytorch.org/torchtune) on the `grammarly/coedit` dataset. |
|
|
|
### Training details |
|
|
|
The exact training script (`lora_finetune_distributed`) and config (`8B_lora.yaml`) are both included in this repository. Specifically, in order to add the dataset, I added the following lines to the config: |
|
|
|
``` |
|
dataset: |
|
_component_: torchtune.datasets.instruct_dataset |
|
source: grammarly/coedit |
|
template: GrammarErrorCorrectionTemplate |
|
column_map: {"sentence": "src", "output": "tgt"} |
|
train_on_input: False |
|
split: train |
|
``` |
|
|
|
### Evaluation results |