Adds missing tokenizer configuration file
#2
by
lysandre
HF staff
- opened
This repository is missing the tokenizer configuration file, and is instead relying on some attributes set within
the transformers
library directly in order to correctly tokenize inputs.
In order to ensure repositories don't depend on internal configuration changes, we're removing these attribute maps
in transformers#29112.
In doing so, we see that the following attributes are currently missing from the configuration and would be
ill-configured without this PR:
{'model_max_length': 1024}
This PR aims to add these attributes and their values to the tokenizer config file.
This will proceed to make this repository more robust by ensuring that:
- the repository does not depend on intra-library code
- clones of this repository continue working as expected even without the correct repository name
- other libraries that would like to leverage this repository do not depend on code within the transformers library
Thanks 🤗
lysandre
changed pull request status to
open
Thank you !
moussaKam
changed pull request status to
merged
Thank you for merging them!