Main Script: QuestionAnswering.py The script uses HuggingFace library for managing the datasets, importing/exporting models and training the models. There are various variables at the start of the script. - train: Training a new model - PEFT: Whether to use PEFT during training - tf32/fp16: Mixed precision training choice - trained_model: Name of trained model (to be pushed to HF Hub) - train_checkpoint: Checkpoint of training (None by default) - squad_shift: Whether to include extra data (squadshift) - base_tokenizer: Tokenizer of base model - base_model: Pre-trained model - test: Testing a model - tokenizer_list/model_list/question_list: Which tokenizer, model and questions to be tested. CUDA is enabled if applicable. Require user to login into HuggingFace Hub (via command line token or through script) if training. Alternative is to not push to hub, a local repository will be created. Huggingface repositories created (models created) - botcon/XLNET_squad_finetuned_large - botcon/XLNET_squadshift_finetuned_large - botcon/LUKE_squad_finetuned_large - botcon/LUKE_squadshift_finetuned_large - botcon/LUKE_squad_what