--- library_name: transformers tags: - generated_from_keras_callback model-index: - name: DistilBERT-base-uncased-english-finetuned-squad results: [] datasets: - rajpurkar/squad language: - en base_model: - distilbert/distilbert-base-uncased pipeline_tag: question-answering --- # DistilBERT-base-uncased-english-finetuned-squad This model was finetuned on squad dataset. Use TFDistilBertForQuestionAnswering to import model. Requires DistilBertTokenizerFast to generate tokens that are accepted by this model. ## Model description Base DistilBERT model finetuned using squad dataset for NLP tasks such as context based question answering. ## Training procedure Trained for 3 epochs. ### Training hyperparameters The following hyperparameters were used during training: - optimizer: Adam with learning_rate=5e-5 - training_precision: float32 ### Training results Loss on final epoch: 0.6417 & validation loss: 1.2772 evaluation yet to be done. ### Framework versions - Transformers 4.44.2 - TensorFlow 2.17.0 - Datasets 3.0.0 - Tokenizers 0.19.1