--- library_name: transformers license: apache-2.0 language: - en base_model: - google/flan-t5-base --- # Model Card for FLAN-T5 QA Study Assistant This model is fine-tuned from the FLAN-T5 model to perform **extractive question-answering** tasks using the SQuAD dataset. It can generate answers based on the given context and question, which makes it useful for educational purposes, personal assistants, or any task requiring context-based question-answering. ## Model Details ### Model Description This is a **question-answering model** fine-tuned on a subset of the SQuAD dataset using the FLAN-T5 model. It is designed to extract answers from a given context based on a corresponding question. - **Developed by:** Tooba Javed - **Funded by [optional]:** Not funded - **Shared by [optional]:** Tooba Javed - **Model type:** Extractive Question-Answering Model (based on FLAN-T5) - **Language(s) (NLP):** English - **License:** Apache-2.0 License - **Finetuned from model [optional]:** google/flan-t5-base ### Model Sources [optional] - **Repository:** [GitHub link to your notebook if applicable] - **Paper [optional]:** N/A - **Demo [optional]:** [Hugging Face link: https://huggingface.co/tootooba/flan-t5-qa-study-assistant] ## Uses ### Direct Use The model is intended to be used for answering questions based on provided context. It can be used in: - **Education**: To help students generate answers from textbooks or lecture notes. - **Customer Support**: Answer common questions from provided documentation or user manuals. - **Personal Assistants**: Assist users by answering general knowledge questions based on given text. ### Downstream Use [optional] The model can be further fine-tuned for domain-specific question-answering tasks, such as: - Legal documents - Medical information - Company-specific internal knowledge bases ### Out-of-Scope Use This model may not work well for tasks that require deep reasoning or understanding beyond surface-level extraction from text. Additionally, it is not designed for tasks like creative writing or opinion generation. ## Bias, Risks, and Limitations - **Bias**: The model is trained on general knowledge from the SQuAD dataset, which may include biases present in Wikipedia-style articles. - **Risks**: Since the model relies on extracting information from provided context, it may return incomplete or misleading answers if the context is ambiguous or lacks the necessary information. - **Limitations**: The model cannot handle complex reasoning or multi-hop question-answering effectively. ### Recommendations Users should ensure that: - They provide relevant and accurate context for better answers. - They are aware of potential biases from the dataset. - The model's limitations are accounted for in critical applications (e.g., legal or medical advice). ## How to Get Started with the Model Use the code below to load the model and tokenizer for question-answering tasks. ```python from transformers import AutoModelForSeq2SeqLM, AutoTokenizer model = AutoModelForSeq2SeqLM.from_pretrained("tootooba/flan-t5-qa-study-assistant").to("cuda") tokenizer = AutoTokenizer.from_pretrained("tootooba/flan-t5-qa-study-assistant") context = "The Eiffel Tower is a wrought-iron lattice tower on the Champ de Mars in Paris, France. It was constructed between 1887 and 1889 as the entrance arch for the 1889 World's Fair." question = "When was the Eiffel Tower constructed?" inputs = tokenizer(question, context, return_tensors="pt", truncation=True, padding=True).to("cuda") outputs = model.generate(inputs.input_ids) answer = tokenizer.decode(outputs[0], skip_special_tokens=True) print("Answer:", answer)