--- license: mit datasets: - gbharti/finance-alpaca - lavita/ChatDoctor-HealthCareMagic-100k - laion/OIG - openai/webgpt_comparisons - taskydata/GPT4Tools - DataProvenanceInitiative/cot_submix_original - 0x70DA/stackoverflow-chat-data language: - en library_name: adapter-transformers pipeline_tag: text-classification --- # Attempt to reproduce Mixture-of-LoRAs classifier Mixture-of-LoRAs: An Efficient Multitask Tuning for Large Language Models https://arxiv.org/pdf/2403.03432 ## Datasets We evenly sample about 10k training data and 2k validation data on each dataset. From `laion/OIG` was taken only: - unified_merged_code_xp3.jsonl - unified_grade_school_math_instructions.jsonl - unified_mathqa_flanv2_kojma_cot.jsonl