|
--- |
|
license: mit |
|
datasets: |
|
- gbharti/finance-alpaca |
|
- lavita/ChatDoctor-HealthCareMagic-100k |
|
- laion/OIG |
|
- openai/webgpt_comparisons |
|
- taskydata/GPT4Tools |
|
- DataProvenanceInitiative/cot_submix_original |
|
- 0x70DA/stackoverflow-chat-data |
|
language: |
|
- en |
|
library_name: adapter-transformers |
|
pipeline_tag: text-classification |
|
--- |
|
|
|
# Attempt to reproduce Mixture-of-LoRAs classifier |
|
|
|
Mixture-of-LoRAs: An Efficient Multitask Tuning for Large Language Models |
|
|
|
https://arxiv.org/pdf/2403.03432 |
|
|
|
## Datasets |
|
|
|
We evenly sample about 10k training data and 2k validation data on each dataset. |
|
|
|
From `laion/OIG` was taken only: |
|
- unified_merged_code_xp3.jsonl |
|
- unified_grade_school_math_instructions.jsonl |
|
- unified_mathqa_flanv2_kojma_cot.jsonl |