PMC-OA Dataset
News: We have released the PMC-OA dataset. You can choose the subset specifically.
P.S. There's something wrong with the huggingface dataset viewer when the dataset scale gets large. So we sample a subset of it to visualize it directly on web. Click PMC-OA-Demo to view it.
Model Zoo
Check it out if you want to load model pretrained on PMC-OA directly.
We plan to release more models pretrained on PMC-OA. Feel free to reach us if the model you want is not included in model zoo for now. Also, we express our thanks to the help from the community.
Model | Link | Provider |
---|---|---|
ViT-L-14 | https://huggingface.co/ryanyip7777/pmc_vit_l_14 | @ryanyip7777 |
Daraset Structure
PMC-OA (seperated images, separated caption).
images.zip
: images folderpmc_oa.jsonl
: dataset file of pmc-oapmc_oa_beta.jsonl
: dataset file of pmc-oa-beta
-
train.jsonl
: metafile of train set-
valid.jsonl
: metafile of valid set- test.jsonl
: metafile of test set
The difference between PMC-OA & PMC-OA-Beta lies in the methods of processing captions. In PMC-OA, we utilize ChatGPT to help us divide compound captions into seperate ones. While PMC-OA-Beta keeps all the compound ones without division.
Sample
A row in pmc_oa.jsonl
is shown bellow,
{
"image": "PMC212319_Fig3_4.jpg",
"caption": "A. Real time image of the translocation of ARF1-GFP to the plasma membrane ...",
}
Explanation to each key
- image: path to the image
- caption: corresponding to the image