Kale's picture
19 9

Kale

Zyn123
Β·

AI & ML interests

None yet

Recent Activity

upvoted an article about 1 month ago
liked a model about 1 month ago
EmergentMethods/gliner_medium_news-v2.1
View all activity

Organizations

None yet

Zyn123's activity

upvoted an article 27 days ago
upvoted an article about 1 month ago
view article
Article

Fine-tune Llama 2 with DPO

β€’ 34
upvoted an article about 1 month ago
view article
Article

How to build a custom text classifier without days of human labeling

By sdiazlor β€’
β€’ 55
upvoted an article about 2 months ago
view article
Article

Fine-tuning LLMs to 1.58bit: extreme quantization made easy

β€’ 205
upvoted 3 articles 3 months ago
view article
Article

Fine-Tune Whisper with πŸ€— Transformers

β€’ 121
view article
Article

Llama-3.1-Storm-8B: Improved SLM with Self-Curation + Model Merging

By akjindal53244 β€’
β€’ 73
upvoted an article 4 months ago
view article
Article

TGI Multi-LoRA: Deploy Once, Serve 30 Models

β€’ 51
upvoted 2 articles 6 months ago
view article
Article

makeMoE: Implement a Sparse Mixture of Experts Language Model from Scratch

By AviSoori1x β€’
β€’ 40
upvoted an article 6 months ago
upvoted an article 7 months ago
Reacted to Molbap's post with πŸ”₯ 8 months ago
view post
Post
5014
πŸš€πŸš€ Exciting times for the document AI community!

We're thrilled to announce the release of some of the largest OCR datasets available to the public.
πŸ”₯ With over 26 million pages , 18 billion text tokens, and 6TB of data, these resources are a significant leap forward for document AI research.

Here's how to access these datasets quickly:

from datasets import load_dataset

pdfa_dataset = load_dataset('pixparse/pdfa-eng-wds', streaming=True)
IDL_dataset = load_dataset('pixparse/idl-wds', streaming=True)

This enables you to stream them directly, integrating seamlessly with your projects using the Hugging Face datasets library. On the hub, you can find them here:

pixparse/pdfa-eng-wds
pixparse/idl-wds

For lean data loading, the new [chug](https://github.com/huggingface/chug) library offers a solution with pdf decoding:


import chug

task_cfg = chug.DataTaskDocReadCfg(
    page_sampling='all',
)
data_cfg = chug.DataCfg(
    source='pixparse/pdfa-eng-wds',
    split='train',
    batch_size=None,
    format='hfids',
    num_workers=0,
)
data_loader = chug.create_loader(
    data_cfg,
    task_cfg,
)
sample = next(iter(data_loader))



We owe a huge thank you to Peter Wyatt, Kate Tasker, Rachel Taketa, Ali Furkan Biten, Ruben Tito, and their colleagues for their contributions. Their work putting these datasets together has been invaluable. πŸ€—

Looking Ahead:

We're on a mission to enhance document AI capabilities, and these datasets are just the beginning. With your engagement and innovation, we're confident in the community's ability to develop robust OCR solutions. We encourage you to explore these datasets, experiment with the code, and contribute to the collective progress in document AI.

For detailed information on usage and licensing, please refer to the dataset cards on the Hugging Face hub.
Β·
liked a Space 8 months ago