AIffl : AI For French Language
non-profit
AI & ML interests
Open source Ai & ML for french.
Recent Activity
AIffl's activity
Post
3519
ππ»ββοΈhey there folks,
periodic reminder : if you are experiencing β οΈ500 errors β οΈ or β οΈ abnormal
we have a thread ππ» https://discord.com/channels/879548962464493619/1295847667515129877
if you can record the problem and share it there , or on the forums in your own post , please dont be shy because i'm not sure but i do think it helps π€π€π€
periodic reminder : if you are experiencing β οΈ500 errors β οΈ or β οΈ abnormal
spaces
behavior on load or launch β οΈwe have a thread ππ» https://discord.com/channels/879548962464493619/1295847667515129877
if you can record the problem and share it there , or on the forums in your own post , please dont be shy because i'm not sure but i do think it helps π€π€π€
Post
830
ππ»ββοΈ hey there folks ,
really enjoying sharing cool genomics and protein datasets on the hub these days , check out our cool new org : https://huggingface.co/seq-to-pheno
scroll down for the datasets, still figuring out how to optimize for discoverability , i do think on that part it will be better than zenodo[dot}org , it would be nice to write a tutorial about that and compare : we already have more downloads than most zenodo datasets from famous researchers !
really enjoying sharing cool genomics and protein datasets on the hub these days , check out our cool new org : https://huggingface.co/seq-to-pheno
scroll down for the datasets, still figuring out how to optimize for discoverability , i do think on that part it will be better than zenodo[dot}org , it would be nice to write a tutorial about that and compare : we already have more downloads than most zenodo datasets from famous researchers !
Post
1454
hey there folks,
twitter is aweful isnt it ? just getting into the habbit of using hf/posts for shares π¦π¦
Tonic/on-device-granite-3.0-1b-a400m-instruct
new granite on device instruct model demo , hope you like it ππ
twitter is aweful isnt it ? just getting into the habbit of using hf/posts for shares π¦π¦
Tonic/on-device-granite-3.0-1b-a400m-instruct
new granite on device instruct model demo , hope you like it ππ
Post
990
if you're encountering 500 errors on spaces that seem to work otherwise , kindly consider screenshotting and sharing the link here : https://discord.com/channels/879548962464493619/1295847667515129877
Post
2739
ππ»ββοΈhey there folks ,
did you know that https://huggingface.co/lmms-lab released a new version of ππLlava on thursday ? Now it has π₯video understanding !
check it out ππ»
collection : lmms-lab/llava-video-661e86f5e8dabc3ff793c944
demo : Tonic/Llava-Video
did you know that https://huggingface.co/lmms-lab released a new version of ππLlava on thursday ? Now it has π₯video understanding !
check it out ππ»
collection : lmms-lab/llava-video-661e86f5e8dabc3ff793c944
demo : Tonic/Llava-Video
Post
1856
ππ»ββοΈ Hey there folks ,
π¦Salamandra release by @mvillegas and team
@BSC_CNS https://huggingface.co/BSC-LT is absolutely impressive so far !
perhaps the largest single training dataset of high quality text to date of 7.8 trillion tokens in 35 European languages and code.
the best part : the data was correctly licenced so it's actually future-proof!
the completions model is really creative and instruct fine tuned version is very good also.
now you can use such models for multi-lingual enterprise applications with further finetunes , long response generation, structured outputs (coding) also works.
check out ππ»
the collection : BSC-LT/salamandra-66fc171485944df79469043a
the repo : https://github.com/langtech-bsc/salamandra
7B-Instruct demo : Tonic/Salamandra-7B
π¦Salamandra release by @mvillegas and team
@BSC_CNS https://huggingface.co/BSC-LT is absolutely impressive so far !
perhaps the largest single training dataset of high quality text to date of 7.8 trillion tokens in 35 European languages and code.
the best part : the data was correctly licenced so it's actually future-proof!
the completions model is really creative and instruct fine tuned version is very good also.
now you can use such models for multi-lingual enterprise applications with further finetunes , long response generation, structured outputs (coding) also works.
check out ππ»
the collection : BSC-LT/salamandra-66fc171485944df79469043a
the repo : https://github.com/langtech-bsc/salamandra
7B-Instruct demo : Tonic/Salamandra-7B
Post
1732
@mlabonne
hey there ππ»ββοΈ I kinda got obsessed with your great model , and i found the endpoint for it in lambda labs, but basically i got rate limited / banned for trying to make my DPO dataset project, i was wondering if you all had an open ai compatible solution for me to make a great "thinking" sft + dpo dataset with all the splits ππ»ππ» kinda desparate , it's true , but was looking forward to a nice write ups πππ
Post
2324
Big Congrats on the BIG RELEASE by
@mlabonne
and team at https://huggingface.co/liquidai ...
testing it out now to make a dataset , i cant hardly wait... but one question ππ» why / wen ? π ππ
check out the blog post : https://www.liquid.ai/liquid-foundation-models
testing it out now to make a dataset , i cant hardly wait... but one question ππ» why / wen ? π ππ
check out the blog post : https://www.liquid.ai/liquid-foundation-models
Post
1242
ππ»ββοΈ Hey there folks,
stepfun-ai/GOT-OCR2_0 is in top trending and spaces of the week for the second week straight !!
This is madness π±
ππcheck out my demo here : Tonic/GOT-OCR
stepfun-ai/GOT-OCR2_0 is in top trending and spaces of the week for the second week straight !!
This is madness π±
ππcheck out my demo here : Tonic/GOT-OCR
Post
1089
ππ»ββοΈHey there folks,
Nvidia just released a small 4B Nemotron-mini model , and it works surprisingly well !
you can check it out here :
base : nvidia/Minitron-4B-Base
instruct : nvidia/Nemotron-Mini-4B-Instruct
demo : https://huggingface.co/spaces/Tonic/Nemotron-Mini-4B
hoep you like it π€π€
Nvidia just released a small 4B Nemotron-mini model , and it works surprisingly well !
you can check it out here :
base : nvidia/Minitron-4B-Base
instruct : nvidia/Nemotron-Mini-4B-Instruct
demo : https://huggingface.co/spaces/Tonic/Nemotron-Mini-4B
hoep you like it π€π€
Post
2731
ππ»ββοΈHey there folks ,
@ucaslcl released a new OCR model , that'sππ»ππ» fantastic : https://huggingface.co/ucaslcl/GOT-OCR2_0
GPU : Tonic/GOT-OCR
Gradio Demo (Image Edit) : Tonic1/ImageEdit-GOT-OCR
Model : https://huggingface.co/ucaslcl/GOT-OCR2_0
Official demo : https://huggingface.co/spaces/ucaslcl/GOT_online
github : https://github.com/Ucas-HaoranWei/GOT-OCR2.0
@ucaslcl released a new OCR model , that'sππ»ππ» fantastic : https://huggingface.co/ucaslcl/GOT-OCR2_0
GPU : Tonic/GOT-OCR
Gradio Demo (Image Edit) : Tonic1/ImageEdit-GOT-OCR
Model : https://huggingface.co/ucaslcl/GOT-OCR2_0
Official demo : https://huggingface.co/spaces/ucaslcl/GOT_online
github : https://github.com/Ucas-HaoranWei/GOT-OCR2.0
Post
1107
ππ»ββοΈ hey there folks ,
made an image similarity demo to test out the mistral-community/pixtral-12b-240910 model .
If anyone knows how to generate captions with it , please do let me know x π
here's the demo : Tonic/Pixtral
hope you like it π€
made an image similarity demo to test out the mistral-community/pixtral-12b-240910 model .
If anyone knows how to generate captions with it , please do let me know x π
here's the demo : Tonic/Pixtral
hope you like it π€
Post
2661
So awesome , now i can deploy a jupyterlab on huggingface and deploy gradio from the jupyterlab
Post
1090
ππ»ββοΈHey there folks,
Did you see the new coding model from @01-ai ?
collection : 01-ai/yi-coder-66bdb00f5bdd611f9a008f30
demo : Tonic/Yi-Coder-9B
achieves SOTA on benchmarks , 125K context window , 55 languages including Docker, Js and many more π
Did you see the new coding model from @01-ai ?
collection : 01-ai/yi-coder-66bdb00f5bdd611f9a008f30
demo : Tonic/Yi-Coder-9B
achieves SOTA on benchmarks , 125K context window , 55 languages including Docker, Js and many more π
Post
2525
ππ»ββοΈhey there folks ,
βοΈInkubaLM has been trained from scratch using 1.9 billion tokens of data for five African languages, along with English and French data, totaling 2.4 billion tokens of data. It is capable of understanding and generating content in five African languages: Swahili, Yoruba, Hausa, isiZulu, and isiXhosa, as well as English and French.
model lelapa/InkubaLM-0.4B
demo Tonic/Inkuba-0.4B
βοΈInkubaLM has been trained from scratch using 1.9 billion tokens of data for five African languages, along with English and French data, totaling 2.4 billion tokens of data. It is capable of understanding and generating content in five African languages: Swahili, Yoruba, Hausa, isiZulu, and isiXhosa, as well as English and French.
model lelapa/InkubaLM-0.4B
demo Tonic/Inkuba-0.4B
Post
791
ππ»ββοΈHey there folks,
just published a demo for Salesforce's new Function Calling Model
- Tonic/Salesforce-Xlam-7b-r
- Tonic/On-Device-Function-Calling
just try em out, and it comes with
just published a demo for Salesforce's new Function Calling Model
Salesforce/xLAM
- Tonic/Salesforce-Xlam-7b-r
- Tonic/On-Device-Function-Calling
just try em out, and it comes with
on-device
version too ! cool ! πPost
734
ππ»ββοΈHey there folks ,
I found this cool (new?) thing by Docker called Testcontainers , and there's an @ollama object that you can use to programmatically serve ephemeral containers and LLMs.
I made a post about it here : https://huggingface.co/blog/Tonic/localai-testcontainers
It's really useful, powerful and fun !
Demo coming soon π€
I found this cool (new?) thing by Docker called Testcontainers , and there's an @ollama object that you can use to programmatically serve ephemeral containers and LLMs.
I made a post about it here : https://huggingface.co/blog/Tonic/localai-testcontainers
It's really useful, powerful and fun !
Demo coming soon π€
Post
1716
ππ»ββοΈ Hey there folks
made a demo for Nvidia Minitron on an A100.
Minitron is a family of small language models (SLMs) obtained by pruning NVIDIA's Nemotron-4 15B model. We prune model embedding size, attention heads, and MLP intermediate dimension, following which, we perform continued training with distillation to arrive at the final models.
Deriving the Minitron 8B and 4B models from the base 15B model using our approach requires up to 40x fewer training tokens per model compared to training from scratch; this results in compute cost savings of 1.8x for training the full model family (15B, 8B, and 4B). Minitron models exhibit up to a 16% improvement in MMLU scores compared to training from scratch, perform comparably to other community models such as Mistral 7B, Gemma 7B and Llama-3 8B, and outperform state-of-the-art compression techniques from the literature. Please refer to our arXiv paper for more details.
Minitron models are for research and development only.
source : nvidia/Minitron-8B-Base
demo : Tonic/Minitron
made a demo for Nvidia Minitron on an A100.
Minitron is a family of small language models (SLMs) obtained by pruning NVIDIA's Nemotron-4 15B model. We prune model embedding size, attention heads, and MLP intermediate dimension, following which, we perform continued training with distillation to arrive at the final models.
Deriving the Minitron 8B and 4B models from the base 15B model using our approach requires up to 40x fewer training tokens per model compared to training from scratch; this results in compute cost savings of 1.8x for training the full model family (15B, 8B, and 4B). Minitron models exhibit up to a 16% improvement in MMLU scores compared to training from scratch, perform comparably to other community models such as Mistral 7B, Gemma 7B and Llama-3 8B, and outperform state-of-the-art compression techniques from the literature. Please refer to our arXiv paper for more details.
Minitron models are for research and development only.
source : nvidia/Minitron-8B-Base
demo : Tonic/Minitron