Raphael Pisoni

4rtemi5

AI & ML interests

None yet

Recent Activity

updated a dataset 9 days ago
4rtemi5/tsb_bert_dataset
updated a dataset 9 days ago
4rtemi5/processed_tsb_bert_dataset
updated a model 9 days ago
4rtemi5/tsb_bert-base-uncased
View all activity

Organizations

Spaces-explorers's profile picture Flax Community's profile picture Italian CLIP Team's profile picture test321's profile picture

4rtemi5's activity

reacted to clem's post with โค๏ธ 3 months ago
view post
Post
4165
Open-source AI creates healthy competition in a field where natural tendencies lead to extreme concentration of power. Imagine a world where only one or two companies could build software. This is the biggest risk and ethical challenge of them all IMO. Let's fight this!
  • 3 replies
ยท
reacted to singhsidhukuldeep's post with ๐Ÿ”ฅ 4 months ago
view post
Post
1222
Remember when @Google launched MediaPipe in an effort to create efficient on-device pipelines?

They've just unlocked the ability to run 7B+ parameter language models directly in your browser. This is a game-changer for on-device AI!

Yes, they are streaming 8.6 GB model files!

Currently, they have Gemma 2B/7B running, but imagine Dynamic LoRA, multimodal support, quantization, and you never leaving Chrome!

This is a significant technical advancement, especially in Memory Optimization:

- Redesigned the model-loading code to work around WebAssembly's 4 GB memory limit.
- Implemented asynchronous loading of transformer stack layers (28 for Gemma 1.1 7B).
- Reduced peak WebAssembly memory usage to less than 1% of previous requirements.

Cross-Platform Compatibility
- Compiled the C++ codebase to WebAssembly for broad browser support.
- Utilized the WebGPU API for native GPU acceleration in browsers.

Here's why this matters:

1. Privacy: No need to send data to remote servers.
2. Cost-Efficiency: Eliminates server expenses.
3. Offline Capabilities: Use powerful AI without an internet connection.

Blog: https://research.google/blog/unlocking-7b-language-models-in-your-browser-a-deep-dive-with-google-ai-edges-mediapipe/
reacted to anakin87's post with ๐Ÿ”ฅ 5 months ago
view post
Post
1667
๐Ÿ’ฌ ๐Ÿ‡ฎ๐Ÿ‡น Phi 3.5 mini ITA: a Small Language Model for Italian

Lately, I've spent some time fine-tuning language models.

Now I am happy to release Phi 3.5 mini ITA: a fine-tuned version of Phi-3.5-mini-instruct to improve performance on the Italian language

๐Ÿ”น Small (3.82 B parameters) but capable model
๐Ÿ”น 128k context length

Chat with it on ๐Ÿค— Spaces: anakin87/Phi-3.5-mini-ITA
Model card: anakin87/Phi-3.5-mini-ITA

๐Ÿ—ƒ๏ธ Data
Supervised fine-tuning using a good mix of English and Italian data:
- mlabonne/FineTome-100k by @mlabonne
- efederici/capybara-claude-15k-ita by @efederici
๐Ÿ™ Thanks to the authors for the datasets.


๐ŸŽฏ Targeted training with Spectrum
I used Spectrum, a relatively new technique for parameter-efficient learning.
The idea is to train only the layers of the model with high Signal-to-Noise Ratio (SNR) and โ„๏ธ freeze the rest.
I trained the top 30% of model layers.

๐Ÿ“ Spectrum paper: https://arxiv.org/abs/2406.06623


๐Ÿ“Š Vibe check and performance on Italian benchmarks seem encouraging
  • 2 replies
ยท
reacted to Xenova's post with ๐Ÿ”ฅ 5 months ago
view post
Post
13945
I can't believe this... Phi-3.5-mini (3.8B) running in-browser at ~90 tokens/second on WebGPU w/ Transformers.js and ONNX Runtime Web! ๐Ÿคฏ Since everything runs 100% locally, no messages are sent to a server โ€” a huge win for privacy!
- ๐Ÿค— Demo: webml-community/phi-3.5-webgpu
- ๐Ÿง‘โ€๐Ÿ’ป Source code: https://github.com/huggingface/transformers.js-examples/tree/main/phi-3.5-webgpu
ยท
liked a Space over 3 years ago