Simeon Emanuilov PRO

s-emanuilov

AI & ML interests

Software Engineer & Ph.D. candidate | Specializing in ML/DL system development & applying AI to solve real-world business problems.

Recent Activity

Organizations

AI Lab - Sofia University's profile picture Scaleflex's profile picture UnfoldAI's profile picture

s-emanuilov's activity

upvoted 2 articles 16 days ago
view article
Article

PaliGemma 2 Mix - New Instruction Vision Language Models by Google

β€’ 65
view article
Article

SigLIP 2: A better multilingual vision language encoder

β€’ 130
upvoted an article 28 days ago
view article
Article

Merge Large Language Models with mergekit

By mlabonne β€’
β€’ 102
replied to their post 29 days ago
view reply

try to reduce gpu_memory_utilization to some lower coefficient

replied to their post about 1 month ago
view reply

Thank you.

I’m also a big fan of Qwen models. However, in this case, I don’t think they are appropriate because I’m not entirely confident in their capabilities regarding multilingual contexts. That’s why I chose Llama.

Overall, I agree that the Qwen series is excellent for most tasks.

posted an update about 1 month ago
view post
Post
5189
Tutorial πŸ’₯ Training a non-English reasoning model with GRPO and Unsloth

I wanted to share my experiment with training reasoning models in languages other than English/Chinese.

Using Llama 3.1 8B as base, GRPO trainer from trl, and Unsloth optimizations, I got a working prototype in Bulgarian after ~5 hours on an L40S GPU. The approach should work for any language where the base model has some pre-training coverage.

Full code and tutorial here: https://unfoldai.com/reasoning-in-a-non-english-language/

The model itself: s-emanuilov/LLMBG-Llama-3.1-8B-BG-Reasoning-v0.1

I hope this helps anyone looking to build reasoning models in their language.
Β·