File size: 2,844 Bytes
7b966b8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4c3e54a
 
7b966b8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
---
license: apache-2.0
language:
- en
- sr
- hr
- bs
datasets:
- teknium/OpenHermes-2.5
- WizardLM/WizardLM_evol_instruct_V2_196k
- draganjovanovich/airoboros-3.0-serbian
- Open-Orca/SlimOrca
- draganjovanovich/prodigy-instruct-reason-function-calling
- iskonai/chatislav-instruct
- mlabonne/FineTome-100k
base_model:
- iskonai/prodigy-sm-base-v0.1
---
# Prodigy SM Instruct v0.1

<img src="https://cdn-uploads.huggingface.co/production/uploads/617bbeec14572ebe9e6ea83f/5sl1RHE054rNkJMA4Z0dp.png" width="70%" height="70%">

**Prodigy SM Instruct v0.1** is an instruction-tuned version of our [Prodigy SM Base v0.1](https://huggingface.co/iskonai/prodigy-sm-base-v0.1) model, optimized for instruction following in Serbian, Croatian, Bosnian and English languages. The model demonstrates exceptional capabilities in function calling and tool usage while maintaining strong performance across all supported languages.

Model was shared as part of the presenatation at DSC Europe 2024 / Belgrade (BREAKOUT SESSIONS & WORKSHOPS).

# ๐ŸŽฎ DEMO TIME! ๐Ÿš€ For next couple of weeks you can test model at: [https://chat.iskon.ai](https://chat.iskon.ai)
or you can download it and use **SGLang**(recomended as it have superior structured output sampling), vLLM, HF text-generation-inference etc.  
proper model card coming soon when we upload final checkpoint, and remove **draft** from the name :)

# Training

The model was instruction-tuned using a carefully curated mix of high-quality datasets in ChatML format.

# Features
- Precise **system prompt** following capabilities in **Serbian** (trained on really large system prompts: 500+ tokens.) 
- Strong instruction following capabilities in Serbian, Croatian, Bosnian and English
- Enhanced function calling and tool usage abilities
- Maintains the base model's strong performance in all supported languages
- Uses ChatML format for conversations

# Use Cases

The model is particularly well-suited for:
- Multilingual applications requiring Serbian, Croatian, Bosnian and English support
- Applications requiring precise instruction following
- Systems utilizing function calling and tool usage
- Conversational AI applications
- Text generation in supported languages

# Limitations

As with all language models:
- Outputs should be verified for critical applications
- Performance in languages other than Serbian, Croatian, Bosnian and English may vary
- Users should be aware of general LLM limitations regarding hallucination and factual accuracy

# What's Next

Stay tuned for our upcoming releases:
- [prodigy-sm-base (llama-3.1)]() **COMING SOON**
- [prodigy-sm-instruct (llama-3.1)]() **COMING SOON**

๐Ÿ“ข We're excited to announce that [iskon.ai](https://Iskon.ai) will soon launch an API platform featuring advanced **Prodigy** series of models, advanced AI tools and much more! ๐Ÿš€