seyabde's picture
Update README.md
1236f5f
|
raw
history blame
1.66 kB
---
license: afl-3.0
language:
- yo
datasets:
- afriqa
- xlsum
- menyo20k_mt
- alpaca-gpt4
---
# Model Description
**mistral_7b_yo_instruct** is a **text generation** model in Yorùbá.
## Intended uses & limitations
#### How to use
```python
import requests
API_URL = "https://i8nykns7vw253vx3.us-east-1.aws.endpoints.huggingface.cloud"
headers = {
"Authorization": "Bearer hf_xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx",
"Content-Type": "application/json"
}
def query(payload):
response = requests.post(API_URL, headers=headers, json=payload)
return response.json()
# Prompt content: "Pẹlẹ o. Bawo ni o se wa?" ("Hello. How are you?")
output = query({
"inputs": "Pẹlẹ o. Bawo ni o se wa?",
})
# Model response: "O dabo. O jẹ ọjọ ti o dara." ("I am safe. It was a good day.")
print(output)
```
#### Eval results
Coming soon
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
#### Training data
This model is fine-tuned on 60k+ instruction-following demonstrations built from an aggregation of datasets ([AfriQA](https://huggingface.co/datasets/masakhane/afriqa), [XLSum](https://huggingface.co/datasets/csebuetnlp/xlsum), [MENYO-20k](https://huggingface.co/datasets/menyo20k_mt)), and translations of [Alpaca-gpt4](https://huggingface.co/datasets/vicgalle/alpaca-gpt4)).
### Use and safety
We emphasize that mistral_7b_yo_instruct is intended only for research purposes and is not ready to be deployed for general use, namely because we have not designed adequate safety measures.