Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ pipeline_tag: text-generation
|
|
16 |
> A [polyglot](https://en.wikipedia.org/wiki/Multilingualism#In_individuals) language model for the [Occident](https://en.wikipedia.org/wiki/Occident).
|
17 |
>
|
18 |
|
19 |
-
**Occiglot-7B-EU5** is a generative language model with 7B parameters supporting the top-5 EU languages (English, Spanish, French, German, and Italian) and trained by the [
|
20 |
It is based on [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) and trained on 293B tokens of additional multilingual and code data with a block size of 8,192 tokens per sample.
|
21 |
Note that the model is a general-purpose base model and was not instruction-fine-tuned nor optimized for chat or other applications. We make an instruction tuned variant available as [occiglot-7b-eu5-instruct](https://huggingface.co/occiglot/occiglot-7b-eu5-instruct)
|
22 |
|
@@ -32,7 +32,7 @@ If you want to train a model for your own language or are working on evaluations
|
|
32 |
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.html)
|
33 |
- **Compute resources:** [HessianAI's 42](https://hessian.ai/)
|
34 |
- **Contributors:** Manuel Brack, Patrick Schramowski, Pedro Ortiz, Malte Ostendorff, Fabio Barth, Georg Rehm, Kristian Kersting
|
35 |
-
- **Research labs:** [SAINT](https://www.dfki.de/en/web/research/research-departments/foundations-of-systems-ai) and [SLT](https://www.dfki.de/en/web/research/research-departments/speech-and-language-technology)
|
36 |
- **Contact:** [Discord](https://discord.gg/wUpvYs4XvM) [hello@occiglot.org](mailto:hello@occiglot.org)
|
37 |
|
38 |
### How to use
|
|
|
16 |
> A [polyglot](https://en.wikipedia.org/wiki/Multilingualism#In_individuals) language model for the [Occident](https://en.wikipedia.org/wiki/Occident).
|
17 |
>
|
18 |
|
19 |
+
**Occiglot-7B-EU5** is a generative language model with 7B parameters supporting the top-5 EU languages (English, Spanish, French, German, and Italian) and trained by the [Occiglot Research Collective](https://occiglot.github.io/occiglot/).
|
20 |
It is based on [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) and trained on 293B tokens of additional multilingual and code data with a block size of 8,192 tokens per sample.
|
21 |
Note that the model is a general-purpose base model and was not instruction-fine-tuned nor optimized for chat or other applications. We make an instruction tuned variant available as [occiglot-7b-eu5-instruct](https://huggingface.co/occiglot/occiglot-7b-eu5-instruct)
|
22 |
|
|
|
32 |
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.html)
|
33 |
- **Compute resources:** [HessianAI's 42](https://hessian.ai/)
|
34 |
- **Contributors:** Manuel Brack, Patrick Schramowski, Pedro Ortiz, Malte Ostendorff, Fabio Barth, Georg Rehm, Kristian Kersting
|
35 |
+
- **Research labs:** [Occiglot](https://occiglot.github.io/occiglot/) with support from [SAINT](https://www.dfki.de/en/web/research/research-departments/foundations-of-systems-ai) and [SLT](https://www.dfki.de/en/web/research/research-departments/speech-and-language-technology)
|
36 |
- **Contact:** [Discord](https://discord.gg/wUpvYs4XvM) [hello@occiglot.org](mailto:hello@occiglot.org)
|
37 |
|
38 |
### How to use
|