File size: 1,179 Bytes
04fad50
 
0a09afb
 
 
 
 
 
 
 
 
 
 
 
ea995ee
 
0a09afb
ea995ee
0a09afb
ea995ee
3839e96
0a09afb
 
 
 
 
 
 
 
ea995ee
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
---
widget:
- text: አዲስ አበባ
  example_title: Example 1
- text:  ኢንግሊዝ ፕሪምየር ሊግ
  example_title: Example 2
- text: ፕሬዚዳንት ዶናልድ ትራምፕ
  example_title: Example 3
language:
- am
metrics:
- perplexity
library_name: transformers
pipeline_tag: text-generation
---

# gpt2-small-amharic-8k-128-v3

This is a smaller version of the [gpt2](https://huggingface.co/openai-community/gpt2) decoder transformer model pretrained from scratch for **1.5 days** on **290 million tokens** of **Amharic** text. 

- It has **29.5 Million parameters**
- The **context size** of this model is **128** tokens.
- It has the same **tokenizer** as gpt2, trained from scratch using the same dataset with a vocabulary size of **8192**.
- This is a base model and hasn't undergone any supervised finetuing yet.

It achieves the following results on the evaluation set:

- `Loss: 3.59`
- `Perplexity: 36.23`

### Demo

You can use the following demo to generate text using gpt2-small-amharic. Please **enter a prompt** and click the **Generate** button to generate completions for the prompt.

https://huggingface.co/spaces/rasyosef/GPT2-Amharic