weiqipedia commited on
Commit
b4220f3
1 Parent(s): 0685c07

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -9
README.md CHANGED
@@ -10,8 +10,7 @@ license: llama3
10
  # LLaMA3 8B SEA-LIONv2
11
 
12
  SEA-LION is a collection of Large Language Models (LLMs) which has been pretrained and instruct-tuned for the Southeast Asia (SEA) region.
13
- This model is continued pre-trained from the [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) model.
14
- This is the card for the LLaMA3 8B SEA-LIONv2 base model.
15
 
16
  SEA-LION stands for <i>Southeast Asian Languages In One Network</i>.
17
 
@@ -20,11 +19,6 @@ SEA-LION stands for <i>Southeast Asian Languages In One Network</i>.
20
 
21
  ### Model Description
22
 
23
- The LLaMA3 8B SEA-LIONv model is a significant leap forward in the field of Natural Language Processing,
24
- specifically trained to understand the SEA regional context.
25
-
26
- For tokenization, the model employs the default tokenizer used in Meta-Llama-3-8B-Instruct.
27
-
28
  The continued pre-training data for LLaMA3 8B SEA-LIONv2 base model encompasses approximately 48B tokens.
29
 
30
  - **Developed by:** Products Pillar, AI Singapore
@@ -33,11 +27,13 @@ The continued pre-training data for LLaMA3 8B SEA-LIONv2 base model encompasses
33
  - **Languages:** English, Indonesian, Thai, Vietnamese, Tamil
34
  - **License:** [LLaMA3 Community License](https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE)
35
 
 
 
36
  ### Benchmark Performance
37
  We evaluated LLaMA3 8B SEA-LIONv2 base model on general language capabilities.
38
 
39
  #### General Language Capabilities
40
- For the evaluation of general language capabilities, we employed the [BHASA evaluation benchmark](https://arxiv.org/abs/2309.06085v2) across a variety of tasks.
41
  These tasks include Question Answering (QA), Sentiment Analysis (Sentiment), Toxicity Detection (Toxicity), Translation in both directions (Eng>Lang & Lang>Eng), Abstractive Summarization (Summ), Causal Reasoning (Causal) and Natural Language Inference (NLI).
42
 
43
  The evaluation was done **five-shot** with native prompts and only a sample of 100-1000 instances for each dataset was used as per the setting described in the paper.
@@ -46,6 +42,8 @@ The evaluation was done **five-shot** with native prompts and only a sample of 1
46
 
47
  To be released soon
48
 
 
 
49
  **English**
50
 
51
  | Model | ARC | BBH | HellaSwag | MMLU | GSM8k | Average |
@@ -85,7 +83,7 @@ LLaMA3 8B SEA-LIONv2 base model was continued pre-trained on 48B tokens of the f
85
  Note:
86
  - All token counts are counted using LLaMA3 tokenizer
87
  - wiki* sources includes Wikipedia, Wiki Books, Wiki Source and Wiki Voyage
88
- - Source of Tamil news is source with permission from [Seithi](https://seithi.mediacorp.sg/)
89
 
90
  ### Infrastructure
91
 
 
10
  # LLaMA3 8B SEA-LIONv2
11
 
12
  SEA-LION is a collection of Large Language Models (LLMs) which has been pretrained and instruct-tuned for the Southeast Asia (SEA) region.
13
+ This is the card for the LLaMA3 8B SEA-LIONv2 base model which has undergone continued pre-training from the [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) model.
 
14
 
15
  SEA-LION stands for <i>Southeast Asian Languages In One Network</i>.
16
 
 
19
 
20
  ### Model Description
21
 
 
 
 
 
 
22
  The continued pre-training data for LLaMA3 8B SEA-LIONv2 base model encompasses approximately 48B tokens.
23
 
24
  - **Developed by:** Products Pillar, AI Singapore
 
27
  - **Languages:** English, Indonesian, Thai, Vietnamese, Tamil
28
  - **License:** [LLaMA3 Community License](https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE)
29
 
30
+ For tokenization, the model employs the default tokenizer used in Meta-Llama-3-8B-Instruct.
31
+
32
  ### Benchmark Performance
33
  We evaluated LLaMA3 8B SEA-LIONv2 base model on general language capabilities.
34
 
35
  #### General Language Capabilities
36
+ For the evaluation of general language capabilities in SEA languages, we employed the [BHASA evaluation benchmark](https://arxiv.org/abs/2309.06085v2) across a variety of tasks.
37
  These tasks include Question Answering (QA), Sentiment Analysis (Sentiment), Toxicity Detection (Toxicity), Translation in both directions (Eng>Lang & Lang>Eng), Abstractive Summarization (Summ), Causal Reasoning (Causal) and Natural Language Inference (NLI).
38
 
39
  The evaluation was done **five-shot** with native prompts and only a sample of 100-1000 instances for each dataset was used as per the setting described in the paper.
 
42
 
43
  To be released soon
44
 
45
+ We also evaluated the model on English capabilities using tasks from the Open LLM Leaderboard.
46
+
47
  **English**
48
 
49
  | Model | ARC | BBH | HellaSwag | MMLU | GSM8k | Average |
 
83
  Note:
84
  - All token counts are counted using LLaMA3 tokenizer
85
  - wiki* sources includes Wikipedia, Wiki Books, Wiki Source and Wiki Voyage
86
+ - Tamil news is sourced with permission from [Seithi](https://seithi.mediacorp.sg/)
87
 
88
  ### Infrastructure
89