diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000000000000000000000000000000000000..55cab133643a2a73e083373d2106533678d0edd5 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,58 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.lz4 filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +# Audio files - uncompressed +*.pcm filter=lfs diff=lfs merge=lfs -text +*.sam filter=lfs diff=lfs merge=lfs -text +*.raw filter=lfs diff=lfs merge=lfs -text +# Audio files - compressed +*.aac filter=lfs diff=lfs merge=lfs -text +*.flac filter=lfs diff=lfs merge=lfs -text +*.mp3 filter=lfs diff=lfs merge=lfs -text +*.ogg filter=lfs diff=lfs merge=lfs -text +*.wav filter=lfs diff=lfs merge=lfs -text +# Image files - uncompressed +*.bmp filter=lfs diff=lfs merge=lfs -text +*.gif filter=lfs diff=lfs merge=lfs -text +*.png filter=lfs diff=lfs merge=lfs -text +*.tiff filter=lfs diff=lfs merge=lfs -text +# Image files - compressed +*.jpg filter=lfs diff=lfs merge=lfs -text +*.jpeg filter=lfs diff=lfs merge=lfs -text +*.webp filter=lfs diff=lfs merge=lfs -text +# Video files - compressed +*.mp4 filter=lfs diff=lfs merge=lfs -text +*.webm filter=lfs diff=lfs merge=lfs -text diff --git a/LLM/LLaMa3-8b-WangchanX-sft-Full/results.json b/LLM/LLaMa3-8b-WangchanX-sft-Full/results.json new file mode 100644 index 0000000000000000000000000000000000000000..2ddc4f0438304099b0d67bd174602b94c0e24f27 --- /dev/null +++ b/LLM/LLaMa3-8b-WangchanX-sft-Full/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "airesearch/LLaMa3-8b-WangchanX-sft-Full"}, "results": {"Math": {"avg_rating": 2.55}, "Reasoning": {"avg_rating": 2.85}, "Extraction": {"avg_rating": 3.15}, "Roleplay": {"avg_rating": 3.8}, "Writing": {"avg_rating": 2.8}, "Social Science": {"avg_rating": 5.1}, "STEM": {"avg_rating": 4.055555555555555}, "Coding": {"avg_rating": 3.6818181818181817}, "Knowledge III": {"avg_rating": 3.6}}} \ No newline at end of file diff --git a/LLM/Meta-Llama-3.1-70B-Instruct/results.json b/LLM/Meta-Llama-3.1-70B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..61581e4b5f41850a23142203229e0dde7defc879 --- /dev/null +++ b/LLM/Meta-Llama-3.1-70B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "meta-llama/Meta-Llama-3.1-70B-Instruct"}, "results": {"Math": {"avg_rating": 6.35}, "Reasoning": {"avg_rating": 5.85}, "Extraction": {"avg_rating": 6.3}, "Roleplay": {"avg_rating": 6.6}, "Writing": {"avg_rating": 7.4}, "Social Science": {"avg_rating": 6.25}, "STEM": {"avg_rating": 6.5}, "Coding": {"avg_rating": 6.409090909090909}, "Knowledge III": {"avg_rating": 4.85}}} \ No newline at end of file diff --git a/LLM/Meta-Llama-3.1-8B-Instruct/results.json b/LLM/Meta-Llama-3.1-8B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..e857738c111e2555055f4e7a1136b955215d642c --- /dev/null +++ b/LLM/Meta-Llama-3.1-8B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "meta-llama/Meta-Llama-3.1-8B-Instruct"}, "results": {"Math": {"avg_rating": 4.25}, "Reasoning": {"avg_rating": 4.2}, "Extraction": {"avg_rating": 4.55}, "Roleplay": {"avg_rating": 6.3}, "Writing": {"avg_rating": 6.25}, "Social Science": {"avg_rating": 7.4}, "STEM": {"avg_rating": 4.305555555555555}, "Coding": {"avg_rating": 4.909090909090909}, "Knowledge III": {"avg_rating": 3.45}}} \ No newline at end of file diff --git a/LLM/Qwen2-72B-Instruct/results.json b/LLM/Qwen2-72B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..ca793d29a653d51e586d83caaa6c2b96df4ef081 --- /dev/null +++ b/LLM/Qwen2-72B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "Qwen/Qwen2-72B-Instruct"}, "results": {"Math": {"avg_rating": 7.25}, "Reasoning": {"avg_rating": 6.35}, "Extraction": {"avg_rating": 6.2}, "Roleplay": {"avg_rating": 7.5}, "Writing": {"avg_rating": 7.2}, "Social Science": {"avg_rating": 8.6}, "STEM": {"avg_rating": 6.833333333333333}, "Coding": {"avg_rating": 7.0}, "Knowledge III": {"avg_rating": 4.75}}} \ No newline at end of file diff --git a/LLM/Qwen2-7B-Instruct/results.json b/LLM/Qwen2-7B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..0569a30522c7ff4cc50e7657731aa4f1aef8a2a4 --- /dev/null +++ b/LLM/Qwen2-7B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "Qwen/Qwen2-7B-Instruct"}, "results": {"Math": {"avg_rating": 4.4}, "Reasoning": {"avg_rating": 4.25}, "Extraction": {"avg_rating": 5.25}, "Roleplay": {"avg_rating": 6.85}, "Writing": {"avg_rating": 6.15}, "Social Science": {"avg_rating": 6.85}, "STEM": {"avg_rating": 4.5}, "Coding": {"avg_rating": 6.2727272727272725}, "Knowledge III": {"avg_rating": 3.1}}} \ No newline at end of file diff --git a/LLM/Sailor-7B-Chat/results.json b/LLM/Sailor-7B-Chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..113379d6745a61ddde5d1ba4208970c94e0ad5b1 --- /dev/null +++ b/LLM/Sailor-7B-Chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "sail/Sailor-7B-Chat"}, "results": {"Math": {"avg_rating": 2.4}, "Reasoning": {"avg_rating": 3.65}, "Extraction": {"avg_rating": 2.65}, "Roleplay": {"avg_rating": 5.55}, "Writing": {"avg_rating": 5.4}, "Social Science": {"avg_rating": 5.45}, "STEM": {"avg_rating": 4.555555555555555}, "Coding": {"avg_rating": 3.5454545454545454}, "Knowledge III": {"avg_rating": 3.05}}} \ No newline at end of file diff --git a/LLM/SeaLLMs-v3-7B-Chat/results.json b/LLM/SeaLLMs-v3-7B-Chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..39ae945f365d5644172bdb88cca38a978ca9660d --- /dev/null +++ b/LLM/SeaLLMs-v3-7B-Chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "SeaLLMs/SeaLLMs-v3-7B-Chat"}, "results": {"Math": {"avg_rating": 4.85}, "Reasoning": {"avg_rating": 2.85}, "Extraction": {"avg_rating": 4.15}, "Roleplay": {"avg_rating": 6.3}, "Writing": {"avg_rating": 6.0}, "Social Science": {"avg_rating": 5.7}, "STEM": {"avg_rating": 5.5}, "Coding": {"avg_rating": 7.045454545454546}, "Knowledge III": {"avg_rating": 3.75}}} \ No newline at end of file diff --git a/LLM/gemini-1.5-flash-001/results.json b/LLM/gemini-1.5-flash-001/results.json new file mode 100644 index 0000000000000000000000000000000000000000..a41fe2720da8e41af25b290722eb465dbfbd8ba4 --- /dev/null +++ b/LLM/gemini-1.5-flash-001/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gemini-1.5-flash-001"}, "results": {"Math": {"avg_rating": 7.7}, "Reasoning": {"avg_rating": 5.85}, "Extraction": {"avg_rating": 6.1}, "Roleplay": {"avg_rating": 7.85}, "Writing": {"avg_rating": 8.0}, "Social Science": {"avg_rating": 7.85}, "STEM": {"avg_rating": 7.888888888888889}, "Coding": {"avg_rating": 7.636363636363637}, "Knowledge III": {"avg_rating": 5.05}}} \ No newline at end of file diff --git a/LLM/gpt-4o-2024-05-13/results.json b/LLM/gpt-4o-2024-05-13/results.json new file mode 100644 index 0000000000000000000000000000000000000000..d94b27e77638099fafead9875f16ad8e9515cc54 --- /dev/null +++ b/LLM/gpt-4o-2024-05-13/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gpt-4o-2024-05-13"}, "results": {"Math": {"avg_rating": 8.5}, "Reasoning": {"avg_rating": 9.0}, "Extraction": {"avg_rating": 7.6}, "Roleplay": {"avg_rating": 7.9}, "Writing": {"avg_rating": 8.15}, "Social Science": {"avg_rating": 8.65}, "STEM": {"avg_rating": 7.861111111111111}, "Coding": {"avg_rating": 8.863636363636363}, "Knowledge III": {"avg_rating": 7.8}}} \ No newline at end of file diff --git a/LLM/gpt-4o-mini-2024-07-18/results.json b/LLM/gpt-4o-mini-2024-07-18/results.json new file mode 100644 index 0000000000000000000000000000000000000000..5eb34a6eb6eea75e53e6b63f5f7f28ed89527abe --- /dev/null +++ b/LLM/gpt-4o-mini-2024-07-18/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gpt-4o-mini-2024-07-18"}, "results": {"Math": {"avg_rating": 7.5}, "Reasoning": {"avg_rating": 6.7}, "Extraction": {"avg_rating": 7.15}, "Roleplay": {"avg_rating": 7.8}, "Writing": {"avg_rating": 8.35}, "Social Science": {"avg_rating": 8.75}, "STEM": {"avg_rating": 7.75}, "Coding": {"avg_rating": 7.318181818181818}, "Knowledge III": {"avg_rating": 6.45}}} \ No newline at end of file diff --git a/LLM/llama-3-typhoon-v1.5-8b-instruct/results.json b/LLM/llama-3-typhoon-v1.5-8b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..c137a19f7e4ac812f67fadb8260c59cef80a0c2c --- /dev/null +++ b/LLM/llama-3-typhoon-v1.5-8b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5-8b-instruct"}, "results": {"Math": {"avg_rating": 3.05}, "Reasoning": {"avg_rating": 5.5}, "Extraction": {"avg_rating": 4.65}, "Roleplay": {"avg_rating": 6.1}, "Writing": {"avg_rating": 7.45}, "Social Science": {"avg_rating": 5.8}, "STEM": {"avg_rating": 4.833333333333333}, "Coding": {"avg_rating": 5.2272727272727275}, "Knowledge III": {"avg_rating": 3.8}}} \ No newline at end of file diff --git a/LLM/llama-3-typhoon-v1.5x-70b-instruct/results.json b/LLM/llama-3-typhoon-v1.5x-70b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..f63771d84c2f5c328149abdd639b999b3f7be205 --- /dev/null +++ b/LLM/llama-3-typhoon-v1.5x-70b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5x-70b-instruct"}, "results": {"Math": {"avg_rating": 5.75}, "Reasoning": {"avg_rating": 6.2}, "Extraction": {"avg_rating": 6.0}, "Roleplay": {"avg_rating": 7.7}, "Writing": {"avg_rating": 7.8}, "Social Science": {"avg_rating": 8.4}, "STEM": {"avg_rating": 6.972222222222222}, "Coding": {"avg_rating": 7.0}, "Knowledge III": {"avg_rating": 4.95}}} \ No newline at end of file diff --git a/LLM/llama-3-typhoon-v1.5x-8b-instruct/results.json b/LLM/llama-3-typhoon-v1.5x-8b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..4f60bdf2458e55d63ab26b88cb1e016ca34e5531 --- /dev/null +++ b/LLM/llama-3-typhoon-v1.5x-8b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5x-8b-instruct"}, "results": {"Math": {"avg_rating": 4.7}, "Reasoning": {"avg_rating": 4.55}, "Extraction": {"avg_rating": 4.65}, "Roleplay": {"avg_rating": 6.4}, "Writing": {"avg_rating": 6.65}, "Social Science": {"avg_rating": 7.5}, "STEM": {"avg_rating": 5.333333333333333}, "Coding": {"avg_rating": 5.090909090909091}, "Knowledge III": {"avg_rating": 3.95}}} \ No newline at end of file diff --git a/LLM/llama3-8b-cpt-sea-lionv2-instruct/results.json b/LLM/llama3-8b-cpt-sea-lionv2-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..7f1b557a7d2e3ed624a1b1980861a172b93b2f14 --- /dev/null +++ b/LLM/llama3-8b-cpt-sea-lionv2-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "aisingapore/llama3-8b-cpt-sea-lionv2-instruct"}, "results": {"Math": {"avg_rating": 3.15}, "Reasoning": {"avg_rating": 4.3}, "Extraction": {"avg_rating": 3.7}, "Roleplay": {"avg_rating": 6.5}, "Writing": {"avg_rating": 6.8}, "Social Science": {"avg_rating": 6.05}, "STEM": {"avg_rating": 5.333333333333333}, "Coding": {"avg_rating": 5.2272727272727275}, "Knowledge III": {"avg_rating": 4.6}}} \ No newline at end of file diff --git a/LLM/openthaigpt-1.0.0-70b-chat/results.json b/LLM/openthaigpt-1.0.0-70b-chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..e8afabaade9d20a9fe2cbd8464fd1748d85f4475 --- /dev/null +++ b/LLM/openthaigpt-1.0.0-70b-chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "openthaigpt/openthaigpt-1.0.0-70b-chat"}, "results": {"Math": {"avg_rating": 2.85}, "Reasoning": {"avg_rating": 4.4}, "Extraction": {"avg_rating": 3.45}, "Roleplay": {"avg_rating": 5.75}, "Writing": {"avg_rating": 4.9}, "Social Science": {"avg_rating": 5.55}, "STEM": {"avg_rating": 5.777777777777778}, "Coding": {"avg_rating": 3.5}, "Knowledge III": {"avg_rating": 4.1}}} \ No newline at end of file diff --git a/LLM/openthaigpt-1.0.0-7b-chat/results.json b/LLM/openthaigpt-1.0.0-7b-chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..5c8fa69d699a0691aaefb56a282ef0c1716e238d --- /dev/null +++ b/LLM/openthaigpt-1.0.0-7b-chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "openthaigpt/openthaigpt-1.0.0-7b-chat"}, "results": {"Math": {"avg_rating": 1.6}, "Reasoning": {"avg_rating": 2.55}, "Extraction": {"avg_rating": 1.75}, "Roleplay": {"avg_rating": 4.1}, "Writing": {"avg_rating": 4.1}, "Social Science": {"avg_rating": 3.0}, "STEM": {"avg_rating": 2.861111111111111}, "Coding": {"avg_rating": 2.3636363636363638}, "Knowledge III": {"avg_rating": 2.15}}} \ No newline at end of file diff --git a/MC/LLaMa3-8b-WangchanX-sft-Full/results.json b/MC/LLaMa3-8b-WangchanX-sft-Full/results.json new file mode 100644 index 0000000000000000000000000000000000000000..aa37b4d5989063b08b6c6cf90f24126b3e4a811d --- /dev/null +++ b/MC/LLaMa3-8b-WangchanX-sft-Full/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "airesearch/LLaMa3-8b-WangchanX-sft-Full"}, "results": {"thaiexam_qa": {"accuracy": 0.3876106194690265}, "m3exam_tha_seacrowd_qa": {"accuracy": 0.4100553505535055}}} \ No newline at end of file diff --git a/MC/Meta-Llama-3.1-70B-Instruct/results.json b/MC/Meta-Llama-3.1-70B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..09e738d4f7d9c708d1b4bbdaf4063b86cf3a4a45 --- /dev/null +++ b/MC/Meta-Llama-3.1-70B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "meta-llama/Meta-Llama-3.1-70B-Instruct"}, "results": {"m3exam_tha_seacrowd_qa": {"accuracy": 0.6051660516605166}, "thaiexam_qa": {"accuracy": 0.5823008849557522}}} \ No newline at end of file diff --git a/MC/Meta-Llama-3.1-8B-Instruct/results.json b/MC/Meta-Llama-3.1-8B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..75d27d0816db514d2a026bdfb66577e0bf8ff40d --- /dev/null +++ b/MC/Meta-Llama-3.1-8B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "meta-llama/Meta-Llama-3.1-8B-Instruct"}, "results": {"thaiexam_qa": {"accuracy": 0.4389380530973451}, "m3exam_tha_seacrowd_qa": {"accuracy": 0.4511070110701107}}} \ No newline at end of file diff --git a/MC/Qwen2-72B-Instruct/results.json b/MC/Qwen2-72B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..cbf976e52dd11f6ee0fa8e903aeeeb3e97bf5006 --- /dev/null +++ b/MC/Qwen2-72B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "Qwen/Qwen2-72B-Instruct"}, "results": {"m3exam_tha_seacrowd_qa": {"accuracy": 0.6305350553505535}, "thaiexam_qa": {"accuracy": 0.5823008849557522}}} \ No newline at end of file diff --git a/MC/Qwen2-7B-Instruct/results.json b/MC/Qwen2-7B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..f9b2e7169758c2e152a010a631a6b55afb9e2423 --- /dev/null +++ b/MC/Qwen2-7B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "Qwen/Qwen2-7B-Instruct"}, "results": {"thaiexam_qa": {"accuracy": 0.4707964601769911}, "m3exam_tha_seacrowd_qa": {"accuracy": 0.4898523985239852}}} \ No newline at end of file diff --git a/MC/Sailor-7B-Chat/results.json b/MC/Sailor-7B-Chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..a25ac599ac2ca2d502a15dbe0de92846c37a2fdf --- /dev/null +++ b/MC/Sailor-7B-Chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "sail/Sailor-7B-Chat"}, "results": {"m3exam_tha_seacrowd_qa": {"accuracy": 0.4363468634686346}, "thaiexam_qa": {"accuracy": 0.4053097345132743}}} \ No newline at end of file diff --git a/MC/SeaLLMs-v3-7B-Chat/results.json b/MC/SeaLLMs-v3-7B-Chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..1652e7d1f3f20bd9d1ee1f1c25949d84a0c1803c --- /dev/null +++ b/MC/SeaLLMs-v3-7B-Chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "SeaLLMs/SeaLLMs-v3-7B-Chat"}, "results": {"m3exam_tha_seacrowd_qa": {"accuracy": 0.5142988929889298}, "thaiexam_qa": {"accuracy": 0.5132743362831859}}} \ No newline at end of file diff --git a/MC/gpt-4o-2024-05-13/results.json b/MC/gpt-4o-2024-05-13/results.json new file mode 100644 index 0000000000000000000000000000000000000000..d99ff3b505d3b9d6038deacf13c837d7e253498b --- /dev/null +++ b/MC/gpt-4o-2024-05-13/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gpt-4o-2024-05-13"}, "results": {"m3exam_tha_seacrowd_qa": {"accuracy": 0.6863468634686347}, "thaiexam_qa": {"accuracy": 0.6389380530973451}}} \ No newline at end of file diff --git a/MC/gpt-4o-mini-2024-07-18/results.json b/MC/gpt-4o-mini-2024-07-18/results.json new file mode 100644 index 0000000000000000000000000000000000000000..0bd153a3b6ea61d5e07e9a3aee6e867652419469 --- /dev/null +++ b/MC/gpt-4o-mini-2024-07-18/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gpt-4o-mini-2024-07-18"}, "results": {"thaiexam_qa": {"accuracy": 0.5451327433628319}, "m3exam_tha_seacrowd_qa": {"accuracy": 0.5784132841328413}}} \ No newline at end of file diff --git a/MC/llama-3-typhoon-v1.5-8b-instruct/results.json b/MC/llama-3-typhoon-v1.5-8b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..b0f79dfc0a4f35badb063479ebe8464804ae9115 --- /dev/null +++ b/MC/llama-3-typhoon-v1.5-8b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5-8b-instruct"}, "results": {"thaiexam_qa": {"accuracy": 0.4123893805309734}, "m3exam_tha_seacrowd_qa": {"accuracy": 0.4483394833948339}}} \ No newline at end of file diff --git a/MC/llama-3-typhoon-v1.5x-70b-instruct/results.json b/MC/llama-3-typhoon-v1.5x-70b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..760ec82623dbb1c7f399ae98dc0ffba9428e6618 --- /dev/null +++ b/MC/llama-3-typhoon-v1.5x-70b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5x-70b-instruct"}, "results": {"m3exam_tha_seacrowd_qa": {"accuracy": 0.599169741697417}, "thaiexam_qa": {"accuracy": 0.5876106194690266}}} \ No newline at end of file diff --git a/MC/llama-3-typhoon-v1.5x-8b-instruct/results.json b/MC/llama-3-typhoon-v1.5x-8b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..49c01ad14e816fa1d75e675f2c826c918b3101fb --- /dev/null +++ b/MC/llama-3-typhoon-v1.5x-8b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5x-8b-instruct"}, "results": {"m3exam_tha_seacrowd_qa": {"accuracy": 0.4381918819188192}, "thaiexam_qa": {"accuracy": 0.4194690265486726}}} \ No newline at end of file diff --git a/MC/llama3-8b-cpt-sea-lionv2-instruct/results.json b/MC/llama3-8b-cpt-sea-lionv2-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..fb00483762847eebed044674d005de30dbac0d69 --- /dev/null +++ b/MC/llama3-8b-cpt-sea-lionv2-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "aisingapore/llama3-8b-cpt-sea-lionv2-instruct"}, "results": {"thaiexam_qa": {"accuracy": 0.3752212389380531}, "m3exam_tha_seacrowd_qa": {"accuracy": 0.4261992619926199}}} \ No newline at end of file diff --git a/MC/openthaigpt-1.0.0-70b-chat/results.json b/MC/openthaigpt-1.0.0-70b-chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..4629a539b473c4ad4bf5175da7d9563f471653b5 --- /dev/null +++ b/MC/openthaigpt-1.0.0-70b-chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "openthaigpt/openthaigpt-1.0.0-70b-chat"}, "results": {"thaiexam_qa": {"accuracy": 0.5008849557522124}, "m3exam_tha_seacrowd_qa": {"accuracy": 0.5170664206642066}}} \ No newline at end of file diff --git a/MC/openthaigpt-1.0.0-7b-chat/results.json b/MC/openthaigpt-1.0.0-7b-chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..f34bb50bccdea8daa64578d86e5d7451719f0491 --- /dev/null +++ b/MC/openthaigpt-1.0.0-7b-chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "openthaigpt/openthaigpt-1.0.0-7b-chat"}, "results": {"m3exam_tha_seacrowd_qa": {"accuracy": 0.2822878228782288}, "thaiexam_qa": {"accuracy": 0.2530973451327433}}} \ No newline at end of file diff --git a/NLG/LLaMa3-8b-WangchanX-sft-Full/results.json b/NLG/LLaMa3-8b-WangchanX-sft-Full/results.json new file mode 100644 index 0000000000000000000000000000000000000000..350106216251761b51246bd7039f62b8fc889c05 --- /dev/null +++ b/NLG/LLaMa3-8b-WangchanX-sft-Full/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "airesearch/LLaMa3-8b-WangchanX-sft-Full"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 37.63697889890035, "ROUGE2": 16.413944846030954, "ROUGEL": 28.198600038522176}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 22.860828678699743, "SacreBLEU": 34.37529930982085, "chrF++": 50.24723442092947}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 48.49722160016288, "SacreBLEU": 31.91991593962373, "chrF++": 55.34127166301678}, "iapp_squad_seacrowd_qa": {"ROUGE1": 80.60864178690281, "ROUGE2": 66.87006649354268, "ROUGEL": 80.37538395013598}}} \ No newline at end of file diff --git a/NLG/Meta-Llama-3.1-70B-Instruct/results.json b/NLG/Meta-Llama-3.1-70B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..f390b9f98fd1ae4c9e35d74eec67048c6e848205 --- /dev/null +++ b/NLG/Meta-Llama-3.1-70B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "meta-llama/Meta-Llama-3.1-70B-Instruct"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 34.15279851230984, "ROUGE2": 13.527627142475689, "ROUGEL": 24.90354379864548}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 28.0767291735307, "SacreBLEU": 38.57267393751737, "chrF++": 53.957636026143504}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 52.75756495966303, "SacreBLEU": 36.91508162495054, "chrF++": 59.46016523399153}, "iapp_squad_seacrowd_qa": {"ROUGE1": 65.23784338666768, "ROUGE2": 52.796325373393, "ROUGEL": 64.62767407946384}}} \ No newline at end of file diff --git a/NLG/Meta-Llama-3.1-8B-Instruct/results.json b/NLG/Meta-Llama-3.1-8B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..5a8b94d4f27e9268b0a6482c5e4653d35b2e8a0c --- /dev/null +++ b/NLG/Meta-Llama-3.1-8B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "meta-llama/Meta-Llama-3.1-8B-Instruct"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 31.280150935761508, "ROUGE2": 11.529087060870982, "ROUGEL": 21.6096544052766}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 21.53499185569734, "SacreBLEU": 30.453791413365572, "chrF++": 46.904976131032285}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 49.534981549467645, "SacreBLEU": 32.279638989338075, "chrF++": 56.42469467923213}, "iapp_squad_seacrowd_qa": {"ROUGE1": 68.99913671626041, "ROUGE2": 55.69258894655722, "ROUGEL": 68.56681583690373}}} \ No newline at end of file diff --git a/NLG/Qwen2-72B-Instruct/results.json b/NLG/Qwen2-72B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..c20374bf99ca94bd994868cdec768ab78788fc4c --- /dev/null +++ b/NLG/Qwen2-72B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "Qwen/Qwen2-72B-Instruct"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 25.69630294614229, "ROUGE2": 8.260340869439057, "ROUGEL": 17.352486052399065}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 21.474236531426097, "SacreBLEU": 33.229982095060386, "chrF++": 51.774302337694}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 44.83361969366853, "SacreBLEU": 29.852007065399075, "chrF++": 58.962206355088185}, "iapp_squad_seacrowd_qa": {"ROUGE1": 35.71235816164353, "ROUGE2": 27.15104533416106, "ROUGEL": 34.979137037450954}}} \ No newline at end of file diff --git a/NLG/Qwen2-7B-Instruct/results.json b/NLG/Qwen2-7B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..28303612f8eb19835edefa7e1491d8786dffa924 --- /dev/null +++ b/NLG/Qwen2-7B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "Qwen/Qwen2-7B-Instruct"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 25.019716118676005, "ROUGE2": 8.819938476123221, "ROUGEL": 16.893508823907176}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 18.429465021232573, "SacreBLEU": 26.773578270493367, "chrF++": 44.35810167800766}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 39.49508634200125, "SacreBLEU": 23.797473235026303, "chrF++": 54.655790877849306}, "iapp_squad_seacrowd_qa": {"ROUGE1": 38.01319643414022, "ROUGE2": 29.173725113229874, "ROUGEL": 37.24609957955968}}} \ No newline at end of file diff --git a/NLG/Sailor-7B-Chat/results.json b/NLG/Sailor-7B-Chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..6ef59435a9c88875a5b2117a3478c7e599a38546 --- /dev/null +++ b/NLG/Sailor-7B-Chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "sail/Sailor-7B-Chat"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 8.082348750838788, "ROUGE2": 1.4226766606322387, "ROUGEL": 6.896338997417523}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 20.09681875599262, "SacreBLEU": 32.529923734701214, "chrF++": 48.34700186743904}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 45.867939537604286, "SacreBLEU": 29.86105422296352, "chrF++": 57.37614705818073}, "iapp_squad_seacrowd_qa": {"ROUGE1": 32.08290873923779, "ROUGE2": 24.236358989805883, "ROUGEL": 31.570119184710787}}} \ No newline at end of file diff --git a/NLG/SeaLLMs-v3-7B-Chat/results.json b/NLG/SeaLLMs-v3-7B-Chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..866f8d4530cd61e7c2d373252ed88f3998da8f08 --- /dev/null +++ b/NLG/SeaLLMs-v3-7B-Chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "SeaLLMs/SeaLLMs-v3-7B-Chat"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 29.968426413809173, "ROUGE2": 10.227472359285144, "ROUGEL": 20.27396351668969}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 23.369448304817407, "SacreBLEU": 32.90676323492414, "chrF++": 49.70573051967939}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 44.66418625932317, "SacreBLEU": 28.795665596226115, "chrF++": 56.82789195857798}, "iapp_squad_seacrowd_qa": {"ROUGE1": 48.51542593783137, "ROUGE2": 39.36421866950317, "ROUGEL": 48.123415223878894}}} \ No newline at end of file diff --git a/NLG/gpt-4o-2024-05-13/results.json b/NLG/gpt-4o-2024-05-13/results.json new file mode 100644 index 0000000000000000000000000000000000000000..0be567cf2425d1fc7874c020449f6468eabc93cc --- /dev/null +++ b/NLG/gpt-4o-2024-05-13/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gpt-4o-2024-05-13"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 26.07540519508083, "ROUGE2": 9.493810579867878, "ROUGEL": 17.461039314397603}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 34.184695263232925, "SacreBLEU": 46.20707168241045, "chrF++": 59.87032517881556}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 52.95793003127012, "SacreBLEU": 37.57233656165935, "chrF++": 62.5830458169773}, "iapp_squad_seacrowd_qa": {"ROUGE1": 37.22409157547258, "ROUGE2": 28.84385857992345, "ROUGEL": 36.59907613378963}}} \ No newline at end of file diff --git a/NLG/gpt-4o-mini-2024-07-18/results.json b/NLG/gpt-4o-mini-2024-07-18/results.json new file mode 100644 index 0000000000000000000000000000000000000000..5ea88cb47c9da3b38b6627d714ae7bfeee2fb13c --- /dev/null +++ b/NLG/gpt-4o-mini-2024-07-18/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gpt-4o-mini-2024-07-18"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 27.390711967187208, "ROUGE2": 9.445645106910009, "ROUGEL": 18.09278006481461}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 30.44345773531864, "SacreBLEU": 42.02386185746654, "chrF++": 56.91612295275716}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 50.45456750112044, "SacreBLEU": 35.04895415588623, "chrF++": 61.04080233081527}, "iapp_squad_seacrowd_qa": {"ROUGE1": 38.86119199633104, "ROUGE2": 30.634787850509333, "ROUGEL": 38.16643099311495}}} \ No newline at end of file diff --git a/NLG/llama-3-typhoon-v1.5-8b-instruct/results.json b/NLG/llama-3-typhoon-v1.5-8b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..58da89b3e8e7aa4c47ad087341d812d2b078f98e --- /dev/null +++ b/NLG/llama-3-typhoon-v1.5-8b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5-8b-instruct"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 29.70552893308527, "ROUGE2": 10.40130095291124, "ROUGEL": 19.73099418924683}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 20.42301736941763, "SacreBLEU": 30.308624386945784, "chrF++": 47.406681011509086}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 44.32516035484208, "SacreBLEU": 28.405122612243485, "chrF++": 56.698085101310255}, "iapp_squad_seacrowd_qa": {"ROUGE1": 56.6595624140093, "ROUGE2": 44.21773783337853, "ROUGEL": 55.9874416305376}}} \ No newline at end of file diff --git a/NLG/llama-3-typhoon-v1.5x-70b-instruct/results.json b/NLG/llama-3-typhoon-v1.5x-70b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..e6efdf4c547c6bd2c2c9824d854d9e9d90a621b4 --- /dev/null +++ b/NLG/llama-3-typhoon-v1.5x-70b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5x-70b-instruct"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 27.94828025074611, "ROUGE2": 10.495022935704329, "ROUGEL": 18.927149051397745}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 24.56695282857076, "SacreBLEU": 35.87021928950534, "chrF++": 52.07290778141319}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 47.93006808151984, "SacreBLEU": 31.98122508709968, "chrF++": 58.99769162518746}, "iapp_squad_seacrowd_qa": {"ROUGE1": 55.09413362900708, "ROUGE2": 43.48289395850847, "ROUGEL": 54.53605035767775}}} \ No newline at end of file diff --git a/NLG/llama-3-typhoon-v1.5x-8b-instruct/results.json b/NLG/llama-3-typhoon-v1.5x-8b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..31b08b2da0273a12caffdf98042f4dec93454c1f --- /dev/null +++ b/NLG/llama-3-typhoon-v1.5x-8b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5x-8b-instruct"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 24.647443491703218, "ROUGE2": 9.292572008638803, "ROUGEL": 16.65766474818548}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 21.75114260265249, "SacreBLEU": 31.94141229827124, "chrF++": 48.65763816553104}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 44.6756443533957, "SacreBLEU": 28.860428237470472, "chrF++": 57.08942685353786}, "iapp_squad_seacrowd_qa": {"ROUGE1": 55.17747628025658, "ROUGE2": 43.73106636235266, "ROUGEL": 54.77648311072716}}} \ No newline at end of file diff --git a/NLG/llama3-8b-cpt-sea-lionv2-instruct/results.json b/NLG/llama3-8b-cpt-sea-lionv2-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..ead2e6a57e7760d93efa0e0c52e2c3a24e389030 --- /dev/null +++ b/NLG/llama3-8b-cpt-sea-lionv2-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "aisingapore/llama3-8b-cpt-sea-lionv2-instruct"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 21.57896237634282, "ROUGE2": 7.9567050604143, "ROUGEL": 14.66884880145703}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 27.246242185807706, "SacreBLEU": 37.798960163282814, "chrF++": 53.07546907542352}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 50.77091046956112, "SacreBLEU": 34.139957925808964, "chrF++": 57.95101492092646}, "iapp_squad_seacrowd_qa": {"ROUGE1": 54.04461363174925, "ROUGE2": 42.74444663261881, "ROUGEL": 53.61474107049092}}} \ No newline at end of file diff --git a/NLG/openthaigpt-1.0.0-70b-chat/results.json b/NLG/openthaigpt-1.0.0-70b-chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..8687d79609408b17a3aff31a356a4a30469f8f3c --- /dev/null +++ b/NLG/openthaigpt-1.0.0-70b-chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "openthaigpt/openthaigpt-1.0.0-70b-chat"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 29.1684873017262, "ROUGE2": 9.692711025177072, "ROUGEL": 21.17699118102632}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 27.71880173693793, "SacreBLEU": 38.90498692430866, "chrF++": 53.54389545440027}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 7.2820993327079, "SacreBLEU": 3.58778640323282, "chrF++": 8.297499265948852}, "iapp_squad_seacrowd_qa": {"ROUGE1": 62.16963279925736, "ROUGE2": 46.67873410863833, "ROUGEL": 61.676715591891366}}} \ No newline at end of file diff --git a/NLG/openthaigpt-1.0.0-7b-chat/results.json b/NLG/openthaigpt-1.0.0-7b-chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..42f9c09d8e9626732b289eb2704d0d85c0c5c0d7 --- /dev/null +++ b/NLG/openthaigpt-1.0.0-7b-chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "openthaigpt/openthaigpt-1.0.0-7b-chat"}, "results": {"xl_sum_tha_seacrowd_t2t": {"ROUGE1": 24.57797140756105, "ROUGE2": 6.576683473636153, "ROUGEL": 16.68568250631098}, "flores200_eng_Latn_tha_Thai_seacrowd_t2t": {"BLEU": 18.43147114138253, "SacreBLEU": 29.731592373901456, "chrF++": 44.90234410450009}, "flores200_tha_Thai_eng_Latn_seacrowd_t2t": {"BLEU": 15.878363204971327, "SacreBLEU": 6.171787547718115, "chrF++": 24.25108108180568}, "iapp_squad_seacrowd_qa": {"ROUGE1": 50.55068681529102, "ROUGE2": 36.76246561062573, "ROUGEL": 50.01850761048598}}} \ No newline at end of file diff --git a/NLU/LLaMa3-8b-WangchanX-sft-Full/results.json b/NLU/LLaMa3-8b-WangchanX-sft-Full/results.json new file mode 100644 index 0000000000000000000000000000000000000000..89a0b3fd2514b6c83246a67f04e45aa452092f2c --- /dev/null +++ b/NLU/LLaMa3-8b-WangchanX-sft-Full/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "airesearch/LLaMa3-8b-WangchanX-sft-Full"}, "results": {"belebele_tha_thai_seacrowd_qa": {"accuracy": 0.6211111111111111}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.507675028079371}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3083832335329341}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.73}}} \ No newline at end of file diff --git a/NLU/Meta-Llama-3.1-70B-Instruct/results.json b/NLU/Meta-Llama-3.1-70B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..cc0f2bef3c99713311e880241f020af70efa01ec --- /dev/null +++ b/NLU/Meta-Llama-3.1-70B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "meta-llama/Meta-Llama-3.1-70B-Instruct"}, "results": {"xcopa_tha_seacrowd_qa": {"accuracy": 0.926}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.8755555555555555}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.4953201048296518}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3347305389221557}}} \ No newline at end of file diff --git a/NLU/Meta-Llama-3.1-8B-Instruct/results.json b/NLU/Meta-Llama-3.1-8B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..3297ae49234d90dd4aba0740fc4dd9f24a60b14c --- /dev/null +++ b/NLU/Meta-Llama-3.1-8B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "meta-llama/Meta-Llama-3.1-8B-Instruct"}, "results": {"belebele_tha_thai_seacrowd_qa": {"accuracy": 0.7488888888888889}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.4545114189442156}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.734}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3349301397205589}}} \ No newline at end of file diff --git a/NLU/Qwen2-72B-Instruct/results.json b/NLU/Qwen2-72B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..86ba8f48568e895116fb095fa793d4c0647e88ae --- /dev/null +++ b/NLU/Qwen2-72B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "Qwen/Qwen2-72B-Instruct"}, "results": {"wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.6008985398727068}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.8611111111111112}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3758483033932135}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.91}}} \ No newline at end of file diff --git a/NLU/Qwen2-7B-Instruct/results.json b/NLU/Qwen2-7B-Instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..f7578d963f8db0420af8c0dc6ebd55ff0a4ec48a --- /dev/null +++ b/NLU/Qwen2-7B-Instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "Qwen/Qwen2-7B-Instruct"}, "results": {"xnli.tha_seacrowd_pairs": {"accuracy": 0.3920159680638722}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.7844444444444445}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.854}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.4706102583302134}}} \ No newline at end of file diff --git a/NLU/Sailor-7B-Chat/results.json b/NLU/Sailor-7B-Chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..d5ab3b96ba1c9fdca2db9eec21cce6c9315fcbc6 --- /dev/null +++ b/NLU/Sailor-7B-Chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "sail/Sailor-7B-Chat"}, "results": {"wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.4552602021714713}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.6811111111111111}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.796}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3349301397205589}}} \ No newline at end of file diff --git a/NLU/SeaLLMs-v3-7B-Chat/results.json b/NLU/SeaLLMs-v3-7B-Chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..c9a05f8335bc9fb6bedfe60f4eeb14df8d3728a7 --- /dev/null +++ b/NLU/SeaLLMs-v3-7B-Chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "SeaLLMs/SeaLLMs-v3-7B-Chat"}, "results": {"belebele_tha_thai_seacrowd_qa": {"accuracy": 0.7822222222222223}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.3957319356046425}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.834}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3405189620758483}}} \ No newline at end of file diff --git a/NLU/gpt-4o-2024-05-13/results.json b/NLU/gpt-4o-2024-05-13/results.json new file mode 100644 index 0000000000000000000000000000000000000000..e44a81e0b2c3c311fda81bbddc381063aaeeac6a --- /dev/null +++ b/NLU/gpt-4o-2024-05-13/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gpt-4o-2024-05-13"}, "results": {"xnli.tha_seacrowd_pairs": {"accuracy": 0.5754491017964072}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.966}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.5675776862598277}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.8977777777777778}}} \ No newline at end of file diff --git a/NLU/gpt-4o-mini-2024-07-18/results.json b/NLU/gpt-4o-mini-2024-07-18/results.json new file mode 100644 index 0000000000000000000000000000000000000000..63deeb8a009b84cd4fe079a8858fee2eda140c6d --- /dev/null +++ b/NLU/gpt-4o-mini-2024-07-18/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "api/gpt-4o-mini-2024-07-18"}, "results": {"belebele_tha_thai_seacrowd_qa": {"accuracy": 0.8477777777777777}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.924}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.5230250842381131}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.5205588822355289}}} \ No newline at end of file diff --git a/NLU/llama-3-typhoon-v1.5-8b-instruct/results.json b/NLU/llama-3-typhoon-v1.5-8b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..4d187dd31d40213e505a4804ed54f4249d0a3954 --- /dev/null +++ b/NLU/llama-3-typhoon-v1.5-8b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5-8b-instruct"}, "results": {"xnli.tha_seacrowd_pairs": {"accuracy": 0.3331337325349301}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.784}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.4129539498315238}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.6866666666666666}}} \ No newline at end of file diff --git a/NLU/llama-3-typhoon-v1.5x-70b-instruct/results.json b/NLU/llama-3-typhoon-v1.5x-70b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..d4d2ef9a4bc810a6b61dcf136ab0b389e600f873 --- /dev/null +++ b/NLU/llama-3-typhoon-v1.5x-70b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5x-70b-instruct"}, "results": {"xcopa_tha_seacrowd_qa": {"accuracy": 0.906}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3614770459081836}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.8555555555555555}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.4062149007862224}}} \ No newline at end of file diff --git a/NLU/llama-3-typhoon-v1.5x-8b-instruct/results.json b/NLU/llama-3-typhoon-v1.5x-8b-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..fc82bcf99c40c34e9e7ef7f3fb1ce95819ac5581 --- /dev/null +++ b/NLU/llama-3-typhoon-v1.5x-8b-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "scb10x/llama-3-typhoon-v1.5x-8b-instruct"}, "results": {"belebele_tha_thai_seacrowd_qa": {"accuracy": 0.7188888888888889}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3335329341317365}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.3845001871958068}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.816}}} \ No newline at end of file diff --git a/NLU/llama3-8b-cpt-sea-lionv2-instruct/results.json b/NLU/llama3-8b-cpt-sea-lionv2-instruct/results.json new file mode 100644 index 0000000000000000000000000000000000000000..fee063193854ec52d068bae3567c032aa04ddaec --- /dev/null +++ b/NLU/llama3-8b-cpt-sea-lionv2-instruct/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "aisingapore/llama3-8b-cpt-sea-lionv2-instruct"}, "results": {"xnli.tha_seacrowd_pairs": {"accuracy": 0.3277445109780439}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.376263571695994}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.5855555555555556}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.684}}} \ No newline at end of file diff --git a/NLU/openthaigpt-1.0.0-70b-chat/results.json b/NLU/openthaigpt-1.0.0-70b-chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..254296694fc07294e0d2c863e2d92170ff1d0dc5 --- /dev/null +++ b/NLU/openthaigpt-1.0.0-70b-chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "openthaigpt/openthaigpt-1.0.0-70b-chat"}, "results": {"wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.3433171096967428}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3590818363273453}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.6944444444444444}, "xcopa_tha_seacrowd_qa": {"accuracy": 0.76}}} \ No newline at end of file diff --git a/NLU/openthaigpt-1.0.0-7b-chat/results.json b/NLU/openthaigpt-1.0.0-7b-chat/results.json new file mode 100644 index 0000000000000000000000000000000000000000..b321b6c0fdd690db891e62dd88de2aa4ddfb365c --- /dev/null +++ b/NLU/openthaigpt-1.0.0-7b-chat/results.json @@ -0,0 +1 @@ +{"config": {"model_name": "openthaigpt/openthaigpt-1.0.0-7b-chat"}, "results": {"xcopa_tha_seacrowd_qa": {"accuracy": 0.526}, "belebele_tha_thai_seacrowd_qa": {"accuracy": 0.3533333333333333}, "xnli.tha_seacrowd_pairs": {"accuracy": 0.3425149700598802}, "wisesight_thai_sentiment_seacrowd_text": {"accuracy": 0.1078247847248221}}} \ No newline at end of file