xlm-roberta-base-sentiment-multilingual-finetuned

Model description

This is a fine-tuned version of the cardiffnlp/twitter-xlm-roberta-base-sentiment-multilingual model, trained on the tyqiangz/multilingual-sentiments dataset. It's designed for multilingual sentiment analysis in English, Malay, and Chinese.

Intended uses & limitations

This model is intended for sentiment analysis tasks in English, Malay, and Chinese. It can classify text into three sentiment categories: positive, negative, and neutral.

Training and evaluation data

The model was trained and evaluated on the tyqiangz/multilingual-sentiments dataset, which includes data in English, Malay, and Chinese.

Training procedure

The model was fine-tuned using the Hugging Face Transformers library.

training_args = TrainingArguments( output_dir="./results", num_train_epochs=5, per_device_train_batch_size=16, per_device_eval_batch_size=64, warmup_steps=500, weight_decay=0.01, logging_dir='./logs', logging_steps=10, evaluation_strategy="epoch", save_strategy="epoch", load_best_model_at_end=True, )

Evaluation results

'eval_accuracy': 0.7528205128205128, 'eval_f1': 0.7511924805177581, 'eval_precision': 0.7506612130427309, 'eval_recall': 0.7528205128205128

Test Score :

Environmental impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

Downloads last month
110
Safetensors
Model size
278M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Dataset used to train terrencewee12/xlm-roberta-base-sentiment-multilingual-finetuned

Evaluation results