arwisyah's picture
Update README.md
1db02db verified
metadata
tags:
  - generated_from_trainer
  - finance
base_model: cardiffnlp/twitter-roberta-base-sentiment
metrics:
  - accuracy
model-index:
  - name: fine-tuned-cardiffnlp-twitter-roberta-base-sentiment-finance-dataset
    results: []
datasets:
  - CJCJ3030/twitter-financial-news-sentiment
language:
  - en
library_name: transformers
pipeline_tag: text-classification
widget:
  - text: UK house sales up 12% in April
  - text: Singapore oil trader convicted of abetting forgery and cheating HSBC
  - text: >-
      ‘There’s money everywhere’: Milken conference-goers look for a dealmaking
      revival
  - text: ETF buying nearly halves in April as US rate cut hopes recede
  - text: >-
      Todd Boehly’s investment house in advanced talks to buy private credit
      firm
  - text: Berkshire Hathaway’s cash pile hits new record as Buffett dumps stocks
  - text: Harvest partnership to bring HK-listed crypto ETFs to Singapore
  - text: Kazakh oligarch Timur Kulibayev sells Mayfair mansion for £35mn
  - text: Deutsche Bank’s DWS inflated client asset inflows by billions of euro
  - text: UBS reports stronger than expected profit in first quarter

fine-tuned-cardiffnlp-twitter-roberta-base-sentiment-finance-dataset

This model is a fine-tuned version of cardiffnlp/twitter-roberta-base-sentiment on an twitter finance news sentiment dataset. It achieves the following results on the evaluation set:

  • Loss: 0.3123
  • Accuracy: 0.8559

10 examples in Inference API are gathered from https://twitter.com/ftfinancenews in early may 2024

Colab Notebook for fine tuning : https://colab.research.google.com/drive/1gvpFbazlxg3AdSldH3w6TYjGUByxqCrh?usp=sharing

Training Data

https://huggingface.co/datasets/CJCJ3030/twitter-financial-news-sentiment/viewer/default/train

Evaluation Data

https://huggingface.co/datasets/CJCJ3030/twitter-financial-news-sentiment/viewer/default/validation

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 120
  • eval_batch_size: 120
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Epoch Step Validation Loss Accuracy
1.0 80 0.3123 0.8559
2.0 160 0.3200 0.8576
3.0 240 0.3538 0.8819
4.0 320 0.3695 0.8882
5.0 400 0.4108 0.8869

Framework versions

  • Transformers 4.40.2
  • Pytorch 2.2.1+cu121
  • Datasets 2.19.1
  • Tokenizers 0.19.1

Citation

@inproceedings{barbieri-etal-2020-tweeteval,
    title = "{T}weet{E}val: Unified Benchmark and Comparative Evaluation for Tweet Classification",
    author = "Barbieri, Francesco  and
      Camacho-Collados, Jose  and
      Espinosa Anke, Luis  and
      Neves, Leonardo",
    booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020",
    month = nov,
    year = "2020",
    address = "Online",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2020.findings-emnlp.148",
    doi = "10.18653/v1/2020.findings-emnlp.148",
    pages = "1644--1650"
}