metadata
language:
- en
tags:
- text-classification
- emotion
- pytorch
license: mit
datasets:
- emotion
metrics:
- accuracy
- precision
- recall
- f1
bert-base-uncased-emotion
Model description
bert-base-uncased
finetuned on the unify-emotion-datasets (https://github.com/sarnthil/unify-emotion-datasets), then transferred to
a small sample of 10K hand-tagged StockTwits messages. Optimized for extracting emotions from financial contexts.
Sequence length 64, learning rate 2e-5, batch size 128, 8 epochs.
For more details, please visit https://github.com/dvamossy/EmTract.
Training data
Data came from https://github.com/sarnthil/unify-emotion-datasets.