Edit model card

Greek Media BERT (uncased)

This model is a domain-adapted version of nlpaueb/bert-base-greek-uncased-v1 on Greek media centric data.

Model description

Details will be updated soon.

Intended uses & limitations

Details will be updated soon.

Training and evaluation data

Details will be updated soon.

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 64
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 3.0

Training results

Details will be updated soon.

Framework versions

  • Transformers 4.21.0.dev0
  • Pytorch 1.12.0+cu116
  • Tensorflow 2.11.0
  • Datasets 2.3.2
  • Tokenizers 0.12.1

Citation

The model has been officially released with the article "PIMA: Parameter-shared Intelligent Media Analytics Framework for Low Resource Language. Dimitrios Zaikis, Nikolaos Stylianou and Ioannis Vlahavas. In the Special Issue: New Techniques of Machine Learning and Deep Learning in Text Classification, Applied Sciences Journal. 2023" (https://www.mdpi.com/2174928).

If you use the model, please cite the following:

@Article{app13053265,
  AUTHOR = {Zaikis, Dimitrios and Stylianou, Nikolaos and Vlahavas, Ioannis},
  TITLE = {PIMA: Parameter-Shared Intelligent Media Analytics Framework for Low Resource Languages},
  JOURNAL = {Applied Sciences},
  VOLUME = {13},
  YEAR = {2023},
  NUMBER = {5},
  ARTICLE-NUMBER = {3265},
  URL = {https://www.mdpi.com/2076-3417/13/5/3265},
  ISSN = {2076-3417},
  DOI = {10.3390/app13053265}
}
Downloads last month
23
Safetensors
Model size
113M params
Tensor type
I64
·
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.