atharva-m commited on
Commit
aa6b55a
·
verified ·
1 Parent(s): b77a245

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -2
README.md CHANGED
@@ -7,6 +7,9 @@ tags:
7
  model-index:
8
  - name: RoBERTa_Sentiment_Analysis
9
  results: []
 
 
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -28,21 +31,28 @@ More information needed
28
 
29
  ## Training and evaluation data
30
 
31
- 'train.csv' of Twitter Sentiment Analysis is used for training and evaluation
32
 
33
  ## Training procedure
34
 
 
 
 
 
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 5e-05
39
  - train_batch_size: 50
40
  - eval_batch_size: 50
41
- - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 500
45
  - num_epochs: 5
 
 
46
 
47
  ### Training results
48
 
 
7
  model-index:
8
  - name: RoBERTa_Sentiment_Analysis
9
  results: []
10
+ language:
11
+ - en
12
+ pipeline_tag: text-classification
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
31
 
32
  ## Training and evaluation data
33
 
34
+ 'train.csv' of Twitter Sentiment Analysis is split into training and evaluation sets (80-20)
35
 
36
  ## Training procedure
37
 
38
+ Pretrained RobertaTokenizerFast is used for tokenizing preprocessed data
39
+
40
+ Pretrained RobertaForSequenceClassification is used as the classification model
41
+
42
+ Hyperparameters are defined in TrainingArguments and Trainer is used to train the model
43
+
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
  - learning_rate: 5e-05
48
  - train_batch_size: 50
49
  - eval_batch_size: 50
 
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
  - lr_scheduler_warmup_steps: 500
53
  - num_epochs: 5
54
+ - weight_decay : 0.0000001
55
+ - report_to="tensorboard"
56
 
57
  ### Training results
58