nicolay-r commited on
Commit
0b2f58b
·
verified ·
1 Parent(s): 41ef857

Updated README.md description

Browse files
Files changed (1) hide show
  1. README.md +52 -54
README.md CHANGED
@@ -10,112 +10,110 @@ pipeline_tag: text2text-generation
10
 
11
  # Model Card for Model ID
12
 
13
- <!-- Provide a quick summary of what the model is/does. -->
14
-
15
-
16
-
17
  ## Model Details
18
 
19
  ### Model Description
20
 
21
- <!-- Provide a longer summary of what this model is. -->
22
-
23
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
 
24
 
25
- - **Developed by:** [More Information Needed]
26
- - **Funded by [optional]:** [More Information Needed]
27
- - **Shared by [optional]:** [More Information Needed]
28
- - **Model type:** [More Information Needed]
29
- - **Language(s) (NLP):** [More Information Needed]
30
- - **License:** [More Information Needed]
31
- - **Finetuned from model [optional]:** [More Information Needed]
32
 
33
- ### Model Sources [optional]
34
-
35
- <!-- Provide the basic links for the model. -->
36
-
37
- - **Repository:** [More Information Needed]
38
- - **Paper [optional]:** [More Information Needed]
39
- - **Demo [optional]:** [More Information Needed]
40
 
41
  ## Uses
42
 
43
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
-
45
  ### Direct Use
46
 
47
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
 
49
  [More Information Needed]
50
 
51
- ### Downstream Use [optional]
52
 
53
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
-
55
- [More Information Needed]
56
 
57
  ### Out-of-Scope Use
58
 
59
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
-
61
- [More Information Needed]
62
-
63
- ## Bias, Risks, and Limitations
64
-
65
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
-
67
- [More Information Needed]
68
 
69
  ### Recommendations
70
 
71
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
-
73
  Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
 
75
  ## How to Get Started with the Model
76
 
77
- Use the code below to get started with the model.
78
 
79
- [More Information Needed]
 
80
 
81
  ## Training Details
82
 
83
  ### Training Data
84
 
85
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
 
 
86
 
87
- [More Information Needed]
 
88
 
89
  ### Training Procedure
90
 
91
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
 
 
 
 
 
 
 
 
 
92
 
 
93
 
94
  #### Training Hyperparameters
95
 
96
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
 
97
 
98
  [More Information Needed]
99
 
100
  ## Evaluation
101
 
102
- <!-- This section describes the evaluation protocols and provides the results. -->
103
-
104
  ### Testing Data, Factors & Metrics
105
 
106
  #### Testing Data
107
 
108
- <!-- This should link to a Dataset Card if possible. -->
109
-
110
- [More Information Needed]
111
 
112
  #### Metrics
113
 
114
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
115
-
116
- [More Information Needed]
117
 
118
  ### Results
119
 
120
- [More Information Needed]
121
-
 
 
 
 
 
 
 
 
 
 
 
 
10
 
11
  # Model Card for Model ID
12
 
 
 
 
 
13
  ## Model Details
14
 
15
  ### Model Description
16
 
17
+ - **Developed by:** Reforged by [nicolay-r](https://github.com/nicolay-r), initial credits for implementation to [scofield7419](https://github.com/scofield7419)
18
+ - **Model type:** [Flan-T5](https://huggingface.co/docs/transformers/en/model_doc/flan-t5)
19
+ - **Language(s) (NLP):** English
20
+ - **License:** [Apache License 2.0](https://github.com/scofield7419/THOR-ISA/blob/main/LICENSE.txt)
21
 
22
+ ### Model Sources
 
 
 
 
 
 
23
 
24
+ - **Repository:** [Reasoning-for-Sentiment-Analysis-Framework](https://github.com/nicolay-r/Reasoning-for-Sentiment-Analysis-Framework)
25
+ - **Paper [optional]:** https://arxiv.org/abs/2404.12342
26
+ - **Demo [optional]:** https://arxiv.org/abs/2404.12342
 
 
 
 
27
 
28
  ## Uses
29
 
 
 
30
  ### Direct Use
31
 
32
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
33
 
34
  [More Information Needed]
35
 
36
+ ### Downstream Use
37
 
38
+ Please refer to the [related section](https://github.com/nicolay-r/Reasoning-for-Sentiment-Analysis-Framework?tab=readme-ov-file#three-hop-chain-of-thought-thor) of the **Reasoning-for-Sentiment-Analysis** Framework
 
 
39
 
40
  ### Out-of-Scope Use
41
 
42
+ This model represent a fine-tuned version of the Flan-T5 on RuSentNE-2023 dataset.
43
+ Since dataset represent three-scale output answers (`positive`, `negative`, `neutral`),
44
+ the behavior in general might be biased to this particular task.
 
 
 
 
 
 
45
 
46
  ### Recommendations
47
 
 
 
48
  Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
49
 
50
  ## How to Get Started with the Model
51
 
52
+ Please proceed with the code from the related [Three-Hop-Reasoning CoT](https://github.com/nicolay-r/Reasoning-for-Sentiment-Analysis-Framework?tab=readme-ov-file#three-hop-chain-of-thought-thor) section.
53
 
54
+ Or following the related section on [Google Colab notebook](https://colab.research.google.com/github/nicolay-r/Reasoning-for-Sentiment-Analysis-Framework/blob/main/Reasoning_for_Sentiment_Analysis_Framework.ipynb
55
+ )
56
 
57
  ## Training Details
58
 
59
  ### Training Data
60
 
61
+ We utilize `train` data which was **automatically translated into English using GoogleTransAPI**.
62
+ The initial source of the texts written in Russian, is from the following repository:
63
+ https://github.com/dialogue-evaluation/RuSentNE-evaluation
64
 
65
+ The translated version on the dataset in English could be automatically downloaded via the following script:
66
+ https://github.com/nicolay-r/Reasoning-for-Sentiment-Analysis-Framework/blob/main/rusentne23_download.py
67
 
68
  ### Training Procedure
69
 
70
+ This model has been trained using the Three-hop-Reasoning framework, proposed in the paper:
71
+ https://arxiv.org/abs/2305.11255
72
+
73
+ For training procedure accomplishing, the reforged version of this framework was used:
74
+ https://github.com/nicolay-r/Reasoning-for-Sentiment-Analysis-Framework
75
+
76
+ Google-colab notebook for reproduction:
77
+ https://colab.research.google.com/github/nicolay-r/Reasoning-for-Sentiment-Analysis-Framework/blob/main/Reasoning_for_Sentiment_Analysis_Framework.ipynb
78
+
79
+ The overall training process took **4 epochs**.
80
 
81
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64e62d11d27a8292c3637f86/JwCP0EIe6q1VVdNrTzPQl.png)
82
 
83
  #### Training Hyperparameters
84
 
85
+ - **Training regime:** All the configuration details were highlighted in the related
86
+ [config](https://github.com/nicolay-r/Reasoning-for-Sentiment-Analysis-Framework/blob/main/config/config.yaml) file
87
 
88
  [More Information Needed]
89
 
90
  ## Evaluation
91
 
 
 
92
  ### Testing Data, Factors & Metrics
93
 
94
  #### Testing Data
95
 
96
+ The direct link to the `test` evaluation data:
97
+ https://github.com/dialogue-evaluation/RuSentNE-evaluation/blob/main/final_data.csv
 
98
 
99
  #### Metrics
100
 
101
+ For the model evaluation, two metrics were used:
102
+ 1. F1_PN -- F1-measure over `positive` and `negative` classes;
103
+ 2. F1_PN0 -- F1-measure over `positive`, `negative`, **and `neutral`** classes;
104
 
105
  ### Results
106
 
107
+ **Result:** F1_PN = 60.024
108
+
109
+ Below is the log of the training process that showcases the final peformance on the RuSentNE-2023 `test` set after 4 epochs (lines 5-6):
110
+ ```tsv
111
+ F1_PN F1_PN0 default mode
112
+ 0 45.523 59.375 59.375 valid
113
+ 1 62.345 70.260 70.260 valid
114
+ 2 62.722 70.704 70.704 valid
115
+ 3 62.721 70.671 70.671 valid
116
+ 4 62.357 70.247 70.247 valid
117
+ 5 60.024 68.171 68.171 test
118
+ 6 60.024 68.171 68.171 test
119
+ ```