wissamantoun commited on
Commit
5ede0fb
·
1 Parent(s): d963694
Files changed (2) hide show
  1. backend/services.py +7 -7
  2. backend/utils.py +8 -13
backend/services.py CHANGED
@@ -190,7 +190,7 @@ class SentimentAnalyzer:
190
  "sar_trial10",
191
  "sa_no_AOA",
192
  ]
193
- self.model_repos = download_models(self.sa_models)
194
  # fmt: off
195
  self.processors = {
196
  "sa_trial5_1": Trial5ArabicPreprocessor(model_name='UBC-NLP/MARBERT'),
@@ -202,12 +202,12 @@ class SentimentAnalyzer:
202
  }
203
 
204
  self.pipelines = {
205
- "sa_trial5_1": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format(self.model_repos["sa_trial5_1"],i), device=-1,return_all_scores =True) for i in range(0,5)],
206
- "sa_no_aoa_in_neutral": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format(self.model_repos["sa_no_aoa_in_neutral"],i), device=-1,return_all_scores =True) for i in range(0,5)],
207
- "sa_cnnbert": [CNNTextClassificationPipeline("{}/train_{}/best_model".format(self.model_repos["sa_cnnbert"],i), device=-1, return_all_scores =True) for i in range(0,5)],
208
- "sa_sarcasm": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format(self.model_repos["sa_sarcasm"],i), device=-1,return_all_scores =True) for i in range(0,5)],
209
- "sar_trial10": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format(self.model_repos["sar_trial10"],i), device=-1,return_all_scores =True) for i in range(0,5)],
210
- "sa_no_AOA": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format(self.model_repos["sa_no_aoa_in_neutral"],i), device=-1,return_all_scores =True) for i in range(0,5)],
211
  }
212
  # fmt: on
213
 
 
190
  "sar_trial10",
191
  "sa_no_AOA",
192
  ]
193
+ download_models(self.sa_models)
194
  # fmt: off
195
  self.processors = {
196
  "sa_trial5_1": Trial5ArabicPreprocessor(model_name='UBC-NLP/MARBERT'),
 
202
  }
203
 
204
  self.pipelines = {
205
+ "sa_trial5_1": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format("sa_trial5_1",i), device=-1,return_all_scores =True) for i in range(0,5)],
206
+ "sa_no_aoa_in_neutral": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format("sa_no_aoa_in_neutral",i), device=-1,return_all_scores =True) for i in range(0,5)],
207
+ "sa_cnnbert": [CNNTextClassificationPipeline("{}/train_{}/best_model".format("sa_cnnbert",i), device=-1, return_all_scores =True) for i in range(0,5)],
208
+ "sa_sarcasm": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format("sa_sarcasm",i), device=-1,return_all_scores =True) for i in range(0,5)],
209
+ "sar_trial10": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format("sar_trial10",i), device=-1,return_all_scores =True) for i in range(0,5)],
210
+ "sa_no_AOA": [pipeline("sentiment-analysis", model="{}/train_{}/best_model".format("sa_no_AOA",i), device=-1,return_all_scores =True) for i in range(0,5)],
211
  }
212
  # fmt: on
213
 
backend/utils.py CHANGED
@@ -1,5 +1,6 @@
1
  import psutil
2
  import os
 
3
 
4
 
5
  def get_current_ram_usage():
@@ -8,31 +9,25 @@ def get_current_ram_usage():
8
 
9
 
10
  def download_models(models):
11
- model_dirs = {}
12
- for model in models:
13
- model_dirs = {
14
- model: model
15
- } # useless i know, but i don't want to change the code
16
  for i in range(0, 5):
17
  curr_dir = f"{model}/train_{i}/best_model/"
18
  os.makedirs(curr_dir)
19
  os.system(
20
- f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/config.json -P {curr_dir}"
21
  )
22
  os.system(
23
- f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/pytorch_model.bin -P {curr_dir}"
24
  )
25
  os.system(
26
- f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/special_tokens_map.json -P {curr_dir}"
27
  )
28
  os.system(
29
- f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/tokenizer_config.json -P {curr_dir}"
30
  )
31
  os.system(
32
- f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/training_args.bin -P {curr_dir}"
33
  )
34
  os.system(
35
- f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/vocab.txt -P {curr_dir}"
36
  )
37
-
38
- return model_dirs
 
1
  import psutil
2
  import os
3
+ from tqdm.auto import tqdm
4
 
5
 
6
  def get_current_ram_usage():
 
9
 
10
 
11
  def download_models(models):
12
+ for model in tqdm(models, desc="Downloading models"):
 
 
 
 
13
  for i in range(0, 5):
14
  curr_dir = f"{model}/train_{i}/best_model/"
15
  os.makedirs(curr_dir)
16
  os.system(
17
+ f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/config.json -P {curr_dir} >/dev/null"
18
  )
19
  os.system(
20
+ f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/pytorch_model.bin -P {curr_dir} >/dev/null"
21
  )
22
  os.system(
23
+ f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/special_tokens_map.json -P {curr_dir} >/dev/null"
24
  )
25
  os.system(
26
+ f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/tokenizer_config.json -P {curr_dir} >/dev/null"
27
  )
28
  os.system(
29
+ f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/training_args.bin -P {curr_dir} >/dev/null"
30
  )
31
  os.system(
32
+ f"wget https://huggingface.co/researchaccount/{model}/resolve/main/train_{i}/best_model/vocab.txt -P {curr_dir} >/dev/null"
33
  )