aliasgerovs commited on
Commit
75ba191
1 Parent(s): 164a7aa
Files changed (2) hide show
  1. nohup.out +29 -282
  2. predictors.py +10 -10
nohup.out CHANGED
@@ -1,179 +1,8 @@
1
- 2024-02-27 18:59:27.464755: I external/local_tsl/tsl/cuda/cudart_stub.cc:31] Could not find cuda drivers on your machine, GPU will not be used.
2
- 2024-02-27 18:59:27.511609: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
3
- 2024-02-27 18:59:27.511670: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
4
- 2024-02-27 18:59:27.513106: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
5
- 2024-02-27 18:59:27.520759: I external/local_tsl/tsl/cuda/cudart_stub.cc:31] Could not find cuda drivers on your machine, GPU will not be used.
6
- 2024-02-27 18:59:27.521010: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
7
- To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
8
- [nltk_data] Downloading package punkt to /root/nltk_data...
9
- [nltk_data] Package punkt is already up-to-date!
10
- [nltk_data] Downloading package punkt to /root/nltk_data...
11
- [nltk_data] Package punkt is already up-to-date!
12
-
13
-
14
-
15
-
16
-
17
-
18
- Traceback (most recent call last):
19
- File "/home/aliasgarov/copyright_checker/app.py", line 197, in <module>
20
- quillbot_model = AutoModelForSequenceClassification.from_pretrained("polygraf-ai/quillbot-detector-roberta-base-28K").to(device)
21
- File "/usr/local/lib/python3.9/dist-packages/transformers/models/auto/auto_factory.py", line 463, in from_pretrained
22
- return model_class.from_pretrained(
23
- File "/usr/local/lib/python3.9/dist-packages/transformers/modeling_utils.py", line 2182, in from_pretrained
24
- raise EnvironmentError(
25
- OSError: polygraf-ai/quillbot-detector-roberta-base-28K does not appear to have a file named pytorch_model.bin, tf_model.h5, model.ckpt or flax_model.msgpack.
26
- 2024-03-27 11:14:47.090743: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
27
- 2024-03-27 11:14:47.090847: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
28
- 2024-03-27 11:14:47.092787: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
29
- 2024-03-27 11:14:47.101435: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
30
- To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
31
- 2024-03-27 11:14:48.367455: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
32
- [nltk_data] Downloading package punkt to /root/nltk_data...
33
- [nltk_data] Package punkt is already up-to-date!
34
- [nltk_data] Downloading package punkt to /root/nltk_data...
35
- [nltk_data] Package punkt is already up-to-date!
36
- [nltk_data] Downloading package stopwords to /root/nltk_data...
37
- [nltk_data] Package stopwords is already up-to-date!
38
- [nltk_data] Downloading package punkt to /root/nltk_data...
39
- [nltk_data] Package punkt is already up-to-date!
40
- [nltk_data] Downloading package punkt to /root/nltk_data...
41
- [nltk_data] Package punkt is already up-to-date!
42
- [nltk_data] Downloading package stopwords to /root/nltk_data...
43
- [nltk_data] Package stopwords is already up-to-date!
44
- error: externally-managed-environment
45
-
46
- × This environment is externally managed
47
- ╰─> To install Python packages system-wide, try apt install
48
- python3-xyz, where xyz is the package you are trying to
49
- install.
50
-
51
- If you wish to install a non-Debian-packaged Python package,
52
- create a virtual environment using python3 -m venv path/to/venv.
53
- Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make
54
- sure you have python3-full installed.
55
-
56
- If you wish to install a non-Debian packaged Python application,
57
- it may be easiest to use pipx install xyz, which will manage a
58
- virtual environment for you. Make sure you have pipx installed.
59
-
60
- See /usr/share/doc/python3.11/README.venv for more information.
61
-
62
- note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages.
63
- hint: See PEP 668 for the detailed specification.
64
- 2024-03-27 11:38:14.074965: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
65
- 2024-03-27 11:38:14.075038: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
66
- 2024-03-27 11:38:14.076732: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
67
- 2024-03-27 11:38:14.085182: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
68
- To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
69
- 2024-03-27 11:38:15.346441: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
70
- [nltk_data] Downloading package punkt to /root/nltk_data...
71
- [nltk_data] Package punkt is already up-to-date!
72
- [nltk_data] Downloading package punkt to /root/nltk_data...
73
- [nltk_data] Package punkt is already up-to-date!
74
- [nltk_data] Downloading package stopwords to /root/nltk_data...
75
- [nltk_data] Package stopwords is already up-to-date!
76
- [nltk_data] Downloading package punkt to /root/nltk_data...
77
- [nltk_data] Package punkt is already up-to-date!
78
- [nltk_data] Downloading package punkt to /root/nltk_data...
79
- [nltk_data] Package punkt is already up-to-date!
80
- [nltk_data] Downloading package stopwords to /root/nltk_data...
81
- [nltk_data] Package stopwords is already up-to-date!
82
- error: externally-managed-environment
83
-
84
- × This environment is externally managed
85
- ╰─> To install Python packages system-wide, try apt install
86
- python3-xyz, where xyz is the package you are trying to
87
- install.
88
-
89
- If you wish to install a non-Debian-packaged Python package,
90
- create a virtual environment using python3 -m venv path/to/venv.
91
- Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make
92
- sure you have python3-full installed.
93
-
94
- If you wish to install a non-Debian packaged Python application,
95
- it may be easiest to use pipx install xyz, which will manage a
96
- virtual environment for you. Make sure you have pipx installed.
97
-
98
- See /usr/share/doc/python3.11/README.venv for more information.
99
-
100
- note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages.
101
- hint: See PEP 668 for the detailed specification.
102
- 2024-03-27 15:11:04.526493: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
103
- 2024-03-27 15:11:04.526578: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
104
- 2024-03-27 15:11:04.528324: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
105
- 2024-03-27 15:11:04.536839: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
106
- To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
107
- 2024-03-27 15:11:05.847612: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
108
- [nltk_data] Downloading package punkt to /root/nltk_data...
109
- [nltk_data] Package punkt is already up-to-date!
110
- [nltk_data] Downloading package punkt to /root/nltk_data...
111
- [nltk_data] Package punkt is already up-to-date!
112
- [nltk_data] Downloading package stopwords to /root/nltk_data...
113
- [nltk_data] Package stopwords is already up-to-date!
114
- [nltk_data] Downloading package punkt to /root/nltk_data...
115
- [nltk_data] Package punkt is already up-to-date!
116
- [nltk_data] Downloading package punkt to /root/nltk_data...
117
- [nltk_data] Package punkt is already up-to-date!
118
- [nltk_data] Downloading package stopwords to /root/nltk_data...
119
- [nltk_data] Package stopwords is already up-to-date!
120
- error: externally-managed-environment
121
-
122
- × This environment is externally managed
123
- ╰─> To install Python packages system-wide, try apt install
124
- python3-xyz, where xyz is the package you are trying to
125
- install.
126
-
127
- If you wish to install a non-Debian-packaged Python package,
128
- create a virtual environment using python3 -m venv path/to/venv.
129
- Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make
130
- sure you have python3-full installed.
131
-
132
- If you wish to install a non-Debian packaged Python application,
133
- it may be easiest to use pipx install xyz, which will manage a
134
- virtual environment for you. Make sure you have pipx installed.
135
-
136
- See /usr/share/doc/python3.11/README.venv for more information.
137
-
138
- note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages.
139
- hint: See PEP 668 for the detailed specification.
140
- /home/aliasgarov/copyright_checker/predictors.py:197: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
141
- probas = F.softmax(tensor_logits).detach().cpu().numpy()
142
- /home/aliasgarov/copyright_checker/predictors.py:197: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
143
- probas = F.softmax(tensor_logits).detach().cpu().numpy()
144
- /home/aliasgarov/copyright_checker/predictors.py:197: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
145
- probas = F.softmax(tensor_logits).detach().cpu().numpy()
146
  /usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
147
  warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
148
- 2024-05-15 06:29:38.253910: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
149
  To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
150
- 2024-05-15 06:29:42.912970: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
151
- [nltk_data] Downloading package punkt to /root/nltk_data...
152
- [nltk_data] Package punkt is already up-to-date!
153
- [nltk_data] Downloading package stopwords to /root/nltk_data...
154
- [nltk_data] Package stopwords is already up-to-date!
155
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
156
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
157
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
158
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
159
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
160
- Traceback (most recent call last):
161
- File "/home/aliasgarov/copyright_checker/app.py", line 4, in <module>
162
- from predictors import predict_bc_scores, predict_mc_scores
163
- File "/home/aliasgarov/copyright_checker/predictors.py", line 93, in <module>
164
- iso_reg = joblib.load("isotonic_regression_model.joblib")
165
- File "/usr/local/lib/python3.9/dist-packages/joblib/numpy_pickle.py", line 658, in load
166
- obj = _unpickle(fobj, filename, mmap_mode)
167
- File "/usr/local/lib/python3.9/dist-packages/joblib/numpy_pickle.py", line 577, in _unpickle
168
- obj = unpickler.load()
169
- File "/usr/lib/python3.9/pickle.py", line 1212, in load
170
- dispatch[key[0]](self)
171
- KeyError: 118
172
- /usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
173
- warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
174
- 2024-05-15 06:35:49.751024: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
175
- To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
176
- 2024-05-15 06:35:50.950991: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
177
  [nltk_data] Downloading package punkt to /root/nltk_data...
178
  [nltk_data] Package punkt is already up-to-date!
179
  [nltk_data] Downloading package stopwords to /root/nltk_data...
@@ -196,85 +25,54 @@ The BetterTransformer implementation does not support padding during training, a
196
  Collecting en_core_web_sm==2.3.1
197
  Using cached en_core_web_sm-2.3.1-py3-none-any.whl
198
  Requirement already satisfied: spacy<2.4.0,>=2.3.0 in /usr/local/lib/python3.9/dist-packages (from en_core_web_sm==2.3.1) (2.3.9)
199
- Requirement already satisfied: srsly<1.1.0,>=1.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.7)
200
- Requirement already satisfied: catalogue<1.1.0,>=0.0.7 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.2)
201
  Requirement already satisfied: preshed<3.1.0,>=3.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (3.0.9)
202
- Requirement already satisfied: requests<3.0.0,>=2.13.0 in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.25.1)
203
- Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (52.0.0)
204
  Requirement already satisfied: thinc<7.5.0,>=7.4.1 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (7.4.6)
205
- Requirement already satisfied: cymem<2.1.0,>=2.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.0.8)
206
- Requirement already satisfied: tqdm<5.0.0,>=4.38.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (4.66.2)
207
- Requirement already satisfied: numpy>=1.15.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.26.4)
208
- Requirement already satisfied: wasabi<1.1.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.10.1)
209
  Requirement already satisfied: murmurhash<1.1.0,>=0.28.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.10)
 
210
  Requirement already satisfied: blis<0.8.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.7.11)
211
- Requirement already satisfied: plac<1.2.0,>=0.9.6 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.1.3)
212
- ✔ Download and installation successful
213
- You can now load the model via spacy.load('en_core_web_sm')
214
- /usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
215
- warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
216
- 2024-05-15 06:39:28.651855: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
217
- To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
218
- 2024-05-15 06:39:29.794203: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
219
- [nltk_data] Downloading package punkt to /root/nltk_data...
220
- [nltk_data] Package punkt is already up-to-date!
221
- [nltk_data] Downloading package stopwords to /root/nltk_data...
222
- [nltk_data] Package stopwords is already up-to-date!
223
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
224
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
225
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
226
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
227
- The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
228
- [nltk_data] Downloading package cmudict to /root/nltk_data...
229
- [nltk_data] Package cmudict is already up-to-date!
230
- [nltk_data] Downloading package punkt to /root/nltk_data...
231
- [nltk_data] Package punkt is already up-to-date!
232
- [nltk_data] Downloading package stopwords to /root/nltk_data...
233
- [nltk_data] Package stopwords is already up-to-date!
234
- [nltk_data] Downloading package wordnet to /root/nltk_data...
235
- [nltk_data] Package wordnet is already up-to-date!
236
- /usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
237
- warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
238
- Collecting en_core_web_sm==2.3.1
239
- Using cached en_core_web_sm-2.3.1-py3-none-any.whl
240
- Requirement already satisfied: spacy<2.4.0,>=2.3.0 in /usr/local/lib/python3.9/dist-packages (from en_core_web_sm==2.3.1) (2.3.9)
241
- Requirement already satisfied: srsly<1.1.0,>=1.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.7)
242
  Requirement already satisfied: wasabi<1.1.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.10.1)
 
243
  Requirement already satisfied: numpy>=1.15.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.26.4)
 
244
  Requirement already satisfied: plac<1.2.0,>=0.9.6 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.1.3)
245
- Requirement already satisfied: cymem<2.1.0,>=2.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.0.8)
246
- Requirement already satisfied: thinc<7.5.0,>=7.4.1 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (7.4.6)
247
- Requirement already satisfied: blis<0.8.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.7.11)
248
  Requirement already satisfied: catalogue<1.1.0,>=0.0.7 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.2)
249
- Requirement already satisfied: requests<3.0.0,>=2.13.0 in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.25.1)
250
- Requirement already satisfied: murmurhash<1.1.0,>=0.28.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.10)
251
- Requirement already satisfied: preshed<3.1.0,>=3.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (3.0.9)
252
- Requirement already satisfied: tqdm<5.0.0,>=4.38.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (4.66.2)
253
- Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (52.0.0)
254
  ✔ Download and installation successful
255
  You can now load the model via spacy.load('en_core_web_sm')
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
256
  huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
257
  To disable this warning, you can either:
258
  - Avoid using `tokenizers` before the fork if possible
259
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
260
  /usr/local/lib/python3.9/dist-packages/torch/cuda/__init__.py:619: UserWarning: Can't initialize NVML
261
  warnings.warn("Can't initialize NVML")
262
- Some characters could not be decoded, and were replaced with REPLACEMENT CHARACTER.
263
  IMPORTANT: You are using gradio version 4.28.3, however version 4.29.0 is available, please upgrade.
264
  --------
265
  Running on local URL: http://0.0.0.0:80
266
- Running on public URL: https://ca11231f7d0d270866.gradio.live
267
 
268
  This share link expires in 72 hours. For free permanent hosting and GPU upgrades, run `gradio deploy` from Terminal to deploy to Spaces (https://huggingface.co/spaces)
269
- ['Multiple factors are helping Russia’s military advance, including a delay in American weaponry and Moscow’s technological innovations on the battlefield.']
270
- huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
271
- To disable this warning, you can either:
272
- - Avoid using `tokenizers` before the fork if possible
273
- - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
274
- huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
275
- To disable this warning, you can either:
276
- - Avoid using `tokenizers` before the fork if possible
277
- - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
278
  huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
279
  To disable this warning, you can either:
280
  - Avoid using `tokenizers` before the fork if possible
@@ -283,28 +81,6 @@ huggingface/tokenizers: The current process just got forked, after parallelism h
283
  To disable this warning, you can either:
284
  - Avoid using `tokenizers` before the fork if possible
285
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
286
- /usr/local/lib/python3.9/dist-packages/optimum/bettertransformer/models/encoder_models.py:301: UserWarning: The PyTorch API of nested tensors is in prototype stage and will change in the near future. (Triggered internally at ../aten/src/ATen/NestedTensorImpl.cpp:178.)
287
- hidden_states = torch._nested_tensor_from_mask(hidden_states, ~attention_mask)
288
- /home/aliasgarov/copyright_checker/predictors.py:247: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
289
- probas = F.softmax(tensor_logits).detach().cpu().numpy()
290
- /home/aliasgarov/copyright_checker/predictors.py:247: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
291
- probas = F.softmax(tensor_logits).detach().cpu().numpy()
292
- WARNING: Invalid HTTP request received.
293
- PLAGIARISM PROCESSING TIME: 21.636404959950596
294
- Original BC scores: AI: 0.9994519352912903, HUMAN: 0.0005480951513163745
295
- Calibration BC scores: AI: 0.8166666666666667, HUMAN: 0.18333333333333335
296
- Input Text: sMultiple factors are helping Russias military advance, including a delay in American weaponry and Moscows technological innovations on the battlefield. /s
297
- Models to Test: ['OpenAI GPT', 'Mistral', 'CLAUDE', 'Gemini', 'Grammar Enhancer']
298
- Original BC scores: AI: 0.9994519352912903, HUMAN: 0.0005480951513163745
299
- Calibration BC scores: AI: 0.8166666666666667, HUMAN: 0.18333333333333335
300
- Starting MC
301
- MC Score: {'OpenAI GPT': 4.604327676532164e-11, 'Mistral': 1.2912608567245758e-11, 'CLAUDE': 3.2367452925959875e-11, 'Gemini': 3.2201130588138284e-11, 'Grammar Enhancer': 0.8166666665431422}
302
- Original BC scores: AI: 0.9994519352912903, HUMAN: 0.0005480951513163745
303
- Calibration BC scores: AI: 0.8166666666666667, HUMAN: 0.18333333333333335
304
- Input Text: sMultiple factors are helping Russias military advance, including a delay in American weaponry and Moscows technological innovations on the battlefield. /s
305
- {'Multiple factors are helping Russia’s military advance, including a delay in American weaponry and Moscow’s technological innovations on the battlefield.': 0.016840771452527072} bc
306
- {'Multiple factors are helping Russia’s military advance, including a delay in American weaponry and Moscow’s technological innovations on the battlefield.': 0.0006924518071430807} quillbot
307
- ['Ilya and OpenAI are going to part ways.', 'This is very sad to me; Ilya is easily one of the greatest minds of our generation, a guiding light of our field, and a dear friend.', 'His brilliance and vision are well known; his warmth and compassion are less well known but no less important.', 'OpenAI would not be what it is without him.', 'Although he has something personally meaningful he is going to go work on, I am forever grateful for what he did here and committed to finishing the mission we started together.', 'I am happy that for so long I got to be close to such genuinely remarkable genius, and someone so focused on getting to the best future for humanity.', 'Jakub is going to be our new Chief Scientist.', 'Jakub is also easily one of the greatest minds of our generation; I am thrilled he is taking the baton here.', 'He has run many of our most important projects, and I am very confident he will lead us to make rapid and safe progress towards our mission of ensuring that AGI benefits everyone.']
308
  huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
309
  To disable this warning, you can either:
310
  - Avoid using `tokenizers` before the fork if possible
@@ -321,32 +97,3 @@ huggingface/tokenizers: The current process just got forked, after parallelism h
321
  To disable this warning, you can either:
322
  - Avoid using `tokenizers` before the fork if possible
323
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
324
- WARNING: Invalid HTTP request received.
325
- WARNING: Invalid HTTP request received.
326
- /home/aliasgarov/copyright_checker/predictors.py:247: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
327
- probas = F.softmax(tensor_logits).detach().cpu().numpy()
328
- /home/aliasgarov/copyright_checker/predictors.py:247: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
329
- probas = F.softmax(tensor_logits).detach().cpu().numpy()
330
- WARNING: Invalid HTTP request received.
331
- WARNING: Invalid HTTP request received.
332
- WARNING: Invalid HTTP request received.
333
- WARNING: Invalid HTTP request received.
334
- WARNING: Invalid HTTP request received.
335
- WARNING: Invalid HTTP request received.
336
- WARNING: Invalid HTTP request received.
337
- WARNING: Invalid HTTP request received.
338
- WARNING: Invalid HTTP request received.
339
- WARNING: Invalid HTTP request received.
340
- WARNING: Invalid HTTP request received.
341
- WARNING: Invalid HTTP request received.
342
- WARNING: Invalid HTTP request received.
343
- WARNING: Invalid HTTP request received.
344
- WARNING: Invalid HTTP request received.
345
- WARNING: Invalid HTTP request received.
346
- WARNING: Invalid HTTP request received.
347
- WARNING: Invalid HTTP request received.
348
- WARNING: Invalid HTTP request received.
349
- WARNING: Invalid HTTP request received.
350
- WARNING: Invalid HTTP request received.
351
- WARNING: Invalid HTTP request received.
352
- WARNING: Invalid HTTP request received.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  /usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
2
  warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
3
+ 2024-05-15 14:59:12.669109: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
4
  To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
5
+ 2024-05-15 14:59:14.457459: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  [nltk_data] Downloading package punkt to /root/nltk_data...
7
  [nltk_data] Package punkt is already up-to-date!
8
  [nltk_data] Downloading package stopwords to /root/nltk_data...
 
25
  Collecting en_core_web_sm==2.3.1
26
  Using cached en_core_web_sm-2.3.1-py3-none-any.whl
27
  Requirement already satisfied: spacy<2.4.0,>=2.3.0 in /usr/local/lib/python3.9/dist-packages (from en_core_web_sm==2.3.1) (2.3.9)
 
 
28
  Requirement already satisfied: preshed<3.1.0,>=3.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (3.0.9)
 
 
29
  Requirement already satisfied: thinc<7.5.0,>=7.4.1 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (7.4.6)
30
+ Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (52.0.0)
 
 
 
31
  Requirement already satisfied: murmurhash<1.1.0,>=0.28.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.10)
32
+ Requirement already satisfied: cymem<2.1.0,>=2.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.0.8)
33
  Requirement already satisfied: blis<0.8.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.7.11)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  Requirement already satisfied: wasabi<1.1.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.10.1)
35
+ Requirement already satisfied: tqdm<5.0.0,>=4.38.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (4.66.2)
36
  Requirement already satisfied: numpy>=1.15.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.26.4)
37
+ Requirement already satisfied: requests<3.0.0,>=2.13.0 in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.25.1)
38
  Requirement already satisfied: plac<1.2.0,>=0.9.6 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.1.3)
 
 
 
39
  Requirement already satisfied: catalogue<1.1.0,>=0.0.7 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.2)
40
+ Requirement already satisfied: srsly<1.1.0,>=1.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.7)
 
 
 
 
41
  ✔ Download and installation successful
42
  You can now load the model via spacy.load('en_core_web_sm')
43
+ Traceback (most recent call last):
44
+ File "/usr/local/lib/python3.9/dist-packages/gradio/queueing.py", line 527, in process_events
45
+ response = await route_utils.call_process_api(
46
+ File "/usr/local/lib/python3.9/dist-packages/gradio/route_utils.py", line 270, in call_process_api
47
+ output = await app.get_blocks().process_api(
48
+ File "/usr/local/lib/python3.9/dist-packages/gradio/blocks.py", line 1847, in process_api
49
+ result = await self.call_function(
50
+ File "/usr/local/lib/python3.9/dist-packages/gradio/blocks.py", line 1433, in call_function
51
+ prediction = await anyio.to_thread.run_sync(
52
+ File "/usr/local/lib/python3.9/dist-packages/anyio/to_thread.py", line 56, in run_sync
53
+ return await get_async_backend().run_sync_in_worker_thread(
54
+ File "/usr/local/lib/python3.9/dist-packages/anyio/_backends/_asyncio.py", line 2144, in run_sync_in_worker_thread
55
+ return await future
56
+ File "/usr/local/lib/python3.9/dist-packages/anyio/_backends/_asyncio.py", line 851, in run
57
+ result = context.run(func, *args)
58
+ File "/usr/local/lib/python3.9/dist-packages/gradio/utils.py", line 788, in wrapper
59
+ response = f(*args, **kwargs)
60
+ File "/home/aliasgarov/copyright_checker/predictors.py", line 119, in update
61
+ corrected_text, corrections = correct_text(text, bias_checker, bias_corrector)
62
+ NameError: name 'bias_checker' is not defined
63
  huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
64
  To disable this warning, you can either:
65
  - Avoid using `tokenizers` before the fork if possible
66
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
67
  /usr/local/lib/python3.9/dist-packages/torch/cuda/__init__.py:619: UserWarning: Can't initialize NVML
68
  warnings.warn("Can't initialize NVML")
 
69
  IMPORTANT: You are using gradio version 4.28.3, however version 4.29.0 is available, please upgrade.
70
  --------
71
  Running on local URL: http://0.0.0.0:80
72
+ Running on public URL: https://a5b565cd42a2675e81.gradio.live
73
 
74
  This share link expires in 72 hours. For free permanent hosting and GPU upgrades, run `gradio deploy` from Terminal to deploy to Spaces (https://huggingface.co/spaces)
75
+ ["OpenAI's chief scientist and co-founder, Ilya Sutskever, is leaving the artificial-intelligence company about six months after he voted to fire Chief Executive Sam Altman only to say he regretted the move days later"]
 
 
 
 
 
 
 
 
76
  huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
77
  To disable this warning, you can either:
78
  - Avoid using `tokenizers` before the fork if possible
 
81
  To disable this warning, you can either:
82
  - Avoid using `tokenizers` before the fork if possible
83
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
84
  huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
85
  To disable this warning, you can either:
86
  - Avoid using `tokenizers` before the fork if possible
 
97
  To disable this warning, you can either:
98
  - Avoid using `tokenizers` before the fork if possible
99
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
predictors.py CHANGED
@@ -78,16 +78,16 @@ text_bc_model = BetterTransformer.transform(text_bc_model)
78
  text_mc_model = BetterTransformer.transform(text_mc_model)
79
  quillbot_model = BetterTransformer.transform(quillbot_model)
80
 
81
- # bias_model_checker = AutoModelForSequenceClassification.from_pretrained(bias_checker_model_name)
82
- # tokenizer = AutoTokenizer.from_pretrained(bias_checker_model_name)
83
- # bias_model_checker = BetterTransformer.transform(bias_model_checker, keep_original_model=False)
84
- # bias_checker = pipeline(
85
- # "text-classification",
86
- # model=bias_checker_model_name,
87
- # tokenizer=bias_checker_model_name,
88
- # )
89
- # gc.collect()
90
- # bias_corrector = pipeline( "text2text-generation", model=bias_corrector_model_name, accelerator="ort")
91
 
92
  # model score calibration
93
  iso_reg = joblib.load("isotonic_regression_model.joblib")
 
78
  text_mc_model = BetterTransformer.transform(text_mc_model)
79
  quillbot_model = BetterTransformer.transform(quillbot_model)
80
 
81
+ bias_model_checker = AutoModelForSequenceClassification.from_pretrained(bias_checker_model_name)
82
+ tokenizer = AutoTokenizer.from_pretrained(bias_checker_model_name)
83
+ bias_model_checker = BetterTransformer.transform(bias_model_checker, keep_original_model=False)
84
+ bias_checker = pipeline(
85
+ "text-classification",
86
+ model=bias_checker_model_name,
87
+ tokenizer=bias_checker_model_name,
88
+ )
89
+ gc.collect()
90
+ bias_corrector = pipeline( "text2text-generation", model=bias_corrector_model_name, accelerator="ort")
91
 
92
  # model score calibration
93
  iso_reg = joblib.load("isotonic_regression_model.joblib")