Spaces:
Running
Running
aliasgerovs
commited on
Commit
•
75ba191
1
Parent(s):
164a7aa
Updated
Browse files- nohup.out +29 -282
- predictors.py +10 -10
nohup.out
CHANGED
@@ -1,179 +1,8 @@
|
|
1 |
-
2024-02-27 18:59:27.464755: I external/local_tsl/tsl/cuda/cudart_stub.cc:31] Could not find cuda drivers on your machine, GPU will not be used.
|
2 |
-
2024-02-27 18:59:27.511609: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
|
3 |
-
2024-02-27 18:59:27.511670: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
|
4 |
-
2024-02-27 18:59:27.513106: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
|
5 |
-
2024-02-27 18:59:27.520759: I external/local_tsl/tsl/cuda/cudart_stub.cc:31] Could not find cuda drivers on your machine, GPU will not be used.
|
6 |
-
2024-02-27 18:59:27.521010: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
|
7 |
-
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
|
8 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
9 |
-
[nltk_data] Package punkt is already up-to-date!
|
10 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
11 |
-
[nltk_data] Package punkt is already up-to-date!
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
Traceback (most recent call last):
|
19 |
-
File "/home/aliasgarov/copyright_checker/app.py", line 197, in <module>
|
20 |
-
quillbot_model = AutoModelForSequenceClassification.from_pretrained("polygraf-ai/quillbot-detector-roberta-base-28K").to(device)
|
21 |
-
File "/usr/local/lib/python3.9/dist-packages/transformers/models/auto/auto_factory.py", line 463, in from_pretrained
|
22 |
-
return model_class.from_pretrained(
|
23 |
-
File "/usr/local/lib/python3.9/dist-packages/transformers/modeling_utils.py", line 2182, in from_pretrained
|
24 |
-
raise EnvironmentError(
|
25 |
-
OSError: polygraf-ai/quillbot-detector-roberta-base-28K does not appear to have a file named pytorch_model.bin, tf_model.h5, model.ckpt or flax_model.msgpack.
|
26 |
-
2024-03-27 11:14:47.090743: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
|
27 |
-
2024-03-27 11:14:47.090847: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
|
28 |
-
2024-03-27 11:14:47.092787: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
|
29 |
-
2024-03-27 11:14:47.101435: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
|
30 |
-
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
|
31 |
-
2024-03-27 11:14:48.367455: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
|
32 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
33 |
-
[nltk_data] Package punkt is already up-to-date!
|
34 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
35 |
-
[nltk_data] Package punkt is already up-to-date!
|
36 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
37 |
-
[nltk_data] Package stopwords is already up-to-date!
|
38 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
39 |
-
[nltk_data] Package punkt is already up-to-date!
|
40 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
41 |
-
[nltk_data] Package punkt is already up-to-date!
|
42 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
43 |
-
[nltk_data] Package stopwords is already up-to-date!
|
44 |
-
error: externally-managed-environment
|
45 |
-
|
46 |
-
× This environment is externally managed
|
47 |
-
╰─> To install Python packages system-wide, try apt install
|
48 |
-
python3-xyz, where xyz is the package you are trying to
|
49 |
-
install.
|
50 |
-
|
51 |
-
If you wish to install a non-Debian-packaged Python package,
|
52 |
-
create a virtual environment using python3 -m venv path/to/venv.
|
53 |
-
Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make
|
54 |
-
sure you have python3-full installed.
|
55 |
-
|
56 |
-
If you wish to install a non-Debian packaged Python application,
|
57 |
-
it may be easiest to use pipx install xyz, which will manage a
|
58 |
-
virtual environment for you. Make sure you have pipx installed.
|
59 |
-
|
60 |
-
See /usr/share/doc/python3.11/README.venv for more information.
|
61 |
-
|
62 |
-
note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages.
|
63 |
-
hint: See PEP 668 for the detailed specification.
|
64 |
-
2024-03-27 11:38:14.074965: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
|
65 |
-
2024-03-27 11:38:14.075038: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
|
66 |
-
2024-03-27 11:38:14.076732: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
|
67 |
-
2024-03-27 11:38:14.085182: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
|
68 |
-
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
|
69 |
-
2024-03-27 11:38:15.346441: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
|
70 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
71 |
-
[nltk_data] Package punkt is already up-to-date!
|
72 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
73 |
-
[nltk_data] Package punkt is already up-to-date!
|
74 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
75 |
-
[nltk_data] Package stopwords is already up-to-date!
|
76 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
77 |
-
[nltk_data] Package punkt is already up-to-date!
|
78 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
79 |
-
[nltk_data] Package punkt is already up-to-date!
|
80 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
81 |
-
[nltk_data] Package stopwords is already up-to-date!
|
82 |
-
error: externally-managed-environment
|
83 |
-
|
84 |
-
× This environment is externally managed
|
85 |
-
╰─> To install Python packages system-wide, try apt install
|
86 |
-
python3-xyz, where xyz is the package you are trying to
|
87 |
-
install.
|
88 |
-
|
89 |
-
If you wish to install a non-Debian-packaged Python package,
|
90 |
-
create a virtual environment using python3 -m venv path/to/venv.
|
91 |
-
Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make
|
92 |
-
sure you have python3-full installed.
|
93 |
-
|
94 |
-
If you wish to install a non-Debian packaged Python application,
|
95 |
-
it may be easiest to use pipx install xyz, which will manage a
|
96 |
-
virtual environment for you. Make sure you have pipx installed.
|
97 |
-
|
98 |
-
See /usr/share/doc/python3.11/README.venv for more information.
|
99 |
-
|
100 |
-
note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages.
|
101 |
-
hint: See PEP 668 for the detailed specification.
|
102 |
-
2024-03-27 15:11:04.526493: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
|
103 |
-
2024-03-27 15:11:04.526578: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
|
104 |
-
2024-03-27 15:11:04.528324: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
|
105 |
-
2024-03-27 15:11:04.536839: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
|
106 |
-
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
|
107 |
-
2024-03-27 15:11:05.847612: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
|
108 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
109 |
-
[nltk_data] Package punkt is already up-to-date!
|
110 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
111 |
-
[nltk_data] Package punkt is already up-to-date!
|
112 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
113 |
-
[nltk_data] Package stopwords is already up-to-date!
|
114 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
115 |
-
[nltk_data] Package punkt is already up-to-date!
|
116 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
117 |
-
[nltk_data] Package punkt is already up-to-date!
|
118 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
119 |
-
[nltk_data] Package stopwords is already up-to-date!
|
120 |
-
error: externally-managed-environment
|
121 |
-
|
122 |
-
× This environment is externally managed
|
123 |
-
╰─> To install Python packages system-wide, try apt install
|
124 |
-
python3-xyz, where xyz is the package you are trying to
|
125 |
-
install.
|
126 |
-
|
127 |
-
If you wish to install a non-Debian-packaged Python package,
|
128 |
-
create a virtual environment using python3 -m venv path/to/venv.
|
129 |
-
Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make
|
130 |
-
sure you have python3-full installed.
|
131 |
-
|
132 |
-
If you wish to install a non-Debian packaged Python application,
|
133 |
-
it may be easiest to use pipx install xyz, which will manage a
|
134 |
-
virtual environment for you. Make sure you have pipx installed.
|
135 |
-
|
136 |
-
See /usr/share/doc/python3.11/README.venv for more information.
|
137 |
-
|
138 |
-
note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages.
|
139 |
-
hint: See PEP 668 for the detailed specification.
|
140 |
-
/home/aliasgarov/copyright_checker/predictors.py:197: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
|
141 |
-
probas = F.softmax(tensor_logits).detach().cpu().numpy()
|
142 |
-
/home/aliasgarov/copyright_checker/predictors.py:197: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
|
143 |
-
probas = F.softmax(tensor_logits).detach().cpu().numpy()
|
144 |
-
/home/aliasgarov/copyright_checker/predictors.py:197: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
|
145 |
-
probas = F.softmax(tensor_logits).detach().cpu().numpy()
|
146 |
/usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
|
147 |
warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
|
148 |
-
2024-05-15
|
149 |
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
|
150 |
-
2024-05-15
|
151 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
152 |
-
[nltk_data] Package punkt is already up-to-date!
|
153 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
154 |
-
[nltk_data] Package stopwords is already up-to-date!
|
155 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
156 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
157 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
158 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
159 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
160 |
-
Traceback (most recent call last):
|
161 |
-
File "/home/aliasgarov/copyright_checker/app.py", line 4, in <module>
|
162 |
-
from predictors import predict_bc_scores, predict_mc_scores
|
163 |
-
File "/home/aliasgarov/copyright_checker/predictors.py", line 93, in <module>
|
164 |
-
iso_reg = joblib.load("isotonic_regression_model.joblib")
|
165 |
-
File "/usr/local/lib/python3.9/dist-packages/joblib/numpy_pickle.py", line 658, in load
|
166 |
-
obj = _unpickle(fobj, filename, mmap_mode)
|
167 |
-
File "/usr/local/lib/python3.9/dist-packages/joblib/numpy_pickle.py", line 577, in _unpickle
|
168 |
-
obj = unpickler.load()
|
169 |
-
File "/usr/lib/python3.9/pickle.py", line 1212, in load
|
170 |
-
dispatch[key[0]](self)
|
171 |
-
KeyError: 118
|
172 |
-
/usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
|
173 |
-
warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
|
174 |
-
2024-05-15 06:35:49.751024: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
|
175 |
-
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
|
176 |
-
2024-05-15 06:35:50.950991: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
|
177 |
[nltk_data] Downloading package punkt to /root/nltk_data...
|
178 |
[nltk_data] Package punkt is already up-to-date!
|
179 |
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
@@ -196,85 +25,54 @@ The BetterTransformer implementation does not support padding during training, a
|
|
196 |
Collecting en_core_web_sm==2.3.1
|
197 |
Using cached en_core_web_sm-2.3.1-py3-none-any.whl
|
198 |
Requirement already satisfied: spacy<2.4.0,>=2.3.0 in /usr/local/lib/python3.9/dist-packages (from en_core_web_sm==2.3.1) (2.3.9)
|
199 |
-
Requirement already satisfied: srsly<1.1.0,>=1.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.7)
|
200 |
-
Requirement already satisfied: catalogue<1.1.0,>=0.0.7 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.2)
|
201 |
Requirement already satisfied: preshed<3.1.0,>=3.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (3.0.9)
|
202 |
-
Requirement already satisfied: requests<3.0.0,>=2.13.0 in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.25.1)
|
203 |
-
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (52.0.0)
|
204 |
Requirement already satisfied: thinc<7.5.0,>=7.4.1 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (7.4.6)
|
205 |
-
Requirement already satisfied:
|
206 |
-
Requirement already satisfied: tqdm<5.0.0,>=4.38.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (4.66.2)
|
207 |
-
Requirement already satisfied: numpy>=1.15.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.26.4)
|
208 |
-
Requirement already satisfied: wasabi<1.1.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.10.1)
|
209 |
Requirement already satisfied: murmurhash<1.1.0,>=0.28.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.10)
|
|
|
210 |
Requirement already satisfied: blis<0.8.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.7.11)
|
211 |
-
Requirement already satisfied: plac<1.2.0,>=0.9.6 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.1.3)
|
212 |
-
[38;5;2m✔ Download and installation successful[0m
|
213 |
-
You can now load the model via spacy.load('en_core_web_sm')
|
214 |
-
/usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
|
215 |
-
warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
|
216 |
-
2024-05-15 06:39:28.651855: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
|
217 |
-
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
|
218 |
-
2024-05-15 06:39:29.794203: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
|
219 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
220 |
-
[nltk_data] Package punkt is already up-to-date!
|
221 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
222 |
-
[nltk_data] Package stopwords is already up-to-date!
|
223 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
224 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
225 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
226 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
227 |
-
The BetterTransformer implementation does not support padding during training, as the fused kernels do not support attention masks. Beware that passing padded batched data during training may result in unexpected outputs. Please refer to https://huggingface.co/docs/optimum/bettertransformer/overview for more details.
|
228 |
-
[nltk_data] Downloading package cmudict to /root/nltk_data...
|
229 |
-
[nltk_data] Package cmudict is already up-to-date!
|
230 |
-
[nltk_data] Downloading package punkt to /root/nltk_data...
|
231 |
-
[nltk_data] Package punkt is already up-to-date!
|
232 |
-
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
233 |
-
[nltk_data] Package stopwords is already up-to-date!
|
234 |
-
[nltk_data] Downloading package wordnet to /root/nltk_data...
|
235 |
-
[nltk_data] Package wordnet is already up-to-date!
|
236 |
-
/usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
|
237 |
-
warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
|
238 |
-
Collecting en_core_web_sm==2.3.1
|
239 |
-
Using cached en_core_web_sm-2.3.1-py3-none-any.whl
|
240 |
-
Requirement already satisfied: spacy<2.4.0,>=2.3.0 in /usr/local/lib/python3.9/dist-packages (from en_core_web_sm==2.3.1) (2.3.9)
|
241 |
-
Requirement already satisfied: srsly<1.1.0,>=1.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.7)
|
242 |
Requirement already satisfied: wasabi<1.1.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.10.1)
|
|
|
243 |
Requirement already satisfied: numpy>=1.15.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.26.4)
|
|
|
244 |
Requirement already satisfied: plac<1.2.0,>=0.9.6 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.1.3)
|
245 |
-
Requirement already satisfied: cymem<2.1.0,>=2.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.0.8)
|
246 |
-
Requirement already satisfied: thinc<7.5.0,>=7.4.1 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (7.4.6)
|
247 |
-
Requirement already satisfied: blis<0.8.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.7.11)
|
248 |
Requirement already satisfied: catalogue<1.1.0,>=0.0.7 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.2)
|
249 |
-
Requirement already satisfied:
|
250 |
-
Requirement already satisfied: murmurhash<1.1.0,>=0.28.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.10)
|
251 |
-
Requirement already satisfied: preshed<3.1.0,>=3.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (3.0.9)
|
252 |
-
Requirement already satisfied: tqdm<5.0.0,>=4.38.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (4.66.2)
|
253 |
-
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (52.0.0)
|
254 |
[38;5;2m✔ Download and installation successful[0m
|
255 |
You can now load the model via spacy.load('en_core_web_sm')
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
256 |
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
|
257 |
To disable this warning, you can either:
|
258 |
- Avoid using `tokenizers` before the fork if possible
|
259 |
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
|
260 |
/usr/local/lib/python3.9/dist-packages/torch/cuda/__init__.py:619: UserWarning: Can't initialize NVML
|
261 |
warnings.warn("Can't initialize NVML")
|
262 |
-
Some characters could not be decoded, and were replaced with REPLACEMENT CHARACTER.
|
263 |
IMPORTANT: You are using gradio version 4.28.3, however version 4.29.0 is available, please upgrade.
|
264 |
--------
|
265 |
Running on local URL: http://0.0.0.0:80
|
266 |
-
Running on public URL: https://
|
267 |
|
268 |
This share link expires in 72 hours. For free permanent hosting and GPU upgrades, run `gradio deploy` from Terminal to deploy to Spaces (https://huggingface.co/spaces)
|
269 |
-
['
|
270 |
-
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
|
271 |
-
To disable this warning, you can either:
|
272 |
-
- Avoid using `tokenizers` before the fork if possible
|
273 |
-
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
|
274 |
-
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
|
275 |
-
To disable this warning, you can either:
|
276 |
-
- Avoid using `tokenizers` before the fork if possible
|
277 |
-
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
|
278 |
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
|
279 |
To disable this warning, you can either:
|
280 |
- Avoid using `tokenizers` before the fork if possible
|
@@ -283,28 +81,6 @@ huggingface/tokenizers: The current process just got forked, after parallelism h
|
|
283 |
To disable this warning, you can either:
|
284 |
- Avoid using `tokenizers` before the fork if possible
|
285 |
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
|
286 |
-
/usr/local/lib/python3.9/dist-packages/optimum/bettertransformer/models/encoder_models.py:301: UserWarning: The PyTorch API of nested tensors is in prototype stage and will change in the near future. (Triggered internally at ../aten/src/ATen/NestedTensorImpl.cpp:178.)
|
287 |
-
hidden_states = torch._nested_tensor_from_mask(hidden_states, ~attention_mask)
|
288 |
-
/home/aliasgarov/copyright_checker/predictors.py:247: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
|
289 |
-
probas = F.softmax(tensor_logits).detach().cpu().numpy()
|
290 |
-
/home/aliasgarov/copyright_checker/predictors.py:247: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
|
291 |
-
probas = F.softmax(tensor_logits).detach().cpu().numpy()
|
292 |
-
WARNING: Invalid HTTP request received.
|
293 |
-
PLAGIARISM PROCESSING TIME: 21.636404959950596
|
294 |
-
Original BC scores: AI: 0.9994519352912903, HUMAN: 0.0005480951513163745
|
295 |
-
Calibration BC scores: AI: 0.8166666666666667, HUMAN: 0.18333333333333335
|
296 |
-
Input Text: sMultiple factors are helping Russias military advance, including a delay in American weaponry and Moscows technological innovations on the battlefield. /s
|
297 |
-
Models to Test: ['OpenAI GPT', 'Mistral', 'CLAUDE', 'Gemini', 'Grammar Enhancer']
|
298 |
-
Original BC scores: AI: 0.9994519352912903, HUMAN: 0.0005480951513163745
|
299 |
-
Calibration BC scores: AI: 0.8166666666666667, HUMAN: 0.18333333333333335
|
300 |
-
Starting MC
|
301 |
-
MC Score: {'OpenAI GPT': 4.604327676532164e-11, 'Mistral': 1.2912608567245758e-11, 'CLAUDE': 3.2367452925959875e-11, 'Gemini': 3.2201130588138284e-11, 'Grammar Enhancer': 0.8166666665431422}
|
302 |
-
Original BC scores: AI: 0.9994519352912903, HUMAN: 0.0005480951513163745
|
303 |
-
Calibration BC scores: AI: 0.8166666666666667, HUMAN: 0.18333333333333335
|
304 |
-
Input Text: sMultiple factors are helping Russias military advance, including a delay in American weaponry and Moscows technological innovations on the battlefield. /s
|
305 |
-
{'Multiple factors are helping Russia’s military advance, including a delay in American weaponry and Moscow’s technological innovations on the battlefield.': 0.016840771452527072} bc
|
306 |
-
{'Multiple factors are helping Russia’s military advance, including a delay in American weaponry and Moscow’s technological innovations on the battlefield.': 0.0006924518071430807} quillbot
|
307 |
-
['Ilya and OpenAI are going to part ways.', 'This is very sad to me; Ilya is easily one of the greatest minds of our generation, a guiding light of our field, and a dear friend.', 'His brilliance and vision are well known; his warmth and compassion are less well known but no less important.', 'OpenAI would not be what it is without him.', 'Although he has something personally meaningful he is going to go work on, I am forever grateful for what he did here and committed to finishing the mission we started together.', 'I am happy that for so long I got to be close to such genuinely remarkable genius, and someone so focused on getting to the best future for humanity.', 'Jakub is going to be our new Chief Scientist.', 'Jakub is also easily one of the greatest minds of our generation; I am thrilled he is taking the baton here.', 'He has run many of our most important projects, and I am very confident he will lead us to make rapid and safe progress towards our mission of ensuring that AGI benefits everyone.']
|
308 |
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
|
309 |
To disable this warning, you can either:
|
310 |
- Avoid using `tokenizers` before the fork if possible
|
@@ -321,32 +97,3 @@ huggingface/tokenizers: The current process just got forked, after parallelism h
|
|
321 |
To disable this warning, you can either:
|
322 |
- Avoid using `tokenizers` before the fork if possible
|
323 |
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
|
324 |
-
WARNING: Invalid HTTP request received.
|
325 |
-
WARNING: Invalid HTTP request received.
|
326 |
-
/home/aliasgarov/copyright_checker/predictors.py:247: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
|
327 |
-
probas = F.softmax(tensor_logits).detach().cpu().numpy()
|
328 |
-
/home/aliasgarov/copyright_checker/predictors.py:247: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.
|
329 |
-
probas = F.softmax(tensor_logits).detach().cpu().numpy()
|
330 |
-
WARNING: Invalid HTTP request received.
|
331 |
-
WARNING: Invalid HTTP request received.
|
332 |
-
WARNING: Invalid HTTP request received.
|
333 |
-
WARNING: Invalid HTTP request received.
|
334 |
-
WARNING: Invalid HTTP request received.
|
335 |
-
WARNING: Invalid HTTP request received.
|
336 |
-
WARNING: Invalid HTTP request received.
|
337 |
-
WARNING: Invalid HTTP request received.
|
338 |
-
WARNING: Invalid HTTP request received.
|
339 |
-
WARNING: Invalid HTTP request received.
|
340 |
-
WARNING: Invalid HTTP request received.
|
341 |
-
WARNING: Invalid HTTP request received.
|
342 |
-
WARNING: Invalid HTTP request received.
|
343 |
-
WARNING: Invalid HTTP request received.
|
344 |
-
WARNING: Invalid HTTP request received.
|
345 |
-
WARNING: Invalid HTTP request received.
|
346 |
-
WARNING: Invalid HTTP request received.
|
347 |
-
WARNING: Invalid HTTP request received.
|
348 |
-
WARNING: Invalid HTTP request received.
|
349 |
-
WARNING: Invalid HTTP request received.
|
350 |
-
WARNING: Invalid HTTP request received.
|
351 |
-
WARNING: Invalid HTTP request received.
|
352 |
-
WARNING: Invalid HTTP request received.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
/usr/lib/python3/dist-packages/requests/__init__.py:87: RequestsDependencyWarning: urllib3 (2.2.1) or chardet (4.0.0) doesn't match a supported version!
|
2 |
warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported "
|
3 |
+
2024-05-15 14:59:12.669109: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
|
4 |
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
|
5 |
+
2024-05-15 14:59:14.457459: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
[nltk_data] Downloading package punkt to /root/nltk_data...
|
7 |
[nltk_data] Package punkt is already up-to-date!
|
8 |
[nltk_data] Downloading package stopwords to /root/nltk_data...
|
|
|
25 |
Collecting en_core_web_sm==2.3.1
|
26 |
Using cached en_core_web_sm-2.3.1-py3-none-any.whl
|
27 |
Requirement already satisfied: spacy<2.4.0,>=2.3.0 in /usr/local/lib/python3.9/dist-packages (from en_core_web_sm==2.3.1) (2.3.9)
|
|
|
|
|
28 |
Requirement already satisfied: preshed<3.1.0,>=3.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (3.0.9)
|
|
|
|
|
29 |
Requirement already satisfied: thinc<7.5.0,>=7.4.1 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (7.4.6)
|
30 |
+
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (52.0.0)
|
|
|
|
|
|
|
31 |
Requirement already satisfied: murmurhash<1.1.0,>=0.28.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.10)
|
32 |
+
Requirement already satisfied: cymem<2.1.0,>=2.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.0.8)
|
33 |
Requirement already satisfied: blis<0.8.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.7.11)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
Requirement already satisfied: wasabi<1.1.0,>=0.4.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (0.10.1)
|
35 |
+
Requirement already satisfied: tqdm<5.0.0,>=4.38.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (4.66.2)
|
36 |
Requirement already satisfied: numpy>=1.15.0 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.26.4)
|
37 |
+
Requirement already satisfied: requests<3.0.0,>=2.13.0 in /usr/lib/python3/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (2.25.1)
|
38 |
Requirement already satisfied: plac<1.2.0,>=0.9.6 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.1.3)
|
|
|
|
|
|
|
39 |
Requirement already satisfied: catalogue<1.1.0,>=0.0.7 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.2)
|
40 |
+
Requirement already satisfied: srsly<1.1.0,>=1.0.2 in /usr/local/lib/python3.9/dist-packages (from spacy<2.4.0,>=2.3.0->en_core_web_sm==2.3.1) (1.0.7)
|
|
|
|
|
|
|
|
|
41 |
[38;5;2m✔ Download and installation successful[0m
|
42 |
You can now load the model via spacy.load('en_core_web_sm')
|
43 |
+
Traceback (most recent call last):
|
44 |
+
File "/usr/local/lib/python3.9/dist-packages/gradio/queueing.py", line 527, in process_events
|
45 |
+
response = await route_utils.call_process_api(
|
46 |
+
File "/usr/local/lib/python3.9/dist-packages/gradio/route_utils.py", line 270, in call_process_api
|
47 |
+
output = await app.get_blocks().process_api(
|
48 |
+
File "/usr/local/lib/python3.9/dist-packages/gradio/blocks.py", line 1847, in process_api
|
49 |
+
result = await self.call_function(
|
50 |
+
File "/usr/local/lib/python3.9/dist-packages/gradio/blocks.py", line 1433, in call_function
|
51 |
+
prediction = await anyio.to_thread.run_sync(
|
52 |
+
File "/usr/local/lib/python3.9/dist-packages/anyio/to_thread.py", line 56, in run_sync
|
53 |
+
return await get_async_backend().run_sync_in_worker_thread(
|
54 |
+
File "/usr/local/lib/python3.9/dist-packages/anyio/_backends/_asyncio.py", line 2144, in run_sync_in_worker_thread
|
55 |
+
return await future
|
56 |
+
File "/usr/local/lib/python3.9/dist-packages/anyio/_backends/_asyncio.py", line 851, in run
|
57 |
+
result = context.run(func, *args)
|
58 |
+
File "/usr/local/lib/python3.9/dist-packages/gradio/utils.py", line 788, in wrapper
|
59 |
+
response = f(*args, **kwargs)
|
60 |
+
File "/home/aliasgarov/copyright_checker/predictors.py", line 119, in update
|
61 |
+
corrected_text, corrections = correct_text(text, bias_checker, bias_corrector)
|
62 |
+
NameError: name 'bias_checker' is not defined
|
63 |
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
|
64 |
To disable this warning, you can either:
|
65 |
- Avoid using `tokenizers` before the fork if possible
|
66 |
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
|
67 |
/usr/local/lib/python3.9/dist-packages/torch/cuda/__init__.py:619: UserWarning: Can't initialize NVML
|
68 |
warnings.warn("Can't initialize NVML")
|
|
|
69 |
IMPORTANT: You are using gradio version 4.28.3, however version 4.29.0 is available, please upgrade.
|
70 |
--------
|
71 |
Running on local URL: http://0.0.0.0:80
|
72 |
+
Running on public URL: https://a5b565cd42a2675e81.gradio.live
|
73 |
|
74 |
This share link expires in 72 hours. For free permanent hosting and GPU upgrades, run `gradio deploy` from Terminal to deploy to Spaces (https://huggingface.co/spaces)
|
75 |
+
["OpenAI's chief scientist and co-founder, Ilya Sutskever, is leaving the artificial-intelligence company about six months after he voted to fire Chief Executive Sam Altman only to say he regretted the move days later"]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
76 |
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
|
77 |
To disable this warning, you can either:
|
78 |
- Avoid using `tokenizers` before the fork if possible
|
|
|
81 |
To disable this warning, you can either:
|
82 |
- Avoid using `tokenizers` before the fork if possible
|
83 |
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
84 |
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
|
85 |
To disable this warning, you can either:
|
86 |
- Avoid using `tokenizers` before the fork if possible
|
|
|
97 |
To disable this warning, you can either:
|
98 |
- Avoid using `tokenizers` before the fork if possible
|
99 |
- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
predictors.py
CHANGED
@@ -78,16 +78,16 @@ text_bc_model = BetterTransformer.transform(text_bc_model)
|
|
78 |
text_mc_model = BetterTransformer.transform(text_mc_model)
|
79 |
quillbot_model = BetterTransformer.transform(quillbot_model)
|
80 |
|
81 |
-
|
82 |
-
|
83 |
-
|
84 |
-
|
85 |
-
|
86 |
-
|
87 |
-
|
88 |
-
|
89 |
-
|
90 |
-
|
91 |
|
92 |
# model score calibration
|
93 |
iso_reg = joblib.load("isotonic_regression_model.joblib")
|
|
|
78 |
text_mc_model = BetterTransformer.transform(text_mc_model)
|
79 |
quillbot_model = BetterTransformer.transform(quillbot_model)
|
80 |
|
81 |
+
bias_model_checker = AutoModelForSequenceClassification.from_pretrained(bias_checker_model_name)
|
82 |
+
tokenizer = AutoTokenizer.from_pretrained(bias_checker_model_name)
|
83 |
+
bias_model_checker = BetterTransformer.transform(bias_model_checker, keep_original_model=False)
|
84 |
+
bias_checker = pipeline(
|
85 |
+
"text-classification",
|
86 |
+
model=bias_checker_model_name,
|
87 |
+
tokenizer=bias_checker_model_name,
|
88 |
+
)
|
89 |
+
gc.collect()
|
90 |
+
bias_corrector = pipeline( "text2text-generation", model=bias_corrector_model_name, accelerator="ort")
|
91 |
|
92 |
# model score calibration
|
93 |
iso_reg = joblib.load("isotonic_regression_model.joblib")
|