dsmueller commited on
Commit
46d1a84
1 Parent(s): fe25472

Update requirements.txt with new package versions

Browse files
Files changed (2) hide show
  1. requirements.txt +98 -23
  2. train_llm.ipynb +22 -11
requirements.txt CHANGED
@@ -1,16 +1,19 @@
1
  absl-py==2.0.0
2
- accelerate==0.25.0
3
  aiofiles==23.2.1
4
  aiohttp==3.9.1
 
5
  aiosignal==1.3.1
6
  albumentations==1.3.1
7
  alembic==1.13.0
8
  altair==5.2.0
9
  annotated-types==0.6.0
10
  anyio==3.7.1
 
11
  arrow==1.3.0
 
12
  attrs==23.1.0
13
- autotrain-advanced==0.6.59
14
  bitsandbytes==0.41.0
15
  Brotli==1.1.0
16
  cachetools==5.3.2
@@ -20,73 +23,128 @@ click==8.1.7
20
  cmaes==0.10.0
21
  codecarbon==2.2.3
22
  colorlog==6.8.0
 
23
  contourpy==1.2.0
24
  cycler==0.12.1
 
25
  datasets==2.14.7
 
 
26
  diffusers==0.21.4
27
  dill==0.3.7
 
 
28
  docstring-parser==0.15
29
  einops==0.6.1
30
  evaluate==0.3.0
 
31
  fastapi==0.104.1
32
  ffmpy==0.3.1
33
  filelock==3.13.1
34
- fonttools==4.47.0
35
- frozenlist==1.4.1
36
  fsspec==2023.10.0
37
  fuzzywuzzy==0.18.0
38
- google-auth==2.25.2
39
- google-auth-oauthlib==1.2.0
 
 
 
 
 
 
40
  gradio==3.41.0
41
  gradio_client==0.5.0
42
- grpcio==1.60.0
 
 
43
  h11==0.14.0
44
- hf_transfer==0.1.4
45
  httpcore==1.0.2
46
  httpx==0.25.2
47
  huggingface-hub==0.19.4
48
  idna==3.6
49
- imageio==2.33.1
50
  importlib-metadata==7.0.0
51
  importlib-resources==6.1.1
52
  inflate64==1.0.0
53
  invisible-watermark==0.2.0
54
  ipadic==1.0.0
 
 
 
 
55
  Jinja2==3.1.2
56
  jiwer==3.0.2
57
  joblib==1.3.1
 
 
 
58
  jsonschema==4.20.0
59
  jsonschema-specifications==2023.11.2
 
 
 
60
  kiwisolver==1.4.5
 
 
 
61
  lazy_loader==0.3
62
  loguru==0.7.0
63
  Mako==1.3.0
64
  Markdown==3.5.1
65
  markdown-it-py==3.0.0
66
  MarkupSafe==2.1.3
 
67
  matplotlib==3.8.2
 
68
  mdurl==0.1.2
 
69
  mpmath==1.3.0
70
  multidict==6.0.4
71
  multiprocess==0.70.15
72
  multivolumefile==0.2.3
 
 
73
  networkx==3.2.1
74
  nltk==3.8.1
75
- numpy==1.26.2
 
 
 
 
 
 
 
 
 
 
 
 
76
  oauthlib==3.2.2
77
- # opencv-python==4.8.1.78
 
78
  opencv-python-headless==4.8.1.78
79
  optuna==3.3.0
80
  orjson==3.9.10
81
  packaging==23.1
82
  pandas==2.1.4
83
- peft==0.7.1
 
 
 
 
84
  Pillow==10.0.0
 
 
 
 
85
  protobuf==4.23.4
86
- psutil==5.9.7
 
 
87
  py-cpuinfo==9.0.0
88
  py7zr==0.20.6
89
- pyarrow==14.0.2
90
  pyarrow-hotfix==0.6
91
  pyasn1==0.5.1
92
  pyasn1-modules==0.3.0
@@ -96,26 +154,28 @@ pydantic==2.4.2
96
  pydantic_core==2.10.1
97
  pydub==0.25.1
98
  Pygments==2.17.2
99
- pyngrok==7.0.3
100
  pynvml==11.5.0
101
  pyparsing==3.1.1
 
102
  pyppmd==1.0.0
103
  python-dateutil==2.8.2
 
104
  python-multipart==0.0.6
105
  pytz==2023.3.post1
106
  PyWavelets==1.5.0
107
  PyYAML==6.0.1
 
108
  pyzstd==0.15.9
109
  qudida==0.0.4
110
  rapidfuzz==2.13.7
111
- referencing==0.32.0
112
  regex==2023.10.3
113
  requests==2.31.0
114
  requests-oauthlib==1.3.1
115
  responses==0.18.0
116
  rich==13.7.0
117
  rouge-score==0.1.2
118
- rpds-py==0.15.2
119
  rsa==4.9
120
  sacremoses==0.0.53
121
  safetensors==0.4.1
@@ -128,29 +188,44 @@ shtab==1.6.5
128
  six==1.16.0
129
  sniffio==1.3.0
130
  SQLAlchemy==2.0.23
 
 
131
  starlette==0.27.0
132
  sympy==1.12
 
133
  tensorboard==2.15.1
134
  tensorboard-data-server==0.7.2
135
  texttable==1.7.0
136
  threadpoolctl==3.2.0
137
- tifffile==2023.12.9
138
  tiktoken==0.5.1
139
  tokenizers==0.15.0
140
  toolz==0.12.0
141
- torch==2.1.2
 
142
  tqdm==4.65.0
143
- transformers==4.36.1
 
 
144
  trl==0.7.4
 
145
  types-python-dateutil==2.8.19.14
146
- typing_extensions==4.9.0
 
 
 
 
147
  tyro==0.6.0
148
  tzdata==2023.3
149
- urllib3==2.1.0
150
  uvicorn==0.22.0
 
151
  websockets==11.0.3
152
  Werkzeug==2.3.6
 
 
 
153
  xgboost==1.7.6
154
  xxhash==3.4.1
155
- yarl==1.9.4
156
  zipp==3.17.0
 
1
  absl-py==2.0.0
2
+ accelerate==0.24.0
3
  aiofiles==23.2.1
4
  aiohttp==3.9.1
5
+ aioitertools==0.11.0
6
  aiosignal==1.3.1
7
  albumentations==1.3.1
8
  alembic==1.13.0
9
  altair==5.2.0
10
  annotated-types==0.6.0
11
  anyio==3.7.1
12
+ appnope==0.1.3
13
  arrow==1.3.0
14
+ asttokens==2.4.1
15
  attrs==23.1.0
16
+ autotrain-advanced==0.6.51
17
  bitsandbytes==0.41.0
18
  Brotli==1.1.0
19
  cachetools==5.3.2
 
23
  cmaes==0.10.0
24
  codecarbon==2.2.3
25
  colorlog==6.8.0
26
+ comm==0.2.0
27
  contourpy==1.2.0
28
  cycler==0.12.1
29
+ dataclasses-json==0.6.3
30
  datasets==2.14.7
31
+ debugpy==1.8.0
32
+ decorator==5.1.1
33
  diffusers==0.21.4
34
  dill==0.3.7
35
+ distro==1.8.0
36
+ dnspython==2.4.2
37
  docstring-parser==0.15
38
  einops==0.6.1
39
  evaluate==0.3.0
40
+ executing==2.0.1
41
  fastapi==0.104.1
42
  ffmpy==0.3.1
43
  filelock==3.13.1
44
+ fonttools==4.46.0
45
+ frozenlist==1.4.0
46
  fsspec==2023.10.0
47
  fuzzywuzzy==0.18.0
48
+ google-api-core==2.14.0
49
+ google-auth==2.24.0
50
+ google-auth-oauthlib==1.1.0
51
+ google-cloud-core==2.3.3
52
+ google-cloud-storage==2.13.0
53
+ google-crc32c==1.5.0
54
+ google-resumable-media==2.6.0
55
+ googleapis-common-protos==1.61.0
56
  gradio==3.41.0
57
  gradio_client==0.5.0
58
+ greenlet==3.0.2
59
+ grpcio==1.59.3
60
+ gunicorn==21.2.0
61
  h11==0.14.0
 
62
  httpcore==1.0.2
63
  httpx==0.25.2
64
  huggingface-hub==0.19.4
65
  idna==3.6
66
+ imageio==2.33.0
67
  importlib-metadata==7.0.0
68
  importlib-resources==6.1.1
69
  inflate64==1.0.0
70
  invisible-watermark==0.2.0
71
  ipadic==1.0.0
72
+ ipykernel==6.27.1
73
+ ipython==8.18.1
74
+ ipywidgets==8.1.1
75
+ jedi==0.19.1
76
  Jinja2==3.1.2
77
  jiwer==3.0.2
78
  joblib==1.3.1
79
+ jsonlines==4.0.0
80
+ jsonpatch==1.33
81
+ jsonpointer==2.4
82
  jsonschema==4.20.0
83
  jsonschema-specifications==2023.11.2
84
+ jupyter_client==8.6.0
85
+ jupyter_core==5.5.0
86
+ jupyterlab-widgets==3.0.9
87
  kiwisolver==1.4.5
88
+ langchain==0.0.345
89
+ langchain-core==0.0.9
90
+ langsmith==0.0.69
91
  lazy_loader==0.3
92
  loguru==0.7.0
93
  Mako==1.3.0
94
  Markdown==3.5.1
95
  markdown-it-py==3.0.0
96
  MarkupSafe==2.1.3
97
+ marshmallow==3.20.1
98
  matplotlib==3.8.2
99
+ matplotlib-inline==0.1.6
100
  mdurl==0.1.2
101
+ mmh3==3.1.0
102
  mpmath==1.3.0
103
  multidict==6.0.4
104
  multiprocess==0.70.15
105
  multivolumefile==0.2.3
106
+ mypy-extensions==1.0.0
107
+ nest-asyncio==1.5.8
108
  networkx==3.2.1
109
  nltk==3.8.1
110
+ numpy==1.25.2
111
+ nvidia-cublas-cu12==12.1.3.1
112
+ nvidia-cuda-cupti-cu12==12.1.105
113
+ nvidia-cuda-nvrtc-cu12==12.1.105
114
+ nvidia-cuda-runtime-cu12==12.1.105
115
+ nvidia-cudnn-cu12==8.9.2.26
116
+ nvidia-cufft-cu12==11.0.2.54
117
+ nvidia-curand-cu12==10.3.2.106
118
+ nvidia-cusolver-cu12==11.4.5.107
119
+ nvidia-cusparse-cu12==12.1.0.106
120
+ nvidia-nccl-cu12==2.18.1
121
+ nvidia-nvjitlink-cu12==12.3.101
122
+ nvidia-nvtx-cu12==12.1.105
123
  oauthlib==3.2.2
124
+ openai==1.3.7
125
+ opencv-python==4.8.1.78
126
  opencv-python-headless==4.8.1.78
127
  optuna==3.3.0
128
  orjson==3.9.10
129
  packaging==23.1
130
  pandas==2.1.4
131
+ pandas-stubs==2.0.3.230814
132
+ parso==0.8.3
133
+ peft==0.6.2
134
+ pexpect==4.9.0
135
+ pi==0.1.2
136
  Pillow==10.0.0
137
+ pinecone-client==2.2.4
138
+ pinecone-text==0.7.1
139
+ platformdirs==4.1.0
140
+ prompt-toolkit==3.0.41
141
  protobuf==4.23.4
142
+ psutil==5.9.6
143
+ ptyprocess==0.7.0
144
+ pure-eval==0.2.2
145
  py-cpuinfo==9.0.0
146
  py7zr==0.20.6
147
+ pyarrow==11.0.0
148
  pyarrow-hotfix==0.6
149
  pyasn1==0.5.1
150
  pyasn1-modules==0.3.0
 
154
  pydantic_core==2.10.1
155
  pydub==0.25.1
156
  Pygments==2.17.2
 
157
  pynvml==11.5.0
158
  pyparsing==3.1.1
159
+ pypdf==3.17.1
160
  pyppmd==1.0.0
161
  python-dateutil==2.8.2
162
+ python-dotenv==1.0.0
163
  python-multipart==0.0.6
164
  pytz==2023.3.post1
165
  PyWavelets==1.5.0
166
  PyYAML==6.0.1
167
+ pyzmq==25.1.2
168
  pyzstd==0.15.9
169
  qudida==0.0.4
170
  rapidfuzz==2.13.7
171
+ referencing==0.31.1
172
  regex==2023.10.3
173
  requests==2.31.0
174
  requests-oauthlib==1.3.1
175
  responses==0.18.0
176
  rich==13.7.0
177
  rouge-score==0.1.2
178
+ rpds-py==0.13.2
179
  rsa==4.9
180
  sacremoses==0.0.53
181
  safetensors==0.4.1
 
188
  six==1.16.0
189
  sniffio==1.3.0
190
  SQLAlchemy==2.0.23
191
+ sse-starlette==1.8.2
192
+ stack-data==0.6.3
193
  starlette==0.27.0
194
  sympy==1.12
195
+ tenacity==8.2.3
196
  tensorboard==2.15.1
197
  tensorboard-data-server==0.7.2
198
  texttable==1.7.0
199
  threadpoolctl==3.2.0
200
+ tifffile==2023.9.26
201
  tiktoken==0.5.1
202
  tokenizers==0.15.0
203
  toolz==0.12.0
204
+ torch==2.1.1
205
+ tornado==6.4
206
  tqdm==4.65.0
207
+ traitlets==5.14.0
208
+ transformers==4.35.2
209
+ triton==2.1.0
210
  trl==0.7.4
211
+ types-jsonschema==4.20.0.0
212
  types-python-dateutil==2.8.19.14
213
+ types-pytz==2023.3.1.1
214
+ types-PyYAML==6.0.12.12
215
+ types-tqdm==4.66.0.5
216
+ typing-inspect==0.9.0
217
+ typing_extensions==4.8.0
218
  tyro==0.6.0
219
  tzdata==2023.3
220
+ urllib3==2.0.7
221
  uvicorn==0.22.0
222
+ wcwidth==0.2.12
223
  websockets==11.0.3
224
  Werkzeug==2.3.6
225
+ wget==3.2
226
+ widgetsnbextension==4.0.9
227
+ wrapt==1.16.0
228
  xgboost==1.7.6
229
  xxhash==3.4.1
230
+ yarl==1.9.3
231
  zipp==3.17.0
train_llm.ipynb CHANGED
@@ -2,7 +2,7 @@
2
  "cells": [
3
  {
4
  "cell_type": "code",
5
- "execution_count": 32,
6
  "metadata": {},
7
  "outputs": [],
8
  "source": [
@@ -17,7 +17,7 @@
17
  },
18
  {
19
  "cell_type": "code",
20
- "execution_count": 33,
21
  "metadata": {},
22
  "outputs": [],
23
  "source": [
@@ -35,7 +35,7 @@
35
  },
36
  {
37
  "cell_type": "code",
38
- "execution_count": 34,
39
  "metadata": {},
40
  "outputs": [
41
  {
@@ -57,7 +57,7 @@
57
  },
58
  {
59
  "cell_type": "code",
60
- "execution_count": 37,
61
  "metadata": {},
62
  "outputs": [
63
  {
@@ -201,28 +201,39 @@
201
  },
202
  {
203
  "cell_type": "code",
204
- "execution_count": 49,
205
  "metadata": {},
206
  "outputs": [
207
  {
208
  "name": "stderr",
209
  "output_type": "stream",
210
  "text": [
211
- "⚠️ WARNING | 2023-12-22 10:39:42 | autotrain.cli.run_dreambooth:<module>:14 - ❌ Some DreamBooth components are missing! Please run `autotrain setup` to install it. Ignore this warning if you are not using DreamBooth or running `autotrain setup` already.\n",
212
- "usage: autotrain <command> [<args>]\n",
213
- "AutoTrain advanced CLI: error: unrecognized arguments: --batch_size 2\n"
 
 
 
 
 
 
 
 
 
 
 
214
  ]
215
  },
216
  {
217
  "ename": "CalledProcessError",
218
- "evalue": "Command '\nautotrain llm --train --trainer sft --project_name ./llms/ams_data_train-100_6abb23dc-cb9d-428e-9079-e47deee0edd9 --model mistralai/Mistral-7B-v0.1 --data_path ./fine_tune_data/ --train_split train_data --valid_split validation_data --repo_id ai-aerospace/ams-data-train-100-4601c8c8-0903-4f18-a6e8-1d2a40a697ce --push_to_hub --token HUGGINGFACE_TOKEN --block_size 2242 --model_max_length 1121 --logging_steps -1 --evaluation_strategy epoch --save_total_limit 1 --save_strategy epoch --fp16 --lr 3e-05 --num_train_epochs 3 --batch_size 2 --warmup_ratio 0.1 --gradient_accumulation 1 --optimizer adamw_torch --scheduler linear --weight_decay 0 --max_grad_norm 1 --seed 42 --use_int4 --use-peft --lora_r 16 --lora_alpha 32 --lora_dropout 0.05\n' returned non-zero exit status 2.",
219
  "output_type": "error",
220
  "traceback": [
221
  "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
222
  "\u001b[0;31mCalledProcessError\u001b[0m Traceback (most recent call last)",
223
- "Cell \u001b[0;32mIn[49], line 40\u001b[0m\n\u001b[1;32m 4\u001b[0m command\u001b[38;5;241m=\u001b[39m\u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\"\"\u001b[39m\n\u001b[1;32m 5\u001b[0m \u001b[38;5;124mautotrain llm --train \u001b[39m\u001b[38;5;130;01m\\\u001b[39;00m\n\u001b[1;32m 6\u001b[0m \u001b[38;5;124m --trainer sft \u001b[39m\u001b[38;5;130;01m\\\u001b[39;00m\n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 36\u001b[0m \u001b[38;5;124m --lora_dropout \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mmodel_params[\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mlora_dropout\u001b[39m\u001b[38;5;124m'\u001b[39m]\u001b[38;5;132;01m}\u001b[39;00m\n\u001b[1;32m 37\u001b[0m \u001b[38;5;124m\"\"\"\u001b[39m\n\u001b[1;32m 39\u001b[0m \u001b[38;5;66;03m# Use subprocess.run() to execute the command\u001b[39;00m\n\u001b[0;32m---> 40\u001b[0m \u001b[43msubprocess\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mrun\u001b[49m\u001b[43m(\u001b[49m\u001b[43mcommand\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mshell\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mTrue\u001b[39;49;00m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mcheck\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mTrue\u001b[39;49;00m\u001b[43m)\u001b[49m\n",
224
  "File \u001b[0;32m/usr/lib/python3.11/subprocess.py:571\u001b[0m, in \u001b[0;36mrun\u001b[0;34m(input, capture_output, timeout, check, *popenargs, **kwargs)\u001b[0m\n\u001b[1;32m 569\u001b[0m retcode \u001b[38;5;241m=\u001b[39m process\u001b[38;5;241m.\u001b[39mpoll()\n\u001b[1;32m 570\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m check \u001b[38;5;129;01mand\u001b[39;00m retcode:\n\u001b[0;32m--> 571\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m CalledProcessError(retcode, process\u001b[38;5;241m.\u001b[39margs,\n\u001b[1;32m 572\u001b[0m output\u001b[38;5;241m=\u001b[39mstdout, stderr\u001b[38;5;241m=\u001b[39mstderr)\n\u001b[1;32m 573\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m CompletedProcess(process\u001b[38;5;241m.\u001b[39margs, retcode, stdout, stderr)\n",
225
- "\u001b[0;31mCalledProcessError\u001b[0m: Command '\nautotrain llm --train --trainer sft --project_name ./llms/ams_data_train-100_6abb23dc-cb9d-428e-9079-e47deee0edd9 --model mistralai/Mistral-7B-v0.1 --data_path ./fine_tune_data/ --train_split train_data --valid_split validation_data --repo_id ai-aerospace/ams-data-train-100-4601c8c8-0903-4f18-a6e8-1d2a40a697ce --push_to_hub --token HUGGINGFACE_TOKEN --block_size 2242 --model_max_length 1121 --logging_steps -1 --evaluation_strategy epoch --save_total_limit 1 --save_strategy epoch --fp16 --lr 3e-05 --num_train_epochs 3 --batch_size 2 --warmup_ratio 0.1 --gradient_accumulation 1 --optimizer adamw_torch --scheduler linear --weight_decay 0 --max_grad_norm 1 --seed 42 --use_int4 --use-peft --lora_r 16 --lora_alpha 32 --lora_dropout 0.05\n' returned non-zero exit status 2."
226
  ]
227
  }
228
  ],
 
2
  "cells": [
3
  {
4
  "cell_type": "code",
5
+ "execution_count": 51,
6
  "metadata": {},
7
  "outputs": [],
8
  "source": [
 
17
  },
18
  {
19
  "cell_type": "code",
20
+ "execution_count": 52,
21
  "metadata": {},
22
  "outputs": [],
23
  "source": [
 
35
  },
36
  {
37
  "cell_type": "code",
38
+ "execution_count": 53,
39
  "metadata": {},
40
  "outputs": [
41
  {
 
57
  },
58
  {
59
  "cell_type": "code",
60
+ "execution_count": 54,
61
  "metadata": {},
62
  "outputs": [
63
  {
 
201
  },
202
  {
203
  "cell_type": "code",
204
+ "execution_count": 50,
205
  "metadata": {},
206
  "outputs": [
207
  {
208
  "name": "stderr",
209
  "output_type": "stream",
210
  "text": [
211
+ "⚠️ WARNING | 2023-12-22 10:41:00 | autotrain.cli.run_dreambooth:<module>:14 - ❌ Some DreamBooth components are missing! Please run `autotrain setup` to install it. Ignore this warning if you are not using DreamBooth or running `autotrain setup` already.\n",
212
+ "Traceback (most recent call last):\n",
213
+ " File \"/home/dsmueller/Repositories/HuggingFace/autotrain-playground/.venv/bin/autotrain\", line 8, in <module>\n",
214
+ " sys.exit(main())\n",
215
+ " ^^^^^^\n",
216
+ " File \"/home/dsmueller/Repositories/HuggingFace/autotrain-playground/.venv/lib/python3.11/site-packages/autotrain/cli/autotrain.py\", line 47, in main\n",
217
+ " command = args.func(args)\n",
218
+ " ^^^^^^^^^^^^^^^\n",
219
+ " File \"/home/dsmueller/Repositories/HuggingFace/autotrain-playground/.venv/lib/python3.11/site-packages/autotrain/cli/run_llm.py\", line 14, in run_llm_command_factory\n",
220
+ " return RunAutoTrainLLMCommand(args)\n",
221
+ " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n",
222
+ " File \"/home/dsmueller/Repositories/HuggingFace/autotrain-playground/.venv/lib/python3.11/site-packages/autotrain/cli/run_llm.py\", line 473, in __init__\n",
223
+ " raise ValueError(\"No GPU/MPS device found. LLM training requires an accelerator\")\n",
224
+ "ValueError: No GPU/MPS device found. LLM training requires an accelerator\n"
225
  ]
226
  },
227
  {
228
  "ename": "CalledProcessError",
229
+ "evalue": "Command '\nautotrain llm --train --trainer sft --project_name ./llms/ams_data_train-100_6abb23dc-cb9d-428e-9079-e47deee0edd9 --model mistralai/Mistral-7B-v0.1 --data_path ./fine_tune_data/ --train_split train_data --valid_split validation_data --repo_id ai-aerospace/ams-data-train-100-4601c8c8-0903-4f18-a6e8-1d2a40a697ce --push_to_hub --token HUGGINGFACE_TOKEN --block_size 2242 --model_max_length 1121 --logging_steps -1 --evaluation_strategy epoch --save_total_limit 1 --save_strategy epoch --fp16 --lr 3e-05 --num_train_epochs 3 --train_batch_size 2 --warmup_ratio 0.1 --gradient_accumulation 1 --optimizer adamw_torch --scheduler linear --weight_decay 0 --max_grad_norm 1 --seed 42 --use_int4 --use-peft --lora_r 16 --lora_alpha 32 --lora_dropout 0.05\n' returned non-zero exit status 1.",
230
  "output_type": "error",
231
  "traceback": [
232
  "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
233
  "\u001b[0;31mCalledProcessError\u001b[0m Traceback (most recent call last)",
234
+ "Cell \u001b[0;32mIn[50], line 40\u001b[0m\n\u001b[1;32m 4\u001b[0m command\u001b[38;5;241m=\u001b[39m\u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\"\"\u001b[39m\n\u001b[1;32m 5\u001b[0m \u001b[38;5;124mautotrain llm --train \u001b[39m\u001b[38;5;130;01m\\\u001b[39;00m\n\u001b[1;32m 6\u001b[0m \u001b[38;5;124m --trainer sft \u001b[39m\u001b[38;5;130;01m\\\u001b[39;00m\n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 36\u001b[0m \u001b[38;5;124m --lora_dropout \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mmodel_params[\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mlora_dropout\u001b[39m\u001b[38;5;124m'\u001b[39m]\u001b[38;5;132;01m}\u001b[39;00m\n\u001b[1;32m 37\u001b[0m \u001b[38;5;124m\"\"\"\u001b[39m\n\u001b[1;32m 39\u001b[0m \u001b[38;5;66;03m# Use subprocess.run() to execute the command\u001b[39;00m\n\u001b[0;32m---> 40\u001b[0m \u001b[43msubprocess\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mrun\u001b[49m\u001b[43m(\u001b[49m\u001b[43mcommand\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mshell\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mTrue\u001b[39;49;00m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mcheck\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mTrue\u001b[39;49;00m\u001b[43m)\u001b[49m\n",
235
  "File \u001b[0;32m/usr/lib/python3.11/subprocess.py:571\u001b[0m, in \u001b[0;36mrun\u001b[0;34m(input, capture_output, timeout, check, *popenargs, **kwargs)\u001b[0m\n\u001b[1;32m 569\u001b[0m retcode \u001b[38;5;241m=\u001b[39m process\u001b[38;5;241m.\u001b[39mpoll()\n\u001b[1;32m 570\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m check \u001b[38;5;129;01mand\u001b[39;00m retcode:\n\u001b[0;32m--> 571\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m CalledProcessError(retcode, process\u001b[38;5;241m.\u001b[39margs,\n\u001b[1;32m 572\u001b[0m output\u001b[38;5;241m=\u001b[39mstdout, stderr\u001b[38;5;241m=\u001b[39mstderr)\n\u001b[1;32m 573\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m CompletedProcess(process\u001b[38;5;241m.\u001b[39margs, retcode, stdout, stderr)\n",
236
+ "\u001b[0;31mCalledProcessError\u001b[0m: Command '\nautotrain llm --train --trainer sft --project_name ./llms/ams_data_train-100_6abb23dc-cb9d-428e-9079-e47deee0edd9 --model mistralai/Mistral-7B-v0.1 --data_path ./fine_tune_data/ --train_split train_data --valid_split validation_data --repo_id ai-aerospace/ams-data-train-100-4601c8c8-0903-4f18-a6e8-1d2a40a697ce --push_to_hub --token HUGGINGFACE_TOKEN --block_size 2242 --model_max_length 1121 --logging_steps -1 --evaluation_strategy epoch --save_total_limit 1 --save_strategy epoch --fp16 --lr 3e-05 --num_train_epochs 3 --train_batch_size 2 --warmup_ratio 0.1 --gradient_accumulation 1 --optimizer adamw_torch --scheduler linear --weight_decay 0 --max_grad_norm 1 --seed 42 --use_int4 --use-peft --lora_r 16 --lora_alpha 32 --lora_dropout 0.05\n' returned non-zero exit status 1."
237
  ]
238
  }
239
  ],