asrimanth commited on
Commit
f6c8286
1 Parent(s): f807dbc

End of training

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +3 -0
  2. README.md +21 -0
  3. checkpoint-1000/optimizer.bin +3 -0
  4. checkpoint-1000/pytorch_model.bin +3 -0
  5. checkpoint-1000/random_states_0.pkl +3 -0
  6. checkpoint-1000/scaler.pt +3 -0
  7. checkpoint-1000/scheduler.bin +3 -0
  8. checkpoint-1500/optimizer.bin +3 -0
  9. checkpoint-1500/pytorch_model.bin +3 -0
  10. checkpoint-1500/random_states_0.pkl +3 -0
  11. checkpoint-1500/scaler.pt +3 -0
  12. checkpoint-1500/scheduler.bin +3 -0
  13. checkpoint-2000/optimizer.bin +3 -0
  14. checkpoint-2000/pytorch_model.bin +3 -0
  15. checkpoint-2000/random_states_0.pkl +3 -0
  16. checkpoint-2000/scaler.pt +3 -0
  17. checkpoint-2000/scheduler.bin +3 -0
  18. checkpoint-2500/optimizer.bin +3 -0
  19. checkpoint-2500/pytorch_model.bin +3 -0
  20. checkpoint-2500/random_states_0.pkl +3 -0
  21. checkpoint-2500/scaler.pt +3 -0
  22. checkpoint-2500/scheduler.bin +3 -0
  23. checkpoint-3000/optimizer.bin +3 -0
  24. checkpoint-3000/pytorch_model.bin +3 -0
  25. checkpoint-3000/random_states_0.pkl +3 -0
  26. checkpoint-3000/scaler.pt +3 -0
  27. checkpoint-3000/scheduler.bin +3 -0
  28. checkpoint-500/optimizer.bin +3 -0
  29. checkpoint-500/pytorch_model.bin +3 -0
  30. checkpoint-500/random_states_0.pkl +3 -0
  31. checkpoint-500/scaler.pt +3 -0
  32. checkpoint-500/scheduler.bin +3 -0
  33. image_0.png +0 -0
  34. image_1.png +0 -0
  35. image_2.png +0 -0
  36. image_3.png +0 -0
  37. pytorch_lora_weights.bin +3 -0
  38. wandb/debug-internal.log +0 -0
  39. wandb/debug.log +27 -0
  40. wandb/run-20230710_174519-01ve2s6v/files/config.yaml +39 -0
  41. wandb/run-20230710_174519-01ve2s6v/files/output.log +7 -0
  42. wandb/run-20230710_174519-01ve2s6v/files/requirements.txt +165 -0
  43. wandb/run-20230710_174519-01ve2s6v/files/wandb-metadata.json +127 -0
  44. wandb/run-20230710_174519-01ve2s6v/files/wandb-summary.json +1 -0
  45. wandb/run-20230710_174519-01ve2s6v/logs/debug-internal.log +0 -0
  46. wandb/run-20230710_174519-01ve2s6v/logs/debug.log +27 -0
  47. wandb/run-20230710_174519-01ve2s6v/run-01ve2s6v.wandb +3 -0
  48. wandb/run-20230710_174519-cqch7cwa/files/config.yaml +187 -0
  49. wandb/run-20230710_174519-cqch7cwa/files/media/images/validation_1000_30bd8224b667e1b90a1e.png +0 -0
  50. wandb/run-20230710_174519-cqch7cwa/files/media/images/validation_1000_42e11b031b9a15f888e4.png +0 -0
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wandb/run-20230710_174519-01ve2s6v/run-01ve2s6v.wandb filter=lfs diff=lfs merge=lfs -text
37
+ wandb/run-20230710_174519-cqch7cwa/run-cqch7cwa.wandb filter=lfs diff=lfs merge=lfs -text
38
+ wandb/run-20230710_174519-qoix60bv/run-qoix60bv.wandb filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ license: creativeml-openrail-m
4
+ base_model: runwayml/stable-diffusion-v1-5
5
+ tags:
6
+ - stable-diffusion
7
+ - stable-diffusion-diffusers
8
+ - text-to-image
9
+ - diffusers
10
+ - lora
11
+ inference: true
12
+ ---
13
+
14
+ # LoRA text2image fine-tuning - asrimanth/person-thumbs-up-lora-no-cap
15
+ These are LoRA adaption weights for runwayml/stable-diffusion-v1-5. The weights were fine-tuned on the Custom dataset dataset. You can find some example images in the following.
16
+
17
+ ![img_0](./image_0.png)
18
+ ![img_1](./image_1.png)
19
+ ![img_2](./image_2.png)
20
+ ![img_3](./image_3.png)
21
+
checkpoint-1000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:746c62e5113737ef8b02c00326e3b5cdfdcb8f95180c91b330f3f3c94926def9
3
+ size 6591685
checkpoint-1000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ac51b3392264e0485a7949271a2e16544b09a8a3473a6ef318301eb391aa62c
3
+ size 3285965
checkpoint-1000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf626b00f0c0a21c1ee2f5631667c9175d425a61775d2fd75a66283361ef7cec
3
+ size 16719
checkpoint-1000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68cff80b680ddf6e7abbef98b5f336b97f9b5963e2209307f639383870e8cc71
3
+ size 557
checkpoint-1000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61808704787602dc93c00aa68d75b73415f1c6bfacfbeb97c14702c3e19cf9f2
3
+ size 563
checkpoint-1500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77a55cfe2d202d7658e497a13e57ebcdacaf27c1a51ee31ac50c39152218cbc0
3
+ size 6591685
checkpoint-1500/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c253af5591389039973a2ea17bd87d63f574889d6f050aca303f6ccd8a61b7f9
3
+ size 3285965
checkpoint-1500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:863bba4edd1ca97dc90e081725c55f3be037349b333f06678cdaad3d9501ed3e
3
+ size 16719
checkpoint-1500/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:203a72d6c29f42a0e2964fdddc8d7a98df1eccee78fea9de0fa416613390f5c6
3
+ size 557
checkpoint-1500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e89f6844bbfca1eeee319b6e2d6f3fd9d9df58879544d455482ad32bb3ad4df6
3
+ size 563
checkpoint-2000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea56829d7f2a0745c8705572654104d7505898e7a2c20c053a47094f61c80742
3
+ size 6591685
checkpoint-2000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec021ef8f4568ce10dddacb2b0c7f00374040c528a45e63243c4e86ca0ab60f1
3
+ size 3285965
checkpoint-2000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fd21d1d8596fbae8788b9060a2386fe2636e634218aff7b075ff9bbd70231b3
3
+ size 16719
checkpoint-2000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd2de9749828adacdf103bf6e9592702bb7067a2c1df27dd62ab38c1eb8c070f
3
+ size 557
checkpoint-2000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00f6a413e75a46219e92e544cae3353a4577b60b728f62a720befd9033032acd
3
+ size 563
checkpoint-2500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c60850f36c09c2c47453cfdbc5a6b2f7a9628541c7ae7d666b4f84deb979b97f
3
+ size 6591685
checkpoint-2500/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:270f1731d5c47267878cc217744089f3c5db3353c18496879195fe47339aae49
3
+ size 3285965
checkpoint-2500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39f28d64d6912cc34eaaa0bb0cd0819897835a6aae2d96c1c19aa15468bb16c1
3
+ size 16719
checkpoint-2500/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fbcebc8f5487b0c117b5dd47f2ea304af3eebf408d297118d9307e1223927e1
3
+ size 557
checkpoint-2500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df00e6adcececef3aacd6addf17f6c85eae895b25ab91e673e9360b2a7f6ee86
3
+ size 563
checkpoint-3000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6f85f4ee5aa71cbe6aec52bac688f8d18060f9fdc6ed4a24e18bcf229ac97dc
3
+ size 6591685
checkpoint-3000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71d305a98f9f86636331f5c95d26bea6c72360da043b17d6f93ac1032e1e82c7
3
+ size 3285965
checkpoint-3000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d12ebe8df23ebc27557c11a245e66d96279ac2a3ff7c8b15aac109084cee9e94
3
+ size 16719
checkpoint-3000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb1f9398b77268202e8e1465734a63d123b1ef11c27f20f2473677e9883a6869
3
+ size 557
checkpoint-3000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:807d7d063623792ef9b8c66da2304691aed0888f14ad6e70a728375b9c90fe9d
3
+ size 563
checkpoint-500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3779ad72103a5ce89a74b972afd45f12daf23fa910cc7e57fbd14d70d3e6667
3
+ size 6591685
checkpoint-500/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:263b95b3ee399314308e8ddb3d24a1292e981f2798a39f8d235998de2aaaa27c
3
+ size 3285965
checkpoint-500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e670e6c021cd2c9f45d7068757a1a37ed11280498d332d84fc9b52b93dce0e7d
3
+ size 16719
checkpoint-500/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3f196a54202bb4ba1220e8c59f42f9cda0702d68ea83147d814c2fb2f36b8f2
3
+ size 557
checkpoint-500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bc0a22eef87625080f7c833304b0844a5e9a28131b856c035cff1e8f8d56121
3
+ size 563
image_0.png ADDED
image_1.png ADDED
image_2.png ADDED
image_3.png ADDED
pytorch_lora_weights.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea51f54bb3ef6955de5e69c155b00b4ead4a80836046b7fb20b4c33f6a11b823
3
+ size 3287771
wandb/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-07-10 17:45:19,956 INFO MainThread:538851 [wandb_setup.py:_flush():76] Current SDK version is 0.15.4
2
+ 2023-07-10 17:45:19,956 INFO MainThread:538851 [wandb_setup.py:_flush():76] Configure stats pid to 538851
3
+ 2023-07-10 17:45:19,957 INFO MainThread:538851 [wandb_setup.py:_flush():76] Loading settings from /u/sragas/.config/wandb/settings
4
+ 2023-07-10 17:45:19,957 INFO MainThread:538851 [wandb_setup.py:_flush():76] Loading settings from /nfs/nfs2/home/sragas/demo/wandb/settings
5
+ 2023-07-10 17:45:19,957 INFO MainThread:538851 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2023-07-10 17:45:19,957 INFO MainThread:538851 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2023-07-10 17:45:19,957 INFO MainThread:538851 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': 'train_text_to_image_lora.py', 'program': 'train_text_to_image_lora.py'}
8
+ 2023-07-10 17:45:19,957 INFO MainThread:538851 [wandb_init.py:_log_setup():507] Logging user logs to /l/vision/v5/sragas/easel_ai/models_no_cap/wandb/run-20230710_174519-qoix60bv/logs/debug.log
9
+ 2023-07-10 17:45:19,957 INFO MainThread:538851 [wandb_init.py:_log_setup():508] Logging internal logs to /l/vision/v5/sragas/easel_ai/models_no_cap/wandb/run-20230710_174519-qoix60bv/logs/debug-internal.log
10
+ 2023-07-10 17:45:19,958 INFO MainThread:538851 [wandb_init.py:init():547] calling init triggers
11
+ 2023-07-10 17:45:19,958 INFO MainThread:538851 [wandb_init.py:init():554] wandb.init called with sweep_config: {}
12
+ config: {}
13
+ 2023-07-10 17:45:19,958 INFO MainThread:538851 [wandb_init.py:init():596] starting backend
14
+ 2023-07-10 17:45:19,958 INFO MainThread:538851 [wandb_init.py:init():600] setting up manager
15
+ 2023-07-10 17:45:19,961 INFO MainThread:538851 [backend.py:_multiprocessing_setup():106] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
16
+ 2023-07-10 17:45:19,965 INFO MainThread:538851 [wandb_init.py:init():606] backend started and connected
17
+ 2023-07-10 17:45:19,968 INFO MainThread:538851 [wandb_init.py:init():703] updated telemetry
18
+ 2023-07-10 17:45:19,969 INFO MainThread:538851 [wandb_init.py:init():736] communicating run to backend with 60.0 second timeout
19
+ 2023-07-10 17:45:20,272 INFO MainThread:538851 [wandb_run.py:_on_init():2176] communicating current version
20
+ 2023-07-10 17:45:20,355 INFO MainThread:538851 [wandb_run.py:_on_init():2185] got version response upgrade_message: "wandb version 0.15.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
21
+
22
+ 2023-07-10 17:45:20,355 INFO MainThread:538851 [wandb_init.py:init():787] starting run threads in backend
23
+ 2023-07-10 17:45:20,484 INFO MainThread:538851 [wandb_run.py:_console_start():2155] atexit reg
24
+ 2023-07-10 17:45:20,484 INFO MainThread:538851 [wandb_run.py:_redirect():2010] redirect: SettingsConsole.WRAP_RAW
25
+ 2023-07-10 17:45:20,485 INFO MainThread:538851 [wandb_run.py:_redirect():2075] Wrapping output streams.
26
+ 2023-07-10 17:45:20,485 INFO MainThread:538851 [wandb_run.py:_redirect():2100] Redirects installed.
27
+ 2023-07-10 17:45:20,486 INFO MainThread:538851 [wandb_init.py:init():828] run started, returning control to user process
wandb/run-20230710_174519-01ve2s6v/files/config.yaml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.8.10
7
+ cli_version: 0.15.4
8
+ framework: huggingface
9
+ huggingface_version: 4.30.2
10
+ is_jupyter_run: false
11
+ is_kaggle_kernel: true
12
+ start_time: 1689025519.940496
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 11
17
+ - 41
18
+ - 49
19
+ - 51
20
+ - 55
21
+ - 71
22
+ - 83
23
+ 2:
24
+ - 1
25
+ - 11
26
+ - 41
27
+ - 49
28
+ - 51
29
+ - 55
30
+ - 71
31
+ - 83
32
+ 3:
33
+ - 23
34
+ 4: 3.8.10
35
+ 5: 0.15.4
36
+ 6: 4.30.2
37
+ 8:
38
+ - 2
39
+ - 5
wandb/run-20230710_174519-01ve2s6v/files/output.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ 07/10/2023 17:45:20 - INFO - __main__ - Distributed environment: MULTI_GPU Backend: nccl
2
+ Num processes: 3
3
+ Process index: 2
4
+ Local process index: 2
5
+ Device: cuda:2
6
+ Mixed precision type: fp16
7
+ Resolving data files: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 229/229 [00:00<00:00, 58947.81it/s]
wandb/run-20230710_174519-01ve2s6v/files/requirements.txt ADDED
@@ -0,0 +1,165 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.4.0
2
+ accelerate==0.20.3
3
+ aiohttp==3.8.4
4
+ aiosignal==1.3.1
5
+ anyio==3.7.0
6
+ appdirs==1.4.4
7
+ argon2-cffi-bindings==21.2.0
8
+ argon2-cffi==21.3.0
9
+ asttokens==2.2.1
10
+ async-lru==2.0.2
11
+ async-timeout==4.0.2
12
+ attrs==23.1.0
13
+ babel==2.12.1
14
+ backcall==0.2.0
15
+ beautifulsoup4==4.12.2
16
+ bleach==6.0.0
17
+ cachetools==5.3.1
18
+ certifi==2023.5.7
19
+ cffi==1.15.1
20
+ charset-normalizer==3.1.0
21
+ click==8.1.3
22
+ cmake==3.26.4
23
+ comm==0.1.3
24
+ datasets==2.13.1
25
+ debugpy==1.6.7
26
+ decorator==5.1.1
27
+ defusedxml==0.7.1
28
+ diffusers==0.18.0.dev0
29
+ dill==0.3.6
30
+ docker-pycreds==0.4.0
31
+ exceptiongroup==1.1.2
32
+ executing==1.2.0
33
+ fastjsonschema==2.17.1
34
+ filelock==3.12.2
35
+ frozenlist==1.3.3
36
+ fsspec==2023.6.0
37
+ ftfy==6.1.1
38
+ gitdb==4.0.10
39
+ gitpython==3.1.31
40
+ google-auth-oauthlib==1.0.0
41
+ google-auth==2.21.0
42
+ grpcio==1.56.0
43
+ huggingface-hub==0.15.1
44
+ idna==3.4
45
+ importlib-metadata==6.7.0
46
+ importlib-resources==5.12.0
47
+ ipykernel==6.24.0
48
+ ipython==8.12.2
49
+ jedi==0.18.2
50
+ jinja2==3.1.2
51
+ json5==0.9.14
52
+ jsonschema==4.17.3
53
+ jupyter-client==8.3.0
54
+ jupyter-core==5.3.1
55
+ jupyter-events==0.6.3
56
+ jupyter-lsp==2.2.0
57
+ jupyter-server-terminals==0.4.4
58
+ jupyter-server==2.7.0
59
+ jupyterlab-pygments==0.2.2
60
+ jupyterlab-server==2.23.0
61
+ jupyterlab==4.0.2
62
+ lit==16.0.6
63
+ markdown==3.4.3
64
+ markupsafe==2.1.3
65
+ matplotlib-inline==0.1.6
66
+ mistune==3.0.1
67
+ mpmath==1.3.0
68
+ multidict==6.0.4
69
+ multiprocess==0.70.14
70
+ mypy-extensions==1.0.0
71
+ nbclient==0.8.0
72
+ nbconvert==7.6.0
73
+ nbformat==5.9.0
74
+ nest-asyncio==1.5.6
75
+ networkx==3.1
76
+ notebook-shim==0.2.3
77
+ numpy==1.24.4
78
+ nvidia-cublas-cu11==11.10.3.66
79
+ nvidia-cuda-cupti-cu11==11.7.101
80
+ nvidia-cuda-nvrtc-cu11==11.7.99
81
+ nvidia-cuda-runtime-cu11==11.7.99
82
+ nvidia-cudnn-cu11==8.5.0.96
83
+ nvidia-cufft-cu11==10.9.0.58
84
+ nvidia-curand-cu11==10.2.10.91
85
+ nvidia-cusolver-cu11==11.4.0.1
86
+ nvidia-cusparse-cu11==11.7.4.91
87
+ nvidia-nccl-cu11==2.14.3
88
+ nvidia-nvtx-cu11==11.7.91
89
+ oauthlib==3.2.2
90
+ overrides==7.3.1
91
+ packaging==23.1
92
+ pandas==2.0.3
93
+ pandocfilters==1.5.0
94
+ parso==0.8.3
95
+ pathtools==0.1.2
96
+ pexpect==4.8.0
97
+ pickleshare==0.7.5
98
+ pillow==10.0.0
99
+ pip==20.0.2
100
+ pkg-resources==0.0.0
101
+ pkgutil-resolve-name==1.3.10
102
+ platformdirs==3.8.0
103
+ prometheus-client==0.17.0
104
+ prompt-toolkit==3.0.39
105
+ protobuf==4.23.3
106
+ psutil==5.9.5
107
+ ptyprocess==0.7.0
108
+ pure-eval==0.2.2
109
+ pyarrow==12.0.1
110
+ pyasn1-modules==0.3.0
111
+ pyasn1==0.5.0
112
+ pycparser==2.21
113
+ pygments==2.15.1
114
+ pyre-extensions==0.0.29
115
+ pyrsistent==0.19.3
116
+ python-dateutil==2.8.2
117
+ python-json-logger==2.0.7
118
+ pytz==2023.3
119
+ pyyaml==6.0
120
+ pyzmq==25.1.0
121
+ regex==2023.6.3
122
+ requests-oauthlib==1.3.1
123
+ requests==2.31.0
124
+ rfc3339-validator==0.1.4
125
+ rfc3986-validator==0.1.1
126
+ rsa==4.9
127
+ safetensors==0.3.1
128
+ send2trash==1.8.2
129
+ sentry-sdk==1.27.0
130
+ setproctitle==1.3.2
131
+ setuptools==44.0.0
132
+ six==1.16.0
133
+ smmap==5.0.0
134
+ sniffio==1.3.0
135
+ soupsieve==2.4.1
136
+ stack-data==0.6.2
137
+ sympy==1.12
138
+ tensorboard-data-server==0.7.1
139
+ tensorboard==2.13.0
140
+ terminado==0.17.1
141
+ tinycss2==1.2.1
142
+ tokenizers==0.13.3
143
+ tomli==2.0.1
144
+ torch==2.0.1
145
+ torchaudio==2.0.2
146
+ torchvision==0.15.2
147
+ tornado==6.3.2
148
+ tqdm==4.65.0
149
+ traitlets==5.9.0
150
+ transformers==4.30.2
151
+ triton==2.0.0
152
+ typing-extensions==4.7.1
153
+ typing-inspect==0.9.0
154
+ tzdata==2023.3
155
+ urllib3==2.0.3
156
+ wandb==0.15.4
157
+ wcwidth==0.2.6
158
+ webencodings==0.5.1
159
+ websocket-client==1.6.1
160
+ werkzeug==2.3.6
161
+ wheel==0.40.0
162
+ xformers==0.0.20
163
+ xxhash==3.2.0
164
+ yarl==1.9.2
165
+ zipp==3.15.0
wandb/run-20230710_174519-01ve2s6v/files/wandb-metadata.json ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.0-73-generic-x86_64-with-glibc2.29",
3
+ "python": "3.8.10",
4
+ "heartbeatAt": "2023-07-10T21:45:20.460414",
5
+ "startedAt": "2023-07-10T21:45:19.811914",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "--pretrained_model_name_or_path=runwayml/stable-diffusion-v1-5",
10
+ "--train_data_dir=/l/vision/v5/sragas/easel_ai/thumbs_up_no_cap_dataset/",
11
+ "--resolution=512",
12
+ "--center_crop",
13
+ "--random_flip",
14
+ "--train_batch_size=2",
15
+ "--gradient_accumulation_steps=4",
16
+ "--num_train_epochs=300",
17
+ "--learning_rate=1e-6",
18
+ "--max_grad_norm=1",
19
+ "--lr_scheduler=cosine",
20
+ "--lr_warmup_steps=500",
21
+ "--output_dir=/l/vision/v5/sragas/easel_ai/models_no_cap/",
22
+ "--checkpointing_steps=500",
23
+ "--report_to=wandb",
24
+ "--validation_prompt=<tom_cruise> #thumbsup",
25
+ "--seed=15",
26
+ "--push_to_hub",
27
+ "--hub_model_id=person-thumbs-up-lora-no-cap"
28
+ ],
29
+ "state": "running",
30
+ "program": "train_text_to_image_lora.py",
31
+ "codePath": "train_text_to_image_lora.py",
32
+ "host": "magnemite",
33
+ "username": "sragas",
34
+ "executable": "/nfs/blitzle/home/data/vision5/sragas/easel_venv/bin/python3",
35
+ "cpu_count": 6,
36
+ "cpu_count_logical": 12,
37
+ "cpu_freq": {
38
+ "current": 1880.4280833333335,
39
+ "min": 1200.0,
40
+ "max": 3900.0
41
+ },
42
+ "cpu_freq_per_core": [
43
+ {
44
+ "current": 2354.544,
45
+ "min": 1200.0,
46
+ "max": 3900.0
47
+ },
48
+ {
49
+ "current": 1200.0,
50
+ "min": 1200.0,
51
+ "max": 3900.0
52
+ },
53
+ {
54
+ "current": 1300.0,
55
+ "min": 1200.0,
56
+ "max": 3900.0
57
+ },
58
+ {
59
+ "current": 1805.495,
60
+ "min": 1200.0,
61
+ "max": 3900.0
62
+ },
63
+ {
64
+ "current": 1200.0,
65
+ "min": 1200.0,
66
+ "max": 3900.0
67
+ },
68
+ {
69
+ "current": 2470.242,
70
+ "min": 1200.0,
71
+ "max": 3900.0
72
+ },
73
+ {
74
+ "current": 1200.0,
75
+ "min": 1200.0,
76
+ "max": 3900.0
77
+ },
78
+ {
79
+ "current": 1200.0,
80
+ "min": 1200.0,
81
+ "max": 3900.0
82
+ },
83
+ {
84
+ "current": 2168.124,
85
+ "min": 1200.0,
86
+ "max": 3900.0
87
+ },
88
+ {
89
+ "current": 2166.389,
90
+ "min": 1200.0,
91
+ "max": 3900.0
92
+ },
93
+ {
94
+ "current": 1200.0,
95
+ "min": 1200.0,
96
+ "max": 3900.0
97
+ },
98
+ {
99
+ "current": 1300.0,
100
+ "min": 1200.0,
101
+ "max": 3900.0
102
+ }
103
+ ],
104
+ "disk": {
105
+ "total": 116.31898498535156,
106
+ "used": 59.75596237182617
107
+ },
108
+ "gpu": "NVIDIA GeForce GTX TITAN X",
109
+ "gpu_count": 3,
110
+ "gpu_devices": [
111
+ {
112
+ "name": "NVIDIA GeForce GTX TITAN X",
113
+ "memory_total": 12884901888
114
+ },
115
+ {
116
+ "name": "NVIDIA GeForce GTX TITAN X",
117
+ "memory_total": 12884901888
118
+ },
119
+ {
120
+ "name": "NVIDIA GeForce GTX TITAN X",
121
+ "memory_total": 12884901888
122
+ }
123
+ ],
124
+ "memory": {
125
+ "total": 62.69387435913086
126
+ }
127
+ }
wandb/run-20230710_174519-01ve2s6v/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
wandb/run-20230710_174519-01ve2s6v/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20230710_174519-01ve2s6v/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-07-10 17:45:19,929 INFO MainThread:538852 [wandb_setup.py:_flush():76] Current SDK version is 0.15.4
2
+ 2023-07-10 17:45:19,929 INFO MainThread:538852 [wandb_setup.py:_flush():76] Configure stats pid to 538852
3
+ 2023-07-10 17:45:19,929 INFO MainThread:538852 [wandb_setup.py:_flush():76] Loading settings from /u/sragas/.config/wandb/settings
4
+ 2023-07-10 17:45:19,929 INFO MainThread:538852 [wandb_setup.py:_flush():76] Loading settings from /nfs/nfs2/home/sragas/demo/wandb/settings
5
+ 2023-07-10 17:45:19,929 INFO MainThread:538852 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2023-07-10 17:45:19,929 INFO MainThread:538852 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2023-07-10 17:45:19,930 INFO MainThread:538852 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': 'train_text_to_image_lora.py', 'program': 'train_text_to_image_lora.py'}
8
+ 2023-07-10 17:45:19,930 INFO MainThread:538852 [wandb_init.py:_log_setup():507] Logging user logs to /l/vision/v5/sragas/easel_ai/models_no_cap/wandb/run-20230710_174519-01ve2s6v/logs/debug.log
9
+ 2023-07-10 17:45:19,930 INFO MainThread:538852 [wandb_init.py:_log_setup():508] Logging internal logs to /l/vision/v5/sragas/easel_ai/models_no_cap/wandb/run-20230710_174519-01ve2s6v/logs/debug-internal.log
10
+ 2023-07-10 17:45:19,930 INFO MainThread:538852 [wandb_init.py:init():547] calling init triggers
11
+ 2023-07-10 17:45:19,930 INFO MainThread:538852 [wandb_init.py:init():554] wandb.init called with sweep_config: {}
12
+ config: {}
13
+ 2023-07-10 17:45:19,930 INFO MainThread:538852 [wandb_init.py:init():596] starting backend
14
+ 2023-07-10 17:45:19,930 INFO MainThread:538852 [wandb_init.py:init():600] setting up manager
15
+ 2023-07-10 17:45:19,934 INFO MainThread:538852 [backend.py:_multiprocessing_setup():106] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
16
+ 2023-07-10 17:45:19,939 INFO MainThread:538852 [wandb_init.py:init():606] backend started and connected
17
+ 2023-07-10 17:45:19,943 INFO MainThread:538852 [wandb_init.py:init():703] updated telemetry
18
+ 2023-07-10 17:45:19,944 INFO MainThread:538852 [wandb_init.py:init():736] communicating run to backend with 60.0 second timeout
19
+ 2023-07-10 17:45:20,209 INFO MainThread:538852 [wandb_run.py:_on_init():2176] communicating current version
20
+ 2023-07-10 17:45:20,400 INFO MainThread:538852 [wandb_run.py:_on_init():2185] got version response upgrade_message: "wandb version 0.15.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
21
+
22
+ 2023-07-10 17:45:20,400 INFO MainThread:538852 [wandb_init.py:init():787] starting run threads in backend
23
+ 2023-07-10 17:45:20,593 INFO MainThread:538852 [wandb_run.py:_console_start():2155] atexit reg
24
+ 2023-07-10 17:45:20,594 INFO MainThread:538852 [wandb_run.py:_redirect():2010] redirect: SettingsConsole.WRAP_RAW
25
+ 2023-07-10 17:45:20,594 INFO MainThread:538852 [wandb_run.py:_redirect():2075] Wrapping output streams.
26
+ 2023-07-10 17:45:20,594 INFO MainThread:538852 [wandb_run.py:_redirect():2100] Redirects installed.
27
+ 2023-07-10 17:45:20,595 INFO MainThread:538852 [wandb_init.py:init():828] run started, returning control to user process
wandb/run-20230710_174519-01ve2s6v/run-01ve2s6v.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c2f32aa537003af2bda504bc679cc9717175ecc6ebfd535e53b34b878a1f6fc
3
+ size 2459037
wandb/run-20230710_174519-cqch7cwa/files/config.yaml ADDED
@@ -0,0 +1,187 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.8.10
7
+ cli_version: 0.15.4
8
+ framework: huggingface
9
+ huggingface_version: 4.30.2
10
+ is_jupyter_run: false
11
+ is_kaggle_kernel: true
12
+ start_time: 1689025519.94471
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 11
17
+ - 41
18
+ - 49
19
+ - 51
20
+ - 55
21
+ - 71
22
+ - 83
23
+ 2:
24
+ - 1
25
+ - 11
26
+ - 41
27
+ - 49
28
+ - 51
29
+ - 55
30
+ - 71
31
+ - 83
32
+ 3:
33
+ - 23
34
+ - 24
35
+ 4: 3.8.10
36
+ 5: 0.15.4
37
+ 6: 4.30.2
38
+ 8:
39
+ - 2
40
+ - 5
41
+ pretrained_model_name_or_path:
42
+ desc: null
43
+ value: runwayml/stable-diffusion-v1-5
44
+ revision:
45
+ desc: null
46
+ value: null
47
+ dataset_name:
48
+ desc: null
49
+ value: null
50
+ dataset_config_name:
51
+ desc: null
52
+ value: null
53
+ train_data_dir:
54
+ desc: null
55
+ value: /l/vision/v5/sragas/easel_ai/thumbs_up_no_cap_dataset/
56
+ image_column:
57
+ desc: null
58
+ value: image
59
+ caption_column:
60
+ desc: null
61
+ value: text
62
+ validation_prompt:
63
+ desc: null
64
+ value: '<tom_cruise> #thumbsup'
65
+ num_validation_images:
66
+ desc: null
67
+ value: 4
68
+ validation_epochs:
69
+ desc: null
70
+ value: 1
71
+ max_train_samples:
72
+ desc: null
73
+ value: null
74
+ output_dir:
75
+ desc: null
76
+ value: /l/vision/v5/sragas/easel_ai/models_no_cap/
77
+ cache_dir:
78
+ desc: null
79
+ value: /l/vision/v5/sragas/hf_models/
80
+ seed:
81
+ desc: null
82
+ value: 15
83
+ resolution:
84
+ desc: null
85
+ value: 512
86
+ center_crop:
87
+ desc: null
88
+ value: true
89
+ random_flip:
90
+ desc: null
91
+ value: true
92
+ train_batch_size:
93
+ desc: null
94
+ value: 2
95
+ num_train_epochs:
96
+ desc: null
97
+ value: 300
98
+ max_train_steps:
99
+ desc: null
100
+ value: 3000
101
+ gradient_accumulation_steps:
102
+ desc: null
103
+ value: 4
104
+ gradient_checkpointing:
105
+ desc: null
106
+ value: false
107
+ learning_rate:
108
+ desc: null
109
+ value: 1.0e-06
110
+ scale_lr:
111
+ desc: null
112
+ value: false
113
+ lr_scheduler:
114
+ desc: null
115
+ value: cosine
116
+ lr_warmup_steps:
117
+ desc: null
118
+ value: 500
119
+ snr_gamma:
120
+ desc: null
121
+ value: null
122
+ use_8bit_adam:
123
+ desc: null
124
+ value: false
125
+ allow_tf32:
126
+ desc: null
127
+ value: false
128
+ dataloader_num_workers:
129
+ desc: null
130
+ value: 0
131
+ adam_beta1:
132
+ desc: null
133
+ value: 0.9
134
+ adam_beta2:
135
+ desc: null
136
+ value: 0.999
137
+ adam_weight_decay:
138
+ desc: null
139
+ value: 0.01
140
+ adam_epsilon:
141
+ desc: null
142
+ value: 1.0e-08
143
+ max_grad_norm:
144
+ desc: null
145
+ value: 1.0
146
+ push_to_hub:
147
+ desc: null
148
+ value: true
149
+ hub_token:
150
+ desc: null
151
+ value: null
152
+ prediction_type:
153
+ desc: null
154
+ value: null
155
+ hub_model_id:
156
+ desc: null
157
+ value: person-thumbs-up-lora-no-cap
158
+ logging_dir:
159
+ desc: null
160
+ value: logs
161
+ mixed_precision:
162
+ desc: null
163
+ value: null
164
+ report_to:
165
+ desc: null
166
+ value: wandb
167
+ local_rank:
168
+ desc: null
169
+ value: 0
170
+ checkpointing_steps:
171
+ desc: null
172
+ value: 500
173
+ checkpoints_total_limit:
174
+ desc: null
175
+ value: null
176
+ resume_from_checkpoint:
177
+ desc: null
178
+ value: null
179
+ enable_xformers_memory_efficient_attention:
180
+ desc: null
181
+ value: false
182
+ noise_offset:
183
+ desc: null
184
+ value: 0
185
+ rank:
186
+ desc: null
187
+ value: 4
wandb/run-20230710_174519-cqch7cwa/files/media/images/validation_1000_30bd8224b667e1b90a1e.png ADDED
wandb/run-20230710_174519-cqch7cwa/files/media/images/validation_1000_42e11b031b9a15f888e4.png ADDED