ACCC1380 commited on
Commit
073a0a1
1 Parent(s): 7520adb

Upload face004/logs/GPU-使用率-温度检测20240529141058/wandb/run-20240529_141120-82eq2cmy/files/wandb-metadata.json with huggingface_hub

Browse files
face004/logs/GPU-使用率-温度检测20240529141058/wandb/run-20240529_141120-82eq2cmy/files/wandb-metadata.json ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.35",
3
+ "python": "3.10.12",
4
+ "heartbeatAt": "2024-05-29T14:11:21.329856",
5
+ "startedAt": "2024-05-29T14:11:20.968446",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "--config_file",
10
+ "/root/lora-scripts/frontend/lora-scripts/frontend/lora-scripts/config/autosave/20240529-141049.toml"
11
+ ],
12
+ "state": "running",
13
+ "program": "/root/lora-scripts/frontend/lora-scripts/frontend/lora-scripts/./sd-scripts/train_network.py",
14
+ "codePathLocal": "sd-scripts/train_network.py",
15
+ "codePath": "sd-scripts/train_network.py",
16
+ "git": {
17
+ "remote": "https://github.com/Akegarasu/lora-scripts",
18
+ "commit": "9c1d6a61b2f72e72aeb51f8a1e133dc975693035"
19
+ },
20
+ "email": null,
21
+ "root": "/root/lora-scripts/frontend/lora-scripts/frontend/lora-scripts",
22
+ "host": "4ac664d71f2b",
23
+ "username": "root",
24
+ "executable": "/kaggle/opt/conda/envs/venv/venv/bin/python3",
25
+ "cpu_count": 2,
26
+ "cpu_count_logical": 4,
27
+ "cpu_freq": {
28
+ "current": 2000.152,
29
+ "min": 0.0,
30
+ "max": 0.0
31
+ },
32
+ "cpu_freq_per_core": [
33
+ {
34
+ "current": 2000.152,
35
+ "min": 0.0,
36
+ "max": 0.0
37
+ },
38
+ {
39
+ "current": 2000.152,
40
+ "min": 0.0,
41
+ "max": 0.0
42
+ },
43
+ {
44
+ "current": 2000.152,
45
+ "min": 0.0,
46
+ "max": 0.0
47
+ },
48
+ {
49
+ "current": 2000.152,
50
+ "min": 0.0,
51
+ "max": 0.0
52
+ }
53
+ ],
54
+ "disk": {
55
+ "/": {
56
+ "total": 8062.387607574463,
57
+ "used": 5607.320209503174
58
+ }
59
+ },
60
+ "gpu": "Tesla T4",
61
+ "gpu_count": 2,
62
+ "gpu_devices": [
63
+ {
64
+ "name": "Tesla T4",
65
+ "memory_total": 16106127360
66
+ },
67
+ {
68
+ "name": "Tesla T4",
69
+ "memory_total": 16106127360
70
+ }
71
+ ],
72
+ "memory": {
73
+ "total": 31.357555389404297
74
+ }
75
+ }