Initial release
Browse files- .gitattributes +1 -0
- LICENSE +201 -0
- README.md +188 -0
- configs/inference.json +101 -0
- configs/inference_autoencoder.json +149 -0
- configs/logging.conf +21 -0
- configs/metadata.json +119 -0
- configs/multi_gpu_train_autoencoder.json +40 -0
- configs/multi_gpu_train_diffusion.json +16 -0
- configs/train_autoencoder.json +206 -0
- configs/train_diffusion.json +157 -0
- docs/README.md +181 -0
- docs/data_license.txt +49 -0
- models/model.pt +3 -0
- models/model_autoencoder.pt +3 -0
- models/model_autoencoder.ts +3 -0
- scripts/__init__.py +12 -0
- scripts/ldm_sampler.py +60 -0
- scripts/ldm_trainer.py +380 -0
- scripts/losses.py +50 -0
- scripts/utils.py +20 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
models/model_autoencoder.ts filter=lfs diff=lfs merge=lfs -text
|
LICENSE
ADDED
@@ -0,0 +1,201 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Apache License
|
2 |
+
Version 2.0, January 2004
|
3 |
+
http://www.apache.org/licenses/
|
4 |
+
|
5 |
+
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
6 |
+
|
7 |
+
1. Definitions.
|
8 |
+
|
9 |
+
"License" shall mean the terms and conditions for use, reproduction,
|
10 |
+
and distribution as defined by Sections 1 through 9 of this document.
|
11 |
+
|
12 |
+
"Licensor" shall mean the copyright owner or entity authorized by
|
13 |
+
the copyright owner that is granting the License.
|
14 |
+
|
15 |
+
"Legal Entity" shall mean the union of the acting entity and all
|
16 |
+
other entities that control, are controlled by, or are under common
|
17 |
+
control with that entity. For the purposes of this definition,
|
18 |
+
"control" means (i) the power, direct or indirect, to cause the
|
19 |
+
direction or management of such entity, whether by contract or
|
20 |
+
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
21 |
+
outstanding shares, or (iii) beneficial ownership of such entity.
|
22 |
+
|
23 |
+
"You" (or "Your") shall mean an individual or Legal Entity
|
24 |
+
exercising permissions granted by this License.
|
25 |
+
|
26 |
+
"Source" form shall mean the preferred form for making modifications,
|
27 |
+
including but not limited to software source code, documentation
|
28 |
+
source, and configuration files.
|
29 |
+
|
30 |
+
"Object" form shall mean any form resulting from mechanical
|
31 |
+
transformation or translation of a Source form, including but
|
32 |
+
not limited to compiled object code, generated documentation,
|
33 |
+
and conversions to other media types.
|
34 |
+
|
35 |
+
"Work" shall mean the work of authorship, whether in Source or
|
36 |
+
Object form, made available under the License, as indicated by a
|
37 |
+
copyright notice that is included in or attached to the work
|
38 |
+
(an example is provided in the Appendix below).
|
39 |
+
|
40 |
+
"Derivative Works" shall mean any work, whether in Source or Object
|
41 |
+
form, that is based on (or derived from) the Work and for which the
|
42 |
+
editorial revisions, annotations, elaborations, or other modifications
|
43 |
+
represent, as a whole, an original work of authorship. For the purposes
|
44 |
+
of this License, Derivative Works shall not include works that remain
|
45 |
+
separable from, or merely link (or bind by name) to the interfaces of,
|
46 |
+
the Work and Derivative Works thereof.
|
47 |
+
|
48 |
+
"Contribution" shall mean any work of authorship, including
|
49 |
+
the original version of the Work and any modifications or additions
|
50 |
+
to that Work or Derivative Works thereof, that is intentionally
|
51 |
+
submitted to Licensor for inclusion in the Work by the copyright owner
|
52 |
+
or by an individual or Legal Entity authorized to submit on behalf of
|
53 |
+
the copyright owner. For the purposes of this definition, "submitted"
|
54 |
+
means any form of electronic, verbal, or written communication sent
|
55 |
+
to the Licensor or its representatives, including but not limited to
|
56 |
+
communication on electronic mailing lists, source code control systems,
|
57 |
+
and issue tracking systems that are managed by, or on behalf of, the
|
58 |
+
Licensor for the purpose of discussing and improving the Work, but
|
59 |
+
excluding communication that is conspicuously marked or otherwise
|
60 |
+
designated in writing by the copyright owner as "Not a Contribution."
|
61 |
+
|
62 |
+
"Contributor" shall mean Licensor and any individual or Legal Entity
|
63 |
+
on behalf of whom a Contribution has been received by Licensor and
|
64 |
+
subsequently incorporated within the Work.
|
65 |
+
|
66 |
+
2. Grant of Copyright License. Subject to the terms and conditions of
|
67 |
+
this License, each Contributor hereby grants to You a perpetual,
|
68 |
+
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
69 |
+
copyright license to reproduce, prepare Derivative Works of,
|
70 |
+
publicly display, publicly perform, sublicense, and distribute the
|
71 |
+
Work and such Derivative Works in Source or Object form.
|
72 |
+
|
73 |
+
3. Grant of Patent License. Subject to the terms and conditions of
|
74 |
+
this License, each Contributor hereby grants to You a perpetual,
|
75 |
+
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
76 |
+
(except as stated in this section) patent license to make, have made,
|
77 |
+
use, offer to sell, sell, import, and otherwise transfer the Work,
|
78 |
+
where such license applies only to those patent claims licensable
|
79 |
+
by such Contributor that are necessarily infringed by their
|
80 |
+
Contribution(s) alone or by combination of their Contribution(s)
|
81 |
+
with the Work to which such Contribution(s) was submitted. If You
|
82 |
+
institute patent litigation against any entity (including a
|
83 |
+
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
84 |
+
or a Contribution incorporated within the Work constitutes direct
|
85 |
+
or contributory patent infringement, then any patent licenses
|
86 |
+
granted to You under this License for that Work shall terminate
|
87 |
+
as of the date such litigation is filed.
|
88 |
+
|
89 |
+
4. Redistribution. You may reproduce and distribute copies of the
|
90 |
+
Work or Derivative Works thereof in any medium, with or without
|
91 |
+
modifications, and in Source or Object form, provided that You
|
92 |
+
meet the following conditions:
|
93 |
+
|
94 |
+
(a) You must give any other recipients of the Work or
|
95 |
+
Derivative Works a copy of this License; and
|
96 |
+
|
97 |
+
(b) You must cause any modified files to carry prominent notices
|
98 |
+
stating that You changed the files; and
|
99 |
+
|
100 |
+
(c) You must retain, in the Source form of any Derivative Works
|
101 |
+
that You distribute, all copyright, patent, trademark, and
|
102 |
+
attribution notices from the Source form of the Work,
|
103 |
+
excluding those notices that do not pertain to any part of
|
104 |
+
the Derivative Works; and
|
105 |
+
|
106 |
+
(d) If the Work includes a "NOTICE" text file as part of its
|
107 |
+
distribution, then any Derivative Works that You distribute must
|
108 |
+
include a readable copy of the attribution notices contained
|
109 |
+
within such NOTICE file, excluding those notices that do not
|
110 |
+
pertain to any part of the Derivative Works, in at least one
|
111 |
+
of the following places: within a NOTICE text file distributed
|
112 |
+
as part of the Derivative Works; within the Source form or
|
113 |
+
documentation, if provided along with the Derivative Works; or,
|
114 |
+
within a display generated by the Derivative Works, if and
|
115 |
+
wherever such third-party notices normally appear. The contents
|
116 |
+
of the NOTICE file are for informational purposes only and
|
117 |
+
do not modify the License. You may add Your own attribution
|
118 |
+
notices within Derivative Works that You distribute, alongside
|
119 |
+
or as an addendum to the NOTICE text from the Work, provided
|
120 |
+
that such additional attribution notices cannot be construed
|
121 |
+
as modifying the License.
|
122 |
+
|
123 |
+
You may add Your own copyright statement to Your modifications and
|
124 |
+
may provide additional or different license terms and conditions
|
125 |
+
for use, reproduction, or distribution of Your modifications, or
|
126 |
+
for any such Derivative Works as a whole, provided Your use,
|
127 |
+
reproduction, and distribution of the Work otherwise complies with
|
128 |
+
the conditions stated in this License.
|
129 |
+
|
130 |
+
5. Submission of Contributions. Unless You explicitly state otherwise,
|
131 |
+
any Contribution intentionally submitted for inclusion in the Work
|
132 |
+
by You to the Licensor shall be under the terms and conditions of
|
133 |
+
this License, without any additional terms or conditions.
|
134 |
+
Notwithstanding the above, nothing herein shall supersede or modify
|
135 |
+
the terms of any separate license agreement you may have executed
|
136 |
+
with Licensor regarding such Contributions.
|
137 |
+
|
138 |
+
6. Trademarks. This License does not grant permission to use the trade
|
139 |
+
names, trademarks, service marks, or product names of the Licensor,
|
140 |
+
except as required for reasonable and customary use in describing the
|
141 |
+
origin of the Work and reproducing the content of the NOTICE file.
|
142 |
+
|
143 |
+
7. Disclaimer of Warranty. Unless required by applicable law or
|
144 |
+
agreed to in writing, Licensor provides the Work (and each
|
145 |
+
Contributor provides its Contributions) on an "AS IS" BASIS,
|
146 |
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
147 |
+
implied, including, without limitation, any warranties or conditions
|
148 |
+
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
149 |
+
PARTICULAR PURPOSE. You are solely responsible for determining the
|
150 |
+
appropriateness of using or redistributing the Work and assume any
|
151 |
+
risks associated with Your exercise of permissions under this License.
|
152 |
+
|
153 |
+
8. Limitation of Liability. In no event and under no legal theory,
|
154 |
+
whether in tort (including negligence), contract, or otherwise,
|
155 |
+
unless required by applicable law (such as deliberate and grossly
|
156 |
+
negligent acts) or agreed to in writing, shall any Contributor be
|
157 |
+
liable to You for damages, including any direct, indirect, special,
|
158 |
+
incidental, or consequential damages of any character arising as a
|
159 |
+
result of this License or out of the use or inability to use the
|
160 |
+
Work (including but not limited to damages for loss of goodwill,
|
161 |
+
work stoppage, computer failure or malfunction, or any and all
|
162 |
+
other commercial damages or losses), even if such Contributor
|
163 |
+
has been advised of the possibility of such damages.
|
164 |
+
|
165 |
+
9. Accepting Warranty or Additional Liability. While redistributing
|
166 |
+
the Work or Derivative Works thereof, You may choose to offer,
|
167 |
+
and charge a fee for, acceptance of support, warranty, indemnity,
|
168 |
+
or other liability obligations and/or rights consistent with this
|
169 |
+
License. However, in accepting such obligations, You may act only
|
170 |
+
on Your own behalf and on Your sole responsibility, not on behalf
|
171 |
+
of any other Contributor, and only if You agree to indemnify,
|
172 |
+
defend, and hold each Contributor harmless for any liability
|
173 |
+
incurred by, or claims asserted against, such Contributor by reason
|
174 |
+
of your accepting any such warranty or additional liability.
|
175 |
+
|
176 |
+
END OF TERMS AND CONDITIONS
|
177 |
+
|
178 |
+
APPENDIX: How to apply the Apache License to your work.
|
179 |
+
|
180 |
+
To apply the Apache License to your work, attach the following
|
181 |
+
boilerplate notice, with the fields enclosed by brackets "[]"
|
182 |
+
replaced with your own identifying information. (Don't include
|
183 |
+
the brackets!) The text should be enclosed in the appropriate
|
184 |
+
comment syntax for the file format. We also recommend that a
|
185 |
+
file or class name and description of purpose be included on the
|
186 |
+
same "printed page" as the copyright notice for easier
|
187 |
+
identification within third-party archives.
|
188 |
+
|
189 |
+
Copyright [yyyy] [name of copyright owner]
|
190 |
+
|
191 |
+
Licensed under the Apache License, Version 2.0 (the "License");
|
192 |
+
you may not use this file except in compliance with the License.
|
193 |
+
You may obtain a copy of the License at
|
194 |
+
|
195 |
+
http://www.apache.org/licenses/LICENSE-2.0
|
196 |
+
|
197 |
+
Unless required by applicable law or agreed to in writing, software
|
198 |
+
distributed under the License is distributed on an "AS IS" BASIS,
|
199 |
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
200 |
+
See the License for the specific language governing permissions and
|
201 |
+
limitations under the License.
|
README.md
ADDED
@@ -0,0 +1,188 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- monai
|
4 |
+
- medical
|
5 |
+
library_name: monai
|
6 |
+
license: apache-2.0
|
7 |
+
---
|
8 |
+
# Model Overview
|
9 |
+
A pre-trained model for volumetric (3D) Brats MRI 3D Latent Diffusion Generative Model.
|
10 |
+
|
11 |
+
This model is trained on BraTS 2016 and 2017 data from [Medical Decathlon](http://medicaldecathlon.com/), using the Latent diffusion model [1].
|
12 |
+
|
13 |
+
![model workflow](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_network.png)
|
14 |
+
|
15 |
+
This model is a generator for creating images like the Flair MRIs based on BraTS 2016 and 2017 data. It was trained as a 3d latent diffusion model and accepts Gaussian random noise as inputs to produce an image output. The `train_autoencoder.json` file describes the training process of the variational autoencoder with GAN loss. The `train_diffusion.json` file describes the training process of the 3D latent diffusion model.
|
16 |
+
|
17 |
+
In this bundle, the autoencoder uses perceptual loss, which is based on ResNet50 with pre-trained weights (the network is frozen and will not be trained in the bundle). In default, the `pretrained` parameter is specified as `False` in `train_autoencoder.json`. To ensure correct training, changing the default settings is necessary. There are two ways to utilize pretrained weights:
|
18 |
+
1. if set `pretrained` to `True`, ImageNet pretrained weights from [torchvision](https://pytorch.org/vision/stable/_modules/torchvision/models/resnet.html#ResNet50_Weights) will be used. However, the weights are for non-commercial use only.
|
19 |
+
2. if set `pretrained` to `True` and specifies the `perceptual_loss_model_weights_path` parameter, users are able to load weights from a local path. This is the way this bundle used to train, and the pre-trained weights are from some internal data.
|
20 |
+
|
21 |
+
Please note that each user is responsible for checking the data source of the pre-trained models, the applicable licenses, and determining if suitable for the intended use.
|
22 |
+
|
23 |
+
#### Example synthetic image
|
24 |
+
An example result from inference is shown below:
|
25 |
+
![Example synthetic image](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_example_generation.png)
|
26 |
+
|
27 |
+
**This is a demonstration network meant to just show the training process for this sort of network with MONAI. To achieve better performance, users need to use larger dataset like [Brats 2021](https://www.synapse.org/#!Synapse:syn25829067/wiki/610865) and have GPU with memory larger than 32G to enable larger networks and attention layers.**
|
28 |
+
|
29 |
+
## MONAI Generative Model Dependencies
|
30 |
+
[MONAI generative models](https://github.com/Project-MONAI/GenerativeModels) can be installed by
|
31 |
+
```
|
32 |
+
pip install lpips==0.1.4
|
33 |
+
git clone https://github.com/Project-MONAI/GenerativeModels.git
|
34 |
+
cd GenerativeModels/
|
35 |
+
git checkout f969c24f88d013dc0045fb7b2885a01fb219992b
|
36 |
+
python setup.py install
|
37 |
+
cd ..
|
38 |
+
```
|
39 |
+
|
40 |
+
## Data
|
41 |
+
The training data is BraTS 2016 and 2017 from the Medical Segmentation Decathalon. Users can find more details on the dataset (`Task01_BrainTumour`) at http://medicaldecathlon.com/.
|
42 |
+
|
43 |
+
- Target: Image Generation
|
44 |
+
- Task: Synthesis
|
45 |
+
- Modality: MRI
|
46 |
+
- Size: 388 3D volumes (1 channel used)
|
47 |
+
|
48 |
+
## Training Configuration
|
49 |
+
If you have a GPU with less than 32G of memory, you may need to decrease the batch size when training. To do so, modify the `train_batch_size` parameter in the [configs/train_autoencoder.json](../configs/train_autoencoder.json) and [configs/train_diffusion.json](../configs/train_diffusion.json) configuration files.
|
50 |
+
|
51 |
+
### Training Configuration of Autoencoder
|
52 |
+
The autoencoder was trained using the following configuration:
|
53 |
+
|
54 |
+
- GPU: at least 32GB GPU memory
|
55 |
+
- Actual Model Input: 112 x 128 x 80
|
56 |
+
- AMP: False
|
57 |
+
- Optimizer: Adam
|
58 |
+
- Learning Rate: 1e-5
|
59 |
+
- Loss: L1 loss, perceptual loss, KL divergence loss, adversarial loss, GAN BCE loss
|
60 |
+
|
61 |
+
#### Input
|
62 |
+
1 channel 3D MRI Flair patches
|
63 |
+
|
64 |
+
#### Output
|
65 |
+
- 1 channel 3D MRI reconstructed patches
|
66 |
+
- 8 channel mean of latent features
|
67 |
+
- 8 channel standard deviation of latent features
|
68 |
+
|
69 |
+
### Training Configuration of Diffusion Model
|
70 |
+
The latent diffusion model was trained using the following configuration:
|
71 |
+
|
72 |
+
- GPU: at least 32GB GPU memory
|
73 |
+
- Actual Model Input: 36 x 44 x 28
|
74 |
+
- AMP: False
|
75 |
+
- Optimizer: Adam
|
76 |
+
- Learning Rate: 1e-5
|
77 |
+
- Loss: MSE loss
|
78 |
+
|
79 |
+
#### Training Input
|
80 |
+
- 8 channel noisy latent features
|
81 |
+
- an int that indicates the time step
|
82 |
+
|
83 |
+
#### Training Output
|
84 |
+
8 channel predicted added noise
|
85 |
+
|
86 |
+
#### Inference Input
|
87 |
+
8 channel noise
|
88 |
+
|
89 |
+
#### Inference Output
|
90 |
+
8 channel denoised latent features
|
91 |
+
|
92 |
+
### Memory Consumption Warning
|
93 |
+
|
94 |
+
If you face memory issues with data loading, you can lower the caching rate `cache_rate` in the configurations within range [0, 1] to minimize the System RAM requirements.
|
95 |
+
|
96 |
+
## Performance
|
97 |
+
|
98 |
+
#### Training Loss
|
99 |
+
![A graph showing the autoencoder training curve](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_train_autoencoder_loss.png)
|
100 |
+
|
101 |
+
![A graph showing the latent diffusion training curve](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_train_diffusion_loss.png)
|
102 |
+
|
103 |
+
## MONAI Bundle Commands
|
104 |
+
|
105 |
+
In addition to the Pythonic APIs, a few command line interfaces (CLI) are provided to interact with the bundle. The CLI supports flexible use cases, such as overriding configs at runtime and predefining arguments in a file.
|
106 |
+
|
107 |
+
For more details usage instructions, visit the [MONAI Bundle Configuration Page](https://docs.monai.io/en/latest/config_syntax.html).
|
108 |
+
|
109 |
+
### Execute Autoencoder Training
|
110 |
+
|
111 |
+
#### Execute Autoencoder Training on single GPU
|
112 |
+
|
113 |
+
```
|
114 |
+
python -m monai.bundle run --config_file configs/train_autoencoder.json
|
115 |
+
```
|
116 |
+
|
117 |
+
Please note that if the default dataset path is not modified with the actual path (it should be the path that contains `Task01_BrainTumour`) in the bundle config files, you can also override it by using `--dataset_dir`:
|
118 |
+
|
119 |
+
```
|
120 |
+
python -m monai.bundle run --config_file configs/train_autoencoder.json --dataset_dir <actual dataset path>
|
121 |
+
```
|
122 |
+
|
123 |
+
#### Override the `train` config to execute multi-GPU training for Autoencoder
|
124 |
+
To train with multiple GPUs, use the following command, which requires scaling up the learning rate according to the number of GPUs.
|
125 |
+
|
126 |
+
```
|
127 |
+
torchrun --standalone --nnodes=1 --nproc_per_node=8 -m monai.bundle run --config_file "['configs/train_autoencoder.json','configs/multi_gpu_train_autoencoder.json']" --lr 8e-5
|
128 |
+
```
|
129 |
+
|
130 |
+
#### Check the Autoencoder Training result
|
131 |
+
The following code generates a reconstructed image from a random input image.
|
132 |
+
We can visualize it to see if the autoencoder is trained correctly.
|
133 |
+
```
|
134 |
+
python -m monai.bundle run --config_file configs/inference_autoencoder.json
|
135 |
+
```
|
136 |
+
|
137 |
+
An example of reconstructed image from inference is shown below. If the autoencoder is trained correctly, the reconstructed image should look similar to original image.
|
138 |
+
|
139 |
+
![Example reconstructed image](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_recon_example.jpg)
|
140 |
+
|
141 |
+
|
142 |
+
### Execute Latent Diffusion Training
|
143 |
+
|
144 |
+
#### Execute Latent Diffusion Model Training on single GPU
|
145 |
+
After training the autoencoder, run the following command to train the latent diffusion model. This command will print out the scale factor of the latent feature space. If your autoencoder is well trained, this value should be close to 1.0.
|
146 |
+
|
147 |
+
```
|
148 |
+
python -m monai.bundle run --config_file "['configs/train_autoencoder.json','configs/train_diffusion.json']"
|
149 |
+
```
|
150 |
+
|
151 |
+
#### Override the `train` config to execute multi-GPU training for Latent Diffusion Model
|
152 |
+
To train with multiple GPUs, use the following command, which requires scaling up the learning rate according to the number of GPUs.
|
153 |
+
|
154 |
+
```
|
155 |
+
torchrun --standalone --nnodes=1 --nproc_per_node=8 -m monai.bundle run --config_file "['configs/train_autoencoder.json','configs/train_diffusion.json','configs/multi_gpu_train_autoencoder.json','configs/multi_gpu_train_diffusion.json']" --lr 8e-5
|
156 |
+
```
|
157 |
+
|
158 |
+
#### Execute inference
|
159 |
+
The following code generates a synthetic image from a random sampled noise.
|
160 |
+
```
|
161 |
+
python -m monai.bundle run --config_file configs/inference.json
|
162 |
+
```
|
163 |
+
|
164 |
+
#### Export checkpoint to TorchScript file
|
165 |
+
|
166 |
+
The Autoencoder can be exported into a TorchScript file.
|
167 |
+
|
168 |
+
```
|
169 |
+
python -m monai.bundle ckpt_export autoencoder_def --filepath models/model_autoencoder.ts --ckpt_file models/model_autoencoder.pt --meta_file configs/metadata.json --config_file configs/inference.json
|
170 |
+
```
|
171 |
+
|
172 |
+
# References
|
173 |
+
[1] Rombach, Robin, et al. "High-resolution image synthesis with latent diffusion models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022. https://openaccess.thecvf.com/content/CVPR2022/papers/Rombach_High-Resolution_Image_Synthesis_With_Latent_Diffusion_Models_CVPR_2022_paper.pdf
|
174 |
+
|
175 |
+
# License
|
176 |
+
Copyright (c) MONAI Consortium
|
177 |
+
|
178 |
+
Licensed under the Apache License, Version 2.0 (the "License");
|
179 |
+
you may not use this file except in compliance with the License.
|
180 |
+
You may obtain a copy of the License at
|
181 |
+
|
182 |
+
http://www.apache.org/licenses/LICENSE-2.0
|
183 |
+
|
184 |
+
Unless required by applicable law or agreed to in writing, software
|
185 |
+
distributed under the License is distributed on an "AS IS" BASIS,
|
186 |
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
187 |
+
See the License for the specific language governing permissions and
|
188 |
+
limitations under the License.
|
configs/inference.json
ADDED
@@ -0,0 +1,101 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"imports": [
|
3 |
+
"$import torch",
|
4 |
+
"$from datetime import datetime",
|
5 |
+
"$from pathlib import Path"
|
6 |
+
],
|
7 |
+
"bundle_root": ".",
|
8 |
+
"model_dir": "$@bundle_root + '/models'",
|
9 |
+
"output_dir": "$@bundle_root + '/output'",
|
10 |
+
"create_output_dir": "$Path(@output_dir).mkdir(exist_ok=True)",
|
11 |
+
"device": "$torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')",
|
12 |
+
"output_postfix": "$datetime.now().strftime('sample_%Y%m%d_%H%M%S')",
|
13 |
+
"spatial_dims": 3,
|
14 |
+
"image_channels": 1,
|
15 |
+
"latent_channels": 8,
|
16 |
+
"latent_shape": [
|
17 |
+
8,
|
18 |
+
36,
|
19 |
+
44,
|
20 |
+
28
|
21 |
+
],
|
22 |
+
"autoencoder_def": {
|
23 |
+
"_target_": "generative.networks.nets.AutoencoderKL",
|
24 |
+
"spatial_dims": "@spatial_dims",
|
25 |
+
"in_channels": "@image_channels",
|
26 |
+
"out_channels": "@image_channels",
|
27 |
+
"latent_channels": "@latent_channels",
|
28 |
+
"num_channels": [
|
29 |
+
64,
|
30 |
+
128,
|
31 |
+
256
|
32 |
+
],
|
33 |
+
"num_res_blocks": 2,
|
34 |
+
"norm_num_groups": 32,
|
35 |
+
"norm_eps": 1e-06,
|
36 |
+
"attention_levels": [
|
37 |
+
false,
|
38 |
+
false,
|
39 |
+
false
|
40 |
+
],
|
41 |
+
"with_encoder_nonlocal_attn": false,
|
42 |
+
"with_decoder_nonlocal_attn": false
|
43 |
+
},
|
44 |
+
"network_def": {
|
45 |
+
"_target_": "generative.networks.nets.DiffusionModelUNet",
|
46 |
+
"spatial_dims": "@spatial_dims",
|
47 |
+
"in_channels": "@latent_channels",
|
48 |
+
"out_channels": "@latent_channels",
|
49 |
+
"num_channels": [
|
50 |
+
256,
|
51 |
+
256,
|
52 |
+
512
|
53 |
+
],
|
54 |
+
"attention_levels": [
|
55 |
+
false,
|
56 |
+
true,
|
57 |
+
true
|
58 |
+
],
|
59 |
+
"num_head_channels": [
|
60 |
+
0,
|
61 |
+
64,
|
62 |
+
64
|
63 |
+
],
|
64 |
+
"num_res_blocks": 2
|
65 |
+
},
|
66 |
+
"load_autoencoder_path": "$@bundle_root + '/models/model_autoencoder.pt'",
|
67 |
+
"load_autoencoder": "$@autoencoder_def.load_state_dict(torch.load(@load_autoencoder_path))",
|
68 |
+
"autoencoder": "$@autoencoder_def.to(@device)",
|
69 |
+
"load_diffusion_path": "$@model_dir + '/model.pt'",
|
70 |
+
"load_diffusion": "$@network_def.load_state_dict(torch.load(@load_diffusion_path))",
|
71 |
+
"diffusion": "$@network_def.to(@device)",
|
72 |
+
"noise_scheduler": {
|
73 |
+
"_target_": "generative.networks.schedulers.DDIMScheduler",
|
74 |
+
"_requires_": [
|
75 |
+
"@load_diffusion",
|
76 |
+
"@load_autoencoder"
|
77 |
+
],
|
78 |
+
"num_train_timesteps": 1000,
|
79 |
+
"beta_start": 0.0015,
|
80 |
+
"beta_end": 0.0195,
|
81 |
+
"beta_schedule": "scaled_linear",
|
82 |
+
"clip_sample": false
|
83 |
+
},
|
84 |
+
"noise": "$torch.randn([1]+@latent_shape).to(@device)",
|
85 |
+
"set_timesteps": "$@noise_scheduler.set_timesteps(num_inference_steps=50)",
|
86 |
+
"inferer": {
|
87 |
+
"_target_": "scripts.ldm_sampler.LDMSampler",
|
88 |
+
"_requires_": "@set_timesteps"
|
89 |
+
},
|
90 |
+
"sample": "$@inferer.sampling_fn(@noise, @autoencoder, @diffusion, @noise_scheduler)",
|
91 |
+
"saver": {
|
92 |
+
"_target_": "SaveImage",
|
93 |
+
"_requires_": "@create_output_dir",
|
94 |
+
"output_dir": "@output_dir",
|
95 |
+
"output_postfix": "@output_postfix"
|
96 |
+
},
|
97 |
+
"generated_image": "$@sample",
|
98 |
+
"run": [
|
99 |
+
"$@saver(@generated_image[0])"
|
100 |
+
]
|
101 |
+
}
|
configs/inference_autoencoder.json
ADDED
@@ -0,0 +1,149 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"imports": [
|
3 |
+
"$import torch",
|
4 |
+
"$from datetime import datetime",
|
5 |
+
"$from pathlib import Path"
|
6 |
+
],
|
7 |
+
"bundle_root": ".",
|
8 |
+
"model_dir": "$@bundle_root + '/models'",
|
9 |
+
"dataset_dir": "@bundle_root",
|
10 |
+
"output_dir": "$@bundle_root + '/output'",
|
11 |
+
"create_output_dir": "$Path(@output_dir).mkdir(exist_ok=True)",
|
12 |
+
"device": "$torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')",
|
13 |
+
"output_orig_postfix": "recon",
|
14 |
+
"output_recon_postfix": "orig",
|
15 |
+
"channel": 0,
|
16 |
+
"spacing": [
|
17 |
+
1.1,
|
18 |
+
1.1,
|
19 |
+
1.1
|
20 |
+
],
|
21 |
+
"spatial_dims": 3,
|
22 |
+
"image_channels": 1,
|
23 |
+
"latent_channels": 8,
|
24 |
+
"infer_patch_size": [
|
25 |
+
144,
|
26 |
+
176,
|
27 |
+
112
|
28 |
+
],
|
29 |
+
"autoencoder_def": {
|
30 |
+
"_target_": "generative.networks.nets.AutoencoderKL",
|
31 |
+
"spatial_dims": "@spatial_dims",
|
32 |
+
"in_channels": "@image_channels",
|
33 |
+
"out_channels": "@image_channels",
|
34 |
+
"latent_channels": "@latent_channels",
|
35 |
+
"num_channels": [
|
36 |
+
64,
|
37 |
+
128,
|
38 |
+
256
|
39 |
+
],
|
40 |
+
"num_res_blocks": 2,
|
41 |
+
"norm_num_groups": 32,
|
42 |
+
"norm_eps": 1e-06,
|
43 |
+
"attention_levels": [
|
44 |
+
false,
|
45 |
+
false,
|
46 |
+
false
|
47 |
+
],
|
48 |
+
"with_encoder_nonlocal_attn": false,
|
49 |
+
"with_decoder_nonlocal_attn": false
|
50 |
+
},
|
51 |
+
"load_autoencoder_path": "$@bundle_root + '/models/model_autoencoder.pt'",
|
52 |
+
"load_autoencoder": "$@autoencoder_def.load_state_dict(torch.load(@load_autoencoder_path))",
|
53 |
+
"autoencoder": "$@autoencoder_def.to(@device)",
|
54 |
+
"preprocessing_transforms": [
|
55 |
+
{
|
56 |
+
"_target_": "LoadImaged",
|
57 |
+
"keys": "image"
|
58 |
+
},
|
59 |
+
{
|
60 |
+
"_target_": "EnsureChannelFirstd",
|
61 |
+
"keys": "image"
|
62 |
+
},
|
63 |
+
{
|
64 |
+
"_target_": "Lambdad",
|
65 |
+
"keys": "image",
|
66 |
+
"func": "$lambda x: x[@channel, :, :, :]"
|
67 |
+
},
|
68 |
+
{
|
69 |
+
"_target_": "AddChanneld",
|
70 |
+
"keys": "image"
|
71 |
+
},
|
72 |
+
{
|
73 |
+
"_target_": "EnsureTyped",
|
74 |
+
"keys": "image"
|
75 |
+
},
|
76 |
+
{
|
77 |
+
"_target_": "Orientationd",
|
78 |
+
"keys": "image",
|
79 |
+
"axcodes": "RAS"
|
80 |
+
},
|
81 |
+
{
|
82 |
+
"_target_": "Spacingd",
|
83 |
+
"keys": "image",
|
84 |
+
"pixdim": "@spacing",
|
85 |
+
"mode": "bilinear"
|
86 |
+
}
|
87 |
+
],
|
88 |
+
"crop_transforms": [
|
89 |
+
{
|
90 |
+
"_target_": "CenterSpatialCropd",
|
91 |
+
"keys": "image",
|
92 |
+
"roi_size": "@infer_patch_size"
|
93 |
+
}
|
94 |
+
],
|
95 |
+
"final_transforms": [
|
96 |
+
{
|
97 |
+
"_target_": "ScaleIntensityRangePercentilesd",
|
98 |
+
"keys": "image",
|
99 |
+
"lower": 0,
|
100 |
+
"upper": 99.5,
|
101 |
+
"b_min": 0,
|
102 |
+
"b_max": 1
|
103 |
+
}
|
104 |
+
],
|
105 |
+
"preprocessing": {
|
106 |
+
"_target_": "Compose",
|
107 |
+
"transforms": "$@preprocessing_transforms + @crop_transforms + @final_transforms"
|
108 |
+
},
|
109 |
+
"dataset": {
|
110 |
+
"_target_": "monai.apps.DecathlonDataset",
|
111 |
+
"root_dir": "@dataset_dir",
|
112 |
+
"task": "Task01_BrainTumour",
|
113 |
+
"section": "validation",
|
114 |
+
"cache_rate": 0.0,
|
115 |
+
"num_workers": 8,
|
116 |
+
"download": false,
|
117 |
+
"transform": "@preprocessing"
|
118 |
+
},
|
119 |
+
"dataloader": {
|
120 |
+
"_target_": "DataLoader",
|
121 |
+
"dataset": "@dataset",
|
122 |
+
"batch_size": 1,
|
123 |
+
"shuffle": true,
|
124 |
+
"num_workers": 0
|
125 |
+
},
|
126 |
+
"saver_orig": {
|
127 |
+
"_target_": "SaveImage",
|
128 |
+
"_requires_": "@create_output_dir",
|
129 |
+
"output_dir": "@output_dir",
|
130 |
+
"output_postfix": "@output_orig_postfix",
|
131 |
+
"resample": false,
|
132 |
+
"padding_mode": "zeros"
|
133 |
+
},
|
134 |
+
"saver_recon": {
|
135 |
+
"_target_": "SaveImage",
|
136 |
+
"_requires_": "@create_output_dir",
|
137 |
+
"output_dir": "@output_dir",
|
138 |
+
"output_postfix": "@output_recon_postfix",
|
139 |
+
"resample": false,
|
140 |
+
"padding_mode": "zeros"
|
141 |
+
},
|
142 |
+
"input_img": "$monai.utils.first(@dataloader)['image'].to(@device)",
|
143 |
+
"recon_img": "$@autoencoder(@input_img)[0][0]",
|
144 |
+
"run": [
|
145 |
+
"$@load_autoencoder",
|
146 |
+
"$@saver_orig(@input_img[0][0])",
|
147 |
+
"$@saver_recon(@recon_img)"
|
148 |
+
]
|
149 |
+
}
|
configs/logging.conf
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[loggers]
|
2 |
+
keys=root
|
3 |
+
|
4 |
+
[handlers]
|
5 |
+
keys=consoleHandler
|
6 |
+
|
7 |
+
[formatters]
|
8 |
+
keys=fullFormatter
|
9 |
+
|
10 |
+
[logger_root]
|
11 |
+
level=INFO
|
12 |
+
handlers=consoleHandler
|
13 |
+
|
14 |
+
[handler_consoleHandler]
|
15 |
+
class=StreamHandler
|
16 |
+
level=INFO
|
17 |
+
formatter=fullFormatter
|
18 |
+
args=(sys.stdout,)
|
19 |
+
|
20 |
+
[formatter_fullFormatter]
|
21 |
+
format=%(asctime)s - %(name)s - %(levelname)s - %(message)s
|
configs/metadata.json
ADDED
@@ -0,0 +1,119 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"schema": "https://github.com/Project-MONAI/MONAI-extra-test-data/releases/download/0.8.1/meta_schema_generator_ldm_20230507.json",
|
3 |
+
"version": "1.0.0",
|
4 |
+
"changelog": {
|
5 |
+
"1.0.0": "Initial release"
|
6 |
+
},
|
7 |
+
"monai_version": "1.2.0rc5",
|
8 |
+
"pytorch_version": "1.13.1",
|
9 |
+
"numpy_version": "1.22.2",
|
10 |
+
"optional_packages_version": {
|
11 |
+
"nibabel": "5.1.0",
|
12 |
+
"lpips": "0.1.4"
|
13 |
+
},
|
14 |
+
"name": "BraTS MRI image latent diffusion generation",
|
15 |
+
"task": "BraTS MRI image synthesis",
|
16 |
+
"description": "A generative model for creating 3D brain MRI from Gaussian noise based on BraTS dataset",
|
17 |
+
"authors": "MONAI team",
|
18 |
+
"copyright": "Copyright (c) MONAI Consortium",
|
19 |
+
"data_source": "http://medicaldecathlon.com/",
|
20 |
+
"data_type": "nibabel",
|
21 |
+
"image_classes": "Flair brain MRI with 1.1x1.1x1.1 mm voxel size",
|
22 |
+
"eval_metrics": {},
|
23 |
+
"intended_use": "This is a research tool/prototype and not to be used clinically",
|
24 |
+
"references": [],
|
25 |
+
"autoencoder_data_format": {
|
26 |
+
"inputs": {
|
27 |
+
"image": {
|
28 |
+
"type": "image",
|
29 |
+
"format": "image",
|
30 |
+
"num_channels": 1,
|
31 |
+
"spatial_shape": [
|
32 |
+
112,
|
33 |
+
128,
|
34 |
+
80
|
35 |
+
],
|
36 |
+
"dtype": "float32",
|
37 |
+
"value_range": [
|
38 |
+
0,
|
39 |
+
1
|
40 |
+
],
|
41 |
+
"is_patch_data": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"outputs": {
|
45 |
+
"pred": {
|
46 |
+
"type": "image",
|
47 |
+
"format": "image",
|
48 |
+
"num_channels": 1,
|
49 |
+
"spatial_shape": [
|
50 |
+
112,
|
51 |
+
128,
|
52 |
+
80
|
53 |
+
],
|
54 |
+
"dtype": "float32",
|
55 |
+
"value_range": [
|
56 |
+
0,
|
57 |
+
1
|
58 |
+
],
|
59 |
+
"is_patch_data": true,
|
60 |
+
"channel_def": {
|
61 |
+
"0": "image"
|
62 |
+
}
|
63 |
+
}
|
64 |
+
}
|
65 |
+
},
|
66 |
+
"generator_data_format": {
|
67 |
+
"inputs": {
|
68 |
+
"latent": {
|
69 |
+
"type": "noise",
|
70 |
+
"format": "image",
|
71 |
+
"num_channels": 8,
|
72 |
+
"spatial_shape": [
|
73 |
+
36,
|
74 |
+
44,
|
75 |
+
28
|
76 |
+
],
|
77 |
+
"dtype": "float32",
|
78 |
+
"value_range": [
|
79 |
+
0,
|
80 |
+
1
|
81 |
+
],
|
82 |
+
"is_patch_data": true
|
83 |
+
},
|
84 |
+
"condition": {
|
85 |
+
"type": "timesteps",
|
86 |
+
"format": "timesteps",
|
87 |
+
"num_channels": 1,
|
88 |
+
"spatial_shape": [],
|
89 |
+
"dtype": "long",
|
90 |
+
"value_range": [
|
91 |
+
0,
|
92 |
+
1000
|
93 |
+
],
|
94 |
+
"is_patch_data": false
|
95 |
+
}
|
96 |
+
},
|
97 |
+
"outputs": {
|
98 |
+
"pred": {
|
99 |
+
"type": "feature",
|
100 |
+
"format": "image",
|
101 |
+
"num_channels": 8,
|
102 |
+
"spatial_shape": [
|
103 |
+
36,
|
104 |
+
44,
|
105 |
+
28
|
106 |
+
],
|
107 |
+
"dtype": "float32",
|
108 |
+
"value_range": [
|
109 |
+
0,
|
110 |
+
1
|
111 |
+
],
|
112 |
+
"is_patch_data": true,
|
113 |
+
"channel_def": {
|
114 |
+
"0": "image"
|
115 |
+
}
|
116 |
+
}
|
117 |
+
}
|
118 |
+
}
|
119 |
+
}
|
configs/multi_gpu_train_autoencoder.json
ADDED
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"device": "$torch.device(f'cuda:{dist.get_rank()}')",
|
3 |
+
"gnetwork": {
|
4 |
+
"_target_": "torch.nn.parallel.DistributedDataParallel",
|
5 |
+
"module": "$@autoencoder_def.to(@device)",
|
6 |
+
"device_ids": [
|
7 |
+
"@device"
|
8 |
+
]
|
9 |
+
},
|
10 |
+
"dnetwork": {
|
11 |
+
"_target_": "torch.nn.parallel.DistributedDataParallel",
|
12 |
+
"module": "$@discriminator_def.to(@device)",
|
13 |
+
"device_ids": [
|
14 |
+
"@device"
|
15 |
+
]
|
16 |
+
},
|
17 |
+
"train#sampler": {
|
18 |
+
"_target_": "DistributedSampler",
|
19 |
+
"dataset": "@train#dataset",
|
20 |
+
"even_divisible": true,
|
21 |
+
"shuffle": true
|
22 |
+
},
|
23 |
+
"train#dataloader#sampler": "@train#sampler",
|
24 |
+
"train#dataloader#shuffle": false,
|
25 |
+
"train#trainer#train_handlers": "$@train#handlers[: -2 if dist.get_rank() > 0 else None]",
|
26 |
+
"initialize": [
|
27 |
+
"$import torch.distributed as dist",
|
28 |
+
"$dist.is_initialized() or dist.init_process_group(backend='nccl')",
|
29 |
+
"$torch.cuda.set_device(@device)",
|
30 |
+
"$monai.utils.set_determinism(seed=123)",
|
31 |
+
"$import logging",
|
32 |
+
"$@train#trainer.logger.setLevel(logging.WARNING if dist.get_rank() > 0 else logging.INFO)"
|
33 |
+
],
|
34 |
+
"run": [
|
35 |
+
"$@train#trainer.run()"
|
36 |
+
],
|
37 |
+
"finalize": [
|
38 |
+
"$dist.is_initialized() and dist.destroy_process_group()"
|
39 |
+
]
|
40 |
+
}
|
configs/multi_gpu_train_diffusion.json
ADDED
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"diffusion": {
|
3 |
+
"_target_": "torch.nn.parallel.DistributedDataParallel",
|
4 |
+
"module": "$@network_def.to(@device)",
|
5 |
+
"device_ids": [
|
6 |
+
"@device"
|
7 |
+
],
|
8 |
+
"find_unused_parameters": true
|
9 |
+
},
|
10 |
+
"run": [
|
11 |
+
"@load_autoencoder",
|
12 |
+
"$@autoencoder.eval()",
|
13 |
+
"$print('scale factor:',@scale_factor)",
|
14 |
+
"$@train#trainer.run()"
|
15 |
+
]
|
16 |
+
}
|
configs/train_autoencoder.json
ADDED
@@ -0,0 +1,206 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"imports": [
|
3 |
+
"$import functools",
|
4 |
+
"$import glob",
|
5 |
+
"$import scripts"
|
6 |
+
],
|
7 |
+
"bundle_root": ".",
|
8 |
+
"device": "$torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')",
|
9 |
+
"ckpt_dir": "$@bundle_root + '/models'",
|
10 |
+
"tf_dir": "$@bundle_root + '/eval'",
|
11 |
+
"dataset_dir": "/workspace/data/medical",
|
12 |
+
"pretrained": false,
|
13 |
+
"perceptual_loss_model_weights_path": null,
|
14 |
+
"train_batch_size": 2,
|
15 |
+
"lr": 1e-05,
|
16 |
+
"train_patch_size": [
|
17 |
+
112,
|
18 |
+
128,
|
19 |
+
80
|
20 |
+
],
|
21 |
+
"channel": 0,
|
22 |
+
"spacing": [
|
23 |
+
1.1,
|
24 |
+
1.1,
|
25 |
+
1.1
|
26 |
+
],
|
27 |
+
"spatial_dims": 3,
|
28 |
+
"image_channels": 1,
|
29 |
+
"latent_channels": 8,
|
30 |
+
"discriminator_def": {
|
31 |
+
"_target_": "generative.networks.nets.PatchDiscriminator",
|
32 |
+
"spatial_dims": "@spatial_dims",
|
33 |
+
"num_layers_d": 3,
|
34 |
+
"num_channels": 32,
|
35 |
+
"in_channels": 1,
|
36 |
+
"out_channels": 1,
|
37 |
+
"norm": "INSTANCE"
|
38 |
+
},
|
39 |
+
"autoencoder_def": {
|
40 |
+
"_target_": "generative.networks.nets.AutoencoderKL",
|
41 |
+
"spatial_dims": "@spatial_dims",
|
42 |
+
"in_channels": "@image_channels",
|
43 |
+
"out_channels": "@image_channels",
|
44 |
+
"latent_channels": "@latent_channels",
|
45 |
+
"num_channels": [
|
46 |
+
64,
|
47 |
+
128,
|
48 |
+
256
|
49 |
+
],
|
50 |
+
"num_res_blocks": 2,
|
51 |
+
"norm_num_groups": 32,
|
52 |
+
"norm_eps": 1e-06,
|
53 |
+
"attention_levels": [
|
54 |
+
false,
|
55 |
+
false,
|
56 |
+
false
|
57 |
+
],
|
58 |
+
"with_encoder_nonlocal_attn": false,
|
59 |
+
"with_decoder_nonlocal_attn": false
|
60 |
+
},
|
61 |
+
"perceptual_loss_def": {
|
62 |
+
"_target_": "generative.losses.PerceptualLoss",
|
63 |
+
"spatial_dims": "@spatial_dims",
|
64 |
+
"network_type": "resnet50",
|
65 |
+
"is_fake_3d": true,
|
66 |
+
"fake_3d_ratio": 0.2,
|
67 |
+
"pretrained": "@pretrained",
|
68 |
+
"pretrained_path": "@perceptual_loss_model_weights_path",
|
69 |
+
"pretrained_state_dict_key": "state_dict"
|
70 |
+
},
|
71 |
+
"dnetwork": "$@discriminator_def.to(@device)",
|
72 |
+
"gnetwork": "$@autoencoder_def.to(@device)",
|
73 |
+
"loss_perceptual": "$@perceptual_loss_def.to(@device)",
|
74 |
+
"doptimizer": {
|
75 |
+
"_target_": "torch.optim.Adam",
|
76 |
+
"params": "$@dnetwork.parameters()",
|
77 |
+
"lr": "@lr"
|
78 |
+
},
|
79 |
+
"goptimizer": {
|
80 |
+
"_target_": "torch.optim.Adam",
|
81 |
+
"params": "$@gnetwork.parameters()",
|
82 |
+
"lr": "@lr"
|
83 |
+
},
|
84 |
+
"preprocessing_transforms": [
|
85 |
+
{
|
86 |
+
"_target_": "LoadImaged",
|
87 |
+
"keys": "image"
|
88 |
+
},
|
89 |
+
{
|
90 |
+
"_target_": "EnsureChannelFirstd",
|
91 |
+
"keys": "image"
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"_target_": "Lambdad",
|
95 |
+
"keys": "image",
|
96 |
+
"func": "$lambda x: x[@channel, :, :, :]"
|
97 |
+
},
|
98 |
+
{
|
99 |
+
"_target_": "AddChanneld",
|
100 |
+
"keys": "image"
|
101 |
+
},
|
102 |
+
{
|
103 |
+
"_target_": "EnsureTyped",
|
104 |
+
"keys": "image"
|
105 |
+
},
|
106 |
+
{
|
107 |
+
"_target_": "Orientationd",
|
108 |
+
"keys": "image",
|
109 |
+
"axcodes": "RAS"
|
110 |
+
},
|
111 |
+
{
|
112 |
+
"_target_": "Spacingd",
|
113 |
+
"keys": "image",
|
114 |
+
"pixdim": "@spacing",
|
115 |
+
"mode": "bilinear"
|
116 |
+
}
|
117 |
+
],
|
118 |
+
"final_transforms": [
|
119 |
+
{
|
120 |
+
"_target_": "ScaleIntensityRangePercentilesd",
|
121 |
+
"keys": "image",
|
122 |
+
"lower": 0,
|
123 |
+
"upper": 99.5,
|
124 |
+
"b_min": 0,
|
125 |
+
"b_max": 1
|
126 |
+
}
|
127 |
+
],
|
128 |
+
"train": {
|
129 |
+
"crop_transforms": [
|
130 |
+
{
|
131 |
+
"_target_": "RandSpatialCropd",
|
132 |
+
"keys": "image",
|
133 |
+
"roi_size": "@train_patch_size",
|
134 |
+
"random_size": false
|
135 |
+
}
|
136 |
+
],
|
137 |
+
"preprocessing": {
|
138 |
+
"_target_": "Compose",
|
139 |
+
"transforms": "$@preprocessing_transforms + @train#crop_transforms + @final_transforms"
|
140 |
+
},
|
141 |
+
"dataset": {
|
142 |
+
"_target_": "monai.apps.DecathlonDataset",
|
143 |
+
"root_dir": "@dataset_dir",
|
144 |
+
"task": "Task01_BrainTumour",
|
145 |
+
"section": "training",
|
146 |
+
"cache_rate": 1.0,
|
147 |
+
"num_workers": 8,
|
148 |
+
"download": false,
|
149 |
+
"transform": "@train#preprocessing"
|
150 |
+
},
|
151 |
+
"dataloader": {
|
152 |
+
"_target_": "DataLoader",
|
153 |
+
"dataset": "@train#dataset",
|
154 |
+
"batch_size": "@train_batch_size",
|
155 |
+
"shuffle": true,
|
156 |
+
"num_workers": 0
|
157 |
+
},
|
158 |
+
"handlers": [
|
159 |
+
{
|
160 |
+
"_target_": "CheckpointSaver",
|
161 |
+
"save_dir": "@ckpt_dir",
|
162 |
+
"save_dict": {
|
163 |
+
"model": "@gnetwork"
|
164 |
+
},
|
165 |
+
"save_interval": 0,
|
166 |
+
"save_final": true,
|
167 |
+
"epoch_level": true,
|
168 |
+
"final_filename": "model_autoencoder.pt"
|
169 |
+
},
|
170 |
+
{
|
171 |
+
"_target_": "StatsHandler",
|
172 |
+
"tag_name": "train_loss",
|
173 |
+
"output_transform": "$lambda x: monai.handlers.from_engine(['g_loss'], first=True)(x)[0]"
|
174 |
+
},
|
175 |
+
{
|
176 |
+
"_target_": "TensorBoardStatsHandler",
|
177 |
+
"log_dir": "@tf_dir",
|
178 |
+
"tag_name": "train_loss",
|
179 |
+
"output_transform": "$lambda x: monai.handlers.from_engine(['g_loss'], first=True)(x)[0]"
|
180 |
+
}
|
181 |
+
],
|
182 |
+
"trainer": {
|
183 |
+
"_target_": "scripts.ldm_trainer.VaeGanTrainer",
|
184 |
+
"device": "@device",
|
185 |
+
"max_epochs": 1500,
|
186 |
+
"train_data_loader": "@train#dataloader",
|
187 |
+
"g_network": "@gnetwork",
|
188 |
+
"g_optimizer": "@goptimizer",
|
189 |
+
"g_loss_function": "$functools.partial(scripts.losses.generator_loss, disc_net=@dnetwork, loss_perceptual=@loss_perceptual)",
|
190 |
+
"d_network": "@dnetwork",
|
191 |
+
"d_optimizer": "@doptimizer",
|
192 |
+
"d_loss_function": "$functools.partial(scripts.losses.discriminator_loss, disc_net=@dnetwork)",
|
193 |
+
"d_train_steps": 5,
|
194 |
+
"g_update_latents": true,
|
195 |
+
"latent_shape": "@latent_channels",
|
196 |
+
"key_train_metric": "$None",
|
197 |
+
"train_handlers": "@train#handlers"
|
198 |
+
}
|
199 |
+
},
|
200 |
+
"initialize": [
|
201 |
+
"$monai.utils.set_determinism(seed=0)"
|
202 |
+
],
|
203 |
+
"run": [
|
204 |
+
"$@train#trainer.run()"
|
205 |
+
]
|
206 |
+
}
|
configs/train_diffusion.json
ADDED
@@ -0,0 +1,157 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"ckpt_dir": "$@bundle_root + '/models'",
|
3 |
+
"train_batch_size": 4,
|
4 |
+
"lr": 1e-05,
|
5 |
+
"train_patch_size": [
|
6 |
+
144,
|
7 |
+
176,
|
8 |
+
112
|
9 |
+
],
|
10 |
+
"latent_shape": [
|
11 |
+
"@latent_channels",
|
12 |
+
36,
|
13 |
+
44,
|
14 |
+
28
|
15 |
+
],
|
16 |
+
"load_autoencoder_path": "$@bundle_root + '/models/model_autoencoder.pt'",
|
17 |
+
"load_autoencoder": "$@autoencoder_def.load_state_dict(torch.load(@load_autoencoder_path))",
|
18 |
+
"autoencoder": "$@autoencoder_def.to(@device)",
|
19 |
+
"network_def": {
|
20 |
+
"_target_": "generative.networks.nets.DiffusionModelUNet",
|
21 |
+
"spatial_dims": "@spatial_dims",
|
22 |
+
"in_channels": "@latent_channels",
|
23 |
+
"out_channels": "@latent_channels",
|
24 |
+
"num_channels": [
|
25 |
+
256,
|
26 |
+
256,
|
27 |
+
512
|
28 |
+
],
|
29 |
+
"attention_levels": [
|
30 |
+
false,
|
31 |
+
true,
|
32 |
+
true
|
33 |
+
],
|
34 |
+
"num_head_channels": [
|
35 |
+
0,
|
36 |
+
64,
|
37 |
+
64
|
38 |
+
],
|
39 |
+
"num_res_blocks": 2
|
40 |
+
},
|
41 |
+
"diffusion": "$@network_def.to(@device)",
|
42 |
+
"optimizer": {
|
43 |
+
"_target_": "torch.optim.Adam",
|
44 |
+
"params": "$@diffusion.parameters()",
|
45 |
+
"lr": "@lr"
|
46 |
+
},
|
47 |
+
"lr_scheduler": {
|
48 |
+
"_target_": "torch.optim.lr_scheduler.MultiStepLR",
|
49 |
+
"optimizer": "@optimizer",
|
50 |
+
"milestones": [
|
51 |
+
100,
|
52 |
+
1000
|
53 |
+
],
|
54 |
+
"gamma": 0.1
|
55 |
+
},
|
56 |
+
"scale_factor": "$scripts.utils.compute_scale_factor(@autoencoder,@train#dataloader,@device)",
|
57 |
+
"noise_scheduler": {
|
58 |
+
"_target_": "generative.networks.schedulers.DDPMScheduler",
|
59 |
+
"_requires_": [
|
60 |
+
"@load_autoencoder"
|
61 |
+
],
|
62 |
+
"beta_schedule": "scaled_linear",
|
63 |
+
"num_train_timesteps": 1000,
|
64 |
+
"beta_start": 0.0015,
|
65 |
+
"beta_end": 0.0195
|
66 |
+
},
|
67 |
+
"inferer": {
|
68 |
+
"_target_": "generative.inferers.LatentDiffusionInferer",
|
69 |
+
"scheduler": "@noise_scheduler",
|
70 |
+
"scale_factor": "@scale_factor"
|
71 |
+
},
|
72 |
+
"loss": {
|
73 |
+
"_target_": "torch.nn.MSELoss"
|
74 |
+
},
|
75 |
+
"train": {
|
76 |
+
"crop_transforms": [
|
77 |
+
{
|
78 |
+
"_target_": "CenterSpatialCropd",
|
79 |
+
"keys": "image",
|
80 |
+
"roi_size": "@train_patch_size"
|
81 |
+
}
|
82 |
+
],
|
83 |
+
"preprocessing": {
|
84 |
+
"_target_": "Compose",
|
85 |
+
"transforms": "$@preprocessing_transforms + @train#crop_transforms + @final_transforms"
|
86 |
+
},
|
87 |
+
"dataset": {
|
88 |
+
"_target_": "monai.apps.DecathlonDataset",
|
89 |
+
"root_dir": "@dataset_dir",
|
90 |
+
"task": "Task01_BrainTumour",
|
91 |
+
"section": "training",
|
92 |
+
"cache_rate": 1.0,
|
93 |
+
"num_workers": 8,
|
94 |
+
"download": false,
|
95 |
+
"transform": "@train#preprocessing"
|
96 |
+
},
|
97 |
+
"dataloader": {
|
98 |
+
"_target_": "DataLoader",
|
99 |
+
"dataset": "@train#dataset",
|
100 |
+
"batch_size": "@train_batch_size",
|
101 |
+
"shuffle": true,
|
102 |
+
"num_workers": 0
|
103 |
+
},
|
104 |
+
"handlers": [
|
105 |
+
{
|
106 |
+
"_target_": "LrScheduleHandler",
|
107 |
+
"lr_scheduler": "@lr_scheduler",
|
108 |
+
"print_lr": true
|
109 |
+
},
|
110 |
+
{
|
111 |
+
"_target_": "CheckpointSaver",
|
112 |
+
"save_dir": "@ckpt_dir",
|
113 |
+
"save_dict": {
|
114 |
+
"model": "@diffusion"
|
115 |
+
},
|
116 |
+
"save_interval": 0,
|
117 |
+
"save_final": true,
|
118 |
+
"epoch_level": true,
|
119 |
+
"final_filename": "model.pt"
|
120 |
+
},
|
121 |
+
{
|
122 |
+
"_target_": "StatsHandler",
|
123 |
+
"tag_name": "train_diffusion_loss",
|
124 |
+
"output_transform": "$lambda x: monai.handlers.from_engine(['loss'], first=True)(x)"
|
125 |
+
},
|
126 |
+
{
|
127 |
+
"_target_": "TensorBoardStatsHandler",
|
128 |
+
"log_dir": "@tf_dir",
|
129 |
+
"tag_name": "train_diffusion_loss",
|
130 |
+
"output_transform": "$lambda x: monai.handlers.from_engine(['loss'], first=True)(x)"
|
131 |
+
}
|
132 |
+
],
|
133 |
+
"trainer": {
|
134 |
+
"_target_": "scripts.ldm_trainer.LDMTrainer",
|
135 |
+
"device": "@device",
|
136 |
+
"max_epochs": 5000,
|
137 |
+
"train_data_loader": "@train#dataloader",
|
138 |
+
"network": "@diffusion",
|
139 |
+
"autoencoder_model": "@autoencoder",
|
140 |
+
"optimizer": "@optimizer",
|
141 |
+
"loss_function": "@loss",
|
142 |
+
"latent_shape": "@latent_shape",
|
143 |
+
"inferer": "@inferer",
|
144 |
+
"key_train_metric": "$None",
|
145 |
+
"train_handlers": "@train#handlers"
|
146 |
+
}
|
147 |
+
},
|
148 |
+
"initialize": [
|
149 |
+
"$monai.utils.set_determinism(seed=0)"
|
150 |
+
],
|
151 |
+
"run": [
|
152 |
+
"@load_autoencoder",
|
153 |
+
"$@autoencoder.eval()",
|
154 |
+
"$print('scale factor:',@scale_factor)",
|
155 |
+
"$@train#trainer.run()"
|
156 |
+
]
|
157 |
+
}
|
docs/README.md
ADDED
@@ -0,0 +1,181 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Model Overview
|
2 |
+
A pre-trained model for volumetric (3D) Brats MRI 3D Latent Diffusion Generative Model.
|
3 |
+
|
4 |
+
This model is trained on BraTS 2016 and 2017 data from [Medical Decathlon](http://medicaldecathlon.com/), using the Latent diffusion model [1].
|
5 |
+
|
6 |
+
![model workflow](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_network.png)
|
7 |
+
|
8 |
+
This model is a generator for creating images like the Flair MRIs based on BraTS 2016 and 2017 data. It was trained as a 3d latent diffusion model and accepts Gaussian random noise as inputs to produce an image output. The `train_autoencoder.json` file describes the training process of the variational autoencoder with GAN loss. The `train_diffusion.json` file describes the training process of the 3D latent diffusion model.
|
9 |
+
|
10 |
+
In this bundle, the autoencoder uses perceptual loss, which is based on ResNet50 with pre-trained weights (the network is frozen and will not be trained in the bundle). In default, the `pretrained` parameter is specified as `False` in `train_autoencoder.json`. To ensure correct training, changing the default settings is necessary. There are two ways to utilize pretrained weights:
|
11 |
+
1. if set `pretrained` to `True`, ImageNet pretrained weights from [torchvision](https://pytorch.org/vision/stable/_modules/torchvision/models/resnet.html#ResNet50_Weights) will be used. However, the weights are for non-commercial use only.
|
12 |
+
2. if set `pretrained` to `True` and specifies the `perceptual_loss_model_weights_path` parameter, users are able to load weights from a local path. This is the way this bundle used to train, and the pre-trained weights are from some internal data.
|
13 |
+
|
14 |
+
Please note that each user is responsible for checking the data source of the pre-trained models, the applicable licenses, and determining if suitable for the intended use.
|
15 |
+
|
16 |
+
#### Example synthetic image
|
17 |
+
An example result from inference is shown below:
|
18 |
+
![Example synthetic image](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_example_generation.png)
|
19 |
+
|
20 |
+
**This is a demonstration network meant to just show the training process for this sort of network with MONAI. To achieve better performance, users need to use larger dataset like [Brats 2021](https://www.synapse.org/#!Synapse:syn25829067/wiki/610865) and have GPU with memory larger than 32G to enable larger networks and attention layers.**
|
21 |
+
|
22 |
+
## MONAI Generative Model Dependencies
|
23 |
+
[MONAI generative models](https://github.com/Project-MONAI/GenerativeModels) can be installed by
|
24 |
+
```
|
25 |
+
pip install lpips==0.1.4
|
26 |
+
git clone https://github.com/Project-MONAI/GenerativeModels.git
|
27 |
+
cd GenerativeModels/
|
28 |
+
git checkout f969c24f88d013dc0045fb7b2885a01fb219992b
|
29 |
+
python setup.py install
|
30 |
+
cd ..
|
31 |
+
```
|
32 |
+
|
33 |
+
## Data
|
34 |
+
The training data is BraTS 2016 and 2017 from the Medical Segmentation Decathalon. Users can find more details on the dataset (`Task01_BrainTumour`) at http://medicaldecathlon.com/.
|
35 |
+
|
36 |
+
- Target: Image Generation
|
37 |
+
- Task: Synthesis
|
38 |
+
- Modality: MRI
|
39 |
+
- Size: 388 3D volumes (1 channel used)
|
40 |
+
|
41 |
+
## Training Configuration
|
42 |
+
If you have a GPU with less than 32G of memory, you may need to decrease the batch size when training. To do so, modify the `train_batch_size` parameter in the [configs/train_autoencoder.json](../configs/train_autoencoder.json) and [configs/train_diffusion.json](../configs/train_diffusion.json) configuration files.
|
43 |
+
|
44 |
+
### Training Configuration of Autoencoder
|
45 |
+
The autoencoder was trained using the following configuration:
|
46 |
+
|
47 |
+
- GPU: at least 32GB GPU memory
|
48 |
+
- Actual Model Input: 112 x 128 x 80
|
49 |
+
- AMP: False
|
50 |
+
- Optimizer: Adam
|
51 |
+
- Learning Rate: 1e-5
|
52 |
+
- Loss: L1 loss, perceptual loss, KL divergence loss, adversarial loss, GAN BCE loss
|
53 |
+
|
54 |
+
#### Input
|
55 |
+
1 channel 3D MRI Flair patches
|
56 |
+
|
57 |
+
#### Output
|
58 |
+
- 1 channel 3D MRI reconstructed patches
|
59 |
+
- 8 channel mean of latent features
|
60 |
+
- 8 channel standard deviation of latent features
|
61 |
+
|
62 |
+
### Training Configuration of Diffusion Model
|
63 |
+
The latent diffusion model was trained using the following configuration:
|
64 |
+
|
65 |
+
- GPU: at least 32GB GPU memory
|
66 |
+
- Actual Model Input: 36 x 44 x 28
|
67 |
+
- AMP: False
|
68 |
+
- Optimizer: Adam
|
69 |
+
- Learning Rate: 1e-5
|
70 |
+
- Loss: MSE loss
|
71 |
+
|
72 |
+
#### Training Input
|
73 |
+
- 8 channel noisy latent features
|
74 |
+
- an int that indicates the time step
|
75 |
+
|
76 |
+
#### Training Output
|
77 |
+
8 channel predicted added noise
|
78 |
+
|
79 |
+
#### Inference Input
|
80 |
+
8 channel noise
|
81 |
+
|
82 |
+
#### Inference Output
|
83 |
+
8 channel denoised latent features
|
84 |
+
|
85 |
+
### Memory Consumption Warning
|
86 |
+
|
87 |
+
If you face memory issues with data loading, you can lower the caching rate `cache_rate` in the configurations within range [0, 1] to minimize the System RAM requirements.
|
88 |
+
|
89 |
+
## Performance
|
90 |
+
|
91 |
+
#### Training Loss
|
92 |
+
![A graph showing the autoencoder training curve](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_train_autoencoder_loss.png)
|
93 |
+
|
94 |
+
![A graph showing the latent diffusion training curve](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_train_diffusion_loss.png)
|
95 |
+
|
96 |
+
## MONAI Bundle Commands
|
97 |
+
|
98 |
+
In addition to the Pythonic APIs, a few command line interfaces (CLI) are provided to interact with the bundle. The CLI supports flexible use cases, such as overriding configs at runtime and predefining arguments in a file.
|
99 |
+
|
100 |
+
For more details usage instructions, visit the [MONAI Bundle Configuration Page](https://docs.monai.io/en/latest/config_syntax.html).
|
101 |
+
|
102 |
+
### Execute Autoencoder Training
|
103 |
+
|
104 |
+
#### Execute Autoencoder Training on single GPU
|
105 |
+
|
106 |
+
```
|
107 |
+
python -m monai.bundle run --config_file configs/train_autoencoder.json
|
108 |
+
```
|
109 |
+
|
110 |
+
Please note that if the default dataset path is not modified with the actual path (it should be the path that contains `Task01_BrainTumour`) in the bundle config files, you can also override it by using `--dataset_dir`:
|
111 |
+
|
112 |
+
```
|
113 |
+
python -m monai.bundle run --config_file configs/train_autoencoder.json --dataset_dir <actual dataset path>
|
114 |
+
```
|
115 |
+
|
116 |
+
#### Override the `train` config to execute multi-GPU training for Autoencoder
|
117 |
+
To train with multiple GPUs, use the following command, which requires scaling up the learning rate according to the number of GPUs.
|
118 |
+
|
119 |
+
```
|
120 |
+
torchrun --standalone --nnodes=1 --nproc_per_node=8 -m monai.bundle run --config_file "['configs/train_autoencoder.json','configs/multi_gpu_train_autoencoder.json']" --lr 8e-5
|
121 |
+
```
|
122 |
+
|
123 |
+
#### Check the Autoencoder Training result
|
124 |
+
The following code generates a reconstructed image from a random input image.
|
125 |
+
We can visualize it to see if the autoencoder is trained correctly.
|
126 |
+
```
|
127 |
+
python -m monai.bundle run --config_file configs/inference_autoencoder.json
|
128 |
+
```
|
129 |
+
|
130 |
+
An example of reconstructed image from inference is shown below. If the autoencoder is trained correctly, the reconstructed image should look similar to original image.
|
131 |
+
|
132 |
+
![Example reconstructed image](https://developer.download.nvidia.com/assets/Clara/Images/monai_brain_image_gen_ldm3d_recon_example.jpg)
|
133 |
+
|
134 |
+
|
135 |
+
### Execute Latent Diffusion Training
|
136 |
+
|
137 |
+
#### Execute Latent Diffusion Model Training on single GPU
|
138 |
+
After training the autoencoder, run the following command to train the latent diffusion model. This command will print out the scale factor of the latent feature space. If your autoencoder is well trained, this value should be close to 1.0.
|
139 |
+
|
140 |
+
```
|
141 |
+
python -m monai.bundle run --config_file "['configs/train_autoencoder.json','configs/train_diffusion.json']"
|
142 |
+
```
|
143 |
+
|
144 |
+
#### Override the `train` config to execute multi-GPU training for Latent Diffusion Model
|
145 |
+
To train with multiple GPUs, use the following command, which requires scaling up the learning rate according to the number of GPUs.
|
146 |
+
|
147 |
+
```
|
148 |
+
torchrun --standalone --nnodes=1 --nproc_per_node=8 -m monai.bundle run --config_file "['configs/train_autoencoder.json','configs/train_diffusion.json','configs/multi_gpu_train_autoencoder.json','configs/multi_gpu_train_diffusion.json']" --lr 8e-5
|
149 |
+
```
|
150 |
+
|
151 |
+
#### Execute inference
|
152 |
+
The following code generates a synthetic image from a random sampled noise.
|
153 |
+
```
|
154 |
+
python -m monai.bundle run --config_file configs/inference.json
|
155 |
+
```
|
156 |
+
|
157 |
+
#### Export checkpoint to TorchScript file
|
158 |
+
|
159 |
+
The Autoencoder can be exported into a TorchScript file.
|
160 |
+
|
161 |
+
```
|
162 |
+
python -m monai.bundle ckpt_export autoencoder_def --filepath models/model_autoencoder.ts --ckpt_file models/model_autoencoder.pt --meta_file configs/metadata.json --config_file configs/inference.json
|
163 |
+
```
|
164 |
+
|
165 |
+
# References
|
166 |
+
[1] Rombach, Robin, et al. "High-resolution image synthesis with latent diffusion models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022. https://openaccess.thecvf.com/content/CVPR2022/papers/Rombach_High-Resolution_Image_Synthesis_With_Latent_Diffusion_Models_CVPR_2022_paper.pdf
|
167 |
+
|
168 |
+
# License
|
169 |
+
Copyright (c) MONAI Consortium
|
170 |
+
|
171 |
+
Licensed under the Apache License, Version 2.0 (the "License");
|
172 |
+
you may not use this file except in compliance with the License.
|
173 |
+
You may obtain a copy of the License at
|
174 |
+
|
175 |
+
http://www.apache.org/licenses/LICENSE-2.0
|
176 |
+
|
177 |
+
Unless required by applicable law or agreed to in writing, software
|
178 |
+
distributed under the License is distributed on an "AS IS" BASIS,
|
179 |
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
180 |
+
See the License for the specific language governing permissions and
|
181 |
+
limitations under the License.
|
docs/data_license.txt
ADDED
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Third Party Licenses
|
2 |
+
-----------------------------------------------------------------------
|
3 |
+
|
4 |
+
/*********************************************************************/
|
5 |
+
i. Multimodal Brain Tumor Segmentation Challenge 2018
|
6 |
+
https://www.med.upenn.edu/sbia/brats2018/data.html
|
7 |
+
/*********************************************************************/
|
8 |
+
|
9 |
+
Data Usage Agreement / Citations
|
10 |
+
|
11 |
+
You are free to use and/or refer to the BraTS datasets in your own
|
12 |
+
research, provided that you always cite the following two manuscripts:
|
13 |
+
|
14 |
+
[1] Menze BH, Jakab A, Bauer S, Kalpathy-Cramer J, Farahani K, Kirby
|
15 |
+
[J, Burren Y, Porz N, Slotboom J, Wiest R, Lanczi L, Gerstner E, Weber
|
16 |
+
[MA, Arbel T, Avants BB, Ayache N, Buendia P, Collins DL, Cordier N,
|
17 |
+
[Corso JJ, Criminisi A, Das T, Delingette H, Demiralp Γ, Durst CR,
|
18 |
+
[Dojat M, Doyle S, Festa J, Forbes F, Geremia E, Glocker B, Golland P,
|
19 |
+
[Guo X, Hamamci A, Iftekharuddin KM, Jena R, John NM, Konukoglu E,
|
20 |
+
[Lashkari D, Mariz JA, Meier R, Pereira S, Precup D, Price SJ, Raviv
|
21 |
+
[TR, Reza SM, Ryan M, Sarikaya D, Schwartz L, Shin HC, Shotton J,
|
22 |
+
[Silva CA, Sousa N, Subbanna NK, Szekely G, Taylor TJ, Thomas OM,
|
23 |
+
[Tustison NJ, Unal G, Vasseur F, Wintermark M, Ye DH, Zhao L, Zhao B,
|
24 |
+
[Zikic D, Prastawa M, Reyes M, Van Leemput K. "The Multimodal Brain
|
25 |
+
[Tumor Image Segmentation Benchmark (BRATS)", IEEE Transactions on
|
26 |
+
[Medical Imaging 34(10), 1993-2024 (2015) DOI:
|
27 |
+
[10.1109/TMI.2014.2377694
|
28 |
+
|
29 |
+
[2] Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby JS,
|
30 |
+
[Freymann JB, Farahani K, Davatzikos C. "Advancing The Cancer Genome
|
31 |
+
[Atlas glioma MRI collections with expert segmentation labels and
|
32 |
+
[radiomic features", Nature Scientific Data, 4:170117 (2017) DOI:
|
33 |
+
[10.1038/sdata.2017.117
|
34 |
+
|
35 |
+
In addition, if there are no restrictions imposed from the
|
36 |
+
journal/conference you submit your paper about citing "Data
|
37 |
+
Citations", please be specific and also cite the following:
|
38 |
+
|
39 |
+
[3] Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby J,
|
40 |
+
[Freymann J, Farahani K, Davatzikos C. "Segmentation Labels and
|
41 |
+
[Radiomic Features for the Pre-operative Scans of the TCGA-GBM
|
42 |
+
[collection", The Cancer Imaging Archive, 2017. DOI:
|
43 |
+
[10.7937/K9/TCIA.2017.KLXWJJ1Q
|
44 |
+
|
45 |
+
[4] Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby J,
|
46 |
+
[Freymann J, Farahani K, Davatzikos C. "Segmentation Labels and
|
47 |
+
[Radiomic Features for the Pre-operative Scans of the TCGA-LGG
|
48 |
+
[collection", The Cancer Imaging Archive, 2017. DOI:
|
49 |
+
[10.7937/K9/TCIA.2017.GJQ7R0EF
|
models/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d780cd5fff4ec886226c5407391a5906c45e388c5d02efbf20da8729b7513e19
|
3 |
+
size 765042309
|
models/model_autoencoder.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a788c35df8e6d7b8c5baf2108bfc6e105a1e6685dfbf564d8d07a66194de8727
|
3 |
+
size 84050405
|
models/model_autoencoder.ts
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39fa19c3e4cd35298337ad4dd8684961d55539a7cd3e29155e12731b254aca3f
|
3 |
+
size 84147155
|
scripts/__init__.py
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Copyright (c) MONAI Consortium
|
2 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
3 |
+
# you may not use this file except in compliance with the License.
|
4 |
+
# You may obtain a copy of the License at
|
5 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
6 |
+
# Unless required by applicable law or agreed to in writing, software
|
7 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
8 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
9 |
+
# See the License for the specific language governing permissions and
|
10 |
+
# limitations under the License.
|
11 |
+
|
12 |
+
from . import ldm_sampler, ldm_trainer, losses, utils
|
scripts/ldm_sampler.py
ADDED
@@ -0,0 +1,60 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Copyright (c) MONAI Consortium
|
2 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
3 |
+
# you may not use this file except in compliance with the License.
|
4 |
+
# You may obtain a copy of the License at
|
5 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
6 |
+
# Unless required by applicable law or agreed to in writing, software
|
7 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
8 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
9 |
+
# See the License for the specific language governing permissions and
|
10 |
+
# limitations under the License.
|
11 |
+
|
12 |
+
from __future__ import annotations
|
13 |
+
|
14 |
+
import torch
|
15 |
+
import torch.nn as nn
|
16 |
+
from monai.utils import optional_import
|
17 |
+
from torch.cuda.amp import autocast
|
18 |
+
|
19 |
+
tqdm, has_tqdm = optional_import("tqdm", name="tqdm")
|
20 |
+
|
21 |
+
|
22 |
+
class LDMSampler:
|
23 |
+
def __init__(self) -> None:
|
24 |
+
super().__init__()
|
25 |
+
|
26 |
+
@torch.no_grad()
|
27 |
+
def sampling_fn(
|
28 |
+
self,
|
29 |
+
input_noise: torch.Tensor,
|
30 |
+
autoencoder_model: nn.Module,
|
31 |
+
diffusion_model: nn.Module,
|
32 |
+
scheduler: nn.Module,
|
33 |
+
conditioning: torch.Tensor | None = None,
|
34 |
+
) -> torch.Tensor:
|
35 |
+
if has_tqdm:
|
36 |
+
progress_bar = tqdm(scheduler.timesteps)
|
37 |
+
else:
|
38 |
+
progress_bar = iter(scheduler.timesteps)
|
39 |
+
|
40 |
+
image = input_noise
|
41 |
+
if conditioning is not None:
|
42 |
+
cond_concat = conditioning.squeeze(1).unsqueeze(-1).unsqueeze(-1).unsqueeze(-1)
|
43 |
+
cond_concat = cond_concat.expand(list(cond_concat.shape[0:2]) + list(input_noise.shape[2:]))
|
44 |
+
|
45 |
+
for t in progress_bar:
|
46 |
+
with torch.no_grad():
|
47 |
+
if conditioning is not None:
|
48 |
+
input_t = torch.cat((image, cond_concat), dim=1)
|
49 |
+
else:
|
50 |
+
input_t = image
|
51 |
+
model_output = diffusion_model(
|
52 |
+
input_t, timesteps=torch.Tensor((t,)).to(input_noise.device).long(), context=conditioning
|
53 |
+
)
|
54 |
+
image, _ = scheduler.step(model_output, t, image)
|
55 |
+
|
56 |
+
with torch.no_grad():
|
57 |
+
with autocast():
|
58 |
+
sample = autoencoder_model.decode_stage_2_outputs(image)
|
59 |
+
|
60 |
+
return sample
|
scripts/ldm_trainer.py
ADDED
@@ -0,0 +1,380 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Copyright (c) MONAI Consortium
|
2 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
3 |
+
# you may not use this file except in compliance with the License.
|
4 |
+
# You may obtain a copy of the License at
|
5 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
6 |
+
# Unless required by applicable law or agreed to in writing, software
|
7 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
8 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
9 |
+
# See the License for the specific language governing permissions and
|
10 |
+
# limitations under the License.
|
11 |
+
|
12 |
+
from __future__ import annotations
|
13 |
+
|
14 |
+
from typing import TYPE_CHECKING, Any, Callable, Iterable, Sequence
|
15 |
+
|
16 |
+
import torch
|
17 |
+
from monai.config import IgniteInfo
|
18 |
+
from monai.engines.utils import IterationEvents, default_metric_cmp_fn, default_prepare_batch
|
19 |
+
from monai.inferers import Inferer, SimpleInferer
|
20 |
+
from monai.transforms import Transform
|
21 |
+
from monai.utils import min_version, optional_import
|
22 |
+
from monai.utils.enums import CommonKeys, GanKeys
|
23 |
+
from torch.optim.optimizer import Optimizer
|
24 |
+
from torch.utils.data import DataLoader
|
25 |
+
|
26 |
+
if TYPE_CHECKING:
|
27 |
+
from ignite.engine import Engine, EventEnum
|
28 |
+
from ignite.metrics import Metric
|
29 |
+
else:
|
30 |
+
Engine, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Engine")
|
31 |
+
Metric, _ = optional_import("ignite.metrics", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Metric")
|
32 |
+
EventEnum, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "EventEnum")
|
33 |
+
from monai.engines.trainer import SupervisedTrainer, Trainer
|
34 |
+
|
35 |
+
|
36 |
+
class VaeGanTrainer(Trainer):
|
37 |
+
"""
|
38 |
+
Generative adversarial network training based on Goodfellow et al. 2014 https://arxiv.org/abs/1406.266,
|
39 |
+
inherits from ``Trainer`` and ``Workflow``.
|
40 |
+
Training Loop: for each batch of data size `m`
|
41 |
+
1. Generate `m` fakes from random latent codes.
|
42 |
+
2. Update discriminator with these fakes and current batch reals, repeated d_train_steps times.
|
43 |
+
3. If g_update_latents, generate `m` fakes from new random latent codes.
|
44 |
+
4. Update generator with these fakes using discriminator feedback.
|
45 |
+
Args:
|
46 |
+
device: an object representing the device on which to run.
|
47 |
+
max_epochs: the total epoch number for engine to run.
|
48 |
+
train_data_loader: Core ignite engines uses `DataLoader` for training loop batchdata.
|
49 |
+
g_network: generator (G) network architecture.
|
50 |
+
g_optimizer: G optimizer function.
|
51 |
+
g_loss_function: G loss function for optimizer.
|
52 |
+
d_network: discriminator (D) network architecture.
|
53 |
+
d_optimizer: D optimizer function.
|
54 |
+
d_loss_function: D loss function for optimizer.
|
55 |
+
epoch_length: number of iterations for one epoch, default to `len(train_data_loader)`.
|
56 |
+
g_inferer: inference method to execute G model forward. Defaults to ``SimpleInferer()``.
|
57 |
+
d_inferer: inference method to execute D model forward. Defaults to ``SimpleInferer()``.
|
58 |
+
d_train_steps: number of times to update D with real data minibatch. Defaults to ``1``.
|
59 |
+
latent_shape: size of G input latent code. Defaults to ``64``.
|
60 |
+
non_blocking: if True and this copy is between CPU and GPU, the copy may occur asynchronously
|
61 |
+
with respect to the host. For other cases, this argument has no effect.
|
62 |
+
d_prepare_batch: callback function to prepare batchdata for D inferer.
|
63 |
+
Defaults to return ``GanKeys.REALS`` in batchdata dict. for more details please refer to:
|
64 |
+
https://pytorch.org/ignite/generated/ignite.engine.create_supervised_trainer.html.
|
65 |
+
g_prepare_batch: callback function to create batch of latent input for G inferer.
|
66 |
+
Defaults to return random latents. for more details please refer to:
|
67 |
+
https://pytorch.org/ignite/generated/ignite.engine.create_supervised_trainer.html.
|
68 |
+
g_update_latents: Calculate G loss with new latent codes. Defaults to ``True``.
|
69 |
+
iteration_update: the callable function for every iteration, expect to accept `engine`
|
70 |
+
and `engine.state.batch` as inputs, return data will be stored in `engine.state.output`.
|
71 |
+
if not provided, use `self._iteration()` instead. for more details please refer to:
|
72 |
+
https://pytorch.org/ignite/generated/ignite.engine.engine.Engine.html.
|
73 |
+
postprocessing: execute additional transformation for the model output data.
|
74 |
+
Typically, several Tensor based transforms composed by `Compose`.
|
75 |
+
key_train_metric: compute metric when every iteration completed, and save average value to
|
76 |
+
engine.state.metrics when epoch completed. key_train_metric is the main metric to compare and save the
|
77 |
+
checkpoint into files.
|
78 |
+
additional_metrics: more Ignite metrics that also attach to Ignite Engine.
|
79 |
+
metric_cmp_fn: function to compare current key metric with previous best key metric value,
|
80 |
+
it must accept 2 args (current_metric, previous_best) and return a bool result: if `True`, will update
|
81 |
+
`best_metric` and `best_metric_epoch` with current metric and epoch, default to `greater than`.
|
82 |
+
train_handlers: every handler is a set of Ignite Event-Handlers, must have `attach` function, like:
|
83 |
+
CheckpointHandler, StatsHandler, etc.
|
84 |
+
decollate: whether to decollate the batch-first data to a list of data after model computation,
|
85 |
+
recommend `decollate=True` when `postprocessing` uses components from `monai.transforms`.
|
86 |
+
default to `True`.
|
87 |
+
optim_set_to_none: when calling `optimizer.zero_grad()`, instead of setting to zero, set the grads to None.
|
88 |
+
more details: https://pytorch.org/docs/stable/generated/torch.optim.Optimizer.zero_grad.html.
|
89 |
+
to_kwargs: dict of other args for `prepare_batch` API when converting the input data, except for
|
90 |
+
`device`, `non_blocking`.
|
91 |
+
amp_kwargs: dict of the args for `torch.cuda.amp.autocast()` API, for more details:
|
92 |
+
https://pytorch.org/docs/stable/amp.html#torch.cuda.amp.autocast.
|
93 |
+
"""
|
94 |
+
|
95 |
+
def __init__(
|
96 |
+
self,
|
97 |
+
device: str | torch.device,
|
98 |
+
max_epochs: int,
|
99 |
+
train_data_loader: DataLoader,
|
100 |
+
g_network: torch.nn.Module,
|
101 |
+
g_optimizer: Optimizer,
|
102 |
+
g_loss_function: Callable,
|
103 |
+
d_network: torch.nn.Module,
|
104 |
+
d_optimizer: Optimizer,
|
105 |
+
d_loss_function: Callable,
|
106 |
+
epoch_length: int | None = None,
|
107 |
+
g_inferer: Inferer | None = None,
|
108 |
+
d_inferer: Inferer | None = None,
|
109 |
+
d_train_steps: int = 1,
|
110 |
+
latent_shape: int = 64,
|
111 |
+
non_blocking: bool = False,
|
112 |
+
d_prepare_batch: Callable = default_prepare_batch,
|
113 |
+
g_prepare_batch: Callable = default_prepare_batch,
|
114 |
+
g_update_latents: bool = True,
|
115 |
+
iteration_update: Callable[[Engine, Any], Any] | None = None,
|
116 |
+
postprocessing: Transform | None = None,
|
117 |
+
key_train_metric: dict[str, Metric] | None = None,
|
118 |
+
additional_metrics: dict[str, Metric] | None = None,
|
119 |
+
metric_cmp_fn: Callable = default_metric_cmp_fn,
|
120 |
+
train_handlers: Sequence | None = None,
|
121 |
+
decollate: bool = True,
|
122 |
+
optim_set_to_none: bool = False,
|
123 |
+
to_kwargs: dict | None = None,
|
124 |
+
amp_kwargs: dict | None = None,
|
125 |
+
):
|
126 |
+
if not isinstance(train_data_loader, DataLoader):
|
127 |
+
raise ValueError("train_data_loader must be PyTorch DataLoader.")
|
128 |
+
|
129 |
+
# set up Ignite engine and environments
|
130 |
+
super().__init__(
|
131 |
+
device=device,
|
132 |
+
max_epochs=max_epochs,
|
133 |
+
data_loader=train_data_loader,
|
134 |
+
epoch_length=epoch_length,
|
135 |
+
non_blocking=non_blocking,
|
136 |
+
prepare_batch=d_prepare_batch,
|
137 |
+
iteration_update=iteration_update,
|
138 |
+
key_metric=key_train_metric,
|
139 |
+
additional_metrics=additional_metrics,
|
140 |
+
metric_cmp_fn=metric_cmp_fn,
|
141 |
+
handlers=train_handlers,
|
142 |
+
postprocessing=postprocessing,
|
143 |
+
decollate=decollate,
|
144 |
+
to_kwargs=to_kwargs,
|
145 |
+
amp_kwargs=amp_kwargs,
|
146 |
+
)
|
147 |
+
self.g_network = g_network
|
148 |
+
self.g_optimizer = g_optimizer
|
149 |
+
self.g_loss_function = g_loss_function
|
150 |
+
self.g_inferer = SimpleInferer() if g_inferer is None else g_inferer
|
151 |
+
self.d_network = d_network
|
152 |
+
self.d_optimizer = d_optimizer
|
153 |
+
self.d_loss_function = d_loss_function
|
154 |
+
self.d_inferer = SimpleInferer() if d_inferer is None else d_inferer
|
155 |
+
self.d_train_steps = d_train_steps
|
156 |
+
self.latent_shape = latent_shape
|
157 |
+
self.g_prepare_batch = g_prepare_batch
|
158 |
+
self.g_update_latents = g_update_latents
|
159 |
+
self.optim_set_to_none = optim_set_to_none
|
160 |
+
|
161 |
+
def _iteration(
|
162 |
+
self, engine: VaeGanTrainer, batchdata: dict | Sequence
|
163 |
+
) -> dict[str, torch.Tensor | int | float | bool]:
|
164 |
+
"""
|
165 |
+
Callback function for Adversarial Training processing logic of 1 iteration in Ignite Engine.
|
166 |
+
Args:
|
167 |
+
engine: `VaeGanTrainer` to execute operation for an iteration.
|
168 |
+
batchdata: input data for this iteration, usually can be dictionary or tuple of Tensor data.
|
169 |
+
Raises:
|
170 |
+
ValueError: must provide batch data for current iteration.
|
171 |
+
"""
|
172 |
+
if batchdata is None:
|
173 |
+
raise ValueError("must provide batch data for current iteration.")
|
174 |
+
|
175 |
+
d_input = engine.prepare_batch(batchdata, engine.state.device, engine.non_blocking, **engine.to_kwargs)[0]
|
176 |
+
g_input = d_input
|
177 |
+
g_output, z_mu, z_sigma = engine.g_inferer(g_input, engine.g_network)
|
178 |
+
|
179 |
+
# Train Discriminator
|
180 |
+
d_total_loss = torch.zeros(1)
|
181 |
+
for _ in range(engine.d_train_steps):
|
182 |
+
engine.d_optimizer.zero_grad(set_to_none=engine.optim_set_to_none)
|
183 |
+
dloss = engine.d_loss_function(g_output, d_input)
|
184 |
+
dloss.backward()
|
185 |
+
engine.d_optimizer.step()
|
186 |
+
d_total_loss += dloss.item()
|
187 |
+
|
188 |
+
# Train Generator
|
189 |
+
engine.g_optimizer.zero_grad(set_to_none=engine.optim_set_to_none)
|
190 |
+
g_loss = engine.g_loss_function(g_output, g_input, z_mu, z_sigma)
|
191 |
+
g_loss.backward()
|
192 |
+
engine.g_optimizer.step()
|
193 |
+
|
194 |
+
return {
|
195 |
+
GanKeys.REALS: d_input,
|
196 |
+
GanKeys.FAKES: g_output,
|
197 |
+
GanKeys.LATENTS: g_input,
|
198 |
+
GanKeys.GLOSS: g_loss.item(),
|
199 |
+
GanKeys.DLOSS: d_total_loss.item(),
|
200 |
+
}
|
201 |
+
|
202 |
+
|
203 |
+
class LDMTrainer(SupervisedTrainer):
|
204 |
+
"""
|
205 |
+
Standard supervised training method with image and label, inherits from ``Trainer`` and ``Workflow``.
|
206 |
+
Args:
|
207 |
+
device: an object representing the device on which to run.
|
208 |
+
max_epochs: the total epoch number for trainer to run.
|
209 |
+
train_data_loader: Ignite engine use data_loader to run, must be Iterable or torch.DataLoader.
|
210 |
+
network: network to train in the trainer, should be regular PyTorch `torch.nn.Module`.
|
211 |
+
optimizer: the optimizer associated to the network, should be regular PyTorch optimizer from `torch.optim`
|
212 |
+
or its subclass.
|
213 |
+
loss_function: the loss function associated to the optimizer, should be regular PyTorch loss,
|
214 |
+
which inherit from `torch.nn.modules.loss`.
|
215 |
+
epoch_length: number of iterations for one epoch, default to `len(train_data_loader)`.
|
216 |
+
non_blocking: if True and this copy is between CPU and GPU, the copy may occur asynchronously
|
217 |
+
with respect to the host. For other cases, this argument has no effect.
|
218 |
+
prepare_batch: function to parse expected data (usually `image`, `label` and other network args)
|
219 |
+
from `engine.state.batch` for every iteration, for more details please refer to:
|
220 |
+
https://pytorch.org/ignite/generated/ignite.engine.create_supervised_trainer.html.
|
221 |
+
iteration_update: the callable function for every iteration, expect to accept `engine`
|
222 |
+
and `engine.state.batch` as inputs, return data will be stored in `engine.state.output`.
|
223 |
+
if not provided, use `self._iteration()` instead. for more details please refer to:
|
224 |
+
https://pytorch.org/ignite/generated/ignite.engine.engine.Engine.html.
|
225 |
+
inferer: inference method that execute model forward on input data, like: SlidingWindow, etc.
|
226 |
+
postprocessing: execute additional transformation for the model output data.
|
227 |
+
Typically, several Tensor based transforms composed by `Compose`.
|
228 |
+
key_train_metric: compute metric when every iteration completed, and save average value to
|
229 |
+
engine.state.metrics when epoch completed. key_train_metric is the main metric to compare and save the
|
230 |
+
checkpoint into files.
|
231 |
+
additional_metrics: more Ignite metrics that also attach to Ignite Engine.
|
232 |
+
metric_cmp_fn: function to compare current key metric with previous best key metric value,
|
233 |
+
it must accept 2 args (current_metric, previous_best) and return a bool result: if `True`, will update
|
234 |
+
`best_metric` and `best_metric_epoch` with current metric and epoch, default to `greater than`.
|
235 |
+
train_handlers: every handler is a set of Ignite Event-Handlers, must have `attach` function, like:
|
236 |
+
CheckpointHandler, StatsHandler, etc.
|
237 |
+
amp: whether to enable auto-mixed-precision training, default is False.
|
238 |
+
event_names: additional custom ignite events that will register to the engine.
|
239 |
+
new events can be a list of str or `ignite.engine.events.EventEnum`.
|
240 |
+
event_to_attr: a dictionary to map an event to a state attribute, then add to `engine.state`.
|
241 |
+
for more details, check: https://pytorch.org/ignite/generated/ignite.engine.engine.Engine.html
|
242 |
+
#ignite.engine.engine.Engine.register_events.
|
243 |
+
decollate: whether to decollate the batch-first data to a list of data after model computation,
|
244 |
+
recommend `decollate=True` when `postprocessing` uses components from `monai.transforms`.
|
245 |
+
default to `True`.
|
246 |
+
optim_set_to_none: when calling `optimizer.zero_grad()`, instead of setting to zero, set the grads to None.
|
247 |
+
more details: https://pytorch.org/docs/stable/generated/torch.optim.Optimizer.zero_grad.html.
|
248 |
+
to_kwargs: dict of other args for `prepare_batch` API when converting the input data, except for
|
249 |
+
`device`, `non_blocking`.
|
250 |
+
amp_kwargs: dict of the args for `torch.cuda.amp.autocast()` API, for more details:
|
251 |
+
https://pytorch.org/docs/stable/amp.html#torch.cuda.amp.autocast.
|
252 |
+
"""
|
253 |
+
|
254 |
+
def __init__(
|
255 |
+
self,
|
256 |
+
device: str | torch.device,
|
257 |
+
max_epochs: int,
|
258 |
+
train_data_loader: Iterable | DataLoader,
|
259 |
+
network: torch.nn.Module,
|
260 |
+
autoencoder_model: torch.nn.Module,
|
261 |
+
optimizer: Optimizer,
|
262 |
+
loss_function: Callable,
|
263 |
+
latent_shape: Sequence,
|
264 |
+
inferer: Inferer,
|
265 |
+
epoch_length: int | None = None,
|
266 |
+
non_blocking: bool = False,
|
267 |
+
prepare_batch: Callable = default_prepare_batch,
|
268 |
+
iteration_update: Callable[[Engine, Any], Any] | None = None,
|
269 |
+
postprocessing: Transform | None = None,
|
270 |
+
key_train_metric: dict[str, Metric] | None = None,
|
271 |
+
additional_metrics: dict[str, Metric] | None = None,
|
272 |
+
metric_cmp_fn: Callable = default_metric_cmp_fn,
|
273 |
+
train_handlers: Sequence | None = None,
|
274 |
+
amp: bool = False,
|
275 |
+
event_names: list[str | EventEnum | type[EventEnum]] | None = None,
|
276 |
+
event_to_attr: dict | None = None,
|
277 |
+
decollate: bool = True,
|
278 |
+
optim_set_to_none: bool = False,
|
279 |
+
to_kwargs: dict | None = None,
|
280 |
+
amp_kwargs: dict | None = None,
|
281 |
+
) -> None:
|
282 |
+
super().__init__(
|
283 |
+
device=device,
|
284 |
+
max_epochs=max_epochs,
|
285 |
+
train_data_loader=train_data_loader,
|
286 |
+
network=network,
|
287 |
+
optimizer=optimizer,
|
288 |
+
loss_function=loss_function,
|
289 |
+
inferer=inferer,
|
290 |
+
optim_set_to_none=optim_set_to_none,
|
291 |
+
epoch_length=epoch_length,
|
292 |
+
non_blocking=non_blocking,
|
293 |
+
prepare_batch=prepare_batch,
|
294 |
+
iteration_update=iteration_update,
|
295 |
+
postprocessing=postprocessing,
|
296 |
+
key_train_metric=key_train_metric,
|
297 |
+
additional_metrics=additional_metrics,
|
298 |
+
metric_cmp_fn=metric_cmp_fn,
|
299 |
+
train_handlers=train_handlers,
|
300 |
+
amp=amp,
|
301 |
+
event_names=event_names,
|
302 |
+
event_to_attr=event_to_attr,
|
303 |
+
decollate=decollate,
|
304 |
+
to_kwargs=to_kwargs,
|
305 |
+
amp_kwargs=amp_kwargs,
|
306 |
+
)
|
307 |
+
|
308 |
+
self.latent_shape = latent_shape
|
309 |
+
self.autoencoder_model = autoencoder_model
|
310 |
+
|
311 |
+
def _iteration(self, engine: LDMTrainer, batchdata: dict[str, torch.Tensor]) -> dict:
|
312 |
+
"""
|
313 |
+
Callback function for the Supervised Training processing logic of 1 iteration in Ignite Engine.
|
314 |
+
Return below items in a dictionary:
|
315 |
+
- IMAGE: image Tensor data for model input, already moved to device.
|
316 |
+
- LABEL: label Tensor data corresponding to the image, already moved to device.
|
317 |
+
- PRED: prediction result of model.
|
318 |
+
- LOSS: loss value computed by loss function.
|
319 |
+
Args:
|
320 |
+
engine: `SupervisedTrainer` to execute operation for an iteration.
|
321 |
+
batchdata: input data for this iteration, usually can be dictionary or tuple of Tensor data.
|
322 |
+
Raises:
|
323 |
+
ValueError: When ``batchdata`` is None.
|
324 |
+
"""
|
325 |
+
if batchdata is None:
|
326 |
+
raise ValueError("Must provide batch data for current iteration.")
|
327 |
+
batch = engine.prepare_batch(batchdata, engine.state.device, engine.non_blocking, **engine.to_kwargs)
|
328 |
+
if len(batch) == 2:
|
329 |
+
images, labels = batch
|
330 |
+
args: tuple = ()
|
331 |
+
kwargs: dict = {}
|
332 |
+
else:
|
333 |
+
images, labels, args, kwargs = batch
|
334 |
+
# put iteration outputs into engine.state
|
335 |
+
engine.state.output = {CommonKeys.IMAGE: images}
|
336 |
+
|
337 |
+
# generate noise
|
338 |
+
noise_shape = [images.shape[0]] + list(self.latent_shape)
|
339 |
+
noise = torch.randn(noise_shape, dtype=images.dtype).to(images.device)
|
340 |
+
engine.state.output = {"noise": noise}
|
341 |
+
|
342 |
+
# Create timesteps
|
343 |
+
timesteps = torch.randint(
|
344 |
+
0, engine.inferer.scheduler.num_train_timesteps, (images.shape[0],), device=images.device
|
345 |
+
).long()
|
346 |
+
|
347 |
+
def _compute_pred_loss():
|
348 |
+
# predicted noise
|
349 |
+
engine.state.output[CommonKeys.PRED] = engine.inferer(
|
350 |
+
inputs=images,
|
351 |
+
autoencoder_model=self.autoencoder_model,
|
352 |
+
diffusion_model=engine.network,
|
353 |
+
noise=noise,
|
354 |
+
timesteps=timesteps,
|
355 |
+
)
|
356 |
+
engine.fire_event(IterationEvents.FORWARD_COMPLETED)
|
357 |
+
# compute loss
|
358 |
+
engine.state.output[CommonKeys.LOSS] = engine.loss_function(
|
359 |
+
engine.state.output[CommonKeys.PRED], noise
|
360 |
+
).mean()
|
361 |
+
engine.fire_event(IterationEvents.LOSS_COMPLETED)
|
362 |
+
|
363 |
+
engine.network.train()
|
364 |
+
engine.optimizer.zero_grad(set_to_none=engine.optim_set_to_none)
|
365 |
+
|
366 |
+
if engine.amp and engine.scaler is not None:
|
367 |
+
with torch.cuda.amp.autocast(**engine.amp_kwargs):
|
368 |
+
_compute_pred_loss()
|
369 |
+
engine.scaler.scale(engine.state.output[CommonKeys.LOSS]).backward()
|
370 |
+
engine.fire_event(IterationEvents.BACKWARD_COMPLETED)
|
371 |
+
engine.scaler.step(engine.optimizer)
|
372 |
+
engine.scaler.update()
|
373 |
+
else:
|
374 |
+
_compute_pred_loss()
|
375 |
+
engine.state.output[CommonKeys.LOSS].backward()
|
376 |
+
engine.fire_event(IterationEvents.BACKWARD_COMPLETED)
|
377 |
+
engine.optimizer.step()
|
378 |
+
engine.fire_event(IterationEvents.MODEL_COMPLETED)
|
379 |
+
|
380 |
+
return engine.state.output
|
scripts/losses.py
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Copyright (c) MONAI Consortium
|
2 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
3 |
+
# you may not use this file except in compliance with the License.
|
4 |
+
# You may obtain a copy of the License at
|
5 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
6 |
+
# Unless required by applicable law or agreed to in writing, software
|
7 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
8 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
9 |
+
# See the License for the specific language governing permissions and
|
10 |
+
|
11 |
+
import torch
|
12 |
+
from generative.losses import PatchAdversarialLoss
|
13 |
+
|
14 |
+
intensity_loss = torch.nn.L1Loss()
|
15 |
+
adv_loss = PatchAdversarialLoss(criterion="least_squares")
|
16 |
+
|
17 |
+
adv_weight = 0.1
|
18 |
+
perceptual_weight = 0.1
|
19 |
+
# kl_weight: important hyper-parameter.
|
20 |
+
# If too large, decoder cannot recon good results from latent space.
|
21 |
+
# If too small, latent space will not be regularized enough for the diffusion model
|
22 |
+
kl_weight = 1e-7
|
23 |
+
|
24 |
+
|
25 |
+
def compute_kl_loss(z_mu, z_sigma):
|
26 |
+
kl_loss = 0.5 * torch.sum(z_mu.pow(2) + z_sigma.pow(2) - torch.log(z_sigma.pow(2)) - 1, dim=[1, 2, 3, 4])
|
27 |
+
return torch.sum(kl_loss) / kl_loss.shape[0]
|
28 |
+
|
29 |
+
|
30 |
+
def generator_loss(gen_images, real_images, z_mu, z_sigma, disc_net, loss_perceptual):
|
31 |
+
recons_loss = intensity_loss(gen_images, real_images)
|
32 |
+
kl_loss = compute_kl_loss(z_mu, z_sigma)
|
33 |
+
p_loss = loss_perceptual(gen_images.float(), real_images.float())
|
34 |
+
loss_g = recons_loss + kl_weight * kl_loss + perceptual_weight * p_loss
|
35 |
+
|
36 |
+
logits_fake = disc_net(gen_images)[-1]
|
37 |
+
generator_loss = adv_loss(logits_fake, target_is_real=True, for_discriminator=False)
|
38 |
+
loss_g = loss_g + adv_weight * generator_loss
|
39 |
+
|
40 |
+
return loss_g
|
41 |
+
|
42 |
+
|
43 |
+
def discriminator_loss(gen_images, real_images, disc_net):
|
44 |
+
logits_fake = disc_net(gen_images.contiguous().detach())[-1]
|
45 |
+
loss_d_fake = adv_loss(logits_fake, target_is_real=False, for_discriminator=True)
|
46 |
+
logits_real = disc_net(real_images.contiguous().detach())[-1]
|
47 |
+
loss_d_real = adv_loss(logits_real, target_is_real=True, for_discriminator=True)
|
48 |
+
discriminator_loss = (loss_d_fake + loss_d_real) * 0.5
|
49 |
+
loss_d = adv_weight * discriminator_loss
|
50 |
+
return loss_d
|
scripts/utils.py
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Copyright (c) MONAI Consortium
|
2 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
3 |
+
# you may not use this file except in compliance with the License.
|
4 |
+
# You may obtain a copy of the License at
|
5 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
6 |
+
# Unless required by applicable law or agreed to in writing, software
|
7 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
8 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
9 |
+
# See the License for the specific language governing permissions and
|
10 |
+
|
11 |
+
import monai
|
12 |
+
import torch
|
13 |
+
|
14 |
+
|
15 |
+
def compute_scale_factor(autoencoder, train_loader, device):
|
16 |
+
with torch.no_grad():
|
17 |
+
check_data = monai.utils.first(train_loader)
|
18 |
+
z = autoencoder.encode_stage_2_inputs(check_data["image"].to(device))
|
19 |
+
scale_factor = 1 / torch.std(z)
|
20 |
+
return scale_factor.item()
|