Flux / ComfyUI_workflow /FluxSchnell_Lora_Text-to-img.json
phamhungd's picture
Upload 8 files
e7570f1 verified
raw
history blame
7.83 kB
{
"last_node_id": 81,
"last_link_id": 137,
"nodes": [
{
"id": 72,
"type": "easy bookmark",
"pos": [
146,
332
],
"size": {
"0": 210,
"1": 82
},
"flags": {
"collapsed": true
},
"order": 0,
"mode": 0,
"title": "🔖 1",
"properties": {},
"widgets_values": [
"1",
1.2
]
},
{
"id": 69,
"type": "CLIPTextEncode",
"pos": [
534,
531
],
"size": {
"0": 400,
"1": 200
},
"flags": {
"collapsed": false
},
"order": 5,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 134
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"links": [
112,
135
],
"slot_index": 0,
"shape": 3
}
],
"properties": {
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"ghibli style, My Neightbor Totoro, Silly boy looking up at the blue sky, wearing a (white shirt and glasses:1.2)"
]
},
{
"id": 81,
"type": "LoraLoaderModelOnly",
"pos": [
535,
397
],
"size": {
"0": 399.364501953125,
"1": 82
},
"flags": {},
"order": 4,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 137
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
136
],
"slot_index": 0,
"shape": 3
}
],
"properties": {
"Node name for S&R": "LoraLoaderModelOnly"
},
"widgets_values": [
"Lora/MyNeighborTotoro_flux/MyNeighborTotoro_flux.safetensors",
1
]
},
{
"id": 53,
"type": "EmptyLatentImage",
"pos": [
539,
783
],
"size": {
"0": 393.8836669921875,
"1": 109.99248504638672
},
"flags": {},
"order": 1,
"mode": 0,
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
85
],
"shape": 3
}
],
"properties": {
"Node name for S&R": "EmptyLatentImage"
},
"widgets_values": [
768,
1024,
1
]
},
{
"id": 68,
"type": "Note",
"pos": [
200,
550
],
"size": {
"0": 312.59844970703125,
"1": 99.95023345947266
},
"flags": {},
"order": 2,
"mode": 0,
"properties": {
"text": ""
},
"widgets_values": [
"ComfyUI Colab:\n- Creator: stablediffusion.vn\n- Website: sdvn.me | comfy.vn | fluxai.vn\n- Document: bit.ly/sdvn-lib\n- Skills training: hungdiffusion.com"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 8,
"type": "VAEDecode",
"pos": [
1319,
397
],
"size": {
"0": 210,
"1": 46
},
"flags": {},
"order": 7,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 66
},
{
"name": "vae",
"type": "VAE",
"link": 88
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
9
],
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "VAEDecode"
}
},
{
"id": 9,
"type": "SaveImage",
"pos": [
1563,
394
],
"size": {
"0": 356.08502197265625,
"1": 499.8143310546875
},
"flags": {},
"order": 8,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 9
}
],
"properties": {
"Node name for S&R": "SaveImage"
},
"widgets_values": [
"ComfyUI"
]
},
{
"id": 52,
"type": "CheckpointLoaderSimple",
"pos": [
198,
402
],
"size": {
"0": 315,
"1": 98
},
"flags": {},
"order": 3,
"mode": 0,
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
137
],
"slot_index": 0,
"shape": 3
},
{
"name": "CLIP",
"type": "CLIP",
"links": [
134
],
"slot_index": 1,
"shape": 3
},
{
"name": "VAE",
"type": "VAE",
"links": [
88
],
"slot_index": 2,
"shape": 3
}
],
"properties": {
"Node name for S&R": "CheckpointLoaderSimple"
},
"widgets_values": [
"Flux_schnell_v1-fp8.safetensors"
]
},
{
"id": 44,
"type": "KSampler",
"pos": [
969,
399
],
"size": {
"0": 318.3961181640625,
"1": 487.5167236328125
},
"flags": {},
"order": 6,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 136
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 112
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 135
},
{
"name": "latent_image",
"type": "LATENT",
"link": 85,
"slot_index": 3
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
66
],
"slot_index": 0,
"shape": 3
}
],
"properties": {
"Node name for S&R": "KSampler"
},
"widgets_values": [
714762845096392,
"randomize",
4,
1,
"euler",
"simple",
1
]
}
],
"links": [
[
9,
8,
0,
9,
0,
"IMAGE"
],
[
66,
44,
0,
8,
0,
"LATENT"
],
[
75,
10,
0,
8,
1,
"VAE"
],
[
76,
12,
0,
44,
0,
"MODEL"
],
[
77,
11,
0,
48,
0,
"CLIP"
],
[
78,
10,
0,
30,
1,
"VAE"
],
[
79,
12,
0,
51,
0,
"MODEL"
],
[
80,
11,
0,
51,
1,
"CLIP"
],
[
81,
10,
0,
51,
2,
"VAE"
],
[
85,
53,
0,
44,
3,
"LATENT"
],
[
88,
52,
2,
8,
1,
"VAE"
],
[
112,
69,
0,
44,
1,
"CONDITIONING"
],
[
134,
52,
1,
69,
0,
"CLIP"
],
[
135,
69,
0,
44,
2,
"CONDITIONING"
],
[
136,
81,
0,
44,
0,
"MODEL"
],
[
137,
52,
0,
81,
0,
"MODEL"
]
],
"groups": [],
"config": {},
"extra": {
"ds": {
"scale": 0.9090909090909095,
"offset": {
"0": -68.71284484863281,
"1": -148.65814208984375
}
}
},
"version": 0.4
}