svjack commited on
Commit
577b892
verified
1 Parent(s): 5bc76c8

Upload 2 files

Browse files
workflows/loras/aniverse_v15Pruned.safetensors-keqing-lora-workflow.json ADDED
@@ -0,0 +1,471 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 10,
3
+ "last_link_id": 17,
4
+ "nodes": [
5
+ {
6
+ "id": 9,
7
+ "type": "SaveImage",
8
+ "pos": {
9
+ "0": 1451,
10
+ "1": 189
11
+ },
12
+ "size": {
13
+ "0": 210,
14
+ "1": 270
15
+ },
16
+ "flags": {},
17
+ "order": 7,
18
+ "mode": 0,
19
+ "inputs": [
20
+ {
21
+ "name": "images",
22
+ "type": "IMAGE",
23
+ "link": 9,
24
+ "label": "images"
25
+ }
26
+ ],
27
+ "outputs": [],
28
+ "properties": {
29
+ "Node name for S&R": "SaveImage"
30
+ },
31
+ "widgets_values": [
32
+ "ComfyUI"
33
+ ]
34
+ },
35
+ {
36
+ "id": 3,
37
+ "type": "KSampler",
38
+ "pos": {
39
+ "0": 943,
40
+ "1": 129
41
+ },
42
+ "size": {
43
+ "0": 315,
44
+ "1": 262
45
+ },
46
+ "flags": {},
47
+ "order": 5,
48
+ "mode": 0,
49
+ "inputs": [
50
+ {
51
+ "name": "model",
52
+ "type": "MODEL",
53
+ "link": 11,
54
+ "label": "model"
55
+ },
56
+ {
57
+ "name": "positive",
58
+ "type": "CONDITIONING",
59
+ "link": 4,
60
+ "label": "positive"
61
+ },
62
+ {
63
+ "name": "negative",
64
+ "type": "CONDITIONING",
65
+ "link": 6,
66
+ "label": "negative"
67
+ },
68
+ {
69
+ "name": "latent_image",
70
+ "type": "LATENT",
71
+ "link": 2,
72
+ "label": "latent_image"
73
+ }
74
+ ],
75
+ "outputs": [
76
+ {
77
+ "name": "LATENT",
78
+ "type": "LATENT",
79
+ "links": [
80
+ 7
81
+ ],
82
+ "slot_index": 0,
83
+ "label": "LATENT"
84
+ }
85
+ ],
86
+ "properties": {
87
+ "Node name for S&R": "KSampler"
88
+ },
89
+ "widgets_values": [
90
+ 581812963887651,
91
+ "randomize",
92
+ 20,
93
+ 8,
94
+ "euler",
95
+ "normal",
96
+ 1
97
+ ]
98
+ },
99
+ {
100
+ "id": 8,
101
+ "type": "VAEDecode",
102
+ "pos": {
103
+ "0": 1116,
104
+ "1": 486
105
+ },
106
+ "size": {
107
+ "0": 210,
108
+ "1": 46
109
+ },
110
+ "flags": {},
111
+ "order": 6,
112
+ "mode": 0,
113
+ "inputs": [
114
+ {
115
+ "name": "samples",
116
+ "type": "LATENT",
117
+ "link": 7,
118
+ "label": "samples"
119
+ },
120
+ {
121
+ "name": "vae",
122
+ "type": "VAE",
123
+ "link": 8,
124
+ "label": "vae"
125
+ }
126
+ ],
127
+ "outputs": [
128
+ {
129
+ "name": "IMAGE",
130
+ "type": "IMAGE",
131
+ "links": [
132
+ 9
133
+ ],
134
+ "slot_index": 0,
135
+ "label": "IMAGE"
136
+ }
137
+ ],
138
+ "properties": {
139
+ "Node name for S&R": "VAEDecode"
140
+ }
141
+ },
142
+ {
143
+ "id": 5,
144
+ "type": "EmptyLatentImage",
145
+ "pos": {
146
+ "0": 1011,
147
+ "1": 627
148
+ },
149
+ "size": {
150
+ "0": 315,
151
+ "1": 106
152
+ },
153
+ "flags": {},
154
+ "order": 0,
155
+ "mode": 0,
156
+ "inputs": [],
157
+ "outputs": [
158
+ {
159
+ "name": "LATENT",
160
+ "type": "LATENT",
161
+ "links": [
162
+ 2
163
+ ],
164
+ "slot_index": 0,
165
+ "label": "LATENT"
166
+ }
167
+ ],
168
+ "properties": {
169
+ "Node name for S&R": "EmptyLatentImage"
170
+ },
171
+ "widgets_values": [
172
+ 512,
173
+ 512,
174
+ 1
175
+ ]
176
+ },
177
+ {
178
+ "id": 7,
179
+ "type": "CLIPTextEncode",
180
+ "pos": {
181
+ "0": 427,
182
+ "1": 600
183
+ },
184
+ "size": {
185
+ "0": 425.27801513671875,
186
+ "1": 180.6060791015625
187
+ },
188
+ "flags": {},
189
+ "order": 4,
190
+ "mode": 0,
191
+ "inputs": [
192
+ {
193
+ "name": "clip",
194
+ "type": "CLIP",
195
+ "link": 17,
196
+ "label": "clip"
197
+ }
198
+ ],
199
+ "outputs": [
200
+ {
201
+ "name": "CONDITIONING",
202
+ "type": "CONDITIONING",
203
+ "links": [
204
+ 6
205
+ ],
206
+ "slot_index": 0,
207
+ "label": "CONDITIONING"
208
+ }
209
+ ],
210
+ "properties": {
211
+ "Node name for S&R": "CLIPTextEncode"
212
+ },
213
+ "widgets_values": [
214
+ "(worst quality, low quality, extra digits, loli, child, male:1.4)), bad_prompt,"
215
+ ]
216
+ },
217
+ {
218
+ "id": 10,
219
+ "type": "LoraLoader",
220
+ "pos": {
221
+ "0": 30,
222
+ "1": 586
223
+ },
224
+ "size": {
225
+ "0": 315,
226
+ "1": 126
227
+ },
228
+ "flags": {},
229
+ "order": 2,
230
+ "mode": 0,
231
+ "inputs": [
232
+ {
233
+ "name": "model",
234
+ "type": "MODEL",
235
+ "link": 10,
236
+ "label": "model"
237
+ },
238
+ {
239
+ "name": "clip",
240
+ "type": "CLIP",
241
+ "link": 15,
242
+ "label": "clip"
243
+ }
244
+ ],
245
+ "outputs": [
246
+ {
247
+ "name": "MODEL",
248
+ "type": "MODEL",
249
+ "links": [
250
+ 11
251
+ ],
252
+ "shape": 3,
253
+ "label": "MODEL",
254
+ "slot_index": 0
255
+ },
256
+ {
257
+ "name": "CLIP",
258
+ "type": "CLIP",
259
+ "links": [
260
+ 16,
261
+ 17
262
+ ],
263
+ "shape": 3,
264
+ "label": "CLIP",
265
+ "slot_index": 1
266
+ }
267
+ ],
268
+ "properties": {
269
+ "Node name for S&R": "LoraLoader"
270
+ },
271
+ "widgets_values": [
272
+ "keqing_lion_optimizer_dim64_loraModel_5e-3noise_token1_4-3-2023.safetensors",
273
+ 1,
274
+ 1
275
+ ]
276
+ },
277
+ {
278
+ "id": 6,
279
+ "type": "CLIPTextEncode",
280
+ "pos": {
281
+ "0": 451,
282
+ "1": 316
283
+ },
284
+ "size": {
285
+ "0": 422.84503173828125,
286
+ "1": 164.31304931640625
287
+ },
288
+ "flags": {},
289
+ "order": 3,
290
+ "mode": 0,
291
+ "inputs": [
292
+ {
293
+ "name": "clip",
294
+ "type": "CLIP",
295
+ "link": 16,
296
+ "label": "clip"
297
+ }
298
+ ],
299
+ "outputs": [
300
+ {
301
+ "name": "CONDITIONING",
302
+ "type": "CONDITIONING",
303
+ "links": [
304
+ 4
305
+ ],
306
+ "slot_index": 0,
307
+ "label": "CONDITIONING"
308
+ }
309
+ ],
310
+ "properties": {
311
+ "Node name for S&R": "CLIPTextEncode"
312
+ },
313
+ "widgets_values": [
314
+ "(Realistic painting style:0.9), masterpiece, best quality, absurdres, looking at viewer, solo, keqing (lantern rite) (genshin impact), official alternate costume, 1girl, keqing (genshin impact), phone, purple hair, solo, skirt, scarf, twintails, hair bun, cellphone, plaid scarf, sweater, purple sweater, long hair, hair ornament, white skirt, looking at viewer, holding phone, holding, cone hair bun, bag, plaid, blush, smartphone, long sleeves, bare shoulders, purple eyes, braid, bow, flower, shoulder bag, hair flower, breasts, frills, bangs, casual, double bun, handbag, hair bow, very long hair, closed mouth, outdoors, cable knit"
315
+ ]
316
+ },
317
+ {
318
+ "id": 4,
319
+ "type": "CheckpointLoaderSimple",
320
+ "pos": {
321
+ "0": 27,
322
+ "1": 154
323
+ },
324
+ "size": {
325
+ "0": 315,
326
+ "1": 98
327
+ },
328
+ "flags": {},
329
+ "order": 1,
330
+ "mode": 0,
331
+ "inputs": [],
332
+ "outputs": [
333
+ {
334
+ "name": "MODEL",
335
+ "type": "MODEL",
336
+ "links": [
337
+ 10
338
+ ],
339
+ "slot_index": 0,
340
+ "label": "MODEL"
341
+ },
342
+ {
343
+ "name": "CLIP",
344
+ "type": "CLIP",
345
+ "links": [
346
+ 15
347
+ ],
348
+ "slot_index": 1,
349
+ "label": "CLIP"
350
+ },
351
+ {
352
+ "name": "VAE",
353
+ "type": "VAE",
354
+ "links": [
355
+ 8
356
+ ],
357
+ "slot_index": 2,
358
+ "label": "VAE"
359
+ }
360
+ ],
361
+ "properties": {
362
+ "Node name for S&R": "CheckpointLoaderSimple"
363
+ },
364
+ "widgets_values": [
365
+ "aniverse_v15Pruned.safetensors"
366
+ ]
367
+ }
368
+ ],
369
+ "links": [
370
+ [
371
+ 2,
372
+ 5,
373
+ 0,
374
+ 3,
375
+ 3,
376
+ "LATENT"
377
+ ],
378
+ [
379
+ 4,
380
+ 6,
381
+ 0,
382
+ 3,
383
+ 1,
384
+ "CONDITIONING"
385
+ ],
386
+ [
387
+ 6,
388
+ 7,
389
+ 0,
390
+ 3,
391
+ 2,
392
+ "CONDITIONING"
393
+ ],
394
+ [
395
+ 7,
396
+ 3,
397
+ 0,
398
+ 8,
399
+ 0,
400
+ "LATENT"
401
+ ],
402
+ [
403
+ 8,
404
+ 4,
405
+ 2,
406
+ 8,
407
+ 1,
408
+ "VAE"
409
+ ],
410
+ [
411
+ 9,
412
+ 8,
413
+ 0,
414
+ 9,
415
+ 0,
416
+ "IMAGE"
417
+ ],
418
+ [
419
+ 10,
420
+ 4,
421
+ 0,
422
+ 10,
423
+ 0,
424
+ "MODEL"
425
+ ],
426
+ [
427
+ 11,
428
+ 10,
429
+ 0,
430
+ 3,
431
+ 0,
432
+ "MODEL"
433
+ ],
434
+ [
435
+ 15,
436
+ 4,
437
+ 1,
438
+ 10,
439
+ 1,
440
+ "CLIP"
441
+ ],
442
+ [
443
+ 16,
444
+ 10,
445
+ 1,
446
+ 6,
447
+ 0,
448
+ "CLIP"
449
+ ],
450
+ [
451
+ 17,
452
+ 10,
453
+ 1,
454
+ 7,
455
+ 0,
456
+ "CLIP"
457
+ ]
458
+ ],
459
+ "groups": [],
460
+ "config": {},
461
+ "extra": {
462
+ "ds": {
463
+ "scale": 0.6830134553650705,
464
+ "offset": {
465
+ "0": 450.1739807128906,
466
+ "1": -9.992695808410645
467
+ }
468
+ }
469
+ },
470
+ "version": 0.4
471
+ }
workflows/loras/niji-鍔ㄦ极浜屾鍏僟3.0.safetensors-keqing-lora-workflow.json ADDED
@@ -0,0 +1,471 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "last_node_id": 10,
3
+ "last_link_id": 17,
4
+ "nodes": [
5
+ {
6
+ "id": 9,
7
+ "type": "SaveImage",
8
+ "pos": {
9
+ "0": 1451,
10
+ "1": 189
11
+ },
12
+ "size": {
13
+ "0": 210,
14
+ "1": 270
15
+ },
16
+ "flags": {},
17
+ "order": 7,
18
+ "mode": 0,
19
+ "inputs": [
20
+ {
21
+ "name": "images",
22
+ "type": "IMAGE",
23
+ "link": 9,
24
+ "label": "images"
25
+ }
26
+ ],
27
+ "outputs": [],
28
+ "properties": {
29
+ "Node name for S&R": "SaveImage"
30
+ },
31
+ "widgets_values": [
32
+ "ComfyUI"
33
+ ]
34
+ },
35
+ {
36
+ "id": 3,
37
+ "type": "KSampler",
38
+ "pos": {
39
+ "0": 943,
40
+ "1": 129
41
+ },
42
+ "size": {
43
+ "0": 315,
44
+ "1": 262
45
+ },
46
+ "flags": {},
47
+ "order": 5,
48
+ "mode": 0,
49
+ "inputs": [
50
+ {
51
+ "name": "model",
52
+ "type": "MODEL",
53
+ "link": 11,
54
+ "label": "model"
55
+ },
56
+ {
57
+ "name": "positive",
58
+ "type": "CONDITIONING",
59
+ "link": 4,
60
+ "label": "positive"
61
+ },
62
+ {
63
+ "name": "negative",
64
+ "type": "CONDITIONING",
65
+ "link": 6,
66
+ "label": "negative"
67
+ },
68
+ {
69
+ "name": "latent_image",
70
+ "type": "LATENT",
71
+ "link": 2,
72
+ "label": "latent_image"
73
+ }
74
+ ],
75
+ "outputs": [
76
+ {
77
+ "name": "LATENT",
78
+ "type": "LATENT",
79
+ "links": [
80
+ 7
81
+ ],
82
+ "slot_index": 0,
83
+ "label": "LATENT"
84
+ }
85
+ ],
86
+ "properties": {
87
+ "Node name for S&R": "KSampler"
88
+ },
89
+ "widgets_values": [
90
+ 1123713238173417,
91
+ "randomize",
92
+ 20,
93
+ 8,
94
+ "euler",
95
+ "normal",
96
+ 1
97
+ ]
98
+ },
99
+ {
100
+ "id": 8,
101
+ "type": "VAEDecode",
102
+ "pos": {
103
+ "0": 1116,
104
+ "1": 486
105
+ },
106
+ "size": {
107
+ "0": 210,
108
+ "1": 46
109
+ },
110
+ "flags": {},
111
+ "order": 6,
112
+ "mode": 0,
113
+ "inputs": [
114
+ {
115
+ "name": "samples",
116
+ "type": "LATENT",
117
+ "link": 7,
118
+ "label": "samples"
119
+ },
120
+ {
121
+ "name": "vae",
122
+ "type": "VAE",
123
+ "link": 8,
124
+ "label": "vae"
125
+ }
126
+ ],
127
+ "outputs": [
128
+ {
129
+ "name": "IMAGE",
130
+ "type": "IMAGE",
131
+ "links": [
132
+ 9
133
+ ],
134
+ "slot_index": 0,
135
+ "label": "IMAGE"
136
+ }
137
+ ],
138
+ "properties": {
139
+ "Node name for S&R": "VAEDecode"
140
+ }
141
+ },
142
+ {
143
+ "id": 5,
144
+ "type": "EmptyLatentImage",
145
+ "pos": {
146
+ "0": 1011,
147
+ "1": 627
148
+ },
149
+ "size": {
150
+ "0": 315,
151
+ "1": 106
152
+ },
153
+ "flags": {},
154
+ "order": 0,
155
+ "mode": 0,
156
+ "inputs": [],
157
+ "outputs": [
158
+ {
159
+ "name": "LATENT",
160
+ "type": "LATENT",
161
+ "links": [
162
+ 2
163
+ ],
164
+ "slot_index": 0,
165
+ "label": "LATENT"
166
+ }
167
+ ],
168
+ "properties": {
169
+ "Node name for S&R": "EmptyLatentImage"
170
+ },
171
+ "widgets_values": [
172
+ 512,
173
+ 512,
174
+ 1
175
+ ]
176
+ },
177
+ {
178
+ "id": 7,
179
+ "type": "CLIPTextEncode",
180
+ "pos": {
181
+ "0": 427,
182
+ "1": 600
183
+ },
184
+ "size": {
185
+ "0": 425.27801513671875,
186
+ "1": 180.6060791015625
187
+ },
188
+ "flags": {},
189
+ "order": 4,
190
+ "mode": 0,
191
+ "inputs": [
192
+ {
193
+ "name": "clip",
194
+ "type": "CLIP",
195
+ "link": 17,
196
+ "label": "clip"
197
+ }
198
+ ],
199
+ "outputs": [
200
+ {
201
+ "name": "CONDITIONING",
202
+ "type": "CONDITIONING",
203
+ "links": [
204
+ 6
205
+ ],
206
+ "slot_index": 0,
207
+ "label": "CONDITIONING"
208
+ }
209
+ ],
210
+ "properties": {
211
+ "Node name for S&R": "CLIPTextEncode"
212
+ },
213
+ "widgets_values": [
214
+ "(worst quality, low quality, extra digits, loli, child, male:1.4)), bad_prompt,"
215
+ ]
216
+ },
217
+ {
218
+ "id": 10,
219
+ "type": "LoraLoader",
220
+ "pos": {
221
+ "0": 30,
222
+ "1": 586
223
+ },
224
+ "size": {
225
+ "0": 315,
226
+ "1": 126
227
+ },
228
+ "flags": {},
229
+ "order": 2,
230
+ "mode": 0,
231
+ "inputs": [
232
+ {
233
+ "name": "model",
234
+ "type": "MODEL",
235
+ "link": 10,
236
+ "label": "model"
237
+ },
238
+ {
239
+ "name": "clip",
240
+ "type": "CLIP",
241
+ "link": 15,
242
+ "label": "clip"
243
+ }
244
+ ],
245
+ "outputs": [
246
+ {
247
+ "name": "MODEL",
248
+ "type": "MODEL",
249
+ "links": [
250
+ 11
251
+ ],
252
+ "shape": 3,
253
+ "label": "MODEL",
254
+ "slot_index": 0
255
+ },
256
+ {
257
+ "name": "CLIP",
258
+ "type": "CLIP",
259
+ "links": [
260
+ 16,
261
+ 17
262
+ ],
263
+ "shape": 3,
264
+ "label": "CLIP",
265
+ "slot_index": 1
266
+ }
267
+ ],
268
+ "properties": {
269
+ "Node name for S&R": "LoraLoader"
270
+ },
271
+ "widgets_values": [
272
+ "keqing_lion_optimizer_dim64_loraModel_5e-3noise_token1_4-3-2023.safetensors",
273
+ 1,
274
+ 1
275
+ ]
276
+ },
277
+ {
278
+ "id": 6,
279
+ "type": "CLIPTextEncode",
280
+ "pos": {
281
+ "0": 451,
282
+ "1": 316
283
+ },
284
+ "size": {
285
+ "0": 422.84503173828125,
286
+ "1": 164.31304931640625
287
+ },
288
+ "flags": {},
289
+ "order": 3,
290
+ "mode": 0,
291
+ "inputs": [
292
+ {
293
+ "name": "clip",
294
+ "type": "CLIP",
295
+ "link": 16,
296
+ "label": "clip"
297
+ }
298
+ ],
299
+ "outputs": [
300
+ {
301
+ "name": "CONDITIONING",
302
+ "type": "CONDITIONING",
303
+ "links": [
304
+ 4
305
+ ],
306
+ "slot_index": 0,
307
+ "label": "CONDITIONING"
308
+ }
309
+ ],
310
+ "properties": {
311
+ "Node name for S&R": "CLIPTextEncode"
312
+ },
313
+ "widgets_values": [
314
+ "(Realistic painting style:0.9), masterpiece, best quality, absurdres, looking at viewer, solo, keqing (lantern rite) (genshin impact), official alternate costume, 1girl, keqing (genshin impact), phone, purple hair, solo, skirt, scarf, twintails, hair bun, cellphone, plaid scarf, sweater, purple sweater, long hair, hair ornament, white skirt, looking at viewer, holding phone, holding, cone hair bun, bag, plaid, blush, smartphone, long sleeves, bare shoulders, purple eyes, braid, bow, flower, shoulder bag, hair flower, breasts, frills, bangs, casual, double bun, handbag, hair bow, very long hair, closed mouth, outdoors, cable knit"
315
+ ]
316
+ },
317
+ {
318
+ "id": 4,
319
+ "type": "CheckpointLoaderSimple",
320
+ "pos": {
321
+ "0": 27,
322
+ "1": 154
323
+ },
324
+ "size": {
325
+ "0": 315,
326
+ "1": 98
327
+ },
328
+ "flags": {},
329
+ "order": 1,
330
+ "mode": 0,
331
+ "inputs": [],
332
+ "outputs": [
333
+ {
334
+ "name": "MODEL",
335
+ "type": "MODEL",
336
+ "links": [
337
+ 10
338
+ ],
339
+ "slot_index": 0,
340
+ "label": "MODEL"
341
+ },
342
+ {
343
+ "name": "CLIP",
344
+ "type": "CLIP",
345
+ "links": [
346
+ 15
347
+ ],
348
+ "slot_index": 1,
349
+ "label": "CLIP"
350
+ },
351
+ {
352
+ "name": "VAE",
353
+ "type": "VAE",
354
+ "links": [
355
+ 8
356
+ ],
357
+ "slot_index": 2,
358
+ "label": "VAE"
359
+ }
360
+ ],
361
+ "properties": {
362
+ "Node name for S&R": "CheckpointLoaderSimple"
363
+ },
364
+ "widgets_values": [
365
+ "niji-鍔ㄦ极浜屾鍏僟3.0.safetensors"
366
+ ]
367
+ }
368
+ ],
369
+ "links": [
370
+ [
371
+ 2,
372
+ 5,
373
+ 0,
374
+ 3,
375
+ 3,
376
+ "LATENT"
377
+ ],
378
+ [
379
+ 4,
380
+ 6,
381
+ 0,
382
+ 3,
383
+ 1,
384
+ "CONDITIONING"
385
+ ],
386
+ [
387
+ 6,
388
+ 7,
389
+ 0,
390
+ 3,
391
+ 2,
392
+ "CONDITIONING"
393
+ ],
394
+ [
395
+ 7,
396
+ 3,
397
+ 0,
398
+ 8,
399
+ 0,
400
+ "LATENT"
401
+ ],
402
+ [
403
+ 8,
404
+ 4,
405
+ 2,
406
+ 8,
407
+ 1,
408
+ "VAE"
409
+ ],
410
+ [
411
+ 9,
412
+ 8,
413
+ 0,
414
+ 9,
415
+ 0,
416
+ "IMAGE"
417
+ ],
418
+ [
419
+ 10,
420
+ 4,
421
+ 0,
422
+ 10,
423
+ 0,
424
+ "MODEL"
425
+ ],
426
+ [
427
+ 11,
428
+ 10,
429
+ 0,
430
+ 3,
431
+ 0,
432
+ "MODEL"
433
+ ],
434
+ [
435
+ 15,
436
+ 4,
437
+ 1,
438
+ 10,
439
+ 1,
440
+ "CLIP"
441
+ ],
442
+ [
443
+ 16,
444
+ 10,
445
+ 1,
446
+ 6,
447
+ 0,
448
+ "CLIP"
449
+ ],
450
+ [
451
+ 17,
452
+ 10,
453
+ 1,
454
+ 7,
455
+ 0,
456
+ "CLIP"
457
+ ]
458
+ ],
459
+ "groups": [],
460
+ "config": {},
461
+ "extra": {
462
+ "ds": {
463
+ "scale": 0.6830134553650705,
464
+ "offset": {
465
+ "0": 450.1739807128906,
466
+ "1": -9.992695808410645
467
+ }
468
+ }
469
+ },
470
+ "version": 0.4
471
+ }