progress.

This commit is contained in:
2026-02-25 17:41:12 -08:00
parent 0474ded71a
commit 0d2a941ecd
1062 changed files with 3291 additions and 3 deletions

View File

@@ -0,0 +1,293 @@
{
"7": {
"inputs": {
"clip_name": "qwen25vl.safetensors",
"type": "qwen_image",
"device": "default"
},
"class_type": "CLIPLoader",
"_meta": {
"title": "Load CLIP"
}
},
"8": {
"inputs": {
"vae_name": "qwen_image_vae.safetensors"
},
"class_type": "VAELoader",
"_meta": {
"title": "Load VAE"
}
},
"14": {
"inputs": {
"width": 1664,
"height": 928,
"batch_size": 1
},
"class_type": "EmptyLatentImage",
"_meta": {
"title": "Empty Latent Image"
}
},
"16": {
"inputs": {
"model_name": "qwen-image-2512-Q8_0.gguf",
"extra_model_name": "none",
"dequant_dtype": "default",
"patch_dtype": "default",
"patch_on_device": false,
"enable_fp16_accumulation": true,
"attention_override": "sageattn"
},
"class_type": "GGUFLoaderKJ",
"_meta": {
"title": "GGUFLoaderKJ"
}
},
"18": {
"inputs": {
"lora_name": "qwen/Qwen-Image-2512-Lightning-4steps-V1.0-fp32.safetensors",
"strength_model": 0.62,
"model": [
"24",
0
]
},
"class_type": "LoraLoaderModelOnly",
"_meta": {
"title": "Load LoRA"
}
},
"19": {
"inputs": {
"upscale_method": "bislerp",
"scale_by": 1.5,
"samples": [
"21",
0
]
},
"class_type": "LatentUpscaleBy",
"_meta": {
"title": "Upscale Latent By"
}
},
"20": {
"inputs": {
"seed": 1024708206384635,
"steps": 6,
"cfg": 1,
"sampler_name": "res_2s",
"scheduler": "bong_tangent",
"denoise": 0.45,
"model": [
"22",
0
],
"positive": [
"27",
0
],
"negative": [
"29",
0
],
"latent_image": [
"19",
0
]
},
"class_type": "KSampler",
"_meta": {
"title": "KSampler"
}
},
"21": {
"inputs": {
"seed": 1097142007423489,
"steps": 6,
"cfg": 1,
"sampler_name": "res_2s",
"scheduler": "bong_tangent",
"denoise": 1,
"model": [
"22",
0
],
"positive": [
"23",
0
],
"negative": [
"29",
0
],
"latent_image": [
"14",
0
]
},
"class_type": "KSampler",
"_meta": {
"title": "KSampler"
}
},
"22": {
"inputs": {
"max_shift": 1.15,
"base_shift": 0.5,
"width": 1024,
"height": 1024,
"model": [
"18",
0
]
},
"class_type": "ModelSamplingFlux",
"_meta": {
"title": "ModelSamplingFlux"
}
},
"23": {
"inputs": {
"threshold": 0.2,
"strength": 4,
"conditioning": [
"27",
0
]
},
"class_type": "ConditioningNoiseInjection",
"_meta": {
"title": "Conditioning Noise Injection"
}
},
"24": {
"inputs": {
"PowerLoraLoaderHeaderWidget": {
"type": "PowerLoraLoaderHeaderWidget"
},
"lora_1": {
"on": true,
"lora": "qwen/qwen-2512-kq5hoyos2/qwen-2512-kq5hoyos2.safetensors",
"strength": 0.45
},
"lora_2": {
"on": true,
"lora": "qwen/qwen-2512-wtp/qwen-2512-wtp.safetensors",
"strength": 0.34
},
"lora_3": {
"on": false,
"lora": "qwen/rebecca_guay_qwen_lora.safetensors",
"strength": 0.13
},
"lora_4": {
"on": true,
"lora": "qwen/qwen-2512-sylvain/qwen-2512-sylvain.safetensors",
"strength": 0.44
},
"lora_5": {
"on": false,
"lora": "qwen/Painterly01b_CE_QWEN_AIT3k.safetensors",
"strength": 0.5
},
" Add Lora": "",
"model": [
"16",
0
],
"clip": [
"7",
0
]
},
"class_type": "Power Lora Loader (rgthree)",
"_meta": {
"title": "Power Lora Loader (rgthree)"
}
},
"25": {
"inputs": {
"samples": [
"20",
0
],
"vae": [
"8",
0
]
},
"class_type": "VAEDecode",
"_meta": {
"title": "VAE Decode"
}
},
"26": {
"inputs": {
"filename_prefix": "auto_background/img",
"images": [
"25",
0
]
},
"class_type": "SaveImage",
"_meta": {
"title": "Save Image"
}
},
"27": {
"inputs": {
"text": "kq5hoyos, a river is running from the top right to the bottom left in a peaceful meadow environment.",
"clip": [
"24",
1
]
},
"class_type": "CLIPTextEncode",
"_meta": {
"title": "CLIP Text Encode (Prompt)"
}
},
"29": {
"inputs": {
"conditioning": [
"23",
0
]
},
"class_type": "ConditioningZeroOut",
"_meta": {
"title": "ConditioningZeroOut"
}
},
"30": {
"inputs": {
"samples": [
"21",
0
],
"vae": [
"8",
0
]
},
"class_type": "VAEDecode",
"_meta": {
"title": "VAE Decode"
}
},
"31": {
"inputs": {
"images": [
"30",
0
]
},
"class_type": "PreviewImage",
"_meta": {
"title": "Preview Image"
}
}
}