{ "id": 1447839, "modelId": 1283256, "name": "v1.0", "createdAt": "2025-02-22T16:24:00.481Z", "updatedAt": "2025-02-22T16:31:00.119Z", "status": "Published", "publishedAt": "2025-02-22T16:31:00.047Z", "trainedWords": [ "chun-li" ], "trainingStatus": null, "trainingDetails": null, "baseModel": "Flux.1 D", "baseModelType": null, "earlyAccessEndsAt": null, "earlyAccessConfig": null, "description": null, "uploadType": "Created", "usageControl": "Download", "air": "urn:air:flux1:lora:civitai:1283256@1447839", "stats": { "downloadCount": 192, "ratingCount": 0, "rating": 0, "thumbsUpCount": 35 }, "model": { "name": "Realistic Chun-Li (Street Fighter) (including multiple outfits) [FLUX]", "type": "LORA", "nsfw": false, "poi": false }, "files": [ { "id": 1349025, "sizeKB": 32875.21484375, "name": "Chun-Li_rank8_bf16-step04000.safetensors", "type": "Model", "pickleScanResult": "Success", "pickleScanMessage": "No Pickle imports", "virusScanResult": "Success", "virusScanMessage": null, "scannedAt": "2025-02-22T16:30:30.274Z", "metadata": { "format": "SafeTensor", "size": null, "fp": null }, "hashes": { "AutoV1": "6A15FD1F", "AutoV2": "3BFCBECAA7", "SHA256": "3BFCBECAA7A8AE169BE0472F81ECB41F3715673312D29C44B102B9EF205A89CD", "CRC32": "B92CD5ED", "BLAKE3": "85AE3280F5FE73DC22FF4874CAB35360F4CECC7229C7B5B2B1896EA0670AFF4F", "AutoV3": "757ADDE3523D" }, "primary": true, "downloadUrl": "https://civitai.com/api/download/models/1447839" } ], "images": [ { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/dc786402-24e5-48a5-956d-d211738520b9/width=832/59431684.jpeg", "nsfwLevel": 1, "width": 832, "height": 1152, "hash": "U7Bfq*b_Hq?HiH%MRpNG0fkDyZNf2uEM9wx]", "type": "image", "metadata": { "hash": "U7Bfq*b_Hq?HiH%MRpNG0fkDyZNf2uEM9wx]", "size": 1148068, "width": 832, "height": 1152 }, "minor": false, "poi": false, "meta": { "seed": 673902776627021, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of of a woman cosplaying as chun-li, a muscular woman with her hair up in double buns with bun covers. she's wearing her blue qipao with a high neck, short puffy sleaves, white sash, dark grey leggings, spiked bracelets, and white knee-high combat boots. She's standing in front of a chinese temple at night, waving and smiling.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_054934_511519730376074\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 673902776627021}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [673902776627021]}, \"34\": {\"inputs\": {\"resolution\": \"832x1152 (0.72)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [848.2301635742188, 200.2738494873047], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [511519730376074, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1152 (0.72)\", 1, 0, 0]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [673902776627021, null, null, null]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of of a woman cosplaying as chun-li, a muscular woman with her hair up in double buns with bun covers. she's wearing her blue qipao with a high neck, short puffy sleaves, white sash, dark grey leggings, spiked bracelets, and white knee-high combat boots. She's standing in front of a chinese temple at night, waving and smiling.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [-180.82670983867317, -154.65756925869454]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of of a woman cosplaying as chun-li, a muscular woman with her hair up in double buns with bun covers. she's wearing her blue qipao with a high neck, short puffy sleaves, white sash, dark grey leggings, spiked bracelets, and white knee-high combat boots. She's standing in front of a chinese temple at night, waving and smiling.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9dc7ddb8-3930-489d-8d46-c46f620649d2/width=832/59435798.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "UDE2IVKjzoxC}@Fx=zt54:OFx^ay58RlJ8W=", "type": "image", "metadata": { "hash": "UDE2IVKjzoxC}@Fx=zt54:OFx^ay58RlJ8W=", "size": 1314460, "width": 832, "height": 1216 }, "minor": false, "poi": false, "meta": { "seed": 255416581629852, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of chun-li, a mature woman with red mascara, large breasts, slim waist, with her hair up in double buns. she's wearing her pale white qipao with blue flowing designs, blue sleeves, a high neck, gold trim, blue leggings, gold and black bracelets, and black slipon kung fu shoes. \\nShe's walking the streets of Chinatown at night. Buildings with neon signs are in the background. A vertical neon signs says \\\"Chun-Li\\\"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_081819_193341951289995\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 255416581629852}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [255416581629852]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0669555664062, 220.04977416992188], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [193341951289995, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [255416581629852, null, null, null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of chun-li, a mature woman with red mascara, large breasts, slim waist, with her hair up in double buns. she's wearing her pale white qipao with blue flowing designs, blue sleeves, a high neck, gold trim, blue leggings, gold and black bracelets, and black slipon kung fu shoes. \\nShe's walking the streets of Chinatown at night. Buildings with neon signs are in the background. A vertical neon signs says \\\"Chun-Li\\\"\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [-4.006548379484157, -130.22871893392812]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of chun-li, a mature woman with red mascara, large breasts, slim waist, with her hair up in double buns. she's wearing her pale white qipao with blue flowing designs, blue sleeves, a high neck, gold trim, blue leggings, gold and black bracelets, and black slipon kung fu shoes. \nShe's walking the streets of Chinatown at night. Buildings with neon signs are in the background. A vertical neon signs says \"Chun-Li\"", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/0446cbb4-65cd-48e9-a42f-083763b5db1e/width=832/59431741.jpeg", "nsfwLevel": 2, "width": 832, "height": 1152, "hash": "U3A9c]^m00tl-ttnI:R-0[E-]~IV0VI]%JV@", "type": "image", "metadata": { "hash": "U3A9c]^m00tl-ttnI:R-0[E-]~IV0VI]%JV@", "size": 1128441, "width": 832, "height": 1152 }, "minor": false, "poi": false, "meta": { "seed": 1056047061739334, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of of a woman cosplaying as chun-li, a muscular woman with large breasts, her long hair flowing down her back. she's wearing her sleevless black halter battle dress with deep cleavage, gold trim and design, a gold girdle with a red rope fastener, black pelvic curtain, and white slip on shoes. Her muscled legs are on display.\\nShe's standing in front of a chinese temple at night, waving and smiling.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_060801_155910549717089\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 1056047061739334}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [1056047061739334]}, \"34\": {\"inputs\": {\"resolution\": \"832x1152 (0.72)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [848.2301635742188, 200.2738494873047], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [155910549717089, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1152 (0.72)\", 1, 0, 0]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [1056047061739334, null, null, null]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of of a woman cosplaying as chun-li, a muscular woman with large breasts, her long hair flowing down her back. she's wearing her sleevless black halter battle dress with deep cleavage, gold trim and design, a gold girdle with a red rope fastener, black pelvic curtain, and white slip on shoes. Her muscled legs are on display.\\nShe's standing in front of a chinese temple at night, waving and smiling.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [-88.92655561915785, -22.042218517856874]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of of a woman cosplaying as chun-li, a muscular woman with large breasts, her long hair flowing down her back. she's wearing her sleevless black halter battle dress with deep cleavage, gold trim and design, a gold girdle with a red rope fastener, black pelvic curtain, and white slip on shoes. Her muscled legs are on display.\nShe's standing in front of a chinese temple at night, waving and smiling.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d1c3364a-b9da-49bf-a1c7-73d227d96979/width=832/59431742.jpeg", "nsfwLevel": 2, "width": 832, "height": 1152, "hash": "U4Avh61i4TIUIT56D+xa00=x.TtQ00^P?cIp", "type": "image", "metadata": { "hash": "U4Avh61i4TIUIT56D+xa00=x.TtQ00^P?cIp", "size": 1190854, "width": 832, "height": 1152 }, "minor": false, "poi": false, "meta": { "seed": 965874288075643, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of of a woman cosplaying as chun-li, a muscular woman large breasts, slim waist, with her hair up in double buns wth gold ribbons. she's wearing her light blue sleeveless top with a high neck and gold designs, skin tight dark blue leggings with gold stripes down the sides that show off her leg muscles, thick black bracelets, and blue tennis shoes with gold trim. She's standing in front of a chinese temple at night, waving and smiling.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_061912_666345988365586\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 965874288075643}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [965874288075643]}, \"34\": {\"inputs\": {\"resolution\": \"832x1152 (0.72)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [848.2301635742188, 200.2738494873047], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [666345988365586, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1152 (0.72)\", 1, 0, 0]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [965874288075643, null, null, null]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of of a woman cosplaying as chun-li, a muscular woman large breasts, slim waist, with her hair up in double buns wth gold ribbons. she's wearing her light blue sleeveless top with a high neck and gold designs, skin tight dark blue leggings with gold stripes down the sides that show off her leg muscles, thick black bracelets, and blue tennis shoes with gold trim. She's standing in front of a chinese temple at night, waving and smiling.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [-86.60011065357213, 26.81584823469545]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of of a woman cosplaying as chun-li, a muscular woman large breasts, slim waist, with her hair up in double buns wth gold ribbons. she's wearing her light blue sleeveless top with a high neck and gold designs, skin tight dark blue leggings with gold stripes down the sides that show off her leg muscles, thick black bracelets, and blue tennis shoes with gold trim. She's standing in front of a chinese temple at night, waving and smiling.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/7ffb777c-2721-438e-a7c1-62308ce05eb1/width=832/59431743.jpeg", "nsfwLevel": 1, "width": 832, "height": 1152, "hash": "U6CF@vs?00^GD4-T?d4?00M}_NtQAaE2j:?G", "type": "image", "metadata": { "hash": "U6CF@vs?00^GD4-T?d4?00M}_NtQAaE2j:?G", "size": 1143266, "width": 832, "height": 1152 }, "minor": false, "poi": false, "meta": { "seed": 1061321669314409, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of of a woman cosplaying as chun-li, a muscular woman with her hair up in double buns. she's wearing her pale white qipao with blue designs, short sleeves, a high neck with a cleavage cutout, short sleaves, blue leggings, gold and black bracelets, and black slipon shoes. She's standing in front of a chinese temple at night, waving and smiling.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_055540_279048681207985\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 1061321669314409}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [1061321669314409]}, \"34\": {\"inputs\": {\"resolution\": \"832x1152 (0.72)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [848.2301635742188, 200.2738494873047], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [279048681207985, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1152 (0.72)\", 1, 0, 0]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [1061321669314409, null, null, null]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of of a woman cosplaying as chun-li, a muscular woman with her hair up in double buns. she's wearing her pale white qipao with blue designs, short sleeves, a high neck with a cleavage cutout, short sleaves, blue leggings, gold and black bracelets, and black slipon shoes. She's standing in front of a chinese temple at night, waving and smiling.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [56.485161018181195, 3.550380146774862]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of of a woman cosplaying as chun-li, a muscular woman with her hair up in double buns. she's wearing her pale white qipao with blue designs, short sleeves, a high neck with a cleavage cutout, short sleaves, blue leggings, gold and black bracelets, and black slipon shoes. She's standing in front of a chinese temple at night, waving and smiling.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/6a079926-5649-4119-b31a-de000f476bb6/width=832/59431765.jpeg", "nsfwLevel": 4, "width": 832, "height": 1152, "hash": "UgLWU:-o~VxZ~pW=xaWWtRj[f6s:t8azt6of", "type": "image", "metadata": { "hash": "UgLWU:-o~VxZ~pW=xaWWtRj[f6s:t8azt6of", "size": 955725, "width": 832, "height": 1152 }, "minor": false, "poi": false, "meta": { "seed": 655368449854588, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of of a woman cosplaying as chun-li, a muscular woman large breasts, slim waist, with her hair up in double buns wth gold ribbons. she's wearing a dark blue bikini with gold designs. She's standing on a beach at sunset. she's smiling at the viewer.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_062129_27402610190614\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 655368449854588}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [655368449854588]}, \"34\": {\"inputs\": {\"resolution\": \"832x1152 (0.72)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [848.2301635742188, 200.2738494873047], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [27402610190614, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1152 (0.72)\", 1, 0, 0]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [655368449854588, null, null, null]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of of a woman cosplaying as chun-li, a muscular woman large breasts, slim waist, with her hair up in double buns wth gold ribbons. she's wearing a dark blue bikini with gold designs. She's standing on a beach at sunset. she's smiling at the viewer.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [-111.029273830013, -34.838670947802925]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of of a woman cosplaying as chun-li, a muscular woman large breasts, slim waist, with her hair up in double buns wth gold ribbons. she's wearing a dark blue bikini with gold designs. She's standing on a beach at sunset. she's smiling at the viewer.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/03deb4a9-dc8b-41ea-9c46-4e29832bbe4b/width=1216/59432066.jpeg", "nsfwLevel": 2, "width": 1216, "height": 832, "hash": "UCCG_fU[h{ys4TbbTLMd~TpJafsk9FxabcXU", "type": "image", "metadata": { "hash": "UCCG_fU[h{ys4TbbTLMd~TpJafsk9FxabcXU", "size": 1348340, "width": 1216, "height": 832 }, "minor": false, "poi": false, "meta": { "seed": 285220768725359, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of chun-li, a muscular, mature woman with large breasts, slim waist, with her hair up in double buns with bun covers. she's wearing her blue qipao with a high neck, short puffy sleaves, white sash, dark grey leggings, black spiked bracelets, and white combat boots. She's standing in front of a chinese temple, doing a high side kick. Her foot is moving so fast that the air around her foot glows blue around her boot. The explosion of power surrounds her boot entirely.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_065412_1120409683952681\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 285220768725359}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [285220768725359]}, \"34\": {\"inputs\": {\"resolution\": \"1216x832 (1.46)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"1216x832 (1.46)\", 1, 0, 0]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0669555664062, 220.04977416992188], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [1120409683952681, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [285220768725359, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of chun-li, a muscular, mature woman with large breasts, slim waist, with her hair up in double buns with bun covers. she's wearing her blue qipao with a high neck, short puffy sleaves, white sash, dark grey leggings, black spiked bracelets, and white combat boots. She's standing in front of a chinese temple, doing a high side kick. Her foot is moving so fast that the air around her foot glows blue around her boot. The explosion of power surrounds her boot entirely.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [56.484881448596575, -91.84008719372973]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of chun-li, a muscular, mature woman with large breasts, slim waist, with her hair up in double buns with bun covers. she's wearing her blue qipao with a high neck, short puffy sleaves, white sash, dark grey leggings, black spiked bracelets, and white combat boots. She's standing in front of a chinese temple, doing a high side kick. Her foot is moving so fast that the air around her foot glows blue around her boot. The explosion of power surrounds her boot entirely.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/a3e14bb5-763e-473d-b5dc-66b7cb907b95/width=832/59432216.jpeg", "nsfwLevel": 4, "width": 832, "height": 1216, "hash": "UFE3#QS~0zt,ujIov}VXLz%K%MOZ9Zn#s:t7", "type": "image", "metadata": { "hash": "UFE3#QS~0zt,ujIov}VXLz%K%MOZ9Zn#s:t7", "size": 1130177, "width": 832, "height": 1216 }, "minor": false, "poi": false, "meta": { "seed": 181891228213810, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of a woman cosplaying as chun-li, a beautiful woman with large breasts, slim waist, with her hair up in double buns and bangs and a few loose strands that frame her face. She's wearing a blue bikini with gold design. She's standing on one foot in a pool. Her other leg is above the water, doing a high kick, as she shows off her strong, flexible legs.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_073314_1092102458085130\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 181891228213810}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [181891228213810]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0669555664062, 220.04977416992188], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [1092102458085130, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [181891228213810, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of a woman cosplaying as chun-li, a beautiful woman with large breasts, slim waist, with her hair up in double buns and bangs and a few loose strands that frame her face. She's wearing a blue bikini with gold design. She's standing on one foot in a pool. Her other leg is above the water, doing a high kick, as she shows off her strong, flexible legs.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [85.567120935932, -211.65919185359886]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of a woman cosplaying as chun-li, a beautiful woman with large breasts, slim waist, with her hair up in double buns and bangs and a few loose strands that frame her face. She's wearing a blue bikini with gold design. She's standing on one foot in a pool. Her other leg is above the water, doing a high kick, as she shows off her strong, flexible legs.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d95ea599-7f8e-4343-b164-5b1790f171a9/width=832/59432230.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "UGCiK=#R^jRk~Cs:xaM|M|o#jcEL57b^RQNG", "type": "image", "metadata": { "hash": "UGCiK=#R^jRk~Cs:xaM|M|o#jcEL57b^RQNG", "size": 1331965, "width": 832, "height": 1216 }, "minor": false, "poi": false, "meta": { "seed": 442073804597973, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photograph of chun-li, a muscular woman with an hourglass figure, dressed as a policewoman, walking the streets of Chinatown at night.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_083508_401857856190986\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 442073804597973}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [442073804597973]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0669555664062, 220.04977416992188], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [401857856190986, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [442073804597973, null, null, null]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photograph of chun-li, a muscular woman with an hourglass figure, dressed as a policewoman, walking the streets of Chinatown at night.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [-4.006548379484157, -130.22871893392812]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photograph of chun-li, a muscular woman with an hourglass figure, dressed as a policewoman, walking the streets of Chinatown at night.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/3c968a58-9408-4e88-8325-60ac3fcfdc90/width=832/59432283.jpeg", "nsfwLevel": 4, "width": 832, "height": 1216, "hash": "UBB3KY*_^k?H~C+],W%MENrrRP-V57NGIUs;", "type": "image", "metadata": { "hash": "UBB3KY*_^k?H~C+],W%MENrrRP-V57NGIUs;", "size": 1301914, "width": 832, "height": 1216 }, "minor": false, "poi": false, "meta": { "seed": 1055627344092713, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photograph of chun-li, a muscular woman with an hourglass figure, dressed as a policewoman, walking the streets of Chinatown at night.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-02-22/v4chunli_083547_930921781389948\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 1055627344092713}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [1055627344092713]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0669555664062, 220.04977416992188], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [930921781389948, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1411.28076171875, 209.82850646972656], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/v4chunli_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [1055627344092713, null, null, null]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\Chun-Li\\\\v4\\\\Chun-Li_rank8_bf16-step04000.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photograph of chun-li, a muscular woman with an hourglass figure, dressed as a policewoman, walking the streets of Chinatown at night.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.573085533011683, \"offset\": [-4.006548379484157, -130.22871893392812]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photograph of chun-li, a muscular woman with an hourglass figure, dressed as a policewoman, walking the streets of Chinatown at night.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null } ], "downloadUrl": "https://civitai.com/api/download/models/1447839" }