{ "id": 1492916, "modelId": 1322295, "name": "v1.0", "createdAt": "2025-03-04T23:25:18.748Z", "updatedAt": "2025-03-05T14:17:37.273Z", "status": "Published", "publishedAt": "2025-03-04T23:31:03.958Z", "trainedWords": [ "alicegoodwin" ], "trainingStatus": null, "trainingDetails": null, "baseModel": "Flux.1 D", "baseModelType": null, "earlyAccessEndsAt": null, "earlyAccessConfig": null, "description": null, "uploadType": "Created", "usageControl": "Download", "air": "urn:air:flux1:lora:civitai:1322295@1492916", "stats": { "downloadCount": 194, "ratingCount": 0, "rating": 0, "thumbsUpCount": 29 }, "model": { "name": "Alice Goodwin - English adult model (~2009-12) [FLUX]", "type": "LORA", "nsfw": false, "poi": true }, "files": [ { "id": 1392863, "sizeKB": 32875.23046875, "name": "AliceGoodwin_rank8_bf16-step04000.safetensors", "type": "Model", "pickleScanResult": "Success", "pickleScanMessage": "No Pickle imports", "virusScanResult": "Success", "virusScanMessage": null, "scannedAt": "2025-03-04T23:30:34.378Z", "metadata": { "format": "SafeTensor", "size": null, "fp": null }, "hashes": { "AutoV1": "4081413D", "AutoV2": "3DD6E80758", "SHA256": "3DD6E8075814236B0455AC971AD0D688F5868437C4EF1F75A8D0F5E5C687AA36", "CRC32": "51C69125", "BLAKE3": "8CBDB3B23CE07DAAA3ACDA4D3CE7FD81A68F8BED0BD5ABDC3AD6B1A9E896E17F", "AutoV3": "F1CB2D6CDAB1" }, "primary": true, "downloadUrl": "https://civitai.com/api/download/models/1492916" } ], "images": [ { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/5f10c075-537e-44fb-aa0e-9ab449826903/width=832/61543592.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "UHDR$u~B0g57EhI;oKWBM|n%={%19uI:xas:", "type": "image", "metadata": { "hash": "UHDR$u~B0g57EhI;oKWBM|n%={%19uI:xas:", "size": 1524457, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 232788944861919, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"closeup photo of alicegoodwin, a woman with brown hair and hazel eyes, looking at the viewer with a smile\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_161122_403014503761973\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 232788944861919}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [232788944861919]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [403014503761973, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [232788944861919, null, null, null]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"closeup photo of alicegoodwin, a woman with brown hair and hazel eyes, looking at the viewer with a smile\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.6303940863128513, \"offset\": [158.84742313058547, 29.130284706255104]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "closeup photo of alicegoodwin, a woman with brown hair and hazel eyes, looking at the viewer with a smile", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/180a4f28-872d-4f13-afdd-4443ca37be2a/width=832/61543658.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "ULG95D_3?v%g~pt6kVWUMcM_niV@IoR*RPay", "type": "image", "metadata": { "hash": "ULG95D_3?v%g~pt6kVWUMcM_niV@IoR*RPay", "size": 1234448, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 45656668175743, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a brunette with hazel eyes, with lipstick, black frame glasses, hair in a low messy bun with a few loose stands that frame her face, wearing a tight, modest button-up collared shirt, navy blue jacket and a black tie that has been loosened slightly. Her large breasts cause the tie to hang in the air with some space between it and her blouse. She is standing at the front of a classroom by the whiteboard, staring at the viewer with shiny, parted lips, mouth open, leaning forward on a podium.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_161739_128867108874209\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 45656668175743}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [45656668175743]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [128867108874209, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [45656668175743, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a brunette with hazel eyes, with lipstick, black frame glasses, hair in a low messy bun with a few loose stands that frame her face, wearing a tight, modest button-up collared shirt, navy blue jacket and a black tie that has been loosened slightly. Her large breasts cause the tie to hang in the air with some space between it and her blouse. She is standing at the front of a classroom by the whiteboard, staring at the viewer with shiny, parted lips, mouth open, leaning forward on a podium.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.6303940863128513, \"offset\": [-97.23442553530653, -11.738001733804445]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a brunette with hazel eyes, with lipstick, black frame glasses, hair in a low messy bun with a few loose stands that frame her face, wearing a tight, modest button-up collared shirt, navy blue jacket and a black tie that has been loosened slightly. Her large breasts cause the tie to hang in the air with some space between it and her blouse. She is standing at the front of a classroom by the whiteboard, staring at the viewer with shiny, parted lips, mouth open, leaning forward on a podium.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/03ebb849-3317-47e5-b951-fc4e48f5ace4/width=832/61543691.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "U8E22$%20e8w5i9F_NtmPo-prC4:Q8S#XS?b", "type": "image", "metadata": { "hash": "U8E22$%20e8w5i9F_NtmPo-prC4:Q8S#XS?b", "size": 1345648, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 231053655705492, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a brunette with hazel eyes, wearing a tomato-red shirt, sitting at a table in a cafe with a mug of coffee in front of her. she is looking at the viewer and smiling. pov from across the table.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_161947_633561516554602\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 231053655705492}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [231053655705492]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [633561516554602, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [231053655705492, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a brunette with hazel eyes, wearing a tomato-red shirt, sitting at a table in a cafe with a mug of coffee in front of her. she is looking at the viewer and smiling. pov from across the table.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.6303940863128513, \"offset\": [-97.23442553530653, -11.738001733804445]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a brunette with hazel eyes, wearing a tomato-red shirt, sitting at a table in a cafe with a mug of coffee in front of her. she is looking at the viewer and smiling. pov from across the table.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/250ec1cd-5a7a-43e7-99b3-75bb168337b6/width=832/61543693.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "UMD+M3=_^$S6~TR%kSIr?Fs+xEj]x[jEw{xa", "type": "image", "metadata": { "hash": "UMD+M3=_^$S6~TR%kSIr?Fs+xEj]x[jEw{xa", "size": 1539052, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 1069298240461819, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a brunette with hazel eyes, with long hair wearing a dark green hoodie. she is hiking in an autumn forest on a sunny day. she is looking at the viewer and smiling.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_162022_43655947893782\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 1069298240461819}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [1069298240461819]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [43655947893782, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [1069298240461819, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a brunette with hazel eyes, with long hair wearing a dark green hoodie. she is hiking in an autumn forest on a sunny day. she is looking at the viewer and smiling.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.6303940863128513, \"offset\": [-97.23442553530653, -11.738001733804445]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a brunette with hazel eyes, with long hair wearing a dark green hoodie. she is hiking in an autumn forest on a sunny day. she is looking at the viewer and smiling.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d0f573d5-2caa-43cc-b68d-cf7e260f364d/width=832/61543694.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "ULHwPl-n%x%2_0jXM{R,%fs+VZs:}?xVEft6", "type": "image", "metadata": { "hash": "ULHwPl-n%x%2_0jXM{R,%fs+VZs:}?xVEft6", "size": 1428244, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 817644094998066, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a brunette with hazel eyes, wearing a vibrantly colored hiking jacket. she is hiking in an autumn forest on a sunny day. she is looking at the viewer and smiling.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_162108_156956021904407\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 817644094998066}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [817644094998066]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [156956021904407, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [817644094998066, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a brunette with hazel eyes, wearing a vibrantly colored hiking jacket. she is hiking in an autumn forest on a sunny day. she is looking at the viewer and smiling.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.6303940863128513, \"offset\": [-97.23442553530653, -11.738001733804445]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a brunette with hazel eyes, wearing a vibrantly colored hiking jacket. she is hiking in an autumn forest on a sunny day. she is looking at the viewer and smiling.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/058224ff-e690-4bd6-a8f3-7a0524f3237a/width=832/61543737.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "UDGuUQ0NTCg3^EXpSP.8-,IqRj%L~qI:9Ext", "type": "image", "metadata": { "hash": "UDGuUQ0NTCg3^EXpSP.8-,IqRj%L~qI:9Ext", "size": 1531144, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 1110886526095044, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a brunette with hazel eyes, with cascading brunette locks and radiant golden highlights, poses against the backdrop of a vividly colored mural in a serene, sun-dappled park on a crisp autumn afternoon. She\\u2019s wearing a white t-shirt and a oversized, distressed denim jacket. She stands confidently with one hand tucked into the pocket of her jacket, while her other hand playfully twirls a silver bracelet around her wrist. Her eyes sparkle with genuine warmth as she flashes a wide, infectious grin, revealing perfectly glossed lips. The soft golden light filters through the leaves of a nearby tree, casting an ethereal glow on her features and adding to the dreamlike quality of the scene.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_163241_857234752006923\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 1110886526095044}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [1110886526095044]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [857234752006923, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [1110886526095044, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a brunette with hazel eyes, with cascading brunette locks and radiant golden highlights, poses against the backdrop of a vividly colored mural in a serene, sun-dappled park on a crisp autumn afternoon. She\\u2019s wearing a white t-shirt and a oversized, distressed denim jacket. She stands confidently with one hand tucked into the pocket of her jacket, while her other hand playfully twirls a silver bracelet around her wrist. Her eyes sparkle with genuine warmth as she flashes a wide, infectious grin, revealing perfectly glossed lips. The soft golden light filters through the leaves of a nearby tree, casting an ethereal glow on her features and adding to the dreamlike quality of the scene.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.6303940863128513, \"offset\": [-97.23442553530653, -11.738001733804445]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a brunette with hazel eyes, with cascading brunette locks and radiant golden highlights, poses against the backdrop of a vividly colored mural in a serene, sun-dappled park on a crisp autumn afternoon. She\u2019s wearing a white t-shirt and a oversized, distressed denim jacket. She stands confidently with one hand tucked into the pocket of her jacket, while her other hand playfully twirls a silver bracelet around her wrist. Her eyes sparkle with genuine warmth as she flashes a wide, infectious grin, revealing perfectly glossed lips. The soft golden light filters through the leaves of a nearby tree, casting an ethereal glow on her features and adding to the dreamlike quality of the scene.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/0e34131a-bcfb-4df1-8a95-0ad58ccc05e3/width=832/61543748.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "UGCP^r_3t,xv~A-nIUoe%LxZ?bxug4t8xaoz", "type": "image", "metadata": { "hash": "UGCP^r_3t,xv~A-nIUoe%LxZ?bxug4t8xaoz", "size": 1229149, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 398678615734645, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a woman with hazel eyes, a sophisticated portrait of a confident woman in a stylish urban setting. She is wearing a well-tailored pantsuit in a rich, deep blue color, paired with a crisp WHITE blouse. She's standing in front of a modern city backdrop with sleek architecture. Dramatic lighting creates depth and contrast, highlighting the woman's features and the suit's sharp lines. DEEP BLUES, SILVERS, and BLACKS to evoke a sense of elegance and power. Focus on capturing the woman's self-assured expression and professional posture.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_165726_449684145756764\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 398678615734645}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [398678615734645]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [449684145756764, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [398678615734645, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a woman with hazel eyes, a sophisticated portrait of a confident woman in a stylish urban setting. She is wearing a well-tailored pantsuit in a rich, deep blue color, paired with a crisp WHITE blouse. She's standing in front of a modern city backdrop with sleek architecture. Dramatic lighting creates depth and contrast, highlighting the woman's features and the suit's sharp lines. DEEP BLUES, SILVERS, and BLACKS to evoke a sense of elegance and power. Focus on capturing the woman's self-assured expression and professional posture.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.5209868481924366, \"offset\": [-16.22033710928679, 58.14530370934153]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a woman with hazel eyes, a sophisticated portrait of a confident woman in a stylish urban setting. She is wearing a well-tailored pantsuit in a rich, deep blue color, paired with a crisp WHITE blouse. She's standing in front of a modern city backdrop with sleek architecture. Dramatic lighting creates depth and contrast, highlighting the woman's features and the suit's sharp lines. DEEP BLUES, SILVERS, and BLACKS to evoke a sense of elegance and power. Focus on capturing the woman's self-assured expression and professional posture.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f95ed340-e26d-4cac-936a-0fd0e4b92c83/width=832/61543808.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "UYI=764n_NIA?HR%tRM{yDRPs:oz%1M|NHs.", "type": "image", "metadata": { "hash": "UYI=764n_NIA?HR%tRM{yDRPs:oz%1M|NHs.", "size": 1314707, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 1006911162452662, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a woman with hazel eyes, wearing a dark green t-shirt tucked into denim jeans. she's standing by the window in her highrise apartment. She is posing holding a cute calico kitten, and smiling\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_170528_62495905334692\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 1006911162452662}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [1006911162452662]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [62495905334692, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [1006911162452662, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a woman with hazel eyes, wearing a dark green t-shirt tucked into denim jeans. she's standing by the window in her highrise apartment. She is posing holding a cute calico kitten, and smiling\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.6934334949441333, \"offset\": [29.05010104498962, -51.9130610085172]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a woman with hazel eyes, wearing a dark green t-shirt tucked into denim jeans. she's standing by the window in her highrise apartment. She is posing holding a cute calico kitten, and smiling", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f3a384e2-b791-4833-bc90-6ea8c281b3d8/width=832/61543850.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "U58qEK.95*t7}+OA10R*?Gwu-U$%13Ny={WB", "type": "image", "metadata": { "hash": "U58qEK.95*t7}+OA10R*?Gwu-U$%13Ny={WB", "size": 1531698, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 294247253488789, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a woman with hazel eyes, wearing round glasses and a high-neck slate blue top. She is standing at the chalkboard of a classroom. The woman is gazing seductively at the viewer with a soft smile and lustful eyes. Her dark red lipstick and piercing hazel eyes create a sense of both seduction and sexiness.\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_170747_302448302577283\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 294247253488789}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [294247253488789]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [302448302577283, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [294247253488789, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a woman with hazel eyes, wearing round glasses and a high-neck slate blue top. She is standing at the chalkboard of a classroom. The woman is gazing seductively at the viewer with a soft smile and lustful eyes. Her dark red lipstick and piercing hazel eyes create a sense of both seduction and sexiness.\"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.6934334949441335, \"offset\": [29.050101044989503, -51.9130610085173]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a woman with hazel eyes, wearing round glasses and a high-neck slate blue top. She is standing at the chalkboard of a classroom. The woman is gazing seductively at the viewer with a soft smile and lustful eyes. Her dark red lipstick and piercing hazel eyes create a sense of both seduction and sexiness.", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null }, { "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/10afa3f4-af4b-45b3-a857-cea05e9dc692/width=832/61543860.jpeg", "nsfwLevel": 1, "width": 832, "height": 1216, "hash": "UHF5:J00%$%2_39E%Mof~qDikWS4%MIUt7t7", "type": "image", "metadata": { "hash": "UHF5:J00%$%2_39E%Mof~qDikWS4%MIUt7t7", "size": 1401393, "width": 832, "height": 1216 }, "minor": false, "poi": true, "meta": { "seed": 943336449883980, "vaes": [ "ae.safetensors" ], "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": [\"33\", 0], \"steps\": 25, \"cfg\": 1.0, \"sampler_name\": \"euler\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"25\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"34\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"6\": {\"inputs\": {\"text\": \"photo of alicegoodwin, a woman with hazel eyes, standing confidently on the corner of a picturesque city street lined with historic brownstones. Her long hair cascades over her shoulders, perfectly complementing her plaid turtleneck top and high-slit skirt that elegantly hugs her figure. She accessorizes with gold hoop earrings and bracelets, her refined style blending modern elegance with classic charm. Behind her, the ivy-covered facade of a Victorian-era building contrasts with sleek glass skyscrapers in the distance, creating a captivating blend of old and new. The glow of a vintage streetlamp adds warmth to the cobblestone pavement, while the bustling city around her feels alive yet tranquil in this timeless urban setting. \", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Positive)\"}}, \"7\": {\"inputs\": {\"text\": \"\", \"clip\": [\"25\", 1]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Negative)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"30\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2025-03-04/img_171400_515798673584164\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"25\": {\"inputs\": {\"PowerLoraLoaderHeaderWidget\": {\"type\": \"PowerLoraLoaderHeaderWidget\"}, \"lora_1\": {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1}, \"\\u2795 Add Lora\": \"\", \"model\": [\"29\", 0], \"clip\": [\"28\", 0]}, \"class_type\": \"Power Lora Loader (rgthree)\", \"_meta\": {\"title\": \"Power Lora Loader (rgthree)\"}}, \"28\": {\"inputs\": {\"clip_name1\": \"t5xxl_fp8_e4m3fn.safetensors\", \"clip_name2\": \"clip_l.safetensors\", \"type\": \"flux\", \"+\": null}, \"class_type\": \"DualCLIPLoader\", \"_meta\": {\"title\": \"DualCLIPLoader\"}}, \"29\": {\"inputs\": {\"unet_name\": \"flux1-dev-fp8.safetensors\", \"weight_dtype\": \"fp8_e4m3fn_fast\", \"+\": null}, \"class_type\": \"UNETLoader\", \"_meta\": {\"title\": \"Load Diffusion Model\"}}, \"30\": {\"inputs\": {\"vae_name\": \"ae.safetensors\", \"+\": null}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}, \"33\": {\"inputs\": {\"seed\": 943336449883980}, \"class_type\": \"Seed (rgthree)\", \"_meta\": {\"title\": \"Seed (rgthree)\"}, \"is_changed\": [943336449883980]}, \"34\": {\"inputs\": {\"resolution\": \"832x1216 (0.68)\", \"batch_size\": 1, \"width_override\": 0, \"height_override\": 0}, \"class_type\": \"SDXLEmptyLatentSizePicker+\", \"_meta\": {\"title\": \"\\ud83d\\udd27 Empty Latent Size Picker\"}}}, \"workflow\": {\"last_node_id\": 39, \"last_link_id\": 64, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1211.7613525390625, 207.52772521972656], \"size\": [210, 46], \"flags\": {\"collapsed\": true}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 54}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [9], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 38, \"type\": \"Reroute\", \"pos\": [944.2876586914062, 218.80259704589844], \"size\": [75, 26], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"\", \"type\": \"*\", \"link\": 61}], \"outputs\": [{\"name\": \"\", \"type\": \"VAE\", \"links\": [54], \"slot_index\": 0}], \"properties\": {\"showOutputText\": false, \"horizontal\": false}}, {\"id\": 30, \"type\": \"VAELoader\", \"pos\": [-522.5015869140625, 230.6754608154297], \"size\": [318.4554138183594, 84.38363647460938], \"flags\": {\"collapsed\": true}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [61], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"ae.safetensors\", null]}, {\"id\": 29, \"type\": \"UNETLoader\", \"pos\": [-545.8380126953125, 272.9270935058594], \"size\": [343.6680603027344, 106], \"flags\": {\"collapsed\": true}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [64], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UNETLoader\"}, \"widgets_values\": [\"flux1-dev-fp8.safetensors\", \"fp8_e4m3fn_fast\", null]}, {\"id\": 28, \"type\": \"DualCLIPLoader\", \"pos\": [-531.428955078125, 319.86724853515625], \"size\": [347.35186767578125, 130], \"flags\": {\"collapsed\": true}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [62], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"DualCLIPLoader\"}, \"widgets_values\": [\"t5xxl_fp8_e4m3fn.safetensors\", \"clip_l.safetensors\", \"flux\", null]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [358.125244140625, 674.3905029296875], \"size\": [442.9576416015625, 94.61177062988281], \"flags\": {\"collapsed\": false}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 36}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [21], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Negative)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"\"], \"color\": \"#322\", \"bgcolor\": \"#533\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [1410.1173095703125, 56.27363967895508], \"size\": [1079.3812255859375, 1249.052001953125], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"%date:yyyy-MM-dd%/img_%date:hhmmss%_%KSampler.seed%\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [847.0668334960938, 217.7231903076172], \"size\": [533.8973999023438, 794.4248046875], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 37}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 21}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 49}, {\"name\": \"seed\", \"type\": \"INT\", \"link\": 47, \"widget\": {\"name\": \"seed\"}}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [7], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [515798673584164, \"randomize\", 25, 1, \"euler\", \"normal\", 1]}, {\"id\": 34, \"type\": \"SDXLEmptyLatentSizePicker+\", \"pos\": [275.15325927734375, 800.7346801757812], \"size\": [259.20001220703125, 170], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [49], \"slot_index\": 0}, {\"name\": \"width\", \"type\": \"INT\", \"links\": [], \"slot_index\": 1}, {\"name\": \"height\", \"type\": \"INT\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"SDXLEmptyLatentSizePicker+\"}, \"widgets_values\": [\"832x1216 (0.68)\", 1, 0, 0]}, {\"id\": 25, \"type\": \"Power Lora Loader (rgthree)\", \"pos\": [-313.511962890625, 235.59800720214844], \"size\": [645.4066772460938, 233.36923217773438], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 64, \"dir\": 3}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 62, \"dir\": 3}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [37], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [35, 36], \"slot_index\": 1, \"shape\": 3, \"dir\": 4}], \"properties\": {\"Show Strengths\": \"Single Strength\"}, \"widgets_values\": [null, {\"type\": \"PowerLoraLoaderHeaderWidget\"}, {\"on\": true, \"lora\": \"!Myloras\\\\AliceGoodwin\\\\v3\\\\AliceGoodwin_rank8_bf16-step04021.safetensors\", \"strength\": 1, \"strengthTwo\": null}, null, \"\"]}, {\"id\": 33, \"type\": \"Seed (rgthree)\", \"pos\": [549.7734985351562, 806.7191772460938], \"size\": [244.56271362304688, 159.79544067382812], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"SEED\", \"type\": \"INT\", \"links\": [47], \"slot_index\": 0, \"shape\": 3, \"dir\": 4}], \"properties\": {}, \"widgets_values\": [943336449883980, null, null, null]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [343.7825012207031, 215.52549743652344], \"size\": [472.1268005371094, 411.44879150390625], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 35}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"title\": \"CLIP Text Encode (Positive)\", \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"photo of alicegoodwin, a woman with hazel eyes, standing confidently on the corner of a picturesque city street lined with historic brownstones. Her long hair cascades over her shoulders, perfectly complementing her plaid turtleneck top and high-slit skirt that elegantly hugs her figure. She accessorizes with gold hoop earrings and bracelets, her refined style blending modern elegance with classic charm. Behind her, the ivy-covered facade of a Victorian-era building contrasts with sleek glass skyscrapers in the distance, creating a captivating blend of old and new. The glow of a vintage streetlamp adds warmth to the cobblestone pavement, while the bustling city around her feels alive yet tranquil in this timeless urban setting. \"], \"color\": \"#232\", \"bgcolor\": \"#353\"}], \"links\": [[4, 6, 0, 3, 1, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [9, 8, 0, 9, 0, \"IMAGE\"], [21, 7, 0, 3, 2, \"CONDITIONING\"], [35, 25, 1, 6, 0, \"CLIP\"], [36, 25, 1, 7, 0, \"CLIP\"], [37, 25, 0, 3, 0, \"MODEL\"], [47, 33, 0, 3, 4, \"INT\"], [49, 34, 0, 3, 3, \"LATENT\"], [54, 38, 0, 8, 1, \"VAE\"], [61, 30, 0, 38, 0, \"*\"], [62, 28, 0, 25, 1, \"CLIP\"], [64, 29, 0, 25, 0, \"MODEL\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.630394086312849, \"offset\": [201.6483972528476, 60.82836605446473]}, \"VHS_latentpreview\": false, \"VHS_latentpreviewrate\": 0, \"ue_links\": []}, \"version\": 0.4, \"widget_idx_map\": {\"3\": {\"sampler_name\": 4, \"scheduler\": 5}, \"33\": {\"seed\": 0}}}}", "steps": 25, "models": [], "prompt": "photo of alicegoodwin, a woman with hazel eyes, standing confidently on the corner of a picturesque city street lined with historic brownstones. Her long hair cascades over her shoulders, perfectly complementing her plaid turtleneck top and high-slit skirt that elegantly hugs her figure. She accessorizes with gold hoop earrings and bracelets, her refined style blending modern elegance with classic charm. Behind her, the ivy-covered facade of a Victorian-era building contrasts with sleek glass skyscrapers in the distance, creating a captivating blend of old and new. The glow of a vintage streetlamp adds warmth to the cobblestone pavement, while the bustling city around her feels alive yet tranquil in this timeless urban setting. ", "denoise": 1, "sampler": "Euler", "cfgScale": 1, "modelIds": [], "scheduler": "normal", "upscalers": [], "versionIds": [], "controlNets": [], "additionalResources": [] }, "availability": "Public", "hasMeta": true, "hasPositivePrompt": true, "onSite": false, "remixOfId": null } ], "downloadUrl": "https://civitai.com/api/download/models/1492916" }