{
  "118": {
    "class_type": "ReferenceLatent",
    "inputs": {
      "conditioning": [
        "112",
        0
      ],
      "latent": [
        "119",
        0
      ]
    }
  },
  "119": {
    "class_type": "VAEEncode",
    "inputs": {
      "pixels": [
        "120",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "120": {
    "class_type": "ImageScaleToTotalPixels",
    "inputs": {
      "image": [
        "121",
        0
      ],
      "megapixels": [
        "135",
        0
      ],
      "upscale_method": "lanczos",
      "resolution_steps": 1
    }
  },
  "129": {
    "class_type": "ImageScaleToTotalPixels",
    "inputs": {
      "image": [
        "128",
        0
      ],
      "megapixels": [
        "135",
        0
      ],
      "upscale_method": "lanczos",
      "resolution_steps": 1
    }
  },
  "130": {
    "class_type": "VAEEncode",
    "inputs": {
      "pixels": [
        "129",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "140": {
    "class_type": "ReferenceLatent",
    "inputs": {
      "conditioning": [
        "131",
        0
      ],
      "latent": [
        "139",
        0
      ]
    }
  },
  "139": {
    "class_type": "VAEEncode",
    "inputs": {
      "pixels": [
        "138",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "138": {
    "class_type": "ImageScaleToTotalPixels",
    "inputs": {
      "image": [
        "137",
        0
      ],
      "megapixels": [
        "135",
        0
      ],
      "upscale_method": "lanczos",
      "resolution_steps": 1
    }
  },
  "144": {
    "class_type": "ReferenceLatent",
    "inputs": {
      "conditioning": [
        "140",
        0
      ],
      "latent": [
        "143",
        0
      ]
    }
  },
  "143": {
    "class_type": "VAEEncode",
    "inputs": {
      "pixels": [
        "142",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "142": {
    "class_type": "ImageScaleToTotalPixels",
    "inputs": {
      "image": [
        "141",
        0
      ],
      "megapixels": [
        "135",
        0
      ],
      "upscale_method": "lanczos",
      "resolution_steps": 1
    }
  },
  "131": {
    "class_type": "ReferenceLatent",
    "inputs": {
      "conditioning": [
        "118",
        0
      ],
      "latent": [
        "130",
        0
      ]
    }
  },
  "128": {
    "class_type": "LoadImage",
    "inputs": {
      "image": "pexels-mimfathi-10919291.jpg",
      "upload": "image"
    }
  },
  "137": {
    "class_type": "LoadImage",
    "inputs": {
      "image": "pexels-peterfazekas-1137340.jpg",
      "upload": "image"
    }
  },
  "121": {
    "class_type": "LoadImage",
    "inputs": {
      "image": "pexels-frank-barning-744539-7353336.jpg",
      "upload": "image"
    }
  },
  "147": {
    "class_type": "VAEDecode",
    "inputs": {
      "samples": [
        "125",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "150": {
    "class_type": "VAEEncode",
    "inputs": {
      "pixels": [
        "156",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "112": {
    "class_type": "ReferenceLatent",
    "inputs": {
      "conditioning": [
        "107",
        0
      ],
      "latent": [
        "150",
        0
      ]
    }
  },
  "154": {
    "class_type": "GetImageSize",
    "inputs": {
      "image": [
        "156",
        0
      ]
    }
  },
  "141": {
    "class_type": "LoadImage",
    "inputs": {
      "image": "pexels-alina-zahorulko-48514961-31445409.jpg",
      "upload": "image"
    }
  },
  "126": {
    "class_type": "UNETLoader",
    "inputs": {
      "unet_name": "flux-2-klein-4b-fp8.safetensors",
      "weight_dtype": "default"
    }
  },
  "153": {
    "class_type": "LoraLoaderModelOnly",
    "inputs": {
      "model": [
        "126",
        0
      ],
      "lora_name": "flux-2-klein-4B-outpaint-lora.safetensors",
      "strength_model": 1
    }
  },
  "125": {
    "class_type": "VAEEncode",
    "inputs": {
      "pixels": [
        "115",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "151": {
    "class_type": "LoadImage",
    "inputs": {
      "image": "pexels-marta-dzedyshko-1042863-2067629.jpg",
      "upload": "image"
    }
  },
  "115": {
    "class_type": "ImageScaleToTotalPixels",
    "inputs": {
      "image": [
        "151",
        0
      ],
      "megapixels": [
        "135",
        0
      ],
      "upscale_method": "lanczos",
      "resolution_steps": 1
    }
  },
  "135": {
    "class_type": "PrimitiveFloat",
    "inputs": {
      "value": 1
    }
  },
  "148": {
    "class_type": "GetImageSize",
    "inputs": {
      "image": [
        "115",
        0
      ]
    }
  },
  "149": {
    "class_type": "ImageScale",
    "inputs": {
      "image": [
        "151",
        0
      ],
      "width": [
        "148",
        0
      ],
      "height": [
        "148",
        1
      ],
      "upscale_method": "lanczos",
      "crop": "center"
    }
  },
  "105": {
    "class_type": "EmptyFlux2LatentImage",
    "inputs": {
      "width": [
        "154",
        0
      ],
      "height": [
        "154",
        1
      ],
      "batch_size": 1
    }
  },
  "117": {
    "class_type": "Image Comparer (rgthree)",
    "inputs": {
      "image_a": [
        "104",
        0
      ],
      "image_b": [
        "161",
        0
      ],
      "prompt": [
        {
          "name": "A",
          "selected": true,
          "url": "/api/view?filename=rgthree.compare._temp_jevlz_00003_.png&type=temp&subfolder=&rand=0.4866416249775115"
        },
        {
          "name": "B",
          "selected": true,
          "url": "/api/view?filename=rgthree.compare._temp_jevlz_00004_.png&type=temp&subfolder=&rand=0.8894862737747382"
        }
      ]
    }
  },
  "9": {
    "class_type": "SaveImage",
    "inputs": {
      "images": [
        "104",
        0
      ],
      "filename_prefix": "Flux2Klein_Refine"
    }
  },
  "104": {
    "class_type": "VAEDecode",
    "inputs": {
      "samples": [
        "169",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "169": {
    "class_type": "LanPaint_KSampler",
    "inputs": {
      "model": [
        "166",
        0
      ],
      "positive": [
        "163",
        0
      ],
      "negative": [
        "136",
        0
      ],
      "latent_image": [
        "168",
        0
      ],
      "seed": 540731376013430,
      "steps": 4,
      "cfg": 1,
      "sampler_name": "euler",
      "scheduler": "simple",
      "denoise": 1,
      "LanPaint_NumSteps": 5,
      "LanPaint_PromptMode": "Image First",
      "LanPaint_Info": "LanPaint KSampler. For more info, visit https://github.com/scraed/LanPaint. If you find it useful, please give a star ⭐️!",
      "Inpainting_mode": "🖼️ Image Inpainting"
    }
  },
  "167": {
    "class_type": "CLIPLoader",
    "inputs": {
      "clip_name": "qwen_3_4b.safetensors",
      "type": "flux2",
      "device": "default"
    }
  },
  "146": {
    "class_type": "CLIPLoader",
    "inputs": {
      "clip_name": "qwen_3_4b.safetensors",
      "type": "flux2",
      "device": "default"
    }
  },
  "107": {
    "class_type": "CLIPTextEncode",
    "inputs": {
      "clip": [
        "146",
        0
      ],
      "text": "Fill the green spaces according to the image"
    }
  },
  "136": {
    "class_type": "ConditioningZeroOut",
    "inputs": {
      "conditioning": [
        "107",
        0
      ]
    }
  },
  "160": {
    "class_type": "CLIPTextEncode",
    "inputs": {
      "clip": [
        "167",
        0
      ],
      "text": "improve details and textures, high quality, photorealistic"
    }
  },
  "163": {
    "class_type": "ReferenceLatent",
    "inputs": {
      "conditioning": [
        "160",
        0
      ],
      "latent": [
        "168",
        0
      ]
    }
  },
  "166": {
    "class_type": "UNETLoader",
    "inputs": {
      "unet_name": "flux-2-klein-4b-fp8.safetensors",
      "weight_dtype": "default"
    }
  },
  "168": {
    "class_type": "LanPaint_KSampler",
    "inputs": {
      "model": [
        "153",
        0
      ],
      "positive": [
        "100",
        0
      ],
      "negative": [
        "136",
        0
      ],
      "latent_image": [
        "105",
        0
      ],
      "seed": 883920378869541,
      "steps": 8,
      "cfg": 1,
      "sampler_name": "euler_ancestral",
      "scheduler": "simple",
      "denoise": 1,
      "LanPaint_NumSteps": 5,
      "LanPaint_PromptMode": "Image First",
      "LanPaint_Info": "LanPaint KSampler. For more info, visit https://github.com/scraed/LanPaint. If you find it useful, please give a star ⭐️!",
      "Inpainting_mode": "🖼️ Image Inpainting"
    }
  },
  "100": {
    "class_type": "FluxGuidance",
    "inputs": {
      "conditioning": [
        "144",
        0
      ],
      "guidance": 4
    }
  },
  "161": {
    "class_type": "VAEDecode",
    "inputs": {
      "samples": [
        "168",
        0
      ],
      "vae": [
        "102",
        0
      ]
    }
  },
  "171": {
    "class_type": "SaveImage",
    "inputs": {
      "images": [
        "161",
        0
      ],
      "filename_prefix": "Flux2klein4b"
    }
  },
  "102": {
    "class_type": "VAELoader",
    "inputs": {
      "vae_name": "flux2-vae.safetensors"
    }
  },
  "156": {
    "class_type": "ImagePadKJ",
    "inputs": {
      "image": [
        "149",
        0
      ],
      "left": 505,
      "right": 574,
      "top": 254,
      "bottom": 85,
      "extra_padding": 0,
      "pad_mode": "color",
      "color": "0, 255, 0"
    }
  },
  "155": {
    "class_type": "PreviewImage",
    "inputs": {
      "images": [
        "156",
        0
      ]
    }
  }
}