diff --git a/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_controlnet_t2i_adapters.json b/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_controlnet_t2i_adapters.json index 568b7400..74cbaaf8 100644 --- a/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_controlnet_t2i_adapters.json +++ b/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_controlnet_t2i_adapters.json @@ -6,125 +6,124 @@ "aws_endpoint_url": "https://my-endpoint.backblaze.com", "aws_bucket_name": "your-bucket", "workflow_json": { - "3": { - "inputs": { - "seed": 891858402356003, - "steps": 20, - "cfg": 8, - "sampler_name": "uni_pc_bh2", - "scheduler": "normal", - "denoise": 1, - "model": [ - "34", - 0 - ], - "positive": [ - "23", - 0 - ], - "negative": [ - "7", - 0 - ], - "latent_image": [ - "5", - 0 - ] - }, - "class_type": "KSampler" + "3": { + "inputs": { + "seed": 891858402356003, + "steps": 20, + "cfg": 8, + "sampler_name": "uni_pc_bh2", + "scheduler": "normal", + "denoise": 1, + "model": [ + "34", + 0 + ], + "positive": [ + "23", + 0 + ], + "negative": [ + "7", + 0 + ], + "latent_image": [ + "5", + 0 + ] }, - "5": { - "inputs": { - "width": 832, - "height": 384, - "batch_size": 1 - }, - "class_type": "EmptyLatentImage" + "class_type": "KSampler" + }, + "5": { + "inputs": { + "width": 832, + "height": 384, + "batch_size": 1 }, - "7": { - "inputs": { - "text": "(hands), text, error, cropped, (worst quality:1.2), (low quality:1.2), normal quality, (jpeg artifacts:1.3), signature, watermark, username, blurry, artist name, monochrome, sketch, censorship, censor, (copyright:1.2)", - "clip": [ - "34", - 1 - ] - }, - "class_type": "CLIPTextEncode" + "class_type": "EmptyLatentImage" + }, + "7": { + "inputs": { + "text": "(hands), text, error, cropped, (worst quality:1.2), (low quality:1.2), normal quality, (jpeg artifacts:1.3), signature, watermark, username, blurry, artist name, monochrome, sketch, censorship, censor, (copyright:1.2)", + "clip": [ + "34", + 1 + ] }, - "8": { - "inputs": { - "samples": [ - "3", - 0 - ], - "vae": [ - "34", - 2 - ] - }, - "class_type": "VAEDecode" + "class_type": "CLIPTextEncode" + }, + "8": { + "inputs": { + "samples": [ + "3", + 0 + ], + "vae": [ + "34", + 2 + ] }, - "9": { - "inputs": { - "filename_prefix": "ComfyUI", - "images": [ - "8", - 0 - ] - }, - "class_type": "SaveImage" + "class_type": "VAEDecode" + }, + "9": { + "inputs": { + "filename_prefix": "ComfyUI", + "images": [ + "8", + 0 + ] }, - "20": { - "inputs": { - "image": "https://upload.wikimedia.org/wikipedia/commons/3/3e/Stereogram_Tut_Shark_Depthmap.png", - "upload": "image" - }, - "class_type": "LoadImage" + "class_type": "SaveImage" + }, + "20": { + "inputs": { + "image": "https://upload.wikimedia.org/wikipedia/commons/3/3e/Stereogram_Tut_Shark_Depthmap.png", + "upload": "image" }, - "23": { - "inputs": { - "strength": 1.0000000000000004, - "conditioning": [ - "24", - 0 - ], - "control_net": [ - "33", - 0 - ], - "image": [ - "20", - 0 - ] - }, - "class_type": "ControlNetApply" + "class_type": "LoadImage" + }, + "23": { + "inputs": { + "strength": 1.0000000000000004, + "conditioning": [ + "24", + 0 + ], + "control_net": [ + "33", + 0 + ], + "image": [ + "20", + 0 + ] }, - "24": { - "inputs": { - "text": "underwater photograph shark\n\n\n\n", - "clip": [ - "34", - 1 - ] - }, - "class_type": "CLIPTextEncode" + "class_type": "ControlNetApply" + }, + "24": { + "inputs": { + "text": "underwater photograph shark\n\n\n\n", + "clip": [ + "34", + 1 + ] }, - "33": { - "inputs": { - "control_net_name": "diff_control_sd15_depth_fp16.safetensors", - "model": [ - "34", - 0 - ] - }, - "class_type": "DiffControlNetLoader" + "class_type": "CLIPTextEncode" + }, + "33": { + "inputs": { + "control_net_name": "diff_control_sd15_depth_fp16.safetensors", + "model": [ + "34", + 0 + ] }, - "34": { - "inputs": { - "ckpt_name": "v1-5-pruned-emaonly.ckpt" - }, - "class_type": "CheckpointLoaderSimple" - } + "class_type": "DiffControlNetLoader" + }, + "34": { + "inputs": { + "ckpt_name": "v1-5-pruned-emaonly.ckpt" + }, + "class_type": "CheckpointLoaderSimple" } } } diff --git a/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_image2image.json b/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_image2image.json index 29745590..39568e9c 100644 --- a/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_image2image.json +++ b/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_image2image.json @@ -6,103 +6,101 @@ "aws_endpoint_url": "https://my-endpoint.backblaze.com", "aws_bucket_name": "your-bucket", "workflow_json": { - { - "3": { - "inputs": { - "seed": 280823642470253, - "steps": 20, - "cfg": 8, - "sampler_name": "dpmpp_2m", - "scheduler": "normal", - "denoise": 0.8700000000000001, - "model": [ - "14", - 0 - ], - "positive": [ - "6", - 0 - ], - "negative": [ - "7", - 0 - ], - "latent_image": [ - "12", - 0 - ] - }, - "class_type": "KSampler" + "3": { + "inputs": { + "seed": 280823642470253, + "steps": 20, + "cfg": 8, + "sampler_name": "dpmpp_2m", + "scheduler": "normal", + "denoise": 0.8700000000000001, + "model": [ + "14", + 0 + ], + "positive": [ + "6", + 0 + ], + "negative": [ + "7", + 0 + ], + "latent_image": [ + "12", + 0 + ] }, - "6": { - "inputs": { - "text": "photograph of victorian woman with wings, sky clouds, meadow grass", - "clip": [ - "14", - 1 - ] - }, - "class_type": "CLIPTextEncode" + "class_type": "KSampler" + }, + "6": { + "inputs": { + "text": "photograph of victorian woman with wings, sky clouds, meadow grass", + "clip": [ + "14", + 1 + ] }, - "7": { - "inputs": { - "text": "watermark, text", - "clip": [ - "14", - 1 - ] - }, - "class_type": "CLIPTextEncode" + "class_type": "CLIPTextEncode" + }, + "7": { + "inputs": { + "text": "watermark, text", + "clip": [ + "14", + 1 + ] }, - "8": { - "inputs": { - "samples": [ - "3", - 0 - ], - "vae": [ - "14", - 2 - ] - }, - "class_type": "VAEDecode" + "class_type": "CLIPTextEncode" + }, + "8": { + "inputs": { + "samples": [ + "3", + 0 + ], + "vae": [ + "14", + 2 + ] }, - "9": { - "inputs": { - "filename_prefix": "ComfyUI", - "images": [ - "8", - 0 - ] - }, - "class_type": "SaveImage" + "class_type": "VAEDecode" + }, + "9": { + "inputs": { + "filename_prefix": "ComfyUI", + "images": [ + "8", + 0 + ] }, - "10": { - "inputs": { - "image": "https://raw.githubusercontent.com/comfyanonymous/ComfyUI/master/input/example.png", - "upload": "image" - }, - "class_type": "LoadImage" + "class_type": "SaveImage" + }, + "10": { + "inputs": { + "image": "https://raw.githubusercontent.com/comfyanonymous/ComfyUI/master/input/example.png", + "upload": "image" }, - "12": { - "inputs": { - "pixels": [ - "10", - 0 - ], - "vae": [ - "14", - 2 - ] - }, - "class_type": "VAEEncode" + "class_type": "LoadImage" + }, + "12": { + "inputs": { + "pixels": [ + "10", + 0 + ], + "vae": [ + "14", + 2 + ] }, - "14": { - "inputs": { - "ckpt_name": "v1-5-pruned-emaonly.ckpt" - }, - "class_type": "CheckpointLoaderSimple" - } + "class_type": "VAEEncode" + }, + "14": { + "inputs": { + "ckpt_name": "v1-5-pruned-emaonly.ckpt" + }, + "class_type": "CheckpointLoaderSimple" } } } diff --git a/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_text2image.json b/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_text2image.json index 9cd5d084..942525a2 100644 --- a/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_text2image.json +++ b/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_text2image.json @@ -6,91 +6,89 @@ "aws_endpoint_url": "https://my-endpoint.backblaze.com", "aws_bucket_name": "your-bucket", "workflow_json": { - { - "3": { - "inputs": { - "seed": 156680208700286, - "steps": 20, - "cfg": 8, - "sampler_name": "euler", - "scheduler": "normal", - "denoise": 1, - "model": [ - "4", - 0 - ], - "positive": [ - "6", - 0 - ], - "negative": [ - "7", - 0 - ], - "latent_image": [ - "5", - 0 - ] - }, - "class_type": "KSampler" + "3": { + "inputs": { + "seed": 156680208700286, + "steps": 20, + "cfg": 8, + "sampler_name": "euler", + "scheduler": "normal", + "denoise": 1, + "model": [ + "4", + 0 + ], + "positive": [ + "6", + 0 + ], + "negative": [ + "7", + 0 + ], + "latent_image": [ + "5", + 0 + ] }, - "4": { - "inputs": { - "ckpt_name": "v1-5-pruned-emaonly.ckpt" - }, - "class_type": "CheckpointLoaderSimple" + "class_type": "KSampler" + }, + "4": { + "inputs": { + "ckpt_name": "v1-5-pruned-emaonly.ckpt" }, - "5": { - "inputs": { - "width": 512, - "height": 512, - "batch_size": 1 - }, - "class_type": "EmptyLatentImage" + "class_type": "CheckpointLoaderSimple" + }, + "5": { + "inputs": { + "width": 512, + "height": 512, + "batch_size": 1 }, - "6": { - "inputs": { - "text": "beautiful scenery nature glass bottle landscape, , purple galaxy bottle", - "clip": [ - "4", - 1 - ] - }, - "class_type": "CLIPTextEncode" + "class_type": "EmptyLatentImage" + }, + "6": { + "inputs": { + "text": "beautiful scenery nature glass bottle landscape, , purple galaxy bottle", + "clip": [ + "4", + 1 + ] }, - "7": { - "inputs": { - "text": "text, watermark", - "clip": [ - "4", - 1 - ] - }, - "class_type": "CLIPTextEncode" + "class_type": "CLIPTextEncode" + }, + "7": { + "inputs": { + "text": "text, watermark", + "clip": [ + "4", + 1 + ] }, - "8": { - "inputs": { - "samples": [ - "3", - 0 - ], - "vae": [ - "4", - 2 - ] - }, - "class_type": "VAEDecode" + "class_type": "CLIPTextEncode" + }, + "8": { + "inputs": { + "samples": [ + "3", + 0 + ], + "vae": [ + "4", + 2 + ] }, - "9": { - "inputs": { - "filename_prefix": "ComfyUI", - "images": [ - "8", - 0 - ] - }, - "class_type": "SaveImage" - } + "class_type": "VAEDecode" + }, + "9": { + "inputs": { + "filename_prefix": "ComfyUI", + "images": [ + "8", + 0 + ] + }, + "class_type": "SaveImage" } } } diff --git a/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_upscale.json b/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_upscale.json index f08c7630..4efa03fb 100644 --- a/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_upscale.json +++ b/build/COPY_ROOT/opt/serverless/docs/example_payloads/raw_upscale.json @@ -6,110 +6,108 @@ "aws_endpoint_url": "https://my-endpoint.backblaze.com", "aws_bucket_name": "your-bucket", "workflow_json": { - { - "3": { - "inputs": { - "seed": 833543590226030, - "steps": 20, - "cfg": 8, - "sampler_name": "euler", - "scheduler": "normal", - "denoise": 1, - "model": [ - "4", - 0 - ], - "positive": [ - "6", - 0 - ], - "negative": [ - "7", - 0 - ], - "latent_image": [ - "5", - 0 - ] - }, - "class_type": "KSampler" + "3": { + "inputs": { + "seed": 833543590226030, + "steps": 20, + "cfg": 8, + "sampler_name": "euler", + "scheduler": "normal", + "denoise": 1, + "model": [ + "4", + 0 + ], + "positive": [ + "6", + 0 + ], + "negative": [ + "7", + 0 + ], + "latent_image": [ + "5", + 0 + ] }, - "4": { - "inputs": { - "ckpt_name": "v1-5-pruned-emaonly.ckpt" - }, - "class_type": "CheckpointLoaderSimple" + "class_type": "KSampler" + }, + "4": { + "inputs": { + "ckpt_name": "v1-5-pruned-emaonly.ckpt" }, - "5": { - "inputs": { - "width": 512, - "height": 512, - "batch_size": 1 - }, - "class_type": "EmptyLatentImage" + "class_type": "CheckpointLoaderSimple" + }, + "5": { + "inputs": { + "width": 512, + "height": 512, + "batch_size": 1 }, - "6": { - "inputs": { - "text": "masterpiece best quality girl standing in victorian clothing", - "clip": [ - "4", - 1 - ] - }, - "class_type": "CLIPTextEncode" + "class_type": "EmptyLatentImage" + }, + "6": { + "inputs": { + "text": "masterpiece best quality girl standing in victorian clothing", + "clip": [ + "4", + 1 + ] }, - "7": { - "inputs": { - "text": "bad hands", - "clip": [ - "4", - 1 - ] - }, - "class_type": "CLIPTextEncode" + "class_type": "CLIPTextEncode" + }, + "7": { + "inputs": { + "text": "bad hands", + "clip": [ + "4", + 1 + ] }, - "8": { - "inputs": { - "samples": [ - "3", - 0 - ], - "vae": [ - "4", - 2 - ] - }, - "class_type": "VAEDecode" + "class_type": "CLIPTextEncode" + }, + "8": { + "inputs": { + "samples": [ + "3", + 0 + ], + "vae": [ + "4", + 2 + ] }, - "9": { - "inputs": { - "filename_prefix": "ComfyUI", - "images": [ - "14", - 0 - ] - }, - "class_type": "SaveImage" + "class_type": "VAEDecode" + }, + "9": { + "inputs": { + "filename_prefix": "ComfyUI", + "images": [ + "14", + 0 + ] }, - "13": { - "inputs": { - "model_name": "RealESRGAN_x4.pth" - }, - "class_type": "UpscaleModelLoader" + "class_type": "SaveImage" + }, + "13": { + "inputs": { + "model_name": "RealESRGAN_x4.pth" }, - "14": { - "inputs": { - "upscale_model": [ - "13", - 0 - ], - "image": [ - "8", - 0 - ] - }, - "class_type": "ImageUpscaleWithModel" - } + "class_type": "UpscaleModelLoader" + }, + "14": { + "inputs": { + "upscale_model": [ + "13", + 0 + ], + "image": [ + "8", + 0 + ] + }, + "class_type": "ImageUpscaleWithModel" } } }