Stable Diffusion img2img

This model uses diffusion-denoising mechanism as first proposed by SDEdit, Stable Diffusion is used for text-guided image-to-image translation. This model uses the weights from Stable Diffusion to generate new images from an input image using StableDiffusionImg2ImgPipeline from diffusers


API

If you're looking for an API, you can choose from your desired programming language.

POST
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 import requests import base64 # Use this function to convert an image file from the filesystem to base64 def image_file_to_base64(image_path): with open(image_path, 'rb') as f: image_data = f.read() return base64.b64encode(image_data).decode('utf-8') # Use this function to fetch an image from a URL and convert it to base64 def image_url_to_base64(image_url): response = requests.get(image_url) image_data = response.content return base64.b64encode(image_data).decode('utf-8') api_key = "YOUR_API_KEY" url = "https://api.segmind.com/v1/sd1.5-img2img" # Request payload data = { "image": image_url_to_base64("https://www.segmind.com/sd-img2img-input.jpeg"), # Or use image_file_to_base64("IMAGE_PATH") "samples": 1, "prompt": "A fantasy landscape, trending on artstation, mystical sky", "negative_prompt": "nude, disfigured, blurry", "scheduler": "DDIM", "num_inference_steps": 25, "guidance_scale": 10.5, "strength": 0.75, "seed": 98877465625, "img_width": 512, "img_height": 512, "base64": False } headers = {'x-api-key': api_key} response = requests.post(url, json=data, headers=headers) print(response.content) # The response is the generated image
RESPONSE
image/jpeg
HTTP Response Codes
200 - OKImage Generated
401 - UnauthorizedUser authentication failed
404 - Not FoundThe requested URL does not exist
405 - Method Not AllowedThe requested HTTP method is not allowed
406 - Not AcceptableNot enough credits
500 - Server ErrorServer had some issue with processing

Attributes


imageimage *

Input Image.


samplesint ( default: 1 ) Affects Pricing

Number of samples to generate.

min : 1,

max : 4


promptstr *

Prompt to render


negative_promptstr ( default: None )

Prompts to exclude, eg. 'bad anatomy, bad hands, missing fingers'


schedulerenum:str ( default: DDIM )

Type of scheduler.

Allowed values:


num_inference_stepsint ( default: 20 ) Affects Pricing

Number of denoising steps.

min : 20,

max : 100


guidance_scalefloat ( default: 7.5 )

Scale for classifier-free guidance

min : 0.1,

max : 25


strengthfloat ( default: 1 )

How much to transform the reference image

min : 0.1,

max : 1


seedint ( default: -1 )

Seed for image generation.


img_widthenum:int ( default: 512 ) Affects Pricing

Image resolution.

Allowed values:


img_heightenum:int ( default: 512 ) Affects Pricing

Image resolution.

Allowed values:


base64boolean ( default: 1 )

Base64 encoding of the output image.

To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.

Stable Diffusion Img2Img

Stable Diffusion Img2Img is a transformative AI model that's revolutionizing the way we approach image-to-image conversion. This model harnesses the power of machine learning to turn concepts into visuals, refine existing images, and translate one image to another with text-guided precision. It's an invaluable asset for creatives, marketers, and developers seeking to push the boundaries of digital imagery.

At the heart of Stable Diffusion Img2Img is a robust algorithm capable of understanding and manipulating visual content at a granular level. It takes an existing image and, guided by textual prompts, morphs it into a new creation that aligns with the user's vision. This model excels in tasks such as style transfer, detail enhancement, and subject transformation, all while maintaining the integrity of the original composition.

Advantages

  1. Text-Guided Imagery: Integrates textual prompts to steer the image transformation process, ensuring outputs are aligned with user intent.

  2. Seamless Style Transfers: Adapts the style of one image to another, enabling a smooth transition that feels natural and intentional..

  3. Detail Enhancement:Amplifies the details within images, bringing clarity and vibrance to visual elements.

  4. Creative Flexibility: Offers a wide range of possibilities, from subtle alterations to complete thematic overhauls..

Use Cases

  1. Creative Artwork: Artists can evolve their work, experimenting with different styles and motifs without starting from scratch.

  2. Marketing Material: Marketers can tailor images to fit brand narratives, ensuring consistency across campaigns.

  3. Product Design: Designers can visualize product variations quickly, streamlining the development process.

  4. Entertainment Media: Content creators in film and gaming can modify and enhance visual assets to fit evolving storylines.

  5. Educational Tools: Educators can create custom visuals to aid in teaching complex concepts.