Models

Here are some popular generative model APIs that you can use in your application.

ControlNet Openpose

This model corresponds to the ControlNet conditioned on Human Pose Estimation.

ControlNet Scribble

This model corresponds to the ControlNet conditioned on Scribble images.

ControlNet HED

This model corresponds to the ControlNet conditioned on HED Boundary.

ControlNet Depth

This model corresponds to the ControlNet conditioned on Depth estimation.

ControlNet Canny

This model corresponds to the ControlNet conditioned on Canny edges.

ControlNet MLSD

This model corresponds to the ControlNet conditioned on M-LSD straight line detection.

Stable Diffusion Inpainting

Stable Diffusion Inpainting is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask.

Stable Diffusion 2.1

This is a model that can be used to generate and modify images based on text prompts. It is a Latent Diffusion Model that uses a fixed, pretrained text encoder (OpenCLIP-ViT/H).

Stable Diffusion img2img

This model uses diffusion-denoising mechanism as first proposed by SDEdit, Stable Diffusion is used for text-guided image-to-image translation. This model uses the weights from Stable Diffusion to generate new images from an input image using StableDiffusionImg2ImgPipeline from diffusers.