If you're looking for an API, you can choose from your desired programming language.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
import requests
import base64
# Use this function to convert an image file from the filesystem to base64
def image_file_to_base64(image_path):
with open(image_path, 'rb') as f:
image_data = f.read()
return base64.b64encode(image_data).decode('utf-8')
# Use this function to fetch an image from a URL and convert it to base64
def image_url_to_base64(image_url):
response = requests.get(image_url)
image_data = response.content
return base64.b64encode(image_data).decode('utf-8')
api_key = "YOUR_API_KEY"
url = "https://api.segmind.com/v1/sd1.5-rcnz"
# Request payload
data = {
"prompt": "a duck, warm vibrant colours, natural lighting, dappled lighting, diffused lighting, absurdres, highres, 8k, uhd, hdr, rtx, unreal 5, octane render, RAW photo, photorealistic, global illumination, subsurface scattering",
"negative_prompt": "beard, EasyNegative, lowres, chromatic aberration, depth of field, motion blur, blurry, bokeh, bad quality, worst quality, multiple arms, badhandv4",
"scheduler": "dpmpp_2m",
"num_inference_steps": 30,
"guidance_scale": 7,
"samples": 1,
"seed": 53094109118,
"img_width": 512,
"img_height": 768,
"base64": False
}
headers = {'x-api-key': api_key}
response = requests.post(url, json=data, headers=headers)
print(response.content) # The response is the generated image
Prompt to render
Prompts to exclude, eg. 'bad anatomy, bad hands, missing fingers'
Type of scheduler.
Allowed values:
Number of denoising steps.
min : 20,
max : 100
Scale for classifier-free guidance
min : 0.1,
max : 25
Number of samples to generate.
min : 1,
max : 4
Seed for image generation.
Width of the image.
Allowed values:
Height of the Image
Allowed values:
Base64 encoding of the output image.
To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.
Unlock the full potential of generative AI with Segmind. Create stunning visuals and innovative designs with total creative control. Take advantage of powerful development tools to automate processes and models, elevating your creative workflow.
Gain greater control by dividing the creative process into distinct steps, refining each phase.
Customize at various stages, from initial generation to final adjustments, ensuring tailored creative outputs.
Integrate and utilize multiple models simultaneously, producing complex and polished creative results.
Deploy Pixelflows as APIs quickly, without server setup, ensuring scalability and efficiency.
Built on the foundation of Stable Diffusion 1.5, the RCNZ Cartoon model excels in interpreting artistic ideas and converting them into vivid 3D cartoons. Its sophisticated design allows for intricate detailing and depth.
3D Cartoon Rendering: Specializes in transforming 2D cartoon concepts into captivating 3D images.
Creative Versatility: Offers a wide range of possibilities for cartoon styles and themes..
High-Quality Outputs: Produces vibrant, detailed, and visually appealing cartoon images.
Animation and Film:Ideal for animators and filmmakers looking to create unique 3D cartoon scenes and characters.
Digital Art: Artists can explore new dimensions in cartoon artistry.
Marketing and Advertising: Useful for creating engaging 3D cartoon visuals for promotional content.
Educational Content: Enhances learning materials with appealing 3D cartoon illustrations.
Take a picture/gif and replace the face in it with a face of your choice. You only need one image of the desired face. No dataset, no training
This model is capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask
The SDXL model is the official upgrade to the v1.5 model. The model is released as open-source software
Take a picture/gif and replace the face in it with a face of your choice. You only need one image of the desired face. No dataset, no training