If you're looking for an API, you can choose from your desired programming language.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
import requests
import base64
# Use this function to convert an image file from the filesystem to base64
def image_file_to_base64(image_path):
with open(image_path, 'rb') as f:
image_data = f.read()
return base64.b64encode(image_data).decode('utf-8')
# Use this function to fetch an image from a URL and convert it to base64
def image_url_to_base64(image_url):
response = requests.get(image_url)
image_data = response.content
return base64.b64encode(image_data).decode('utf-8')
api_key = "YOUR_API_KEY"
url = "https://api.segmind.com/v1/flux-dev"
# Request payload
data = {
"prompt": "detailed cinematic dof render of an old dusty detailed CRT monitor on a wooden desk in a dim room with items around, messy dirty room. On the screen are the letters “FLUX dev” glowing softly. High detail hard surface render",
"samples": 1,
"guidance": 3.5,
"steps": 25,
"prompt_strength": 0.8,
"aspect_ratio": "1:1",
"seed": 46588,
"output_format": "webp",
"output_quality": 80
}
headers = {'x-api-key': api_key}
response = requests.post(url, json=data, headers=headers)
print(response.content) # The response is the generated image
Text prompt for image generation
Number of samples to generate.
min : 1,
max : 4
guidance
min : 0,
max : 10
number of steps
min : 1,
max : 50
Prompt Strength
min : 0,
max : 1
Type of scheduler.
Allowed values:
Seed for random number generation
An enumeration.
Allowed values:
Quality when saving the output images, from 0 to 100. 100 is best quality, 0 is lowest quality. Not relevant for .png outputs
min : 0,
max : 100
To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.
Unlock the full potential of generative AI with Segmind. Create stunning visuals and innovative designs with total creative control. Take advantage of powerful development tools to automate processes and models, elevating your creative workflow.
Gain greater control by dividing the creative process into distinct steps, refining each phase.
Customize at various stages, from initial generation to final adjustments, ensuring tailored creative outputs.
Integrate and utilize multiple models simultaneously, producing complex and polished creative results.
Deploy Pixelflows as APIs quickly, without server setup, ensuring scalability and efficiency.
The Flux Dev model by Black Forest Labs is an advanced AI model designed for generating and transforming visual content from simple text inputs. Utilizing cutting-edge AI and machine learning techniques, Flux Dev provides users with the ability to create high-quality images effortlessly.
To use the Flux Dev model:
Input Text Prompt: Provide a textual description of the desired image. The model processes this input to generate a corresponding visual output.
Run the Model: Execute the model with your text input. The AI algorithm interprets the description to produce an image.
Review Outputs: Evaluate the generated images for quality and relevance to your input.
Graphic Design: Automate the creation of graphics based on simple text descriptions, saving time on repetitive design tasks.
Advertising: Generate visual content tailored to marketing campaigns, quickly producing assets that align with brand messages.
Content Creation: Assist writers and content creators in visualizing their narratives by generating illustrative images from textual descriptions.
Web Development: Enhance websites with unique, dynamically generated images that improve user engagement and aesthetic appeal.
Research and Development: Utilize the model for experimental purposes in AI research, testing the boundaries of text-to-image generation capabilities.
Story Diffusion turns your written narratives into stunning image sequences.
Take a picture/gif and replace the face in it with a face of your choice. You only need one image of the desired face. No dataset, no training
This model is capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask
Take a picture/gif and replace the face in it with a face of your choice. You only need one image of the desired face. No dataset, no training