Fashion Video Generator - SeeDance 1 Lite + SegFit 1.3
Convert static fashion flat-lays and mannequin images into captivating, cinematic videos. Powered by Bytedance's latest Video generator SeeDance 1.0 Lite.
If you're looking for an API, here is a sample code in NodeJS to help you out.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
const axios = require('axios');
const api_key = "YOUR API KEY";
const url = "https://api.segmind.com/workflows/68a20ae2c49c91c2edbbb3fd-v1";
const data = {
Attire_image: "publicly accessible image link",
Resolution: "480p" // Options: 480p, 720p
};
axios.post(url, data, {
headers: {
'x-api-key': api_key,
'Content-Type': 'application/json'
}
}).then((response) => {
console.log(response.data);
});
1
2
3
4
5
{
"poll_url": "<base_url>/requests/<some_request_id>",
"request_id": "some_request_id",
"status": "QUEUED"
}
You can poll the above link to get the status and output of your request.
1
2
3
{
"Wan2.2_Output": "any user input string"
}
Attributes
To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.
Fashion video generator - Powered by SeeDance 1.0 Lite and SegFit 1.3
Last updated: 17 Aug 2025 by Rohit
This is a workflow that is similar to Fashion video generator here, with one big difference. We have replaced the Wan 2.2 model with the latest Image to video model as of today from Bytedance, SeeDance 1.0 Lite, which is the faster and more economical version of SeeDance Pro.
This workflows generates 480p or 720p videos with just one image of the attire: flat-lay, actual model photo or Mannequin image, no prompts required. Perfect for fashion brands and social media agencies looking to present the designs in an engaging way.
To help you understand about this workflow: we used Flux Kontext Max to create an image that acts as a starting point for Segmind's SegFit 1.3 virtual try on modal (our best V-TON model to date) to then replace the attire with the exact attire shared by the user. We use SegFit make sure that the attire in the final output is 100% consistent with the one shared by the user as the input. The output from SegFit is then given as starting frame for SeeDance 1.0 Lite image to video model to generate high quality video output at either 480p or 720p.
Key Models
- Image Transformation: We use Flux Kontext Max to convert boring white background flat-lays and mannequin images into photo-realistic images that look real.
- Virtual Try-On: Uses Segmind SegFit v1.3 for realistic outfit visualization on different models.
- Prompt Generation: We use Claude 3.7 Sonnet to create scene-appropriate text prompts that guide video creation, ensuring the output aligns with desired visual storytelling. It analyzes the image to determine optimal camera movements like push-ins, pans, and tracking shots to convey the best narrative for fashion presentations
- Video Creation: Converts processed images into engaging fashion videos using SeeDance 1.0 Lite, generating cinematic video for life like content.
Use Cases
This is perfect for fashion brands to generate captivating videos for website and social media, helping them increase engagement and reach. Realistic videos with models helps attract attention to the product and makes ads and social media posts more effective. Brands and agencies can also experiment with different model ethnicities, backgrounds and add audio tracks to improve the content's effectiveness. Influencers can also use this workflow to generate content for them without having to physically shoot for each attire.
Models Used in the Pixelflow
segfit-v1.3
SegFit v1.3 enables hyper-realistic virtual try-ons, enhancing online fashion retail experiences without physical photoshoots.

seedance-v1-lite-image-to-video
Seedance 1.0 transforms text and images into engaging 720p dynamic videos with cinematic storytelling.

flux-kontext-max
FLUX.1 Kontext [max] transforms textual descriptions into stunning, high-fidelity images with seamless typography integration.

claude-3.7-sonnet
Claude 3.7 Sonnet is a large language model (LLM) launched by Anthropic AI. It is considered state-of-the-art, outperforming previous versions of Claude and competing models in a variety of tasks
