LLaVA 13B

LLaVA 13B is a Vision-language model which allows both image and text as inputs.


Pricing

Serverless Pricing

Buy credits that can be used anywhere on Segmind

Input: $0.300, Output: $0.300 per million tokens

LLaVA 13B

LLaVA 13B is a vision-language model (VLM) trained on OSS LLM-generated instruction following data. Its state-of-the-art architecture enables seamless interaction between visual content and textual prompts. FireLLaVA supports multi-image and multi-prompt generation. You can seamlessly integrate multiple images into your queries, enhancing context and specificity.

Applications

  • Image Captioning: Generate descriptive captions for images, enriching content across social media, e-commerce, and more.

  • Visual Question Answering (VQA): Pose questions about images, and FireLLaVA provides accurate answers.

  • Creative Writing: Fuel your imagination by combining visual cues with textual prompts

Cookie settings

We use cookies to enhance your browsing experience, analyze site traffic, and personalize content. By clicking "Accept all", you consent to our use of cookies.