LLaVA 13B

LLaVA 13B is a Vision-language model which allows both image and text as inputs.


Pricing

Serverless Pricing

Buy credits that can be used anywhere on Segmind

Input: $0.3, Output: $0.3 per million tokens

LLaVA 13B

LLaVA 13B is a vision-language model (VLM) trained on OSS LLM-generated instruction following data. Its state-of-the-art architecture enables seamless interaction between visual content and textual prompts. FireLLaVA supports multi-image and multi-prompt generation. You can seamlessly integrate multiple images into your queries, enhancing context and specificity.

Applications

  • Image Captioning: Generate descriptive captions for images, enriching content across social media, e-commerce, and more.

  • Visual Question Answering (VQA): Pose questions about images, and FireLLaVA provides accurate answers.

  • Creative Writing: Fuel your imagination by combining visual cues with textual prompts