Llama 3.2 90B Vision Instruct

Experience the cutting edge of AI with Llama 3.2-90B Vision-Instruct. This 90B parameter multimodal LLM excels at image understanding, reasoning, captioning, and more.

Playground

Try the model in real time below.

loading...

Click or Drag-n-Drop

PNG, JPG or GIF, Up-to 5mb

Please send a message from the prompt textbox to see a response here.

FEATURES

PixelFlow allows you to use all these features

Unlock the full potential of generative AI with Segmind. Create stunning visuals and innovative designs with total creative control. Take advantage of powerful development tools to automate processes and models, elevating your creative workflow.

Segmented Creation Workflow

Gain greater control by dividing the creative process into distinct steps, refining each phase.

Customized Output

Customize at various stages, from initial generation to final adjustments, ensuring tailored creative outputs.

Layering Different Models

Integrate and utilize multiple models simultaneously, producing complex and polished creative results.

Workflow APIs

Deploy Pixelflows as APIs quickly, without server setup, ensuring scalability and efficiency.

Llama 3.2-90B Vision-Instruct

The Llama 3.2-90B Vision-Instruct is a multimodal large language model (LLM) developed by Meta. It is engineered to process both textual and visual inputs, providing advanced capabilities in areas such as image understanding and reasoning

Key Features of Llama 3.2-90B Vision-Instruct

  • Parameter Count: The model consists of 90 billion parameters (88.8 billion).

  • Input Modalities: Supports text and image inputs, enabling versatile applications.

  • Output Modality: Generates text outputs, making it suitable for a wide range of tasks.

  • Architecture: Built upon the Llama 3.1 text-only model, enhanced with a vision adapter. The vision adapter employs cross-attention layers to integrate image encoder representations into the core LLM.

  • Context Length: Features a 128k context length.

Technical Specifications

  • Training Data: Trained on a dataset of 6 billion image and text pairs.

  • Data Cutoff: The pretraining data has a cutoff of December 2023.

  • Instruction Tuning: Fine-tuned using publicly available vision instruction datasets and over 3 million synthetically generated examples, combining supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF)

Intended Use Cases

The model is optimized for visual recognition, image reasoning, captioning, and question answering about image.

F.A.Q.

Frequently Asked Questions

Take creative control today and thrive.

Start building with a free account or consult an expert for your Pro or Enterprise needs. Segmind's tools empower you to transform your creative visions into reality.

Pixelflow Banner

Cookie settings

We use cookies to enhance your browsing experience, analyze site traffic, and personalize content. By clicking "Accept all", you consent to our use of cookies.