Meta AIReleased September 25, 2024

Llama 3.2 90B Vision

Open Source#11 Arena Rank90B parameters

Context

128K

Input

Free (open)

Key Specifications

🏆

Arena Rank

#11

📐

Context Window

128K

📥

Input Price

per 1M tokens

Free (open)

📤

Output Price

per 1M tokens

Free (open)

🧠

Parameters

90B

🔓

Open Source

Yes

Best For

Image understandingvisual QAmultimodal tasks

About Llama 3.2 90B Vision

Llama 3.2 90B Vision is Meta's first open-source multimodal model, capable of understanding both text and images. With 90 billion parameters, it can analyze charts, diagrams, photographs, and documents while maintaining strong text-only performance. This model represents Meta's push into multimodal AI, enabling the open-source community to build applications that understand visual content without relying on proprietary APIs.

Built byMeta AI

Pricing per 1M tokens

Input Tokens

Free (open)

Output Tokens

Free (open)

Frequently Asked Questions

What is Llama 3.2 90B Vision?
Llama 3.2 90B Vision is Meta's first open-source multimodal model, capable of understanding both text and images. With 90 billion parameters, it can analyze charts, diagrams, photographs, and documents while maintaining strong text-only performance. This model represents Meta's push into multimodal AI, enabling the open-source community to build applications that understand visual content without relying on proprietary APIs.
How much does Llama 3.2 90B Vision cost?
Llama 3.2 90B Vision costs Free (open) per 1 million input tokens and Free (open) per 1 million output tokens. Pricing is based on token usage, making it cost-effective for both small and large-scale applications.
What is Llama 3.2 90B Vision's context window?
Llama 3.2 90B Vision has a context window of 128K tokens. This determines how much text the model can process in a single request — larger context windows allow the model to handle longer documents, maintain more conversation history, and reason over bigger codebases.
Is Llama 3.2 90B Vision open source?
Yes, Llama 3.2 90B Vision is open source. This means the model weights are publicly available, allowing developers and organizations to download, fine-tune, and self-host the model on their own infrastructure. Open-source models offer greater flexibility and data privacy control.
What is Llama 3.2 90B Vision best for?
Llama 3.2 90B Vision is best suited for: Image understanding, visual QA, multimodal tasks. These use cases leverage the model's specific strengths in terms of capability, speed, and cost-effectiveness within Meta AI's model lineup.