Skip to main content
CohereReleased November 2, 2023

Cohere Embed v3

Undisclosed parameters

Cohere Embed v3 is Cohere's entry in a crowded field. Context window: 0.512K tokens.

Context

512 tokens

Input

$0.10/1M tokens

Key Specifications

🏆

Arena Rank

Not disclosed

📐

Context Window

512 tokens

📥

Input Price

per 1M tokens

$0.10/1M tokens

📤

Output Price

per 1M tokens

N/A (embeddings)

🧠

Parameters

Undisclosed

🔒

Open Source

No

Best For

SearchRAGsemantic similarityclustering

About Cohere Embed v3

Cohere Embed v3, developed by Cohere, is an embedding model with a 512-token input limit designed for semantic search, retrieval-augmented generation, and clustering applications. The model generates dense vector representations of text that capture semantic meaning, enabling similarity-based search across document collections. Embed v3 supports 100+ languages and produces compact embeddings optimized for vector database storage and retrieval. It outperforms previous generations on the MTEB benchmark across multiple retrieval and classification tasks. Priced at $0.10 per million tokens, it offers cost-effective embedding generation for production search pipelines. The model serves as the foundation for enterprise search systems, recommendation engines, and RAG architectures. Embed v3 remains widely deployed despite the release of v4, due to its mature ecosystem of integrations and proven production reliability.

Built byCohere

Pricing per 1M tokens

Input Tokens

$0.10/1M tokens

Output Tokens

N/A (embeddings)

Frequently Asked Questions

What is Cohere Embed v3?
Cohere Embed v3, developed by Cohere, is an embedding model with a 512-token input limit designed for semantic search, retrieval-augmented generation, and clustering applications. The model generates dense vector representations of text that capture semantic meaning, enabling similarity-based search across document collections. Embed v3 supports 100+ languages and produces compact embeddings optimized for vector database storage and retrieval. It outperforms previous generations on the MTEB benchmark across multiple retrieval and classification tasks. Priced at $0.10 per million tokens, it offers cost-effective embedding generation for production search pipelines. The model serves as the foundation for enterprise search systems, recommendation engines, and RAG architectures. Embed v3 remains widely deployed despite the release of v4, due to its mature ecosystem of integrations and proven production reliability.
How much does Cohere Embed v3 cost?
Cohere Embed v3 costs $0.10/1M tokens per 1M input tokens and N/A (embeddings) per 1M output tokens. You pay only for what you use, which keeps costs predictable.
What is Cohere Embed v3's context window?
Cohere Embed v3 has a context window of 512 tokens tokens. This determines how much text the model can process in a single request — bigger windows mean longer documents and richer conversation history.
Is Cohere Embed v3 open source?
No, Cohere Embed v3 is a proprietary model available through Cohere's API. You get managed infrastructure, regular updates, and support as part of the package.
What is Cohere Embed v3 best for?
Cohere Embed v3 is best suited for: Search, RAG, semantic similarity, clustering. These use cases play to the model's strengths in capability, speed, and cost within Cohere's lineup.