
Cohere's English-optimized text embedding model with 1,024-dimensional outputs for semantic search, RAG, and classification.
embed-english-v3.0 generates semantic embeddings from English text, producing 1,024-dimensional vectors suitable for retrieval-augmented generation, classification, clustering, and semantic search tasks. It achieves state-of-the-art performance on MTEB and BEIR benchmarks and was trained on approximately 1 billion English training pairs. The model supports batches of up to 96 inputs with 512 tokens maximum per input, and supports both text and image embedding. Pricing is $0.10 per million tokens. A notable limitation is that it requires explicit input_type specification to differentiate between search documents, queries, classification, and clustering tasks.
Bottom line: *Use for high-quality English semantic search and RAG applications with cost-efficient API pricing; avoid if you need multilingual support or extensive context windows beyond 512 tokens.*
Rate-limited access for learning and prototyping
Usage-based billing per token
View full pricing on cohere.com →
Pricing may have changed since last verified. Check the official site for current plans.
No community ratings yet. Be the first to rate this tool!
No community benchmarks yet. Be the first to share a real-world data point.
No reviews yet. Be the first to share your experience.
No discussion yet. Sign in to start the conversation.
Spotted incorrect or missing data? Join our community of contributors.
Sign Up to ContributeBe the first to contribute. General notes, observations, gotchas, and tips from people who use this tool day-to-day.
Be the first to contribute. Concrete time/cost savings, with context. e.g. "Cut my code review backlog from 4h to 45m per week."
embed-english-v3.0 is a specialized text embedding model from Cohere designed specifically for English language inputs. It transforms text phrases, sentences, and paragraphs into 1,024-dimensional dense vectors that capture semantic meaning, enabling applications to perform similarity searches, retrieve relevant documents, classify text, and cluster similar items.
The model was trained on approximately 1 billion English language pairs and achieves state-of-the-art performance on widely-used benchmarks including the Massive Text Embedding Benchmark (MTEB) across 56 datasets for retrieval, classification, and clustering, as well as the BEIR benchmark for zero-shot dense retrieval tasks. These benchmark results demonstrate strong generalization across diverse domains.
API access supports batch processing of up to 96 texts per request, with each input limited to 512 tokens maximum. The model requires users to specify an input_type parameter that distinguishes between search_document, search_query, classification, clustering, and image inputs to optimize embeddings for specific use cases. For image inputs, both text and image can be embedded together. Output embeddings are available in multiple formats including float, int8, uint8, binary, ubinary, and base64 representations to optimize for different storage and retrieval requirements.
The model is available through Cohere’s API with usage-based pricing of $0.10 per million tokens, with a free tier available for learning and prototyping subject to rate limits. Deployment options include Cohere’s managed API platform, AWS SageMaker, Azure AI Foundry, and OCI Generative AI service.
Share a real-world data point. Plausibility-checked by our AI moderator before publishing.