Cohere is a Toronto-based enterprise AI platform founded in 2019 by Aidan Gomez, a co-author of the seminal “Attention Is All You Need” paper that introduced the Transformer architecture. Cohere specializes in secure, customizable language models and tools for business applications including retrieval-augmented generation (RAG), semantic search, and AI agents. By early 2026, the company achieved $240 million ARR with 50%+ quarter-over-quarter growth and approximately 70% gross margins.1)
The Command family of generative models is optimized for enterprise workloads:
The Embed model generates dense vector embeddings for semantic search and retrieval tasks. Deployable via APIs or through Cohere's Model Vault for secure on-premises inference.
Rerank 4 (released December 2025) offers 32K context for enterprise search and RAG pipelines. It uses a cross-encoder architecture that performs cross-attention between queries and documents for high-precision relevance scoring.3)
Command R and R+ are specifically optimized for Retrieval-Augmented Generation, integrating with external data sources, tools, and APIs. The models support grounding responses in enterprise data with citation anchoring, reducing hallucination in knowledge-intensive applications.
Cohere prioritizes regulated industries with multiple deployment options:
Cohere is backed by investors including NVIDIA, AMD, and Salesforce. The company closed a $100 million funding round in September 2025 and has been positioning for a potential 2026 IPO. Revenue grew from under $100 million to $240 million ARR, with strong unit economics driven by efficient model architectures.6)