Notion enhances workspace search with Cohere Rerank
Notion
Notion is a single space where you can think, write, and plan. Capture thoughts, manage projects, or even run an entire company — and do it exactly the way you want.
Overview
Notion, a connected workspace platform, recently launched Notion AI, a personalized assistant seamlessly integrated into the product to help teams work faster, write better, and think bigger. Notion AI taps into customer knowledge across their workspace and connected tools like Slack and Google Drive. To enhance its search capabilities in their workspace environments, Notion integrated Cohere Rerank into their platform.
The Challenge
Notion's connected workspace enables teams to collaboratively write, plan, and organize their work in one central hub that integrates with their other tools. However, as organizations grow, so does the volume of information and knowledge accumulated across their Notion workspace, which can make it increasingly challenging for employees to quickly find the relevant content and answers they need.
Notion needed a solution that would:
- Enhance the accuracy of search results across large datasets.
- Improve the speed of information retrieval for workspaces with high volumes of content including thousands of documents and databases.
- Maintain high relevance in search results without increasing storage costs by relying on embeddings or other costly solutions.
Additionally, as a global company with a diverse user base, Notion required a solution capable of supporting multiple languages and handling non-English datasets effectively.
The Solution
To address these challenges, Notion deployed Cohere Rerank and integrated the model directly into their workspace search.
Cohere Rerank provided key advantages:
- Enhanced search relevance: Cohere’s ranking technology boosted the relevance of search results by applying Rerank before the generative model processes the user query. This ensured that users received the most accurate and contextually relevant results.
- Efficient document search: Notion’s workspaces often contain fewer than 1,000 documents. By skipping traditional embedding models for smaller workspaces and leveraging Rerank, Notion avoided embedding and vector search, reducing both complexity and costs.
- Scalable cloud infrastructure: Notion integrated the solution using Amazon SageMaker, allowing for auto-scaling capabilities. This infrastructure ensured that Notion could dynamically adjust computational resources based on user demand, optimizing performance during peak and off-peak hours.
- Multilingual support: Cohere’s solution natively supports many languages, ensuring that Notion’s global user base, more than half of whom work with multilingual datasets in EMEA and APAC, could benefit from the same enhanced search capabilities.
Notion software engineer Abhishek Modi explains, “One big part of the search pipeline is precision. With Cohere Rerank, we no longer have to worry about it and can focus on other parts like recall.”
The Impact
Using Cohere Rerank, Notion is able to deliver substantial benefits:
- Improved search performance: With Rerank, Notion significantly enhanced the speed and accuracy of their search results. Modi explains, “Cohere Rerank is very good at telling you, given a question how relevant is a given document to that question. And it's much more accurate than the embedding models which will take you from the 100,000 documents to the 200 documents.”
- Cost savings: By eliminating the need for embedding and vector search in many instances, Notion reduced both operational costs (associated with embedding and storage) and complexity, while maintaining high search relevance. Additionally, Cohere Rerank is used to combine search results from different sources for their customers like results from Slack and GitHub, which occurs quite frequently and provides another layer of cost savings while retaining high precision answers.
- Scalability: The implementation of Amazon Sagemaker’s auto-scaling features ensured that Notion’s search functionality could seamlessly handle variable traffic loads, providing an even more reliable performance across global markets.
Notion was also interested in a close relationship with their LLM provider. Modi explains, “we wanted to have a closer relationship with whomever was going to build this with us, because an issue with a lot of the open source things is if you need to fix something, you need to go fix it yourself. The fact that we're able to partner with Cohere and work together to improve the solution was quite important for us.”
And when it comes to performance metrics, Notion’s primary goal is to improve user experience month-on-month, so they focus KPIs on whether a negative experience, in this case a wrong or not-so-accurate answer, can be improved the next time it is asked. Latest figures show millions of Notion users have tried Notion AI features, contributing significant revenue growth to the company. A figure that is growing rapidly month-on-month.
Notion’s use of Rerank helped them scale their tool’s search functionality while maintaining cost efficiency and delivering a top-tier user experience, driving greater satisfaction among their users.
"Cohere is a key part of what makes Notion AI work. Cohere Rerank gives us both the speed and quality we need, and it’s consistently improving. It’s been essential for getting our AI Connectors out the door quickly."
— Cofounder & CTO