Enterprise search remains a significant challenge within organizations, with teams reportedly wasting up to 32 days a year navigating through multiple tools to find answers.
At Quench, we address this problem by building a robust enterprise search engine that seamlessly integrates with the tools teams already use – such as Notion, Google Drive, and Slack – via simple pre-built integrations and APIs.
In our work with businesses, we've learned that delivering correct answers is paramount. Enterprises are understandably wary of inaccuracies or hallucinations from Retrieval-Augmented Generation (RAG) solutions.
With this in mind, we conducted a comprehensive analysis of enterprise search tools, comparing Quench against Sana Labs, NotebookLM, and Google Drive Gemini by running over 300 queries using content from two sample datasets. You can find our methodology at the end.
Our evaluation focused on two key metrics that matter most to businesses:
Here are our key findings:
We evaluated four enterprise search tools - Quench, Sana Labs, NotebookLM, and Google Drive Gemini - by running 324 queries using content from two of our datasets.
We assessed performance based on:
Our evaluation metric "preferred" represents the ratio of cases where one tool was rated better than another, excluding instances where both performed equally.
For example, if Quench is 4.2x more preferred than Sana Labs in correctness, it means that for every time Sana Labs was rated better, Quench was rated better 4.2 times as often.
On fact-based queries alone:
Out of 163 fact-based queries, Quench achieved a 89.8% correctness rate, significantly outperforming the other enterprise search tools. This is also higher than its overall correctness score of 86.3% across all 324 queries.
Sana Labs demonstrated a correctness score of 50.9% for fact-based queries, which is slightly higher than its overall score of 48.1%.
In contrast, NotebookLM’s correctness decreased from 37.0% overall to 32.5% for fact-based queries, indicating a drop in accuracy for this subset of queries.
Meanwhile, Google Drive Gemini's accuracy remained stable, with a correctness score of 35.0% and 35.8% for fact-based and overall queries respectively.
NotebookLM tends to be rated more complete than Quench with NotebookLM returning a complete answer 68% of the time, whereas Quench only returned a complete answer with a success rate of 63%.
We noted that NotebookLM tended to be more verbose and cited a lot more of the documents that have been uploaded.
This makes NotebookLM ideal if your workflow involves open-ended research, brainstorming, and creative ideation.
However, there is a tradeoff as it leads to lower correctness scores and a higher potential for hallucinations and incorrect results.
Here is how each tool responded to the following question for a synthetic dataset:
The organization ensures that employees' interests are protected when they report issues by valuing honesty and ensuring there is no retaliation against employees for being forthright. Any retaliatory actions taken by managers are addressed seriously, and managers are held accountable for inappropriate behavior, including receiving necessary retraining [1][3]. Additionally, the organization has implemented effective tracking systems that promote transparency regarding employee performance, which helps in maintaining a culture of accountability and straightforwardness [2]. This approach encourages employees to report any concerns without fear of negative repercussions, supporting a positive workplace culture.
The organization ensures that employees' interests are protected when they report issues by:
Providing Direct Access to Leadership: Employees can directly message the CEO, COO, or CFO to report any issues, ensuring that their concerns are heard at the highest levels.
Encouraging Transparency: The organization fosters an environment where transparency is valued, and employees are encouraged to speak up without fear of retribution.
Addressing Coercion and Manipulation: Any use of threats, coercion, or manipulation by managers is taken seriously and addressed promptly to protect employees' interests
NotebookLM produced an extremely detailed response that included several correct elements but also retrieved information that, according to our curated reference answers, isn’t correct for the question. It included mentions of the following which were irrelevant to the question:
Because the system aims for maximum completeness by presenting as much sourced material as possible, it may include contradictory or irrelevant information, which ultimately affects correctness – a critical factor for enterprise search applications.
"The document does not specify how the organization protects the interests of employees who report issues."
Quench demonstrated superior performance on correctness for the following reasons:
Sana Labs performed well and mentioned direct access to leadership which is missing from Quench’s answer. However, it lacks key points including retraining of managers and tracking systems, which were essential elements mentioned in the reference claims.
NotebookLM provided excessive detail that included both correct and irrelevant information, reducing its overall correctness score.
Google Drive Gemini incorrectly claimed the document did not contain information on the topic, when in fact it did.
Quench delivers superior results because of our innovative approach to RAG. While traditional systems often struggle with incomplete or disconnected information, Quench’s proprietary method delivers accurate and complete results.
When you use Quench, every piece of information is enriched with its full context.
In an enterprise setting, knowledge is often poorly organised, missing context at scale. We leverage a contextual RAG approach that adds the necessary context to your assets, from documents to transcripts from recordings, creating self-contained knowledge blocks that are complete with relevant background information.
This addresses a common challenge with raw transcripts for example, which often contain fragmented dialogue and irrelevant information.
By reconstructing these into coherent, context-rich segments, we ensure each citation retrieved has the complete information needed to properly answer a question.
Quench goes beyond basic search with:
Enterprise environments have unique communication challenges.
Each organization has its own acronyms, naming conventions, and terminology that generalist LLMs don't understand.
Names like "Husayn" might be incorrectly transcribed as "Hussein," "Husain," or "Usain" in meeting recordings.
When someone asks "What did Husayn say about our Q2 goals?" standard systems might miss the answer entirely because they don't recognize the name variation.
Our extensive testing shows that effective enterprise search requires customization. Quench recognizes that your business is not only another dataset, but a unique environment with its own language. This explains why Quench consistently delivers more accurate, contextually-rich answers compared to other solutions.
NotebookLM is ideal if your workflow involves manually managing document collections and detailed citations - making it a great tool for open-ended research, brainstorming, and creative ideation.
On the other hand, Quench is purpose-built for fast, fact-based enterprise search. It automatically captures and indexes an organization's full spectrum of data - from Slack messages and meeting recordings to Notion pages - delivering swift, accurate, and context-rich answers with robust citations. In environments where speed and accuracy are critical for decision-making, Quench offers a more targeted solution.
If you are curious about how Quench delivers high precision, or you want to benchmark our precision against your internal solution, we are up for the challenge.
You can sign up here and a member of our team will be happy to help you get started.