Rocket.Chat Labs #1: What happens when you throw 1.2M messages at AI search

Pavithra Sudhakar
March 18, 2026
·
min read

Welcome to Rocket.Chat Labs: our way of showing what is cooking in the R&D kitchen. No polished demos on curated datasets. No slide decks dressed up as evidence. Just real engineering, real data, and the honest story of what we found. 

The first in a series where we open up our R&D process: what we built, how we tested it, and what we actually found. We are starting with something that goes to the heart of why Rocket.Chat exists: the problem of finding what matters, when it matters, inside a sea of messages.

The problem: Critical information, buried in plain sight

If you run any messaging platform at scale, you know this problem intimately. 

Information goes missing. Not because it was never shared. Because it was shared in the wrong channel, or the right channel three months ago, or in a thread that branched off a conversation you were never part of. The knowledge is there. It is just buried.

Someone answered your exact question six weeks ago. Someone documented the fix. But when you need it, you cannot find it. You do not remember which channel it was in.

Traditional keyword search only works when you already know what you are looking for. In some organizations this means duplicated work. In others, it means something far more consequential, such as the following: 

01 The commander who walked in without the full picture

Critical updates from three field teams exist in separate channels: a status report filed two days ago, a flag in a side thread, an update nobody escalated. None of it surfaces in time. Decisions get made on a partial picture.

02 The analyst who almost missed the pattern

The signals are spread across weeks of messages and threads from multiple teams. Manually piecing it together takes days. By the time the pattern is visible, the window to act has narrowed. The information existed. It just was not findable at the speed required.

03 The decision maker flying blind

Four active workstreams. Each team communicates. No single place to surface what matters. Operating on status updates hours old, filtered through layers of relay. The ground truth is sitting in the channels. They just cannot see it.

The problem is not that organizations fail to generate information.

The problem is that critical intel is distributed across channels, and without the ability to connect the dots in real time, the people who need the full picture are always working with a partial one.

Scattered knowledge is a mission risk.

How Intelligent Search works: Intent search, not a keyword search

Before we get into what happens at scale, it helps to understand what Intelligent Search actually does, because it is fundamentally different from keyword search.

Traditional search matches words. 

Type "VPN outage" and it looks for those exact words. If someone wrote "the tunnel went down" instead, you get nothing. 

This works for filing cabinets. It does not work for the way people actually communicate in critical and high-stakes situations. 

Here’s how Intelligent Search works: 

01 Vector embeddings

Messages are converted into numerical representations where meaning is encoded as position in a high-dimensional space. Similar concepts end up near each other, regardless of exact wording.

02 Intent matching

"Authentication service is down" and "users can't log in" land near each other in this space. A plain-language question surfaces conversations that mean the same thing, not just conversations that use the same words.

03 Retrieval only

The system retrieves content that has been indexed. It surfaces what exists in the corpus. It does not generate, invent, or synthesize. When a result feels off, it is because the closest match in the corpus genuinely was not close. The system returned the best available answer. It did not invent one.

"The semantic search model does not hallucinate results. Everything it surfaces comes from your actual message history. The challenge at scale is not accuracy. It is ranking. As the corpus grows, more content competes for relevance, and surfacing the best match first becomes harder. That is the engineering problem we set out to measure."

Devanshu Sharma, Lead Research Engineer, Rocket.Chat R&D

Why scale makes it harder

Search quality is not just a technology problem. It is a scale problem.

A system that works well at 50,000 documents can quietly struggle at 500,000. 

Most vendors benchmark on small, clean datasets and extrapolate. We wanted to know what actually happens when you push past a million documents. So we built a benchmark designed to feel like actual operational conditions.

Our R&D team built a benchmark designed to stress Intelligent Search the way actual operations would. The corpus combines a large public conversational dataset with actual Rocket.Chat message exports, totaling 1,198,202 indexed documents. This is heterogeneous, conversational, real-world data across multiple domains. Not a clean dataset built to produce favorable numbers.

We evaluated retrieval quality using 1,395 labeled queries across seven domains, with ground truth defined per query. The benchmark dataset is publicly available on HuggingFace so anyone can download it and reproduce our results. The transparency is deliberate.

As for the embedding model, we used all-MiniLM-L6-v2, a compact, fast encoder as an honest baseline. The goal was to measure what the system actually does under real conditions. A stronger embedding model would likely push MAP toward 0.5 to 0.7. We didn't swap models to improve the number before publishing. This is a baseline. Not a demo.

“We wanted the benchmark to feel like a real operational environment, not a controlled experiment. That means noisy data, varied domains, and messages that look like things people actually send. Not queries designed to make the system look good.”

Devanshu Sharma, Lead Research Engineer, Rocket.Chat R&D

What we found: The honest numbers

We are going to be more candid than most vendors would be. Here is what the system did.

Under sustained load, with concurrent ingestion and query traffic at million-document scale:

  • API response times stayed in the low single-digit milliseconds at the 95th percentile
  • Error rates remained near baseline with no sustained bursts
  • Worker failures were near zero
  • The system absorbed traffic spikes through backpressure and batching without degradation

For defense organizations, government agencies, and critical infrastructure operators, this matters more than any retrieval benchmark. 

A communication platform is mission-essential infrastructure. An AI feature that introduces latency spikes or instability into that stack is a liability, not a feature.

Intelligent Search does not compromise the operational envelope that teams depend on.

At around 380,000 documents, the system delivered strong results: a Mean Reciprocal Rank of 0.72, meaning the first relevant result was typically near the top of the list.

At the full 1.2 million document scale, that metric shifted to 0.56, still finding the right content, but with the top-ranked result sitting a bit further down the list on average.

This is expected behavior. 

As the corpus tripled in size, the search space grew and more content competed for the top result. The system consistently retrieved the correct content. The challenge at 1.2 million documents was consistently ranking the best match first. This is expected behavior, and it gives us a precise target to improve against.

"A MRR of 0.56 over 1.2 million real-world conversational documents is a credible baseline, especially given the messiness and scale of the corpus. More importantly, it gives us a reproducible reference point for improvement. From here, changes such as stronger embedding models, hybrid retrieval, or reranking pipelines can be evaluated against the same baseline rather than judged by intuition or a smaller, cleaner test set."

- Devanshu Sharma, Lead Research Engineer, Rocket.Chat R&D

3 things worth paying attention to

First: the transparency itself is the point. 

The organizations we serve operate under a higher standard of accountability. They do not adopt technology based on marketing demos. They need evidence. We are publishing real benchmarks against real data because trust in AI tooling should be earned with data, not slide decks. For teams that go through rigorous procurement and accreditation processes, this is the kind of evidence that matters.

Second: the operational stability story is underrated. 

Most conversations about AI search focus on retrieval quality. But for teams running communication infrastructure that supports mission-critical operations, the question is not just "does it find the right thing?" It is "does adding this feature introduce risk into the systems my people depend on?" Low single-digit millisecond latency and near-zero error rates at million-document scale is our answer.

Third: this is a baseline, not a finish line. 

We are sharing these results precisely because they represent the starting point for Intelligent Search. The benchmark gives us, and our community, a rigorous foundation for measuring every improvement we make from here. Better embedding models, hybrid retrieval strategies, reranking pipelines, all of these optimizations can now be evaluated against a real, public baseline. That is how engineering-led improvement works: you measure, you share, you iterate.

Why it matters

If you are evaluating AI-powered communication platforms: ask your vendors to show their benchmarks. Not a demo on a hundred documents. A test against a million. Not a controlled query on clean data, but a realistic mix of incident reports, operational threads, and cross-team coordination. 

And ask whether adding AI search changes the operational profile of the system your people depend on every day.

We have done that work. The dataset is public. Anyone can check our work. And there is a lot more to come.

The full technical paper, including detailed metric tables and Grafana telemetry snapshots, is available from our R&D team. The benchmark dataset is published at huggingface.co/datasets/dnouv/intelligent-search-benchmark.

This is Rocket.Chat Labs. There is more coming. 

Intelligent Search is just the beginning. We have also been building native federation based on the Matrix protocol, and we plan to share that story next: the technical decisions, the trade-offs, and why we believe we have built the most complete sovereign communications platform in the market. 

That thread connects everything we are doing. Whether it is intelligent search, federated messaging, or the next capability we ship, the goal stays the same: give organizations full control over their communications, without compromises.

Rocket.Chat Labs is a series exploring how we build, test, and evolve the Rocket.Chat Secure CommsOS. From AI capabilities to native federation and beyond, we are opening up the engineering behind the platform. Follow along for there is a lot more in the kitchen.

Frequently asked questions about <anything>

Pavithra is a Product Marketing Manager at Rocket.Chat. She represents the voice of the customers and helps shape the voice of the product. She is highly passionate about bringing new offerings to the market. When she isn’t donning the hat of a Product Marketer, she tries her hand at multiple cuisines, lives a hundred different fictional characters through books, and enjoys playing badminton.
Pavithra Sudhakar
Related Article:
Team collaboration: 5 reasons to improve it and 6 ways to master it
Want to collaborate securely with your team?
Deploy Rocket.Chat on-premise or in the cloud and keep your conversations private.
  • Digital sovereignty
  • Federation capabilities
  • Scalable and white-labeled
Talk to sales
Looking for a HIPAA-ready communications platform?
Enable patients and healthcare providers to securely communicate without exposing their data.
  • Highly scalable and secure
  • Full patient conversation history
  • HIPAA-ready
Talk to sales
Secure communication
for mission-critical operations
Built to operate securely in the most restricted environments.
  • On-premise and air-gapped ready
  • Full control over sensitive data
  • Secure cross-agency collaboration
Talk to sales
Talk to sales
Want to customize Rocket.Chat according to your own preferences?
See behind the engine and change the code how you see fit.
  • Open source code
  • Highly secure and scalable
  • Unmatched flexibility
Talk to sales
Looking for a secure collaboration platform?
Keep your conversations private while enjoying a seamless collaboration experience with Rocket.Chat.
  • End-to-end encryption
  • Cloud or on-prem deployment
  • Supports compliance with HIPAA, GDPR, FINRA, and more
Talk to sales
Want to build a highly secure in-app chat experience?
Use Rocket.Chat’s APIs, frameworks, and managed backend to build a secure in-app or live chat experience for your customers.
  • Supports compliance with HIPAA, GDPR, FINRA, and more
  • Highly secure and flexible
  • On-prem or cloud deployment
Talk to sales

Relevant articles

See All
No items found.

Our best content, once a week

Share this on:
White house icon with rounded edges on a dark circle background, representing a home or homepage button.
Man with glasses in a video call interface and a blurred chat message with a lock icon indicating secure or encrypted communication.

Get your free, personalized demo now!

Build the most secure chat experience for your team or customers

Book demo
White house icon with rounded edges on a dark circle background, representing a home or homepage button.
Chat conversation showing Maj. Carter sharing a patrol route PDF, Sgt. Alvarez sending a voice confirmation audio message, and Maj. Carter starting a secure video call, with security icons for key and lock.

Get your free demo now!

Tailored to your security, deployment, and compliance needs.

Talk to salesTalk to sales