How Gong Achieves 10x Cost Savings with Pinecone Serverless for Smart Trackers

Gong is a revenue intelligence platform that analyzes billions of customer interactions to help sales teams improve performance. To power Smart Trackers—its patented AI system for detecting and classifying concepts in sales conversations—Gong adopted Pinecone as its core vector database, storing billions of sentence-level embeddings across real conversations. Migrating to Pinecone Serverless delivered a 10x reduction in infrastructure costs while sustaining peak search performance across a massive corpus.

Impact

10x

Infrastructure cost reduction

Billions

Vectors stored

Challenge

Gong needed to track and classify complex, contextually varied concepts across billions of conversation sentences without relying on brittle keyword matching—and to do so at a cost and latency that made real-time user interaction feasible.

Solution

Gong deployed Pinecone as the vector database for Smart Trackers, storing billions of sentence embeddings to enable semantic retrieval that powers an active learning loop—then migrated to Pinecone Serverless to achieve 10x cost reduction at equivalent performance.

Tools & Technologies

What Leaders Say

Users want to track different concepts that occur in conversations, and simple keywords do not work.

Jacob Eckel, VP, R&D Division Manager, Gong

Our choice to work with Pinecone wasn’t just based on technology; it was rooted in their commitment to our success. They listened, understood, and delivered beyond our expectations.

Jacob Eckel, VP, R&D Division Manager, Gong
Get the full story.

Sign up to read complete case studies, access detailed metrics, and unlock all use cases.

Full Story

Gong has built one of the richest repositories of sales conversation data in the world, analyzing calls, emails, and meetings to surface coaching insights and deal intelligence. As early as 2020, Gong was among the first revenue platforms to deploy a vector database for semantic search at scale—recognizing that keyword matching couldn’t capture the nuance and contextual variation in human conversation.

The core challenge was conceptual tracking at enterprise scale. Gong’s Smart Trackers system allows sales managers to define concepts—like a competitor mention or a pricing objection—by providing a handful of example sentences. The system then automatically retrieves semantically similar examples from billions of real conversation records to train a live classification model. Traditional keyword search fails here: conversations are dynamic, and the same concept surfaces in countless phrasings. Only vector similarity search could reliably find them.

Gong selected Pinecone as its vector database partner, storing billions of sentence-level embeddings and enabling the fast similarity searches that power Smart Trackers’ active learning loop. Users label a few sentences, the system retrieves semantically similar candidates from the corpus, and the model improves continuously through feedback. When Pinecone introduced its serverless architecture, Gong migrated and achieved a 10x cost reduction while maintaining performance at peak loads.

The business impact is direct: Smart Trackers allows Gong users to build and fine-tune concept-detection models without any technical expertise. Sales managers, not data scientists, define what matters. The system’s accuracy and speed—sustained by Pinecone’s retrieval infrastructure—make this self-service model practical at scale.

Gong’s architecture demonstrates a maturing pattern in AI product development: managed vector databases as the retrieval backbone for active learning systems. As the corpus of conversation data grows, so does the value of the underlying infrastructure—and the competitive moat built by Gong’s proprietary dataset.

Similar Cases

1
1up
10x faster
response generation speed for rfps and compliance questionnaires

1up, a sales knowledge automation platform, integrated Pinecone's vector database to power a RAG-based system that delivers real-time, highly accurate answers to complex sales queries. The solution replaced a slow, home-grown embedding system and achieved 10x faster response generation for RFPs and compliance questionnaires. Sales reps can now handle high volumes of queries with confidence, reducing reliance on colleagues and accelerating the go-to-market process.

Sales TechnologyAAWS
R(
RAG (Retrieval-Augmented Generation)
T
TaskUs
20%
average handle time reduction

TaskUs is a leading outsourced digital services company providing next-generation customer experience (CX) for innovative global brands. To move beyond flat-file embedding storage and scaling limitations, TaskUs built TaskGPT—a proprietary GenAI platform—with Pinecone as the core vector database for semantic search, RAG-based knowledge retrieval, and client-specific recommendations. The result: a 20% reduction in average handle time and a 5% increase in customer satisfaction across client deployments.

Business Process OutsourcingABAmazon BedrockPPinecone
D
Delphi
>100M
vectors stored

Delphi is an AI platform that enables coaches, creators, and experts to deploy interactive “Digital Minds”—always-on conversational agents trained on their unique content. Scaling from proof of concept to a commercial platform with thousands of customers required a vector database that could support millions of isolated namespaces, billions of vectors, and sub-second retrieval under variable load. Delphi selected Pinecone, achieving P95 query latency of 100ms and keeping retrieval under 30% of total response time—freeing the engineering team to build product rather than manage infrastructure.

TechnologyPPinecone
A
Assembled
~95%
ticket handling time reduction

Assembled is a workforce management and customer support optimization platform serving enterprises like Stripe, Etsy, and DoorDash. To power Assembled Assist, the company built a hybrid RAG pipeline combining Pinecone vector search with Algolia keyword retrieval and LLMs from OpenAI and Anthropic. Support tasks that previously took 40 minutes now complete in 2 minutes—a 95% reduction in handling time.

Customer Support TechnologyACAnthropic ClaudeAAlgolia
I
InpharmD
80%
data storage cost savings

InpharmD's AI assistant, Sherlock, leverages Pinecone's vector database to deliver fast, accurate drug information to healthcare professionals. By embedding 30 million medical documents into a RAG pipeline, InpharmD achieved 70% better query accuracy, 95x faster first response times, and 80% cost savings on data storage.

HealthcareSSherlockAAWS
CC
Chipper Cash
95%+
selfie verification accuracy

Chipper Cash, a fintech serving over five million customers across Africa, deployed a Pinecone-powered facial similarity search system to detect and block fraudulent duplicate sign-ups in real time. The solution slashed identity verification latency from up to 20 minutes down to under 2 seconds, and reduced fraudulent sign-ups by 10x across all markets.

Financial ServicesGCGoogle CloudSSnowflake