Read the Report — State of Applied AI →
TechnologyProduct Development

How Allspice Improved Ingredient Matching from 20% to 97% with Pinecone

Allspice, a food technology startup building a kitchen operating system for consumers and recipe publishers, deployed Pinecone’s vector database to solve the inherent messiness of ingredient data that traditional text search could not handle. The implementation raised ingredient matching accuracy from roughly 20% to 97%, enabling the launch of recipe importing as a core product feature and expanding into a platform-wide semantic layer for search, recommendations, and conversational AI.

Impact

20% → 97%

Ingredient matching accuracy

1 afternoon

Time to validate pipeline

110,000

Total embeddings managed

Challenge

Allspice’s recipe importing pipeline could not reach production because traditional text search failed to handle the inherent messiness of ingredient data — variations in phrasing, modifiers, and spelling kept matching accuracy at roughly 20% — while bolt-on vector capabilities in their existing search stack degraded overall system performance when storing large embeddings.

Solution

Allspice deployed Pinecone as a dedicated vector database, embedding its 10,000-entry ingredient database with OpenAI’s text-embedding-3-large model and building a fully decoupled semantic layer that scales independently — then expanded it across ingredient matching, recipe similarity, fuzzy search, chatbot normalization, and FAQ retrieval across the entire platform.

Tools & Technologies

What Leaders Say

The benefit of vectors has always been flexibility. Instead of carefully managing search params in Typesense, trying to balance always receiving a result with only receiving relevant results, Pinecone removes all that complexity with a simple query.

William Templeton, Co-founder and CTO, Allspice

Pinecone helps bridge a fundamental gap in modern AI systems between strictly structured data types and unstructured natural language input. It provides a semantic layer between those two worlds, allowing us to measure similarity and meaning without requiring exact matches or rigid schemas.

William Templeton, Co-founder and CTO, Allspice

I would have never tried Pinecone without a cloud-hosted, serverless option. I needed something that I could set up in an afternoon and get working in a basic pipeline to evaluate the effectiveness of my solution to my problems.

William Templeton, Co-founder and CTO, Allspice
Get the full context.

Sign up to read complete case studies, access detailed metrics, and unlock all use cases.

Full Story

Allspice is a food technology company building a comprehensive kitchen operating system that serves both consumers and recipe publishers. On the consumer side, the platform helps home cooks discover recipes, manage pantry inventory, and generate automated shopping lists. For publishers, Allspice provides interactive tools that increase engagement and unlock revenue streams beyond traditional display advertising. At the center of both experiences is the ability to understand food the way a human does — recognizing that “one bunch of cilantro” and “fresh cilantro, chopped” are the same ingredient despite entirely different wording.

As Allspice expanded recipe importing into a primary product feature, it hit a fundamental wall. Ingredient data is inherently messy: modifier-heavy phrases, spelling variations, parsing inconsistencies, and informal descriptions make deterministic matching unreliable. The team had built on a NoSQL stack using Google Cloud Firestore and Typesense, which worked well for traditional structured search but could not bridge the semantic gap between how ingredients appear in source recipes and how they are stored in Allspice’s proprietary database. Attempting to use Typesense’s vector capabilities created additional problems — storing large embeddings alongside small documents degraded overall system performance. With ingredient matching sitting at roughly 20% accuracy, recipe importing was not viable as a product feature.

Allspice turned to Pinecone as a dedicated, purpose-built vector database. The team embedded their proprietary ingredient database of approximately 10,000 entries using OpenAI’s text-embedding-3-large model and immediately validated the approach. Pinecone’s serverless, managed infrastructure kept semantic search fully decoupled from the rest of the stack, allowing vector workloads to scale independently without rearchitecting existing layers. What began as a single ingredient-matching pipeline expanded iteratively: recipe-level embeddings enabled a “more recipes like this” feature, fuzzy recipe search added flexible retrieval alongside structured filtering, and Pinecone-hosted Llama embeddings supported chatbot data normalization by mapping free-form user inputs to structured internal representations. The platform now manages approximately 110,000 total recipe and ingredient embeddings.

The results transformed the product. Ingredient matching accuracy jumped from 20% to 97%, moving the recipe importing pipeline from unusable to production-ready. With that foundation in place, Allspice was able to launch one of its core features — a capability that had been blocked entirely before the vector layer existed. The team set up Pinecone and validated a working pipeline in a single afternoon, a speed that was essential for a startup where idea validation determines what ships. User satisfaction with recipe search improved significantly, and publishers gained new engagement surfaces that increased time on site and enabled revenue from grocery exports, subscriptions, and affiliate commissions.

Allspice is now expanding Pinecone’s role into its AI and conversational systems. Planned work includes FAQ classification and retrieval to match user questions against publisher-approved content, and broader use of vector retrieval to reduce LLM token usage as chatbot query volume grows. The stack — Pinecone for semantic retrieval, Gemini 2.5 and GPT-4.1-mini for generation, and Firestore for document storage — keeps Allspice model-agnostic while grounding its AI features in reliable, high-accuracy retrieval.

Similar Cases

1
1up
10x faster
response generation speed for rfps and compliance questionnaires

1up, a sales knowledge automation platform, integrated Pinecone's vector database to power a RAG-based system that delivers real-time, highly accurate answers to complex sales queries. The solution replaced a slow, home-grown embedding system and achieved 10x faster response generation for RFPs and compliance questionnaires. Sales reps can now handle high volumes of queries with confidence, reducing reliance on colleagues and accelerating the go-to-market process.

TechnologyAAWSPPinecone
C
CustomGPT.ai
>400M
vectors stored

CustomGPT.ai built a RAG-as-a-Service platform on Pinecone storing over 400M vectors, achieving sub-20ms query latency and the #1 ranking in an independent RAG accuracy benchmark.

TechnologyPPinecone
D
Delphi
>100M
vectors stored

Delphi is an AI platform that enables coaches, creators, and experts to deploy interactive “Digital Minds”—always-on conversational agents trained on their unique content. Scaling from proof of concept to a commercial platform with thousands of customers required a vector database that could support millions of isolated namespaces, billions of vectors, and sub-second retrieval under variable load. Delphi selected Pinecone, achieving P95 query latency of 100ms and keeping retrieval under 30% of total response time—freeing the engineering team to build product rather than manage infrastructure.

TechnologyPPinecone
TX
Terminal X
0.68 to 0.91
f1 retrieval accuracy improvement

Terminal X is a vertical AI platform for institutional investors that acts as a 24/7 research agent, processing millions of financial documents for hedge funds, asset managers, and private equity firms. By rebuilding its retrieval architecture on Pinecone’s vector database, Terminal X improved F1 retrieval accuracy from 0.68 to 0.91, cut average latency by over 35%, and doubled deployment velocity. Users now save approximately three hours per day, and investment memo preparation dropped from two days to half a day.

Financial ServicesTechnologyPPinecone
Z
ZoomInfo
>50%
increase in user engagement

ZoomInfo, a B2B go-to-market intelligence platform with hundreds of millions of professional contact records, needed a vector database to power real-time personalized contact recommendations for sales and marketing teams. The company deployed Pinecone’s serverless vector database with Dedicated Read Nodes to run semantic search over 390 million contact embeddings with sub-second latency. The result was a 50% increase in user engagement, a 2x improvement in recommendation relevancy, and 50x more peak request capacity.

TechnologyPPinecone
A
Assembled
~95%
ticket handling time reduction

Assembled is a workforce management and customer support optimization platform serving enterprises like Stripe, Etsy, and DoorDash. To power Assembled Assist, the company built a hybrid RAG pipeline combining Pinecone vector search with Algolia keyword retrieval and LLMs from OpenAI and Anthropic. Support tasks that previously took 40 minutes now complete in 2 minutes—a 95% reduction in handling time.

TechnologyAAlgoliaOLOpenAI LLMs
G
Gong
10x
infrastructure cost reduction

Gong is a revenue intelligence platform that analyzes billions of customer interactions to help sales teams improve performance. To power Smart Trackers—its patented AI system for detecting and classifying concepts in sales conversations—Gong adopted Pinecone as its core vector database, storing billions of sentence-level embeddings across real conversations. Migrating to Pinecone Serverless delivered a 10x reduction in infrastructure costs while sustaining peak search performance across a massive corpus.

TechnologyPPinecone
P
Pfizer
93%
database reduction

Pfizer achieved a 93% database reduction and 20% cost avoidance by migrating their global SAP environment to S/4HANA on IBM Power10 infrastructure.

PharmaceuticalsTechnologyICIBM ConsultingIPIBM Power Virtual Server