Download the latest version of the CrateDB Architecture Guide

Download Now
Skip to content
Use cases

AI-Ready Analytics & Vector Search

CrateDB provides a real-time analytics foundation that feeds AI and machine learning workflows. It supports vector search alongside traditional SQL analytics, enabling similarity search, hybrid queries, and AI-driven applications on fresh, high-volume data, without introducing additional system.

AI and machine learning applications depend on fast access to large volumes of relevant data. As models move closer to production, they require real-time context, hybrid queries, and the ability to combine analytical data with vector embeddings. AI-ready analytics and vector search make it possible to enrich AI workflows with fresh data and power intelligent applications that reason over both similarity and structure.

Analytics as the Foundation for AI

CrateDB delivers fast, scalable analytics on the data that feeds AI and machine learning systems. This ensures models operate on current, high-quality information rather than stale or batch-processed data.

Vector Search with Context

Vector embeddings can be queried alongside structured and semi-structured data. Similarity search combines naturally with filters, aggregations, and time-based analysis, enabling richer and more precise AI-driven queries.
cr-quote-image

Real-Time Data for Intelligent Applications

Fresh data flows directly into AI-powered applications, supporting use cases such as recommendations, anomaly detection, and semantic search. Models can react to changing conditions as they happen.
cr-quote-image

One Platform, Hybrid Workloads

CrateDB unifies traditional analytics and vector search in a single system. This reduces architectural complexity while supporting advanced AI use cases that span analytics, search, and machine learning.
cr-quote-image

Where Traditional Systems Fall Short

Modern AI and ML workloads demand real-time access to structured, semi-structured, unstructured, and vector data. Traditional architectures often can't keep up:

  • Vector databases lack full SQL analytics, aggregations, and joins.
  • OLAP or search engines can’t natively store or query high-dimensional embeddings.
  • Multiple systems introduce latency, duplicated data, and operational overhead.
The result is slower model iteration, delayed insights, and complex ETL pipelines moving data between disconnected platforms.

Is CrateDB the Right Database for Your Workloads?

Answer 10 questions. Get your score. No email required.

Talk
Unlocking the Power of Semantic Search
Unlocking the Power of Semantic Search

Unlock the power of semantic search by watching this ingsightful webinar where Simon Prickett, Senior Product Evangelist CrateDB, highlights CrateDB's ability to integrate various data types (text, geospatial, vectors) for hybrid search using SQL, enabling faster, more contextually relevant results.

Webinar
From-Documents-to-Dialogue-Unlocking-PDF-Data-with-a-Smart-Chatbot
From Documents to Dialogue: Unlocking PDF Data with a Smart Chatbot

In this webinar recording Simon Prickett reveals how to unlock text and image data trapped in PDF files and search it using the power of AI and CrateDB.

Website
semantic search
Vector search with CrateDB

CrateDB maximizes the potential of vector data with a single, scalable database that can be queried with SQL and streamlines data management, significantly reducing development time and total cost of ownership.

Documentation
icon_documentation
Search

Based on Apache Lucene, CrateDB offers native BM25 term search and vector search, all using SQL. By combining it, also using SQL, you can implement powerful single-query hybrid search.

Webinar
Recommendation for Pickcenter
Faster Fixes, Better Outcomes: How AI Empowers Operators on the Shop Floor

In this webinar recording, TGW Logistics shows you how combining digital twins with generative AI can help you solve real-world operational challenges—like reducing downtime, streamlining maintenance, and empowering your teams with instant access to the knowledge they need. 

White Paper
DATA ENGINEERING ESSENTIALS FOR THE AI ERA
Data Engineering Essentials for the AI era

Download this report to discover how to build a future-proof data backbone for real-time AI success. 

Webinar
The OEE Whisperer Demo
The OEE Whisperer

Meet “The OEE Whisperer” – a groundbreaking AI-powered voice assistant built to transform your factory floor. Speak to your factory in plain language and get real-time, predictive insights instantly. 

Ebook
Context Data
Unlocking the Power of Knowledge Assistants with CrateDB

As a cutting-edge real-time analytics database, CrateDB provides the foundation for building chatbots and knowledge assistants that are not only fast and reliable but also intelligent and scalable. 

Demo
Building-an-AI-Chatbot-with-CrateDB-and-LangChain-play
Building an AI Chatbot with CrateDB and LangChain

This video shows step by step how to build an AI-powered chatbot using LangChain to connect to the different LLMs and CrateDB to store embeddings and run similarity searches against them.

Additional resources

Want to know more?

FAQ

Vector search finds results based on semantic similarity rather than exact matches, comparing vector embeddings that capture the meaning of data (e.g., text, documents, images, videos).

CrateDB supports vector embeddings natively and allows similarity queries (e.g. k-nearest neighbors), combined with filters, aggregations, and full-text or time-series data, all in SQL.

Full-text or keyword search (e.g. BM25) matches exact or approximate lexical similarity (words, phrases). Vector search matches semantic similarity (e.g. meaning, context, concepts, embeddings). CrateDB allows hybrid search, combining vector and full-text search in the same query.

Yes. CrateDB is built to scale: high ingestion rates, storage of structured + vector + unstructured data, and efficient querying even with large embedding volumes.

CrateDB’s value proposition is to unify these: instead of separate vector stores, search engines, and analytics/OLAP systems, you can do embeddings + filtering + aggregations + full-text + search in one system. This reduces latency, complexity, and operational overhead.

Examples include: semantic search over documents, recommendation systems (matching embeddings), anomaly detection in real time, combining text search + vector similarity (hybrid search), powering chatbots or AI features that need fast access to embeddings + metadata + analytics together. 

RAG Pipelines, short for Retrieval Augmented Generation Pipelines, are a crucial component of generative AI, that combines the vast knowledge of large language models (LLMs) with the specific context of your private data.

A RAG Pipeline works by breaking down your data (text, PDFs, images, etc.) into smaller chunks, creating a unique "fingerprint" for each chunk called an embedding, and storing these embeddings in a database. When you ask a question, the system identifies the most relevant chunks based on your query and feeds this information to the LLM, ensuring accurate and context-aware answers. They operate through a streamlined process involving data preparation, data retrieval, and response generation.

  1. Phase 1: Data Preparation
    During the data preparation phase, raw data such as text, audio, etc., is extracted and divided into smaller chunks. These chunks are then translated into embeddings and stored in a vector database. It is important to store the chunks and their metadata together with the embeddings in order to reference back to the actual source of information in the retrieval phase.

  2. Phase 2: Data Retrieval
    The retrieval phase is initiated by a user prompt or question. An embedding of this prompt is created and used to search for the most similar pieces of content in the vector database. The relevant data extracted from the source data is used as context, along with the original question, for the Large Language Model (LLM) to generate a response.


Retrieval augmented generation (RAG) Pipeline

While this is a simplified representation of the process, the real-world implementation involves more intricate steps. Questions such as how to properly chunk and extract information from sources like PDF files or documentation and how to define and measure relevance for re-ranking results are part of broader considerations.