Text data refinement
This page allows you to configure text processing parameters that determine how documents are processed and stored for optimal retrieval and knowledge base creation.
Select the RAG definition
In this step, configure the data retrieval model that determines how your knowledge base will store, index, and retrieve content to generate accurate, context-aware responses to user queries.
ZBrain supports two RAG storage models:
Vector store – default, chunk-and-embedding index for semantic similarity search
Knowledge graph – entity–relationship graph stored in ZBrain’s graph database

Select the one that best matches your data and query needs.
Vector store
Splits each document into chunks, converts the chunks into high-dimensional embeddings, and saves them in a vector database. At query time, ZBrain performs a semantic-similarity search and supplies the matched chunks to the LLM.
Unstructured text
Rapid prototyping
Knowledge graph
Extracts entities and relationships from every chunk, stores them as nodes and edges, and also embeds the chunk text so you can fall back on vector similarity. The query engine can traverse the graph, run vector search, or perform both operations.
Information with critical inter-entity relationships, such as product-component hierarchies, chronological timelines, or complex organizational structures.
Next steps
To finish configuring your knowledge base, choose the retrieval model that best suits your data and follow the step-by-step guide to configure the knowledge base:
Knowledge Graph
Follow the appropriate guide, complete the remaining configuration steps, and publish your knowledge base to make it available for querying in ZBrain.
Last updated