Understand how Synaptiq retrieves and uses your uploaded documents to generate accurate, grounded answers.
When a customer asks your Synaptiq AI agent a question, the answer does not come from the AI's general training data. It comes from the specific documents you uploaded to your knowledge base. This page explains how that works so you can structure your content for the best possible results.
This approach is called Retrieval-Augmented Generation (RAG), and it is the core of what makes Synaptiq accurate and trustworthy instead of creative and unreliable.
Large language models are trained on vast amounts of text from the internet, but they do not know anything about your specific products, pricing, or policies. Left on their own, they will either make things up or give generic answers.
RAG solves this by adding a retrieval step before generation:
Think of it like the difference between asking someone to answer from memory versus handing them the reference manual and asking them to look it up first. The second approach is far more reliable.
When you upload a document, Synaptiq does not store it as a single blob. It goes through a multi-step processing pipeline.
The raw text is extracted from your document. For PDFs, this means parsing the PDF structure. For DOCX files, it means reading the underlying XML. For TXT and MD files, the content is used as-is.
The extracted text is split into smaller sections called chunks. Chunking is essential because when a customer asks a question, the AI does not need your entire 30-page product guide. It needs the two or three paragraphs that are actually relevant.
Synaptiq chunks your documents intelligently:
This is why document formatting matters. A well-structured document with clear headings produces clean, topic-focused chunks. A wall of unformatted text produces chunks with mixed topics, leading to less precise retrieval.
Each chunk is converted into a vector embedding, which is a numerical representation of the chunk's meaning. Two chunks about similar topics will have similar embeddings, even if they use different words.
For example, a chunk about "monthly subscription cost" and a customer question about "how much does it cost per month" will have very similar embeddings, allowing the system to match them even though the exact wording differs.
These embeddings are stored in a vector database that enables fast similarity searches across all your documents.
Here is the step-by-step flow when a real conversation happens:
The customer's question is converted into the same type of vector embedding used for your document chunks. This puts the question and your content into the same mathematical space where similarity can be measured.
Synaptiq searches your vector database for the chunks whose embeddings are most similar to the question's embedding. It retrieves the top matching chunks, typically between 3 and 8 depending on how much relevant content exists.
Each retrieved chunk receives a relevance score between 0 and 1. Chunks that closely match the question's intent score higher. Chunks that are only tangentially related score lower.
Only chunks above a minimum relevance threshold are passed to the AI. This prevents low-quality matches from polluting the answer.
The qualifying chunks are assembled into a context window along with the conversation history and the customer's current question. The AI receives all of this as input.
The AI generates a response using the retrieved context. It is instructed to base its answer on the provided documents and to cite which documents it drew from.
Every answer Synaptiq generates carries an internal confidence score. This score reflects how well the retrieved document chunks matched the question and how much of the answer is grounded in the source material.
You can see confidence indicators in the test interface when evaluating your knowledge base coverage.
Synaptiq is designed to say "I don't know" rather than fabricate an answer. This happens when:
When the AI does not know the answer, it acknowledges this honestly and offers to connect the customer with a human team member or suggests they rephrase their question.
This is a feature, not a bug. A sales AI that never says "I don't know" is one that sometimes lies. Synaptiq prioritizes trust.
If the AI is saying "I don't know" too often, the fix is almost always to add more content to your knowledge base:
Synaptiq uses a strict priority hierarchy:
This means keeping your knowledge base current is critical. The AI trusts your documents above all other sources of information. Outdated documents lead to outdated answers.
Understanding how RAG works has direct consequences for how you should write and organize your documents:
Was this page helpful?