Prerequisites
An Embeddings Model must be configured in Settings before you can use the Knowledge Base. If you don’t see an embeddings selector here, add one first.Uploading documents
Drag and drop files into the upload area or click to browse. Uploaded documents are:- Processed and chunked into passages
- Embedded using the selected embeddings model
- Stored in the vector database (Qdrant)
Retrieval parameters
These parameters control how the agent searches your documents when answering a question.Search type
| Type | Description | Best for |
|---|---|---|
similarity | Pure semantic similarity — finds the most relevant chunks | Specific, focused questions |
mmr | Maximal Marginal Relevance — balances relevance with diversity | Broad questions, avoiding repetitive passages |
Parameters
| Parameter | Default | Description |
|---|---|---|
| K | 4 | Number of document chunks retrieved per query |
| Score Threshold | none | Minimum similarity score (0.0–1.0). Chunks below this are excluded. |
| Fetch K | — | MMR only: candidates to evaluate before selecting the final K |
| Lambda Mult | — | MMR only: diversity weight (0 = max diversity, 1 = max relevance) |
How retrieval works at runtime
When the agent receives a message:- The query is embedded using the same embeddings model
- The vector database finds the K most similar document chunks
- Those chunks are injected into the agent’s prompt as context
- The LLM uses this context to generate a grounded answer
Next steps
Embeddings Models
Configure the embeddings model required for the knowledge base.
Memory
Add conversation memory alongside document retrieval.