Knowledge
Enable context-aware responses with RAG-powered retrieval.Overview
Knowledge connects your agents to your data sources, enabling Retrieval-Augmented Generation (RAG). Instead of relying solely on the LLM’s training data, agents can search your knowledge bases and incorporate relevant information into responses.How It Works
The RAG Pipeline
Key Capabilities
Extensive Connectors
Connect to 100+ data sources:- Cloud Storage: Google Drive, OneDrive, Dropbox
- Collaboration: Confluence, SharePoint, Notion
- Business Apps: Salesforce, Jira, HubSpot
- Communication: Slack, Microsoft Teams
- Databases: SQL, NoSQL, custom APIs
Intelligent Search
- Hybrid search: Combine keyword and semantic matching
- Multi-vector: Weight different embedding strategies
- Programmable extraction: Custom chunking pipelines
Enterprise Security
- Access control integration
- Permission-aware retrieval
- Audit logging
Setting Up Knowledge
Step 1: Create a Search AI Application
- Navigate to Knowledge in your app
- Click + Connect Knowledge
- Select Create new Search AI app or link existing
Step 2: Add Data Sources
Choose your content sources:Step 3: Configure Ingestion
Set up how content is processed:Step 4: Link to Agent
- Open your agent configuration
- Navigate to Knowledge
- Select the Search AI application
- Configure retrieval settings
Three-Stage Workflow
1. Ingestion
Bring content into the knowledge base. Methods:- Web crawling
- Directory indexing
- Third-party connectors
- File uploads
- API integrations
- Content extraction
- Format normalization
- Metadata enrichment
2. Enhancement & Embedding
Prepare content for semantic search. Chunking: Split documents into searchable segments3. Retrieval
Fetch relevant content at query time. Strategies:- Vector similarity search
- Hybrid (vector + keyword)
- Reranking for relevance
Retrieval Configuration
Basic Settings
Advanced Options
Answer Generation
Extractive
Pull exact quotes from source documents.Generative
Synthesize answers using retrieved context.Knowledge in Agent Prompts
Retrieved knowledge is injected into agent context:Best Practices
Quality Content
- Keep documentation up-to-date
- Use clear, structured writing
- Include relevant metadata
Appropriate Chunking
| Content Type | Strategy | Chunk Size |
|---|---|---|
| FAQs | Question-answer pairs | Small (200-300 tokens) |
| Documentation | Sections/headers | Medium (400-600 tokens) |
| Long articles | Semantic breaks | Variable |
Retrieval Tuning
- Start with top_k=5, adjust based on results
- Use score thresholds to filter low-quality matches
- Enable reranking for better precision
Grounding Instructions
Tell agents how to use knowledge:Monitoring
Track knowledge retrieval performance:- Hit rate: How often relevant content is found
- Latency: Search response times
- Relevance: User satisfaction with answers
- Coverage: Questions without good matches