Search AI
Conversational search with LLM and generative AI capabilities.Overview
Search AI enables natural language search across your enterprise content:- Connect 50+ data sources
- Retrieve relevant information with semantic search
- Generate accurate, contextual answers
- Cite sources for transparency
Architecture
Content Sources
Supported Connectors
| Category | Sources |
|---|---|
| Documents | PDF, Word, Excel, PowerPoint, TXT, Markdown |
| Knowledge bases | Confluence, SharePoint, Notion, Guru |
| Help centers | Zendesk Guide, Intercom, Freshdesk, ServiceNow |
| Websites | Web crawling, sitemap indexing, RSS feeds |
| Cloud storage | Google Drive, Dropbox, Box, OneDrive, S3 |
| Databases | SQL (PostgreSQL, MySQL), MongoDB, Elasticsearch |
| Code | GitHub, GitLab, Bitbucket repositories |
| Custom | REST API, file upload, webhooks |
Adding a Content Source
- Navigate to Search AI → Sources
- Click Add Source
- Select connector type
- Configure authentication
- Set sync schedule
- Map content structure
Connector Configuration
Confluence
SharePoint
Web Crawling
Custom API
Index Configuration
Chunking Strategies
How content is split affects retrieval quality.| Strategy | Best For | Chunk Size |
|---|---|---|
| Fixed size | General documents | 500-1000 tokens |
| Paragraph | Well-structured articles | Natural breaks |
| Semantic | Mixed content | Variable |
| Heading-based | Documentation with sections | Per section |
| Sentence | FAQs, Q&A content | 1-3 sentences |
Chunking Configuration
Embedding Models
Select the model that generates vector representations.| Model | Provider | Dimensions | Languages | Best For |
|---|---|---|---|---|
text-embedding-3-small | OpenAI | 1536 | Multi | Cost-effective |
text-embedding-3-large | OpenAI | 3072 | Multi | High accuracy |
embed-english-v3.0 | Cohere | 1024 | English | English content |
embed-multilingual-v3.0 | Cohere | 1024 | 100+ | International |
bge-large-en-v1.5 | Open source | 1024 | English | Self-hosted |
multilingual-e5-large | Open source | 1024 | Multi | Self-hosted |
Embedding Configuration
Index Settings
Query Processing
Query Understanding Pipeline
Query Configuration
Retrieval Strategies
Hybrid Search
Combine semantic and keyword search for best results.Re-ranking
Improve result quality with a second-stage ranker.Contextual Retrieval
Include surrounding context for better answers.Filtering
Apply metadata filters to narrow results.Answer Generation
Generation Configuration
Response Modes
| Mode | Description | Use Case |
|---|---|---|
| Generative | LLM synthesizes answer from sources | Complex questions |
| Extractive | Returns relevant passages directly | Simple lookups |
| Hybrid | Generated answer + source excerpts | Detailed responses |
| Direct | Returns top chunk if high confidence | FAQ-style |
Fallback Handling
Workbench
Testing Queries
- Navigate to Search AI → Workbench
- Enter test queries
- View retrieved chunks with scores
- Review generated answers
- Iterate on configuration
Debug Panel
| Tab | Information |
|---|---|
| Query | Processed query, expansions, filters |
| Retrieval | All retrieved chunks with scores |
| Reranking | Before/after reranking comparison |
| Generation | Prompt sent to LLM, token usage |
| Sources | Final sources used in answer |
Quality Metrics
| Metric | Description |
|---|---|
| Retrieval precision | % of retrieved chunks that are relevant |
| Answer accuracy | Manual or LLM-judged correctness |
| Citation accuracy | % of claims properly cited |
| Latency | End-to-end response time |
| Fallback rate | % of queries with no/low-confidence answers |
Integration
API Access
Virtual Assistant Integration
Embed Search AI in Automation AI:Widget Embed
Best Practices
Content Quality
- Keep source documents up to date
- Use clear headings and structure
- Remove duplicate content
- Maintain consistent terminology
- Add metadata for better filtering
Retrieval Tuning
- Start with hybrid search (70/30 split)
- Use reranking for better precision
- Set appropriate similarity thresholds
- Test with real user queries
- Monitor retrieval metrics
Answer Generation
- Provide clear system instructions
- Require source citations
- Handle knowledge gaps gracefully
- Test edge cases
- Monitor answer quality
Related
- Automation AI — Virtual assistant integration
- Agent AI — Agent knowledge assist
- Integrations — Enterprise connections
- Getting Started — Initial setup