Skip to main content
This guide covers the configuration of retrieval strategies, answer generation, and search results in Search AI. These settings determine how content is retrieved from your index and how responses are delivered to users. Navigation: Configuration menu in Search AI

Retrieval Strategies

Configure the chunk retrieval strategy and corresponding thresholds for finding relevant content.

Retrieval Methods

StrategyDescriptionBest For
Vector RetrievalUses cosine similarity between query vector and chunk vectors. Scores range from 0 (no match) to 1 (complete match)Semantic similarity matching, contextual queries
Hybrid RetrievalCombines keyword-based matching with vector-based scoring, leveraging strengths of both approachesBalanced precision and recall, improved accuracy

Qualification Criteria

ParameterDescriptionRangeDefault
Similarity Score ThresholdMinimum similarity score for a chunk to be considered relevant. Higher values require more similarity0-10020
Proximity ThresholdHow closely retrieved chunks should be to the highest-ranking chunk. Lower values mean closer chunks0-5020
Top K ChunksMaximum number of qualified chunks for answer generation1-10020
Token Budget for ChunksMaximum tokens to send to LLM. Must account for prompt, query, context, and response within LLM’s context window1-1,000,00020,000
Default Configuration Summary:
  • Retrieval Mechanism: Hybrid Retrieval
  • Similarity Score: 20
  • Proximity Threshold: 20
  • Top K Chunks: 20
  • Token budget for chunks: 20,000

Answer Generation

Configure how responses are composed and delivered to users.

Answer Components

ComponentDescription
Answer TextThe generated response addressing the user’s question
Snippet ReferenceLink to source as citation for further reading

Answer Types

TypeDescriptionConfiguration
ExtractiveTop chunk retrieved is directly presented as-is without text changesConfigure Response Length (tokens)
GenerativeTop chunks are sent to configured LLM, which generates a paraphrased answerRequires LLM integration and enabled Answer Generation in GenAI Tools

Generative Answer Configuration

Chunk Settings:
SettingDescriptionDefaultMax
Token Budget for ChunksTotal tokens for chunks sent to LLM20,0001,000,000
Enable Document Level ProcessingSend full documents instead of just chunks for richer contextDisabled-
Token Budget for DocumentsMaximum tokens when sending full documents50,000100,000
Chunk Order Options:
OrderDescriptionUse Case
Most to Least RelevantHighest relevance first, then decreasingStandard prioritization
Least to Most RelevantLowest relevance first, most relevant at end followed by queryWhen recency in context matters
LLM Configuration:
SettingDescription
Select Generative ModelChoose from configured LLM models
Answer PromptSelect prompt template for answer generation
TemperatureControls randomness (lower = more deterministic, higher = more creative)
Response LengthExpected answer length in tokens
Feedback Configuration: Enable feedback mechanism to allow users to rate answers. Feedback data appears in Answer Insights analytics.

Response Streaming

Enable real-time token-by-token response delivery for Web/Mobile SDK channels, reducing perceived latency for longer answers. Note: Streaming is configured via prompt settings and is currently supported only for OpenAI and Azure OpenAI models. Not available for API-based responses.

Search Results

Search results display a ranked list of documents/chunks by relevance, providing broader information compared to direct answers.

When to Use Search Results vs Answers

Use CaseRecommended
Direct, specific questionsAnswers
Broad topic explorationSearch Results
Complex queries requiring comparisonsSearch Results
Debugging/troubleshooting with multiple sourcesSearch Results

Configuration Settings

SettingDescriptionRangeDefault
Number of Search ResultsMaximum chunks displayed1-10020
Note: When both search results and extractive answers are enabled, the top result is omitted from search results (shown as the answer) to avoid redundancy.

Filters (Facets)

Filters enable users to narrow results based on specific criteria, useful for large result sets. Filter Types:
TypeDescription
StaticFixed, predefined filter values
DynamicValues derived from search results
Filter UI Options:
UI TypeAvailabilitySelection
TabsStatic filters onlySingle value, string fields only
Single SelectDynamic filtersOne value at a time
Multi SelectDynamic filtersMultiple values concurrently

Creating Filters

  1. Provide unique Filter Name
  2. Select Filter Type (Static or Dynamic)
  3. Choose Field for filtering
  4. Select Filter UI style
Filter Rules:
  • Only one tab-style filter can be enabled at a time
  • Two filters cannot use the same field concurrently
  • Filters apply only if search results contain the specified field

Default Filters

Every application includes default filters that can be updated, deleted, or disabled as needed.

Search Results Access

Currently available via Search API only.

Configuration Checklist

StepSettingLocation
1Choose retrieval strategyConfiguration > Retrieval Strategies
2Set qualification criteria thresholdsConfiguration > Retrieval Strategies
3Select answer type (Extractive/Generative)Configuration > Answer Generation
4Configure LLM settings (if Generative)Configuration > Answer Generation
5Enable/configure search resultsConfiguration > Search Results
6Set up filtersConfiguration > Search Results

Quick Reference: Default Values

ParameterDefault Value
Retrieval StrategyHybrid
Similarity Score Threshold20
Proximity Threshold20
Top K Chunks20
Token Budget (Chunks)20,000
Token Budget (Documents)50,000
Search Results Count20