GETTING STARTED
SearchAssist Overview
SearchAssist Introduction
Onboarding SearchAssist
Build your first App
Glossary
Release Notes
Current Version
Recent Updates
Previous Versions

CONCEPTS
Managing Sources
Introduction
Files
Web Pages
FAQs
Structured Data 
Connectors
Introduction to Connectors
SharePoint Connector
Confluence Connector
Zendesk Connector
ServiceNow Connector
Salesforce Connector
Azure Storage Connector
Google Drive Connector
Dropbox Connector
Oracle Knowledge Connector
Virtual Assistants
Managing Indices
Introduction
Index Fields
Traits
Workbench
Introduction to Workbench
Field Mapping
Entity Extraction
Traits Extraction
Keyword Extraction
Exclude Document
Semantic Meaning
Snippet Extraction
Custom LLM Prompts
Index Settings
Index Languages
Managing Chunks
Chunk Browser
Managing Relevance
Introduction
Weights
Highlighting
Presentable
Synonyms
Stop Words
Search Relevance
Spell Correction
Prefix Search
Personalizing Results
Introduction
Answer Snippets
Introduction
Extractive Model
Generative Model
Enabling Both Models
Simulation and Testing
Debugging
Best Practices and Points to Remember
Troubleshooting Answers
Answer Snippets Support Across Content Sources
Result Ranking
Facets
Business Rules
Introduction
Contextual Rules
NLP Rules
Engagement
Small Talk
Bot Actions
Designing Search Experience
Introduction
Search Interface
Result Templates
Testing
Preview and Test
Debug Tool
Running Experiments
Introduction
Experiments
Analyzing Search Performance
Overview
Dashboard
User Engagement
Search Insights
Result Insights
Answer Insights

ADMINISTRATION
General Settings
Credentials
Channels
Collaboration
Integrations
OpenAI Integration
Azure OpenAI Integration
Billing and Usage
Plan Details
Usage Logs
Order and Invoices

SearchAssist PUBLIC APIs
API Introduction
API List

SearchAssist SDK

HOW TOs
Use Custom Fields to Filter Search Results and Answers
Add Custom Metadata to Ingested Content
Write Painless Scripts
Configure Business Rules for Generative Answers

Best Practices and Points to Remember

Generative Answers

  • Generative answers provide the best user experience and accuracy, especially for multilingual use cases.
  • Using a 16k LLM model with top 10 chunks is recommended in Generative Answers
  • It is recommended to edit the Generative AI prompt according to your use case, e.g., mentioning details of the use case and defining how you want the LLM to answer(“Answer like a helpful assistant/answer like a customer service representative/etc). You can also describe the profile of who is asking the question to improve the LLM response further. 
  • It is not recommended to ask the LLM to make any logical deductions or mathematical calculations. 
  • Response time for answers depends on the number of business rules, total content indexed, chunk size, no of chunks sent to LLM, prompt, LLM model used, etc. Typically, you can expect  5-10 seconds per query in our cloud.

Extractive Answers

  • The chunk size is not customizable and depends on the format of the document.
  • This extraction model has certain known limitations and doesn’t work with all different types of formats. 
  • The extracted content in the chunks is presented as it is to the user as the answer.
  • Implement manual review or validation processes to ensure the correctness of the answers. 
  • Due to the way chunks are generated for extractive answers, they may lack or have limited contextual knowledge. 
  • The quality of extractive answers heavily depends on the quality and relevance of the source text.

When both models are enabled

  • Two sets of chunks are generated for the source content as per the chunking strategy for both models. 
  • When a generative answer is presented to the user, only the chunks generated by the Generative model(Plain text extraction model) are used for the answer. 
  • When an extractive answer is presented to the user, only the chunks generated by the Extractive model(Rule-based extraction model) are used for the answer.
  • The precedence of the models can be selected by the user in Answer Snippets.

Best Practices and Points to Remember

Generative Answers

  • Generative answers provide the best user experience and accuracy, especially for multilingual use cases.
  • Using a 16k LLM model with top 10 chunks is recommended in Generative Answers
  • It is recommended to edit the Generative AI prompt according to your use case, e.g., mentioning details of the use case and defining how you want the LLM to answer(“Answer like a helpful assistant/answer like a customer service representative/etc). You can also describe the profile of who is asking the question to improve the LLM response further. 
  • It is not recommended to ask the LLM to make any logical deductions or mathematical calculations. 
  • Response time for answers depends on the number of business rules, total content indexed, chunk size, no of chunks sent to LLM, prompt, LLM model used, etc. Typically, you can expect  5-10 seconds per query in our cloud.

Extractive Answers

  • The chunk size is not customizable and depends on the format of the document.
  • This extraction model has certain known limitations and doesn’t work with all different types of formats. 
  • The extracted content in the chunks is presented as it is to the user as the answer.
  • Implement manual review or validation processes to ensure the correctness of the answers. 
  • Due to the way chunks are generated for extractive answers, they may lack or have limited contextual knowledge. 
  • The quality of extractive answers heavily depends on the quality and relevance of the source text.

When both models are enabled

  • Two sets of chunks are generated for the source content as per the chunking strategy for both models. 
  • When a generative answer is presented to the user, only the chunks generated by the Generative model(Plain text extraction model) are used for the answer. 
  • When an extractive answer is presented to the user, only the chunks generated by the Extractive model(Rule-based extraction model) are used for the answer.
  • The precedence of the models can be selected by the user in Answer Snippets.