Skip to main content
Back to XO GPT Module

Live Model Versions

The table below lists all currently deployed XO GPT models.
For non-English languages, XO GPT supports industry-established generic use cases. For additional language-specific support, use the Agent Platform.
XO GPT ModelSupported FeatureVersionBase ModelLanguagesRegionsDeployed
Answer Generation ModelAnswer Generationv3.0Llama 3.1 8B InstructEnglish, French, German, Japanese, Polish, SpanishUS, DE, EU6 May 2025
Conversation Summarization ModelConversation Summarizationv2.0Mistral 7B Instruct v0.2English, French, German, Japanese, Polish, Simplified Chinese, Spanish, Traditional Chinese, TurkishUS, DE, JP23 Sep 2025 (US/DE), 20 Dec 2024 (JP)
Response Rephrasing ModelRephrase Dialog Responsesv1.0Mistral 7B Instruct v0.2EnglishUS, DE1 Jun 2024 (US), 3 Sep 2024 (DE)
User Query Paraphrasing ModelRephrase User Queryv1.0Mistral 7B Instruct v0.2EnglishUS, DE1 Jun 2024 (US), 3 Sep 2024 (DE)
DialogGPT ModelDialogGPT - Conversation Orchestrationv1.1Llama-3.1-8B-InstructEnglish, French, German, Japanese, Polish, SpanishUS, DE26 May 2025

Model Building Process

XO GPT Model Building Process

Data Collection

Training data is gathered from conversations across multiple domains—healthcare, banking, e-commerce, IT support, finance, and more. Training data sources:
  1. Synthetic data generated using Azure OpenAI GPT-4.
  2. Human experts manually develop data based on the problem, challenges, and expected outcomes.
  3. A separate team of human annotators evaluates the data for relevancy, scenario coverage, and correctness.
  4. No customer data is used for training or evaluation.
Data profile: Multiple samples per language across various categories and use cases. All training data is versioned in Kore.ai’s XO GPT Data Repository and is proprietary to Kore.ai.

Data Processing

Raw data is cleaned to remove irrelevant content, standardize formats, and prepare text for tokenization and normalization. This ensures compatibility with the base model’s requirements. Fine-tuning techniques used:
TechniqueDescription
Memory Efficiency4-bit precision loading and double quantization reduce memory usage while maintaining accuracy.
Low-Rank Adaptation (LoRA)Applied to specific model layers; parameters such as rank, scaling factor, and dropout are tuned to minimize overfitting.
Optimized Training ParametersCarefully selected learning rate, batch size, and epochs balance training efficiency with performance.
Advanced OptimizationState-of-the-art optimizer with optional warm-up steps, early stopping, and LR scheduling.
Task-Specific AdaptationModels are fine-tuned for causal language modeling to target the specific conversational task.

Model Evaluation

Metrics: Accuracy, fluency, hallucinations, robustness, AI safety, and bias. Validation techniques: Cross-validation and hold-out validation to ensure generalization to unseen data. Evaluation data: Synthetic data from GPT models and human experts, covering diverse topics and challenging inputs (typos, poor grammar, profanity). Notes: Internal benchmarks are based on synthetic data and may not generalize to all scenarios. Real-world results may vary based on hardware, network, and implementation specifics.

Model Benchmarks

For version-specific benchmarking details, see:

Model Roadmap

Maintenance

The model is reviewed, updated, and retrained regularly. Bug fixes and performance improvements are addressed on an ongoing basis; new features are added quarterly.

Expansion

  • Multilingual Proficiency — New languages beyond English, French, Spanish, Japanese, Turkish, and German, refined through expert feedback.
  • New Summary Templates — Custom templates such as Stepwise and PRA (Problem-Resolution-Action), developed on demand.

FOR EVALUATION PURPOSES ONLY This document contains proprietary information of Kore.ai Inc. and is provided exclusively for evaluation. It does not grant any licenses, rights, or permissions regarding intellectual property. DISCLAIMER: XO GPT is an advanced AI model that may require improvements over time. Outputs may occasionally be unpredictable, inaccurate, biased, or unexpected. Thoroughly test the model and adjust it for your specific use cases.