Virbe Documentation
  • Say hi to Virtual Beings!
  • Getting started
    • Introduction to Virbe platform
      • Virbe-hosted launch
      • Azure-hosted deployment
  • Dashboard Management
    • Dashboard Architecture
    • Overview
    • Profiles
      • General
      • Language
      • Persona
      • Settings
        • Web widget profile settings
        • Metahuman kiosk profile settings
      • Deployment
        • Web widget deployment
        • Metahuman kiosk deployment
    • Configurations
      • Speech-to-Text
      • Text-to-Speech
      • AI Models
      • Conversational Engines
    • Personas
    • Knowledge Base
      • Best Practices for Knowledge Base Content
    • Data tables
    • Conversation Flows
      • Getting started with Conversation Flows
      • Managing nodes
      • Nodes
        • Logic nodes
          • Route by Profile
          • If/Else Router
          • Intent Matcher
        • Flow control nodes
          • Go to Flow
          • Checkpoint
        • Response nodes
          • Text
          • LLM Response
          • Call ConvAI
          • Behavior
        • Action nodes
          • Collect User Data
          • Quick Reply
          • Cards from Table
        • Context nodes
          • Find Record(s)
        • Integration nodes
          • Call Webhook
          • Custom Action
      • Troubleshooting Conversation Flows
  • Touchpoints
    • Kiosk Apps
      • Metahuman Kiosk
        • Hardware setup
          • Recommended setup
          • Computer
          • Screen
          • Microphones
          • IoT
          • Peripherals
          • Operating System
            • Windows
              • Login & power management
              • Other
              • OpenHab
            • Mac
            • Remote Access
        • Customer Experience
          • VAD & press-to-talk
          • Microphone settings
          • Avatars
    • Web Integration
Powered by GitBook
On this page

Was this helpful?

  1. Dashboard Management
  2. Configurations

AI Models

PreviousText-to-SpeechNextConversational Engines

Last updated 2 months ago

Was this helpful?

AI Models

Add and configure Large Language Models for your virtual being that will govern the response generation as well as conversation flow logic. Supported engines:

  • OpenAI

  • Azure OpenAI

  • Anthropic

Where to find credentials?

Log in to a relevant service provider and find a section related to using the service by API. Below you'll find an example of such a section in OpenAI's dashboard:

You can configure multiple AI models and set one as default. Each model requires its own set of credentials and configurations.

In Conversation Flows editor you'll be able to use different AI models for different nodes requiring LLMs, so that you can use faster and cheaper models for logic processing and larger models for high quality conversational responses.

Ai models configuration
API keys section in OpenAI's dashboard