General Purpose¶
Configure Large Language Models (LLMs) for your AI for Work deployment.
AI for Work provides secure and seamless integration with leading large language models (LLMs), enabling you to leverage powerful AI capabilities for your virtual assistants and workflows. Choose from pre-configured models from industry-leading providers or integrate your own custom models to suit your specific business needs.
What is LLM Configuration?¶
LLM Configuration allows you to connect and manage various language models for use across AI for Work features. This includes both pre-built integrations with commercial providers and the flexibility to bring your own models.
Key Capabilities:
- Pre-Built Integrations: Access leading AI services from OpenAI, Azure OpenAI, and Google Gemini with minimal configuration.
- Embedding Models: Configure specialized models to generate embeddings that power semantic search and other AI features.
- Flexible Authorization: Support for API keys, custom headers, and various authentication methods.
- Enterprise-Grade Security: Maintain control over your data with policy guidelines and secure credential management.
Supported LLM Providers and Models¶
AI for Work supports a range of models from leading providers, each offering different capabilities and pricing tiers. Choose the right model based on your use case requirements, performance needs, and budget considerations.
Pre-Built LLM Providers¶
Provider | Available Models | Access Method | Recommended For |
OpenAI | GPT-5-mini, GPT-5-nano, GPT-5, GPT-4.1, GPT-4.1- mini, GPT-4o, GPT-3.5-turbo, o3, o3-mini, o4-mini | Direct API integration through OpenAI API key | General-purpose tasks, conversational AI, complex reasoning, and content generation |
Azure OpenAI | GPT-5-mini, GPT-5-nano, GPT-5, GPT-4.1, GPT-4.1- mini, GPT-4o, GPT-3.5-turbo, o3, o3-mini, o4-mini | Azure Portal integration through Azure OpenAI Service | Enterprise deployments requiring Azure infrastructure, enhanced security, and compliance |
Google Gemini | Gemini 2.5 Pro (Recommended), Gemini 2.5 Flash (Recommended),
Gemini 2.5 Flash Lite, Gemini 2.0 Flash, Gemini 2.0 Flash Lite |
Google Vertex AI integration | Multi-modal tasks, fast response times, and Google Cloud-based deployments |
Model Tier Recommendations¶
Understanding model tiers helps you select the right balance between capability and cost:
- Basic Tier Models: Ideal for high-volume, straightforward tasks such as classification, simple Q&A, and routine customer inquiries. These models offer faster response times and lower costs.
- Examples: GPT-4o-mini, Gemini 2.5 Flash, Gemini 2.0 Flash.
- Standard Tier Models: Best for complex reasoning, multi-step workflows, and tasks requiring deeper contextual understanding. These models are recommended as the default for all system prompts and orchestrator.
- Examples: GPT-4.1, GPT-4o, Gemini 2.5 Pro.
- Premium Tier Models: Reserved for advanced reasoning tasks, complex problem-solving, and scenarios requiring the highest level of AI capability.
- Examples: o3, o3-mini, o4-mini.
Configuration Methods¶
AI for Work provides two primary methods for configuring LLMs, each designed to meet different organizational needs and deployment scenarios.
Pre-Built LLM Configuration¶
Pre-built integrations offer the fastest path to leveraging leading AI models with minimal technical setup. These integrations are optimized for ease of use while maintaining enterprise-grade security and compliance standards.
Configuration Steps:
- Navigate to Admin Console > Assist Configuration > General Purpose.
- Click New and select your desired LLM provider from the dropdown menu:
- Enter the required authorization details:
- Integration Name: A unique identifier for this configuration (e.g., "OpenAI-Production").
- API Key: Your provider-specific API key obtained from their portal.
- Model Name: Select the specific model from the dropdown menu.
- Review the Policy Guidelines carefully:
- Click Save to activate the configuration.
- Verify the integration appears in the General LLM Integrations list with an active status.
Provider-Specific Configuration Notes:
OpenAI:
- Obtain your API key from the OpenAI Platform dashboard (OpenAI).
- API keys should be stored securely and rotated regularly according to your security policies.
- Monitor usage through the OpenAI dashboard to track costs and ensure rate limits are appropriate.
Azure OpenAI:
- Access your Azure Portal and navigate to Azure OpenAI Service.
- Locate Keys and Endpoints in the resource settings.
- Copy the endpoint URL and one of the available API keys.
- Ensure your Azure subscription has the necessary quotas and permissions.
- Configure network security rules if your deployment requires VNet integration.
Google Gemini:
- Access Google Cloud Console and enable Vertex AI API.
- Create a service account with appropriate permissions.
- Generate and download the API credentials.
- Ensure selected models are enabled in Google Model Garden.
- Configure project-level access and billing settings.
Custom LLM Configuration¶
Custom LLM configuration provides maximum flexibility for organizations with unique requirements, specialized models, or specific hosting preferences. This approach supports virtually any LLM that exposes an API endpoint.
Configuration Steps:
- Navigate to Admin Console > Assist Configuration > General Purpose.
- Click New and select Custom LLM from the dropdown.
- Enter the integration details: Basic Configuration:
- API Configuration:
- Method: Select the HTTP method (typically POST or GET) from the dropdown.
- Max Request Tokens: Set the maximum number of tokens for each request to control costs and response size.
- Authentication:
- Auth Type: Choose your authentication method (API Key, Bearer Token, Custom Header).
- Configure authentication credentials based on your model's requirements.
- Configure custom headers:
- Test the connection:
- Enter a sample test payload in the provided field.
- Click Test to verify connectivity and model response.
- Review the response to ensure proper configuration:
- Review and accept:
- Read the Policy Guidelines.
- Select the checkbox to acknowledge compliance.
- Click Save to finalize the configuration.
- Confirm the integration appears in the General LLM Integrations list with an active status.
Embedding Models Configuration¶
Embedding models generate vector representations of text, enabling semantic search, similarity matching, and other advanced AI features. AI for Work supports both pre-built and custom embedding models to power these capabilities.
What are Embeddings?
Embeddings are numerical representations of text that capture semantic meaning. Unlike traditional keyword matching, embeddings allow AI systems to understand context, intent, and relationships between concepts. This enables more intelligent search results, better content recommendations, and improved natural language understanding.
Note
Embedding models are required for the attachments functionality, where end users can attach files and ask questions about them.
Managing LLM Integrations¶
Once configured, your LLM integrations require ongoing management to ensure optimal performance, security, and cost efficiency.
Viewing Active Integrations¶
All configured models appear in the General LLM Integrations list, which provides:
- Integration Name: Quick identification of each configured model.
- Provider: The LLM provider (OpenAI, Azure OpenAI, Gemini, Custom).
- Model Name: The specific model version in use.
- Status: Active, inactive, or error states.
- Last Updated: Timestamp of the most recent configuration change.
Modifying Existing Integrations¶
To update an existing LLM configuration:
- Locate the integration in the General LLM Integrations list.
- Click the edit icon or select the integration name.
- Modify the necessary fields (API key, model version, headers, etc.).
- Test the connection after making changes.
- Save the updated configuration.
Removing Integrations¶
To deactivate or remove an LLM integration:
- Locate the integration in the General LLM Integrations list.
- Click the delete or deactivate option.
- Confirm the action when prompted.
- Verify that dependent features are configured to use alternative models.