LLM Provider Configuration
This guide explains how to configure Large Language Model (LLM) providers in Simba Intelligence. LLM providers power the AI capabilities that enable natural language querying, data source creation, and intelligent analysis features.Overview
What are LLM Providers?
LLM providers are external AI services that power Simba Intelligence’s natural language capabilities. Think of them as the “brains” behind the system that understand your questions, analyze your data requirements, and generate intelligent responses. When you ask “What were our top-selling products last quarter?” in the Playground, an LLM provider:- Understands your natural language question
- Analyzes your available data sources
- Generates appropriate SQL queries
- Interprets the results into meaningful insights
Supported LLM Providers
Simba Intelligence supports multiple LLMs with structured output support. Older models like GPT-3.5 do not support this and cannot be used. The below LLMs have been tested with the Data Source Agent and Query.| Provider | Model | Status | Quality | Cost | Notes |
|---|---|---|---|---|---|
| 🟢 Vertex AI | Gemini 2.0 Flash | ✅ Works | Standard | $ | Quality trade-offs |
| 🟢 Vertex AI | Gemini 2.5 Flash | ✅ Works | High | $$ | Most tested, reliable ¹ |
| 🟢 Vertex AI | Gemini 2.5 Flash Lite | ❌ Limited | Unstable | $ | Fast and cheap, but unstable |
| 🟢 Vertex AI | Gemini 2.5 Pro | ✅ Works | High | $$$ | High quality and cost, slow ² |
| 🔵 Azure OpenAI | GPT-4o | ❌ Limited | Unstable | $$$ | Unreliable query generation |
| 🔵 Azure OpenAI | GPT-4.1 | ✅ Works | High | $$ | Solid performance |
| 🔵 Azure OpenAI | GPT-4.1-mini | ✅ Works | Standard | $ | Quality trade-offs |
| 🔵 Azure OpenAI | GPT-5.2 | ✅ Works | High | $$ | Good quality, slow |
| 🟠 AWS Bedrock | Nova Pro | ✅ Works | Standard | $$ | Adequate performance |
| 🟠 AWS Bedrock | Claude Sonnet 4 | ✅ Works | High | $$$ | High quality, high cost |
Provider Capabilities
💬 Chat Capability- Natural language understanding and generation
- Query interpretation and response creation
- Business insight generation
- Semantic search and similarity matching
- Intelligent caching of similar queries
- Content understanding for better results
- Dashboard image analysis
- Screenshot-to-data-source conversion
- Visual mockup interpretation
Prerequisites
Required Permissions
To configure LLM providers, you need:- Supervisor role in Simba Intelligence
- Access to LLM Configuration interface (
/llm-configuration) - Administrative privileges for system-wide AI configuration
AI Provider Account Requirements
For each provider you want to use: Google Vertex AI:- Google Cloud Platform account with billing enabled
- Vertex AI API enabled in your GCP project
- Service account with Vertex AI permissions
- Sufficient API quota for your expected usage
- Azure OpenAI resource
- API key with appropriate usage limits
- Sufficient credits/quota for your organization’s needs
- AWS account with Bedrock access enabled
- IAM user or role with Bedrock permissions
- Model access granted for Claude or other desired models
- Appropriate usage quotas configured
Accessing LLM Configuration
Navigation to Configuration Interface
- Log into Simba Intelligence with supervisor credentials
- Navigate to LLM Configuration:
- Option 1: Visit directly at
http://your-domain/llm-configuration - Option 2: User menu → “LLM Configuration”
- Option 1: Visit directly at
- Verify access: You should see a tabbed interface with available providers
Configuration Interface Overview
The LLM Configuration interface provides:- Provider tabs: Separate tabs for each supported AI provider
- Configuration forms: Provider-specific credential and parameter forms
- Capability management: Enable/disable different AI capabilities per provider
- Testing tools: Built-in connection testing and validation
- Status indicators: Real-time status of provider configurations
Google Vertex AI Configuration
Google Vertex AI provides the most comprehensive AI capabilities, including advanced vision analysis for dashboard image processing.Prerequisites for Vertex AI
Google Cloud Platform setup:- Create or select a GCP project with billing enabled
- Enable Vertex AI API:
- Create a service account:
- Grant necessary permissions:
- Generate service account key:
Vertex AI Configuration Process
- Access Vertex AI tab in LLM Configuration interface
-
Enter service account JSON:
- Copy the complete contents of your service account JSON file
- Paste into the credentials text area
- The JSON should include all required fields:
-
Configure global parameters:
-
Enable and configure capabilities:
Chat Capability:
Embeddings Capability:Vision Capability:
-
Save configuration:
- Click “Save” to store the configuration
- Verify all capabilities show as “Active”
Vertex AI Cost Management
Understanding costs:- Chat usage: Charged per input/output token
- Embeddings: Charged per text embedding generated
- Vision: Charged per image analyzed
- Model selection: Different models have different pricing
- Monitor usage in Google Cloud Console
- Set up billing alerts for unexpected usage
- Use semantic caching to reduce redundant API calls
- Choose appropriate models for different use cases
Azure OpenAI Configuration
Azure OpenAI offers the proven GPT models with enterprise-grade security and compliance.Azure OpenAI Configuration
Prerequisites:- Azure subscription with OpenAI resource created
- Deployed models in your Azure OpenAI resource
- API key and endpoint details
- Access Azure OpenAI tab in LLM Configuration interface
-
Enter Azure-specific details:
-
Configure deployments:
-
Set capability parameters:
AWS Bedrock Configuration
AWS Bedrock provides access to various foundation models including Anthropic’s Claude, with enterprise-grade AWS integration.Prerequisites for AWS Bedrock
AWS account setup:- AWS account with Bedrock access enabled in your region
- IAM credentials with appropriate Bedrock permissions
- Model access granted for desired models (Claude, etc.)
- Sufficient service quotas for your expected usage
Bedrock Configuration Process
- Access AWS Bedrock tab in LLM Configuration interface
-
Enter IAM credentials:
-
Configure global parameters:
-
Enable chat capability:
-
Configure embeddings (if available):
Bedrock Model Selection
Available model families:- Anthropic Claude: Excellent for reasoning and analysis
- AI21 Labs Jurassic: Strong language understanding
- Cohere Command: Multilingual capabilities
- Amazon Titan: AWS-native models with competitive performance
- Performance requirements: Response quality and speed
- Cost considerations: Different models have different pricing
- Regional availability: Not all models available in all regions
- Compliance requirements: Some models may have specific compliance certifications
Managing Capabilities
Understanding Capability Types
Chat Capability:- Purpose: Natural language understanding and generation
- Used for: Query interpretation, response generation, insight creation
- Configuration: Model selection, temperature, token limits
- Purpose: Converting text to numerical representations for similarity matching
- Used for: Semantic caching, query similarity detection, content understanding
- Configuration: Model selection, dimension settings
- Purpose: Image analysis and understanding
- Used for: Dashboard mockup analysis, screenshot interpretation
- Configuration: Model selection, image processing limits
Capability Configuration Best Practices
Chat capability optimization:- Temperature settings:
0.0-0.3: Deterministic, factual responses0.4-0.7: Balanced creativity and accuracy (recommended)0.8-1.0: Creative but potentially less accurate
- Token limits: Set based on expected query/response complexity
- Model selection: Balance performance, cost, and feature requirements
- Dimension selection: Higher dimensions = better accuracy but higher cost
- Model compatibility: Ensure embedding model matches your use case
- Caching strategy: Configure appropriate cache retention for embeddings
- Primary provider: Choose most reliable provider for critical capabilities
- Fallback providers: Configure secondary providers for redundancy
- Capability specialization: Use different providers for different capabilities
Testing and Validation
Manual Validation Procedures
After configuring providers:-
Test basic chat capability:
- Go to Playground interface
- Ask a simple question about your data
- Verify natural language response is generated
-
Test embeddings (if configured):
- Ask similar questions multiple times
- Verify responses improve with semantic caching
- Check cache hit rates in monitoring
-
Test vision capability (Vertex AI):
- Go to Data Agent interface
- Upload a dashboard screenshot
- Verify image analysis produces relevant insights
-
Performance validation:
- Monitor response times during testing
- Check API usage in provider dashboards
- Verify no rate limiting or quota issues
Troubleshooting Common Issues
Authentication failures:Security Best Practices
Credential Management
Secure storage:- Never store credentials in code or configuration files
- Use environment variables or secure secret management systems
- Encrypt credentials at rest using appropriate encryption
- Limit credential access to authorized personnel only
- API keys: Rotate every 90 days or according to security policy
- Service accounts: Regenerate keys quarterly
- Access review: Regularly audit who has access to credentials
- Emergency procedures: Have procedures for immediate credential revocation
Network Security
Data privacy:- Understand data flow: Know what data is sent to each provider
- Regional compliance: Use appropriate regions for data sovereignty
- Data retention: Understand provider data retention policies
Access Control
Role-based access:- Limit configuration access: Only supervisors can configure LLM providers
- Separate development/production: Use different credentials for different environments
- Audit configuration changes: Log all changes to LLM provider configurations
- Emergency access: Maintain emergency procedures for credential issues
Cost Management and Optimization
Understanding AI Provider Costs
Cost factors:- Token usage: Both input (prompts) and output (responses) tokens are charged
- Model selection: Premium models cost more than basic models
- Capability type: Chat, embeddings, and vision have different pricing
- Request frequency: High-volume usage may trigger different pricing tiers
Cost Optimization Strategies
Provider selection:- Cost comparison: Compare pricing across providers for your use cases
- Feature optimization: Use expensive features (like vision) only when necessary
- Load balancing: Distribute load to take advantage of different pricing models
Ready to configure your first LLM provider? Start with the provider that best matches your organization’s requirements and follow the step-by-step instructions above. Remember: a well-configured AI foundation is essential for optimal Simba Intelligence performance.

