Quick Start Guide (Production Focus)
This guide is for enterprise teams standing up Simba Intelligence in a production (or production-like) environment. It intentionally avoids developer workstation / sample data instructions and instead walks you through the minimum sequence to become operational with your own infrastructure and data.Overview of the Flow
- Deploy the platform
- Configure an approved LLM / AI provider
- Register at least one production data connector and create a data connection
- Generate an initial governed data source using the Data Source Agent
- Enable business users to explore via the Playground
- Validate operational readiness & plan next hardening steps
1. Deploy the Platform
Deployment steps (container orchestration, Helm, configuration files, secrets, scaling parameters, ingress, TLS, etc.) are documented in the dedicated Installation Guide. ➡️ Follow: Installation Guide When complete you should have:- Platform services healthy
- Administrative (supervisor) credentials provisioned
- Network access from the platform to required outbound AI endpoints and internal data sources
2. Configure an LLM / AI Provider
LLM configuration enables natural language understanding, semantic modeling, and embeddings. Use production credentials (not personal developer keys).- Sign in with a supervisor / administrator role.
- Navigate to:
/llm-configurationin the UI. - Select your provider (e.g., Vertex AI, Azure OpenAI, Bedrock).
- Provide the required credentials (service account JSON, API key + endpoint, or similar) via the secure form.
- Enable the ncessary Chat + Embeddings services and save.
- Vertex AI: see Configuration Reference – Google Vertex AI
- Azure OpenAI: see Configuration Reference – Azure OpenAI
- AWS Bedrock: see Configuration Reference – AWS Bedrock
- Store original secrets in your enterprise vault (e.g., HashiCorp Vault, AWS Secrets Manager); rotate per policy.
- Use least-privilege IAM role for service accounts.
- Monitor usage limits / quotas early to baseline cost.
3. Create a Data Connection
Connect an authoritative data source (e.g., Snowflake, PostgreSQL, SQL Server, BigQuery, etc.). Avoid pointing to non-governed sandbox datasets for your first production rollout.- Navigate to:
Data Connectionsin the UI. - Click Create Connection.
- Choose connector type.
- Supply credentials (service user / managed identity preferred) + network parameters (host / account / warehouse / catalog / database / schema as applicable).
- (Recommended) Scope access to read-only analytical schemas for the initial rollout.
- Test Connection → Save.
- Allow initial metadata discovery to complete (time depends on catalog size).
- Snowflake: Provide account locator / region, warehouse sized for concurrency testing, role with
USAGE+SELECT. - PostgreSQL / SQL Server: Prefer read replica endpoints to avoid production OLTP impact.
- BigQuery: Service account with dataset-level
roles/bigquery.dataViewerand projectmetadataViewer.
4. Create an Initial AI-Generated Data Source
Use the Data Source Agent to translate business intent into a governed data source definition over your connected warehouse.-
Navigate to:
Data Source Agent(/data-source-agent). - Select the connection you just registered.
-
(Optional) Provide a clear objective (concise, business-oriented). Examples:
- (Optional) List critical tables or schemas to prioritize (helps reduce exploration time).
- Start the agent. It will inspect schema metadata, propose transformations / joins, and assemble the data source definition.
- Review the generated fields; rename or hide any that should not be exposed. Approve to publish.
- Keep the first data source narrow; optimize clarity over breadth.
- Enforce naming conventions (e.g.,
revenue_total,customer_churn_rate). - Capture data lineage output for audit (export if needed).
5. Explore with the Playground
The Playground allows natural language querying and iterative exploration over approved data sources.- Navigate to:
Playground(/playground). - Select the published data source created above.
- Ask domain-focused questions or use the suggested questions. Examples:
- “What is total revenue this quarter vs the same quarter last year?”
- “Show top 15 customers by trailing 90 day revenue and percent growth vs prior 90 days.”
- “Break down subscription churn by region and customer segment.”
- “List products with declining monthly revenue for the last 4 months.”
- “Identify anomalies in daily order volume over the past 30 days.”
- “Compare average customer lifetime value between enterprise and mid-market segments.”
6. Validate Operational Readiness
Confirm the platform is stable, secure, and delivering expected value before scaling access. Readiness Checklist:- ✅ LLM provider status healthy; usage metrics monitored
- ✅ At least one production data connection
- ✅ Initial curated data source published and reviewed
- ✅ Successful natural language queries returning accurate SQL and results
- ✅ Access controls / roles applied (admin vs analyst vs viewer)
- ✅ Logging & monitoring integrated (infrastructure + application + AI usage)
- ✅ Backup / recovery strategy documented (see Operations Guide)
- Add additional connections (e.g., finance, customer success, product analytics) incrementally.
- Implement tagging / classification for sensitive fields (PII, financial, etc.).
- Configure performance monitoring & query cost dashboards.
- Set up automated model / provider health alerts.
Troubleshooting (Production Context)
Common early issues: LLM Provider Authentication Fails- Verify service account / key has required model permissions.
- Confirm outbound firewall rules allow provider endpoints.
- Limit initial schema scope; exclude large archival schemas.
- Schedule off-peak discovery windows for very large catalogs.
- Enable caching in each data source
- Add row-level / column filters to high-volume tables.
- Validate source table freshness; confirm no stale replicas.
- Check semantic field definitions for unintended aggregations.
Reference Links
- Installation: Installation Guide
- Data Connection: Data Connection Management
- Data Source Agent: Data Agent User Guide
- Playground: Playground User Guide
- Administration: Administrator Guide
- Operations:
- Security & Auth: Authentication & Security
You’re now ready to onboard additional teams. Keep the initial surface area small, validate quality, then scale connections, data sources, and user groups iteratively.

