Guide April 18, 2026 · 17 mins · The D23 Team

How D23 Onboards a New Customer to Managed Apache Superset in 7 Days

Walk through D23's 7-day onboarding sprint: infrastructure setup, data connections, dashboard templates, and production-ready analytics. From kickoff to first dashboards.

How D23 Onboards a New Customer to Managed Apache Superset in 7 Days

How D23 Onboards a New Customer to Managed Apache Superset in 7 Days

Most companies spend 6–12 weeks getting a business intelligence platform off the ground. They provision infrastructure, hire DBAs to tune database connections, build security models from scratch, and iterate on dashboard templates while stakeholders grow impatient.

D23 does it in 7 days.

This isn’t marketing hyperbole. It’s the result of running the same onboarding sprint dozens of times—with data and engineering leaders at scale-ups, mid-market companies, private equity portfolio firms, and venture capital groups. We’ve distilled the process into a repeatable operational framework that moves from kickoff to production dashboards faster than most teams can schedule their first standup.

This article walks you through exactly how that sprint works: what happens each day, why we’ve structured it this way, and what your team should expect to have running by day seven.

Why Seven Days?

Seven days isn’t arbitrary. It’s the minimum viable timeline to move from zero to production-grade analytics without cutting corners on security, data governance, or performance.

Here’s why the typical timeline is so much longer: most BI deployments treat onboarding as a linear process. You configure the database, then build security, then design dashboards, then test. If any step hits a snag—and they always do—everything downstream gets delayed.

D23’s sprint model runs in parallel. While your infrastructure is being provisioned, we’re building data models. While dashboards are being designed, security policies are being tested. Your team is never blocked waiting for the previous step to finish.

The seven-day window also creates operational urgency. Everyone knows the deadline. Decisions get made faster. Scope stays tight. Stakeholders show up because they know their window to influence the initial setup is narrow.

And practically: seven days is long enough to surface real problems (bad data quality, unclear KPI definitions, missing database credentials) but short enough that you don’t lose momentum or let scope creep derail the project.

Day 1: Infrastructure & Access Setup

The sprint kicks off with a half-day onboarding call. This isn’t a generic product demo. It’s a working session with your engineering lead, data lead, and whoever owns the databases you’ll be connecting.

We cover three things:

Database inventory and credentials. We need to know what databases you’re querying, where they live (cloud data warehouse, on-prem Postgres, etc.), and what credentials we’ll use. If you don’t have a dedicated analytics user account yet, we create one together and set appropriate permissions. This is also where we discuss data governance and security best practices for Superset deployments to ensure your data connections follow least-privilege access patterns.

Authentication and user provisioning. We set up your D23 instance with your authentication method—SSO via Okta, Azure AD, or standard email/password. We create initial user accounts for your core team (typically 3–5 people at kickoff) and assign roles. D23 uses Apache Superset’s role-based access control (RBAC) under the hood, which means we can lock down who sees which dashboards and data sources from day one.

Slack or email integration. We connect your communication channel so the D23 team can stay in sync with you throughout the week without scheduling 10 separate calls.

After the call, our infrastructure team provisions your D23 instance. This includes:

  • Spinning up a managed Superset environment on our infrastructure (or your VPC if you prefer)
  • Configuring Apache Superset configuration settings for your specific needs (cache settings, query timeouts, feature flags)
  • Setting up automated backups and monitoring
  • Provisioning SSL certificates and DNS

By end of day one, you have a live, empty D23 instance with your team logged in and your first database connection tested. Nothing flashy, but everything’s live.

Day 2: Data Connections & Semantic Layer

Day two is where we move from infrastructure to data. Your database is already connected (we tested it on day one), but now we’re building the semantic layer—the translation between raw SQL tables and business-friendly datasets that your team will actually query.

This is critical because it’s the difference between your analysts writing SQL directly (slow, error-prone, not scalable) and your entire company being able to explore data through a self-serve interface.

We start by cataloging your data sources. We ask:

  • What tables contain your core business metrics (revenue, signups, churn, etc.)?
  • What’s the grain of each table? (daily? hourly? per-transaction?)
  • What fields are dimensions (categorical: region, product, customer segment) vs. measures (numeric: revenue, count, duration)?
  • Are there any tricky calculations or business logic we need to encode? (e.g., “revenue only counts if the transaction is marked complete”)

We then create datasets in Superset that wrap these tables. A dataset in Superset is a logical entity that includes:

  • The underlying table(s)
  • Pre-calculated columns (computed fields like “revenue per user” or “days since signup”)
  • Access controls (who can see this dataset)
  • Caching rules (how often the data refreshes)

For a mid-market company, this typically means 5–15 datasets. For a data-heavy organization, it might be 30–50. The key is that we build only what you’ll actually use in the first 30 days, not a complete data dictionary.

We also discuss Superset multi-tenancy and roles if you’re a platform company embedding analytics into your product or if you’re a PE firm managing multiple portfolio companies. Superset’s RBAC model lets us create separate “tenants” within a single instance, each with their own datasets, dashboards, and user groups.

By end of day two, your team can log into D23, pick a dataset, and start building charts. Nothing’s optimized yet—we’ll tune query performance later—but the data is accessible.

Day 3: Template Dashboards & KPI Framework

Day three is about establishing the dashboard architecture. We’re not building custom dashboards for every user yet. Instead, we’re creating template dashboards that cover the core KPIs your business tracks.

We typically create 3–5 template dashboards:

1. Executive Dashboard. High-level business metrics: revenue, growth rate, customer count, churn, unit economics. This is what your CEO/board sees. It’s usually 6–10 charts, refreshing daily, with minimal drilling capability (we keep it simple so it loads fast and tells a clear story).

2. Operational Dashboard. Real-time or near-real-time metrics for day-to-day operations. For a SaaS company: signups, activation, feature adoption, support tickets. For a marketplace: transaction volume, seller activity, buyer retention. This is what ops/product teams check multiple times a day.

3. Finance Dashboard. Revenue, bookings, ARR, cash burn, cohort analysis, unit economics. Usually monthly refresh, heavy on historical trend lines.

4. Data Quality Dashboard. (Optional but recommended) Metrics on your own data pipeline: row counts, update frequency, null rates, data freshness. This catches data problems before they corrupt downstream analytics.

5. Product/Feature Analytics. (If applicable) Feature usage, funnel conversion, user cohorts, A/B test results. This one changes most frequently as your product roadmap evolves.

Each dashboard follows a consistent design pattern:

  • Clear title and last-refresh timestamp
  • Filters at the top (date range, segment, region, etc.)
  • Metrics in descending order of importance
  • Drill-down capability for charts (click a bar to filter other charts on the dashboard)
  • Mobile-responsive layout (charts stack on smaller screens)

We also establish naming conventions and a shared dashboard repository so your team can find things. If you’re using Apache Superset’s official onboarding guidebook, you’ll see similar best practices.

By end of day three, your team has working dashboards they can actually use. They’re not perfect—we’ll iterate—but they’re live and answering real business questions.

Day 4: Performance Tuning & Caching Strategy

Day four is invisible to most users but critical for production readiness. We’re optimizing query performance and setting up caching so dashboards load in seconds, not minutes.

Here’s what we do:

Query analysis. We run each dashboard’s queries and look for bottlenecks. Are we scanning billions of rows? Are there missing indexes on the database side? Are we doing expensive joins that could be pre-computed?

For each slow query, we have options:

  • Database optimization. Work with your DBA to add indexes or materialized views. This is the gold standard but takes longer and requires database access.
  • Superset caching. Cache query results so the second time someone runs the same query, we return the cached result instantly. We configure cache TTL (time-to-live) based on how fresh the data needs to be. Executive dashboards might cache for 1 hour; operational dashboards for 5 minutes.
  • Pre-aggregation. Build summary tables upstream in your data warehouse (using dbt, SQL, or your ETL tool) so Superset queries against smaller, pre-computed datasets.
  • Query simplification. Sometimes the dashboard is asking for something expensive. We redesign the chart to ask a simpler question that’s still useful.

We also configure Superset’s query timeout and result size limits to prevent runaway queries from breaking the system.

By end of day four, your dashboards load in under 5 seconds even during peak traffic. Your team doesn’t have to wait for analytics.

Day 5: Security, Governance & Access Control

Day five is about locking down who sees what. This is non-negotiable for any production system, especially if you’re handling sensitive financial data, customer PII, or competitive metrics.

We implement:

Role-based access control (RBAC). We create roles that map to your org structure:

  • Admin. Full access to all dashboards, datasets, and settings. Usually just your data lead and one engineer.
  • Editor. Can view dashboards and create new ones, but can’t modify system settings. Product managers, analysts, finance team.
  • Viewer. Read-only access to assigned dashboards. Most of the company falls here.
  • Custom roles. If you need fine-grained control (e.g., “finance team can only see finance dashboards”), we build those.

Row-level security (RLS). If you’re a multi-tenant company or you have regional/divisional separation, we use Superset’s RLS feature to ensure each user only sees their own data. For example, a regional sales manager sees only their region’s numbers, even though all regions live in the same table.

Dataset-level permissions. We tag datasets as public, internal, or restricted, and assign permissions accordingly. Your finance team can access the revenue dataset; your support team cannot.

Audit logging. We enable Superset’s audit log so we can track who viewed what dashboard, when, and from where. This is required for compliance in many industries.

We also discuss data governance best practices with your team: data classification (what’s public vs. sensitive), retention policies, and what happens when someone leaves the company (account deactivation, dashboard ownership transfer).

By end of day five, your system is locked down. Users can only see what they’re supposed to see, and you have an audit trail of who accessed what.

Day 6: Embedded Analytics & API Integration

Day six is for teams that want to embed D23 dashboards into their product or integrate analytics into their application stack.

If you’re a product company, embedded analytics is a game-changer. Instead of your customers logging into a separate BI tool, they see dashboards directly in your app. It’s faster, more seamless, and increases feature adoption.

D23 supports embedded analytics through:

Embedded dashboards. We generate an embed URL that you drop into an iframe in your app. The dashboard is fully interactive (users can filter, drill down, export) but appears native to your product.

Superset API. We expose Superset’s REST API so you can programmatically query dashboards, create charts, or fetch data. This is how you’d build custom analytics experiences on top of Superset.

MCP server for analytics. If you’re using Claude or another LLM, D23 can expose an MCP (Model Context Protocol) server that lets the LLM query your dashboards and data directly. This enables natural-language analytics: “What’s our churn rate this month?” gets answered by the LLM querying Superset and returning the result.

We also discuss text-to-SQL capabilities. If your team is comfortable with AI-assisted queries, we can enable a feature where users type a natural-language question (“Show me revenue by region for Q4”) and Superset generates the SQL and returns the answer. This requires some setup and testing to ensure the AI doesn’t hallucinate bad queries, but it’s powerful for non-technical users.

By end of day six, your product team has everything they need to embed analytics. If you’re not embedding, this day is optional—we skip it and move to day seven.

Day 7: Testing, Documentation & Handoff

Day seven is about ensuring everything works and your team can run it independently.

We do:

User acceptance testing (UAT). Your team logs in and uses the dashboards like they would in production. We watch for bugs, confusing UI, slow queries, or missing data. We fix anything critical on the spot; non-critical issues go into a backlog.

Documentation. We create:

  • Dashboard inventory. A spreadsheet listing every dashboard, what it shows, who owns it, refresh frequency, and how to interpret it.
  • Data dictionary. For each dataset, we document the tables, fields, calculations, and any quirks (e.g., “revenue doesn’t include refunds”).
  • Runbook for common tasks. How to add a new user, create a dashboard, refresh data, debug a slow query.
  • FAQ. Common questions and answers.

We also record a 20-minute walkthrough video showing your team how to navigate D23, create a simple chart, and build a dashboard from scratch.

Handoff meeting. We sit down with your team and walk through the system. We answer questions, demo a few workflows, and make sure everyone’s comfortable.

Support plan. We define how you’ll get help going forward. D23 provides 24/7 support for critical issues (system down, data corruption) and business-hours support for questions. We also schedule a weekly 30-minute check-in for the first month to catch any issues early.

By end of day seven, your team owns the system. They can create dashboards, add users, and troubleshoot basic issues. D23 is there if they hit something complex, but the day-to-day operation is theirs.

What Makes Seven Days Possible

This timeline works because of a few structural choices:

Managed infrastructure. We handle all the ops work (patching, monitoring, backups, scaling). You don’t have to. If you were self-hosting Superset, this timeline would be 3–4 weeks just for infrastructure.

Pre-built templates. We don’t start from scratch. We have battle-tested dashboard templates, dataset patterns, and security configurations from dozens of customers. We adapt them to your business, not build from zero.

Parallel workstreams. Different parts of the team work on different things simultaneously. Your DBA tunes the database while our team builds dashboards. Your security team reviews access controls while our team sets up caching.

Ruthless scope discipline. We build what you need in 30 days, not what you might need in 2 years. That’s not a limitation; it’s a feature. You’ll iterate and add more dashboards later. The goal is to get to value quickly, not to boil the ocean.

Experienced team. Our team has done this 50+ times. We know the common pitfalls (bad data quality, unclear KPI definitions, scope creep) and how to navigate them. We move fast because we’ve seen it all.

Real Timeline Variability

Seven days is the target, but real projects vary:

Faster (4–5 days): If your data is already clean, well-documented, and in a cloud data warehouse (Snowflake, BigQuery, Redshift), and your team is available and aligned on KPIs, we sometimes finish in 4–5 days.

Slower (10–14 days): If your data is messy (missing values, inconsistent schemas, undocumented tables), your databases are on-prem and require VPN setup, or your team is distributed and hard to sync, it might take 10–14 days. This is still much faster than typical BI deployments, but it’s honest about real-world constraints.

Variable scope: Some customers want just an executive dashboard (2–3 days). Others want embedded analytics and API integration (add 2–3 days). We scope based on what you actually need.

Comparison to Alternatives

How does this compare to other managed BI platforms?

Looker (Google Cloud). Looker is powerful and feature-rich, but onboarding typically takes 8–12 weeks. You need a dedicated Looker developer. Licensing is expensive ($2–5K per user per year). D23 is built on open-source Apache Superset, so you own your data model and aren’t locked into Looker’s LookML language.

Tableau. Similar story: 10–16 weeks to production, high licensing costs, steep learning curve. Tableau is great if you have a large team of dedicated analysts, but it’s overkill for most mid-market companies.

Metabase. Metabase is simpler and faster to deploy (2–3 weeks), but it’s less flexible than Superset. You can’t customize the UI, embed easily, or integrate with LLMs. Metabase is good for basic dashboarding; D23 is for companies that want a BI platform that scales with them.

Mode. Mode is SQL-focused and great for analysts, but it’s not a full BI platform. It doesn’t have role-based access control, embedded analytics, or LLM integration. It’s more of a query editor than a BI suite.

Power BI. Power BI is cheap and integrates well with Microsoft, but it’s enterprise-focused and requires significant IT overhead. Onboarding is 8–12 weeks. It’s best for large organizations with dedicated BI teams.

D23 sits in the middle: faster than enterprise platforms (Looker, Tableau, Power BI), more flexible than simple tools (Metabase), and purpose-built for modern data stacks (cloud data warehouses, dbt, Airflow, etc.).

Day-by-Day Checklist

If you’re planning to work with D23, here’s what to have ready:

Before day 1:

  • Database credentials for your analytics database(s)
  • List of who should have access (names, emails, roles)
  • Your SSO provider details (Okta, Azure AD, etc.) if you use one
  • Slack or email for async communication

Day 1:

  • Attend the onboarding call (30 min)
  • Provide database credentials
  • Create analytics user accounts on your databases
  • Confirm your team can log into D23

Day 2:

  • Document your core business metrics and KPIs
  • List the tables/datasets you want to analyze
  • Clarify any complex calculations or business logic
  • Test that your team can browse datasets in D23

Day 3:

  • Review template dashboards
  • Provide feedback (missing charts, unclear metrics, design changes)
  • Confirm dashboard owners and stakeholders
  • Test dashboard filters and interactivity

Day 4:

  • Monitor dashboard performance (should be fast)
  • Report any slow queries
  • Confirm caching strategy (how fresh should data be?)
  • No action needed; this is mostly backend work

Day 5:

  • Review access controls (who can see what)
  • Confirm role definitions match your org
  • Test that row-level security works correctly
  • Sign off on security and governance policies

Day 6 (optional):

  • If embedding: provide your app’s domain for iframe setup
  • If using APIs: test API calls with sample queries
  • If using MCP: test natural-language queries
  • If not embedding: skip this day

Day 7:

  • Attend UAT session (1–2 hours)
  • Review documentation
  • Ask questions
  • Confirm your team is ready to operate independently
  • Schedule first support check-in

Beyond Day 7: The First 30 Days

The sprint gets you to production, but the real work is adoption. In the first 30 days:

Week 2: Your team creates custom dashboards for their specific needs. We’re available to help, but they’re driving. You’ll probably discover data quality issues or KPI misalignments that we iterate on.

Week 3: You integrate D23 into your decision-making workflows. The executive dashboard becomes part of the weekly board meeting. The operational dashboard gets checked daily. You start making decisions based on data, not hunches.

Week 4: You identify power users and invest in their growth. These are the people who’ll evangelize D23 across your company. We might run a training session for them on advanced features (custom SQL, complex filters, embedding).

By day 30, D23 isn’t a new tool anymore—it’s how your company talks about data.

Conclusion: Speed Without Sacrifice

The seven-day onboarding sprint is fast, but it doesn’t cut corners. You get:

  • Production-grade infrastructure with monitoring and backups
  • Secure, governed data access with role-based controls
  • Dashboards that answer real business questions
  • Query performance that won’t frustrate your team
  • Documentation and training so your team can operate independently
  • A vendor (D23) that stays involved and helps you iterate

This is possible because D23 is purpose-built for this workflow. We’re not a generic Superset consultant; we’re a managed Superset service that has optimized every step of onboarding. We’ve learned what works, automated what we can, and focused on what matters: getting you from zero to production-grade analytics as fast as possible without sacrificing quality, security, or usability.

If you’re evaluating D23 managed Apache Superset or any BI platform, ask them about their onboarding timeline and process. If they can’t articulate a clear, parallel workflow, you’ll end up in the 8–16 week category with most vendors.

D23 does it in 7 days. And we’ve documented exactly how.