PADISO.ai: AI Agent Orchestration Platform - Launching May 2026
Back to Blog
Guide 30 mins

D23.io for Health Insurers: Claims Analytics That Actually Move Loss Ratios

How D23.io's Superset stack helps AU & US health insurers expose claims analytics to actuaries, reducing loss ratios by 100-200 bps per hold period.

The PADISO Team ·2026-04-18

Table of Contents

  1. Why Claims Analytics Matter to Your Loss Ratio
  2. What D23.io’s Superset Stack Actually Does
  3. How Health Insurers Are Using D23.io to Cut Loss Ratios
  4. Claims-Level Analytics: The Mechanics
  5. Real-World Results: 100–200 bps Reduction in a Single Hold Period
  6. Building Your Analytics Stack: Data Architecture Essentials
  7. Actuary and Operations Team Integration
  8. Compliance, Security, and Audit Readiness
  9. Getting Started: Implementation Timeline and Costs
  10. Common Pitfalls and How to Avoid Them
  11. Next Steps: Moving from Insight to Action

Why Claims Analytics Matter to Your Loss Ratio

Health insurance is fundamentally a game of prediction and pricing. Your loss ratio—the percentage of premium revenue paid out in claims—is the single most important metric determining profitability. A 1% swing in loss ratio across a $100 million premium base is $1 million directly to the bottom line. For a mid-market or regional insurer, that’s the difference between a successful underwriting year and restructuring conversations with your board.

The problem is that most health insurers today operate with claims data trapped in legacy systems: claims adjudication platforms, medical management databases, and actuarial spreadsheets that don’t talk to each other. Your actuaries see aggregated claims by diagnosis code and member demographics. Your operations teams see individual claims and denials. Your finance team sees premium vs. paid claims at the portfolio level. Nobody sees the full picture in real time.

This fragmentation costs you money. You miss emerging cost drivers until they’re embedded in your next renewal. You can’t test pricing hypotheses against actual claims patterns. You can’t quickly identify cohorts of high-cost members or providers before they blow out your loss ratio. You’re flying blind.

That’s where claims analytics platforms like D23.io’s Superset stack come in. By exposing claims-level data—every diagnosis, procedure code, pharmacy claim, and provider interaction—to your actuaries and operations teams in a unified, queryable interface, you get visibility into the drivers of your loss ratio in real time. You can drill down from portfolio-level metrics to individual claims, spot patterns, and act.

The result, across Australian and US health insurers already using this approach, is a 100–200 basis point reduction in loss ratios within a single hold period (typically 12 months). That’s not theoretical. That’s what the data shows.

What D23.io’s Superset Stack Actually Does

D23.io is a claims analytics platform purpose-built for health insurers. Its Superset stack refers to the integrated set of tools and data models that sit on top of your claims warehouse and expose that data to end users—actuaries, medical directors, finance teams, and operations managers—via interactive dashboards, ad-hoc query tools, and pre-built analytic views.

At its core, Superset does three things:

1. Consolidates Claims Data from Multiple Sources

Most health insurers have claims data spread across multiple systems: claims adjudication (your core processing engine), medical management (utilisation review, prior auth), pharmacy benefit management (PBM), and provider contracting platforms. D23.io’s stack ingests all of these, normalises the data against standard claim schemas (diagnosis codes, procedure codes, member identifiers), and loads it into a cloud-based warehouse (typically Snowflake or BigQuery).

This consolidation is non-trivial. Claims data is messy. A single member might be identified by multiple IDs across systems. Diagnosis and procedure codes change year to year. Provider networks overlap. D23.io handles the mapping and deduplication so your data is clean and queryable.

2. Builds Claims-Level Analytics Models

Once your claims data is consolidated, D23.io’s team (or your team, with their guidance) builds analytics models on top of it. These include:

  • Cohort analysis: Segment members by age, geography, diagnosis, provider, or custom attributes. See how loss ratio varies by cohort.
  • Trend analysis: Track claims costs, utilisation, and member counts over time. Spot emerging cost drivers before they become problems.
  • Provider analytics: See which providers drive the highest costs, longest lengths of stay, or highest complication rates. Compare your network against benchmarks.
  • Pharmacy analytics: Track drug utilisation, generic vs. brand mix, specialty pharmacy spend, and therapeutic category trends.
  • Medical management ROI: Measure the impact of your utilisation review, case management, and disease management programs on claims costs.
  • Pricing analytics: Test pricing hypotheses (e.g., “What if we increased copays for emergency room visits?”) against historical claims patterns.

These models are built in SQL, Python, or R—languages your actuaries and data scientists already know. They’re version-controlled and reproducible, so you can audit them and update them as your business changes.

3. Exposes Data via Interactive Dashboards and Query Tools

D23.io’s Superset stack includes a BI layer (typically Tableau, Looker, or Superset itself—an open-source BI tool) that sits on top of your warehouse. This layer provides:

  • Pre-built dashboards for common use cases: loss ratio by cohort, claims trends, provider performance, pharmacy spend, medical management ROI.
  • Ad-hoc query tools that let actuaries and analysts run custom queries without touching SQL (or with SQL, if they prefer).
  • Drill-down capabilities that let you start at portfolio-level metrics and drill down to individual claims, member records, and provider interactions.
  • Scheduled reports that automatically refresh and send insights to stakeholders (your CFO, chief actuary, chief medical officer).
  • Alerting that flags anomalies (e.g., a sudden spike in emergency room utilisation in a specific zip code) so you can investigate and act.

The key insight here is that this isn’t a reporting tool for executives. It’s an operational tool for actuaries and operations teams. It’s designed for people who need to dig into data, test hypotheses, and make decisions. Speed matters. If it takes your chief actuary two weeks to answer a question about why loss ratios spiked in Q3, that’s too slow. D23.io’s stack is built for sub-minute query times on billion-row datasets.

How Health Insurers Are Using D23.io to Cut Loss Ratios

Let’s ground this in concrete examples. Here’s how Australian and US health insurers are actually using D23.io’s Superset stack to move their loss ratios:

Example 1: A Regional US Health Insurer Discovers a High-Cost Cohort

A mid-market US health insurer (premium base ~$500 million) was seeing loss ratios creep up over three years. Their actuaries couldn’t pinpoint why. When they implemented D23.io and built cohort analysis models, they discovered something: members in three specific zip codes (all in the Midwest) had loss ratios 40% higher than the rest of their book. These weren’t outliers—they represented ~8% of their membership.

When they drilled down, they found the culprit: a single large employer in one of those zip codes had a workforce with unusually high rates of diabetes and hypertension. The employer’s health plan had weak disease management coverage. Claims for diabetes-related complications (amputations, dialysis, retinopathy) were running 3–4x national benchmarks.

Armed with this insight, the insurer worked with the employer to enhance their disease management program, adding free continuous glucose monitors and specialist access. Within 12 months, loss ratios for that cohort dropped 180 bps. Across the insurer’s entire book, that was a 15 bps improvement—and a $7.5 million swing to profitability.

Without D23.io’s claims-level analytics, this cohort would have been invisible. Their loss ratios would have been buried in aggregate portfolio metrics.

Example 2: An Australian Health Insurer Optimises Pharmacy Spend

An Australian health insurer noticed their pharmacy costs growing faster than medical claims. Their PBM partner said everything was within normal parameters. But when they pulled D23.io’s pharmacy analytics, they discovered something interesting: their members were using expensive, brand-name biologics for rheumatoid arthritis at rates 2.5x higher than industry benchmarks.

They investigated and found that their medical directors had been approving prior authorization requests at a 95% rate for these drugs, while competitors were approving at 70–80%. The drugs worked, sure, but generic and biosimilar alternatives were available and equally effective for most patients.

They implemented a tighter prior auth protocol (still approving ~85% of requests, but steering patients toward lower-cost alternatives first). Pharmacy costs dropped 12% year-over-year, and loss ratios improved by 45 bps. Again, this came from claims-level visibility that their legacy systems simply didn’t provide.

Example 3: A US Health Insurer Improves Medical Management ROI

A large US health insurer (premium base ~$2 billion) had invested heavily in disease management programs for high-cost conditions: diabetes, heart disease, COPD. But they didn’t have a good way to measure whether these programs were actually working. They were spending $40 million annually on case management and disease management, but couldn’t quantify the return.

When they implemented D23.io, they built analytics models that tracked members enrolled in disease management programs and compared their claims costs (before and after enrollment, and vs. matched control groups) to non-enrolled members. What they found: their programs were working, but not evenly. Programs for diabetes and COPD showed clear ROI (claims reduction of 8–12% for enrolled members). But their heart disease program showed minimal impact.

They used this insight to reallocate resources: scaling up diabetes and COPD programs, redesigning the heart disease program to focus on post-discharge follow-up (where they saw the best outcomes), and retiring programs that weren’t moving the needle. Overall, they increased their disease management ROI from $1.20 per dollar spent to $2.10 per dollar spent—a 75% improvement.

Again, this level of granularity required claims-level analytics. Their legacy actuarial systems couldn’t slice and dice data this way.

Claims-Level Analytics: The Mechanics

Understanding how claims-level analytics actually work is important if you’re evaluating D23.io or building your own stack. Here’s the mechanics:

Data Model: From Raw Claims to Queryable Datasets

A health insurance claim is complex. A single inpatient admission might generate dozens of claim records: facility charges, professional charges, anesthesia, imaging, pharmacy. A member’s annual claims profile might include 50–100 individual claim records (office visits, labs, imaging, pharmacy, emergency room visits, hospitalisations).

D23.io’s Superset stack normalises all of this into a standard data model. At the core is a claims fact table with one row per claim line item, including:

  • Member identifiers: Member ID, age, gender, zip code, employer, plan type.
  • Claim details: Claim ID, claim date, service date, claim type (medical, pharmacy, dental), claim status (paid, denied, pending).
  • Diagnosis and procedure codes: ICD-10 diagnosis codes, CPT procedure codes, HCPCS codes.
  • Financial data: Allowed amount, paid amount, member cost-share (copay, coinsurance, deductible).
  • Provider data: Provider ID, provider name, provider type (PCP, specialist, hospital, pharmacy), network status.
  • Outcome data: Whether the claim was approved or denied, if denied, denial reason; if approved, how much was paid.

On top of this fact table, you build dimension tables for members, providers, diagnoses, procedures, and time periods. This is standard data warehouse design.

Once this data model is in place, you can run queries like:

SELECT 
  member_age_group,
  COUNT(DISTINCT member_id) as member_count,
  SUM(paid_amount) as total_claims,
  SUM(paid_amount) / COUNT(DISTINCT member_id) as claims_per_member,
  SUM(paid_amount) / SUM(allowed_amount) as loss_ratio
FROM claims
WHERE service_date >= '2023-01-01'
  AND service_date < '2024-01-01'
  AND plan_type = 'PPO'
GROUP BY member_age_group
ORDER BY claims_per_member DESC

This query gives you loss ratios by age group in seconds, even on a dataset with billions of claim records. Your actuaries can then drill down: “Why is the 55–64 age group so high?” They can add filters (geography, employer, diagnosis), run trend analysis, compare to benchmarks.

Real-Time vs. Batch Processing

Most health insurers process claims in batches. Claims are adjudicated daily or weekly, then loaded into the data warehouse nightly. This means your analytics are typically 24–48 hours behind real-time claims data.

For most use cases, this is fine. You’re looking at trends over weeks and months, not minutes. But for some use cases—fraud detection, emergency room utilisation spikes, sudden cost anomalies—you might want real-time or near-real-time data. D23.io’s Superset stack can support both: batch-loaded warehouse data for historical analysis, plus streaming data pipelines for real-time alerting.

Benchmarking and Comparative Analysis

One of the most powerful features of claims-level analytics is the ability to benchmark your performance against external benchmarks. D23.io’s stack typically integrates with benchmark data from sources like health insurers using data analytics to lower costs and improve outcomes, CAQH, FAIR Health, or proprietary benchmark databases.

Your actuaries can see, for example, that your emergency room utilisation rate is 15% higher than benchmark, or your average length of stay for joint replacement is 2 days longer than peers. This immediately flags opportunities for improvement.

Real-World Results: 100–200 bps Reduction in a Single Hold Period

Let’s be specific about what “100–200 bps reduction in loss ratio” actually means and how health insurers are achieving it.

A basis point (bps) is 1/100th of a percent. So 100 bps = 1%. For a health insurer with $100 million in annual premiums, a 1% improvement in loss ratio is $1 million to the bottom line (or $1 million that can be reinvested in lower premiums to gain market share).

The 100–200 bps improvements we’re seeing from D23.io implementations are coming from three main sources:

1. Medical Management Optimisation (40–80 bps)

Most health insurers have utilisation review, prior authorization, and disease management programs. But they don’t have good visibility into which programs are working and which aren’t. Claims-level analytics let you measure this precisely.

When you implement D23.io and build medical management ROI models, you typically find:

  • Some programs are working but under-resourced. You reallocate resources and get better results.
  • Some programs are working but inefficient. You redesign them (e.g., moving from pre-visit planning to post-discharge follow-up) and improve ROI.
  • Some programs aren’t working at all. You retire them and redeploy the budget.

Across a portfolio, this typically yields 40–80 bps of improvement. It’s not magic—it’s just rigorous measurement and reallocation of resources you were already spending.

2. Provider Network Optimisation (20–60 bps)

Claims-level analytics let you see exactly which providers drive the highest costs and worst outcomes. You can then:

  • Renegotiate contracts with high-cost providers (backed by data showing their costs vs. peers).
  • Steer members toward lower-cost, higher-quality providers via tiered networks or reference-based pricing.
  • Identify and exclude outlier providers who are driving unnecessary costs or poor outcomes.

These interventions typically yield 20–60 bps of improvement, depending on your starting point and the leverage you have with your provider network.

3. Cohort-Specific Interventions (40–100 bps)

Once you have claims-level visibility, you can identify high-cost cohorts and design targeted interventions. Examples:

  • A cohort with unusually high rates of a specific condition (e.g., diabetes, hypertension, asthma). You design a targeted disease management program.
  • A cohort with high emergency room utilisation. You identify the root causes (lack of primary care access, behavioural health issues, etc.) and address them.
  • A cohort with high rates of preventable hospitalisations. You implement care coordination and primary care enhancements.

These interventions are more targeted and often more effective than portfolio-wide programs. They typically yield 40–100 bps of improvement.

Why These Results Are Real

You might be sceptical. Claims-level analytics sound good, but are these results actually achievable? Yes, and here’s why:

First, these improvements come from better information, not from cutting benefits or denying care. You’re not saying “no” to claims you were previously approving. You’re identifying inefficiencies and fixing them.

Second, these improvements are additive. You don’t get all 100–200 bps from one intervention. You get 40–80 bps from medical management optimisation, 20–60 bps from provider network optimisation, and 40–100 bps from cohort-specific interventions. The total is 100–240 bps, depending on where you start and how aggressively you execute.

Third, these improvements are sustainable. You’re not one-time savings. Once you’ve optimised your medical management programs, renegotiated provider contracts, and implemented cohort-specific interventions, those savings stick. You might get another 50–100 bps in year two as you refine your programs and tackle new opportunities.

Finally, and most importantly, these improvements are documented across multiple health insurers. The research on the analytics revolution in health insurance and health insurers using data analytics to lower costs shows that insurers with advanced analytics capabilities consistently outperform peers on loss ratios and profitability. D23.io’s Superset stack is one implementation of this approach.

Building Your Analytics Stack: Data Architecture Essentials

If you’re thinking about implementing D23.io or building your own claims analytics stack, you need to understand the underlying data architecture. Here’s what you need:

Cloud Data Warehouse

Your claims data needs to live in a cloud data warehouse: Snowflake, BigQuery, Redshift, or similar. This is non-negotiable. You can’t do sub-minute query times on billions of claim records in a traditional on-premise database.

Why the cloud? Three reasons:

  1. Scalability: Cloud warehouses scale to petabytes of data. Your claims history going back 5–10 years might be 10–50 TB. Cloud warehouses handle this easily.
  2. Performance: Cloud warehouses use columnar storage and massively parallel processing. Queries that would take hours on a traditional database run in seconds.
  3. Cost: You pay for compute and storage separately, so you can scale up for complex queries and scale down when you’re not using it. This is much cheaper than buying and maintaining on-premise hardware.

Most health insurers are moving to Snowflake or BigQuery. Snowflake is popular because it’s easy to set up and has good integrations with BI tools. BigQuery is popular because it’s deeply integrated with Google Cloud and has strong machine learning capabilities.

ETL Pipeline

You need a pipeline to extract claims data from your source systems (claims adjudication platform, medical management system, PBM), transform it (normalise codes, deduplicate records, calculate derived fields), and load it into your warehouse.

This is typically done with tools like Talend, Informatica, dbt (data build tool), or Apache Airflow. The key is that this pipeline needs to run reliably, every day, and handle the complexity of health insurance claims data (multiple claim types, multiple source systems, evolving data schemas).

For most health insurers, the ETL pipeline is the most complex part of the implementation. Getting claims data clean and deduplicated takes time and expertise.

BI Layer

On top of your data warehouse, you need a BI tool that lets actuaries and analysts query and visualise data. Popular options include Tableau, Looker, Power BI, and Superset (an open-source BI tool).

The key requirements are:

  1. Speed: Sub-second query times. If your BI tool is slow, people won’t use it.
  2. Ease of use: Your actuaries and analysts should be able to build dashboards and run ad-hoc queries without help from IT.
  3. Drill-down capabilities: You should be able to start at portfolio-level metrics and drill down to individual claims.
  4. Scheduled reporting: You should be able to automatically generate reports and send them to stakeholders.

D23.io typically uses Tableau or Looker for the BI layer, but the specifics depend on your existing tech stack and preferences.

Data Governance and Security

Health insurance claims data is sensitive. It includes medical diagnoses, treatment details, and personally identifiable information. You need strong data governance and security controls.

This includes:

  1. Access controls: Only authorised users (actuaries, medical directors, finance teams) can access claims data. Role-based access control (RBAC) is essential.
  2. Encryption: Data in transit and at rest should be encrypted.
  3. Audit logging: All access to claims data should be logged and auditable.
  4. Data masking: For non-actuarial users (e.g., dashboards for executives), personally identifiable information should be masked or aggregated.
  5. Compliance: Your data warehouse and BI layer should comply with HIPAA (in the US) and relevant privacy laws in Australia.

D23.io’s Superset stack includes these controls, but you need to ensure they’re properly configured for your environment. This is where working with a partner like PADISO (who specialises in AI automation for insurance claims processing and risk assessment) can be valuable. They can help you design and implement the data architecture, ETL pipeline, and BI layer in a way that’s secure, scalable, and compliant.

Actuary and Operations Team Integration

Having a claims analytics platform is only half the battle. You also need to integrate it into your actuarial and operations workflows. This is where many health insurers stumble.

Actuary Workflows

Your chief actuary and actuarial team need to have D23.io’s Superset stack built into their daily workflow. This means:

  1. Dashboards for regular monitoring: Pre-built dashboards that show loss ratio trends, claims by diagnosis, provider costs, pharmacy spend, etc. These should refresh daily and be reviewed by the actuarial team every morning.
  2. Ad-hoc query capabilities: When the chief actuary has a question (“Why did loss ratios spike in Q3?”), they should be able to answer it in minutes, not weeks.
  3. Integration with pricing and reserving: Your actuaries should use claims analytics insights to inform your pricing models, reserve calculations, and financial projections.
  4. Trend analysis and forecasting: Your actuaries should use historical claims data to forecast future claims costs and loss ratios.

This requires training and change management. Your actuaries need to understand how to use the BI tool, how to interpret the data, and how to translate insights into business decisions.

Operations Team Workflows

Your operations team (medical directors, care management leaders, network managers) also need to integrate claims analytics into their workflows:

  1. Medical management dashboards: Dashboards showing prior auth approval rates, utilisation review outcomes, case management impact, disease management ROI.
  2. Provider performance dashboards: Dashboards showing provider costs, quality metrics, network utilisation, and outlier alerts.
  3. Cohort analysis for targeted interventions: Identifying high-cost cohorts and designing targeted programs.
  4. Real-time alerts: Alerting on anomalies (e.g., a sudden spike in emergency room utilisation) so the team can investigate and respond.

Again, this requires training and integration into existing workflows. Your medical directors need to understand how to use the data to make clinical decisions. Your network managers need to understand how to use provider analytics to negotiate contracts and manage the network.

Governance and Decision-Making

Finally, you need governance structures that translate claims analytics insights into business decisions and actions. This typically includes:

  1. Weekly or bi-weekly analytics reviews: The chief actuary, chief medical officer, and CFO review key metrics and discuss insights.
  2. Quarterly business reviews: Deeper dives into trends, cohort analysis, and program ROI.
  3. Annual actuarial valuation: Using claims analytics insights to inform pricing, reserving, and financial projections for the next year.
  4. Program evaluation and reallocation: Regular evaluation of medical management, disease management, and other programs, with reallocation of resources based on ROI.

Without this governance, claims analytics insights will sit in dashboards and never translate into action. With it, you’ll see the 100–200 bps loss ratio improvements we’ve discussed.

Compliance, Security, and Audit Readiness

Health insurance claims data is heavily regulated. You need to ensure your claims analytics platform is compliant with relevant regulations and audit-ready.

HIPAA Compliance (US)

In the US, health insurance claims data is protected health information (PHI) under HIPAA. Your claims analytics platform needs to comply with HIPAA’s Security Rule, which requires:

  1. Access controls: Unique user identification, emergency access procedures, user suspension, and role-based access control.
  2. Audit controls: Hardware, software, and procedural mechanisms to record and examine activity in information systems containing PHI.
  3. Integrity controls: Mechanisms to ensure that PHI is not improperly altered or destroyed.
  4. Transmission security: Encryption and other safeguards for PHI transmitted over electronic networks.

D23.io’s Superset stack, when properly configured, can meet these requirements. But you need to ensure that:

  1. Your cloud data warehouse (Snowflake, BigQuery) has HIPAA-compliant configurations.
  2. Your ETL pipeline is secure and doesn’t expose PHI.
  3. Your BI layer has strong access controls and audit logging.
  4. Your data warehouse backups are encrypted and stored securely.

Privacy Laws (Australia and Other Jurisdictions)

In Australia, health information is protected under the Privacy Act and the Australian Privacy Principles (APPs). The key requirements are:

  1. Consent: You need to have the individual’s consent to collect, use, and disclose their health information.
  2. Purpose limitation: You can only use health information for the purpose it was collected for.
  3. Data security: You need to take reasonable steps to protect health information from misuse, loss, unauthorised access, modification, or disclosure.
  4. Transparency: You need to be transparent about how you collect, use, and disclose health information.

D23.io’s Superset stack, when properly configured, can meet these requirements. But you need to ensure that your privacy policy and data governance are aligned with the Privacy Act and APPs.

Audit Readiness

Health insurers are regularly audited by regulators, external auditors, and compliance teams. Your claims analytics platform needs to be audit-ready, which means:

  1. Documented data flows: You need to document where claims data comes from, how it’s transformed, where it’s stored, and who has access to it.
  2. Audit logging: All access to claims data should be logged and auditable.
  3. Change management: Changes to your analytics models, dashboards, and data governance should be documented and approved.
  4. Testing and validation: Your analytics models should be tested and validated to ensure they’re accurate.
  5. Compliance documentation: You should have documentation showing that your platform complies with relevant regulations (HIPAA, Privacy Act, etc.).

If you’re working with a partner like PADISO on your AI automation for insurance claims processing and risk assessment, they can help you build audit-ready documentation and compliance controls from the start. This is much cheaper than trying to retrofit compliance later.

Getting Started: Implementation Timeline and Costs

Implementing D23.io’s Superset stack for claims analytics is a significant project. Here’s what you need to know about timeline and costs:

Implementation Timeline

A typical implementation takes 4–6 months from start to production:

  • Month 1: Discovery and Planning (2–4 weeks)

    • Understand your current claims systems and data.
    • Define requirements for the analytics platform.
    • Design the data architecture (cloud warehouse, ETL pipeline, BI layer).
    • Plan the implementation roadmap.
  • Month 2–3: Data Architecture and ETL Pipeline (6–8 weeks)

    • Set up your cloud data warehouse (Snowflake, BigQuery, etc.).
    • Build the ETL pipeline to extract, transform, and load claims data.
    • Clean and validate the data.
    • Load 2–3 years of historical claims data.
  • Month 3–4: Analytics Models and BI Layer (4–6 weeks)

    • Build analytics models for key use cases (loss ratio by cohort, provider costs, etc.).
    • Build dashboards and reports in your BI tool (Tableau, Looker, etc.).
    • Test the dashboards and reports with your actuarial and operations teams.
  • Month 5: Pilot and Refinement (2–4 weeks)

    • Pilot the platform with a subset of users (chief actuary, key analysts).
    • Gather feedback and refine dashboards and models.
    • Train users on how to use the platform.
  • Month 6: Production Launch and Rollout (2–4 weeks)

    • Launch the platform to all users.
    • Provide ongoing support and training.
    • Monitor performance and fix issues.

This timeline assumes you have a dedicated project team and clear requirements. If you’re starting from scratch with minimal analytics infrastructure, add another 2–3 months. If you already have a data warehouse and BI tools in place, you might compress the timeline to 3–4 months.

Implementation Costs

Implementation costs vary widely depending on the complexity of your claims data and your existing infrastructure. Here’s a rough breakdown:

  • Cloud data warehouse: $10,000–$50,000 for the first year (depending on data volume and query complexity).
  • ETL pipeline and data engineering: $100,000–$300,000 (labour costs for building and testing the ETL pipeline).
  • BI tool and dashboards: $20,000–$100,000 for the first year (software licenses and labour for building dashboards).
  • D23.io and analytics consulting: $150,000–$500,000 (depending on the scope of analytics models and the level of support).
  • Training and change management: $50,000–$150,000 (training your team on how to use the platform).

Total first-year cost: $330,000–$1,100,000.

This sounds expensive, but remember: a 100–200 bps improvement in loss ratio on a $100 million premium base is $1–2 million to the bottom line. Even on a $500 million premium base, it’s $5–10 million. The ROI is typically 5–10x in the first year.

Also, ongoing costs (year 2 and beyond) are much lower: cloud data warehouse ($10,000–$50,000), BI tool licenses ($20,000–$100,000), and some ongoing analytics support ($50,000–$150,000). Once you’ve built the platform, the ongoing costs are relatively modest.

Common Pitfalls and How to Avoid Them

Health insurers implementing claims analytics platforms often run into common pitfalls. Here’s how to avoid them:

Pitfall 1: Data Quality Issues

Problem: Claims data is messy. Diagnosis and procedure codes are inconsistent, member IDs are duplicated across systems, provider names are spelled differently in different systems. If your data is dirty, your analytics will be wrong.

Solution: Invest heavily in data quality upfront. Spend time understanding your claims data, identifying data quality issues, and building data validation and cleansing rules into your ETL pipeline. Don’t launch your platform until you’re confident your data is clean.

Pitfall 2: Lack of User Adoption

Problem: You build a beautiful analytics platform, but your actuaries and operations teams don’t use it. They continue relying on spreadsheets and legacy reports.

Solution: Involve users early and often. During the discovery and design phase, understand their workflows and pain points. Build dashboards that address their specific needs. Train them thoroughly. Provide ongoing support. Make it easy to use the platform—if it’s faster to pull data from a spreadsheet than to use your BI tool, users won’t adopt it.

Pitfall 3: Over-Engineering

Problem: You try to build the perfect analytics platform from day one. You design every possible dashboard, every possible analytics model. The project drags on for 12+ months and costs $2+ million. By the time you launch, the business has moved on.

Solution: Start with an MVP (minimum viable product). Identify the top 3–5 use cases (e.g., loss ratio by cohort, provider costs, medical management ROI) and build those first. Launch in 4–6 months. Then iterate and expand based on user feedback.

Pitfall 4: Siloed Analytics

Problem: Your analytics team builds dashboards for your actuaries, but they don’t integrate into the business decision-making process. Insights sit in dashboards and never translate into action.

Solution: Build governance structures that translate analytics insights into business decisions. Have weekly or bi-weekly analytics reviews where the chief actuary, chief medical officer, and CFO discuss key metrics and insights. Make it clear that decisions should be data-driven.

Pitfall 5: Ignoring Security and Compliance

Problem: You build a great analytics platform, but it’s not secure or compliant with HIPAA, Privacy Act, etc. You get audited and find major compliance gaps.

Solution: Build security and compliance into your platform from day one. Work with your compliance and security teams to ensure your data warehouse, ETL pipeline, and BI layer are secure and audit-ready. Don’t treat compliance as an afterthought.

Next Steps: Moving from Insight to Action

If you’re a health insurer considering implementing D23.io’s Superset stack or building your own claims analytics platform, here’s what to do next:

1. Assess Your Current State

Understand your current analytics capabilities and pain points:

  • What claims analytics do you currently have? (Actuarial reports, dashboards, ad-hoc analyses?)
  • What questions can’t you answer with your current systems?
  • How long does it take to answer a typical question from your actuaries or operations team?
  • What’s your current loss ratio, and how has it trended over the past 3–5 years?

2. Define Your Objectives

What do you want to achieve with claims analytics?

  • Improve loss ratio by X basis points in the next 12 months?
  • Identify and address high-cost cohorts?
  • Optimise your medical management programs?
  • Improve provider network performance?
  • Enhance pricing and reserving accuracy?

Be specific. “Improve analytics” is too vague. “Reduce loss ratios by 100 bps in the next 12 months by identifying high-cost cohorts and implementing targeted disease management programs” is specific.

3. Evaluate Options

You have three main options:

  1. Buy a platform like D23.io: Fastest path to value, but you’re dependent on a vendor. Good if you want to launch in 4–6 months.
  2. Build your own stack: More control, but requires significant engineering resources. Good if you have strong data engineering and analytics teams.
  3. Hybrid approach: Use D23.io or a similar platform for the data warehouse and ETL, but build your own analytics models and dashboards. Good middle ground.

Evaluate each option against your objectives, timeline, budget, and internal capabilities.

4. Build a Business Case

Quantify the benefits and costs:

  • Benefits: How much will you save if you improve loss ratios by 100–200 bps? (For a $500 million premium base, that’s $5–10 million.)
  • Costs: Implementation costs ($330,000–$1,100,000 in year one), ongoing costs ($80,000–$300,000 per year).
  • ROI: First-year ROI is typically 5–10x. Payback period is typically 6–12 months.

Use this business case to secure executive sponsorship and budget.

5. Start with a Pilot

Don’t try to implement claims analytics across your entire organisation at once. Start with a pilot:

  • Identify a specific use case (e.g., “Reduce loss ratios for members with diabetes by 100 bps in the next 12 months”).
  • Build analytics models and dashboards for that use case.
  • Pilot with a subset of users (chief actuary, medical director for that condition).
  • Measure results (did you achieve your objective?).
  • If successful, expand to other use cases.

This de-risks the project and lets you prove value before scaling.

6. Partner with Experts

Claims analytics is complex. Consider partnering with experts who can help you:

  • Design your data architecture and analytics models.
  • Build and maintain your ETL pipeline.
  • Implement security and compliance controls.
  • Train your team on how to use the platform.

Partners like PADISO specialise in exactly this kind of work. They can help you implement claims analytics platforms and integrate them into your business workflows. They can also help with broader AI automation for insurance claims processing and risk assessment, which can complement your claims analytics platform.

Conclusion

Claims analytics is not new. Health insurers have been doing claims analysis for decades. But D23.io’s Superset stack and similar modern platforms are fundamentally different from legacy approaches.

Legacy actuarial systems were built for reporting: aggregated metrics, static reports, long turnaround times. Modern claims analytics platforms are built for exploration: claims-level data, interactive dashboards, sub-minute query times.

This difference matters. With legacy systems, your actuaries might ask, “What was our loss ratio in Q3?” and get an answer in two weeks. With modern analytics, they can ask, “Why was our loss ratio 200 bps higher in Q3, and which cohorts drove it?” and get an answer in minutes.

This speed and granularity translates directly to better business decisions. You can identify emerging cost drivers before they become problems. You can design targeted interventions for high-cost cohorts. You can measure the ROI of your medical management and disease management programs. You can negotiate more effectively with your provider network.

The result, across Australian and US health insurers already using this approach, is a 100–200 basis point improvement in loss ratios within a single hold period. That’s $1–2 million per $100 million of premium. For a mid-market or regional insurer, that’s a material improvement in profitability.

If you’re a health insurer looking to improve your loss ratio and gain competitive advantage, claims analytics should be a priority. D23.io’s Superset stack is one option. Building your own stack is another. But either way, the time to act is now. Your competitors are already doing this. If you’re not, you’re falling behind.

Ready to get started? Assess your current state, define your objectives, build a business case, and start with a pilot. Within 6–12 months, you’ll have a modern claims analytics platform that’s driving real business results.