The Honest Answer: Cloud Can Be More Expensive

A company migrates its on-premises SQL Server warehouse to Azure SQL. Year 1 cloud cost: $180K. Previous on-premises cost: $120K. The cloud is 50% more expensive — because: the VM runs 24/7 (on-premises hardware was already paid for), SQL Server licensing in the cloud is expensive (no hybrid benefit applied), storage isn't tiered (everything on premium storage instead of hot/warm/cold), and nobody right-sized the compute (the VM has 64 cores because that's what the on-premises server had — even though peak usage is 20% of capacity). After optimization: cloud cost drops to $85K/year. Savings: 29% vs. on-premises + eliminated hardware refresh ($300K saved every 5 years) + elastic scaling + managed operations. The lesson: cloud is cheaper when optimized. Cloud is more expensive when lifted-and-shifted without architecture changes.

Cloud doesn't automatically save money. Cloud saves money when you architect for the cloud — auto-scaling, right-sized compute, storage tiering, and reserved instances. Lift-and-shift without optimization often costs 30-50% more than on-premises. — Xylity Data Engineering Practice

TCO Components: What to Include

ComponentOn-PremisesCloud WarehouseLakehouse
ComputeServer hardware (amortized)VM/cluster hourly rateCapacity units (CU/DBU)
StorageSAN/NAS (amortized)Managed storage ($/GB/mo)Object storage ($/GB/mo)
SoftwareDatabase license + ETL licenseService fee (often includes license)Platform fee (often includes everything)
OperationsDBA + storage admin + backupReduced (managed service)Minimal (fully managed)
NetworkingLAN (minimal)Egress charges + VPNSame as cloud WH
FacilitiesRack space, power, coolingNone (provider)None (provider)
DR/HASecondary site + replicationBuilt-in (geo-redundant)Built-in (geo-redundant)
Refresh cycle$500K-3M every 4-5 yearsNone (always current)None (always current)

On-Premises TCO: The Full Picture

On-premises costs that organizations frequently undercount: hardware refresh (the $1.5M Teradata node amortized over 5 years = $300K/year — but the refresh also requires 6 months of planning and execution, consuming 2 FTEs for half a year), software licensing (SQL Server Enterprise at $15K/core × 16 cores = $240K/year; Oracle at $47K/core × 8 cores = $376K/year; Teradata per-TB pricing at $30K/TB × 20TB = $600K/year), operations staff (1-2 DBAs at $130-180K each, storage admin at $120K — total $380-480K/year in salary + benefits), facilities (rack space, power, cooling — often buried in IT overhead but real: $10-30K/year per rack), and opportunity cost (the capabilities you can't have: AI/ML, real-time analytics, elastic scaling, self-service — unquantified but significant). Total on-premises TCO for a 5TB warehouse with 50 users: $400-800K/year depending on platform and staffing.

Cloud Warehouse TCO: Synapse, Snowflake, BigQuery

Azure Synapse Dedicated Pool: DW1000c (~$12/hour) × 12 hours/day × 22 business days = $3,168/month = $38K/year for compute. Storage: 5TB at $0.025/GB/month = $1,500/year. Total: ~$40K/year (plus networking and governance). Snowflake: Medium warehouse (~$4/credit) × 8 hours/day × 22 days × 4 credits/hour = $2,816/month = $34K/year compute. Storage: 5TB at $23/TB/month = $1,380/year. Total: ~$35K/year. Optimization impact: Auto-suspend (compute pauses when idle — reduces cost 40-60%), right-sizing (match compute to actual workload, not peak capacity), and reserved capacity (1-year commitment saves 30-40%). Optimized cloud warehouse TCO: $25-50K/year for a 5TB, 50-user deployment — vs. $400-800K on-premises.

Lakehouse TCO: Fabric, Databricks

Microsoft Fabric F64: $5,995/month = $72K/year. Includes: compute (Spark, SQL, Power BI), storage (OneLake), governance (Purview integration), and Power BI capacity. With reserved instance (1-year): ~$43K/year (40% savings). Databricks: Jobs cluster ($0.15-0.40/DBU × 8-16 DBUs × 12 hours/day) + SQL Warehouse ($0.22-0.55/DBU × 8-32 DBUs × 8 hours/day) + cloud storage ($1,500/year for 5TB). Total: $50-120K/year depending on workload mix. With reserved capacity: 30-40% savings.

Lakehouse cost advantage: The lakehouse includes data engineering capabilities (Spark, pipelines, streaming) that would cost $30-60K/year separately with cloud warehouse + Azure Data services. The all-inclusive nature of Fabric and Databricks means the lakehouse TCO comparison should include the eliminated cost of separate data engineering tools.

Side-by-Side: 5TB, 50 Users, 3-Year TCO

ComponentOn-PremisesCloud WarehouseLakehouse (Fabric)
Year 1 (setup + run)$600K (existing, amortized)$150K (migration + run)$200K (migration + run)
Year 2$500K$50K$55K
Year 3$500K + $1.5M refresh$50K$55K
3-Year Total$3.1M$250K$310K
CapabilitiesBI onlyBI + some self-serviceBI + ML + streaming + self-service

The 3-year insight: On-premises includes a hardware refresh in Year 3 that alone exceeds the entire 3-year cloud or lakehouse cost. Even without the refresh: annual operating cost is 5-10x higher on-premises due to licensing, staffing, and facilities. The lakehouse costs slightly more than cloud warehouse but includes data engineering capabilities that the cloud warehouse requires separate services for — making lakehouse the better value for organizations that need analytics + engineering.

Hidden Costs: What TCO Calculators Miss

Cloud hidden costs: Egress charges (moving data out of the cloud — $0.05-0.09/GB adds up at scale), cross-region replication (DR in a second region doubles storage cost), premium support ($10K-100K/year for enterprise support plans), and training (the team needs cloud skills they may not have — $5K-15K per person for certification and training). On-premises hidden costs: Opportunity cost (what revenue/efficiency could you achieve with AI/ML/real-time that the legacy platform can't provide?), insurance cost (the risk of hardware failure, data loss, or security breach on aging infrastructure), and migration tax (the longer you wait, the more expensive the eventual migration — every year adds data volume, dependencies, and complexity).

6 Cost Optimization Levers

1

Reserved Instances

1-year or 3-year commitments save 30-60% on compute — the single largest cost reduction lever for predictable workloads.

2

Auto-Scaling and Auto-Pause

Scale compute to zero when not in use. The warehouse that runs 24/7 at full capacity costs 3x more than one that scales during business hours and pauses overnight.

3

Storage Tiering

Hot storage for frequently accessed data (current month). Cool for archived data (older than 90 days). Archive for compliance retention. Tiering reduces storage cost 50-70%.

4

Right-Sizing Compute

Profile actual usage: if peak CPU is 30% of the provisioned capacity, downsize. Right-sizing reduces compute cost 40-60% for most workloads.

5

Workload Isolation

Separate ETL compute from query compute. ETL runs on cheaper batch compute; queries run on interactive compute sized for actual concurrency.

6

Data Lifecycle Management

Archive data older than the analytical window. A warehouse that retains 7 years of data costs 3.5x more than one that retains 2 years and archives the rest.

The FinOps Practice: Continuous Cloud Cost Optimization

Cloud cost optimization isn't a one-time project — it's an ongoing practice (FinOps). The FinOps cycle: inform (visibility into cloud spend — dashboards showing cost by service, by team, by workload; allocated to business units, not just a single IT line item), optimize (monthly optimization reviews — identify unused resources, right-size over-provisioned compute, apply new reserved instance opportunities), and operate (budget alerts, spending anomaly detection, governance policies that prevent cost surprises). For data platforms specifically: monitor compute utilization (Spark clusters, SQL warehouses running at 20% utilization should be right-sized), storage growth rate (is data lifecycle management working? is archived data actually moving to cold tier?), and per-query cost (which reports cost the most to run? are they worth it?). Organizations practicing FinOps reduce cloud data platform costs 20-30% annually through continuous optimization — beyond the initial architecture savings.

Making the Business Case: CFO-Ready TCO Presentation

The TCO presentation that gets CFO approval: start with the problem (hardware refresh is $1.5M in 18 months; annual operating cost is growing 8% per year; the team can't hire Teradata specialists), present the comparison (3-year TCO: on-premises $X, cloud warehouse $Y, lakehouse $Z — with all cost components itemized), show the capability delta (not just cost savings — what new business capabilities does the modern platform enable that the legacy platform can't?), include risk quantification (what's the cost of doing nothing? hardware failure risk, security breach on unpatched systems, inability to hire talent), and propose phased investment (not $1M on day one — Phase 1 at $200K proves the concept, Phase 2 at $300K migrates core workloads, Phase 3 at $200K completes the transformation). The phased approach reduces perceived risk — the CFO approves Phase 1 with a gate review before Phase 2. Each phase delivers measurable results that justify the next.

When Cloud Is NOT Cheaper: Scenarios to Watch

Cloud costs more than on-premises in specific scenarios: always-on, compute-heavy workloads (a Spark cluster running 24/7 at 100% utilization costs more than equivalent on-premises hardware — cloud's advantage is elasticity, which always-on workloads don't use), massive egress (if the architecture requires moving 10TB+ out of the cloud monthly — to on-premises users, to partner systems, or across regions — egress charges dominate the bill), unoptimized deployment (cloud resources provisioned at peak capacity and never right-sized — the $200K/year on-premises bill becomes a $300K/year cloud bill without optimization), and licensing traps (SQL Server Enterprise licensing in the cloud without Azure Hybrid Benefit costs 2-3x more than on-premises — always apply hybrid benefit for existing license holders). The honest TCO analysis accounts for these scenarios. If your workload is 24/7 compute-heavy with high egress and you don't plan to optimize — on-premises may genuinely be cheaper. For the 80% of organizations with variable workloads, elastic demand, and willingness to optimize: cloud is 30-50% cheaper.

The Xylity Approach

We analyze data platform TCO with the full-cost framework — compute, storage, licensing, operations, facilities, refresh cycles, and opportunity cost. Our data architects and Fabric/Databricks specialists design the optimized architecture — reserved instances, auto-scaling, storage tiering, and right-sized compute — ensuring cloud migration delivers the 30-50% cost reduction that justifies the investment.

Continue building your understanding with these related resources from our consulting practice.

Cloud Saves Money — When Architected Right

Full TCO analysis, 6 optimization levers, 3-year comparison. Data platform cost framework that proves the business case.

Start Your TCO Analysis →