Enterprises leverage cloud providers’ compute and storage services for their ad-hoc data analytics, streaming analytics and ML use cases as cloud data lakes provide significant cost advantages, agility and scale from the get-go. Proof of Concepts (POCs) for data-driven initiatives start easily and without any huge upfront bill. But over time as projects mature or ad hoc queries take longer or model iteration cycles increase, the seemingly endless supply of underlying resources leads to wasteful expenditure on compute and resources.
The usage comes with cost unpredictability and lacks financial governance and thus negatively impacts TCO. In the cloud, rising costs are not necessarily bad; it means that the data team is using more services, which theoretically means the team is doing more “good stuff” and hopefully is delivering business value. TCO optimization makes sure that wasteful spending is identified and eventually eliminated. Cloud data lake platforms should be able to help enterprises keep check on this wasteful spending to lower TCO. Admins should be able to do the following for optimizing TCO within their data lake platforms:
- Control and design the infrastructure spend at will, override policy, preference, or autonomous self-learning
- Leverage built-in capabilities to optimize clusters for lower infrastructure spend based on custom-defined parameters
- Monitor total costs at the application, user, account, cluster, cluster-instance level to drive accountability and meaningful discussions across teams
- Identify areas of cost optimization to drive maximum performance for the lowest TCO
SAVING $230M+ ANNUALLY IN CLOUD COSTS
Watch this 6-minute overview to see how Qubole’s open data lake platform saved $230m+ in cloud costs for customers in 2019 alone.
As platforms provide these core TCO focussed capabilities, it should be autonomous and policy-based TCO optimization without sacrificing Service Level Agreements (SLAs).
With Qubole, the open data lake platform, enterprises address all 4 key above requirements for optimizing TCO by:
- Reducing costs continuously in an automated manner based on set or default policy, preference, and autonomous self-learning.
- Optimizing the consumption of resources consistently like performance improvements to the underlying engine so that jobs are completed efficiently.
- Finding and consuming lower-priced resources on a continual basis with workload-aware autoscaling; admin-defined heterogeneous cluster configurations and only provision resources when needed, whether On-demand or Spot.
- Eliminating unnecessary resource consumption with aggressive downscaling, optimized upscaling and at-will shut down.
- Throttling queries based on monetary limits based on the budget set by the administrator.
- Providing insights for user, job and cluster level cost metrics in a multi-tenant environment to do data-driven show back discussion.
In summary, a cloud data lake platform should be able to understand what is currently happening and build a financial profile of your cloud spending, help put measures in place to control spending and optimize by taking the advantage of cloud data platform facilities to reduce costs and improve overall TCO.