|By Brad Vaughan||
|April 16, 2011 11:45 AM EDT||
A classic use of ROI or its twin TCO is in the Microsoft Economics of the Cloud, Nov 2010 paper. The conclusion is you can improve TCO by up to 80% by using applications in public cloud versus on-premise deployment. The basics of the calculation being:
- improved utilization (10% to 90%) enabled by virtualization/consolidation & elasticity
- the economies (power, operations, HW purchase etc..) of scale of multi-tenant cloud scale hosting
Given most costs in a DC are directly linked to the amount of infrastructure deployed, then improving utilization from 10% to 90% sounds like the primary justification for the 80% improvement. The misuse of the information is not more evident that when Phil Wainewright writes that the strength of this "research" is enough to put the nail in the coffin of the concept of private cloud. Definitive words indeed.. The problem I have with this conclusion is it is a black & white, monolithic view of "what is cloud". This is combined with TCO/ROI modeling that uses some pretty broad assumptions to underpin the cost model. It is often good marketing or publicity to offer polarized view of the issues, but it does not provide a real-world executable decision making capability (read "value") for future consumer of cloud services (public or private). So why are people needing to kill the concept of the "private cloud"? James Urquhart tweeted it best (4/12/2011):
- "I don't hear ANYONE who isn't a public cloud provider or SI who bet the farm on public cloud make any claims about "false clouds". Period."
- "Oh, wait. There may be one or two startups and journalists in there...all of which stand to gain from public cloud. Sorry about that. :\ "
If you take that approach and as a result just build a "cloud" for the sake of "cloud" then you are making "the BIG mistake". Implementing a framework as a product is doomed to fail. If you implemented SOA this way then disaster, ITIL would equal chaos, Prince2 would create inertia, Web2.0 would have resulted in mega-$$$. These concepts, whether they be architectural, process or other are meant to guide execution, not be implemented blindly. So how should ROI be used? So when people ask the question. "What's is the ROI of the cloud?" it is not an issue of "What is an ROI?" or "What is the benefit of Cloud?" or even "What data goes into a ROI calculation?". It is about how to answer the question of why, what and how to adopt the cloud. Most of the Cloud ROI (return-on-investment) or TCO (total cost of ownership) discussions are like the whitepaper from Microsoft. Comparing side by side a complex cloud deployment with a traditional infrastructure deployment. In reality, it's too difficult to develop a model to cater "true total cost of ownership", you quickly have to jump to broad assumptions, and narrow scope to make it manageable. If you start you model as a greenfield cloud deployment, your model has radical inaccuracies as you try apply this to brownfield or legacy enterprises. Try starting based on data of a legacy deployment and you have huge problems dealing with the depreciation of assets. Brownfield models also have the challenge of dealing with the elasticity of the delivery assets or opportunity costs; for example, you can manage 100 or 150 servers with the same team, or your existing 20% utilized asset can possible only support 2X or maybe as much as 10X the workload. You then overlay this with the changing economics of real estate facilities, HVAC, compute. The result is, you end up with a model that can have error factors upward of 100% It's too complex a problem to solve without a huge dataset to validate the variables, dependencies, etc... Armada takes a Fast Track approach to solving the problem. You are looking at cloud as a reference framework to help develop a solution that returns the business value. You calculate ROI based on a specific situation and end-state solution. A ROI needs to have a pay back of less than a year, so long-term theoretical modeling has no significant value. So how do you do it? Remember three things;
- You must have a triggering event
- Use scenario analysis and not lifecycle modeling
- Apply the 80/20 rule to data, and only the stuff that impacts your costs
Most of the time being a technical architect in consulting creates looks of skepticism from engineers in enterprise customers. Fair enough, when I was in that seat I felt the same way. When I gave up internal politics for politics of "revenue/pipeline", "everyone is a salesperson" and "whitepapers and webinars" a few things became pretty crystal clear. The most important thing is, don't waste time doing anything unless there is a pain point, problem to solve, triggering event. Wants are good, but needs are better. This is important in ROI calculation. The triggering event is the anchor point for the evaluation and defines where you are looking for the biggest "return" in the ROI.. The triggering event can be something specific like;
- "we will run out of datacenter space in 6 months"
- "it takes us 6 months to deploy an environment"
- "we are on CNN if our primary datacenter fails because we have no DR"
Alternatively, it can be softer and described as the business goal, business driver like:
- "we need to reduce operational management costs"
- "we need to improve our infrastructure utilization"
These things are scoping statements for the project and then the ROI is applied to the return for this project.
You scope the project, but if you try and calculate the return based on lifecycle costs over a long term, you will be scratching your head forever. If the ROI is not 1-3 years, then you are probably not doing it. Most likely it needs to be in less than a year. Scenario analysis is fairly simple, but a little time consuming. It is, however, a step down the direction of implementation, rather than a detour into developing a business case that will never be used or validated later. You create three (3) scenarios:
- Business as usual - sometimes this is the "no decision, decision" or just solve the problem the way you have in the past
- Option 1 - the "go big or go home" scenario, build the pure play cloud solution
- Option 2 - the "pragmatic solution", or sometimes called the cheap solution. This is often the winner, but generally can be folded into option 1 after a subsequent round of funding.
Gather the requirements. Design the end-state architectures for three options and price out the implementation and on-going costs. You are already starting the design, so when the solution is green lighted, you are ready to go..
80/20 Rule of Data
A basic premise of the Fast Track method is to make decisions based on readily available information. Creating data and model takes time and effort for little return. In the time it takes to do this, IT services are evolving and changing. So in collecting data for a ROI analysis, use what is available, don't over process it and limit yourself to the data that impacts your business. From Gartner and other models we know that the biggest contributors to ROI/TCO are;
- Hardware Costs (storage, compute, network)
- Hardware Maintenance/Support
- Software License (applications licenses, tools licenses etc..)
- Software Maintenance/Support
- Management & Operations (people, benefits etc..)
- Facilities (real estate, hvac, security etc..)
- Development/Customization/System Integration
- Opportunity Cost (increase costs in existing infrastructure by reducing its scale)
Focus on capturing this information to support the scope of your project. If your project is not looking for value in reduction of power costs, then don't include it in the model. Just deliver the value you have visbility and control over. You should try and be as complete as possible, without creating an environment of political inertia. So with this approach its easy to capture a return on investment (ROI) calculation. I need to add, that David Linthicum wrote a very relevant post that reinforced that ROI does not make a business case. You need to also include the soft value factors, which for the cloud revolve around agility and time-to-market. Hard to define or place a value, but critical to the final assessment.
Between 2005 and 2020, data volumes will grow by a factor of 300 – enough data to stack CDs from the earth to the moon 162 times. This has come to be known as the ‘big data’ phenomenon. Unfortunately, traditional approaches to handling, storing and analyzing data aren’t adequate at this scale: they’re too costly, slow and physically cumbersome to keep up. Fortunately, in response a new breed of technology has emerged that is cheaper, faster and more scalable. Yet, in meeting these new needs they...
Dec. 4, 2016 12:30 AM EST Reads: 1,773
The cloud promises new levels of agility and cost-savings for Big Data, data warehousing and analytics. But it’s challenging to understand all the options – from IaaS and PaaS to newer services like HaaS (Hadoop as a Service) and BDaaS (Big Data as a Service). In her session at @BigDataExpo at @ThingsExpo, Hannah Smalltree, a director at Cazena, provided an educational overview of emerging “as-a-service” options for Big Data in the cloud. This is critical background for IT and data professionals...
Dec. 3, 2016 11:00 PM EST Reads: 4,131
"Once customers get a year into their IoT deployments, they start to realize that they may have been shortsighted in the ways they built out their deployment and the key thing I see a lot of people looking at is - how can I take equipment data, pull it back in an IoT solution and show it in a dashboard," stated Dave McCarthy, Director of Products at Bsquare Corporation, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 3, 2016 11:00 PM EST Reads: 949
@DevOpsSummit taking place June 6-8, 2017 at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. @DevOpsSummit at Cloud Expo New York Call for Papers is now open.
Dec. 3, 2016 09:30 PM EST Reads: 1,764
The cloud competition for database hosts is fierce. How do you evaluate a cloud provider for your database platform? In his session at 18th Cloud Expo, Chris Presley, a Solutions Architect at Pythian, gave users a checklist of considerations when choosing a provider. Chris Presley is a Solutions Architect at Pythian. He loves order – making him a premier Microsoft SQL Server expert. Not only has he programmed and administered SQL Server, but he has also shared his expertise and passion with b...
Dec. 3, 2016 08:00 PM EST Reads: 3,953
In his session at 19th Cloud Expo, Claude Remillard, Principal Program Manager in Developer Division at Microsoft, contrasted how his team used config as code and immutable patterns for continuous delivery of microservices and apps to the cloud. He showed how the immutable patterns helps developers do away with most of the complexity of config as code-enabling scenarios such as rollback, zero downtime upgrades with far greater simplicity. He also demoed building immutable pipelines in the cloud ...
Dec. 3, 2016 08:00 PM EST Reads: 1,744
As data explodes in quantity, importance and from new sources, the need for managing and protecting data residing across physical, virtual, and cloud environments grow with it. Managing data includes protecting it, indexing and classifying it for true, long-term management, compliance and E-Discovery. Commvault can ensure this with a single pane of glass solution – whether in a private cloud, a Service Provider delivered public cloud or a hybrid cloud environment – across the heterogeneous enter...
Dec. 3, 2016 06:15 PM EST Reads: 1,513
"IoT is going to be a huge industry with a lot of value for end users, for industries, for consumers, for manufacturers. How can we use cloud to effectively manage IoT applications," stated Ian Khan, Innovation & Marketing Manager at Solgeniakhela, in this SYS-CON.tv interview at @ThingsExpo, held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 3, 2016 05:30 PM EST Reads: 4,045
Today we can collect lots and lots of performance data. We build beautiful dashboards and even have fancy query languages to access and transform the data. Still performance data is a secret language only a couple of people understand. The more business becomes digital the more stakeholders are interested in this data including how it relates to business. Some of these people have never used a monitoring tool before. They have a question on their mind like “How is my application doing” but no id...
Dec. 3, 2016 05:15 PM EST Reads: 2,138
@GonzalezCarmen has been ranked the Number One Influencer and @ThingsExpo has been named the Number One Brand in the “M2M 2016: Top 100 Influencers and Brands” by Onalytica. Onalytica analyzed tweets over the last 6 months mentioning the keywords M2M OR “Machine to Machine.” They then identified the top 100 most influential brands and individuals leading the discussion on Twitter.
Dec. 3, 2016 05:15 PM EST Reads: 2,001
In IT, we sometimes coin terms for things before we know exactly what they are and how they’ll be used. The resulting terms may capture a common set of aspirations and goals – as “cloud” did broadly for on-demand, self-service, and flexible computing. But such a term can also lump together diverse and even competing practices, technologies, and priorities to the point where important distinctions are glossed over and lost.
Dec. 3, 2016 04:30 PM EST Reads: 1,476
Predictive analytics tools monitor, report, and troubleshoot in order to make proactive decisions about the health, performance, and utilization of storage. Most enterprises combine cloud and on-premise storage, resulting in blended environments of physical, virtual, cloud, and other platforms, which justifies more sophisticated storage analytics. In his session at 18th Cloud Expo, Peter McCallum, Vice President of Datacenter Solutions at FalconStor, discussed using predictive analytics to mon...
Dec. 3, 2016 04:00 PM EST Reads: 4,862
All clouds are not equal. To succeed in a DevOps context, organizations should plan to develop/deploy apps across a choice of on-premise and public clouds simultaneously depending on the business needs. This is where the concept of the Lean Cloud comes in - resting on the idea that you often need to relocate your app modules over their life cycles for both innovation and operational efficiency in the cloud. In his session at @DevOpsSummit at19th Cloud Expo, Valentin (Val) Bercovici, CTO of Soli...
Dec. 3, 2016 03:30 PM EST Reads: 1,591
Information technology is an industry that has always experienced change, and the dramatic change sweeping across the industry today could not be truthfully described as the first time we've seen such widespread change impacting customer investments. However, the rate of the change, and the potential outcomes from today's digital transformation has the distinct potential to separate the industry into two camps: Organizations that see the change coming, embrace it, and successful leverage it; and...
Dec. 3, 2016 03:15 PM EST Reads: 3,225
Without a clear strategy for cost control and an architecture designed with cloud services in mind, costs and operational performance can quickly get out of control. To avoid multiple architectural redesigns requires extensive thought and planning. Boundary (now part of BMC) launched a new public-facing multi-tenant high resolution monitoring service on Amazon AWS two years ago, facing challenges and learning best practices in the early days of the new service. In his session at 19th Cloud Exp...
Dec. 3, 2016 03:00 PM EST Reads: 485
The Internet of Things (IoT) promises to simplify and streamline our lives by automating routine tasks that distract us from our goals. This promise is based on the ubiquitous deployment of smart, connected devices that link everything from industrial control systems to automobiles to refrigerators. Unfortunately, comparatively few of the devices currently deployed have been developed with an eye toward security, and as the DDoS attacks of late October 2016 have demonstrated, this oversight can ...
Dec. 3, 2016 02:45 PM EST Reads: 715
Extracting business value from Internet of Things (IoT) data doesn’t happen overnight. There are several requirements that must be satisfied, including IoT device enablement, data analysis, real-time detection of complex events and automated orchestration of actions. Unfortunately, too many companies fall short in achieving their business goals by implementing incomplete solutions or not focusing on tangible use cases. In his general session at @ThingsExpo, Dave McCarthy, Director of Products...
Dec. 3, 2016 02:45 PM EST Reads: 530
Regulatory requirements exist to promote the controlled sharing of information, while protecting the privacy and/or security of the information. Regulations for each type of information have their own set of rules, policies, and guidelines. Cloud Service Providers (CSP) are faced with increasing demand for services at decreasing prices. Demonstrating and maintaining compliance with regulations is a nontrivial task and doing so against numerous sets of regulatory requirements can be daunting task...
Dec. 3, 2016 02:45 PM EST Reads: 727
Fact: storage performance problems have only gotten more complicated, as applications not only have become largely virtualized, but also have moved to cloud-based infrastructures. Storage performance in virtualized environments isn’t just about IOPS anymore. Instead, you need to guarantee performance for individual VMs, helping applications maintain performance as the number of VMs continues to go up in real time. In his session at Cloud Expo, Dhiraj Sehgal, Product and Marketing at Tintri, sha...
Dec. 3, 2016 02:45 PM EST Reads: 787
Machine Learning helps make complex systems more efficient. By applying advanced Machine Learning techniques such as Cognitive Fingerprinting, wind project operators can utilize these tools to learn from collected data, detect regular patterns, and optimize their own operations. In his session at 18th Cloud Expo, Stuart Gillen, Director of Business Development at SparkCognition, discussed how research has demonstrated the value of Machine Learning in delivering next generation analytics to impr...
Dec. 3, 2016 02:15 PM EST Reads: 6,957