|By Bryan Semple||
|July 14, 2011 08:00 AM EDT||
Organizations that are seeking to deploy cloud-based business models for their infrastructure face unique capacity management challenges. This article will review these challenges to enable cloud providers, either public or private, to avoid the pitfalls of improper capacity management. While the article discusses both types of clouds, the needs of private cloud providers will be especially highlighted due to the unique challenges they face with this business model.
IT Becomes a Business within a Business
For years, we have been hearing "IT must behave more like a business." A hypervisor's ability to deliver utility computing moves this vision closer to reality. Using hypervisors to deploy an infrastructure cloud fundamentally changes the relationship between IT and their customers. Application portability combined with competing cloud offerings from companies like Amazon change the dynamic between application owners and corporate IT. If end users can't explicitly move their applications between cloud providers, they can at least compare prices and service levels between providers. Right or wrong, Amazon S3 becomes a measuring stick in price, performance and service for IT organizations.
To deliver an Amazon S3 experience requires significant retooling in IT processes. Capacity management is one of the areas requiring retooling. Virtualization alone causes changes to capacity management (for more information, see http://www.vkernel.com/solutions/capacity-planning). But virtualization deployed as an infrastructure cloud adds nuances to the capacity management problem. Capacity management for cloud providers is unique for the following five reasons:
- Capacity monitoring in addition to planning
- Chargeback is mandatory
- Efficiency drives return on assets
- Tenant reporting requirements are unique
- Optimization is a value add
Variable Demand Drives Criticality of Capacity Monitoring
Cloud deployments of virtualization technology introduce many operational changes for IT administrators. The first is a change in the amount of control IT has over the loads deployed on their hardware. With cloud deployments, either public or private, end users deploy applications using self-service portals as they see fit, load them as they desire and consume resources at whatever pace they need.
Hence, unlike the careful P2V sizing process undertaken for the first wave of virtualization where applications are sized, scheduled and deployed in a methodical manner, clouds have no careful sizing or timing that the cloud operator is aware of. Applications of unknown sizes appear, consume resources, and may just as quickly disappear. Without adequate capacity, these applications will fail to perform to customer expectations. Without real-time capacity monitoring, application deployment can dramatically impact other applications.
Hence capacity planning, a well-thought-out process for making sure there is sufficient capacity en masse, must be supplemented with capacity monitoring. Capacity monitoring is a real-time process that takes raw performance and utilization data and transforms it into actionable information concerning system-level capacity requirements. Without capacity monitoring, system administrators are left to interpret real-time utilization metrics from individual virtual machines. VKernel's research has shown that properly monitoring capacity in real time involves collecting over 20 metrics per VM at least 10 times per hour, and keeping this information for at least 30 days. A 100 VM environment would require about 17 million data points to accurately monitor capacity in the environment. This capacity monitoring, however, is a must-have to augment standard capacity planning and prevent performance issues from impacting the cloud.
For any cloud where resources can be deployed in a self-service fashion, charging back for resources becomes a necessity. Without a method to chargeback or show back, self-service clouds would quickly find themselves at capacity since resources are essentially free.
But chargeback is a tricky area. For commercial cloud and private cloud operators, charging back for allocated resources is fairly straightforward. But, since the private chargeback operator is simply shifting costs insides the company and not impacting the bottom line, the motivations for chargeback are different. The public cloud operator is indifferent to allocated resources and utilized resources. If the public cloud operator is charging for an allocated resource pool, they make money. In fact, the higher the ratio between allocated and utilized, the more over allocation of resources is possible and the higher the profit margins. For private cloud operators, however, the goal is to actually lower the costs for the company. Hence, the private cloud operator wants the allocated resource usage to be very close to actual usage to drive resource efficiency. Highlighting the difference between actual resource usage and allocated resource usage shows internal business units the amount of corporate resources they are wasting. This motivation can then be used to right size environments and reduce overall IT spend.
While chargeback is important, cloud operators need to be mindful of what they charge. For public operators, there are competitive pressures. For private operators, charging provides a way to directly compare internal IT costs vs. external costs such as Amazon.
Is a simple $/CPU comparison between internal clouds and Amazon a fair comparison? Does Amazon contain the same level of compliance? Of control? Is the company comfortable with data outside the company firewall?
More important, chargeback for private cloud operators is primarily a means to minimize the difference between allocated and utilized resources to drive up efficiencies and VM densities. Chargeback or showback becomes a control mechanism as opposed to an actual financial transfer mechanism. Hence the rate of chargeback is not as important as the difference between allocated and actual usage.
Even here, the challenges for private cloud operators are greater. Let's assume a private cloud operator hosts 200 internal customers. Assume each of these internal customers is wasting 50% of their resource allocation. On an individual basis, the absolute value of the wasted resources may be insignificant. But across all 200 customers, the magnitude of the IT spend could be quite large. The greater good theory for IT would require that IT actually reduce resource usage for all the internal customers to claim additional savings for the company despite what the internal customers want. Private cloud operators must operate not only for their internal customers' needs, but also for the company's needs.
Setting rates for chargeback is the final tricky area for cloud operators. For a public cloud operator, the rate needs to be competitive, provide some profit margin and match customer's value. Easy enough. But what about private cloud operators? Once again, being a private operator makes things difficult. What are the rates for chargeback for a private cloud operator? Is the goal to set rates to make an internal profit when 50% of the VMs slots are filled? 75%? But if the internal cost center is making a profit, is that the right thing to do from a budgeting standpoint? Is the goal cost reclamation or efficiency?
The net of this is that for cloud operators, chargeback is critical. For private cloud operators, chargeback's purpose needs to be clearly defined to align IT not only with their customer's goals, but also the broader corporate goals.
Capacity Planning Impacts Revenue and Cost
Customers expecting to use a cloud service have high expectations with regards to time to deploy a service. For public clouds, this expectation will be a nearly instant deployment after the service request. Private cloud operators may not have quite an instantaneous expectation for their customers. Either way, the "acceptable" wait time of several weeks to deploy a new server is gone. Immediate is the word, not days.
To enable this immediate capability, sufficient capacity must be on hand to deploy new virtual machines based on both steady state and unexpected increases in demands. To meet this accelerated time duration, a high degree of capacity planning must take place to predict future capacity needs ahead of demand and allow for the slower process of procuring and installing physical servers, networks and storage.
It's easy to meet rapid deployment expectations by over procuring hardware. The danger here is that over procurement impacts cash flows and profitability for a cloud. Having large amounts of depreciating assets sitting around is not a sound business strategy. If these assets are plugged in and configured, the added power costs worsen an already bad situation. Under procuring hardware is just as bad since cloud operators will be unable to meet customer needs should their systems not have available capacity. Hence the goal is to have a solid understanding of consumption and then apply a safety factor to allow for unexpected demand.
Understanding capacity needs across the entire IT infrastructure is important. But it's just as important to understand where there are available VM slots for the best placement of VMs from a performance and a capacity standpoint. Utilizing available VM slot reporting ensures the performance of the running VMs and increases the VM density per host, which is critical to achieving a high return on assets.
Capacity planning is critical to cloud operators to generate a high return on assets while also meeting customer demand for near instantaneous deployment requests.
With a standard virtualized environment, the IT organization may report on environment status to a few senior IT leaders. Not so with an infrastructure cloud. For public clouds and private clouds, there is a greater expectation of visibility into the environment. Reporting for cloud tenants could involve availability, resource allocation, resource utilization, current charges, and pricing plan. The amount of information revealed to a customer could depend on the business philosophy or type of cloud. For a private cloud, revealing differences between allocated resources and utilized resources and the savings a customer could achieve by reducing their resources allocation makes sense for a company trying to save money. For a public cloud provider, however, suggesting ways to reduce resource allocations may not be in the provider's interest.
Beyond questions around what type of information to provide is the manner in which information is provided. For public cloud operators, online portals are most likely the reporting distribution mechanism of choice. For private clouds, however, information needs to flow seamlessly into the enterprises existing reporting infrastructure. This could involve connections with SharePoint, with IT service management frameworks, internal portals, or simple email distribution of reports on an ongoing basis.
Optimization Is a Value Add
Many times, when virtual machines are first deployed, they are over allocated CPU, memory and storage. In a cloud deployment, this over allocation does not cost the cloud operator revenue. In fact, just the opposite occurs. The cloud operator deploys the requested resources, but despite not being used, the cloud operator still collects the revenue. As discussed earlier, the cloud operator can decide whether to reveal this to the tenant or not.
While the cloud operator may not care about wasted resources to an application, the end customer does as does the firm's CFO. Decreasing requested CPU, memory and storage reduces monthly tenant costs. Hence, optimization becomes an optional value-add service for the provider to offer tenants to reduce spending. This would be the equivalent of your cell phone company contacting you to suggest a lower monthly plan. While it lowers the monthly revenue of the cloud operator, it dramatically improves customer loyalty.
Because of the increased need for chargeback, monitoring, capacity planning, and reporting, capacity management takes on added urgency for cloud operators. Capacity management performed correctly enables cloud operators to maximize compute cycles delivered to a customer at the lowest possible cost and, thus, capacity management is a key building block for any cloud implementation.
SYS-CON Events announced today that Peak 10, Inc., a national IT infrastructure and cloud services provider, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. Peak 10 provides reliable, tailored data center and network services, cloud and managed services. Its solutions are designed to scale and adapt to customers’ changing business needs, enabling them to lower costs, improve performance and focus inter...
May. 1, 2016 12:00 AM EDT Reads: 1,053
You think you know what’s in your data. But do you? Most organizations are now aware of the business intelligence represented by their data. Data science stands to take this to a level you never thought of – literally. The techniques of data science, when used with the capabilities of Big Data technologies, can make connections you had not yet imagined, helping you discover new insights and ask new questions of your data. In his session at @ThingsExpo, Sarbjit Sarkaria, data science team lead ...
Apr. 30, 2016 10:45 PM EDT Reads: 677
So, you bought into the current machine learning craze and went on to collect millions/billions of records from this promising new data source. Now, what do you do with them? Too often, the abundance of data quickly turns into an abundance of problems. How do you extract that "magic essence" from your data without falling into the common pitfalls? In her session at @ThingsExpo, Natalia Ponomareva, Software Engineer at Google, will provide tips on how to be successful in large scale machine lear...
Apr. 30, 2016 10:00 PM EDT Reads: 1,091
If there is anything we have learned by now, is that every business paves their own unique path for releasing software- every pipeline, implementation and practices are a bit different, and DevOps comes in all shapes and sizes. Software delivery practices are often comprised of set of several complementing (or even competing) methodologies – such as leveraging Agile, DevOps and even a mix of ITIL, to create the combination that’s most suitable for your organization and that maximize your busines...
Apr. 30, 2016 08:15 PM EDT Reads: 1,800
Struggling to keep up with increasing application demand? Learn how Platform as a Service (PaaS) can streamline application development processes and make resource management easy.
Apr. 30, 2016 08:00 PM EDT Reads: 2,060
Whether your IoT service is connecting cars, homes, appliances, wearable, cameras or other devices, one question hangs in the balance – how do you actually make money from this service? The ability to turn your IoT service into profit requires the ability to create a monetization strategy that is flexible, scalable and working for you in real-time. It must be a transparent, smoothly implemented strategy that all stakeholders – from customers to the board – will be able to understand and comprehe...
Apr. 30, 2016 04:45 PM EDT Reads: 1,055
See storage differently! Storage performance problems have only gotten worse and harder to solve as applications have become largely virtualized and moved to a cloud-based infrastructure. Storage performance in a virtualized environment is not just about IOPS, it is about how well that potential performance is guaranteed to individual VMs for these apps as the number of VMs keep going up real time. In his session at 18th Cloud Expo, Dhiraj Sehgal, in product and marketing at Tintri, will discu...
Apr. 30, 2016 04:15 PM EDT Reads: 755
Machine Learning helps make complex systems more efficient. By applying advanced Machine Learning techniques such as Cognitive Fingerprinting, wind project operators can utilize these tools to learn from collected data, detect regular patterns, and optimize their own operations. In his session at 18th Cloud Expo, Stuart Gillen, Director of Business Development at SparkCognition, will discuss how research has demonstrated the value of Machine Learning in delivering next generation analytics to im...
Apr. 30, 2016 04:15 PM EDT Reads: 1,667
Up until last year, enterprises that were looking into cloud services usually undertook a long-term pilot with one of the large cloud providers, running test and dev workloads in the cloud. With cloud’s transition to mainstream adoption in 2015, and with enterprises migrating more and more workloads into the cloud and in between public and private environments, the single-provider approach must be revisited. In his session at 18th Cloud Expo, Yoav Mor, multi-cloud solution evangelist at Cloudy...
Apr. 30, 2016 03:30 PM EDT Reads: 1,438
There is an ever-growing explosion of new devices that are connected to the Internet using “cloud” solutions. This rapid growth is creating a massive new demand for efficient access to data. And it’s not just about connecting to that data anymore. This new demand is bringing new issues and challenges and it is important for companies to scale for the coming growth. And with that scaling comes the need for greater security, gathering and data analysis, storage, connectivity and, of course, the...
Apr. 30, 2016 03:15 PM EDT Reads: 763
This is not a small hotel event. It is also not a big vendor party where politicians and entertainers are more important than real content. This is Cloud Expo, the world's longest-running conference and exhibition focused on Cloud Computing and all that it entails. If you want serious presentations and valuable insight about Cloud Computing for three straight days, then register now for Cloud Expo.
Apr. 30, 2016 02:30 PM EDT Reads: 1,715
Redis is not only the fastest database, but it has become the most popular among the new wave of applications running in containers. Redis speeds up just about every data interaction between your users or operational systems. In his session at 18th Cloud Expo, Dave Nielsen, Developer Relations at Redis Labs, will shares the functions and data structures used to solve everyday use cases that are driving Redis' popularity.
Apr. 30, 2016 02:00 PM EDT Reads: 1,082
SYS-CON Events announced today that Stratoscale, the software company developing the next generation data center operating system, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. Stratoscale is revolutionizing the data center with a zero-to-cloud-in-minutes solution. With Stratoscale’s hardware-agnostic, Software Defined Data Center (SDDC) solution to store everything, run anything and scale everywhere...
Apr. 30, 2016 01:15 PM EDT Reads: 1,532
[session] Don’t Forget the Ops: Build Operations into Your Cloud By @BMCSoftware | @CloudExpo #Cloud
Many private cloud projects were built to deliver self-service access to development and test resources. While those clouds delivered faster access to resources, they lacked visibility, control and security needed for production deployments. In their session at 18th Cloud Expo, Steve Anderson, Product Manager at BMC Software, and Rick Lefort, Principal Technical Marketing Consultant at BMC Software, will discuss how a cloud designed for production operations not only helps accelerate developer...
Apr. 30, 2016 01:00 PM EDT Reads: 1,125
Angular 2 is a complete re-write of the popular framework AngularJS. Programming in Angular 2 is greatly simplified – now it's a component-based well-performing framework. This immersive one-day workshop at 18th Cloud Expo, led by Yakov Fain, a Java Champion and a co-founder of the IT consultancy Farata Systems and the product company SuranceBay, will provide you with everything you wanted to know about Angular 2.
Apr. 30, 2016 12:45 PM EDT Reads: 1,703
SYS-CON Events announced today that Men & Mice, the leading global provider of DNS, DHCP and IP address management overlay solutions, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. The Men & Mice Suite overlay solution is already known for its powerful application in heterogeneous operating environments, enabling enterprises to scale without fuss. Building on a solid range of diverse platform support,...
Apr. 30, 2016 12:00 PM EDT Reads: 2,295
SYS-CON Events announced today that Enzu, a leading provider of cloud hosting solutions, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. Enzu’s mission is to be the leading provider of enterprise cloud solutions worldwide. Enzu enables online businesses to use its IT infrastructure to their competitive advantage. By offering a suite of proven hosting and management services, Enzu wants companies to foc...
Apr. 30, 2016 12:00 PM EDT Reads: 970
Peak 10, Inc., has announced the implementation of IT service management, a business process alignment initiative based on the widely adopted Information Technology Infrastructure Library (ITIL) framework. The implementation of IT service management enhances Peak 10’s current service-minded approach to IT delivery by propelling the company to deliver higher levels of personalized and prompt service. The majority of Peak 10’s operations employees have been trained and certified in the ITIL frame...
Apr. 30, 2016 12:00 PM EDT Reads: 919
You deployed your app with the Bluemix PaaS and it's gaining some serious traction, so it's time to make some tweaks. Did you design your application in a way that it can scale in the cloud? Were you even thinking about the cloud when you built the app? If not, chances are your app is going to break. Check out this webcast to learn various techniques for designing applications that will scale successfully in Bluemix, for the confidence you need to take your apps to the next level and beyond.
Apr. 30, 2016 11:30 AM EDT Reads: 1,468
SYS-CON Events announced today that Ericsson has been named “Gold Sponsor” of SYS-CON's @ThingsExpo, which will take place on June 7-9, 2016, at the Javits Center in New York, New York. Ericsson is a world leader in the rapidly changing environment of communications technology – providing equipment, software and services to enable transformation through mobility. Some 40 percent of global mobile traffic runs through networks we have supplied. More than 1 billion subscribers around the world re...
Apr. 30, 2016 11:00 AM EDT Reads: 878