|By Bryan Semple||
|July 14, 2011 08:00 AM EDT||
Organizations that are seeking to deploy cloud-based business models for their infrastructure face unique capacity management challenges. This article will review these challenges to enable cloud providers, either public or private, to avoid the pitfalls of improper capacity management. While the article discusses both types of clouds, the needs of private cloud providers will be especially highlighted due to the unique challenges they face with this business model.
IT Becomes a Business within a Business
For years, we have been hearing "IT must behave more like a business." A hypervisor's ability to deliver utility computing moves this vision closer to reality. Using hypervisors to deploy an infrastructure cloud fundamentally changes the relationship between IT and their customers. Application portability combined with competing cloud offerings from companies like Amazon change the dynamic between application owners and corporate IT. If end users can't explicitly move their applications between cloud providers, they can at least compare prices and service levels between providers. Right or wrong, Amazon S3 becomes a measuring stick in price, performance and service for IT organizations.
To deliver an Amazon S3 experience requires significant retooling in IT processes. Capacity management is one of the areas requiring retooling. Virtualization alone causes changes to capacity management (for more information, see http://www.vkernel.com/solutions/capacity-planning). But virtualization deployed as an infrastructure cloud adds nuances to the capacity management problem. Capacity management for cloud providers is unique for the following five reasons:
- Capacity monitoring in addition to planning
- Chargeback is mandatory
- Efficiency drives return on assets
- Tenant reporting requirements are unique
- Optimization is a value add
Variable Demand Drives Criticality of Capacity Monitoring
Cloud deployments of virtualization technology introduce many operational changes for IT administrators. The first is a change in the amount of control IT has over the loads deployed on their hardware. With cloud deployments, either public or private, end users deploy applications using self-service portals as they see fit, load them as they desire and consume resources at whatever pace they need.
Hence, unlike the careful P2V sizing process undertaken for the first wave of virtualization where applications are sized, scheduled and deployed in a methodical manner, clouds have no careful sizing or timing that the cloud operator is aware of. Applications of unknown sizes appear, consume resources, and may just as quickly disappear. Without adequate capacity, these applications will fail to perform to customer expectations. Without real-time capacity monitoring, application deployment can dramatically impact other applications.
Hence capacity planning, a well-thought-out process for making sure there is sufficient capacity en masse, must be supplemented with capacity monitoring. Capacity monitoring is a real-time process that takes raw performance and utilization data and transforms it into actionable information concerning system-level capacity requirements. Without capacity monitoring, system administrators are left to interpret real-time utilization metrics from individual virtual machines. VKernel's research has shown that properly monitoring capacity in real time involves collecting over 20 metrics per VM at least 10 times per hour, and keeping this information for at least 30 days. A 100 VM environment would require about 17 million data points to accurately monitor capacity in the environment. This capacity monitoring, however, is a must-have to augment standard capacity planning and prevent performance issues from impacting the cloud.
For any cloud where resources can be deployed in a self-service fashion, charging back for resources becomes a necessity. Without a method to chargeback or show back, self-service clouds would quickly find themselves at capacity since resources are essentially free.
But chargeback is a tricky area. For commercial cloud and private cloud operators, charging back for allocated resources is fairly straightforward. But, since the private chargeback operator is simply shifting costs insides the company and not impacting the bottom line, the motivations for chargeback are different. The public cloud operator is indifferent to allocated resources and utilized resources. If the public cloud operator is charging for an allocated resource pool, they make money. In fact, the higher the ratio between allocated and utilized, the more over allocation of resources is possible and the higher the profit margins. For private cloud operators, however, the goal is to actually lower the costs for the company. Hence, the private cloud operator wants the allocated resource usage to be very close to actual usage to drive resource efficiency. Highlighting the difference between actual resource usage and allocated resource usage shows internal business units the amount of corporate resources they are wasting. This motivation can then be used to right size environments and reduce overall IT spend.
While chargeback is important, cloud operators need to be mindful of what they charge. For public operators, there are competitive pressures. For private operators, charging provides a way to directly compare internal IT costs vs. external costs such as Amazon.
Is a simple $/CPU comparison between internal clouds and Amazon a fair comparison? Does Amazon contain the same level of compliance? Of control? Is the company comfortable with data outside the company firewall?
More important, chargeback for private cloud operators is primarily a means to minimize the difference between allocated and utilized resources to drive up efficiencies and VM densities. Chargeback or showback becomes a control mechanism as opposed to an actual financial transfer mechanism. Hence the rate of chargeback is not as important as the difference between allocated and actual usage.
Even here, the challenges for private cloud operators are greater. Let's assume a private cloud operator hosts 200 internal customers. Assume each of these internal customers is wasting 50% of their resource allocation. On an individual basis, the absolute value of the wasted resources may be insignificant. But across all 200 customers, the magnitude of the IT spend could be quite large. The greater good theory for IT would require that IT actually reduce resource usage for all the internal customers to claim additional savings for the company despite what the internal customers want. Private cloud operators must operate not only for their internal customers' needs, but also for the company's needs.
Setting rates for chargeback is the final tricky area for cloud operators. For a public cloud operator, the rate needs to be competitive, provide some profit margin and match customer's value. Easy enough. But what about private cloud operators? Once again, being a private operator makes things difficult. What are the rates for chargeback for a private cloud operator? Is the goal to set rates to make an internal profit when 50% of the VMs slots are filled? 75%? But if the internal cost center is making a profit, is that the right thing to do from a budgeting standpoint? Is the goal cost reclamation or efficiency?
The net of this is that for cloud operators, chargeback is critical. For private cloud operators, chargeback's purpose needs to be clearly defined to align IT not only with their customer's goals, but also the broader corporate goals.
Capacity Planning Impacts Revenue and Cost
Customers expecting to use a cloud service have high expectations with regards to time to deploy a service. For public clouds, this expectation will be a nearly instant deployment after the service request. Private cloud operators may not have quite an instantaneous expectation for their customers. Either way, the "acceptable" wait time of several weeks to deploy a new server is gone. Immediate is the word, not days.
To enable this immediate capability, sufficient capacity must be on hand to deploy new virtual machines based on both steady state and unexpected increases in demands. To meet this accelerated time duration, a high degree of capacity planning must take place to predict future capacity needs ahead of demand and allow for the slower process of procuring and installing physical servers, networks and storage.
It's easy to meet rapid deployment expectations by over procuring hardware. The danger here is that over procurement impacts cash flows and profitability for a cloud. Having large amounts of depreciating assets sitting around is not a sound business strategy. If these assets are plugged in and configured, the added power costs worsen an already bad situation. Under procuring hardware is just as bad since cloud operators will be unable to meet customer needs should their systems not have available capacity. Hence the goal is to have a solid understanding of consumption and then apply a safety factor to allow for unexpected demand.
Understanding capacity needs across the entire IT infrastructure is important. But it's just as important to understand where there are available VM slots for the best placement of VMs from a performance and a capacity standpoint. Utilizing available VM slot reporting ensures the performance of the running VMs and increases the VM density per host, which is critical to achieving a high return on assets.
Capacity planning is critical to cloud operators to generate a high return on assets while also meeting customer demand for near instantaneous deployment requests.
With a standard virtualized environment, the IT organization may report on environment status to a few senior IT leaders. Not so with an infrastructure cloud. For public clouds and private clouds, there is a greater expectation of visibility into the environment. Reporting for cloud tenants could involve availability, resource allocation, resource utilization, current charges, and pricing plan. The amount of information revealed to a customer could depend on the business philosophy or type of cloud. For a private cloud, revealing differences between allocated resources and utilized resources and the savings a customer could achieve by reducing their resources allocation makes sense for a company trying to save money. For a public cloud provider, however, suggesting ways to reduce resource allocations may not be in the provider's interest.
Beyond questions around what type of information to provide is the manner in which information is provided. For public cloud operators, online portals are most likely the reporting distribution mechanism of choice. For private clouds, however, information needs to flow seamlessly into the enterprises existing reporting infrastructure. This could involve connections with SharePoint, with IT service management frameworks, internal portals, or simple email distribution of reports on an ongoing basis.
Optimization Is a Value Add
Many times, when virtual machines are first deployed, they are over allocated CPU, memory and storage. In a cloud deployment, this over allocation does not cost the cloud operator revenue. In fact, just the opposite occurs. The cloud operator deploys the requested resources, but despite not being used, the cloud operator still collects the revenue. As discussed earlier, the cloud operator can decide whether to reveal this to the tenant or not.
While the cloud operator may not care about wasted resources to an application, the end customer does as does the firm's CFO. Decreasing requested CPU, memory and storage reduces monthly tenant costs. Hence, optimization becomes an optional value-add service for the provider to offer tenants to reduce spending. This would be the equivalent of your cell phone company contacting you to suggest a lower monthly plan. While it lowers the monthly revenue of the cloud operator, it dramatically improves customer loyalty.
Because of the increased need for chargeback, monitoring, capacity planning, and reporting, capacity management takes on added urgency for cloud operators. Capacity management performed correctly enables cloud operators to maximize compute cycles delivered to a customer at the lowest possible cost and, thus, capacity management is a key building block for any cloud implementation.
DevOps Summit 2015 New York, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that it is now accepting Keynote Proposals. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long development cycles that produce software that is obsolete...
Dec. 18, 2014 09:45 PM EST Reads: 845
“DevOps is really about the business. The business is under pressure today, competitively in the marketplace to respond to the expectations of the customer. The business is driving IT and the problem is that IT isn't responding fast enough," explained Mark Levy, Senior Product Marketing Manager at Serena Software, in this SYS-CON.tv interview at DevOps Summit, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 18, 2014 08:00 PM EST Reads: 1,124
“We help people build clusters, in the classical sense of the cluster. We help people put a full stack on top of every single one of those machines. We do the full bare metal install," explained Greg Bruno, Vice President of Engineering and co-founder of StackIQ, in this SYS-CON.tv interview at 15th Cloud Expo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 18, 2014 02:30 PM EST Reads: 825
The cloud is becoming the de-facto way for enterprises to leverage common infrastructure while innovating and one of the biggest obstacles facing public cloud computing is security. In his session at 15th Cloud Expo, Jeff Aliber, a global marketing executive at Verizon, discussed how the best place for web security is in the cloud. Benefits include: Functions as the first layer of defense Easy operation –CNAME change Implement an integrated solution Best architecture for addressing network-l...
Dec. 18, 2014 02:00 PM EST Reads: 871
Mobile commerce traffic is surpassing desktop, yet less than 20% of sales in the U.S. are mobile commerce sales. In his session at 15th Cloud Expo, Dan Franklin, Segment Manager, Commerce, at Verizon Digital Media Services, defined mobile devices and discussed how next generation means simplification. It means taking your digital content and turning it into instantly gratifying experiences.
Dec. 18, 2014 12:00 PM EST Reads: 1,016
“In the past year we've seen a lot of stabilization of WebRTC. You can now use it in production with a far greater degree of certainty. A lot of the real developments in the past year have been in things like the data channel, which will enable a whole new type of application," explained Peter Dunkley, Technical Director at Acision, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 18, 2014 11:30 AM EST Reads: 889
SYS-CON Events announced today that Windstream, a leading provider of advanced network and cloud communications, has been named “Silver Sponsor” of SYS-CON's 16th International Cloud Expo®, which will take place on June 9–11, 2015, at the Javits Center in New York, NY. Windstream (Nasdaq: WIN), a FORTUNE 500 and S&P 500 company, is a leading provider of advanced network communications, including cloud computing and managed services, to businesses nationwide. The company also offers broadband, p...
Dec. 18, 2014 11:00 AM EST Reads: 2,112
The major cloud platforms defy a simple, side-by-side analysis. Each of the major IaaS public-cloud platforms offers their own unique strengths and functionality. Options for on-site private cloud are diverse as well, and must be designed and deployed while taking existing legacy architecture and infrastructure into account. Then the reality is that most enterprises are embarking on a hybrid cloud strategy and programs. In this Power Panel at 15th Cloud Expo (http://www.CloudComputingExpo.com...
Dec. 18, 2014 10:30 AM EST Reads: 2,240
Verizon Enterprise Solutions is simplifying the cloud-purchasing experience for its clients, with the launch of Verizon Cloud Marketplace, a key foundational component of the company's robust ecosystem of enterprise-class technologies. The online storefront will initially feature pre-built cloud-based services from AppDynamics, Hitachi Data Systems, Juniper Networks, PfSense and Tervela. Available globally to enterprises using Verizon Cloud, Verizon Cloud Marketplace provides a one-stop shop fo...
Dec. 18, 2014 10:30 AM EST Reads: 1,794
Leysin American School is an exclusive, private boarding school located in Leysin, Switzerland. Leysin selected an OpenStack-powered, private cloud as a service to manage multiple applications and provide development environments for students across the institution. Seeking to meet rigid data sovereignty and data integrity requirements while offering flexible, on-demand cloud resources to users, Leysin identified OpenStack as the clear choice to round out the school's cloud strategy. Additional...
Dec. 18, 2014 10:30 AM EST Reads: 1,840
The Internet of Things is not new. Historically, smart businesses have used its basic concept of leveraging data to drive better decision making and have capitalized on those insights to realize additional revenue opportunities. So, what has changed to make the Internet of Things one of the hottest topics in tech? In his session at @ThingsExpo, Chris Gray, Director, Embedded and Internet of Things, discussed the underlying factors that are driving the economics of intelligent systems. Discover ...
Dec. 18, 2014 10:15 AM EST Reads: 2,068
The move in recent years to cloud computing services and architectures has added significant pace to the application development and deployment environment. When enterprise IT can spin up large computing instances in just minutes, developers can also design and deploy in small time frames that were unimaginable a few years ago. The consequent move toward lean, agile, and fast development leads to the need for the development and operations sides to work very closely together. Thus, DevOps become...
Dec. 18, 2014 10:00 AM EST Reads: 1,884
"Our premise is Docker is not enough. That's not a bad thing - we actually love Docker. At ActiveState all our products are based on open source technology and Docker is an up-and-coming piece of open source technology," explained Bart Copeland, President & CEO of ActiveState Software, in this SYS-CON.tv interview at DevOps Summit at Cloud Expo®, held Nov 4-6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 18, 2014 10:00 AM EST Reads: 1,830
"BSQUARE is in the business of selling software solutions for smart connected devices. It's obvious that IoT has moved from being a technology to being a fundamental part of business, and in the last 18 months people have said let's figure out how to do it and let's put some focus on it, " explained Dave Wagstaff, VP & Chief Architect, at BSQUARE Corporation, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4-6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 18, 2014 10:00 AM EST Reads: 1,768
SYS-CON Events announced today that AIC, a leading provider of OEM/ODM server and storage solutions, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. AIC is a leading provider of both standard OTS, off-the-shelf, and OEM/ODM server and storage solutions. With expert in-house design capabilities, validation, manufacturing and production, AIC's broad selection of products are highly flexible and are conf...
Dec. 18, 2014 09:45 AM EST Reads: 1,720
SYS-CON Events announced today that IDenticard will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. IDenticard™ is the security division of Brady Corp (NYSE: BRC), a $1.5 billion manufacturer of identification products. We have small-company values with the strength and stability of a major corporation. IDenticard offers local sales, support and service to our customers across the United States and Canada...
Dec. 18, 2014 09:30 AM EST Reads: 1,878
"People are a lot more knowledgeable about APIs now. There are two types of people who work with APIs - IT people who want to use APIs for something internal and the product managers who want to do something outside APIs for people to connect to them," explained Roberto Medrano, Executive Vice President at SOA Software, in this SYS-CON.tv interview at Cloud Expo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 18, 2014 09:00 AM EST Reads: 1,208
“We are a managed services company. We have taken the key aspects of the cloud and the purposed data center and merged the two together and launched the Purposed Cloud about 18–24 months ago," explained Chetan Patwardhan, CEO of Stratogent, in this SYS-CON.tv interview at 15th Cloud Expo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 18, 2014 09:00 AM EST Reads: 1,174
The Internet of Things is a misnomer. That implies that everything is on the Internet, and that simply should not be - especially for things that are blurring the line between medical devices that stimulate like a pacemaker and quantified self-sensors like a pedometer or pulse tracker. The mesh of things that we manage must be segmented into zones of trust for sensing data, transmitting data, receiving command and control administrative changes, and peer-to-peer mesh messaging. In his session a...
Dec. 17, 2014 11:15 PM EST Reads: 1,265