|By Kjel Hanson||
|April 5, 2012 10:30 AM EDT||
Cloud computing has found its way into many organizations as business leaders and IT departments look to capitalize on the many benefits that cloud offers. As your company considers moving all or part of its IT operation to the cloud, a key decision is whether to rely on public cloud, virtual private cloud, or a combination. Finalizing a cloud strategy must start with understanding your objectives and how they best align with the value each offering can provide.
The public cloud can be characterized by IT resources delivered via the Internet using a standardized, self-service, pay-per-use methodology. Public clouds are designed to provide compute resources virtually at will - similar to that of a utility. Public clouds are highly standardized, allow limited customization, and their respective resources can be oversubscribed and massively shared. Workloads requiring inexpensive storage or compute cycles where known response time to the user community is not critical can be a fit with the public cloud.
Virtual private clouds offer scalable compute resources similar to that of public clouds, but in a more controlled environment. Virtual private cloud providers, especially those with managed services around hosted applications, bring insight into the workload and impacts to the infrastructure. Virtual private cloud providers have the flexibility to customize solutions to meet security and performance requirements. They can also identify where customer data is stored, as in a specific data center or country. The setup allows for more customization and delivers a higher-degree of privacy and security.
As you determine which methodology makes the most sense for your business, here are the three major assessment areas to consider and help guide you in your decision.
When it comes to accessing more computing resources, both virtual private and public clouds are designed to provide highly elastic compute power and data storage. When you need more resources, you can request and receive them almost immediately. However, there is a tradeoff since public cloud customers are competing for the same pool of resources. This can impact the cloud experience with unexpected bursts in demand or seasonal type activity. Virtual private cloud providers are able to introduce a level of segmentation to protect workload for a predictable user experience, but still provide the resiliency and flexibility the cloud provides for availability.
Like the public cloud, virtual private cloud services rely on virtualized computing resources to provide elasticity and scale. However, each customer is given its own private pool of resources rather than sharing them. Resources can be expanded, but it is done in a more controlled manner.
Virtual private clouds can offer a degree of elasticity, but also a higher degree of stability than public clouds. This is why virtual private clouds are more attractive for production environments, where the ability to scale is important, but uptime is just as critical.
Another key component to availability is access to the compute resources in the cloud. Traditionally access to the public cloud is done via the Internet. Virtual private cloud providers can be more accommodating for those customers that want to leverage the private line wide area networks currently deployed. With the potential to leverage the Internet as an alternate path to the environment with a dynamic reroute across a hardware-based VPN solution should any carrier issues arise.
Like any utility, public clouds are easily accessible by the masses. Security controls are in place, but with limits as to how much they can control risk. Public clouds thus can be attractive targets for hackers who enjoy the challenge of breaking into public clouds, which they can then use anonymously to attack other sites.
Virtual private clouds offer more security since computing resources are more logically separated. Where virtual private cloud providers are hosting known applications, tighter security at the network layer can be deployed to further reduce the risk of unnecessary traffic. Security zones and firewall rule sets can be deployed to address multi-tenancy concerns of cloud offerings.
As stated above on availability, there is also a higher degree of security with access to the cloud resources and connectivity. Companies accessing the virtual private cloud via virtual private networks or dedicated circuit can beneficial for firms in highly regulated arenas where enterprise data needs to be protected carefully to demonstrate financial and operational stability to regulators and investors.
By design, public clouds give users direct control over the volume of computing resources provisioned: you simply provision what you need when you need it. But, you cannot control what other customers in the resource pool access, which may affect your environment and minimize performance predictability.
Public clouds also make modifications to the underlying infrastructure more challenging. For example, if a technical change is needed, such as a software patch or hardware swap, that change impacts everyone because customers are not isolated from each other. Also there is no coordination with the application MTRs running on top of the infrastructure and how the updates may impact functionality. In addition, customers must diligently control the level of computing resources they contract for, monitoring the resources they need and use, and then requesting resources to be turned off when no longer needed; providing less control over computing costs.
Conversely, a virtual private cloud gives you more control over the performance of the technology environment. Customers can work jointly with virtual private cloud providers to adhere to change control policies that may already be established. Resource allocation and load balancing can be finely tuned based on each customer's environment, usage patterns, and resource consumption.
The environment is also more resilient as more sophisticated redundancy and failover capabilities can be incorporated. Virtual private clouds can also more easily provide degrees of data backup for various data retention policies. Customized solutions for disaster recovery customers based on recovery point and recovery time objectives can all be taken into the design criteria for a solution.
Utility and Consistency Requirements Dictate the Ultimate Choice
If your business requires basic computing resources where uptime and system control are not mission-critical, public clouds can serve as an inexpensive method for rapid provisioning of IT infrastructure. As is the case with most utility companies, public cloud providers offer a serviceable, raw platform at a low cost.
But if you want scalability benefits with more control, virtual private cloud services are much more likely to meet your requirements. Virtual private clouds essentially provide a more consistent experience because providers are more in tune with how their customers use the infrastructure and can plan accordingly. This allows for application performance SLAs where customers can shift their focus away from managing the infrastructure and concentrate on their business. Customers receive the benefits of scale and can leverage the cost savings that cloud provides without all the management issues.
DevOps is often described as a combination of technology and culture. Without both, DevOps isn't complete. However, applying the culture to outdated technology is a recipe for disaster; as response times grow and connections between teams are delayed by technology, the culture will die. A Nutanix Enterprise Cloud has many benefits that provide the needed base for a true DevOps paradigm.
Mar. 26, 2017 12:30 AM EDT Reads: 1,794
As businesses adopt functionalities in cloud computing, it’s imperative that IT operations consistently ensure cloud systems work correctly – all of the time, and to their best capabilities. In his session at @BigDataExpo, Bernd Harzog, CEO and founder of OpsDataStore, will present an industry answer to the common question, “Are you running IT operations as efficiently and as cost effectively as you need to?” He will expound on the industry issues he frequently came up against as an analyst, and...
Mar. 26, 2017 12:00 AM EDT Reads: 4,079
Keeping pace with advancements in software delivery processes and tooling is taxing even for the most proficient organizations. Point tools, platforms, open source and the increasing adoption of private and public cloud services requires strong engineering rigor - all in the face of developer demands to use the tools of choice. As Agile has settled in as a mainstream practice, now DevOps has emerged as the next wave to improve software delivery speed and output. To make DevOps work, organization...
Mar. 26, 2017 12:00 AM EDT Reads: 1,672
The financial services market is one of the most data-driven industries in the world, yet it’s bogged down by legacy CPU technologies that simply can’t keep up with the task of querying and visualizing billions of records. In his session at 20th Cloud Expo, Jared Parker, Director of Financial Services at Kinetica, will discuss how the advent of advanced in-database analytics on the GPU makes it possible to run sophisticated data science workloads on the same database that is housing the rich inf...
Mar. 25, 2017 10:45 PM EDT Reads: 3,596
For organizations that have amassed large sums of software complexity, taking a microservices approach is the first step toward DevOps and continuous improvement / development. Integrating system-level analysis with microservices makes it easier to change and add functionality to applications at any time without the increase of risk. Before you start big transformation projects or a cloud migration, make sure these changes won’t take down your entire organization.
Mar. 25, 2017 09:45 PM EDT Reads: 3,543
Apache Hadoop is emerging as a distributed platform for handling large and fast incoming streams of data. Predictive maintenance, supply chain optimization, and Internet-of-Things analysis are examples where Hadoop provides the scalable storage, processing, and analytics platform to gain meaningful insights from granular data that is typically only valuable from a large-scale, aggregate view. One architecture useful for capturing and analyzing streaming data is the Lambda Architecture, represent...
Mar. 25, 2017 08:45 PM EDT Reads: 5,914
My team embarked on building a data lake for our sales and marketing data to better understand customer journeys. This required building a hybrid data pipeline to connect our cloud CRM with the new Hadoop Data Lake. One challenge is that IT was not in a position to provide support until we proved value and marketing did not have the experience, so we embarked on the journey ourselves within the product marketing team for our line of business within Progress. In his session at @BigDataExpo, Sum...
Mar. 25, 2017 08:45 PM EDT Reads: 2,740
SYS-CON Events announced today that Ocean9will exhibit at SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Ocean9 provides cloud services for Backup, Disaster Recovery (DRaaS) and instant Innovation, and redefines enterprise infrastructure with its cloud native subscription offerings for mission critical SAP workloads.
Mar. 25, 2017 05:15 PM EDT Reads: 1,922
Adding public cloud resources to an existing application can be a daunting process. The tools that you currently use to manage the software and hardware outside the cloud aren’t always the best tools to efficiently grow into the cloud. All of the major configuration management tools have cloud orchestration plugins that can be leveraged, but there are also cloud-native tools that can dramatically improve the efficiency of managing your application lifecycle.
Mar. 25, 2017 04:00 PM EDT Reads: 2,825
DevOps is often described as a combination of technology and culture. Without both, DevOps isn't complete. However, applying the culture to outdated technology is a recipe for disaster; as response times grow and connections between teams are delayed by technology, the culture will die. A Nutanix Enterprise Cloud has many benefits that provide the needed base for a true DevOps paradigm. In his Day 3 Keynote at 20th Cloud Expo, Chris Brown, a Solutions Marketing Manager at Nutanix, will explore t...
Mar. 25, 2017 03:00 PM EDT Reads: 2,715
SYS-CON Events announced today that Technologic Systems Inc., an embedded systems solutions company, will exhibit at SYS-CON's @ThingsExpo, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Technologic Systems is an embedded systems company with headquarters in Fountain Hills, Arizona. They have been in business for 32 years, helping more than 8,000 OEM customers and building over a hundred COTS products that have never been discontinued. Technologic Systems’ pr...
Mar. 25, 2017 01:45 PM EDT Reads: 3,281
SYS-CON Events announced today that CA Technologies has been named “Platinum Sponsor” of SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY, and the 21st International Cloud Expo®, which will take place October 31-November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. CA Technologies helps customers succeed in a future where every business – from apparel to energy – is being rewritten by software. From ...
Mar. 25, 2017 01:30 PM EDT Reads: 1,699
The taxi industry never saw Uber coming. Startups are a threat to incumbents like never before, and a major enabler for startups is that they are instantly “cloud ready.” If innovation moves at the pace of IT, then your company is in trouble. Why? Because your data center will not keep up with frenetic pace AWS, Microsoft and Google are rolling out new capabilities In his session at 20th Cloud Expo, Don Browning, VP of Cloud Architecture at Turner, will posit that disruption is inevitable for c...
Mar. 25, 2017 01:15 PM EDT Reads: 2,037
SYS-CON Events announced today that Cloudistics, an on-premises cloud computing company, has been named “Bronze Sponsor” of SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Cloudistics delivers a complete public cloud experience with composable on-premises infrastructures to medium and large enterprises. Its software-defined technology natively converges network, storage, compute, virtualization, and management into a ...
Mar. 25, 2017 12:45 PM EDT Reads: 1,872
Deep learning has been very successful in social sciences and specially areas where there is a lot of data. Trading is another field that can be viewed as social science with a lot of data. With the advent of Deep Learning and Big Data technologies for efficient computation, we are finally able to use the same methods in investment management as we would in face recognition or in making chat-bots. In his session at 20th Cloud Expo, Gaurav Chakravorty, co-founder and Head of Strategy Development ...
Mar. 25, 2017 12:45 PM EDT Reads: 3,476
What if you could build a web application that could support true web-scale traffic without having to ever provision or manage a single server? Sounds magical, and it is! In his session at 20th Cloud Expo, Chris Munns, Senior Developer Advocate for Serverless Applications at Amazon Web Services, will show how to build a serverless website that scales automatically using services like AWS Lambda, Amazon API Gateway, and Amazon S3. We will review several frameworks that can help you build serverle...
Mar. 25, 2017 12:30 PM EDT Reads: 1,801
The explosion of new web/cloud/IoT-based applications and the data they generate are transforming our world right before our eyes. In this rush to adopt these new technologies, organizations are often ignoring fundamental questions concerning who owns the data and failing to ask for permission to conduct invasive surveillance of their customers. Organizations that are not transparent about how their systems gather data telemetry without offering shared data ownership risk product rejection, regu...
Mar. 25, 2017 12:30 PM EDT Reads: 5,057
SYS-CON Events announced today that Loom Systems will exhibit at SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Founded in 2015, Loom Systems delivers an advanced AI solution to predict and prevent problems in the digital business. Loom stands alone in the industry as an AI analysis platform requiring no prior math knowledge from operators, leveraging the existing staff to succeed in the digital era. With offices in S...
Mar. 25, 2017 12:30 PM EDT Reads: 1,166
SYS-CON Events announced today that Interoute, owner-operator of one of Europe's largest networks and a global cloud services platform, has been named “Bronze Sponsor” of SYS-CON's 20th Cloud Expo, which will take place on June 6-8, 2017 at the Javits Center in New York, New York. Interoute is the owner-operator of one of Europe's largest networks and a global cloud services platform which encompasses 12 data centers, 14 virtual data centers and 31 colocation centers, with connections to 195 add...
Mar. 25, 2017 12:00 PM EDT Reads: 930
Interoute has announced the integration of its Global Cloud Infrastructure platform with Rancher Labs’ container management platform, Rancher. This approach enables enterprises to accelerate their digital transformation and infrastructure investments. Matthew Finnie, Interoute CTO commented “Enterprises developing and building apps in the cloud and those on a path to Digital Transformation need Digital ICT Infrastructure that allows them to build, test and deploy faster than ever before. The int...
Mar. 25, 2017 11:15 AM EDT Reads: 829