|By Jon Shende||
|August 31, 2010 07:45 AM EDT||
[Adopted from my BLOG December 2009]
Lately in the IT community all the hype is on Cloud Computing. We have small start-ups offering several variations of Cloud services as well as some of the established big players (Google, Amazon, IBM, Novell (aimed at cloud service providers),Sun) stepping up their offerings of cloud services.
But what exactly is Cloud Computing? Is it Virtualization? Is it services that we accessed via a web browser over the years, something totally new, or is it all of these,but just rebranded?
The term Cloud Computing started gaining traction when Google and IBM launched a university initiative to address internet scale computing back in 2007.
These services has been evolving since the 90s and its previous incarnations can be said to be Grid and Utility computing and the Software as a Service offerings we saw around a decade ago.
Cloud Computing Journal - the Web's most widely Web resource on Cloud Computing
In a nutshell we can draw an analogy which can be stated as this: think of a utility service you use, say for example electricity. You get your meter read every few weeks and you receive a bill for energy consumed between readings.
The same underlying premise can be applied to a cloud service, an end user can subscribe for any of the offered cloud services and based on service usage from the provider be billed for consumption of that particular service or series of services for its specified time-frame.
Once can safely state that Cloud Computing as an on-demand, self-service, pay-as you go utility, evolved from a combination of grid computing, virtualization, and automation.
Experts estimate that this industry will grow to a 42 billion dollar industry by 2012, however the implementation and usage of cloud computing models and services is not without issues.
Most business managers will most likely consider the Capex and Opex aspects, especially in this economy. How much money an IT department can save yet still maintain operational efficiencies and security is a primary focus; by implementing one or more cloud computing services, an enterprise can obtain the scale and flexibility it needs and potentially save time as well with the concepts of dynamic provisioning of needed services.
One Cloud Computing claim is to lower costs, increase business agility and help increase the velocity at which applications can be deployed, however a good question to consider is can one expect its implementation to be disruptive and to what length?
In order to engage cloud computing services, business models will have to be adjusted or downright changed, in order to effectively and efficiently managing the utility aspect of computing power used in everyday operations and the manner in which management will be able to utilize resources.
As with any implementation, standards and regulation needs to be formulated and implemented in order to ensure that both vendor and the tenant are in compliance and within governance of an agreed format of policies.
As of now there are no formal standards directed solely toward cloud computing however NIST has proposed a potential framework standard called Standards Acceleration to Jumpstart Adoption of Cloud Computing (SAJAAC). With this, every effort should be made to ensure the confidentiality, availability and integrity of data held within a cloud computing environment going forward.
The National Institute for Science and Technology (NIST) defines cloud computing as "a pay-per-use model for enabling available, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, services) that can be rapidly provisioned and released with minimal management effort or service provider interaction." 
Cloud Computing Models: Cloud models can be one of the following three:
- Infrastructure as a Service (IaaS) e.g. Tier 3, Amazon EC2,while the subscriber does not control the cloud infrastructure they do have control over select portions of network e.g. firewalls, operating system, deployed applications and storage.
- Platform as a Service (PaaS) this goes back to the early 70's when it was referred to as Framework as a Service. What is does is simply to provide different combinations of services to a subscriber supporting an application development life-cycle e.g. Google's App Engine which will let a subscriber run web applications on Google's infrastructure or Azure. In essence the subscriber will use programming(.Net Java python) and tools supplied by the service provider with no underlying responsibility for the cloud deployed network, severs, operating system and storage etc.
- Software as a Service (SaaS) e.g. Facebook, Salesforce.com, applications running on a cloud infrastructure that can be accessed via a web browser interface.
Bear in mind that there can be dependencies and a relationship between the models as Infrastructure as a Service can be stated as the foundation of Cloud Computing services, upon which Platform as a Service and then Software as a Service is built upon.
these services can be implemented by the end user in four different manners :
1) Private Cloud aka a corporate cloud refers to proprietary computing architecture providing hosted services to a limited number of people which resides behind a corporate firewall, in other words a single tenant.
By using private clouds enterprises/tenants will receive the same economies of scale and bi-directional scaling as that of the public cloud user.
However being a single enterprise or division within that enterprise will ensure the additional benefit of more control and security for data held within the private cloud, as on-premise data centers can be converted into private clouds by implementing virtualization technologies from companies such as Microsoft, Citrix, VMware, Novell and Sun.
2) Public Cloud is the cloud that is provided for lease external to an entity's physical location e.g. Amazon's EC2.
This deployment facilitates rapid scaling via virtualization technologies (which enables cloud user resources an ability to rapidly start up and shut down,) and can be utilized by multiple tenants however; within this deployment users have no access to dedicated resources.
This results in users giving up a certain amount of control over the process, which in turn can raise security and compliance issues.
3) Hybrid Cloud is a mixture of the public and private. This can be appealing for a company that chose to store non confidential data externally say using Simple Storage Service (S3) whilst keeping private data in-house.
4) Managed Cloud -In this offering the physical infrastructure in operation is owned by the subscriber and can be housed within the physical premises of the subscriber.
However the service provider will control portions of management and security of the service utility.
Some Deployment Concerns
As with the deployment of any IT system there will be challenges and cause for concern. Certain scenarios will have to be anticipated and use cases as well as processes to mitigate these concerns need to be clarified. Some examples of concerns are as follows:
1) Software licenses: software is typically subscribed as those that are proprietary licensed or those that are free and open source licensed. Software licenses govern usage or redistribution of software which are in most cases copyright protected.
Something to consider is how software that you want to deploy into a cloud is licensed.
Is the software you want to deploy licensed on a per server basis or not and how will easy or difficult will deploying your software of choice into a cloud be?
Will proprietary software solutions need to be confined to dedicated hosting environments? Most likely yes, at least for now or until you can get a vendor who can securely provide the software you need on a pay as you go basis.
Because of how software licensing is structured early cloud users have been found to use more open source software.
2) Single point of failure: a mission critical application is deployed via a single vendor; issues at the vendor's site may severely impact the availability of resources for the tenant.
The vendor may claim to have multiple, remote backup locations completely powered However when it comes to ensuring that, in addition, redundant cloud administration and infrastructure software are in place, the vendor may fall short.
3) Portability: a cause for concern is that each vendor may utilize different applications APIs and formats for data. This in turn may limit application and data portability to other environments, as they are likely to be using proprietary APIs thus causing a "Lock In" situation where it will be easy to sign a contract and use a vendor service however transferring out to another vendor could have major issues.
Of course as the cloud computing environment evolves this may soon be remedied.
4) Security: the elephant in the cloud room. The most common fear with using a cloud deployment is a loss of control and security of data.
Granted this is still a system built on hard and software platforms and as such is still susceptible to the traditional security attacks (DOS, DDOS etc.), conversely a point for consideration should be that any security measure will be more cost effective when implemented on a larger scale.
Any good IT manager has voiced concerns over whether employees/administrators at the cloud provider can be trusted to not look at data or even modify it or, whether other customers sharing the cloud can hack data or access it without leaving an audit trail.
From this a tenant can ask about methods the vendor is employing to protect data such as high physical security as well as what types of monitoring, intrusion detection and firewall equipment are in place at their centers.
Even worse is whether competitors could find out sensitive information such as customer orders, pricing and cost information, and negatively impact business. And of course what about privacy concerns and government regulations?
Other issues of concern can be:
- What levels of protection in place to protect one customer from accessing another customer's data or application within a shared cloud space?
- Who will be liable for security breaches and how will the law regarding this in any one jurisdiction ensure compliance?
- How well will a vendor system integrate with a tenant's security systems?
5) Scalability: Every user/potential user of the cloud constantly hear of the substantial savings they will realize by utilizing cloud-based resources.
In order to take full advantage of the scalability of the cloud there should be a means of ensuring that there is some form of dynamic measurement and resource management for applications held within a cloud.
Scalability within the cloud can be had by composing the service from other scalable services as can be seen with Google App Engine.
6) Auditing: With the cloud one has to consider how compliance with ISO standards, Sarbanes-Oxley, HIPAA ,PCI-DSS etc. will impact certain data from being deployed. This more so when considering the attractiveness of data to unauthorized entities and the methods they could use to gain access to that data.
Any IT manager will also tell you that without proper planning the cost of an audit can be higher than expected.
As of this writing, I am not aware of any formulated standards for auditing within the cloud, however I must state that for a business, auditing within the cloud may be an attractive option as, this can be done live with no down time or interruption to business processes.
7) Compliance: There are no standards in place as of yet, but the National Institute of Standards and Technology (NIST )and others are working toward that end.
8) Other Data Access: what happens to my data if the vendor revokes my access or there is a system malfunction? This is a common question that anyone thinking of using the cloud should ask. Remember the data loss for T-mobile customers using the services of Microsoft subsidiary - Danger? "Microsoft said any data that users had on their devices and is no longer there has almost certainly been permanently lost" Here there was no revocation of access but an alleged system glitch.
A personal example occurred a few days ago when I tried to access a Gmail account I kept just for research and online backup.
The system message intimated that I had violated the "Terms of Agreement". What?! The Gmail account was hardly ever used to send email and the Google docs account was used as a second online backup for some of my documents and files. If this was not a secondary backup or not a backup at all, I would have lost access all my uploaded documents and files, with no recourse for resolution but filling out a form and hoping for contact from the support center.
In order to address and mitigate these issues the tenant should ensure that workarounds and backup plans are worked into their Service Level Agreements (SLAs) with the vendor.
Service Level Agreements (SLAs)
A service level agreement a part of a service contract where the level of service is formally defined. In practice, the term is sometimes used to refer to the contracted delivery time (of the service) or performance.  Whilst there may not be much flexibility with a vendor in defining an SLA, I am confident that the laws of supply and demand will shift this more toward the tenant in the near future.
Cloud computing vendors are getting into this business to affect their bottom line and shareholder value if publicly traded. At the end of the day their focus will be on making a profit on services offered.
In light of this most tenants may feel as though they are getting into an arrangement where it appears as though vendors create the SLAs for their own protection against litigation, with minimal assurances to a tenant.
That being said, this does not mean that an IT manager cannot make the SLA work as a tool to chose an appropriate service provider. An IT manager's main concern will be the security of data and of course, the traditional interpretation of the CIA triad (Confidentiality, Integrity and Availability) may not be applicable within their cloud service.
To start an IT manager can focus on the following when hammering out their SLA with a vendor:
1) Data Protection: where there is a clear definition as to who will have access to the data and the levels of protection in effect for their data.
Some questions that can be asked are:
- How will data be encrypted?
- How will compliance be addressed?
- What are the levels of access control?
- Will there be sub-contractors or third party providers processing the data?
- Where are backups stored?
- How is the data center secured?
- What happens to the data if service providers are switched?
- What processes are in place to mitigate legal inquiries about a customer's data?
- How often are audits done and what types of auditing tools are in place?
- What happens to my data if there is an investigation taking place on another tenant sharing services and how will you ensure my access to my data in the event of equipment seizure by federal entities?
- How is data deletion handled?
2) Continuity: one has to consider what happens in the event of an outage or another related event that causes data to become unavailable.
Some questions to consider here are
- How will the vendor define a services outage?
- Will there be scheduled vendor downtime for maintenance etc.?
- Will there be an alternative vendor hot site or vendor site prepped to take on load of access in the event of a vendor outage?
- Are there tools in motion which will determine the severity of a vendor outage?
- How will the tenant be compensated in the event of a vendor an outage?
- Define levels of redundancy in place to minimize vendor outages?
3) Costs: on cost to consider are:
- How is the vendor's fee structured and is taxes and external fees accounted for in a vendor quote?
- Will there be or are there current licensing fees above and beyond stated vendor service fees?
- Will there be any hidden or add on costs for vendor support?
- How does the vendor structure their charges? Is it based upon usage, traffic or storage limit
- Does the vendor offer price protection?
It is expected that Cloud Computing will the wave of the future in terms of computing, it is only logical that the cloud's economies of scale and flexibility will impact how technology evolves and how users of technologies implement these technologies.
However in terms of security the massive availability of resources and data within a cloud does present a very attractive target for attackers.
That being said, we can assume that cloud-based defences may be more robust, scalable and cost-effective, in an effort to mitigate security concerns regarding multiple tenants, encryption, trust and compliance.
Part of a cloud service is the API. However when it comes to integration between vendors this may pose a problem for tenants, given the fact that cloud APIs are not yet standardized. This means that each vendor has a specific APIs for managing its services that will lock customers to their vendors due to vendor proprietary technology.
The work around here would be to look for vendors that use standard APIs wherever possible. This is a viable option as standard APIs are already implemented for access to storage as well as deploying and scaling applications.
In terms of auditing and forensics, dedicated, pay-per-use forensic images of virtual machines can be obtained by an auditor without having to take infrastructure offline. This of course results in less down-time for auditing as well as it can provide cost-effective storage for logs without deterring system performance.
All of which will increase the return on investment as well as decrease operational costs normally involved with in house systems processing the same data as in the cloud.
Of course Cloud Computing is still in its infancy and whilst some proposals may look good in theory, only time will tell how we proceed and evolve with this system of computing.
 Cloud Connect
 Cloud Security Alliance
Cloud Computing journal
European Network and Information Security Agency.
VictorOps is making on-call suck less with the only collaborative alert management platform on the market. With easy on-call scheduling management, a real-time incident timeline that gives you contextual relevance around your alerts and powerful reporting features that make post-mortems more effective, VictorOps helps your IT/DevOps team solve problems faster.
Mar. 1, 2015 05:00 PM EST Reads: 1,198
Skeuomorphism usually means retaining existing design cues in something new that doesn’t actually need them. However, the concept of skeuomorphism can be thought of as relating more broadly to applying existing patterns to new technologies that, in fact, cry out for new approaches. In his session at DevOps Summit, Gordon Haff, Senior Cloud Strategy Marketing and Evangelism Manager at Red Hat, will discuss why containers should be paired with new architectural practices such as microservices ra...
Mar. 1, 2015 04:00 PM EST Reads: 1,425
Roberto Medrano, Executive Vice President at SOA Software, had reached 30,000 page views on his home page - http://RobertoMedrano.SYS-CON.com/ - on the SYS-CON family of online magazines, which includes Cloud Computing Journal, Internet of Things Journal, Big Data Journal, and SOA World Magazine. He is a recognized executive in the information technology fields of SOA, internet security, governance, and compliance. He has extensive experience with both start-ups and large companies, having been ...
Mar. 1, 2015 04:00 PM EST Reads: 1,202
The industrial software market has treated data with the mentality of “collect everything now, worry about how to use it later.” We now find ourselves buried in data, with the pervasive connectivity of the (Industrial) Internet of Things only piling on more numbers. There’s too much data and not enough information. In his session at @ThingsExpo, Bob Gates, Global Marketing Director, GE’s Intelligent Platforms business, to discuss how realizing the power of IoT, software developers are now focu...
Mar. 1, 2015 03:15 PM EST Reads: 1,322
Operational Hadoop and the Lambda Architecture for Streaming Data Apache Hadoop is emerging as a distributed platform for handling large and fast incoming streams of data. Predictive maintenance, supply chain optimization, and Internet-of-Things analysis are examples where Hadoop provides the scalable storage, processing, and analytics platform to gain meaningful insights from granular data that is typically only valuable from a large-scale, aggregate view. One architecture useful for capturing...
Mar. 1, 2015 02:00 PM EST Reads: 1,304
SYS-CON Events announced today that Vitria Technology, Inc. will exhibit at SYS-CON’s @ThingsExpo, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Vitria will showcase the company’s new IoT Analytics Platform through live demonstrations at booth #330. Vitria’s IoT Analytics Platform, fully integrated and powered by an operational intelligence engine, enables customers to rapidly build and operationalize advanced analytics to deliver timely business outcomes ...
Mar. 1, 2015 01:45 PM EST Reads: 1,188
DevOps is about increasing efficiency, but nothing is more inefficient than building the same application twice. However, this is a routine occurrence with enterprise applications that need both a rich desktop web interface and strong mobile support. With recent technological advances from Isomorphic Software and others, it is now feasible to create a rich desktop and tuned mobile experience with a single codebase, without compromising performance or usability.
Mar. 1, 2015 01:15 PM EST Reads: 1,089
SYS-CON Events announced today Arista Networks will exhibit at SYS-CON's DevOps Summit 2015 New York, which will take place June 9-11, 2015, at the Javits Center in New York City, NY. Arista Networks was founded to deliver software-driven cloud networking solutions for large data center and computing environments. Arista’s award-winning 10/40/100GbE switches redefine scalability, robustness, and price-performance, with over 3,000 customers and more than three million cloud networking ports depl...
Mar. 1, 2015 01:00 PM EST Reads: 1,548
The speed of software changes in growing and large scale rapid-paced DevOps environments presents a challenge for continuous testing. Many organizations struggle to get this right. Practices that work for small scale continuous testing may not be sufficient as the requirements grow. In his session at DevOps Summit, Marc Hornbeek, Sr. Solutions Architect of DevOps continuous test solutions at Spirent Communications, will explain the best practices of continuous testing at high scale, which is r...
Mar. 1, 2015 01:00 PM EST Reads: 1,172
Thanks to Docker, it becomes very easy to leverage containers to build, ship, and run any Linux application on any kind of infrastructure. Docker is particularly helpful for microservice architectures because their successful implementation relies on a fast, efficient deployment mechanism – which is precisely one of the features of Docker. Microservice architectures are therefore becoming more popular, and are increasingly seen as an interesting option even for smaller projects, instead of bein...
Mar. 1, 2015 12:00 PM EST Reads: 2,545
Security can create serious friction for DevOps processes. We've come up with an approach to alleviate the friction and provide security value to DevOps teams. In her session at DevOps Summit, Shannon Lietz, Senior Manager of DevSecOps at Intuit, will discuss how DevSecOps got started and how it has evolved. Shannon Lietz has over two decades of experience pursuing next generation security solutions. She is currently the DevSecOps Leader for Intuit where she is responsible for setting and driv...
Mar. 1, 2015 12:00 PM EST Reads: 2,355
The explosion of connected devices / sensors is creating an ever-expanding set of new and valuable data. In parallel the emerging capability of Big Data technologies to store, access, analyze, and react to this data is producing changes in business models under the umbrella of the Internet of Things (IoT). In particular within the Insurance industry, IoT appears positioned to enable deep changes by altering relationships between insurers, distributors, and the insured. In his session at @Things...
Mar. 1, 2015 12:00 PM EST Reads: 1,241
SYS-CON Events announced today that Open Data Centers (ODC), a carrier-neutral colocation provider, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place June 9-11, 2015, at the Javits Center in New York City, NY. Open Data Centers is a carrier-neutral data center operator in New Jersey and New York City offering alternative connectivity options for carriers, service providers and enterprise customers.
Mar. 1, 2015 12:00 PM EST Reads: 1,894
Even as cloud and managed services grow increasingly central to business strategy and performance, challenges remain. The biggest sticking point for companies seeking to capitalize on the cloud is data security. Keeping data safe is an issue in any computing environment, and it has been a focus since the earliest days of the cloud revolution. Understandably so: a lot can go wrong when you allow valuable information to live outside the firewall. Recent revelations about government snooping, along...
Mar. 1, 2015 11:00 AM EST Reads: 6,946
In his session at DevOps Summit, Tapabrata Pal, Director of Enterprise Architecture at Capital One, will tell a story about how Capital One has embraced Agile and DevOps Security practices across the Enterprise – driven by Enterprise Architecture; bringing in Development, Operations and Information Security organizations together. Capital Ones DevOpsSec practice is based upon three "pillars" – Shift-Left, Automate Everything, Dashboard Everything. Within about three years, from 100% waterfall, C...
Mar. 1, 2015 11:00 AM EST Reads: 2,713
PubNub on Monday has announced that it is partnering with IBM to bring its sophisticated real-time data streaming and messaging capabilities to Bluemix, IBM’s cloud development platform. “Today’s app and connected devices require an always-on connection, but building a secure, scalable solution from the ground up is time consuming, resource intensive, and error-prone,” said Todd Greene, CEO of PubNub. “PubNub enables web, mobile and IoT developers building apps on IBM Bluemix to quickly add sc...
Mar. 1, 2015 10:00 AM EST Reads: 4,730
Data-intensive companies that strive to gain insights from data using Big Data analytics tools can gain tremendous competitive advantage by deploying data-centric storage. Organizations generate large volumes of data, the vast majority of which is unstructured. As the volume and velocity of this unstructured data increases, the costs, risks and usability challenges associated with managing the unstructured data (regardless of file type, size or device) increases simultaneously, including end-to-...
Mar. 1, 2015 09:45 AM EST Reads: 2,192
The excitement around the possibilities enabled by Big Data is being tempered by the daunting task of feeding the analytics engines with high quality data on a continuous basis. As the once distinct fields of data integration and data management increasingly converge, cloud-based data solutions providers have emerged that can buffer your organization from the complexities of this continuous data cleansing and management so that you’re free to focus on the end goal: actionable insight.
Mar. 1, 2015 09:30 AM EST Reads: 1,657
With several hundred implementations of IoT-enabled solutions in the past 12 months alone, this session will focus on experience over the art of the possible. Many can only imagine the most advanced telematics platform ever deployed, supporting millions of customers, producing tens of thousands events or GBs per trip, and hundreds of TBs per month. With the ability to support a billion sensor events per second, over 30PB of warm data for analytics, and hundreds of PBs for an data analytics arc...
Mar. 1, 2015 09:00 AM EST Reads: 1,264
Between the compelling mockups and specs produced by your analysts and designers, and the resulting application built by your developers, there is a gulf where projects fail, costs spiral out of control, and applications fall short of requirements. In his session at DevOps Summit, Charles Kendrick, CTO and Chief Architect at Isomorphic Software, will present a new approach where business and development users collaborate – each using tools appropriate to their goals and expertise – to build mo...
Mar. 1, 2015 09:00 AM EST Reads: 2,908