|By Andreas Grabner||
|November 4, 2016 06:15 PM EDT||
DevOps is speeding towards the IT world like a freight train and the hype around it is deafening. There is no reason to be afraid of this change as it is the natural reaction to the agile movement that revolutionized development just a few years ago. By definition, DevOps is the natural alignment of IT performance to business profitability. The relevance of this has yet to be quantified but it has been suggested that the route to the CEO's chair will come from the IT leaders that successfully make the transition to a DevOps model. If this still seems foreign to you, I recommend reading up on DevOps Blog from IT Revolution, the OpsCode Blog, and check out The Phoenix Project.
Despite all the talk around simple monitoring tools, breaking through the walls between Dev and Ops still poses a real challenge. This is because of a misunderstanding around Operations real purpose - extracting real value from its resources. According to Kevin Behr the definition of Operations is the act of harvesting value from IT resources. Anything that prevents this from happening is a detriment to the business. This means that firefighting and war room sessions are a hindrance to the DevOps model. The following screenshots are good examples of a war room scenario.
Unexpected crashes of websites upon new rollouts still lead to "War Room" situations - despite all the good efforts of DevOps and Agile Delivery/Deployment
Successfully riding the DevOps train: Many of our production customers that made it through the firefighting mode applied the principals of DevOps with a special focus on Application Performance. In this article we describe the steps and milestones companies need to go through in order to level-up their Operations and Engineering Teams to provide more value out of the existing resources.
The Foundation of DevOps
CAMS (Culture, Automation, Measurement, Sharing) are four key areas that are core to the DevOps movement. Culture is the hardest to change but is also the most important because it means a change in the way in which the different teams work together and share the responsibility for the end users of their application. It promotes the usage of development practices in operations to automate deployment. It also allows developers to learn from "the real world" Ops experience and with that mutual exchange it breaks down the walls.
The Lack of Performance Focus
An interesting fact based on the feedback we get from operation teams worldwide: The root cause for about 80% of site crashes or performance problems is related to only about 20% of problem patterns. Want to learn more? Check out blogs such as Top Performance Landmines in Production and I am sure you'll find some issues you already ran into yourself.
Looking at these common problem patterns it is clear that despite all the DevOps efforts lots of performance and scalability-related problems still make it into a release deployment. Why is that? Because our organizations are still very much driven by business requirements that need numerous new features being pushed in ever shorter release cycles. Teams keep growing and are being spread around the world. In order to keep up with the pace, third-party components are included in the code in place of in-house innovation. This "natural" evolution however is also the root cause for firefights and limiting the benefits of DevOps because there is too much focus on pushing functionality through the Deployment Pipeline but not enough focus on Performance.
More developers across more locations including more untested 3rd party code with less time to focus on performance
Plugging Performance into DevOps
In order to focus developers on performance to avoid War Room scenarios you must plug performance into the four pillars of CAMS:
- Culture: Performance as Key Requirement in Dev, Test and Ops
- Automation: Automated Performance Tests already in Continuous Integration
- Measurement: Measure Key Performance Metrics in CI, Test and Ops
- Sharing: Share the same tools and same performance data across Dev, Test and Ops
There are several key milestones to consider:
Milestone 1: Level-Up Performance to Increase Feedback Between Ops and Development
The first step in any DevOps initiative is to get the Ops teams and the Dev teams talking in order to relieve constraints on the business. This might be easy for small teams to accomplish but the larger the organization, the more difficult it becomes as constraints are greater in larger organizations. Operations has to diminish these constraints on the business. This is where APM solutions can help. Beware as not all solutions are created equal. As mentioned in the previous DevOps blog the drive to diminish constraint needs to be applied across the delivery chain. Monitoring just does not cut it here. There needs to be something that not only starts the process but allows the teams to continue to mature and grow. Simple monitoring tools fall short because they only help extinguish fires in operations. How does this continue to drive down constraints? Remember, firefighting is not a part of operations, which means operations should not be looking at fire extinguishers for their DevOps strategies.
Milestone 2: Level-Up Performance Thinking of Engineering
Both Operations and Test Teams have a good understanding of Performance as they deal with it every day. These teams need to educate engineering on the importance of performance as it is a key requirement to software engineering and how it plays a role in large-scale environments under heavy load.
The Ops team shares data with engineering to highlight the performance behavior of their applications under real production load. This helps engineers to prevent these top performance problems from entering production and with that eliminating the need for firefights.
The test teams do their share by providing automated performance test frameworks and educating engineering on how to automate testing for these performance problem patterns.
Milestone 3: Level-Up Load and Capacity Testing
With development executing its own performance tests it's time to level up the test team as well. On one side there is more time to focus on large-scale load tests that need to be executed in a production- like environment. This helps to find any "data-driven", scalability, and "third-party impacted" performance problems. Close collaboration with Ops ensures that tests can be executed either in the prod environment or in a staged environment that mirrors production. Executing these tests in collaboration with Ops allows the teams to become more confident when releasing a new version and also helps with proper capacity planning steps.
Running tests against the production system gives better input for capacity planning and uncovers heavy load application issues
Milestone 4: Level-Up Performance Test Automation
The "traditional" testing teams are used to execute performance and scalability tests in their own environments at the end of a milestone. The goal is to provide these test frameworks and environments to engineering so that these basic performance tests can be executed automatically in the CI environment. In order for this to work you need to make sure that:
- These test frameworks are easy to use and accepted by developers
- Deliver performance metrics to detect the common problem patterns
- These are fully integrated into continuous integration
Automatic Integration Tests run in C/I to detect performance regressions on metrics such as # of SQL Calls, Page Load Time, # of JS files or Images ...
What's Next? Build a Performance Center of Excellence
Many of our customers who jumped on the DevOps train a while back are now promoting a performance culture in their organizations. In the next few blogs we will cover their best practices and tips on either building a separate "Performance Center of Excellence" Team or up-leveling the existing DevOps teams to deliver software with high confidence and fewer War Room weekends.
Without a clear strategy for cost control and an architecture designed with cloud services in mind, costs and operational performance can quickly get out of control. To avoid multiple architectural redesigns requires extensive thought and planning. Boundary (now part of BMC) launched a new public-facing multi-tenant high resolution monitoring service on Amazon AWS two years ago, facing challenges and learning best practices in the early days of the new service. In his session at 19th Cloud Exp...
Dec. 4, 2016 09:00 AM EST Reads: 543
"Venafi has a platform that allows you to manage, centralize and automate the complete life cycle of keys and certificates within the organization," explained Gina Osmond, Sr. Field Marketing Manager at Venafi, in this SYS-CON.tv interview at DevOps at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 4, 2016 08:45 AM EST Reads: 785
Effectively SMBs and government programs must address compounded regulatory compliance requirements. The most recent are Controlled Unclassified Information and the EU's GDPR have Board Level implications. Managing sensitive data protection will likely result in acquisition criteria, demonstration requests and new requirements. Developers, as part of the pre-planning process and the associated supply chain, could benefit from updating their code libraries and design by incorporating changes. In...
Dec. 4, 2016 08:30 AM EST Reads: 969
"Coalfire is a cyber-risk, security and compliance assessment and advisory services firm. We do a lot of work with the cloud service provider community," explained Ryan McGowan, Vice President, Sales (West) at Coalfire Systems, Inc., in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 4, 2016 08:30 AM EST Reads: 760
Regulatory requirements exist to promote the controlled sharing of information, while protecting the privacy and/or security of the information. Regulations for each type of information have their own set of rules, policies, and guidelines. Cloud Service Providers (CSP) are faced with increasing demand for services at decreasing prices. Demonstrating and maintaining compliance with regulations is a nontrivial task and doing so against numerous sets of regulatory requirements can be daunting task...
Dec. 4, 2016 08:15 AM EST Reads: 750
CloudJumper, a Workspace as a Service (WaaS) platform innovator for agile business IT, has been recognized with the Customer Value Leadership Award for its nWorkSpace platform by Frost & Sullivan. The company was also featured in a new report(1) by the industry research firm titled, “Desktop-as-a-Service Buyer’s Guide, 2016,” which provides a comprehensive comparison of DaaS providers, including CloudJumper, Amazon, VMware, and Microsoft.
Dec. 4, 2016 08:15 AM EST Reads: 705
Businesses and business units of all sizes can benefit from cloud computing, but many don't want the cost, performance and security concerns of public cloud nor the complexity of building their own private clouds. Today, some cloud vendors are using artificial intelligence (AI) to simplify cloud deployment and management. In his session at 20th Cloud Expo, Ajay Gulati, Co-founder and CEO of ZeroStack, will discuss how AI can simplify cloud operations. He will cover the following topics: why clou...
Dec. 4, 2016 08:00 AM EST Reads: 689
Fact: storage performance problems have only gotten more complicated, as applications not only have become largely virtualized, but also have moved to cloud-based infrastructures. Storage performance in virtualized environments isn’t just about IOPS anymore. Instead, you need to guarantee performance for individual VMs, helping applications maintain performance as the number of VMs continues to go up in real time. In his session at Cloud Expo, Dhiraj Sehgal, Product and Marketing at Tintri, sha...
Dec. 4, 2016 08:00 AM EST Reads: 804
In his keynote at 18th Cloud Expo, Andrew Keys, Co-Founder of ConsenSys Enterprise, provided an overview of the evolution of the Internet and the Database and the future of their combination – the Blockchain. Andrew Keys is Co-Founder of ConsenSys Enterprise. He comes to ConsenSys Enterprise with capital markets, technology and entrepreneurial experience. Previously, he worked for UBS investment bank in equities analysis. Later, he was responsible for the creation and distribution of life sett...
Dec. 4, 2016 06:15 AM EST Reads: 6,971
20th Cloud Expo, taking place June 6-8, 2017, at the Javits Center in New York City, NY, will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy.
Dec. 4, 2016 05:30 AM EST Reads: 1,753
More and more companies are looking to microservices as an architectural pattern for breaking apart applications into more manageable pieces so that agile teams can deliver new features quicker and more effectively. What this pattern has done more than anything to date is spark organizational transformations, setting the foundation for future application development. In practice, however, there are a number of considerations to make that go beyond simply “build, ship, and run,” which changes how...
Dec. 4, 2016 04:45 AM EST Reads: 4,968
WebRTC is the future of browser-to-browser communications, and continues to make inroads into the traditional, difficult, plug-in web communications world. The 6th WebRTC Summit continues our tradition of delivering the latest and greatest presentations within the world of WebRTC. Topics include voice calling, video chat, P2P file sharing, and use cases that have already leveraged the power and convenience of WebRTC.
Dec. 4, 2016 04:30 AM EST Reads: 1,554
Amazon has gradually rolled out parts of its IoT offerings, but these are just the tip of the iceberg. In addition to optimizing their backend AWS offerings, Amazon is laying the ground work to be a major force in IoT - especially in the connected home and office. In his session at @ThingsExpo, Chris Kocher, founder and managing director of Grey Heron, explained how Amazon is extending its reach to become a major force in IoT by building on its dominant cloud IoT platform, its Dash Button strat...
Dec. 4, 2016 04:00 AM EST Reads: 6,226
Let’s face it, embracing new storage technologies, capabilities and upgrading to new hardware often adds complexity and increases costs. In his session at 18th Cloud Expo, Seth Oxenhorn, Vice President of Business Development & Alliances at FalconStor, discussed how a truly heterogeneous software-defined storage approach can add value to legacy platforms and heterogeneous environments. The result reduces complexity, significantly lowers cost, and provides IT organizations with improved efficienc...
Dec. 4, 2016 04:00 AM EST Reads: 4,959
Internet-of-Things discussions can end up either going down the consumer gadget rabbit hole or focused on the sort of data logging that industrial manufacturers have been doing forever. However, in fact, companies today are already using IoT data both to optimize their operational technology and to improve the experience of customer interactions in novel ways. In his session at @ThingsExpo, Gordon Haff, Red Hat Technology Evangelist, will share examples from a wide range of industries – includin...
Dec. 4, 2016 03:45 AM EST Reads: 1,570
"We build IoT infrastructure products - when you have to integrate different devices, different systems and cloud you have to build an application to do that but we eliminate the need to build an application. Our products can integrate any device, any system, any cloud regardless of protocol," explained Peter Jung, Chief Product Officer at Pulzze Systems, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 4, 2016 02:15 AM EST Reads: 886
When it comes to cloud computing, the ability to turn massive amounts of compute cores on and off on demand sounds attractive to IT staff, who need to manage peaks and valleys in user activity. With cloud bursting, the majority of the data can stay on premises while tapping into compute from public cloud providers, reducing risk and minimizing need to move large files. In his session at 18th Cloud Expo, Scott Jeschonek, Director of Product Management at Avere Systems, discussed the IT and busin...
Dec. 4, 2016 02:00 AM EST Reads: 3,786
Between 2005 and 2020, data volumes will grow by a factor of 300 – enough data to stack CDs from the earth to the moon 162 times. This has come to be known as the ‘big data’ phenomenon. Unfortunately, traditional approaches to handling, storing and analyzing data aren’t adequate at this scale: they’re too costly, slow and physically cumbersome to keep up. Fortunately, in response a new breed of technology has emerged that is cheaper, faster and more scalable. Yet, in meeting these new needs they...
Dec. 4, 2016 12:30 AM EST Reads: 1,788
The cloud promises new levels of agility and cost-savings for Big Data, data warehousing and analytics. But it’s challenging to understand all the options – from IaaS and PaaS to newer services like HaaS (Hadoop as a Service) and BDaaS (Big Data as a Service). In her session at @BigDataExpo at @ThingsExpo, Hannah Smalltree, a director at Cazena, provided an educational overview of emerging “as-a-service” options for Big Data in the cloud. This is critical background for IT and data professionals...
Dec. 3, 2016 11:00 PM EST Reads: 4,167
"Once customers get a year into their IoT deployments, they start to realize that they may have been shortsighted in the ways they built out their deployment and the key thing I see a lot of people looking at is - how can I take equipment data, pull it back in an IoT solution and show it in a dashboard," stated Dave McCarthy, Director of Products at Bsquare Corporation, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 3, 2016 11:00 PM EST Reads: 972