|By Marten Terpstra||
|April 10, 2014 01:00 PM EDT||
We love analogies. No matter what the topic, analogies are a great way to explain something in a different context to make a specific point with a frame of reference that may be more familiar to those we are making a point to. There is one that seems to come back over and over again in our industry, the one that compares the network to the power grid, network connections to power plugs. I had not heard it for a while but at Interop last week, I heard it used twice in booth demonstrations as part of plug and play pitches. And I really do not like that analogy.
The comparison to power comes from the angle of it just being there and available. Plug something in and it turns on. Its universal and it’s simple, anyone can use it. It’s highly standardized, you can buy bits and pieces that work together in any hardware store. It’s reliable. You just get a bill each month, pay it, and in many places you can even pick your provider different from the owner of the cable that actually plugs into your house. It’s all good, it’s mostly invisible.
Except that all of these come with their own challenges. The overall power grid reliability is well below 99.99%. Across the US, the average outage minutes per year is hovering in the 2 hour range, in the Northeast almost double that. There are only very people I know that do not have a backup generator, just in case, and get to use it at least once a year, ignoring the smaller outages where we just live in the dark for a few minutes. It is very standardized, with plenty of standard variations from country to country, thank goodness for $30 converters. And the cost, well, it’s rising faster than inflation (almost twice as fast in fact).
Now, this article is not meant to be a complaining session about power. When power works and you forget about the bill, it is very much invisible, which is the only piece of the analogy I do like, maybe.
Networks should be invisible, but not in the same sense power should be invisible. When comparing networks to power it always comes down to “it’s just there, plentiful, accessible, and just works”. And (not surprising), I think the network can do much better than “just be there”. Networks need to migrate away from an entity that is architected, designed, implemented, managed and debugged separately from the applications it serves. Networks have to become an integral part of the application infrastructure.
Now, any two (or more) distinctly different entities need to be glued together. No different for applications, servers and networks. Applications are glued to servers by means of the OS. As a consumer of an application, or even an administrator, the amount of work you have to do to put these two together is (usually) fairy minimal. The creator of the OS has provided tons of hooks and tools to give the application what it needs, the application provides to and asks what it wants from the OS.
The glue between the server and the network is much harsher, much less refined. There is an expectation of matching configuration, very little is actually transacted between the two to get them to work together nicely. Yes there is DHCP and perhaps you can even count LLDP in that, but the amount of alignment of information between the two so they can work together seamlessly without human intervention is minimal. And if something is not quite working the way it should, you need heavily certified network engineers and tons of homegrown tools and scripts to figure out why. That is by no means a knock on network certifications (that is for @cloudtoad to comment on), but it should not be that hard. The exchange is extremely focused on data plane traffic for transportation, very little provisioning, debug and other feedback is exchanged and reacted to.
We need to get to a network that is invisible enough to allow you to create coarse provisioning policies created from application templates. There should be no need for me to configure individual ports and protocols on ports. There should be no need for me to configure how network components are tied together and provide optimized connectivity based on the needs of its customers. The network needs to provide debugging tools that interact with applications in feedback loops. It needs to provide enough details for orchestration systems to adjust their placement of apps. Or in Plexxi’s case, take in information from orchestration systems to change the L1, L2 and L3 behavior of the network to the requested demand. And that orchestration system could be a specialized Hadoop controller or a far more generic cloud orchestrator.
The network needs to cease to exist as a separate entity, it needs to become an integrated part of the application infrastructure that uses its services. That’s the kind of invisible we need to achieve.
[Today's Fun Fact: An Iranian man paid $500 to a wizard to become invisible. The man then walked into a bank and grabbed money from people's hands believing he was invisible. He was promptly arrested and convicted. Conclusion: pay more or pick a better wizard next time.]
Interested in leveraging automation technologies and a cloud architecture to make developers more productive? Learn how PaaS can benefit your organization to help you streamline your application development, allow you to use existing infrastructure and improve operational efficiencies. Begin charting your path to PaaS with OpenShift Enterprise.
Oct. 9, 2015 02:00 AM EDT Reads: 555
The buzz continues for cloud, data analytics and the Internet of Things (IoT) and their collective impact across all industries. But a new conversation is emerging - how do companies use industry disruption and technology enablers to lead in markets undergoing change, uncertainty and ambiguity? Organizations of all sizes need to evolve and transform, often under massive pressure, as industry lines blur and merge and traditional business models are assaulted and turned upside down. In this new da...
Oct. 9, 2015 02:00 AM EDT Reads: 281
Internet of Things (IoT) will be a hybrid ecosystem of diverse devices and sensors collaborating with operational and enterprise systems to create the next big application. In their session at @ThingsExpo, Bramh Gupta, founder and CEO of robomq.io, and Fred Yatzeck, principal architect leading product development at robomq.io, discussed how choosing the right middleware and integration strategy from the get-go will enable IoT solution developers to adapt and grow with the industry, while at th...
Oct. 9, 2015 02:00 AM EDT Reads: 2,192
Overgrown applications have given way to modular applications, driven by the need to break larger problems into smaller problems. Similarly large monolithic development processes have been forced to be broken into smaller agile development cycles. Looking at trends in software development, microservices architectures meet the same demands. Additional benefits of microservices architectures are compartmentalization and a limited impact of service failure versus a complete software malfunction....
Oct. 9, 2015 02:00 AM EDT Reads: 214
Data loss happens, even in the cloud. In fact, if your company has adopted a cloud application in the past three years, data loss has probably happened, whether you know it or not. In his session at 17th Cloud Expo, Bryan Forrester, Senior Vice President of Sales at eFolder, will present how common and costly cloud application data loss is and what measures you can take to protect your organization from data loss.
Oct. 9, 2015 02:00 AM EDT Reads: 573
The Internet of Things (IoT) is growing rapidly by extending current technologies, products and networks. By 2020, Cisco estimates there will be 50 billion connected devices. Gartner has forecast revenues of over $300 billion, just to IoT suppliers. Now is the time to figure out how you’ll make money – not just create innovative products. With hundreds of new products and companies jumping into the IoT fray every month, there’s no shortage of innovation. Despite this, McKinsey/VisionMobile data...
Oct. 9, 2015 02:00 AM EDT Reads: 205
The web app is agile. The REST API is agile. The testing and planning are agile. But alas, data infrastructures certainly are not. Once an application matures, changing the shape or indexing scheme of data often forces at best a top down planning exercise and at worst includes schema changes that force downtime. The time has come for a new approach that fundamentally advances the agility of distributed data infrastructures. Come learn about a new solution to the problems faced by software organ...
Oct. 9, 2015 02:00 AM EDT Reads: 878
Between the compelling mockups and specs produced by analysts, and resulting applications built by developers, there exists a gulf where projects fail, costs spiral, and applications disappoint. Methodologies like Agile attempt to address this with intensified communication, with partial success but many limitations. In his session at DevOps Summit, Charles Kendrick, CTO and Chief Architect at Isomorphic Software, will present a revolutionary model enabled by new technologies. Learn how busine...
Oct. 9, 2015 02:00 AM EDT Reads: 282
“In the past year we've seen a lot of stabilization of WebRTC. You can now use it in production with a far greater degree of certainty. A lot of the real developments in the past year have been in things like the data channel, which will enable a whole new type of application," explained Peter Dunkley, Technical Director at Acision, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Oct. 9, 2015 01:45 AM EDT Reads: 7,010
There are many considerations when moving applications from on-premise to cloud. It is critical to understand the benefits and also challenges of this migration. A successful migration will result in lower Total Cost of Ownership, yet offer the same or higher level of robustness. Migration to cloud shifts computing resources from your data center, which can yield significant advantages provided that the cloud vendor an offer enterprise-grade quality for your application.
Oct. 9, 2015 12:30 AM EDT Reads: 247
JFrog has announced a powerful technology for managing software packages from development into production. JFrog Artifactory 4 represents disruptive innovation in its groundbreaking ability to help development and DevOps teams deliver increasingly complex solutions on ever-shorter deadlines across multiple platforms JFrog Artifactory 4 establishes a new category – the Universal Artifact Repository – that reflects JFrog's unique commitment to enable faster software releases through the first pla...
Oct. 9, 2015 12:30 AM EDT Reads: 634
IT data is typically silo'd by the various tools in place. Unifying all the log, metric and event data in one analytics platform stops finger pointing and provides the end-to-end correlation. Logs, metrics and custom event data can be joined to tell the holistic story of your software and operations. For example, users can correlate code deploys to system performance to application error codes.
Oct. 9, 2015 12:00 AM EDT Reads: 222
Through WebRTC, audio and video communications are being embedded more easily than ever into applications, helping carriers, enterprises and independent software vendors deliver greater functionality to their end users. With today’s business world increasingly focused on outcomes, users’ growing calls for ease of use, and businesses craving smarter, tighter integration, what’s the next step in delivering a richer, more immersive experience? That richer, more fully integrated experience comes ab...
Oct. 9, 2015 12:00 AM EDT Reads: 1,133
As-a-service models offer huge opportunities, but also complicate security. It may seem that the easiest way to migrate to a new architectural model is to let others, experts in their field, do the work. This has given rise to many as-a-service models throughout the industry and across the entire technology stack, from software to infrastructure. While this has unlocked huge opportunities to accelerate the deployment of new capabilities or increase economic efficiencies within an organization, i...
Oct. 8, 2015 11:00 PM EDT Reads: 267
SYS-CON Events announced today that Dyn, the worldwide leader in Internet Performance, will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Dyn is a cloud-based Internet Performance company. Dyn helps companies monitor, control, and optimize online infrastructure for an exceptional end-user experience. Through a world-class network and unrivaled, objective intelligence into Internet condit...
Oct. 8, 2015 10:00 PM EDT Reads: 595
The last decade was about virtual machines, but the next one is about containers. Containers enable a service to run on any host at any time. Traditional tools are starting to show cracks because they were not designed for this level of application portability. Now is the time to look at new ways to deploy and manage applications at scale. In his session at @DevOpsSummit, Brian “Redbeard” Harrington, a principal architect at CoreOS, will examine how CoreOS helps teams run in production. Attende...
Oct. 8, 2015 09:45 PM EDT Reads: 1,253
Containers are revolutionizing the way we deploy and maintain our infrastructures, but monitoring and troubleshooting in a containerized environment can still be painful and impractical. Understanding even basic resource usage is difficult - let alone tracking network connections or malicious activity. In his session at DevOps Summit, Gianluca Borello, Sr. Software Engineer at Sysdig, will cover the current state of the art for container monitoring and visibility, including pros / cons and li...
Oct. 8, 2015 09:30 PM EDT Reads: 216
Containers are changing the security landscape for software development and deployment. As with any security solutions, security approaches that work for developers, operations personnel and security professionals is a requirement. In his session at @DevOpsSummit, Kevin Gilpin, CTO and Co-Founder of Conjur, will discuss various security considerations for container-based infrastructure and related DevOps workflows.
Oct. 8, 2015 09:15 PM EDT Reads: 231
There are so many tools and techniques for data analytics that even for a data scientist the choices, possible systems, and even the types of data can be daunting. In his session at @ThingsExpo, Chris Harrold, Global CTO for Big Data Solutions for EMC Corporation, will show how to perform a simple, but meaningful analysis of social sentiment data using freely available tools that take only minutes to download and install. Participants will get the download information, scripts, and complete en...
Oct. 8, 2015 09:15 PM EDT Reads: 288
Manufacturing has widely adopted standardized and automated processes to create designs, build them, and maintain them through their life cycle. However, many modern manufacturing systems go beyond mechanized workflows to introduce empowered workers, flexible collaboration, and rapid iteration. Such behaviors also characterize open source software development and are at the heart of DevOps culture, processes, and tooling.
Oct. 8, 2015 09:00 PM EDT Reads: 1,093