Welcome!

@CloudExpo Authors: Yeshim Deniz, Pat Romanski, Elizabeth White, Liz McMillan, Zakia Bouachraoui

Related Topics: Containers Expo Blog, Microservices Expo, @CloudExpo

Containers Expo Blog: Article

Infrastructure 2.0 in the News: The Beat Goes On

Virtualization by itself isn’t enough; it needs an infrastructure 2.0 network

Infrastructure 2.0 Session at Virtualization Expo

This week InformationWeek featured my recent infrastructure 2.0 interview with Alexander Wolfe. As the infrastructure 2.0 Working Group is in stealth mode, we focused on the high level tech issues resulting from the collision between virtual and static (physical) infrastructure:

InformationWeek: Let's dive right in. Define the issue, as you see it.

Greg Ness: If you look at networks today, the way they're operated, run, and configured hasn't really changed for the last several decades. There's a contrast between today's networks and today's systems. The systems are getting increasingly automated, while the networks are still very much in silos. If I were to articulate this in one sentence, I'd say that today's networks are run like yesterday's businesses.

You can read the entire interview here.

CNET blogger James Urquhart (Cisco’s Cloud Computing Market Strategist) referred to the interview in his blog entitled Understanding Infrastructure 2.0.

James did a great job highlighting the problems around manual address management at the core of the network:

"Take a look at the core TCP/IP and Ethernet networks that we all use today, and how enterprise IT manages those services. Not long ago, I wrote an article that described how most corporations relied heavily on manual labor to manage everything from IP addresses and domain names to routing and switching configuration. At the time, I cited a survey that indicated that a full 63 percent of enterprises were still using spreadsheets to manage IP addresses."

Unfortunately as we move into an era of data center virtualization and cloud computing, spreadsheets don’t cut it anymore.

You can read the rest of James’ excellent blog here, where he also points readers to several of his favorite infrastructure 2.0 blogs.

The common theme from both of these news sites is disconnect between virtualized, dynamic systems and static, manually managed networks. IT automation will require network automation and network automation will require DNS, DHCP and IPAM automation. Virtualization by itself isn’t enough; it needs an infrastructure 2.0 network.

You can follow my rants in real-time via Twitter at Archimedius.

More Stories By Greg Ness

Gregory Ness is the VP of Marketing of Vidder and has over 30 years of experience in marketing technology, B2B and consumer products and services. Prior to Vidder, he was VP of Marketing at cloud migration pioneer CloudVelox. Before CloudVelox he held marketing leadership positions at Vantage Data Centers, Infoblox (BLOX), BlueLane Technologies (VMW), Redline Networks (JNPR), IntruVert (INTC) and ShoreTel (SHOR). He has a BA from Reed College and an MA from The University of Texas at Austin. He has spoken on virtualization, networking, security and cloud computing topics at numerous conferences including CiscoLive, Interop and Future in Review.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


CloudEXPO Stories
Most modern computer languages embed a lot of metadata in their application. We show how this goldmine of data from a runtime environment like production or staging can be used to increase profits. Adi conceptualized the Crosscode platform after spending over 25 years working for large enterprise companies like HP, Cisco, IBM, UHG and personally experiencing the challenges that prevent companies from quickly making changes to their technology, due to the complexity of their enterprise. An accomplished expert in Enterprise Architecture, Adi has also served as CxO advisor to numerous Fortune executives.
Cloud computing is a goal aspired to by all organizations, yet those in regulated industries and many public sector organizations are challenged in adopting cloud technologies. The ability to use modern application development capabilities such as containers, serverless computing, platform-based services, IoT and others are potentially of great benefit for these organizations but doing so in a public cloud-consistent way is the challenge.
"Calligo is a cloud service provider with data privacy at the heart of what we do. We are a typical Infrastructure as a Service cloud provider but it's been designed around data privacy," explained Julian Box, CEO and co-founder of Calligo, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
In his session at 21st Cloud Expo, Michael Burley, a Senior Business Development Executive in IT Services at NetApp, described how NetApp designed a three-year program of work to migrate 25PB of a major telco's enterprise data to a new STaaS platform, and then secured a long-term contract to manage and operate the platform. This significant program blended the best of NetApp’s solutions and services capabilities to enable this telco’s successful adoption of private cloud storage and launching of virtual storage services to its enterprise market.
When building large, cloud-based applications that operate at a high scale, it’s important to maintain a high availability and resilience to failures. In order to do that, you must be tolerant of failures, even in light of failures in other areas of your application. “Fly two mistakes high” is an old adage in the radio control airplane hobby. It means, fly high enough so that if you make a mistake, you can continue flying with room to still make mistakes. In his session at 18th Cloud Expo, Lee Atchison, Principal Cloud Architect and Advocate at New Relic, will discuss how this same philosophy can be applied to highly scaled applications, and can dramatically increase your resilience to failure.