|By Jörg-Peter Elbers, Achim Autenrieth||
|February 1, 2013 10:30 AM EST||
Using OpenFlow to extend software-defined networking (SDN) to the optical layer is a compelling prospect for enterprises seeking to achieve joint orchestration of information technology (IT) and network resources for cloud services, to virtualize the network and to more simply manage interconnections of distributed data centers that require synchronization.
Today's fragmented, specialized management and control approaches are fraught with proprietary protocols and management systems, limited scalability and configuration complexities. With an OpenFlow-enabled transport network, an enterprise could instead engage in a kind of "one-stop shopping" for control of cloud computing, storage and networking resources - all via one, unified application programming interface (API). The benefits could include significantly simplified configuration, management and scaling of large-scale enterprise infrastructures through integration and automation.
That's a new role for OpenFlow, demanding strategic tailoring of the protocol for the optical transport domain. Demonstration and development of the capability are closely watched by enterprises that are under incessant pressure to cost-effectively meet ever-increasing demand for bandwidth and services.
Virtualization's New Frontier
Servers and storage have been virtualized in the enterprise; the next great frontier for virtualization is the network.
Because of the substantial cost savings and performance benefits that it can deliver, SDN-based virtualization is of prime interest to enterprises for a wide range of applications. OpenFlow has emerged as one of the most popular SDN protocols. Web 2.0 network operators and national research and education network (NREN) operators, especially, like OpenFlow.
With OpenFlow, an abstraction of the network's packet switches can be generated and flow-forwarding behavior can be specified across an infrastructure via an external controller. Operations can be substantially automated and streamlined by breaking up the monolithically integrated control and forwarding paradigm of today's switches.
Using OpenFlow, could SDN be extended across layers and create a scenario in which - with a single instruction - the controller could jointly create virtual machines and enable enterprise network administrators to reserve computing, networking and storage resources in one stroke?
It is an obviously compelling notion for enterprise network staffs who desperately need to simplify operations. However, the problem is OpenFlow deployment and development has largely been limited to the electrical packet layer, whereas the interconnection beyond the data center is typically comprised of optical transport technology. Furthermore, the optical domain is where things get hazy for many enterprise network administrators. Their comfort zone tends to be packets - not wavelengths and optics.
The result is that cloud computing is currently decoupled from the transport networking control and operation. The network exists as a static, separated entity in today's cloud implementations. There is no interaction between cloud computing processes and the statically configured network. The two are not in any way interoperable; they speak different languages.
Converging cloud computing and networking requires a more dynamic mode of control and operation, but enterprises largely have judged integrating management of the optical network into the data-center environment to be too complex.
To extend OpenFlow from its established role in the electrical packet domain to the optical layer (and, thereby, extend SDN across multiple network layers), a range of optical-specific concerns must be tackled.
Crafting and Experimenting
Within the European Commission's FP7 ICT Work Programme is a collaborative project, "OpenFlow in Europe - Linking Infrastructure and Applications" (OFELIA), that provides researchers with a test bed in which to experiment with SDN applications and virtual multi-layer networks over shared network infrastructure.
Via standardized, secure interfaces through GÉANT, a high-bandwidth interconnection of European R&E networks, researchers develop, run and control experiments using packet switches and application servers at the University of Essex and seven other test-bed facilities throughout Europe.
OFELIA hosts a prototype implementation of dynamic control of wavelength-switched optical networks via OpenFlow. Bandwidth, latency and power consumption can be adjusted to meet the specific requirements of specific applications.
To make it happen, key OpenFlow additions had to be engineered in order for the protocol to effectively control the optical domain. Optical-specific considerations were required to adapt OpenFlow from the packet world. A packet can travel from any ingress to any egress port in an electrical switch or from any time slot in a time-division multiplexing (TDM) device. The optical domain, however, introduces strict switching constraints, with regard to wavelength continuity, optical impairments, optical power leveling on the line side, etc.
Augmenting OpenFlow to address those optical-specific concerns has resulted in an OFELIA prototype that demonstrates a truly transparent, wavelength-switched optical network. The research community is able to experiment with the capability via a flexible, Web-services approach; commercial enterprises, too, are interested in trialing the capability for their specific applications and environments.
OpenFlow is not sufficient in itself to enable the complete transformation that enterprise network administrators envision, to SDN-enable virtualization across all layers of their infrastructures. The additions to OpenFlow that were engineered for the OFELIA test bed provide only the bridge between the optical layer and packet layer and allow integration into a cloud operating system such as OpenStack.
But that is one very important bridge, and the promise for enterprise network administrators is considerable. The OpenFlow innovation could seamlessly integrate the optical transport network under a common management umbrella with an enterprise's routers and switches - all via one familiar interface. Management of the optical domain could become as simple as the management of Ethernet boxes - using an encapsulation of virtual resources that enterprise network administrators could manage via typical and familiar infrastructure. That's a significant breakthrough. With many enterprises already considering usage of an OpenFlow-based control for their packet networks, extending the framework to the wavelength-switched optical layer would be a natural migration.
Virtualization has developed over phases in enterprise networking. First, resource virtualization inside data centers delivered economic savings through enhanced utilization, scalability and redundancy. Data-center virtualization conveyed greater infrastructure flexibility, higher availability and better workload balancing. The next frontier, network virtualization, promises true platform agility and, with it, a host of long-sought-after enterprise capabilities: capacity on-demand, adaptive infrastructure and dynamic service automation, among them. Adapting OpenFlow and extending SDN to the optical transport domain comprise an important step toward that vision.
The Internet of Things (IoT) is growing rapidly by extending current technologies, products and networks. By 2020, Cisco estimates there will be 50 billion connected devices. Gartner has forecast revenues of over $300 billion, just to IoT suppliers. Now is the time to figure out how you’ll make money – not just create innovative products. With hundreds of new products and companies jumping into the IoT fray every month, there’s no shortage of innovation. Despite this, McKinsey/VisionMobile data...
Oct. 8, 2015 07:30 PM EDT Reads: 161
The buzz continues for cloud, data analytics and the Internet of Things (IoT) and their collective impact across all industries. But a new conversation is emerging - how do companies use industry disruption and technology enablers to lead in markets undergoing change, uncertainty and ambiguity? Organizations of all sizes need to evolve and transform, often under massive pressure, as industry lines blur and merge and traditional business models are assaulted and turned upside down. In this new da...
Oct. 8, 2015 07:30 PM EDT Reads: 232
Containers are changing the security landscape for software development and deployment. As with any security solutions, security approaches that work for developers, operations personnel and security professionals is a requirement. In his session at @DevOpsSummit, Kevin Gilpin, CTO and Co-Founder of Conjur, will discuss various security considerations for container-based infrastructure and related DevOps workflows.
Oct. 8, 2015 07:15 PM EDT Reads: 187
As-a-service models offer huge opportunities, but also complicate security. It may seem that the easiest way to migrate to a new architectural model is to let others, experts in their field, do the work. This has given rise to many as-a-service models throughout the industry and across the entire technology stack, from software to infrastructure. While this has unlocked huge opportunities to accelerate the deployment of new capabilities or increase economic efficiencies within an organization, i...
Oct. 8, 2015 07:15 PM EDT Reads: 228
There are so many tools and techniques for data analytics that even for a data scientist the choices, possible systems, and even the types of data can be daunting. In his session at @ThingsExpo, Chris Harrold, Global CTO for Big Data Solutions for EMC Corporation, will show how to perform a simple, but meaningful analysis of social sentiment data using freely available tools that take only minutes to download and install. Participants will get the download information, scripts, and complete en...
Oct. 8, 2015 07:15 PM EDT Reads: 251
Containers are revolutionizing the way we deploy and maintain our infrastructures, but monitoring and troubleshooting in a containerized environment can still be painful and impractical. Understanding even basic resource usage is difficult - let alone tracking network connections or malicious activity. In his session at DevOps Summit, Gianluca Borello, Sr. Software Engineer at Sysdig, will cover the current state of the art for container monitoring and visibility, including pros / cons and li...
Oct. 8, 2015 07:00 PM EDT Reads: 177
IT data is typically silo'd by the various tools in place. Unifying all the log, metric and event data in one analytics platform stops finger pointing and provides the end-to-end correlation. Logs, metrics and custom event data can be joined to tell the holistic story of your software and operations. For example, users can correlate code deploys to system performance to application error codes.
Oct. 8, 2015 06:45 PM EDT Reads: 201
Between the compelling mockups and specs produced by analysts, and resulting applications built by developers, there exists a gulf where projects fail, costs spiral, and applications disappoint. Methodologies like Agile attempt to address this with intensified communication, with partial success but many limitations. In his session at DevOps Summit, Charles Kendrick, CTO and Chief Architect at Isomorphic Software, will present a revolutionary model enabled by new technologies. Learn how busine...
Oct. 8, 2015 06:45 PM EDT Reads: 253
There are many considerations when moving applications from on-premise to cloud. It is critical to understand the benefits and also challenges of this migration. A successful migration will result in lower Total Cost of Ownership, yet offer the same or higher level of robustness. Migration to cloud shifts computing resources from your data center, which can yield significant advantages provided that the cloud vendor an offer enterprise-grade quality for your application.
Oct. 8, 2015 06:00 PM EDT Reads: 225
Internet of Things (IoT) will be a hybrid ecosystem of diverse devices and sensors collaborating with operational and enterprise systems to create the next big application. In their session at @ThingsExpo, Bramh Gupta, founder and CEO of robomq.io, and Fred Yatzeck, principal architect leading product development at robomq.io, discussed how choosing the right middleware and integration strategy from the get-go will enable IoT solution developers to adapt and grow with the industry, while at th...
Oct. 8, 2015 06:00 PM EDT Reads: 2,159
Manufacturing has widely adopted standardized and automated processes to create designs, build them, and maintain them through their life cycle. However, many modern manufacturing systems go beyond mechanized workflows to introduce empowered workers, flexible collaboration, and rapid iteration. Such behaviors also characterize open source software development and are at the heart of DevOps culture, processes, and tooling.
Oct. 8, 2015 06:00 PM EDT Reads: 1,069
The web app is agile. The REST API is agile. The testing and planning are agile. But alas, data infrastructures certainly are not. Once an application matures, changing the shape or indexing scheme of data often forces at best a top down planning exercise and at worst includes schema changes that force downtime. The time has come for a new approach that fundamentally advances the agility of distributed data infrastructures. Come learn about a new solution to the problems faced by software organ...
Oct. 8, 2015 06:00 PM EDT Reads: 781
The last decade was about virtual machines, but the next one is about containers. Containers enable a service to run on any host at any time. Traditional tools are starting to show cracks because they were not designed for this level of application portability. Now is the time to look at new ways to deploy and manage applications at scale. In his session at @DevOpsSummit, Brian “Redbeard” Harrington, a principal architect at CoreOS, will examine how CoreOS helps teams run in production. Attende...
Oct. 8, 2015 05:45 PM EDT Reads: 1,224
“All our customers are looking at the cloud ecosystem as an important part of their overall product strategy. Some see it evolve as a multi-cloud / hybrid cloud strategy, while others are embracing all forms of cloud offerings like PaaS, IaaS and SaaS in their solutions,” noted Suhas Joshi, Vice President – Technology, at Harbinger Group, in this exclusive Q&A with Cloud Expo Conference Chair Roger Strukhoff.
Oct. 8, 2015 05:00 PM EDT Reads: 437
Can call centers hang up the phones for good? Intuitive Solutions did. WebRTC enabled this contact center provider to eliminate antiquated telephony and desktop phone infrastructure with a pure web-based solution, allowing them to expand beyond brick-and-mortar confines to a home-based agent model. It also ensured scalability and better service for customers, including MUY! Companies, one of the country's largest franchise restaurant companies with 232 Pizza Hut locations. This is one example of...
Oct. 8, 2015 04:30 PM EDT Reads: 7,468
Secure Cloud through Automated Compliance | @CloudExpo @CloudRaxak #Cloud #BigData #DevOps #Microservices
Cloud computing delivers on-demand resources that provide businesses with flexibility and cost-savings. The challenge in moving workloads to the cloud has been the cost and complexity of ensuring the initial and ongoing security and regulatory (PCI, HIPAA, FFIEC) compliance across private and public clouds. Manual security compliance is slow, prone to human error, and represents over 50% of the cost of managing cloud applications. Determining how to automate cloud security compliance is critical...
Oct. 8, 2015 04:00 PM EDT Reads: 300
SYS-CON Events announced today that VividCortex, the monitoring solution for the modern data system, will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. The database is the heart of most applications, but it’s also the part that’s hardest to scale, monitor, and optimize even as it’s growing 50% year over year. VividCortex is the first unified suite of database monitoring tools specifically desi...
Oct. 8, 2015 04:00 PM EDT Reads: 461
Saviynt Inc. has announced the availability of the next release of Saviynt for AWS. The comprehensive security and compliance solution provides a Command-and-Control center to gain visibility into risks in AWS, enforce real-time protection of critical workloads as well as data and automate access life-cycle governance. The solution enables AWS customers to meet their compliance mandates such as ITAR, SOX, PCI, etc. by including an extensive risk and controls library to detect known threats and b...
Oct. 8, 2015 03:00 PM EDT Reads: 203
You have your devices and your data, but what about the rest of your Internet of Things story? Two popular classes of technologies that nicely handle the Big Data analytics for Internet of Things are Apache Hadoop and NoSQL. Hadoop is designed for parallelizing analytical work across many servers and is ideal for the massive data volumes you create with IoT devices. NoSQL databases such as Apache HBase are ideal for storing and retrieving IoT data as “time series data.”
Oct. 8, 2015 02:45 PM EDT Reads: 494
Clearly the way forward is to move to cloud be it bare metal, VMs or containers. One aspect of the current public clouds that is slowing this cloud migration is cloud lock-in. Every cloud vendor is trying to make it very difficult to move out once a customer has chosen their cloud. In his session at 17th Cloud Expo, Naveen Nimmu, CEO of Clouber, Inc., will advocate that making the inter-cloud migration as simple as changing airlines would help the entire industry to quickly adopt the cloud wit...
Oct. 8, 2015 02:30 PM EDT Reads: 646