|By Marten Terpstra||
|May 16, 2014 12:45 PM EDT||
IP Multicast is one of those technologies that most everyone loves to hate. It’s almost the perfect example of how complicated we have made networking. Getting IP Multicast to run depends on several protocols that are all somewhat intertwined or dependent on each, their relationship sometimes explicit, sometimes implicit.
Even trying to describe the basic operation is complicated.
When an application or service provides information using IP multicast, it simply starts sending it onto a specific multicast group. The multicast router for the subnet of the sender sees the incoming multicast packet and will initially have no forwarding information for that stream in its forwarding hardware. The packet is passed onto the CPU of that router, which will encapsulate this packet and send it towards a special multicast router designated the Rendez-vous Point (RP). When the RP has installed the multicast routes for this group, it will tell the multicast router on the sender’s segment to stop sending. When it does, this router installs its own multicast routes for the source tree (the tree specific to this sender) and the shared tree (the one towards the RP) without any outgoing interfaces, and the traffic is dropped at this first router. But, the network (well at least the part between the sender and the RP) is now aware of this multicast stream. And who is sending.
Now when we want to join this IP Multicast group, the first action is send an IGMP join out on the subnet you are attached to. The IP Multicast router that serves this subnet sees the join and determines where RP can be found. It takes the client join, and sends it towards the RP, using the unicast routing table as its guide. Every multicast router along the way registers that there is a listener on the interface this join came in on and passes it along towards the IP. All along this path, the unicast routing entry for the RP is used to create the tree towards the listener.
Once received by the RP, the shared tree and the source tree towards the sender have been joined. We have an end to end path between sender and receiver, with the RP in the middle of it all. All that is left is to send a join from the RP towards the router on the sender’s subnet to essentially tell it to start passing the actual multicast along the path towards the RP (the source tree), where the RP will then push it out onto the shared tree towards the destination. Voila, it’s as simple as that.
But wait, we are not done. Once the packets start to flow from source to destination, the multicast router closest to the destination will send another join message for this group, but this time towards the sender. It is only now that it can do this because those first few data packets actually indicate who the sender is. That join is passed router to router to router towards the router on the sender’s subnet, and once arrived, that router will now also start sending the multicast data along that path towards the receiver. The receiving subnet router sees that stream appearing and will now send a prune message onto the shared tree towards the RP, indicating it no longer needs the multicast stream through the RP.
If you are not familiar with IP Multicast and after reading the above are not confused, congratulations, your brain is very well wired for complex networking.
If you step away from how IGMP and PIM implement this today as above, the most fundamental of IP multicast topologies is that you need to build a forwarding tree that is rooted in the source, with the destinations as its leaves. At each intermediate node in the tree, the packets are replicated to its branches, therefore creating the least amount of duplication. And by using a tree, it is loop free, packets won’t swirl around the network bringing it to its knees.
The challenging part though is that the tree is based on the unicast forwarding topology. From a leaf on this tree towards the sender, each step is identical to how a unicast IP packet would be forwarded. The forwarding topologies are connected and dependent on each other. IP Multicast is built on top of a unicast routed infrastructure, and unicast routing changes can have dramatic impacts to the multicast forwarding topologies.
I mentioned here before that I once spent a wonderful 2 weeks in Delhi working on a network where surveillance cameras created an aggregate 8Gbit/sec worth of multicast data, with a requirement that any unicast change would have limited impact to these streams. Believe me, it is extremely hard to engineer and tune, and we had the luxury of hijacking a really large network night after night to simulate failures.
SDN based architectures have the opportunity to change all this. Multicast forwarding was designed the way it was designed to work on arbitrary network topologies, with random senders and receivers coming and going. It builds trees on the fly and on demand. For many networks, topologies are not arbitrary, and those applications that consume/produce lots of multicast do not have randomly placed senders and receivers that come and go as they please. Many of them are well known or placed in fairly static and fixed topologies.
A controller with a global view of the network can create multicast topologies ahead of time. It knows all possible replication points and can create distribution trees among them. It can create different distribution trees for different multicast groups. It can create them independent of the unicast forwarding. It can calculate backup topologies in case portions of the tree fail. And it can do all of that guaranteeing there are no loops and optimal replication. When applications indicate their participation in specific multicast streams as senders or listeners to this controller, it can optimize very specifically based on those participants. The possibilities are endless.
We had a customer visit us yesterday that has very significant multicast needs and we walked him through some of these possibilities. He left with a huge smile on his face. And that smile on his face was not because he really liked what we built (even though he did), but it was because we showed him that if you remove legacy network thinking and constraints, networking can yet again be extremely exciting and creates solutions that he did not think were possible, in a fairly simple and straightforward way. And that, in turn, is truly exciting to us.
DevOps Summit 2015 New York, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that it is now accepting Keynote Proposals. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long development cycles that produce software that is obsolete...
Jan. 28, 2015 01:15 PM EST Reads: 2,423
BMC Software plans to acquire assets of CDB Software, Inc., a mainframe data management company that has developed utilities for managing IBM DB2 databases with virtually no outage. Focusing on the availability of mission-critical applications is strategic for BMC as it continues to help its customers transform IT into a competitive advantage for their business. CDB's technology complements BMC's existing mainframe data management portfolio, which includes software utilities for DB2 administrat...
Jan. 28, 2015 01:00 PM EST Reads: 969
Amazon, Google and Facebook are household names in part because of their mastery of Big Data. But what about organizations without billions of dollars to spend on Big Data tools - how can they extract value from their data? In his session at 6th Big Data Expo®, Ali Ghodsi, Co-Founder and Head of Engineering at Databricks, discussed how the zero management cost and scalability of the cloud is addressing the challenges and pain points that data engineers face when working with Big Data. He also s...
Jan. 28, 2015 01:00 PM EST Reads: 2,299
IBM has announced software that allows people to hide or anonymize their personal information on the Web, ensuring protection from identity theft and other misuse. Developed by researchers at IBM's laboratory in Zurich, Switzerland, the software – called Identity Mixer – will enable consumers to purchase goods and services on the Internet without disclosing personal information. As consumers hand over personal details in exchange for downloading music or subscribing to online newsletters, they...
Jan. 28, 2015 01:00 PM EST Reads: 353
In this scenarios approach Joe Thykattil, Technology Architect & Sales at TimeWarner / Navisite, presented examples that will allow business-savvy professionals to make informed decisions based on a sound business model. This model covered the technology options in detail as well as a financial analysis. The TCO (Total Cost of Ownership) and ROI (Return on Investment) demonstrated how to start, develop and formulate a business case that will allow both small and large scale projects to achieve...
Jan. 28, 2015 01:00 PM EST Reads: 1,910
IBM has announced a new strategic technology services agreement with Anthem, Inc., a health benefits company in the U.S. IBM has been selected to provide operational services for Anthem's mainframe and data center server and storage infrastructure for the next five years. Among the benefits of the relationship, Anthem has the ability to leverage IBM Cloud solutions that will help increase the ease, availability and speed of adding infrastructure to support new business requirements.
Jan. 28, 2015 01:00 PM EST Reads: 1,396
The term culture has had a polarizing effect among DevOps supporters. Some propose that culture change is critical for success with DevOps, but are remiss to define culture. Some talk about a DevOps culture but then reference activities that could lead to culture change and there are those that talk about culture change as a set of behaviors that need to be adopted by those in IT. There is no question that businesses successful in adopting a DevOps mindset have seen departmental culture change, ...
Jan. 28, 2015 12:45 PM EST Reads: 2,345
The Internet of Things will greatly expand the opportunities for data collection and new business models driven off of that data. In her session at @ThingsExpo, Esmeralda Swartz, CMO of MetraTech, discussed how for this to be effective you not only need to have infrastructure and operational models capable of utilizing this new phenomenon, but increasingly service providers will need to convince a skeptical public to participate. Get ready to show them the money!
Jan. 28, 2015 12:30 PM EST Reads: 2,874
Connected devices and the Internet of Things are getting significant momentum in 2014. In his session at Internet of @ThingsExpo, Jim Hunter, Chief Scientist & Technology Evangelist at Greenwave Systems, examined three key elements that together will drive mass adoption of the IoT before the end of 2015. The first element is the recent advent of robust open source protocols (like AllJoyn and WebRTC) that facilitate M2M communication. The second is broad availability of flexible, cost-effective ...
Jan. 28, 2015 12:00 PM EST Reads: 2,420
"Our premise is Docker is not enough. That's not a bad thing - we actually love Docker. At ActiveState all our products are based on open source technology and Docker is an up-and-coming piece of open source technology," explained Bart Copeland, President & CEO of ActiveState Software, in this SYS-CON.tv interview at DevOps Summit at Cloud Expo®, held Nov 4-6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Jan. 28, 2015 11:45 AM EST Reads: 2,768
The Internet of Things will put IT to its ultimate test by creating infinite new opportunities to digitize products and services, generate and analyze new data to improve customer satisfaction, and discover new ways to gain a competitive advantage across nearly every industry. In order to help corporate business units to capitalize on the rapidly evolving IoT opportunities, IT must stand up to a new set of challenges. In his session at @ThingsExpo, Jeff Kaplan, Managing Director of THINKstrateg...
Jan. 28, 2015 11:45 AM EST Reads: 2,637
Cloud and Big Data present unique dilemmas: embracing the benefits of these new technologies while maintaining the security of your organization's assets. When an outside party owns, controls and manages your infrastructure and computational resources, how can you be assured that sensitive data remains private and secure? How do you best protect data in mixed use cloud and big data infrastructure sets? Can you still satisfy the full range of reporting, compliance and regulatory requirements? In...
Jan. 28, 2015 11:45 AM EST Reads: 2,551
The 3rd International Internet of @ThingsExpo, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that its Call for Papers is now open. The Internet of Things (IoT) is the biggest idea since the creation of the Worldwide Web more than 20 years ago.
Jan. 28, 2015 11:30 AM EST Reads: 2,912
The BPM world is going through some evolution or changes where traditional business process management solutions really have nowhere to go in terms of development of the road map. In this demo at 15th Cloud Expo, Kyle Hansen, Director of Professional Services at AgilePoint, shows AgilePoint’s unique approach to dealing with this market circumstance by developing a rapid application composition or development framework.
Jan. 28, 2015 11:30 AM EST Reads: 2,094
Eighty-five percent of companies store information in some sort of unstructured manner. In this demo at 15th Cloud Expo, Mark Fronczak, Product Manager at Solgenia, discussed their enterprise content management solution, which was created to help companies organize and take control of their digital assets.
Jan. 28, 2015 11:30 AM EST Reads: 1,659
Fundamentally, SDN is still mostly about network plumbing. While plumbing may be useful to tinker with, what you can do with your plumbing is far more intriguing. A rigid interpretation of SDN confines it to Layers 2 and 3, and that's reasonable. But SDN opens opportunities for novel constructions in Layers 4 to 7 that solve real operational problems in data centers. "Data center," in fact, might become anachronistic - data is everywhere, constantly on the move, seemingly always overflowing. Net...
Jan. 28, 2015 11:15 AM EST Reads: 2,636
Leysin American School is an exclusive, private boarding school located in Leysin, Switzerland. Leysin selected an OpenStack-powered, private cloud as a service to manage multiple applications and provide development environments for students across the institution. Seeking to meet rigid data sovereignty and data integrity requirements while offering flexible, on-demand cloud resources to users, Leysin identified OpenStack as the clear choice to round out the school's cloud strategy. Additional...
Jan. 28, 2015 11:15 AM EST Reads: 2,581
DevOps is all about agility. However, you don't want to be on a high-speed bus to nowhere. The right DevOps approach controls velocity with a tight feedback loop that not only consists of operational data but also incorporates business context. With a business context in the decision making, the right business priorities are incorporated, which results in a higher value creation. In his session at DevOps Summit, Todd Rader, Solutions Architect at AppDynamics, discussed key monitoring techniques...
Jan. 28, 2015 11:00 AM EST Reads: 2,612
Advanced Persistent Threats (APTs) are increasing at an unprecedented rate. The threat landscape of today is drastically different than just a few years ago. Attacks are much more organized and sophisticated. They are harder to detect and even harder to anticipate. In the foreseeable future it's going to get a whole lot harder. Everything you know today will change. Keeping up with this changing landscape is already a daunting task. Your organization needs to use the latest tools, methods and ex...
Jan. 28, 2015 11:00 AM EST Reads: 3,022