|By Marten Terpstra||
|May 16, 2014 12:45 PM EDT||
IP Multicast is one of those technologies that most everyone loves to hate. It’s almost the perfect example of how complicated we have made networking. Getting IP Multicast to run depends on several protocols that are all somewhat intertwined or dependent on each, their relationship sometimes explicit, sometimes implicit.
Even trying to describe the basic operation is complicated.
When an application or service provides information using IP multicast, it simply starts sending it onto a specific multicast group. The multicast router for the subnet of the sender sees the incoming multicast packet and will initially have no forwarding information for that stream in its forwarding hardware. The packet is passed onto the CPU of that router, which will encapsulate this packet and send it towards a special multicast router designated the Rendez-vous Point (RP). When the RP has installed the multicast routes for this group, it will tell the multicast router on the sender’s segment to stop sending. When it does, this router installs its own multicast routes for the source tree (the tree specific to this sender) and the shared tree (the one towards the RP) without any outgoing interfaces, and the traffic is dropped at this first router. But, the network (well at least the part between the sender and the RP) is now aware of this multicast stream. And who is sending.
Now when we want to join this IP Multicast group, the first action is send an IGMP join out on the subnet you are attached to. The IP Multicast router that serves this subnet sees the join and determines where RP can be found. It takes the client join, and sends it towards the RP, using the unicast routing table as its guide. Every multicast router along the way registers that there is a listener on the interface this join came in on and passes it along towards the IP. All along this path, the unicast routing entry for the RP is used to create the tree towards the listener.
Once received by the RP, the shared tree and the source tree towards the sender have been joined. We have an end to end path between sender and receiver, with the RP in the middle of it all. All that is left is to send a join from the RP towards the router on the sender’s subnet to essentially tell it to start passing the actual multicast along the path towards the RP (the source tree), where the RP will then push it out onto the shared tree towards the destination. Voila, it’s as simple as that.
But wait, we are not done. Once the packets start to flow from source to destination, the multicast router closest to the destination will send another join message for this group, but this time towards the sender. It is only now that it can do this because those first few data packets actually indicate who the sender is. That join is passed router to router to router towards the router on the sender’s subnet, and once arrived, that router will now also start sending the multicast data along that path towards the receiver. The receiving subnet router sees that stream appearing and will now send a prune message onto the shared tree towards the RP, indicating it no longer needs the multicast stream through the RP.
If you are not familiar with IP Multicast and after reading the above are not confused, congratulations, your brain is very well wired for complex networking.
If you step away from how IGMP and PIM implement this today as above, the most fundamental of IP multicast topologies is that you need to build a forwarding tree that is rooted in the source, with the destinations as its leaves. At each intermediate node in the tree, the packets are replicated to its branches, therefore creating the least amount of duplication. And by using a tree, it is loop free, packets won’t swirl around the network bringing it to its knees.
The challenging part though is that the tree is based on the unicast forwarding topology. From a leaf on this tree towards the sender, each step is identical to how a unicast IP packet would be forwarded. The forwarding topologies are connected and dependent on each other. IP Multicast is built on top of a unicast routed infrastructure, and unicast routing changes can have dramatic impacts to the multicast forwarding topologies.
I mentioned here before that I once spent a wonderful 2 weeks in Delhi working on a network where surveillance cameras created an aggregate 8Gbit/sec worth of multicast data, with a requirement that any unicast change would have limited impact to these streams. Believe me, it is extremely hard to engineer and tune, and we had the luxury of hijacking a really large network night after night to simulate failures.
SDN based architectures have the opportunity to change all this. Multicast forwarding was designed the way it was designed to work on arbitrary network topologies, with random senders and receivers coming and going. It builds trees on the fly and on demand. For many networks, topologies are not arbitrary, and those applications that consume/produce lots of multicast do not have randomly placed senders and receivers that come and go as they please. Many of them are well known or placed in fairly static and fixed topologies.
A controller with a global view of the network can create multicast topologies ahead of time. It knows all possible replication points and can create distribution trees among them. It can create different distribution trees for different multicast groups. It can create them independent of the unicast forwarding. It can calculate backup topologies in case portions of the tree fail. And it can do all of that guaranteeing there are no loops and optimal replication. When applications indicate their participation in specific multicast streams as senders or listeners to this controller, it can optimize very specifically based on those participants. The possibilities are endless.
We had a customer visit us yesterday that has very significant multicast needs and we walked him through some of these possibilities. He left with a huge smile on his face. And that smile on his face was not because he really liked what we built (even though he did), but it was because we showed him that if you remove legacy network thinking and constraints, networking can yet again be extremely exciting and creates solutions that he did not think were possible, in a fairly simple and straightforward way. And that, in turn, is truly exciting to us.
SYS-CON Events announced today that IBM Cloud Data Services has been named “Bronze Sponsor” of SYS-CON's 17th Cloud Expo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. IBM Cloud Data Services offers a portfolio of integrated, best-of-breed cloud data services for developers focused on mobile computing and analytics use cases.
Oct. 6, 2015 10:00 PM EDT Reads: 626
SYS-CON Events announced today that G2G3 will exhibit at SYS-CON's @DevOpsSummit Silicon Valley, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Based on a collective appreciation for user experience, design, and technology, G2G3 is uniquely qualified and motivated to redefine how organizations and people engage in an increasingly digital world.
Oct. 6, 2015 09:30 PM EDT Reads: 345
In his session at @ThingsExpo, Tony Shan, Chief Architect at CTS, will explore the synergy of Big Data and IoT. First he will take a closer look at the Internet of Things and Big Data individually, in terms of what, which, why, where, when, who, how and how much. Then he will explore the relationship between IoT and Big Data. Specifically, he will drill down to how the 4Vs aspects intersect with IoT: Volume, Variety, Velocity and Value. In turn, Tony will analyze how the key components of IoT ...
Oct. 6, 2015 08:00 PM EDT Reads: 310
When it comes to IoT in the enterprise, namely the commercial building and hospitality markets, a benefit not getting the attention it deserves is energy efficiency, and IoT’s direct impact on a cleaner, greener environment when installed in smart buildings. Until now clean technology was offered piecemeal and led with point solutions that require significant systems integration to orchestrate and deploy. There didn't exist a 'top down' approach that can manage and monitor the way a Smart Buildi...
Oct. 6, 2015 05:00 PM EDT Reads: 251
SYS-CON Events announced today that Cloud Raxak has been named “Media & Session Sponsor” of SYS-CON's 17th Cloud Expo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Raxak Protect automates security compliance across private and public clouds. Using the SaaS tool or managed service, developers can deploy cloud apps quickly, cost-effectively, and without error.
Oct. 6, 2015 04:40 PM EDT
As-a-service models offer huge opportunities, but also complicate security. It may seem that the easiest way to migrate to a new architectural model is to let others, experts in their field, do the work. This has given rise to many as-a-service models throughout the industry and across the entire technology stack, from software to infrastructure. While this has unlocked huge opportunities to accelerate the deployment of new capabilities or increase economic efficiencies within an organization, i...
Oct. 6, 2015 03:00 PM EDT Reads: 119
“All our customers are looking at the cloud ecosystem as an important part of their overall product strategy. Some see it evolve as a multi-cloud / hybrid cloud strategy, while others are embracing all forms of cloud offerings like PaaS, IaaS and SaaS in their solutions,” noted Suhas Joshi, Vice President – Technology, at Harbinger Group, in this exclusive Q&A with Cloud Expo Conference Chair Roger Strukhoff.
Oct. 6, 2015 02:45 PM EDT Reads: 370
SYS-CON Events announced today that ProfitBricks, the provider of painless cloud infrastructure, will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. ProfitBricks is the IaaS provider that offers a painless cloud experience for all IT users, with no learning curve. ProfitBricks boasts flexible cloud servers and networking, an integrated Data Center Designer tool for visual control over the...
Oct. 6, 2015 01:00 PM EDT Reads: 739
You have your devices and your data, but what about the rest of your Internet of Things story? Two popular classes of technologies that nicely handle the Big Data analytics for Internet of Things are Apache Hadoop and NoSQL. Hadoop is designed for parallelizing analytical work across many servers and is ideal for the massive data volumes you create with IoT devices. NoSQL databases such as Apache HBase are ideal for storing and retrieving IoT data as “time series data.”
Oct. 6, 2015 12:45 PM EDT Reads: 458
Clearly the way forward is to move to cloud be it bare metal, VMs or containers. One aspect of the current public clouds that is slowing this cloud migration is cloud lock-in. Every cloud vendor is trying to make it very difficult to move out once a customer has chosen their cloud. In his session at 17th Cloud Expo, Naveen Nimmu, CEO of Clouber, Inc., will advocate that making the inter-cloud migration as simple as changing airlines would help the entire industry to quickly adopt the cloud wit...
Oct. 6, 2015 12:30 PM EDT Reads: 585
As the world moves towards more DevOps and microservices, application deployment to the cloud ought to become a lot simpler. The microservices architecture, which is the basis of many new age distributed systems such as OpenStack, NetFlix and so on, is at the heart of Cloud Foundry - a complete developer-oriented Platform as a Service (PaaS) that is IaaS agnostic and supports vCloud, OpenStack and AWS. In his session at 17th Cloud Expo, Raghavan "Rags" Srinivas, an Architect/Developer Evangeli...
Oct. 6, 2015 12:15 PM EDT Reads: 110
Secure Cloud through Automated Compliance | @CloudExpo @CloudRaxak #Cloud #BigData #DevOps #Microservices
Cloud computing delivers on-demand resources that provide businesses with flexibility and cost-savings. The challenge in moving workloads to the cloud has been the cost and complexity of ensuring the initial and ongoing security and regulatory (PCI, HIPAA, FFIEC) compliance across private and public clouds. Manual security compliance is slow, prone to human error, and represents over 50% of the cost of managing cloud applications. Determining how to automate cloud security compliance is critical...
Oct. 6, 2015 12:00 PM EDT Reads: 248
SYS-CON Events announced today that VividCortex, the monitoring solution for the modern data system, will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. The database is the heart of most applications, but it’s also the part that’s hardest to scale, monitor, and optimize even as it’s growing 50% year over year. VividCortex is the first unified suite of database monitoring tools specifically desi...
Oct. 6, 2015 12:00 PM EDT Reads: 412
Organizations already struggle with the simple collection of data resulting from the proliferation of IoT, lacking the right infrastructure to manage it. They can't only rely on the cloud to collect and utilize this data because many applications still require dedicated infrastructure for security, redundancy, performance, etc. In his session at 17th Cloud Expo, Emil Sayegh, CEO of Codero Hosting, will discuss how in order to resolve the inherent issues, companies need to combine dedicated a...
Oct. 6, 2015 12:00 PM EDT Reads: 440
Culture is the most important ingredient of DevOps. The challenge for most organizations is defining and communicating a vision of beneficial DevOps culture for their organizations, and then facilitating the changes needed to achieve that. Often this comes down to an ability to provide true leadership. As a CIO, are your direct reports IT managers or are they IT leaders? The hard truth is that many IT managers have risen through the ranks based on their technical skills, not their leadership ab...
Oct. 6, 2015 11:00 AM EDT Reads: 850
Learn how Backup as a Service can help your customer base protect their data. In his session at 17th Cloud Expo, Stefaan Vervaet, Director of Strategic Alliances at HGST, will discuss the challenges of data protection in an era of exploding storage requirements, show you the benefits of a backup service for your cloud customers, and explain how the HGST Active Archive and CommVault are already enabling this service today with customer examples.
Oct. 6, 2015 11:00 AM EDT Reads: 680
SYS-CON Events announced today that Key Information Systems, Inc. (KeyInfo), a leading cloud and infrastructure provider offering integrated solutions to enterprises, will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Key Information Systems is a leading regional systems integrator with world-class compute, storage and networking solutions and professional services for the most advanced softwa...
Oct. 6, 2015 11:00 AM EDT Reads: 326
Apps and devices shouldn't stop working when there's limited or no network connectivity. Learn how to bring data stored in a cloud database to the edge of the network (and back again) whenever an Internet connection is available. In his session at 17th Cloud Expo, Bradley Holt, Developer Advocate at IBM Cloud Data Services, will demonstrate techniques for replicating cloud databases with devices in order to build offline-first mobile or Internet of Things (IoT) apps that can provide a better, ...
Oct. 6, 2015 10:45 AM EDT Reads: 452
SYS-CON Events announced today that Harbinger Systems will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Harbinger Systems is a global company providing software technology services. Since 1990, Harbinger has developed a strong customer base worldwide. Its customers include software product companies ranging from hi-tech start-ups in Silicon Valley to leading product companies in the US a...
Oct. 6, 2015 10:00 AM EDT Reads: 755
As more and more data is generated from a variety of connected devices, the need to get insights from this data and predict future behavior and trends is increasingly essential for businesses. Real-time stream processing is needed in a variety of different industries such as Manufacturing, Oil and Gas, Automobile, Finance, Online Retail, Smart Grids, and Healthcare. Azure Stream Analytics is a fully managed distributed stream computation service that provides low latency, scalable processing of ...
Oct. 6, 2015 10:00 AM EDT Reads: 734