Welcome!

SDN Journal Authors: Elizabeth White, Liz McMillan, John Walsh, Sven Olav Lund, Simon Hill

Related Topics: SDN Journal, Java IoT, Linux Containers, Containers Expo Blog, @CloudExpo, Cloud Security

SDN Journal: Blog Post

Fabric Engineering Is More than Traffic Engineering

Traffic engineering has taken on a very specific meaning, very much related to the way traffic is mapped onto MPLS

It is human nature to try and relate new information and new ways of doing things to something that we know, something we are familiar with. Often when we talk about the way we fit traffic onto a Plexxi mesh network, the reaction is “I know what you mean, you are doing traffic engineering like we (used to) do in MPLS”. The response to that is usually “kinda, but not really”.

In the most basic meaning, everything that has to do with the placement of traffic onto links, routing and forwarding choices being programmed, etc., would be part of Traffic Engineering. But like too many words and phrases in our networking dictionary, traffic engineering has taken on a very specific meaning, very much related to the way traffic is mapped onto MPLS and like networks. For us, its a bit different. We build ethernet and IP networks that use packet by packet forwarding rules and while individual flows may have a hop by hop crafted path through the network, a very large portion of traffic travels using regular L2 and L3 forwarding tables. We just construct those tables differently than the typical network. Dare I say we engineer the tables to ensure forwarding occurs the way we calculated it to go.

Mike Bushong yesterday talked about equal cost paths and shortest path algorithms. We have discussed our views several times in this and other forums, and it is because we have a fundamental believe that  it can be done better. A Plexxi network consists of switches that are connected together using a variety of optical technologies with a mesh of 10GbE Ethernet connections. These point to point connections between switches for the basis of the L2 and L3 connectivity that exists, on top of which Plexxi Control Engineers the forwarding behavior of the fabric (skipping the part where Control can actually change those point to point connections for now).

So far this still smells like Traffic Engineering. I have a network, I have traffic I need to put on the network. The goal is simple, maximize the available network capacity. Use it as best as you can. All of it. While providing the best possible service for the providers and consumers of the traffic. But the way we do it is just a little different.

Our fabric is engineered based on a set of information sources. The most obvious of those sources is the actual network that is constructed. How our switches are connected together, and what initial L1/L2 topography (mesh of point to point networks) is created as part of our default connectivity. Our second source of information is the set of Affinities defined for the network. What application relationships are explicitly described, what is the requested network behavior for them and most importantly, where can I find these application components on the network. The latter is learned of course, not defined by the operator. The third component is actual traffic utilization on the fabric. What links are in use and how much of it.

These three inputs go into what we call the fitting engine. Heavy duty graph theory that we have blogged about before. The goal: resolve all Affinity constraints on top of the network infrastructure provided, and provide Non Equal Weighted Egress Based Multipath Trees for all non Affinitized traffic to ensure it gets the best service and spreads traffic as best as we can across all available paths between any two points in the fabric. That last one is a mouthful and not really what we call it, but at NFD7 yesterday I used it jokingly as the best descriptive term that articulates all it actually is. Last, a set of backup paths are calculated for link and switch failures. This can number 100s of paths easily. The results of these calculations are passed back to the switches in a multi phase commit fashion, ensuring that all switches will start using the same forwarding directions at the same time. Some of these are used in the switches as explicit flow like rules (similar to perhaps OpenFlow), others are used to populate forwarding tables when new MAC addresses are learned, or ARPs are resolved.

The result is a very carefully constructed description of connectivity from anywhere to anywhere on the fabric, satisfying the needs of Affinities, and balancing the remainder of the traffic across all available bandwidth. And once completed, the next set of traffic statistics that are entered into the overall traffic modeling will provide an even better tuned view of reality. When new Affinities are defined, you can run partial computations to just layer these new requirements on top of the previously calculated topologies.

Traffic Engineering has a very 90’s meaning attached to it. At least for me it does. Maybe this is Traffic Engineering 3.0 for the Datacenter and Data Center Interconnects (see last week’s blog post). Maybe the name does not matter. It’s a very mathematically engineered fabric. Mike said it well at NFD7: we are really a math company, masquerading as a software company, masquerading as a hardware company.

[Today's fun fact: Coconuts kill more people than sharks. Each year 150 people die from falling coconuts. Umbrellas make for safer shade than coconut trees.]

The post Fabric Engineering is more than Traffic Engineering appeared first on Plexxi.

Read the original blog entry...

More Stories By Marten Terpstra

Marten Terpstra is a Product Management Director at Plexxi Inc. Marten has extensive knowledge of the architecture, design, deployment and management of enterprise and carrier networks.

@CloudExpo Stories
Agile has finally jumped the technology shark, expanding outside the software world. Enterprises are now increasingly adopting Agile practices across their organizations in order to successfully navigate the disruptive waters that threaten to drown them. In our quest for establishing change as a core competency in our organizations, this business-centric notion of Agile is an essential component of Agile Digital Transformation. In the years since the publication of the Agile Manifesto, the conn...
In his session at 21st Cloud Expo, James Henry, Co-CEO/CTO of Calgary Scientific Inc., introduced you to the challenges, solutions and benefits of training AI systems to solve visual problems with an emphasis on improving AIs with continuous training in the field. He explored applications in several industries and discussed technologies that allow the deployment of advanced visualization solutions to the cloud.
"MobiDev is a software development company and we do complex, custom software development for everybody from entrepreneurs to large enterprises," explained Alan Winters, U.S. Head of Business Development at MobiDev, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
In his session at 21st Cloud Expo, Carl J. Levine, Senior Technical Evangelist for NS1, will objectively discuss how DNS is used to solve Digital Transformation challenges in large SaaS applications, CDNs, AdTech platforms, and other demanding use cases. Carl J. Levine is the Senior Technical Evangelist for NS1. A veteran of the Internet Infrastructure space, he has over a decade of experience with startups, networking protocols and Internet infrastructure, combined with the unique ability to it...
Vulnerability management is vital for large companies that need to secure containers across thousands of hosts, but many struggle to understand how exposed they are when they discover a new high security vulnerability. In his session at 21st Cloud Expo, John Morello, CTO of Twistlock, addressed this pressing concern by introducing the concept of the “Vulnerability Risk Tree API,” which brings all the data together in a simple REST endpoint, allowing companies to easily grasp the severity of the ...
Enterprises are adopting Kubernetes to accelerate the development and the delivery of cloud-native applications. However, sharing a Kubernetes cluster between members of the same team can be challenging. And, sharing clusters across multiple teams is even harder. Kubernetes offers several constructs to help implement segmentation and isolation. However, these primitives can be complex to understand and apply. As a result, it’s becoming common for enterprises to end up with several clusters. Thi...
While some developers care passionately about how data centers and clouds are architected, for most, it is only the end result that matters. To the majority of companies, technology exists to solve a business problem, and only delivers value when it is solving that problem. 2017 brings the mainstream adoption of containers for production workloads. In his session at 21st Cloud Expo, Ben McCormack, VP of Operations at Evernote, discussed how data centers of the future will be managed, how the p...
"NetApp is known as a data management leader but we do a lot more than just data management on-prem with the data centers of our customers. We're also big in the hybrid cloud," explained Wes Talbert, Principal Architect at NetApp, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
"Space Monkey by Vivent Smart Home is a product that is a distributed cloud-based edge storage network. Vivent Smart Home, our parent company, is a smart home provider that places a lot of hard drives across homes in North America," explained JT Olds, Director of Engineering, and Brandon Crowfeather, Product Manager, at Vivint Smart Home, in this SYS-CON.tv interview at @ThingsExpo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
The question before companies today is not whether to become intelligent, it’s a question of how and how fast. The key is to adopt and deploy an intelligent application strategy while simultaneously preparing to scale that intelligence. In her session at 21st Cloud Expo, Sangeeta Chakraborty, Chief Customer Officer at Ayasdi, provided a tactical framework to become a truly intelligent enterprise, including how to identify the right applications for AI, how to build a Center of Excellence to oper...
"IBM is really all in on blockchain. We take a look at sort of the history of blockchain ledger technologies. It started out with bitcoin, Ethereum, and IBM evaluated these particular blockchain technologies and found they were anonymous and permissionless and that many companies were looking for permissioned blockchain," stated René Bostic, Technical VP of the IBM Cloud Unit in North America, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Conventi...
Coca-Cola’s Google powered digital signage system lays the groundwork for a more valuable connection between Coke and its customers. Digital signs pair software with high-resolution displays so that a message can be changed instantly based on what the operator wants to communicate or sell. In their Day 3 Keynote at 21st Cloud Expo, Greg Chambers, Global Group Director, Digital Innovation, Coca-Cola, and Vidya Nagarajan, a Senior Product Manager at Google, discussed how from store operations and ...
"Infoblox does DNS, DHCP and IP address management for not only enterprise networks but cloud networks as well. Customers are looking for a single platform that can extend not only in their private enterprise environment but private cloud, public cloud, tracking all the IP space and everything that is going on in that environment," explained Steve Salo, Principal Systems Engineer at Infoblox, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Conventio...
Large industrial manufacturing organizations are adopting the agile principles of cloud software companies. The industrial manufacturing development process has not scaled over time. Now that design CAD teams are geographically distributed, centralizing their work is key. With large multi-gigabyte projects, outdated tools have stifled industrial team agility, time-to-market milestones, and impacted P&L stakeholders.
"ZeroStack is a startup in Silicon Valley. We're solving a very interesting problem around bringing public cloud convenience with private cloud control for enterprises and mid-size companies," explained Kamesh Pemmaraju, VP of Product Management at ZeroStack, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
"We're focused on how to get some of the attributes that you would expect from an Amazon, Azure, Google, and doing that on-prem. We believe today that you can actually get those types of things done with certain architectures available in the market today," explained Steve Conner, VP of Sales at Cloudistics, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
"Codigm is based on the cloud and we are here to explore marketing opportunities in America. Our mission is to make an ecosystem of the SW environment that anyone can understand, learn, teach, and develop the SW on the cloud," explained Sung Tae Ryu, CEO of Codigm, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Data scientists must access high-performance computing resources across a wide-area network. To achieve cloud-based HPC visualization, researchers must transfer datasets and visualization results efficiently. HPC clusters now compute GPU-accelerated visualization in the cloud cluster. To efficiently display results remotely, a high-performance, low-latency protocol transfers the display from the cluster to a remote desktop. Further, tools to easily mount remote datasets and efficiently transfer...
High-velocity engineering teams are applying not only continuous delivery processes, but also lessons in experimentation from established leaders like Amazon, Netflix, and Facebook. These companies have made experimentation a foundation for their release processes, allowing them to try out major feature releases and redesigns within smaller groups before making them broadly available. In his session at 21st Cloud Expo, Brian Lucas, Senior Staff Engineer at Optimizely, discussed how by using ne...
Gemini is Yahoo’s native and search advertising platform. To ensure the quality of a complex distributed system that spans multiple products and components and across various desktop websites and mobile app and web experiences – both Yahoo owned and operated and third-party syndication (supply), with complex interaction with more than a billion users and numerous advertisers globally (demand) – it becomes imperative to automate a set of end-to-end tests 24x7 to detect bugs and regression. In th...