Welcome!

SDN Journal Authors: Liz McMillan, Pat Romanski, Elizabeth White, ManageEngine IT Matters, Rishi Bhargava

Related Topics: SDN Journal, Java IoT, Microservices Expo, Containers Expo Blog, @CloudExpo, @BigDataExpo

SDN Journal: Article

Using OpenFlow to Extend Software Defined Networking

OpenFlow’s Role in One-Stop Shopping for Control of Enterprise Cloud, Storage and Network Resources

Using OpenFlow to extend software-defined networking (SDN) to the optical layer is a compelling prospect for enterprises seeking to achieve joint orchestration of information technology (IT) and network resources for cloud services, to virtualize the network and to more simply manage interconnections of distributed data centers that require synchronization.

Today's fragmented, specialized management and control approaches are fraught with proprietary protocols and management systems, limited scalability and configuration complexities. With an OpenFlow-enabled transport network, an enterprise could instead engage in a kind of "one-stop shopping" for control of cloud computing, storage and networking resources - all via one, unified application programming interface (API). The benefits could include significantly simplified configuration, management and scaling of large-scale enterprise infrastructures through integration and automation.

That's a new role for OpenFlow, demanding strategic tailoring of the protocol for the optical transport domain. Demonstration and development of the capability are closely watched by enterprises that are under incessant pressure to cost-effectively meet ever-increasing demand for bandwidth and services.

Virtualization's New Frontier
Servers and storage have been virtualized in the enterprise; the next great frontier for virtualization is the network.

Because of the substantial cost savings and performance benefits that it can deliver, SDN-based virtualization is of prime interest to enterprises for a wide range of applications. OpenFlow has emerged as one of the most popular SDN protocols. Web 2.0 network operators and national research and education network (NREN) operators, especially, like OpenFlow.

With OpenFlow, an abstraction of the network's packet switches can be generated and flow-forwarding behavior can be specified across an infrastructure via an external controller. Operations can be substantially automated and streamlined by breaking up the monolithically integrated control and forwarding paradigm of today's switches.

Using OpenFlow, could SDN be extended across layers and create a scenario in which - with a single instruction - the controller could jointly create virtual machines and enable enterprise network administrators to reserve computing, networking and storage resources in one stroke?

It is an obviously compelling notion for enterprise network staffs who desperately need to simplify operations. However, the problem is OpenFlow deployment and development has largely been limited to the electrical packet layer, whereas the interconnection beyond the data center is typically comprised of optical transport technology. Furthermore, the optical domain is where things get hazy for many enterprise network administrators. Their comfort zone tends to be packets - not wavelengths and optics.

The result is that cloud computing is currently decoupled from the transport networking control and operation. The network exists as a static, separated entity in today's cloud implementations. There is no interaction between cloud computing processes and the statically configured network. The two are not in any way interoperable; they speak different languages.

Converging cloud computing and networking requires a more dynamic mode of control and operation, but enterprises largely have judged integrating management of the optical network into the data-center environment to be too complex.

To extend OpenFlow from its established role in the electrical packet domain to the optical layer (and, thereby, extend SDN across multiple network layers), a range of optical-specific concerns must be tackled.

Crafting and Experimenting
Within the European Commission's FP7 ICT Work Programme is a collaborative project, "OpenFlow in Europe - Linking Infrastructure and Applications" (OFELIA), that provides researchers with a test bed in which to experiment with SDN applications and virtual multi-layer networks over shared network infrastructure.

Via standardized, secure interfaces through GÉANT, a high-bandwidth interconnection of European R&E networks, researchers develop, run and control experiments using packet switches and application servers at the University of Essex and seven other test-bed facilities throughout Europe.

OFELIA hosts a prototype implementation of dynamic control of wavelength-switched optical networks via OpenFlow. Bandwidth, latency and power consumption can be adjusted to meet the specific requirements of specific applications.

To make it happen, key OpenFlow additions had to be engineered in order for the protocol to effectively control the optical domain. Optical-specific considerations were required to adapt OpenFlow from the packet world. A packet can travel from any ingress to any egress port in an electrical switch or from any time slot in a time-division multiplexing (TDM) device. The optical domain, however, introduces strict switching constraints, with regard to wavelength continuity, optical impairments, optical power leveling on the line side, etc.

Augmenting OpenFlow to address those optical-specific concerns has resulted in an OFELIA prototype that demonstrates a truly transparent, wavelength-switched optical network. The research community is able to experiment with the capability via a flexible, Web-services approach; commercial enterprises, too, are interested in trialing the capability for their specific applications and environments.

OpenFlow is not sufficient in itself to enable the complete transformation that enterprise network administrators envision, to SDN-enable virtualization across all layers of their infrastructures. The additions to OpenFlow that were engineered for the OFELIA test bed provide only the bridge between the optical layer and packet layer and allow integration into a cloud operating system such as OpenStack.

But that is one very important bridge, and the promise for enterprise network administrators is considerable. The OpenFlow innovation could seamlessly integrate the optical transport network under a common management umbrella with an enterprise's routers and switches - all via one familiar interface. Management of the optical domain could become as simple as the management of Ethernet boxes - using an encapsulation of virtual resources that enterprise network administrators could manage via typical and familiar infrastructure. That's a significant breakthrough. With many enterprises already considering usage of an OpenFlow-based control for their packet networks, extending the framework to the wavelength-switched optical layer would be a natural migration.

Virtualization has developed over phases in enterprise networking. First, resource virtualization inside data centers delivered economic savings through enhanced utilization, scalability and redundancy. Data-center virtualization conveyed greater infrastructure flexibility, higher availability and better workload balancing. The next frontier, network virtualization, promises true platform agility and, with it, a host of long-sought-after enterprise capabilities: capacity on-demand, adaptive infrastructure and dynamic service automation, among them. Adapting OpenFlow and extending SDN to the optical transport domain comprise an important step toward that vision.

More Stories By Jörg-Peter Elbers

Jörg-Peter Elbers is VP Advanced Technology in the CTO office at ADVA Optical Networking in Munich, Germany, and is globally responsible for technology strategy, new product concepts, standardization, and research activities.

Prior to joining ADVA in 2007, he was Director of Technology in the Optical Product Unit of Marconi (now Ericsson). From 1999 to 2001, he worked at Siemens AG, last as Director of Network Architecture in Siemens Optical Networks.

More Stories By Achim Autenrieth

Achim Autenrieth is Principle Research Engineer Advanced Technology in the CTO Office at ADVA Optical Networking, where he is working on the design and evaluation of multilayer networks, control plane, and SDN concepts. Achim is a member of IEEE and VDE/ITG, he authored or co-authored more than 70 reviewed and invited scientific publications and he is technical program committee member of ECOC, DRCN, and RNDM.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@CloudExpo Stories
"Operations is sort of the maturation of cloud utilization and the move to the cloud," explained Steve Anderson, Product Manager for BMC’s Cloud Lifecycle Management, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
The cloud competition for database hosts is fierce. How do you evaluate a cloud provider for your database platform? In his session at 18th Cloud Expo, Chris Presley, a Solutions Architect at Pythian, gave users a checklist of considerations when choosing a provider. Chris Presley is a Solutions Architect at Pythian. He loves order – making him a premier Microsoft SQL Server expert. Not only has he programmed and administered SQL Server, but he has also shared his expertise and passion with b...
Unless your company can spend a lot of money on new technology, re-engineering your environment and hiring a comprehensive cybersecurity team, you will most likely move to the cloud or seek external service partnerships. In his session at 18th Cloud Expo, Darren Guccione, CEO of Keeper Security, revealed what you need to know when it comes to encryption in the cloud.
We're entering the post-smartphone era, where wearable gadgets from watches and fitness bands to glasses and health aids will power the next technological revolution. With mass adoption of wearable devices comes a new data ecosystem that must be protected. Wearables open new pathways that facilitate the tracking, sharing and storing of consumers’ personal health, location and daily activity data. Consumers have some idea of the data these devices capture, but most don’t realize how revealing and...
What are the successful IoT innovations from emerging markets? What are the unique challenges and opportunities from these markets? How did the constraints in connectivity among others lead to groundbreaking insights? In her session at @ThingsExpo, Carmen Feliciano, a Principal at AMDG, will answer all these questions and share how you can apply IoT best practices and frameworks from the emerging markets to your own business.
Ask someone to architect an Internet of Things (IoT) solution and you are guaranteed to see a reference to the cloud. This would lead you to believe that IoT requires the cloud to exist. However, there are many IoT use cases where the cloud is not feasible or desirable. In his session at @ThingsExpo, Dave McCarthy, Director of Products at Bsquare Corporation, will discuss the strategies that exist to extend intelligence directly to IoT devices and sensors, freeing them from the constraints of ...
You think you know what’s in your data. But do you? Most organizations are now aware of the business intelligence represented by their data. Data science stands to take this to a level you never thought of – literally. The techniques of data science, when used with the capabilities of Big Data technologies, can make connections you had not yet imagined, helping you discover new insights and ask new questions of your data. In his session at @ThingsExpo, Sarbjit Sarkaria, data science team lead ...
SYS-CON Events announced today the Kubernetes and Google Container Engine Workshop, being held November 3, 2016, in conjunction with @DevOpsSummit at 19th Cloud Expo at the Santa Clara Convention Center in Santa Clara, CA. This workshop led by Sebastian Scheele introduces participants to Kubernetes and Google Container Engine (GKE). Through a combination of instructor-led presentations, demonstrations, and hands-on labs, students learn the key concepts and practices for deploying and maintainin...
Extracting business value from Internet of Things (IoT) data doesn’t happen overnight. There are several requirements that must be satisfied, including IoT device enablement, data analysis, real-time detection of complex events and automated orchestration of actions. Unfortunately, too many companies fall short in achieving their business goals by implementing incomplete solutions or not focusing on tangible use cases. In his general session at @ThingsExpo, Dave McCarthy, Director of Products...
Cloud analytics is dramatically altering business intelligence. Some businesses will capitalize on these promising new technologies and gain key insights that’ll help them gain competitive advantage. And others won’t. Whether you’re a business leader, an IT manager, or an analyst, we want to help you and the people you need to influence with a free copy of “Cloud Analytics for Dummies,” the essential guide to this explosive new space for business intelligence.
Traditional IT, great for stable systems of record, is struggling to cope with newer, agile systems of engagement requirements coming straight from the business. In his session at 18th Cloud Expo, William Morrish, General Manager of Product Sales at Interoute, outlined ways of exploiting new architectures to enable both systems and building them to support your existing platforms, with an eye for the future. Technologies such as Docker and the hyper-convergence of computing, networking and sto...
With an estimated 50 billion devices connected to the Internet by 2020, several industries will begin to expand their capabilities for retaining end point data at the edge to better utilize the range of data types and sheer volume of M2M data generated by the Internet of Things. In his session at @ThingsExpo, Don DeLoach, CEO and President of Infobright, discussed the infrastructures businesses will need to implement to handle this explosion of data by providing specific use cases for filterin...
IoT generates lots of temporal data. But how do you unlock its value? You need to discover patterns that are repeatable in vast quantities of data, understand their meaning, and implement scalable monitoring across multiple data streams in order to monetize the discoveries and insights. Motif discovery and deep learning platforms are emerging to visualize sensor data, to search for patterns and to build application that can monitor real time streams efficiently. In his session at @ThingsExpo, ...
Enterprise networks are complex. Moreover, they were designed and deployed to meet a specific set of business requirements at a specific point in time. But, the adoption of cloud services, new business applications and intensifying security policies, among other factors, require IT organizations to continuously deploy configuration changes. Therefore, enterprises are looking for better ways to automate the management of their networks while still leveraging existing capabilities, optimizing perf...
Early adopters of IoT viewed it mainly as a different term for machine-to-machine connectivity or M2M. This is understandable since a prerequisite for any IoT solution is the ability to collect and aggregate device data, which is most often presented in a dashboard. The problem is that viewing data in a dashboard requires a human to interpret the results and take manual action, which doesn’t scale to the needs of IoT.
When building large, cloud-based applications that operate at a high scale, it’s important to maintain a high availability and resilience to failures. In order to do that, you must be tolerant of failures, even in light of failures in other areas of your application. “Fly two mistakes high” is an old adage in the radio control airplane hobby. It means, fly high enough so that if you make a mistake, you can continue flying with room to still make mistakes. In his session at 18th Cloud Expo, Lee...
Internet of @ThingsExpo has announced today that Chris Matthieu has been named tech chair of Internet of @ThingsExpo 2016 Silicon Valley. The 6thInternet of @ThingsExpo will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Continuous testing helps bridge the gap between developing quickly and maintaining high quality products. But to implement continuous testing, CTOs must take a strategic approach to building a testing infrastructure and toolset that empowers their team to move fast. Download our guide to laying the groundwork for a scalable continuous testing strategy.
What does it look like when you have access to cloud infrastructure and platform under the same roof? Let’s talk about the different layers of Technology as a Service: who cares, what runs where, and how does it all fit together. In his session at 18th Cloud Expo, Phil Jackson, Lead Technology Evangelist at SoftLayer, an IBM company, spoke about the picture being painted by IBM Cloud and how the tools being crafted can help fill the gaps in your IT infrastructure.
CenturyLink has announced that application server solutions from GENBAND are now available as part of CenturyLink’s Networx contracts. The General Services Administration (GSA)’s Networx program includes the largest telecommunications contract vehicles ever awarded by the federal government. CenturyLink recently secured an extension through spring 2020 of its offerings available to federal government agencies via GSA’s Networx Universal and Enterprise contracts. GENBAND’s EXPERiUS™ Application...