|By Michael Bushong||
|June 24, 2014 09:45 AM EDT||
Compute started its major architectural transition several years ago with the introduction of virtualization. If you pay attention to any of the IT noise today, it should be clear that storage and networking are going through their own architectural evolutions as well. But another shift is also underway: applications are fundamentally changing as well.
An interesting dynamic in all of this is that it is near impossible for each of the four major IT areas to undergo simultaneous, coordinated evolution. Change is hard enough on its own, but changing multiple variable at once makes it difficult to anchor to anything substantial. And when change does occur along multiple fronts at the same time, the task of determining causation for new found results is challenging at best.
Understanding that business underlies much of the change, the best that the industry can collectively do is to take some things as fixed and then change around that. And so we evolve each of the silos somewhat independently, trying hard to keep in mind that the environment into which they plug is also changing. At our best, we try to intersect the various changes. But when we miss, we tend to aim towards slotting into current architectural paradigms, because that allows us the best chance to make a meaningful difference in production environments. Indeed, always playing out ahead of the horizon might be great for making future progress, but it makes building a business around the resulting innovation nigh impossible.
And so IT as a whole continues the trudge forward.
Each of the major infrastructure silos—compute, storage, and networking—has to operate with some fixed environment in mind. The most basic thing to attach to is application infrastructure. While virtualization has made compute containers that make applications portable (the reason dynamic and change are so prominent in most marketing materials), the applications themselves have remained largely static.
Except, of course, that they haven’t.
Anyone watching the major web-scale players (think: Facebook, Google, Twitter, and the like) will know that the architecture for their applications is actually significantly different than what most enterprises currently think of. Applications tend to be flatter and more distributed, typically running on bare metal to avoid some of the virtualization overhead that exists in a pure hypervisor environment.
This new breed of scale-out applications would appear to mark the beginning of the application evolution. If this is truly a trend that will only increase, how will it impact the evolution of the other silos?
The prevailing chatter across the whole of IT is about software-defined everything. The view is that compute, storage, and networking will all work in cahoots to meet application requirements. The rise of controller-based architectures is prominent in both networking and storage roadmaps, and the compute side of the house embraced central control awhile back.
But what happens if the underlying assumption that applications emerge largely unscathed turns out not to be true?
It could be that the future of datacenter architectures will hinge not on the supporting infrastructure but on the applications themselves. If this is true, we could see a re-emergence of something that we haven’t really talked about in quite awhile: the operating system itself. Sure, there is still talk about Linux and all the server tools that come with it, but the actual operating system hasn’t materially changed in quite some time.
Learning from web-scale applications
If we learn anything from the web-scale companies pushing the boundaries for application performance, it should be that the future is not necessarily about the containers in which applications run. It could be about the underlying OS itself. What if the reason massively scaled companies are embracing bare metal isn’t only about the cost? There is certainly a performance aspect to it as well.
One somewhat uncomfortable conclusion here would be that all the infrastructure work involved in handling application portability across a containerized infrastructure could be somewhat transient. I don’t mean to suggest that it is not useful; there will be a relatively long transition to any kind of new application architecture. And even if there is a transition, the persistence of mainframes should tell us all that no change is absolute or all-encompassing. But a scenario where pockets of new-era applications co-exist in data centers with legacy applications seems likely. We are already seeing this with Hadoop, but I would expect to see more applications built on new architectures.
But this does mean that future-proofing datacenter investments requires a bit more nuance than just buying and planning on scale. Highly-distributed application architectures are even more dependent on east-west traffic. For every 1 byte of traffic going in and out of the datacenter, close to 1 Gigabyte transits the datacenter fabric in some current applications. This ratio likely gets even more aggressive over time.
The bottom line
How all of this plays out is anyone’s guess. We will certainly end up with a hybrid environment supporting all kinds of application architectures making use of various underlying infrastructure architectures. But none of us should be surprised when the industry starts talking a bit more broadly about the role of the operating system going forward. And if you are making plans based on a set of assumptions anchored to current architectures, it might be worth expanding the strategic aperture some to consider how this impacts current plans.
And as a final thought, if the operating system changes, is it still defined by the server, or do we end up with large, distributed operating systems? Put differently, what is the definition of the underlying platform? Are we looking at a new era of platform that includes all of compute, storage, networking, and applications? The implications would be dramatic.
[Today’s fun fact: Potatoes have more chromosomes than humans. I wonder if that means Mr. Potato Head has the combined chromosomes or just a subset of a potato’s.]
SYS-CON Events announced today that IBM Cloud Data Services has been named “Bronze Sponsor” of SYS-CON's 17th Cloud Expo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. IBM Cloud Data Services offers a portfolio of integrated, best-of-breed cloud data services for developers focused on mobile computing and analytics use cases.
Oct. 6, 2015 10:00 PM EDT Reads: 626
SYS-CON Events announced today that G2G3 will exhibit at SYS-CON's @DevOpsSummit Silicon Valley, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Based on a collective appreciation for user experience, design, and technology, G2G3 is uniquely qualified and motivated to redefine how organizations and people engage in an increasingly digital world.
Oct. 6, 2015 09:30 PM EDT Reads: 345
In his session at @ThingsExpo, Tony Shan, Chief Architect at CTS, will explore the synergy of Big Data and IoT. First he will take a closer look at the Internet of Things and Big Data individually, in terms of what, which, why, where, when, who, how and how much. Then he will explore the relationship between IoT and Big Data. Specifically, he will drill down to how the 4Vs aspects intersect with IoT: Volume, Variety, Velocity and Value. In turn, Tony will analyze how the key components of IoT ...
Oct. 6, 2015 08:00 PM EDT Reads: 310
When it comes to IoT in the enterprise, namely the commercial building and hospitality markets, a benefit not getting the attention it deserves is energy efficiency, and IoT’s direct impact on a cleaner, greener environment when installed in smart buildings. Until now clean technology was offered piecemeal and led with point solutions that require significant systems integration to orchestrate and deploy. There didn't exist a 'top down' approach that can manage and monitor the way a Smart Buildi...
Oct. 6, 2015 05:00 PM EDT Reads: 251
SYS-CON Events announced today that Cloud Raxak has been named “Media & Session Sponsor” of SYS-CON's 17th Cloud Expo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Raxak Protect automates security compliance across private and public clouds. Using the SaaS tool or managed service, developers can deploy cloud apps quickly, cost-effectively, and without error.
Oct. 6, 2015 04:40 PM EDT
As-a-service models offer huge opportunities, but also complicate security. It may seem that the easiest way to migrate to a new architectural model is to let others, experts in their field, do the work. This has given rise to many as-a-service models throughout the industry and across the entire technology stack, from software to infrastructure. While this has unlocked huge opportunities to accelerate the deployment of new capabilities or increase economic efficiencies within an organization, i...
Oct. 6, 2015 03:00 PM EDT Reads: 119
“All our customers are looking at the cloud ecosystem as an important part of their overall product strategy. Some see it evolve as a multi-cloud / hybrid cloud strategy, while others are embracing all forms of cloud offerings like PaaS, IaaS and SaaS in their solutions,” noted Suhas Joshi, Vice President – Technology, at Harbinger Group, in this exclusive Q&A with Cloud Expo Conference Chair Roger Strukhoff.
Oct. 6, 2015 02:45 PM EDT Reads: 370
SYS-CON Events announced today that ProfitBricks, the provider of painless cloud infrastructure, will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. ProfitBricks is the IaaS provider that offers a painless cloud experience for all IT users, with no learning curve. ProfitBricks boasts flexible cloud servers and networking, an integrated Data Center Designer tool for visual control over the...
Oct. 6, 2015 01:00 PM EDT Reads: 739
You have your devices and your data, but what about the rest of your Internet of Things story? Two popular classes of technologies that nicely handle the Big Data analytics for Internet of Things are Apache Hadoop and NoSQL. Hadoop is designed for parallelizing analytical work across many servers and is ideal for the massive data volumes you create with IoT devices. NoSQL databases such as Apache HBase are ideal for storing and retrieving IoT data as “time series data.”
Oct. 6, 2015 12:45 PM EDT Reads: 458
Clearly the way forward is to move to cloud be it bare metal, VMs or containers. One aspect of the current public clouds that is slowing this cloud migration is cloud lock-in. Every cloud vendor is trying to make it very difficult to move out once a customer has chosen their cloud. In his session at 17th Cloud Expo, Naveen Nimmu, CEO of Clouber, Inc., will advocate that making the inter-cloud migration as simple as changing airlines would help the entire industry to quickly adopt the cloud wit...
Oct. 6, 2015 12:30 PM EDT Reads: 585
As the world moves towards more DevOps and microservices, application deployment to the cloud ought to become a lot simpler. The microservices architecture, which is the basis of many new age distributed systems such as OpenStack, NetFlix and so on, is at the heart of Cloud Foundry - a complete developer-oriented Platform as a Service (PaaS) that is IaaS agnostic and supports vCloud, OpenStack and AWS. In his session at 17th Cloud Expo, Raghavan "Rags" Srinivas, an Architect/Developer Evangeli...
Oct. 6, 2015 12:15 PM EDT Reads: 110
Secure Cloud through Automated Compliance | @CloudExpo @CloudRaxak #Cloud #BigData #DevOps #Microservices
Cloud computing delivers on-demand resources that provide businesses with flexibility and cost-savings. The challenge in moving workloads to the cloud has been the cost and complexity of ensuring the initial and ongoing security and regulatory (PCI, HIPAA, FFIEC) compliance across private and public clouds. Manual security compliance is slow, prone to human error, and represents over 50% of the cost of managing cloud applications. Determining how to automate cloud security compliance is critical...
Oct. 6, 2015 12:00 PM EDT Reads: 248
SYS-CON Events announced today that VividCortex, the monitoring solution for the modern data system, will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. The database is the heart of most applications, but it’s also the part that’s hardest to scale, monitor, and optimize even as it’s growing 50% year over year. VividCortex is the first unified suite of database monitoring tools specifically desi...
Oct. 6, 2015 12:00 PM EDT Reads: 412
Organizations already struggle with the simple collection of data resulting from the proliferation of IoT, lacking the right infrastructure to manage it. They can't only rely on the cloud to collect and utilize this data because many applications still require dedicated infrastructure for security, redundancy, performance, etc. In his session at 17th Cloud Expo, Emil Sayegh, CEO of Codero Hosting, will discuss how in order to resolve the inherent issues, companies need to combine dedicated a...
Oct. 6, 2015 12:00 PM EDT Reads: 440
Culture is the most important ingredient of DevOps. The challenge for most organizations is defining and communicating a vision of beneficial DevOps culture for their organizations, and then facilitating the changes needed to achieve that. Often this comes down to an ability to provide true leadership. As a CIO, are your direct reports IT managers or are they IT leaders? The hard truth is that many IT managers have risen through the ranks based on their technical skills, not their leadership ab...
Oct. 6, 2015 11:00 AM EDT Reads: 850
Learn how Backup as a Service can help your customer base protect their data. In his session at 17th Cloud Expo, Stefaan Vervaet, Director of Strategic Alliances at HGST, will discuss the challenges of data protection in an era of exploding storage requirements, show you the benefits of a backup service for your cloud customers, and explain how the HGST Active Archive and CommVault are already enabling this service today with customer examples.
Oct. 6, 2015 11:00 AM EDT Reads: 680
SYS-CON Events announced today that Key Information Systems, Inc. (KeyInfo), a leading cloud and infrastructure provider offering integrated solutions to enterprises, will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Key Information Systems is a leading regional systems integrator with world-class compute, storage and networking solutions and professional services for the most advanced softwa...
Oct. 6, 2015 11:00 AM EDT Reads: 326
Apps and devices shouldn't stop working when there's limited or no network connectivity. Learn how to bring data stored in a cloud database to the edge of the network (and back again) whenever an Internet connection is available. In his session at 17th Cloud Expo, Bradley Holt, Developer Advocate at IBM Cloud Data Services, will demonstrate techniques for replicating cloud databases with devices in order to build offline-first mobile or Internet of Things (IoT) apps that can provide a better, ...
Oct. 6, 2015 10:45 AM EDT Reads: 452
SYS-CON Events announced today that Harbinger Systems will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Harbinger Systems is a global company providing software technology services. Since 1990, Harbinger has developed a strong customer base worldwide. Its customers include software product companies ranging from hi-tech start-ups in Silicon Valley to leading product companies in the US a...
Oct. 6, 2015 10:00 AM EDT Reads: 755
As more and more data is generated from a variety of connected devices, the need to get insights from this data and predict future behavior and trends is increasingly essential for businesses. Real-time stream processing is needed in a variety of different industries such as Manufacturing, Oil and Gas, Automobile, Finance, Online Retail, Smart Grids, and Healthcare. Azure Stream Analytics is a fully managed distributed stream computation service that provides low latency, scalable processing of ...
Oct. 6, 2015 10:00 AM EDT Reads: 734