Welcome!

SDN Journal Authors: Liz McMillan, Yeshim Deniz, Elizabeth White, Pat Romanski, TJ Randall

Related Topics: @CloudExpo, Microservices Expo, Containers Expo Blog, Cloud Security, @DXWorldExpo, SDN Journal

@CloudExpo: Article

Six Characteristics of Best-in-Class Object Storage

Six “must haves” that are required to create best-in-class object storage solutions and why they matter to you

The adoption of cloud storage technology for a broad range of consumer and business applications is transforming the storage landscape by transitioning away from traditional disk arrays to object-based storage systems that have the scalability, availability, resiliency and accessibility to enable cloud-scale storage and instant access.

A recent IDC report predicts that the market for File- and Object-Based Storage (FOBS) will experience an annual growth rate of 24.5% through 2017, reaching $38 billion. "Increased versatility will result in more diverse use cases for FOBS," said IDC.

Software based, object storage is not saddled with the cost, complexity and vendor lock-in of legacy storage arrays or the scalability limitations of traditional file system storage. But all object storage systems are not created equal. Here are six "must haves" that are required to create best-in-class object storage solutions.

1. Granular, Automated Scalability:
Best-in-class object stores should support highly flexible scalability, spanning the addition of a single disk all the way up to multiple nodes to extend the capacity or performance of the solution.

Why this matters to you: Granular scalability lets you scale as you grow and eliminates the need to over purchase hardware because of the technical limitations of the storage solution.

2. No Single Point of Failure:
The most efficient object storage systems are built on a symmetric architecture where all nodes run the same code, resulting in high availability and unprecedented scalability, eliminating any single point of failure.

Why this matters to you: When you hear management node, controller node, or database this means more management and the addition of single points of failure that can critically impact performance, stability and fault tolerance. In highly available object storage solutions all nodes do the same thing so that if one fails, the others can immediately remedy the issue. This also eliminates the need for specialized hardware that needs to be physically shipped if an issue is discovered.

3. Flexible Data Protection on a Per-Object Basis:
Data protection flexibility is critical as no single data protection scheme can be optimized for every use case. Object storage systems need both replication and erasure coding as well as the ability to move between them, all available in the same cluster to ensure comprehensive, efficient data protection.

Why this matters to you: One size fits all just does not work in real life. Different use cases require different combinations of replication and erasure coding. Object solutions that constrain the transition from one protection scheme to the other or lock the protection scheme to specific hardware ultimately hinder growth and your ability to optimize resources. Support for both protection schemes on the same server means you can optimize for access, data protection and resource utilization system wide - without constraint.

4. Support for Large and Small Files:
Object stores must be designed with the versatility and flexibility to handle a broad range of applications and workloads without performance impact, equally adept at storing and accessing billions of small files, documents and emails or very large files like high-definition videos.

Why this matters to you: This is primarily about performance from an access perspective. The variation in file sizes will continue. While compression algorithms get more efficient in making files smaller, technological advancements will continue to add to the complexity and depth of some file types resulting in larger files. An object storage solution that ensures rapid access and efficient storage, regardless of file size or object count will increase the number of use cases reducing the number of point solutions you need to purchase.

5. Continuous Integrity Checks and Fast Volume Recovery:
Best of breed solutions continuously check content integrity from a protection scheme and content perspective. If a bad disk is discovered, recovery should be distributed with the rate of repair accelerating as the storage solution grows.

Why this matters to you: Content you store should always be available. Some object solutions only check data integrity on reads - the worst time to ensure data integrity. Others rely on specialized nodes to identify and repair issues that limit scale.

6. Instant Content Lookup and Retrieval:
Best-of-breed solutions allow queries against the object store based on object attributes or customizable metadata "tags" stored with the object.

Why this matters to you: As the amount of content grows from millions to billions of objects and management resources change (hardware migration and employee turnover) efficient content lookup and retrieval becomes a challenge. Some object solutions store metadata in a database, which introduces an additional layer of complexity between content requests and content delivery - a textbook bottleneck. Databases also become unwieldy with size and require investment in specialized management resources. By storing metadata with the object, content is self-contained and security, authentication and all identifying information is always available regardless of application, employee turnover, technological obsolescence or even time.

More Stories By Paul Carpentier

Paul Carpentier is CTO and Founder of Caringo. Known as the father of the Content Addressing concept, He invented the patent pending scalable and upgradeable security that is at the heart of Caringo. He was the architect of SequeLink — the first client/server middleware product to connect heterogeneous front ends running over multiple networks to multiple databases on the server side.

Paul founded Wave Research and conceived FileWave, the first fully automated, model-driven software distribution and management system. At FilePool, he invented the technology that created the Content Addressed Storage industry. FilePool, was sold to EMC who turned CAS into a multi-billion dollar marketplace. Caringo CAStor, based on two of Mr. Carpentier's six patents promises to revolutionize the data storage business in much the same manner that CAS created a whole new marketplace.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


CloudEXPO Stories
Sanjeev Sharma Joins November 11-13, 2018 @DevOpsSummit at @CloudEXPO New York Faculty. Sanjeev Sharma is an internationally known DevOps and Cloud Transformation thought leader, technology executive, and author. Sanjeev's industry experience includes tenures as CTO, Technical Sales leader, and Cloud Architect leader. As an IBM Distinguished Engineer, Sanjeev is recognized at the highest levels of IBM's core of technical leaders.
DXWorldEXPO LLC announced today that Kevin Jackson joined the faculty of CloudEXPO's "10-Year Anniversary Event" which will take place on November 11-13, 2018 in New York City. Kevin L. Jackson is a globally recognized cloud computing expert and Founder/Author of the award winning "Cloud Musings" blog. Mr. Jackson has also been recognized as a "Top 100 Cybersecurity Influencer and Brand" by Onalytica (2015), a Huffington Post "Top 100 Cloud Computing Experts on Twitter" (2013) and a "Top 50 Cloud Computing Blogger for IT Integrators" by CRN (2015). Mr. Jackson's professional career includes service in the US Navy Space Systems Command, Vice President J.P. Morgan Chase, Worldwide Sales Executive for IBM and NJVC Vice President, Cloud Services. He is currently part of a team responsible for onboarding mission applications to the US Intelligence Community cloud computing environment (IC ...
When applications are hosted on servers, they produce immense quantities of logging data. Quality engineers should verify that apps are producing log data that is existent, correct, consumable, and complete. Otherwise, apps in production are not easily monitored, have issues that are difficult to detect, and cannot be corrected quickly. Tom Chavez presents the four steps that quality engineers should include in every test plan for apps that produce log output or other machine data. Learn the steps so your team's apps not only function but also can be monitored and understood from their machine data when running in production.
The best way to leverage your Cloud Expo presence as a sponsor and exhibitor is to plan your news announcements around our events. The press covering Cloud Expo and @ThingsExpo will have access to these releases and will amplify your news announcements. More than two dozen Cloud companies either set deals at our shows or have announced their mergers and acquisitions at Cloud Expo. Product announcements during our show provide your company with the most reach through our targeted audiences.
When building large, cloud-based applications that operate at a high scale, it's important to maintain a high availability and resilience to failures. In order to do that, you must be tolerant of failures, even in light of failures in other areas of your application. "Fly two mistakes high" is an old adage in the radio control airplane hobby. It means, fly high enough so that if you make a mistake, you can continue flying with room to still make mistakes. In his session at 18th Cloud Expo, Lee Atchison, Principal Cloud Architect and Advocate at New Relic, discussed how this same philosophy can be applied to highly scaled applications, and can dramatically increase your resilience to failure.