SlideShare une entreprise Scribd logo
1  sur  11
Télécharger pour lire hors ligne
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE:
CHOOSING THE STORAGE INFRASTRUCTURE THAT BEST FITS
YOUR BUSINESS NEEDS
BY
JON TOIGO
CHAIRMAN
THE DATA MANAGEMENT INSTITUTE
SUMMARY
Software-Defined Storage (SDS) has become a meme in industry and trade press discussions of
storage technology lately, though the term itself lacks rigorous technical definition. Essentially,
SDS is touted as a model for building storage that will work better with virtualized workloads
running under server hypervisor technology than do “legacy” NAS and SAN infrastructure.
Regardless of the veracity of these claims, the business-savvy IT planner should base his or her
choice of storage infrastructure not on trendy memes, but on traditional selection criteria: cost,
availability, and simplicity.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 2
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
INTRODUCTION
IT planners have been hearing a lot about software-defined storage (SDS) lately -- and how it is
supposed to replace all of the “legacy storage” that companies have deployed for the past
decade or so, including Storage Area Networks (SANs) and Network Attached Storage (NAS)
appliances. Unfortunately, many of the arguments advanced by evangelists for adopting SDS
are only partially grounded in fact or have nothing whatsoever to do with improvements in
storage performance, allocation or utilization efficiency.
The purpose of this paper is to quickly review the arguments for SDS, to provide the proper
context for analysis of the technologies, and to help guide the business-savvy IT manager or
planner to basic common-sense criteria for evaluating the various products available to
implement the SDS model if it is deemed suitable. In the final analysis, the kind of storage
infrastructure that is appropriate for a data center, branch office or small business equipment
room should be determined by what the application workload requires, first and foremost.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 3
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
WHY SDS?
Software-defined storage is touted by evangelists as the fix for what ails traditional or legacy
storage – as the next step in storage evolution. It is unclear what that means beyond the
obvious marketing appeal of flowery rhetoric. If there is any sort of evolution in storage, it has
been the movement away from isolated “islands” of storage that created obstacles to accessing
and sharing stored data, and toward storage infrastructure models that enabled greater data
sharing.
In the 1960’s and 70’s, the predominant forms of storage were internal storage (storage devices
mounted inside the server frame itself) and direct-attached storage (a crate of disk drives
connected to the outside of the server cabinet via an external extension of the server
mainboard bus). From those implementation models, storage technology expanded in two
directions: one supporting the storage and sharing of file-based data across a network; the
other supporting the storage and sharing of block data from database and transaction-oriented
systems. Block data storage “evolved” from island architectures (internal and directly attached
storage arrays that were accessed only through the server itself, making stored data difficult to
access, share or scale efficiently) to shareable arrays – that is, arrays with multiple ports for
connecting many server hosts. Scaling these systems, however, became problematic, so
engineers figured out ways to attach storage arrays into a switched fabric (a sort of rudimentary
network) using a serial SCSI protocol such as Fibre Channel or SCSI over IP (iSCSI). The ultimate
goal was to enable the storage infrastructure to be shared by all applications by creating a true
“storage area network” or SAN.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 4
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
However, the holy grail of an open and heterogeneous SAN never appeared in the market.
Vendors preferred to keep their storage equipment sales profitable by adding functionality to
the proprietary controllers installed in their external arrays, and never fully participated in any
sort of scheme to enable different equipment from different vendors to be managed in concert.
Absent common management, the fabric SAN was incredibly pricey to own because of all the
proprietary value-add software on the proprietary controllers, and to operate because of the
lack of a universal management scheme and the need to embed expensive storage experts in IT
staff.
The above situation already existed when server virtualization took hold in the early 2000s. This
made legacy storage an easy target for server hypervisor vendors to blame for application
performance issues that arose following the consolidation of application workloads on fewer
server platforms.
Truth be told, whatever inefficiencies that already existed in SANs and NAS, server virtualization
workloads just made them worse. Consolidating applications onto fewer servers dramatically
changed the amount of I/O emanating from a single server, requiring on average the addition of
7 to 16 I/O connections per server. Hypervisor computing also changed the traffic patterns on
networks and fabric interconnects and switching systems.
Hypervisor vendors assured customers that many good things would accrue to this model. For
one, virtual machines could be replicated from host to host and joined into highly available
clusters. Virtual machines could also move around server to server to distribute load more
efficiently and to provide high availability.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 5
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
Only, consolidating a lot of virtual machines into a single server host, then migrating workload
at will from one physical server to another, created challenges in terms of maintaining
connections between applications and their stored data. When an app moved locations, often
intervention was required by an administrator to provide the application with a new route to
access its storage shares from a new server perch. Moreover, we were introduced to a new
problem that resulted from combining the I/O from multiple workloads in the same server –
something that the industry has started to refer to as the “I/O blender effect.” In operation,
every hosted application will simply write its data out at random into a shared I/O path.
Without special technology to intervene, this randomized I/O will eventually clog memory and
disk storage devices and slow applications down to a crawl.
With application performance issues becoming the number one complaint of hypervisor
computing customers, hypervisor vendors pushed several explanations that blamed “legacy
storage” inefficiencies and began arguing for the “ripping and replacing” of older storage
infrastructure and its replacement with something new, which they came to call software-
defined storage.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 6
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
HYPERVISOR VENDOR ARGUMENTS OFTEN FLAWED, BUT SDS IS VIABLE
NONETHELESS
Surveying the literature, hypervisor vendors have developed five reasons to de-provision legacy
storage. Some are valid arguments while others are based on half-truths or downright
falsehoods.
1. “Slow application performance due to ‘legacy SAN (and NAS) infrastructure’” – This is
often a falsehood. You can see whether the storage I/O path is causing application
delays rather simply by checking storage I/O queue depths. In many cases, queue
depths are not significant (meaning that there is no data waiting in line to be written to
disk). When this queue depth reading is combined with a processor cycle or processor
activity measurement showing high rates of processor cycling, a more logical conclusion
is that a log jam exists in “raw I/O” – the communications path between CPU and
memory where application processing occurs. If that is the case, the application
slowdown is a function of either application code or hypervisor code, not storage at all.
2. “High cost of storage (OPEX & CAPEX) due to proprietary storage” – Legacy storage was
and is without a doubt expensive. Hardware vendors argue that value-add software
functionality added to array controllers are differentiators of their products and
represent much R&D that needs to be rewarded. Unfortunately, value-add often adds
configuration complexity and expensive administrative staff skills requirements, and
sometimes interferes with unified management of the kit as part of infrastructure.
3. “Poor utilization efficiency and management difficulty due to lack of common
management” – Per #2, proprietary value-add storage is difficult to manage in common,
and this has long been a reason to argue for infrastructure that might deliver better
management.
4. “Lack of agility due to legacy infrastructure” – Allocating and de-allocating storage
resources and services is much more difficult with proprietary and complicated value-
add storage. However, this has nothing whatsoever to do with application performance.
5. “DAS storage outperforms SAN” – This is not true and makes no sense since both NAS
and SAN are also direct attached storage configurations. NAS is storage directly-attached
to a thin server that provides network-based access to data. A SAN is direct-attached
storage with a physical or fabric-layer switch that makes and breaks server/storage
connections at high speed, giving the illusion of network attachment.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 7
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
Valid or not, the arguments of hypervisor vendors regarding need to abandon legacy storage
have resonated. In its place, vendors are recommending something called software-defined
storage – which is, once again, storage that is physically, directly attached to virtual server hosts.
The only difference between this and direct-attached storage in the past is that value-add
storage services (software) are not located on the controller of the external array, but are
instead instantiated in a software layer in the server, usually as part of the hypervisor vendor’s
software stack.
Vendors can’t agree, of course, on a single definition for software defined storage. Instead, it is
simply a panacea – a cure-all for everything that is wrong with “legacy storage” for use in a
software-defined data center. The key difference is where the intelligence is located: is it on
the array controller or is it somewhere in the server hypervisor software stack?
Actually, there are two sets of discriminators to consider when looking at the many SDS
products from hypervisor vendors and independent software developers that have appeared in
the market of late.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 8
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
Some SDS solutions are hypervisor-dedicated or hypervisor-dependent. Certainly, this is the
case with VMware’s Virtual SAN, which only works with the company’s proprietary hypervisor.
To a lesser extent, Microsoft’s Clustered Storage Spaces is proprietary to Microsoft, though they
do claim to be able to share their storage with VMware (you just convert your VMware
workload into Microsoft VHD format and import it into Hyper-V so you can share the Microsoft
SDS infrastructure!).
The other end of the spectrum are the many third party SDS offerings, including StorMagic’s
SvSAN technology, that are hypervisor-agnostic. Vendors of these solutions are earnestly trying
to make their software-defined storage infrastructure work with multiple hypervisors and, in
some cases, with non-virtualized workload as well. Some implement a common storage
environment into which the data from different hypervisor workloads can be written, while
others enable common management of storage infrastructures created with their software but
dedicated to specific hypervisor workloads and their data. Either way, they are more robust
than hypervisor-dedicated solutions and offer a better fit for companies that are deploying
more than one hypervisor or that may do so in the future.
The other differentiator to consider is hardware dependency and hardware agnosticism.
Despite the ideology of software-defined (“use any hardware you want”), the truth is that some
software-defined solutions have fairly rigid requirements when it comes to hardware
components and topology. Some of the latest “hyper-converged infrastructure appliances” for
example are just another kind of proprietary storage array. Logically, these products would be
placed on the hardware-dependent side of the SDS solution spectrum. Hypervisor-dependent
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 9
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
SDS products like VMware’s EVO:RAIL are also hardware dependent, and even their VMware
Virtual SAN carries with it some fairly exacting and expensive hardware and topology
requirements, including a minimum of three storage nodes.
In the middle of this spectrum are a number of third party virtual SAN SDS solutions, which
unlike their hypervisor software cousins, do not require special configurations or limit what kind
of disk or what kind of flash devices you can use. StorMagic, for example, provides a solution
that can embrace a wide range of hardware components.
On the hardware-agnostic end of the spectrum are storage virtualization software products. To
SDS purists, storage virtualization is not software-defined storage because SDS seeks to
decouple storage capacity from storage value add services functionality and performance;
capacity virtualization is not part of their SDS model.
In general, if you are considering an SDS solution, your storage infrastructure requirements are
usually best served by pursuing hardware- and hypervisor-agnostic SDS solutions. That decision
is more likely to enable you to customize your storage gear selections to your needs and budget
and to consolidate the management of different storage infrastructures supporting different
workloads.
If you are operating a large shop, such as a managed hosting environment or a large enterprise
data center, moving to a Virtual SAN may actually be a step backward on the evolutionary chart.
A real SAN, assuming that one is ever delivered to market, will deliver all of the capabilities and
cost metrics that you need for such environments.
On the other hand, if yours is a small to medium sized business or if you are tasked to provide
storage infrastructure for remote or branch office operations, big centralized storage may not be
your best or most strategic infrastructure choice. You may be attracted to VMware’s virtual SAN
or Microsoft’s all-in-one Cluster Storage Spaces SDS solution as a template for building your
storage.
That is, until you find out what it will cost.
The fact is that VMware’s Virtual SAN only works with VMware vSphere workloads and requires
you to limit your hardware to a specific list of supported equipment. The same is true of
Microsoft’s Clustered Storage Spaces model, which is heavily dependent on SAS storage. Such
equipment requirements may not be a deal breaker for firms that want a simple single vendor
solution, but they do limit options. Planners may not be able to take advantage of newer
storage technologies or leverage existing investments in storage equipment that they have
today.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 10
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
More to the point, the cost of implementing a hypervisor dependent storage stack can be
prohibitive. A recent lab evaluation published in a popular computing publication placed the
cost of a basic Virtual SAN implementation, which requires three nodes minimum, at $11 to 14K
in software licenses per node and about that much again for the hardware that is approved by
VMware for use in their infrastructure. The prices for Microsoft are significantly lower, but
there is still a minimum 3 node requirement and a preference for PCIe flash and SAS disk drives
rather than SATA.
One significant benefit of using a third party Software Defined Storage solution is that it can
insulate the planner from hardware lock-ins and usually does not require a three-node
minimum hardware cluster. SDS can be implemented initially as a much less expensive two
node cluster.
For the smaller shop, or the ROBO environment, the independent software vendor’s SDS
solution might be just the thing to test the value of software defined storage to application
performance.
50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 11
© COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED.
CONCLUSION
The fit for software-defined storage, whether from your hypervisor vendor or from an
independent developer like StorMagic, should be a function of its fit with deployed (and likely
to be deployed) applications and their requirements, with strategic goals for infrastructure and
data management, and with budget, staff skills and other practical boundary conditions. It is a
good idea to determine what your needs are before you seek out an SDS solution.
It is worth noting that, in the absence of uniform standards or guarantees of interoperability
between different software stacks, IT planners may be looking at a need to deploy and manage
different SDS solutions to meet different requirements. A product aimed at large enterprise
shops and hosting environments may not be well suited to deployment in a ROBO environment,
for example.
It is a good idea to try the SDS solution before you buy it. StorMagic offers a 60 day trial version
of its product, StorMagic SvSAN, that you can access from the company website
(http://www.stormagic.com/60-day-free-trial/).

Contenu connexe

Tendances

Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-SeagateLowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Michael K. Connolly, MBATM
 
Red Hat Software Defined Storage
Red Hat Software Defined StorageRed Hat Software Defined Storage
Red Hat Software Defined Storage
DLT Solutions
 

Tendances (20)

Software-Defined Storage (SDS)
Software-Defined Storage (SDS)Software-Defined Storage (SDS)
Software-Defined Storage (SDS)
 
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-SeagateLowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
Lowering Cloud and Data Center TCO with SAS Based Storage - Xyratex-Seagate
 
Red Hat Software Defined Storage
Red Hat Software Defined StorageRed Hat Software Defined Storage
Red Hat Software Defined Storage
 
G11.2014 magic quadrant for general-purpose disk
G11.2014   magic quadrant for general-purpose diskG11.2014   magic quadrant for general-purpose disk
G11.2014 magic quadrant for general-purpose disk
 
Presidio Data Center Practice Overview
Presidio Data Center Practice OverviewPresidio Data Center Practice Overview
Presidio Data Center Practice Overview
 
DataCore + Fujitsu Business Solutions
DataCore + Fujitsu Business SolutionsDataCore + Fujitsu Business Solutions
DataCore + Fujitsu Business Solutions
 
Software defined storage
Software defined storageSoftware defined storage
Software defined storage
 
Engineering the Enterprise Storage Infrastructure with the IBM DS8000
Engineering the Enterprise Storage Infrastructure with the IBM DS8000Engineering the Enterprise Storage Infrastructure with the IBM DS8000
Engineering the Enterprise Storage Infrastructure with the IBM DS8000
 
Workload Centric Scale-Out Storage for Next Generation Datacenter
Workload Centric Scale-Out Storage for Next Generation DatacenterWorkload Centric Scale-Out Storage for Next Generation Datacenter
Workload Centric Scale-Out Storage for Next Generation Datacenter
 
Advantages of Mainframe Replication With Hitachi VSP
Advantages of Mainframe Replication With Hitachi VSPAdvantages of Mainframe Replication With Hitachi VSP
Advantages of Mainframe Replication With Hitachi VSP
 
The Whats, Whys and Hows of Database as a Service
The Whats, Whys and Hows of Database as a ServiceThe Whats, Whys and Hows of Database as a Service
The Whats, Whys and Hows of Database as a Service
 
IMEXresearch software defined storage
IMEXresearch software defined storageIMEXresearch software defined storage
IMEXresearch software defined storage
 
Comparing Cost of Dell EMC Centera and HPE/SUSE/iTernity iCAS
Comparing Cost of Dell EMC Centera and HPE/SUSE/iTernity iCASComparing Cost of Dell EMC Centera and HPE/SUSE/iTernity iCAS
Comparing Cost of Dell EMC Centera and HPE/SUSE/iTernity iCAS
 
Application Report: Virtualizing Tier-1 Workloads using FC SANs
Application Report: Virtualizing Tier-1 Workloads using FC SANsApplication Report: Virtualizing Tier-1 Workloads using FC SANs
Application Report: Virtualizing Tier-1 Workloads using FC SANs
 
Storage Analytics: Transform Storage Infrastructure Into a Business Enabler
Storage Analytics: Transform Storage Infrastructure Into a Business EnablerStorage Analytics: Transform Storage Infrastructure Into a Business Enabler
Storage Analytics: Transform Storage Infrastructure Into a Business Enabler
 
UNIVERGE BLUE Backup as a Service Case Study: B & A
UNIVERGE BLUE Backup as a Service Case Study: B & AUNIVERGE BLUE Backup as a Service Case Study: B & A
UNIVERGE BLUE Backup as a Service Case Study: B & A
 
Build the Optimal Mainframe Storage Architecture
Build the Optimal Mainframe Storage ArchitectureBuild the Optimal Mainframe Storage Architecture
Build the Optimal Mainframe Storage Architecture
 
Solution Brief HPE StoreOnce backup with Veeam
Solution Brief HPE StoreOnce backup with VeeamSolution Brief HPE StoreOnce backup with Veeam
Solution Brief HPE StoreOnce backup with Veeam
 
Dynamic Hyper-Converged Future Proof Your Data Center
Dynamic Hyper-Converged Future Proof Your Data CenterDynamic Hyper-Converged Future Proof Your Data Center
Dynamic Hyper-Converged Future Proof Your Data Center
 
Resource guide-30th-edition
Resource guide-30th-editionResource guide-30th-edition
Resource guide-30th-edition
 

En vedette (6)

Miriplatin 141977-79-9-api
Miriplatin 141977-79-9-apiMiriplatin 141977-79-9-api
Miriplatin 141977-79-9-api
 
Note de décryptage droits humains et climat
Note de décryptage droits humains et climatNote de décryptage droits humains et climat
Note de décryptage droits humains et climat
 
Communiqué de presse du RC&D "Enfin un accord, mais..."
Communiqué de presse du RC&D "Enfin un accord, mais..."Communiqué de presse du RC&D "Enfin un accord, mais..."
Communiqué de presse du RC&D "Enfin un accord, mais..."
 
RTAGNO Resume
RTAGNO ResumeRTAGNO Resume
RTAGNO Resume
 
Proyecto ajedrez
Proyecto ajedrezProyecto ajedrez
Proyecto ajedrez
 
Rapport de stage
Rapport de stageRapport de stage
Rapport de stage
 

Similaire à 50 Shades of Grey in Software-Defined Storage

Hyper-converged infrastructure
Hyper-converged infrastructureHyper-converged infrastructure
Hyper-converged infrastructure
Igor Malts
 
GigaOm-sector-roadmap-cloud-analytic-databases-2017
GigaOm-sector-roadmap-cloud-analytic-databases-2017GigaOm-sector-roadmap-cloud-analytic-databases-2017
GigaOm-sector-roadmap-cloud-analytic-databases-2017
Jeremy Maranitch
 
Carlson Companies is one of the largest privately held compani.docx
Carlson Companies is one of the largest privately held compani.docxCarlson Companies is one of the largest privately held compani.docx
Carlson Companies is one of the largest privately held compani.docx
wendolynhalbert
 

Similaire à 50 Shades of Grey in Software-Defined Storage (20)

Building a Highly Available Data Infrastructure
Building a Highly Available Data InfrastructureBuilding a Highly Available Data Infrastructure
Building a Highly Available Data Infrastructure
 
Hyper-converged infrastructure
Hyper-converged infrastructureHyper-converged infrastructure
Hyper-converged infrastructure
 
Storage Area Networks Unit 1 Notes
Storage Area Networks Unit 1 NotesStorage Area Networks Unit 1 Notes
Storage Area Networks Unit 1 Notes
 
The Journey Toward the Software-Defined Data Center
The Journey Toward the Software-Defined Data CenterThe Journey Toward the Software-Defined Data Center
The Journey Toward the Software-Defined Data Center
 
vNet SAS Focus Group for NextIO
vNet SAS Focus Group for NextIOvNet SAS Focus Group for NextIO
vNet SAS Focus Group for NextIO
 
White Paper: Rethink Storage: Transform the Data Center with EMC ViPR Softwar...
White Paper: Rethink Storage: Transform the Data Center with EMC ViPR Softwar...White Paper: Rethink Storage: Transform the Data Center with EMC ViPR Softwar...
White Paper: Rethink Storage: Transform the Data Center with EMC ViPR Softwar...
 
Removing Storage Related Barriers to Server and Desktop Virtualization
Removing Storage Related Barriers to Server and Desktop VirtualizationRemoving Storage Related Barriers to Server and Desktop Virtualization
Removing Storage Related Barriers to Server and Desktop Virtualization
 
2015 CTO Predictions
2015 CTO Predictions2015 CTO Predictions
2015 CTO Predictions
 
How Savvy Firms Choose the best Hyperconverged Infrastructure for their Business
How Savvy Firms Choose the best Hyperconverged Infrastructure for their BusinessHow Savvy Firms Choose the best Hyperconverged Infrastructure for their Business
How Savvy Firms Choose the best Hyperconverged Infrastructure for their Business
 
Cloud Infrastructure for Your Data Center
Cloud Infrastructure for Your Data CenterCloud Infrastructure for Your Data Center
Cloud Infrastructure for Your Data Center
 
Software Defined Storage Accelerates Storage Cost Reduction
Software Defined Storage Accelerates Storage Cost ReductionSoftware Defined Storage Accelerates Storage Cost Reduction
Software Defined Storage Accelerates Storage Cost Reduction
 
Connect July-Aug 2014
Connect July-Aug 2014Connect July-Aug 2014
Connect July-Aug 2014
 
Performance,cost and reliability through hybrid cloud storage
Performance,cost and reliability through hybrid cloud storagePerformance,cost and reliability through hybrid cloud storage
Performance,cost and reliability through hybrid cloud storage
 
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
Positioning IBM Flex System 16 Gb Fibre Channel Fabric for Storage-Intensive ...
 
Storage Virtualization isn’t About Storage
Storage Virtualization isn’t About StorageStorage Virtualization isn’t About Storage
Storage Virtualization isn’t About Storage
 
Product Brief Storage Virtualization isn’t About Storage
Product Brief Storage Virtualization isn’t About StorageProduct Brief Storage Virtualization isn’t About Storage
Product Brief Storage Virtualization isn’t About Storage
 
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage SwitzerlandWhy is Virtualization Creating Storage Sprawl? By Storage Switzerland
Why is Virtualization Creating Storage Sprawl? By Storage Switzerland
 
GigaOm-sector-roadmap-cloud-analytic-databases-2017
GigaOm-sector-roadmap-cloud-analytic-databases-2017GigaOm-sector-roadmap-cloud-analytic-databases-2017
GigaOm-sector-roadmap-cloud-analytic-databases-2017
 
Gridstore's Software-Defined-Storage Architecture
Gridstore's Software-Defined-Storage ArchitectureGridstore's Software-Defined-Storage Architecture
Gridstore's Software-Defined-Storage Architecture
 
Carlson Companies is one of the largest privately held compani.docx
Carlson Companies is one of the largest privately held compani.docxCarlson Companies is one of the largest privately held compani.docx
Carlson Companies is one of the largest privately held compani.docx
 

Dernier

Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
Joaquim Jorge
 

Dernier (20)

Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 

50 Shades of Grey in Software-Defined Storage

  • 1. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE: CHOOSING THE STORAGE INFRASTRUCTURE THAT BEST FITS YOUR BUSINESS NEEDS BY JON TOIGO CHAIRMAN THE DATA MANAGEMENT INSTITUTE SUMMARY Software-Defined Storage (SDS) has become a meme in industry and trade press discussions of storage technology lately, though the term itself lacks rigorous technical definition. Essentially, SDS is touted as a model for building storage that will work better with virtualized workloads running under server hypervisor technology than do “legacy” NAS and SAN infrastructure. Regardless of the veracity of these claims, the business-savvy IT planner should base his or her choice of storage infrastructure not on trendy memes, but on traditional selection criteria: cost, availability, and simplicity.
  • 2. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 2 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. INTRODUCTION IT planners have been hearing a lot about software-defined storage (SDS) lately -- and how it is supposed to replace all of the “legacy storage” that companies have deployed for the past decade or so, including Storage Area Networks (SANs) and Network Attached Storage (NAS) appliances. Unfortunately, many of the arguments advanced by evangelists for adopting SDS are only partially grounded in fact or have nothing whatsoever to do with improvements in storage performance, allocation or utilization efficiency. The purpose of this paper is to quickly review the arguments for SDS, to provide the proper context for analysis of the technologies, and to help guide the business-savvy IT manager or planner to basic common-sense criteria for evaluating the various products available to implement the SDS model if it is deemed suitable. In the final analysis, the kind of storage infrastructure that is appropriate for a data center, branch office or small business equipment room should be determined by what the application workload requires, first and foremost.
  • 3. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 3 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. WHY SDS? Software-defined storage is touted by evangelists as the fix for what ails traditional or legacy storage – as the next step in storage evolution. It is unclear what that means beyond the obvious marketing appeal of flowery rhetoric. If there is any sort of evolution in storage, it has been the movement away from isolated “islands” of storage that created obstacles to accessing and sharing stored data, and toward storage infrastructure models that enabled greater data sharing. In the 1960’s and 70’s, the predominant forms of storage were internal storage (storage devices mounted inside the server frame itself) and direct-attached storage (a crate of disk drives connected to the outside of the server cabinet via an external extension of the server mainboard bus). From those implementation models, storage technology expanded in two directions: one supporting the storage and sharing of file-based data across a network; the other supporting the storage and sharing of block data from database and transaction-oriented systems. Block data storage “evolved” from island architectures (internal and directly attached storage arrays that were accessed only through the server itself, making stored data difficult to access, share or scale efficiently) to shareable arrays – that is, arrays with multiple ports for connecting many server hosts. Scaling these systems, however, became problematic, so engineers figured out ways to attach storage arrays into a switched fabric (a sort of rudimentary network) using a serial SCSI protocol such as Fibre Channel or SCSI over IP (iSCSI). The ultimate goal was to enable the storage infrastructure to be shared by all applications by creating a true “storage area network” or SAN.
  • 4. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 4 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. However, the holy grail of an open and heterogeneous SAN never appeared in the market. Vendors preferred to keep their storage equipment sales profitable by adding functionality to the proprietary controllers installed in their external arrays, and never fully participated in any sort of scheme to enable different equipment from different vendors to be managed in concert. Absent common management, the fabric SAN was incredibly pricey to own because of all the proprietary value-add software on the proprietary controllers, and to operate because of the lack of a universal management scheme and the need to embed expensive storage experts in IT staff. The above situation already existed when server virtualization took hold in the early 2000s. This made legacy storage an easy target for server hypervisor vendors to blame for application performance issues that arose following the consolidation of application workloads on fewer server platforms. Truth be told, whatever inefficiencies that already existed in SANs and NAS, server virtualization workloads just made them worse. Consolidating applications onto fewer servers dramatically changed the amount of I/O emanating from a single server, requiring on average the addition of 7 to 16 I/O connections per server. Hypervisor computing also changed the traffic patterns on networks and fabric interconnects and switching systems. Hypervisor vendors assured customers that many good things would accrue to this model. For one, virtual machines could be replicated from host to host and joined into highly available clusters. Virtual machines could also move around server to server to distribute load more efficiently and to provide high availability.
  • 5. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 5 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. Only, consolidating a lot of virtual machines into a single server host, then migrating workload at will from one physical server to another, created challenges in terms of maintaining connections between applications and their stored data. When an app moved locations, often intervention was required by an administrator to provide the application with a new route to access its storage shares from a new server perch. Moreover, we were introduced to a new problem that resulted from combining the I/O from multiple workloads in the same server – something that the industry has started to refer to as the “I/O blender effect.” In operation, every hosted application will simply write its data out at random into a shared I/O path. Without special technology to intervene, this randomized I/O will eventually clog memory and disk storage devices and slow applications down to a crawl. With application performance issues becoming the number one complaint of hypervisor computing customers, hypervisor vendors pushed several explanations that blamed “legacy storage” inefficiencies and began arguing for the “ripping and replacing” of older storage infrastructure and its replacement with something new, which they came to call software- defined storage.
  • 6. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 6 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. HYPERVISOR VENDOR ARGUMENTS OFTEN FLAWED, BUT SDS IS VIABLE NONETHELESS Surveying the literature, hypervisor vendors have developed five reasons to de-provision legacy storage. Some are valid arguments while others are based on half-truths or downright falsehoods. 1. “Slow application performance due to ‘legacy SAN (and NAS) infrastructure’” – This is often a falsehood. You can see whether the storage I/O path is causing application delays rather simply by checking storage I/O queue depths. In many cases, queue depths are not significant (meaning that there is no data waiting in line to be written to disk). When this queue depth reading is combined with a processor cycle or processor activity measurement showing high rates of processor cycling, a more logical conclusion is that a log jam exists in “raw I/O” – the communications path between CPU and memory where application processing occurs. If that is the case, the application slowdown is a function of either application code or hypervisor code, not storage at all. 2. “High cost of storage (OPEX & CAPEX) due to proprietary storage” – Legacy storage was and is without a doubt expensive. Hardware vendors argue that value-add software functionality added to array controllers are differentiators of their products and represent much R&D that needs to be rewarded. Unfortunately, value-add often adds configuration complexity and expensive administrative staff skills requirements, and sometimes interferes with unified management of the kit as part of infrastructure. 3. “Poor utilization efficiency and management difficulty due to lack of common management” – Per #2, proprietary value-add storage is difficult to manage in common, and this has long been a reason to argue for infrastructure that might deliver better management. 4. “Lack of agility due to legacy infrastructure” – Allocating and de-allocating storage resources and services is much more difficult with proprietary and complicated value- add storage. However, this has nothing whatsoever to do with application performance. 5. “DAS storage outperforms SAN” – This is not true and makes no sense since both NAS and SAN are also direct attached storage configurations. NAS is storage directly-attached to a thin server that provides network-based access to data. A SAN is direct-attached storage with a physical or fabric-layer switch that makes and breaks server/storage connections at high speed, giving the illusion of network attachment.
  • 7. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 7 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. Valid or not, the arguments of hypervisor vendors regarding need to abandon legacy storage have resonated. In its place, vendors are recommending something called software-defined storage – which is, once again, storage that is physically, directly attached to virtual server hosts. The only difference between this and direct-attached storage in the past is that value-add storage services (software) are not located on the controller of the external array, but are instead instantiated in a software layer in the server, usually as part of the hypervisor vendor’s software stack. Vendors can’t agree, of course, on a single definition for software defined storage. Instead, it is simply a panacea – a cure-all for everything that is wrong with “legacy storage” for use in a software-defined data center. The key difference is where the intelligence is located: is it on the array controller or is it somewhere in the server hypervisor software stack? Actually, there are two sets of discriminators to consider when looking at the many SDS products from hypervisor vendors and independent software developers that have appeared in the market of late.
  • 8. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 8 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. Some SDS solutions are hypervisor-dedicated or hypervisor-dependent. Certainly, this is the case with VMware’s Virtual SAN, which only works with the company’s proprietary hypervisor. To a lesser extent, Microsoft’s Clustered Storage Spaces is proprietary to Microsoft, though they do claim to be able to share their storage with VMware (you just convert your VMware workload into Microsoft VHD format and import it into Hyper-V so you can share the Microsoft SDS infrastructure!). The other end of the spectrum are the many third party SDS offerings, including StorMagic’s SvSAN technology, that are hypervisor-agnostic. Vendors of these solutions are earnestly trying to make their software-defined storage infrastructure work with multiple hypervisors and, in some cases, with non-virtualized workload as well. Some implement a common storage environment into which the data from different hypervisor workloads can be written, while others enable common management of storage infrastructures created with their software but dedicated to specific hypervisor workloads and their data. Either way, they are more robust than hypervisor-dedicated solutions and offer a better fit for companies that are deploying more than one hypervisor or that may do so in the future. The other differentiator to consider is hardware dependency and hardware agnosticism. Despite the ideology of software-defined (“use any hardware you want”), the truth is that some software-defined solutions have fairly rigid requirements when it comes to hardware components and topology. Some of the latest “hyper-converged infrastructure appliances” for example are just another kind of proprietary storage array. Logically, these products would be placed on the hardware-dependent side of the SDS solution spectrum. Hypervisor-dependent
  • 9. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 9 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. SDS products like VMware’s EVO:RAIL are also hardware dependent, and even their VMware Virtual SAN carries with it some fairly exacting and expensive hardware and topology requirements, including a minimum of three storage nodes. In the middle of this spectrum are a number of third party virtual SAN SDS solutions, which unlike their hypervisor software cousins, do not require special configurations or limit what kind of disk or what kind of flash devices you can use. StorMagic, for example, provides a solution that can embrace a wide range of hardware components. On the hardware-agnostic end of the spectrum are storage virtualization software products. To SDS purists, storage virtualization is not software-defined storage because SDS seeks to decouple storage capacity from storage value add services functionality and performance; capacity virtualization is not part of their SDS model. In general, if you are considering an SDS solution, your storage infrastructure requirements are usually best served by pursuing hardware- and hypervisor-agnostic SDS solutions. That decision is more likely to enable you to customize your storage gear selections to your needs and budget and to consolidate the management of different storage infrastructures supporting different workloads. If you are operating a large shop, such as a managed hosting environment or a large enterprise data center, moving to a Virtual SAN may actually be a step backward on the evolutionary chart. A real SAN, assuming that one is ever delivered to market, will deliver all of the capabilities and cost metrics that you need for such environments. On the other hand, if yours is a small to medium sized business or if you are tasked to provide storage infrastructure for remote or branch office operations, big centralized storage may not be your best or most strategic infrastructure choice. You may be attracted to VMware’s virtual SAN or Microsoft’s all-in-one Cluster Storage Spaces SDS solution as a template for building your storage. That is, until you find out what it will cost. The fact is that VMware’s Virtual SAN only works with VMware vSphere workloads and requires you to limit your hardware to a specific list of supported equipment. The same is true of Microsoft’s Clustered Storage Spaces model, which is heavily dependent on SAS storage. Such equipment requirements may not be a deal breaker for firms that want a simple single vendor solution, but they do limit options. Planners may not be able to take advantage of newer storage technologies or leverage existing investments in storage equipment that they have today.
  • 10. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 10 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. More to the point, the cost of implementing a hypervisor dependent storage stack can be prohibitive. A recent lab evaluation published in a popular computing publication placed the cost of a basic Virtual SAN implementation, which requires three nodes minimum, at $11 to 14K in software licenses per node and about that much again for the hardware that is approved by VMware for use in their infrastructure. The prices for Microsoft are significantly lower, but there is still a minimum 3 node requirement and a preference for PCIe flash and SAS disk drives rather than SATA. One significant benefit of using a third party Software Defined Storage solution is that it can insulate the planner from hardware lock-ins and usually does not require a three-node minimum hardware cluster. SDS can be implemented initially as a much less expensive two node cluster. For the smaller shop, or the ROBO environment, the independent software vendor’s SDS solution might be just the thing to test the value of software defined storage to application performance.
  • 11. 50 SHADES OF GREY IN SOFTWARE-DEFINED STORAGE 11 © COPYRIGHT 2015 BY THE DATA MANAGEMENT INSTITUTE, LLC. ALL RIGHTS RESERVED. CONCLUSION The fit for software-defined storage, whether from your hypervisor vendor or from an independent developer like StorMagic, should be a function of its fit with deployed (and likely to be deployed) applications and their requirements, with strategic goals for infrastructure and data management, and with budget, staff skills and other practical boundary conditions. It is a good idea to determine what your needs are before you seek out an SDS solution. It is worth noting that, in the absence of uniform standards or guarantees of interoperability between different software stacks, IT planners may be looking at a need to deploy and manage different SDS solutions to meet different requirements. A product aimed at large enterprise shops and hosting environments may not be well suited to deployment in a ROBO environment, for example. It is a good idea to try the SDS solution before you buy it. StorMagic offers a 60 day trial version of its product, StorMagic SvSAN, that you can access from the company website (http://www.stormagic.com/60-day-free-trial/).