By Gavin McLaughlin, VP Worldwide Marketing, X-IO Technologies
Confronted with ever-increasing amounts of data, IT departments have never been more challenged to find greater capacity and higher resilience in less time and under budget. IT managers are battling to get the most out of their hardware at a time when budget constraints are hitting almost every IT department.
However, there are ways they can improve and boost the performance of their existing storage hardware by using the right tool for the job. There are four key methods businesses can use, and each has its strengths and weaknesses.
Most traditional storage vendors offer the ability to add Solid State Disks (SSDs) to existing Storage Area Network (SAN) storage. However, while this is technically possible, it comes with limitations on the number of drives that can be deployed due to possible back-end bandwidth limitations. It’s also important to stress that traditional storage arrays have been designed with one challenge in mind: getting the most from Hard Disk Drives (HDDs). Adding SSDs into an existing architecture can often cause serious damage to the whole system.
While this option presents itself as a simple and relatively low-risk approach to address bottle-necks, due to its use of existing storage management practices, it can prove a costly way of addressing the issue and restrict future growth due to the technical challenges involved.
Opening servers and adding flash cards, such as PCI Express Cards, to the storage design is often thought to be the simplest way to address performance challenges. This approach can be useful for single-application, single-server environments, but adding flash cards to servers is not well suited to enterprise solutions. In fact, adding flash cards to the server can significantly limit the growth and functionality of any end-to-end virtualized solution, due to the constraints found in virtualized environments where tools such as vMotion cannot be deployed as a result of captive storage.
Like option 1, adding flash cards to servers significantly hinders expansion. If growth is required in the long or short term, storage and/or application administrators will need to manually move data be-tween tier 0 and tier 1 and/or utilize a costly data movement tool. This solution can restrict direct data and solution growth, but is perceived to be low risk because only the host configuration is being modified, not the enterprise storage architecture itself.
Adding flash cards to servers is also a high-cost approach because of the cost (per/GB) of the cards themselves. Given the budget constraints they are operating under, IT departments may want to think twice about adopting this approach.
In the last year or so, we have seen a plethora of all-flash arrays (AFAs) come into the storage market. While AFAs can provide extremely high input/output (I/O) rates, they are really designed for one job — high I/O, low latency data delivery. In some cases, such as low latency trading, the cost outlay (often in excess of $100k/TB) can be justified. However, in most cases, projected costs of $100k/TB simply minimize and distort the cost justification for standard enterprise class projects.
Utilizing all-flash arrays is probably the most unbalanced approach when it comes to building on existing hardware. Adding all-flash to existing systems requires extremely high capital and operating expenditure, yet it doesn’t deliver the flexibility you might expect. With IT budgets always being squeezed, businesses need to look at their long-term development plans before jumping into the deep-end with a solution that does not necessarily support future growth.
By combining flash and HDDs into a single pool, the most appropriate tool can be utilized for the data in hand, be it cache memory or solid state. However, even within hybrid storage architectures, there are different approaches to incorporating flash and controlling data access patterns. The key element here is the ability to move data between different tiers, in real time, without any manual intervention. This should be based on I/O activity rather than workload predictions based on file activity.
In essence, hybrid storage with real-time tiering is the only methodology to address the storage architecture challenges presented by new technologies, providing a true balance of cost, growth and risk.
Whether implementing a whole system refresh or adding to an existing system, the trick is to ignore the hype, look at the business needs and make a decision based on cost, risk and anticipated growth.
Photo courtesy of Shutterstock.
Ethics and Artificial Intelligence: Driving Greater Equality
FEATURE | By James Maguire,
December 16, 2020
AI vs. Machine Learning vs. Deep Learning
FEATURE | By Cynthia Harvey,
December 11, 2020
Huawei’s AI Update: Things Are Moving Faster Than We Think
FEATURE | By Rob Enderle,
December 04, 2020
Keeping Machine Learning Algorithms Honest in the ‘Ethics-First’ Era
ARTIFICIAL INTELLIGENCE | By Guest Author,
November 18, 2020
Key Trends in Chatbots and RPA
FEATURE | By Guest Author,
November 10, 2020
FEATURE | By Samuel Greengard,
November 05, 2020
ARTIFICIAL INTELLIGENCE | By Guest Author,
November 02, 2020
How Intel’s Work With Autonomous Cars Could Redefine General Purpose AI
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
October 29, 2020
Dell Technologies World: Weaving Together Human And Machine Interaction For AI And Robotics
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
October 23, 2020
The Super Moderator, or How IBM Project Debater Could Save Social Media
FEATURE | By Rob Enderle,
October 16, 2020
FEATURE | By Cynthia Harvey,
October 07, 2020
ARTIFICIAL INTELLIGENCE | By Guest Author,
October 05, 2020
CIOs Discuss the Promise of AI and Data Science
FEATURE | By Guest Author,
September 25, 2020
Microsoft Is Building An AI Product That Could Predict The Future
FEATURE | By Rob Enderle,
September 25, 2020
Top 10 Machine Learning Companies 2021
FEATURE | By Cynthia Harvey,
September 22, 2020
NVIDIA and ARM: Massively Changing The AI Landscape
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
September 18, 2020
Continuous Intelligence: Expert Discussion [Video and Podcast]
ARTIFICIAL INTELLIGENCE | By James Maguire,
September 14, 2020
Artificial Intelligence: Governance and Ethics [Video]
ARTIFICIAL INTELLIGENCE | By James Maguire,
September 13, 2020
IBM Watson At The US Open: Showcasing The Power Of A Mature Enterprise-Class AI
FEATURE | By Rob Enderle,
September 11, 2020
Artificial Intelligence: Perception vs. Reality
FEATURE | By James Maguire,
September 09, 2020
Datamation is the leading industry resource for B2B data professionals and technology buyers. Datamation's focus is on providing insight into the latest trends and innovation in AI, data security, big data, and more, along with in-depth product recommendations and comparisons. More than 1.7M users gain insight and guidance from Datamation every year.
Advertise with TechnologyAdvice on Datamation and our other data and technology-focused platforms.
Advertise with Us
Property of TechnologyAdvice.
© 2025 TechnologyAdvice. All Rights Reserved
Advertiser Disclosure: Some of the products that appear on this
site are from companies from which TechnologyAdvice receives
compensation. This compensation may impact how and where products
appear on this site including, for example, the order in which
they appear. TechnologyAdvice does not include all companies
or all types of products available in the marketplace.