According to International Data Corp. (IDC) of Framingham, MA, organizations have been battling year-over-year information growth rates of 50 to 60 percent. This trend is expected to continue well into the next decade. Yet these same organizations are being asked to do more with less in terms of budgets and personnel.
Just adding more storage is no longer an effective option. So what is?
“Storing more effectively allows IT organizations to consolidate multiple application tiers into a single common infrastructure to optimize resources, lower costs, and improve service levels,” said Scott Delandy, senior product manager of EMC Corp. of Hopkinton, MA. “A key benefit is reduced complexity. At the same time, IT organizations can simplify storage management across the enterprise by leveraging automation tools and advanced capabilities, as well as improve energy efficiency by reducing the number of storage systems and networks.”
According to EMC, developing and implementing a tiered storage strategy is the right starting point for managing information growth. In many organizations, after all, most growth is absorbed by tier-one storage – the most expensive tier. By identifying inactive and static data and moving it from tier-one to lower-cost tiers, tier-one resources are freed up while reducing costs.
“Deploying tiered storage allows an organization to store data on the right storage type at the right time for the right cost point,” said Delandy. “This aligns storage capabilities to application and business requirements and ultimately lowers total cost of ownership while optimizing service levels.”
But storage tiering can’t be done haphazardly. It doesn’t work for IT to apply some arbitrary classification based on time or capacity. That leads to such situations as the quarterly or annual reports being delayed as the data needed for them has been archived by an over-enthusiastic storage admin who neglected to survey business units about access patterns.
“The first step to implementing a tiered storage strategy is to classify applications based on service levels and business impact,” said Delandy. “Key application criteria to consider when classifying application tiers include performance needs, availability requirements, recovery point objectives and recovery time objectives, as well as cost points across storage tiers.”
The number of tiers deployed, however, is dependent on the environment. Some may require fewer tiers, others more. That lets you can create different service-level categories for primary storage, archival storage, backup and recovery, business continuity, and so forth.
Depending on the environment, you can also factor in the costs associated with each category. That permits the organization to better understand the cost of information for each application and align the information infrastructure with the associated business requirements.
EMC lays out the basic steps as follows:
The first step to storing more effectively is to classify data and applications based on business requirements. In assessing this on a per application basis, companies typically find that a “one size fits all” approach to storing data doesn’t make sense.
Extract static or infrequently accessed data from the primary storage environment and move it to an active archive. Such data can be identified during the classification process in the previous step. A key criterion of an active archive is that when the information is needed, it is online and readily accessible to support daily requests or longer-term audit cycles. Hence, the term “active archive” – data is archived, yet remains online and can be accessed relatively quickly via a cost-effective storage solution.
One of the key drivers impacting storage costs is the amount of data that needs to be stored and backed up. Reducing or eliminating duplicate data is a key step to storing more effectively.
By extending the tiered storage methodology to backup/recovery operations, organizations can improve the speed and reliability of their backup and recovery processes. High-capacity disk-based technologies, for example, are helping users reduce backup windows by taking the pressure off tapes.
Creating full copies of production data has been a common practice to assist with backup, recovery, and testing of application environments. As information continues to grow at a fast rate, the storage requirements and costs associated with this have escalated as well. The use of snapshots provides an affordable alternative for protecting and re-purposing production data via local replication.
In a typical environment, many application and/or file servers run at utilization rates as low as 15 percent.
“By deploying server and file virtualization techniques, customers can drastically improve utilization by consolidating physical servers – in many cases at the rate of 20:1,” said Delandy. “The overall effects are lower capital costs, simplified management, and increased utilization of the storage infrastructure.”
This article was first published on EnterpriseITPlanet.com.
Ethics and Artificial Intelligence: Driving Greater Equality
FEATURE | By James Maguire,
December 16, 2020
AI vs. Machine Learning vs. Deep Learning
FEATURE | By Cynthia Harvey,
December 11, 2020
Huawei’s AI Update: Things Are Moving Faster Than We Think
FEATURE | By Rob Enderle,
December 04, 2020
Keeping Machine Learning Algorithms Honest in the ‘Ethics-First’ Era
ARTIFICIAL INTELLIGENCE | By Guest Author,
November 18, 2020
Key Trends in Chatbots and RPA
FEATURE | By Guest Author,
November 10, 2020
FEATURE | By Samuel Greengard,
November 05, 2020
ARTIFICIAL INTELLIGENCE | By Guest Author,
November 02, 2020
How Intel’s Work With Autonomous Cars Could Redefine General Purpose AI
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
October 29, 2020
Dell Technologies World: Weaving Together Human And Machine Interaction For AI And Robotics
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
October 23, 2020
The Super Moderator, or How IBM Project Debater Could Save Social Media
FEATURE | By Rob Enderle,
October 16, 2020
FEATURE | By Cynthia Harvey,
October 07, 2020
ARTIFICIAL INTELLIGENCE | By Guest Author,
October 05, 2020
CIOs Discuss the Promise of AI and Data Science
FEATURE | By Guest Author,
September 25, 2020
Microsoft Is Building An AI Product That Could Predict The Future
FEATURE | By Rob Enderle,
September 25, 2020
Top 10 Machine Learning Companies 2021
FEATURE | By Cynthia Harvey,
September 22, 2020
NVIDIA and ARM: Massively Changing The AI Landscape
ARTIFICIAL INTELLIGENCE | By Rob Enderle,
September 18, 2020
Continuous Intelligence: Expert Discussion [Video and Podcast]
ARTIFICIAL INTELLIGENCE | By James Maguire,
September 14, 2020
Artificial Intelligence: Governance and Ethics [Video]
ARTIFICIAL INTELLIGENCE | By James Maguire,
September 13, 2020
IBM Watson At The US Open: Showcasing The Power Of A Mature Enterprise-Class AI
FEATURE | By Rob Enderle,
September 11, 2020
Artificial Intelligence: Perception vs. Reality
FEATURE | By James Maguire,
September 09, 2020
Datamation is the leading industry resource for B2B data professionals and technology buyers. Datamation's focus is on providing insight into the latest trends and innovation in AI, data security, big data, and more, along with in-depth product recommendations and comparisons. More than 1.7M users gain insight and guidance from Datamation every year.
Advertise with TechnologyAdvice on Datamation and our other data and technology-focused platforms.
Advertise with Us
Property of TechnologyAdvice.
© 2025 TechnologyAdvice. All Rights Reserved
Advertiser Disclosure: Some of the products that appear on this
site are from companies from which TechnologyAdvice receives
compensation. This compensation may impact how and where products
appear on this site including, for example, the order in which
they appear. TechnologyAdvice does not include all companies
or all types of products available in the marketplace.