SHARE
Facebook X Pinterest WhatsApp

IBM To Help Build Massive Data Grid To Recreate ‘Big Bang’

IBM is joining the CERN openlab for DataGrid applications to help create a massive data-management system built on Grid computing. IBM’s storage virtualization and file management technology will play a key role in the collaboration, which aims to create a data file system “far larger than exists today” to help scientists at the renowned particle […]

Written By
thumbnail Paul Shread
Paul Shread
Apr 2, 2003
Datamation content and product recommendations are editorially independent. We may make money when you click on links to our partners. Learn More

IBM is joining the CERN openlab for DataGrid applications to help create a massive data-management system built on Grid computing.

IBM’s storage virtualization and file management technology will play a key role in the collaboration, which aims to create a data file system “far larger than exists today” to help scientists at the renowned particle physics research center understand some of the most fundamental questions about the nature of matter and the universe, according to IBM and CERN, the European Organization for Nuclear Research.

Conceived in IBM Research, the storage virtualization technology, known as Storage Tank, is designed to provide scalable, high-performance and highly available management of huge amounts of data using a single file namespace regardless of where or on what operating system the data reside.

IBM and CERN will work together to extend Storage Tank’s capabilities so it can manage and provide access from any location worldwide to the unprecedented amounts of data – billions of gigabytes a year – that CERN’s Large Hadron Collider (LHC) is expected to produce when it goes online in 2007. The LHC is the next-generation particle accelerator designed to recreate the conditions shortly after the Big Bang to help researchers understand the initial seconds when the universe was formed.

The CERN community – which is credited with inventing the World Wide Web in 1990 – hopes to push the Internet even further with Grid computing and the massive data processing requirements for the LHC. CERN openlab is a collaboration between CERN and leading industrial partners that will create and implement data-intensive Grid-computing technologies to aid LHC scientists. Because the same issues facing CERN are becoming increasingly important to the IT industry, the CERN openlab and its partners – which include Enterasys, HP and Intel – are working together to explore advanced computing and data management solutions.

By 2005, the CERN openlab collaboration with IBM is expected to be able to handle up to a petabyte (a million gigabytes) of data.

“We are delighted that IBM is joining the CERN openlab for DataGrid applications,” said Wolfgang von Ruden, Information Technology Division leader and head of the CERN openlab. “Together with IBM, we aim to achieve a one petabyte storage solution and integrate it with the Grid that CERN is building to handle the extreme data challenges of the LHC project.”

“CERN’s scientists and colleagues want to be able to get to their data wherever it may be – local or remote and regardless of which operating system on which it may reside,” said Jai Menon, a fellow at IBM’s Almaden Research Center in San Jose and co-director of IBM’s Storage Systems Institute joint program between IBM Research and the company’s product division. “This is the perfect environment for us to enhance Storage Tank to meet the demanding requirements of large-scale Grid computing systems.”

As part of the agreement, several leading storage management experts from IBM’s Almaden and Haifa (Israel) Research Labs will work with the CERN openlab team. IBM will also give CERN the system’s initial 20 terabytes of high-performance disk storage and six Linux-based servers, and IBM Switzerland will provide additional support.

Storage Tank employs policy-based storage management and includes clustering and specialized protocols that detect network failures to enable very high reliability and availability.

In this initiative, IBM is following a collaboration strategy initiated in 2001 with the European Union-sponsored European Data Grid project, which is also led by CERN.

  SEE ALL
ARTICLES
 
thumbnail Paul Shread

Paul Shread has covered nearly every aspect of enterprise technology in his 20+ years in IT journalism, including an award-winning series on software-defined data centers. He wrote a column on small business technology for Time.com, and covered financial markets for 10 years, from the dot-com boom and bust to the 2007-2009 financial crisis. He holds a market analyst certification.

Recommended for you...

10 Top Storage Certifications for Data Professionals In 2024
Anina Ot
Mar 27, 2024
What is Data Compression & How Does it Work?
Kashyap Vyas
Sep 19, 2023
What Is a Data Pipeline? Examples & Types
Anina Ot
Sep 18, 2023
Top 10 Data Center Certifications
Kashyap Vyas
Aug 22, 2023
Datamation Logo

Datamation is the leading industry resource for B2B data professionals and technology buyers. Datamation's focus is on providing insight into the latest trends and innovation in AI, data security, big data, and more, along with in-depth product recommendations and comparisons. More than 1.7M users gain insight and guidance from Datamation every year.

Property of TechnologyAdvice. © 2025 TechnologyAdvice. All Rights Reserved

Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.