IDC analysts said in a new report that the practice of putting two or more processors on a single chip will lead to new revenue opportunities for vendors who can sell the concept without rubbing customers the wrong way.
Multi-core processing has distinct advantages over its single-core ancestors. Putting more processors, or cores, on one chip yields certain performance benefits (though not double the power of one chip, as some might assume) without consuming the power associated with two distinct processors.
Matt Eastwood, vice president of IDC's server research, said in his report that the emergence of multi-core ''may be one of the most significant industry developments of the past 40 years.'' He believes the impact on the IT infrastructure will accelerate with each generation of multi-core processing technology, causing a big market shift by 2009.
''Users expect significant price/performance improvements from these systems and anticipate that multi-core will further accelerate the adoption of server virtualization technologies,'' Eastwood said. ''Technology suppliers that prepare for this transition will be rewarded with increased revenue opportunities over the next few years.''
Computing experts expect multi-core processing to drive significant cost savings and performance improvements for customers using modular blade systems and virtualization software, for which pricing has always been tricky.
Detractors have said a combination of dual-core pricing and virtualization could be a tricky proposition in terms of accurate licensing rates. Virtualization allows part of a processor to be dedicated to one task, which can cloud the pricing issue.
The question is: If each chip boasts two sockets, how would this affect the way vendors charge customers? This is the same question that plagues the prospect of utility computing scenarios, which allows users to pay according to the computing power they use. This approach gives users greater control over how they spend money on resources such as CPUs.