According to a senior technical manager at chipmaker AMD, the new dual core Intel Xeon 5100 server processors are not all they're cracked up to be because of a major design drawback. The problem, according to the AMD technical guru, is that the so-called dual core processors share a single memory cache unlike the dual core AMD Opteron processor, which has separate dedicated cache for each core.
Michael Apthorpe, senior technical manager for AMD in Australia and New Zealand, an electronics engineer with more than 25 years experience in chip design starting from his days at Mitsubishi Electric, says that any dual core processor design that involves both cores sharing a single memory cache has both performance and power consumption disadvantages.
"There are two types of cache with processors and they are known as exclusive and inclusive cache," says Apthorpe. "With inclusive cache, you have one allocation of RAM. What happens is that if you're running a program and all of sudden you change programs, you have to stop and flush the entire cache and reallocate and reload it. That takes clock cycles to do and that's our competitor's product."
The difference with AMD's Opteron dual core processor range, according to Apthorpe, is that each core on the processor has its own dedicated cache. "That means with our processors, they never have to stop and ask the cache to flush and wait for it to reload," he says. "Therefore, the processor always has full access to the cache which really speeds things up.
"When you go to multiple processors, the problem gets even more pronounced. If you have just one cache and one processor is dominating the cache utilisation and then you get a request on the other processor, the same thing happens. That is it has to stop, flush, reallocate and reload. This all takes clock pulses so again it hamstrings the performance of the system. That's why our competitor has to put so much more cache in their systems because they have to make up for the latency. They port information out of their main RAM into the cache of the processor and they try to make up the latency which they create by stopping, flushing and reloading the processor cache."
Is this the reason why the Woodcrest/Coroe are prone to cache trashing?
Is this the main reason why AMD is going for a large and shared L3 cache?
http://www.itwire.com.au/content/view/4785/53/