"There is a place where high speed, versus low latency, will be an actual advantage (as opposed to high 'synthetic' benchmark tests that really don't relate at all to the real world - always reminds me of a manufacturer of high-performance heads for cars who reminds us: "Remember: We don’t race flow benches!") - any operations that require large, sustained, reads from and writes to RAM - like, as I mentioned, video transcoding... I always consider my 'pass/fail' system stress test to be: watch/pause one HDTV stream off a networked ATSC tuner, while recording a second stream off a PCI NTSC tuner, while transcoding and 'de-commercialing' a third stream to an NAS media server... But, for the vast majority of people, for the vast majority of use, this is not the case. What's going on behind the scenes: the task scheduler is scurrying around, busier than a centipede learning to tap-dance, counting 'ticks': ...tick... yo - over there, you gotta finish up, your tick is over, push your environment, that's a good fella; oops - cache snoops says we've got an incoherency - grab me a meg for him from over there; ...tick... you - get me the address of the block being used by {F92BFB9B-59E9-4B65-8AA3-D004C26BA193}, will 'ya; yeah - UAC says he has permission - I dunno - we'll just have to trust him; damnit - everybody listen up, we've got a pending interrupt request, everyone drop what you're doing, and you - over there - query interrupt handler for a vector - this is important!!! ...tick.... This is why (aside from the obvious matter of access architecture) that swap files are optimized in 4k 'chunks'..."