Three Xeon E5 Server Systems From Intel, Tyan, And Supermicro

Intel R2208GZ4GC: Layout And Overview

Third-up is Intel's R2208GZ4GC "Grizzly Pass" kit, which the company touts as being highly customizable. To that end, we see that the front of the chassis is partitioned off into three sections to facilitate different drive options. The kit came with both 10 GbE and LSI RAID controller options, which I removed for testing and price comparisons. Our test mule has eight 2.5" hot-swappable bays, but the chassis does allow for a second block of hot-swappable bays if you want to configure it for 16.

Like Supermicro and Tyan, Intel makes room for a slim optical drive. Most 2U enclosures designed with high density in mind sacrifice this space (along with front-panel I/O) to cram as many as 24 drives into the front of the chassis. All three of the samples sent to us include two front-panel USB ports and a slim optical bay, though. Whereas the Supermicro chassis also features a front-accessible DE-9 serial connector, Intel's solution includes an HD-15 VGA connector. Intel's front-panel connections are tailored for KVM.

It's pretty apparent that Intel is using a heavily customized implementation, and its PCB is absolutely huge. Officially, the S2600GZ4 motherboard is a proprietary 16.5 x 16.5" form factor. Just to give you an idea of this board's size, a typical quad-socket AMD G34-based platform with 32 DIMM slots fits in a 16.5" x 13" form factor. The benefit of such a large PCB is that you can see Intel's enclosure is built to exploit the additional size.

There are five 80 mm fans in the middle of the chassis that blow air through thermally-sensitive components. A clear plastic duct guides air through passive CPU heat sinks. Whereas Supermicro's shroud channeled air all the way through to the back of its enclosure, Intel employs a shorter duct, since its chassis uses much of the space around back for PCIe rises and power supplies.

Two 80 PLUS Platinum-rated 750 W power supplies connect directly to the motherboard. This is a significant difference between Intel's implementation and the competition from Tyan and Supermicro, which utilize an intermediate distribution board for power. 

Both redundant power supplies are removed by pushing on the teal lever and pulling the handle back. Intel uses a side-by-side configuration and does exhaust air from above the PSUs.

Create a new thread in the US Reviews comments forum about this subject
This thread is closed for comments
20 comments
Comment from the forums
    Your comment
  • mayankleoboy1
    the charts are looking strange. they need to be reduced in size a bit....
    6
  • EzioAs
    Anonymous said:
    the charts are looking strange. they need to be reduced in size a bit....


    I agree. Just reduce it a little bit but don't make it too hard to see
    0
  • dogman_1234
    Cool. Now, can we compare these to Opteron systems?
    0
  • TheBigTroll
    no comparison needed. intel usually wins
    4
  • willard
    TheBigTrollno comparison needed. intel usually wins

    Usually? The E5s absolutely crush AMD's best offerings. AMD's top of the line server chips are about equal in performance to Intel's last generation of chips, which are now more than two years old. It's even more lopsided than Sandy Bridge vs. Bulldozer.
    4
  • Malovane
    dogman_1234Cool. Now, can we compare these to Opteron systems?


    As an AMD fan, I wish we could. But while Magny-Cours was competitive with the last gen Xeons, AMD doesn't really have anything that stacks up against the E5. In pretty much every workload, E5 dominates the 62xx or the 61xx series by 30-50%. The E5 is even price competitive at this point.

    We'll just have to see how Piledriver does.
    2
  • jaquith
    Hmm...in comparison my vote is the Dell PowerEdge R720 http://www.dell.com/us/business/p/poweredge-r720/pd?oc=bectj3&model_id=poweredge-r720 it's better across the board i.e. no comparison. None of this 'testing' is applicable to these servers.
    0
  • lilcinw
    Finally we have some F@H benches!! Thank you!

    Having said that I would suggest you include expected PPD for the given TPF since that is what folders look at when deciding on hardware. Or you could just devote 48 hours from each machine to generate actual results for F@H and donate those points to your F@H team (yes Tom's has a team [40051] and visibility is our biggest problem).
    3
  • dogman_1234
    lilcinwFinally we have some F@H benches!! Thank you!Having said that I would suggest you include expected PPD for the given TPF since that is what folders look at when deciding on hardware. Or you could just devote 48 hours from each machine to generate actual results for F@H and donate those points to your F@H team (yes Tom's has a team [40051] and visibility is our biggest problem).

    The issue is that other tech sites promote their teams. We do not have a promotive site. Even while mentioning F@H, some people do not agree with it or will never want to participate. It is a mentality. However, it is a choice!
    1
  • lilcinw
    I don't expect promotion at this point, just basic recognition would be appreciated.
    2
  • Luscious
    I've seen a few 2U 4-node blade servers from Asus and Supermicro running dual-socket E5's - just the thought of populating a single 42U rack with those things makes the mind boggle (168 processors, 1344 cores, 2688 threads, 42TB of total RAM and 1008TB of storage).

    F@H on such a monster? Do the math and you'll see that just after one year of 24/7 operation you would rack up over 3 billion points, putting you in the top 10 for teams and no.1 spot for single user.

    That's assuming, of course, that you've forked out $20k for your monthly power bill to run that fully-stocked 42U rack and paid $240k to your utility company for the entire year. Then there's the cost of the hardware itself - around $26k for each 2U server, or around a cool $600,000.

    SPEND MONEY FAST
    1
  • utomo88
    We need somebody to design new server which is powerful and can handle a lot of memory and harddisk and affordable price.
    all powerful server are expensive now.
    I believe market for cheap but powerful server are big, and no one is working on this area.
    I know the profit is not big, but by big quantity it mean big money too :)
    0
  • bit_user
    Would be cool to test both a E5 Xeon and a Sandybridge-E of the same clock speed, in the same C602 motherboard (Supermicro's X9SRA), to see if DDIO has any performance impact on 3D and GPGPU benchmarks.
    0
  • bit_user
    Also, how about posting measuring the scaling from 1x 4core E5 Xeon -> 1x 8core E5 Xeon -> 2x 4core E5 Xeon?

    The point is that memory is directly connected to 1 CPU only. Adding a 2nd CPU doubles aggregate bandwidth, but could actually hurt performance, if the software isn't written to carefully to localize data and manage affinity between threads & CPUs.
    0
  • _zxzxzx_
    Would be nice to see how these compare to the AMD systems.
    0
  • saturn85
    nice folding@home benchmark.
    great work.
    0
  • centosfan
    Why don't you include HP DL series and the Dell R series. Those two companies have about 70% market share on the 2U Inter server market. I don't understand why you would exclude them. Most companies wouldn't buy anything but a Dell, HP or and IBM.
    0
  • pjkenned
    centosfanWhy don't you include HP DL series and the Dell R series. Those two companies have about 70% market share on the 2U Inter server market. I don't understand why you would exclude them. Most companies wouldn't buy anything but a Dell, HP or and IBM.


    That is something that we are looking at. This was more of a look at what is out there for barebones kits. I totally agree that these types of comparisons would be great.
    1
  • DVFinn
    I've been using Supermicro bare bones servers for a few years now and I love them. When it comes to performance per dollar there's no way to compare these systems with off the shelf dell or HP offerings. I've specced them out 20 ways and the bottom line is I can build nearly 2x the horsepower at significantly lower cost vs. a fully configured server from one of the big vendors. My newest SQL servers are 8xSSD Raid10 data sets and simple sata mirrors for the OS. 128GB RAM, dual, quad-core Xeons, adaptec 6805 controllers. About $5k each for the full build, and their chassis are so easy to work in the whole build takes less than 30 minutes before I'm loading the OS.
    0
  • razor512
    utomo88We need somebody to design new server which is powerful and can handle a lot of memory and harddisk and affordable price.all powerful server are expensive now. I believe market for cheap but powerful server are big, and no one is working on this area.I know the profit is not big, but by big quantity it mean big money too


    That is already done (but as more of a work around) build a standard PC.
    Many high end gaming motherboards work well in a server environment, and can easily handle a high traffic website.
    Most web hosting does not need a super powerful server (which is why virtualization is so popular). If you are running a relatively small business and are not doing anything that is hugely CPU bound (eg, rendering) then you can save a bit of money with a decent desktop PC.
    0