Top Tips on How to Squeeze More From Your IT Systems?

jpishgar

Splendid
Overlord Emeritus
Hey all!

We've got a big feature article coming up based on the above title, and we'd like to address a few questions the IT Pros in the crowd have and share a bit of your own insights with the community.

Every business has IT systems, whether they are racks for HPC computing or ranks of SAN boxes. You might have one server in a closet, or a whole data center humming along in the desert. One thing remains consistent - everyone out there is constantly searching for more ways to wring more out of their IT dollar.

Virtualization is one way (and we cover this here on the forums in Business Computing), but that's a pretty big generalization. Surely, some approaches to virtualization prove more effective than others.

We also hear a lot of buzz about converged architectures, where IT has "piles" of compute, storage, and networking infrastructure that can be dynamically allocated to the organization as needed. What we want to know:

• What has been your experience with converged architecture?

• Has it been helpful?

• Do you need more info on the subject?


If you've got some pearls of wisdom and advice, we encourage you to share some of your top tips. If you find you are lacking on advice, tell us what you'd like to hear about within this context. This is your change to help shape the Tom's coverage, so we can in return offer you better, more actionable information.

Thanks in advance!

Yours,
William Van Winkle, Contributing Writer
Joe Pishgar, Senior Community Manager

 

williamvw

Distinguished
Feb 4, 2009
144
0
18,680
Let me see if I can goose this discussion a bit with some more specific questions. Please feel free to address any of the following:

1. How do you go about determining if your current infrastructure is sufficient for the business you anticipate conducting 12 to 24 months forward?

2. How do you go about planning flexibility into your infrastructure? For example, say you made a big infrastructure investment six months ago, and today you have a need to change things up – you need more storage bandwidth or a change in application types demands higher compute capabilities. Did you plan for such eventualities, and, if so, how?

3. Do you use cloud resources in order to help manage workloads? If so, can you describe how in general terms?

4. There is this idea of “silos” in data centers, that “this app runs over here” or “this segment of the business runs over there.” Converged infrastructure, which shares compute, storage, and networking resources from a centralized pool as needed, is increasingly mentioned as a fix for the inefficiencies of “siloing.” Is it? Have you witnessed these inefficiencies in your own operations? And have you tried some sort of converged infrastructure as a remedy? If so, did it work?

5. What areas of your IT infrastructure now benefit most from virtualization?

Thanks for taking a couple minutes to consider this!
 

daglesj

Distinguished
Jul 14, 2007
485
21
18,785
Buy 2008 spec Skt771 Dell Xeon Workstations (T5400/T7400) and kit them out with dual CPUs and as much ECC ram as you can find. Cheap as chips to buy, built like tanks and will run pretty much anything (VMs etc) forever.

Crazy machines for the money. You don't always have to spend a fortune. Hunt out your local Enterprise IT Clearance guy for bargains.
 

genz

Distinguished
1. KI's differ from project to project, but basically come down to 3 things. Bandwidth and Capacity of either processors, non-volatile storage or network infrastructure. Whilst the latter would be bringing things outside of the scope of unified computing (besides the limitations of single system servers as far as networking performance), keeping growth manageable has always been about having a big cache pool and larger bandwidth, as well as keeping 3x failovers (in addition to tape backup etc), so that in periods of extreme unforeseen demand, you have a large pool of cache that should soak up the missing bandwidth and you can cannibalize one of the failovers to provide 50% more storage (as half of the storage has to act as a failover to the new data).

I don't work on absolutely mission critical systems though. Lots of the data I store is very low demand. I work on unified webservers in a university that also double as virtualisation hosts, entry access and frontline account dataservers (there are secondaries that run through peak hours as well as tertiaries that run during updates etc). Rather than run servers for each task, I now run all as all but with the critical account data on a separate low power server. Power efficiency is much better thanks to the fact that I can disable and even shut down all but two systems per rack during off peak hours and have the others on network startup with a response time of around 10 seconds. (I have one on standby for a 2 second resume and the rest fully shut down)

2. Yes, we do. The answer to that simply is to requisition the biggest, densest parts now and leave slots, racks and connectivity to spare. If we need 50TB of space, we need at least 2.5TB SSD cache and we have a SAS capacity of 25 sockets, we get 4TB HDDs, but we only half fill all our storage racks. Parallel SSDs obviously increase bandwidth so that is on a per case basis but the 120% rule applies. We have no intention of doubling our storage needs this year, but storage racks are cheap, RAID will always justify it's cost provided we get one's with good throughput and downtime for installation of storage racks are not. Our primary servers use 16GB RAM sticks in a 128GB configuration, because again, only a quarter of total needs is sated. We want the hardware not only to be able to handle massive growth in users, for example if there is a news article about the Uni that drives a spike in traffic, but also a massive growth in the needs of the virtualized machine demands. Of course when DDR4 comes out, like DDR3 was, it will not be cost effective to move to straight away and we will review the servers then.

3. IMO you should always have a off-site cloud service like AWS for emergencies, but only as a last ditch "fan+poo" solution. This is not my job in my workplace however so I probably cannot help you with anything up-to-the-minute.

4. Yes, yes and more yes. But not always. Most of the tasks that an aisle would carry out are a particular resource intensive. For us Virtualisation is storage intensive, web-service and database is granularly CPU intensive (due to DB's strong crypto) so can be distributed well over spare CPU capacity but is storage bandwidth and capacity intensive, so what we did is separate our systems into two 'silos' for 4 (technically 3) tasks. The databases themselves are on their own aisle, but the processing is on another. It would have been possible to completely consolidate the two into a 'super aisle' of sorts, and probably saved a system or two worth of unused cycles and bits in the process but security makes it impractical to partition away the protected data, and simpler to secure that data further on a physical level.

5. I would say reliability and running costs as well as investment. The fact that our load balancing and redundancy now scales across nearly every system we have due to the fact that most are dual or triple purposed, means that a worst case cascade failure like a power regulator failing and causing a blowout (god forbid) does not always mean too much downtime as we can now fire up previously single purpose failovers or backups and have them carry out all of the jobs. This in turn means we need less backup systems, and we can run less systems at lowest-demand hours as our idle power draw is much lower across the board.