Introduction & Goals of the Build

The market for network attached storage (NAS) devices has registered huge gains over the last few years. In keeping up with the market trends, the coverage of NAS units has also seen an uptick on AnandTech since the middle of 2010. Followers of our NAS reviews have seen the standard Intel NASPT benchmarks and file transfer test results along with a qualitative coverage of the NAS’s operating system / user interface. The reviews briefly touch upon miscellaneous factors such as power consumption. The feedback from the readers as well as the industry pointed towards some essential NAS aspects such as performance under loading from multiple clients being ignored. Towards the end of 2011, we started evaluating approaches to cover these aspects.

Our goal was to simulate a SMB (Small to Medium Business) / SOHO (Small Office / Home Office) type environment for the NAS under test. From the viewpoint of our testing, we consider a SMB as any setup with 5 - 25 distinct clients for the NAS. Under ideal circumstances, we could have had multiple PCs accessing the NAS at the same time. However, we wanted a testbed which didn’t require too much space or consume a lot of power. It was also necessary that the testbed be easily administered. These requirements ruled out the possibility of multiple distinct physical machines making up the testbed.

In order to set up multiple virtual machines (VMs), we wanted to build a multi-processor workstation. One of the primary challenges when running a large number of VMs on a single machine is the paucity of resources. It is important not to be disk bound. Therefore, we set out with the intent of providing each VM with its own processor core, physical primary disk and network port. After taking a look at the options, we decided to build a dual processor workstation capable of running up to 12 VMs. In the first four sections, we will take a look at the hardware options that we chose for the build.

Following the discussion of the hardware aspects, we have a section on the software infrastructure. This includes details of the host and guest operating systems, the benchmarking software and scripts used in the testing process. We initially gave a trial run of the new test components on two different NAS units, the Synology DS211+ and the Thecus N4800. Results from the new test components are presented in the two sections preceding the concluding remarks.

Hardware Build - Motherboard, CPUs and Coolers


View All Comments

  • mfed3 - Wednesday, September 5, 2012 - link

    someone didn't read the title of the article or the article itself. the purpose is to set up a testbed, not build a system with this software target in mind. Reply
  • Zink - Wednesday, September 5, 2012 - link

    At the same time this system seems extremely over the top for the uses mentioned. It seems likely that the same tests could be run with much less hardware. I know the testbed as specced can be used for much more than testing NAS performance but the only use discussed is simulating the network utilization of a SMB environment.
    The SSDs are justified because a single HDD was "not suitable" for 12VMs but it seems there are intermediate solutions such as RAIDing two 512GB SSDs that would provide buckets of performance and a cleaner solution than 14 individual disks. I also do not understand how having a physical CPU core per VM is needed to “ensure smooth operation” if network benchmarking software is I/O bound and runs fine on a Pentium 4. Assuming you really do need 64GB of RAM for shared files and Windows VMs then it seems a 1P 2011 board would be more than up to running these benchmarks. Switch to Linux VMs for Dynamo and you could try running the benches from an even lighter system such as an i7-3770.
    On the network side would it not also be possible to virtualize the physical LAN? The clients could connect together over the internal network and the host OS on the tested perform the switch’s role and stress the NAS over a single aggregated link? For testing NAS performance specifically, what would the effect be of removing the VMs entirely and just running multiple Iometer sessions over a single aggregated link or letting Iometer use the multiple NICs from the host OS?
    NAS benchmarking would be an interesting application to try to optimize a system for. A simpler system would help you out with reducing power consumption, increasing reliability and reducing cost. You could run some experiments by changing the system configuration and benching again to see if the same NAS performance can be generated. Figuring out what other kinds of systems generate the same results would also make it possible for other editors to bench NAS units without having to purchase 14 SSDs.
    Sorry for complaining about the system configuration, I know you built it to test other hardware and not as a project in itself but I find the testbed more interesting than the NAS performance.
  • ganeshts - Thursday, September 6, 2012 - link

    Zink, Thanks for your comment. Let me try to address your concerns one-by-one, starting with the premise that the current set of tests are not the only ones we propose to run in the testbed. That premise accounts for devoting a single physical core to each VM.

    As for the single disk for each VM vs. RAIDed SSDs, that was one of the ideas we considered. However, we decided to isolate the VMs from each other as much as possible. In fact, if you re-check the build, the DRAM is the only 'hardware component' that is shared.

    We didn't go with the 'virtualizing the physical LAN' because that puts an upper limit to the number of clients which can be set up for benchmarking purpose (dependent on the host resources). In the current case, using an external switch and one physical LAN port for each VM more accurately represents real world usage. Also, in case we want to increase the number of clients, it is a simple matter of connecting more physical machines to the switch.

    Multiple IOMeter sessions: As far as we could test out / understand, IOMeter doesn't allow multiple simultaneous sessions on a given machine. One can create multiple workers, but synchronizing across them is a much more difficult job than synchronizing the dynamo processes across multiple machines. I am also not sure if the workers on one machine can operate through different network interfaces.

    As noted by another reader, 12 VMs haven't been able to max out the N4800 from Thecus. The next time around, we will probably go with the RAIDing 512 GB SSD option for storage of the VM disks. Physical NICs are probably going to remain (along with one physical CPU core or, probably, thread, for each VM).
  • bobbozzo - Thursday, September 6, 2012 - link

    Hi Ganesh,

    Could you post power consumption for the server with the CPUs loaded (with Prime95 or whatever)?

    I'm thinking of building something like this for a webserver.

  • ganeshts - Friday, September 7, 2012 - link

    Power consumption with Prime95 set for maximum power consumption was 202 W with all CPU cores 100% loaded. Note that the BIOS has a TDP limit of 70W before throttling the cores down.

    However, I noticed that RAM usage in that particular scenario was only 4 GB in total out of the 64 GB available. It is possible that higher DRAM activity might result in more power usage.
  • Stahn Aileron - Wednesday, September 5, 2012 - link

    Just out of curiosity, when you run with multiple clients accessing the NAS, are they all running the (exact?) same type of workload? Or is each VM/client set to use a slightly (if not entirely) different workload?

    I'm curious since, from a home network PoV, I can see multiple access coming from say:

    -One (or more) client(s) streaming a movie (or maybe music)
    -Another (or several) doing copy (reads) from the NAS
    -Others doing writes to the NAS
    -Maybe even one client (I can't really imagine more) doing a torrent (I don't like the idea of a client using a mounted shared network device as the primary drive for torrenting, but you never know. Also, some NASes feature built-in torrent functionality as a feature.)

    I'm just wondering how much the workload from each client differs from one another, if at all, when conducting your tests/benchmarks.

    Also, for the NASes that do RAID, will you be testing how array degradation and/or rebuilding impacts client usage benchmarks?
  • ganeshts - Wednesday, September 5, 2012 - link


    Thanks for your feedback. This is exactly what I am looking at from our readers.

    As for your primary question, in our benchmark case, all the VMs are running the same type of workload at a given time. The type of workload is given in the title of each graph.

    It should be possible to set up an IOMeter benchmark ICF file with the type of multiple workloads that you are mentioning. I will try to frame one and try to get it processed for the next NAS review.

    Ref. array degradation / rebuild process : Right now, we present results indicating the time taken to rebuild the array when there is no access to the NAS. I will set up a NASPT run when rebuild is in progress to get a feel of how the rebuild process affects the NAS performance.
  • Stahn Aileron - Thursday, September 6, 2012 - link

    Glad to be of some help. To be honest, benchmarking and running tests (troubleshooting) is something I used to do in the Navy as Avionics Technician. I actually do kind of miss it (especially being a tech geek.) Reminiscing aside...

    Back on-topic: what I described in my previous post was more of a home user secenario. Is there anything else you would also need/want to consider in a more work-oriented "dissimilar multi-client workload" benchmark/test? If this was a SOHO environment, I would add the following to my previous post:

    -DB access (not sure how you want to distribute the read/write workload, though I suppose leaning heavier to reads).

    I mention this now because my previous post for read/writes was more along the lines of sequential instead of random. I would guess DB access would be more random-ish in nature.

    For other work-oriented scenarios in a "dissimilar multi-client workload" benchmark, I'm not sure what else could be added. I'm mainly just a power-user. I dunno is people would really use an NAS for, say, an Exchange Server's storage or maybe a locally-hosted website. (Some NASes come with Web service funtions and features, no?)

    I'm just throwing out ideas for consideration. I don't xpect you to implement everything and anything since you don't have the time to do that. Time is your most precious resource during testing and benchmarking, after all.

    Thank you all for running a wonderful website and to Ganesh for a quick reply.

    Oh, one last thing: does disk fragmentation matter in regards to NASes? Would it affect NAS perfomance? Do any NASes defrag themselves?

    This is more of a long-term issue, so you can't really test it readily I'm guessing. (Unless you happen to have a fragmented dataset you could clone to the NAS somehow...) I haven't heard much about disk fragmentation since the advent of SSDs in the consumer space. That, and higher perfomance HDDs. This is mainly just a curiosity for me. (I do have a more personal reason for my interest, but it's a long story...)
  • insz - Wednesday, September 5, 2012 - link

    Interesting article. Would it be possible to add some pics of the final setup? It'd be interesting to see what the testbed would look like assembled and wired up. Reply
  • ganeshts - Friday, September 7, 2012 - link

    I didn't add the pics to the article because the setup wasn't 'photogenic' after final assembly and placement in my work area :) (as the album below shows). Doesn't matter, I will just link it in this comments section

    2012 AnandTech SMB / SOHO NAS Testbed :

    Individual images:

Log in

Don't have an account? Sign up now