If you insist on doing this in a single box, consider running Linux. There are quite a number of good vendors for reliable, high-density Linux boxes and storage systems.
Linux is definitely the way to go. Unfortunately, using a ready-made solution will definitely exceed the budget.
DIY would be better cost wise, but it comes with the compromise of self-support. And it's likely to exceed the budget as well.
I'm surprised nobody has mentioned 3 TB drives, you can get 7200 RPM 3TB drives for $179 or 5400 RPM 3TB drives for $149 these days, why not reduce the number of bays you need by ~50%.
At this level, enterprise disks would be required with a hardware RAID card for stability reasons (RAID6 given the member count), and recommended if done via a software implementation (i.e. multiple arrays in Z-RAID2 + 2 hot spares per array due to both the member count, and it's to be used for remote access <I treating it as a remote system too = no person available 24/7 to fix a fault>).
A Hitachi 3TB Ultrastar goes for $350 each from newegg (
here).
2 arrays @ 7 members each + 2 hot spares per (RAID6 or Z-RAID2), gives a usable capacity of 30TB.
That doesn't leave much, and it certainly won't cover both the server and a backup solution (assuming a JBOD configuration, there's another $1800 in 3TB consumer grade Hitachi Deskstars <not my favorite brand either for consumer disks due to failures>).
- Total disks = $8100 (no boards, enclosures, or cables).
Keep in mind RAID is not a replacement for backup...
Absolutely.
Given the amount of data involved, I'd skip trying to use anything but a proper backup (there are inexpensive ways of doing it for the requirements, but eSATA may not be an option; i.e. single eSATA port would take ~2 days @ 200MB/s). Adding more disks (consumer units would work here due to the lower duty cycle) on a second set of non-RAID controllers would probably be the cheapest way to go (may need another disk enclosure; say 3TB disks and a
24 bay Norco enclosure).
There's also a bit of a concern with SATA due to the cable length limitations (1.0 meters), as the SFF-8088 to fanout adapters used tend to introduce too much contact resistance (results in unstable arrays - seen this before). In such cases, shorter cables are needed, which means a custom order (say 0.5 to 0.75 meters).
RAID rebuild time is going to be more affected by RAID-6 instead of RAID-5, by software RAID instead of hardware RAID than by 3 TB vs 2/1 TB. The double parity overhead of RAID-6 makes it very slow, but for a budget 30 TB file system I can't think of a better trade-off.
RAID 6 or Z-RAID2 would be the way to go in terms of redundancy vs. performance trade-off for the given capacity requirement IMO as well.
Advantages of 3 TB drives - chassis needs fewer bays, future expandability with the same bays is better, administration is easier (fewer drives, fewer potential points of failure), power draw is significantly less - if you've shopped for colo's lately you'll find that power is pretty expensive these days, moreso than rackspace or arguably bandwidth.
There are cases where using smaller capacity drives has an advantage (no need to keep buying enclosures), and additional performance due to the additional parallelism from additional members in a set.
But in this case, the budget may not work out (need to go back and see if an additional enclosure + 2TB disks is cheaper, but I don't think it will be).
RAID-5 would be ridiculous especially if you were to go with 31 1TB drives, that would be so risky it's ludicrous. Two drives go out and you just lost all your data. Thats 3% of your drives can fail and you're SOL.
Exactly. Even if someone was there 24/7 to catch a failed disk (i.e. no hot spare), there's still the risk of a second failure during the rebuild process, which is becoming more critical with increased platter densities.
If you want to save some money, build your own ZFS box based on OpenSolaris or FreeBSD.
Given the budget, I don't see away around this approach.
Edit: And now the most important recommendation:
If you want really good recommendations and background information, drop the user nanofrog a message. He's the storage expert in this forum.
Gee. Thanks for the mess you got me into...