Which RAID for a Personal Fileserver? 898
Dredd2Kad asks: "I'm tired of HD failures. I've suffered through a few of them. Even with backups, they are still a pain to recover from. I've got all fairly inexpensive but reliable hardware picked out, but I'm just not sure which RAID level to implement. My goals are to build a file server that can live through a drive failure with no loss of data, and will be easy to rebuild. Ideally, in the event of a failure, I'd just like to remove the bad hard drive and install a new one and be done with it. Is this possible? How many drives to I need to get this done, 2,4 or 5? What size should they be? I know when you implement RAID, your usable drive space is N% of the total drive space depending on the RAID level."
RAID 1 (Score:5, Informative)
For simplicity and low expense, even though you lose a full drive worth of capacity, go with RAID 1.
You might want to read The Tech Report's recent article [techreport.com] mentioned on Slashdot [slashdot.org] if you haven't already.
Re:RAID 1 (Score:5, Informative)
Software raid (Score:3)
Re:Software raid (Score:3, Informative)
Depends what kind of RAID you're doing. If it's just a mirror, writes are slowed slightly, and read performance is significantly improved over a single drive. Don't even bother trying to do RAID 5 in software. Buy a 3ware Escalade controller or a SCSI RAID contr
Re:Software raid (Score:5, Informative)
I've run software RAID-5 on Linux for several years on two of my home fileservers.
The only problem I ever encountered were hardware failures (Promise *ack* *spit* PCI IDE cards) and one drive failure. Performance is not really an issue for home use; I can easily saturate my 100Mbps network card.
My Fileserver: AMD Duron 1300MHz, 768MB RAM
This device was built from 4x 160GB 7200rpm SW RAID-5 for online storage (including all of my digital photos, and my collection of CD's ripped to MP3).
For backup I have an old Celeron 433, 512MB RAM box with 4x 120GB 5400rpm SW RAID-5
The main fileserver is rsynced to the backup server once a week. CPU on the backup server is a bottleneck; the Celeron is a bit underpowered for rsync, but it works ;)
My $0.02:
- Software RAID is perfectly usable, especially for typical home use. Performance is adequate.
- With RAID-5 you "lose" only one disk to parity so it is quite cheap to build
- Yes, I'd really like a 3Ware Escalade but if the card fails I need to get a new one pronto; software RAID sets can be migrated to most PCs.
Re:Software raid (Score:5, Interesting)
Hardware controllers with batter backed RAM (note; not all controllers have this), will have an edge over software solutions on ALL writes - no matter which RAID level you use.
Don't even bother trying to do RAID 5 in software
SW RAID is usually a lot faster than HW RAID solutions, when you factor out the battery-backed RAM part. Any HW RAID controller with battery backed memory will lose big-time to SW RAID on even moderately faster CPUs (like 500MHz P-IIIs), especially on RAID-5 which is compute intensive, an even more on RAID-6 which is also compute intensive but not XOR based.
Modern HW RAID controllers have reasonably fast CPUs with XOR accelerators built in - therefore they can do RAID-5 as fast as the pure SW solution. But this is not the case with older controllers.
I know of people who use 3ware cards for large RAID-5 servers, but only use the 3ware cards as "dumb" IDE controllers, and leave the RAID-5 handling to SW-RAID. The reason? Their benchmarks indicate that this is significantly faster.
And when you think about it, it makes sense. Nobody puts a GHz processor on a RAID controller. Even a slow-by-todays-standards P-III is able to XOR more than a gigabyte of data per second - much much more than anything you put thru most file servers out there.
So, the "HW RAID is faster than SW RAID" is true in one scenario only; when you have write-intensive workloads and a HW RAID controller with battery backed cache.
In *all* other cases, SW RAID will be a win, performance wise.
For a personal file server, I wouldn't hesitate to run RAID-5 in plain software. It's as fast or faster than any HW RAID controller in the sub-$3K price range, it's reliable, and the flexibility beats the heck out of any HW based solution out there (mixing IDE/SCSI, allowing a cryptographic layer between the RAID layer and the physical disks, etc. etc...)
Re:Software raid (Score:4, Insightful)
If your HW RAID controller dies, you have to get another one of the same controller, and hope that you can re-import your config w/o losing all your data. If your running SW RAID and your SCSI/IDE controller dies, you can replace it w/ whatever is cheap/available at the time. As long as the failure itself didn't bork your data, you shouldn't have to do much, if anything, to see your data again.
If you can afford to get the top of the line SCSI RAID controller from a good vendor it's probably the better option, but if cost is an issue, IDE SW RAID is the only way to go.
Re:Software raid (Score:4, Insightful)
Re:Software raid (Score:5, Informative)
So, several hundred users using IMAP and POP3 to collect mail, SMTP to send mail, and the 100k or so incoming messages do add up to a lot of work, and it handles it flawlessly.
$ cat
Personalities : [linear] [raid0] [raid1] [raid5] [multipath]
read_ahead 1024 sectors
md0 : active raid5 hdc2[2] hdb2[1] hda2[0]
351100416 blocks level 5, 64k chunk, algorithm 2 [3/3] [UUU]
unused devices:
$ df -h
Filesystem Size Used Avail Use% Mounted on
/dev/md0 330G 11G 302G 4%
/dev/hda1 122M 8.0M 108M 7%
none 499M 0 499M 0%
Re:Software raid (Score:3, Interesting)
Performance with an IDE raid controller is pathetic. You can't get much more than 22MB/s. I can hit 68MB/s reading and 31MB/s on one system with 4 7200 8 MB cache IDE drives. (This system has
Re:Software raid (Score:4, Informative)
As for forking $ for RAID cards, I've had really good experiences w/ the MegaRaid cards from LSI Logic [lsilogic.com] - really, really good tech support and exceptionally inexpensive cards.
Re:Software raid (Score:5, Informative)
My system, for example:
36gig 15k (3.6ms) rpm scsi:
250 gig 5k (9.5ms) rpm ide:
Who needs 3.6ms access time for their music and videos? What will that gain you? I can tell you what 3.6 ms access time gives you for a root partition, though: blazingly fast startup of the system, of X, of programs, and compilation.
All of my media is in
As I demonstrated, you can get a small 10k rpm scsi drive with access time 70% better of that for all but the nicest IDE drives (which cost notably more than scsi drives), brand new and with shipping, for 30$. After re-looking at pricewatch, I found the same thing for only 20$, including shipping. You can get a new scsi controller for 20$ also, inc. shipping, that will do 40mb/s (plenty for one drive). A new cable will cost you about 6$. That's 50$ for a root partition that will give you a 70% speed boost over a 7200 RPM ide drive.
Why would one *not* do something like that, unless they really don't care about speed at all? And if they don't care about speed, why raid for reasons other than redundancy?
Re:Software raid (Score:4, Insightful)
The real advantage of software over hardware RAID is that you don't need to keep a spare RAID card around. With hardware RAID, when your RAID card fails you'll need exactly the same make & model card to read your data.
With Linux software RAID, you can read the drive set on any system with the raid modules.
Re:RAID 1 (Score:5, Interesting)
The benefits are that you get the same protection as with RAID 1, but lose the speed penalty, all without needing special hardware or spare CPU power for expensive CRC calculations.
With a 4 drive RAID 1+0, you'll get read performance of 2x-4x a single drive, while writes will be from 1x-2x. In theory, that is. In reality, if using a RAID PCI card or motherboard solution hooked to the south bridge, you'll most likely max out the read speed.
Anyhow, it's a very cheap solution that doesn't tax your CPU too much even if done through software (like with a highpoint controller), and it does give you piece of mind.
The worst downside is that you will have to take the system down to change a drive (correct me if I'm wrong, but I've never seen a hot-swappable RAID 1+0 solution), and the performance before you do that will take a substantial hit.
Raid 4/5 is nice because it doesn't waste a lot of drive space, but it comes at the price of very slow writes, and very high CPU use unless you also get a hardware controller with an onboard CPU.
Regards,
--
*Art
Re:RAID 1 (Score:5, Interesting)
If you're having two drives fail before you can get one replaced you need better hardware or a better failure notification system, or both.
And, speaking from personal experience, and both the theoretical and real-world benchmark tests, I can say quite firmly that the software RAID 1+0 on my dual P3 1GHz fileserver does give a 'spead boost'. Not the theoretical maximum of 4x read 2x write, obviously, but certainly a noticable speed boost.
And finally, you complain about a writing performance hit under RAID 1? Have you ever even used or benchmarked a RAID 5 system? Computing parity information, unelss you have a *very* expensive RAID 5 controller, puts RAID 5 well behind every other type of RAID when it comes to writing speed.
Like seriously man, have you ever even experimented with different RAID setups, or are you just extrapolating these ideas from something you read on the web?
Re:RAID 1 (Score:4, Informative)
I had heard that the new VM for linux supports snapshots so I will probably be looking into that soon but I haven't messed with my file server in over 3 years. It just works (TM).
Re:RAID 1 (Score:5, Informative)
My fileserver has a mirrored pair of drives in front mounted, hot-swap bays. I have a third drive on my workstation and I sync to that every time I add significant amounts of data to my server. The mirroring protects against drive failure and the third drive protects against server failure, operator error, filesystem corruption or other problems that can wipe out a RAID array.
Lastly, the stuff that changes often and is worth the most to me - small documents and other things I create - gets a nightly sync to the server's boot drive and I keep a month's worth of revisions. This lets me "go back in time" to retrieve things if I need to. Considering the relatively small size of this type of material, this doesn't take up a lot of space. I think the whole month's worth of revisions only takes up 10GB or so.
The hot swap bays let me yank a drive out on my way out of the house if the place catches on fire. Yes, I know I should be storing that third drive at a friend's house, but it's too inconvenient to retrieve it every time I want to backup my array. So a fire may destroy everything if I'm not home or can't safely pull a drive on my way out. I'm comfortable with that.
Re:RAID 1 (Score:5, Informative)
At a different site.
KFG
Re:RAID 1 (Score:5, Funny)
> At a different site.
In a galaxy far, far away..
Re:RAID 1 (Score:5, Informative)
DO NOT RELY ON RAID TO PROTECT YOUR DATA. If you do, you will lose it some day. Raid only protects against hardware failure. There are plenty of other ways you can lose data and one of them will catch up to you eventually.
If you can't afford to lose it, back it up to another drive on another computer. If you really can't afford to lose it no matter what, store your backup drive with a friend.
Re:RAID 1 (Score:5, Interesting)
I was even thinking of buying the app until I surfed to the company's site and found it was >$2K US. Screw that. If it happens again, I may not reciver my stuff.
I didn't have anything critical on there, but it woudl have been very time consuming to re-rip my CDs again.
jason
Re:RAID 1 (Score:5, Interesting)
Amen. I have vivid memories of typing rm -rf * in the wrong directory (and that was WITH pwd in my prompt). It took an entire week to duplicate the work lost.
Combining the rm command and lack of sleep is like combining a loaded gun and your forehead. You can only do it so often before you destroy something valuable.
Re:RAID 1 (Score:5, Interesting)
Re:RAID 1 (Score:5, Interesting)
You can resolve this issue with high-capacity, portable storage. I keep all most critical stuff (software, licenses, photos, pr0n, etc.) on my 40GB portable drive. Forget those keychain things. The FireLite SmartDisk [smartdisk.com] is a USB 2.0, aluminum encased laptop drive. It draws power from USB - it even worked on my old USB 1.l system. They provide a special power cable, in case your old USB ports aren't pushing enough power. I toss thing in backpack every day and lug it all over - it has yet to show signs of weakness.
I totally agree with your configuration. For my Linux server, I've been using Linux (RH7.2) Software RAID-1 mirrored for ~3 years without a single issue.
Re:RAID 1 (Score:4, Funny)
Re:RAID 1 (Score:3, Informative)
It'll be cheaper than tape, but more work.
RAID 5 can be appropriate (Score:3, Interesting)
Re:RAID 1 (Score:4, Informative)
Re:RAID 1 (Score:3, Insightful)
Lets see... 2.5 TB ~ 600 DVDs
And you store all of those DVDs where? And you access them quickly how?
Re:RAID 1 (Score:3, Interesting)
Re:RAID 1 (Score:3, Funny)
I would say:
And you access them quickly how with the one hand available left?
Re:RAID 1 (Score:5, Funny)
You've heard of usenet? It's just this guy's hard drive being posted over and over.
Re:RAID 1 (Score:5, Interesting)
The only semi-common RAID I know of that could handle two drives failing at the same time would be RAID 10, A mirrored set of striped drives, and then only if one side of the mirror died.
For your diligence bit, I've actually worked with a machine that had a drive fail in the RAID 5 set and then as the hot spare came online and started rebuilding the data needed to keep the R in RAID another drive died. The whole set was then completely unusable and somebody probably would have been fired if there weren't a set of recent backups around. As it was a couple people got to work about 12 more hours on top of their 8 for the day to make sure the machine was running again by the next day.
Thus my moral, RAID isn't a replacement for backups, as there still can be failures. RAID will reduce the frequency with which you need said backups, hopefully to never, but it can still fail. Nothing replaces a good backup.
Oh, and also another good reason for RAID 5 instead of 1, there should be a bit of speedup since there's multiple disks involved, assuming, of course, your RAID card can handle all the XORs.
Re:RAID 1 (Score:5, Funny)
Of course it hasn't failed. See, there is some sort of weird universal law that applies in these situations. If you take the time to put in some redundancy (raid 5, or just mirroring) then none of your hard discs will ever fail. However, if you had used the exact same physical drives in a non-redundant fashion, you can bet your buttons they would have failed. It's just one of those things... like the drives know your situation or something. If they know that their failure won't really cause you major headaches, they figure it's just not worth their time to fail. They only strike when they know they're going to hurt you.
Re:RAID 5 (Score:4, Informative)
Your description seems more to fit a RAID 1+0 which is something completely different.
And then, you don't seem to know anything about probabilities:
"If you have 15 drives, and two fail, the chances of them being consecutive are very low."
Correct. But the probability of two consecutive drives failing ist still just as high as with 3 drives! It is just much more probable that from 15 drives two fail at the same time (that's just 2/15) than from three drives (2/3). Still, it could be better to have more drives, just because you could have a better "feeling" of how many drives fail before it comes to the fatal crash.
But for RAID 5, this is irrelevant anyways, because any two drives failing will screw your data. And with 15 drives, the probability for that is much higher (and I would even say that 15 drives is a bit too much for RAID 5, use a RAID level where more than one drive can fail without data loss).
Just remember the RAID song (Score:5, Funny)
RAID 1, is equally fun,
but RAID 5 keeps you alive!
Re:Just remember the RAID song (Score:5, Insightful)
RAID 0, you need a hero,
RAID 1, is equally fun,
but RAID 5 keeps you alive!
RAID 5 - better keep an extra drive
Or you'll be down until the replacement arrives
RAID 10 is better my friend
Work doesn't stop when the drive comes to an end
Re:Just remember the RAID song (Score:5, Informative)
Re:Just remember the RAID song (Score:3, Informative)
Or you'll be down until the replacement arrives
Um, really? Software RAID 5 does require downtime, but hardware implemented RAID 5 allows for hot swapping out of the bad drive, assuming you have a decent controller card.
Re:Just remember the RAID song (Score:5, Informative)
RAID 0: This is a striped set, there is no redundancy. One drive goes, everything's gone. Useable space = 100%
RAID 1: This is a mirrored set. Typically this involves 2 drives. One drive is an exact copy of the second. If a drive fails, you replace it and rebuild the set. Life goes on. Useable space = 50%. Most IDE raid cards only support RAID 0 AND 1.
RAID 5: This is a striped set with parity. You get the performance associated with a striped set. Particularly on reads. If you have 4 drives, there are 4 stripes. 3 of those stripes are data stripes, the 4th is parity. Lose 1 drive and the parity information is used to rebuild the set. Useable space = (n-1)/n. To do this in hardware is typically fairly expensive.
There's a lot of hardware solutions out there. It can also be done in software. Windows supports creating disk sets in software. Other options include the purchase of a Snap! server, or other brand of NAS. If you've got a little $ to throw around, NAS is the way to go. Plug it into your network, minimal setup, and your off and running. Not very upgradeable, and somewhat problematic if your drive does actually die, but I use them at the office for a zero maintenence file server.
Re:Snap Appliance (Score:3, Informative)
RAID -1 (Score:5, Informative)
RAID 5 or RAID 10 (Score:5, Informative)
Quick overview:
RAID 5 - Requires at least 3 HDs (many times implemented with 5 - can be used with up to 24 I believe). Data is not mirrored but can be reconstructed after drive failure using the remaining disks and the parity data (very similiar to how PAR files can reconstruct damaged/missing RAR files for the Newsgroup pirates out there). % of total space available dependent on number of drives used.
RAID 10 - High performance, but expensive. You get ~50% of the total HD space as it is fully mirrored. So, 1 TB total disk space nets you 500 GB total storage space. Your data is mirrored so if one drive fails you do not lose everything. However, if you experience multiple drive failure you can be in big trouble.
Re:RAID 5 or RAID 10 (Score:4, Informative)
Specifically the setup is as follows
1 == 2
3 == 4
5 == 6
7 == 8
Setting up a RAID in this way will allow you to experience multiple drive failures while still keeping the raid alive. The most detremental in this scenario is if you lose two drives on the same deveice. Meaning if you lost drives 1 & 2 you expereince a more of a problem as opposed to losing drive 1 & 4.
Just my 2 cents, poke holes where necessary
Re:RAID 5 or RAID 10 (Score:3, Insightful)
Give me a break, your argument is worthless when the costs of very reasonable hard drives are in the range of $100 each. (200GB drives for $101 at pricewatch). To even consider RAID-5 you would have be wanting to do this with a minimum of 3 drives. Probably more. 3 drives comes to $300.
If this is getting too rich for you as your post seems to claim, you shouldn't be even talking about RAID-5. Not because of the price of
raid and ide channels (Score:5, Informative)
Re:raid and ide channels (Score:3, Informative)
Hardware (Score:3, Funny)
Re:Hardware (Score:3, Insightful)
"Good, Fast, Cheap: Pick any 2."
Raid 1, 0+1, or 5.. (Score:4, Informative)
Raid 1 is the safest.. just mirroring the drives, but it results in no speed increase..
Raid 0+1 does mirrored stripe sets -- you get the speed advantages of raid 0 with the full protection of raid 1.
Raid 5 is good middle ground. Raid 5 stores 1 drive's worth of parity. When you lose a drive, your system goes down (if you don't have a hot spare), but you throw another disk in and it'll come back up. You also get some speed increase over a normal drive setup. With RAID 5, you only lose a single drive's worth of capacity no matter how many drives are in your array, whereas with raid 1, you lose 50%.
Re:Raid 1, 0+1, or 5.. (Score:3, Informative)
Actually, with any proper implementation of RAID 5 you wouldn't lose functionality during a single drive failure, but you would suffer a performance hit because every read would require the drive controller to reconstruct the missing data from the checksums.
Replace the bad drive very quickly, though, becaus
Re:Raid 1, 0+1, or 5.. (Score:3, Insightful)
Re:Raid 1, 0+1, or 5.. (Score:3, Informative)
Example: two SATA drives
RAID-0: Write Speed: 2x, Read Speed: 2x
RAID-1: Write Speed: 1x, Read Speed: 2x
Basically, when doing a write, the driver can use the same buffer and stream the write data to both drives synchronously meaning no slowdown. A proper read driver will read alternate chunks simultaneously from the two drives,
Re:Raid 1, 0+1, or 5.. (Score:3, Informative)
You are unlikely to get double read performance from a RAID 1 setup. It's theoretically possible, but in practice it doesn't happen (take a look at the recently posted review [tech-report.com] at Tech Report). It's actually easier to get good performance with RAID 1 using software RAID as the OS is in a much better position to schedule reads efficiently than a RAID controller.
Old PC + 4 channel raid controller = easy (Score:5, Informative)
Here's what I came up with: Total cost about $1200 (probably less by now).
0) Red Hat Linux, ext3 filesystem.
1) 3Ware Escalade 7506-4LP card (64 bit card, but fits in 32bit slot)
2) 4x 250Gb Western Digital drives
3) Big fan.
At RAID 5 This yields 750gigs (715Gb after crappy GB conversion).
The 3Ware software has a nice web monitor interface and does daily or weekly integrity checks. It emails me if there is a problem - I did have one drive die already and replaced it easily.
Pat Niemeyer
Author of Learning Java, O'Reilly & Associates
Re:Old PC + 4 channel raid controller = easy (Score:4, Informative)
1) The raid card is well worth the $200 - it will all just work out of the box, looking like one big disk. I forgot to install the 3ware drivers on my first pass and the raid still set up and worked just fine out of firmware - the monitor just wasn't there. Don't mess with software... just buy the card.
2) Put it in the basement! It's always cooler down there... and a little noise won't matter. Temperature is the key to the life span of disks.
3) Also - it's really important that you churn through the data on a regular basis so that the raid can detect bad sectors and repair them in a timely matter or warn you of impending disk failure. The 3ware software does this automatically... but regular backups would accomplish the same. I'm a bit torn on how much reading through the data to do, since I don't use my raid heavily and the read cycle is actually the largest usage... If I overdo it I'm probably shortening the life of the disks.
Pat Niemeyer
Author of Learning Java, O'Reilly & Associates
All boils down to money (Score:3)
Bottom line, you need to figure out how much you're willing to spend on this and then go from there and see what your options are. RAID5 is the hotness, but it's very expensive (easily over $10K for large capacity devices).
Here it comes... (Score:3, Funny)
If you just run Gentoo, you can type "emerge new_harddrive" and it takes care of everything by the end of the month!
or..
Your shit PEECEE WINTEL crap parts made in china are no match for real quality Mac hardware, which are fully integrated with the UNIX UNDERPINNINGS that have the Best GUI Ever(tm) on top.
Disclaimer: I love trolls.
Dear Slashdot (Score:5, Funny)
which is better, SCSI or IDE?
Googleless in VA
My choice (Score:5, Insightful)
Raid's great, but an rm -rf is still an rm -rf, thus the third drive
Raid 5 (Score:3, Interesting)
Whichever damn raid level you want! (Score:5, Informative)
If you have a stack of 6 drives and believe not a single one is ever going to fail, go for level 0.
If you are a government contractor and are required to handle simultaneous failures of 75% of your drives, either mirror them all or go with 5+1 or a raid 10 setup.
All in all, its a poor question to ask slashdot. You need to let us know what you consider an acceptable failure, and by the time you have that figured out determining what raid level you need is easy.
RAID 5 or 6 (Score:3, Informative)
So, RAID 5 or 6 would be the best (RAID 6 is worth the extra bit of space for the 2nd calculation, and really helps when you can test the pairity bits against another pairity to create the lost data.)
There will be some slow down associated with RAID, but it wont be as bad with 5 or 6 and generally, you can live through it with the thought of having relativly robust file servers.
Software RAID? (Score:5, Insightful)
1) You don't need drives that are the same size.
I've done hardware RAID, had a drive fail 2 years down the road and not been able to find an 18GB SCSI drive to re-insert to the array. That has the potential to jack your entire array. With software RAID, you buy a 36G drive, partition it so that 1 partition fits your array, and off you go
2) It's a personal file server, so speed is less important than cost (i'm guessing). With software RAID you can mix all sorts of wonderous things together. IDE drives from the basement, SCSI-320 drives you stole from work and nearly everything in between. It's for flexible, and has no associated controller cost.
3) It's easy as heck. You can configure it in Disk Druid/fdisk, and it works quite easily in any major distribution (I've done it in Slack, Debian, RH, Fedora and Mandrake).
The major downside is that you cannot (as least I don't know how to) hot-swap drives. But again, this is a personal file server. Spend your money on pizza and beer, screw the SCA hot-swap drives that are going to cost you an arm and a leg.
That's just my $0.02...flame away
Re:Software RAID? (Score:3, Informative)
The limitations and versatility are not determined by the "software or hardware" ("hardware" being software on a dedicated raid controller) but by the design of the specific software under consideration.
True, the software raid in Linux is quite versatile, but there is no reason why a raid controller could not work with two disks of different sizes and use part of one dis
Re:Software RAID? (Score:5, Informative)
Many benchmarks show the exact opposite, except when dealing with high-end RAID cards. Why? Because the average CPU on a system with a RAID is going to be much more powerful than anything you're likely to find on a low- to medium-range hardware adapter. I use software RAID on a number of FreeBSD servers and it absolutely flies.
The major downside is that you cannot (as least I don't know how to) hot-swap drives.
That's a function of the hardware and OS. One of the above-mentioned FreeBSD servers is in a nice IBM server case with hot-swappable front-access LVD drives. The swap process is:
There's no reason you can't do hot-swappable software RAID. If there is, then someone forgot to tell me server.
Which RAID level to run ... (Score:3, Informative)
RAID 0 stripes the data across 2 or more drives and therefore offers no redundancy (in fact, in a two-disk stripe you mutiply danger of data loss x4 compared to two individual drives -- because you not only double the possibility of failure with two disks as opposed to one, but stand to lose all of the data on both drives should one fail). In any event, no point in discussing it further since redundancy is the point.
RAID 1 offers redundancy by exactly duplicating the contents of a drive onto another drive, and needs exactly two drives. This is considered the most "fail-safe" method of RAID array although offers no performance benefits whatsoever.
RAID 10 (or 1+0 or 0+1) is a combination of RAID 0 and 1 and is nearly always done with four drives, although technically it can be done with six or eight (if your controller supports them). It offers both performance benefit and redundancy, although the cost of the "wasted" drive space is quite high.
RAID 3 involves using 3 or more drives, one of which contains parity information to rebuild the lost drive should any of the other drives fail. This is one of the least popular RAID formats and has more or less been totally replaced by RAID 5.
RAID 5 involves using 3 or more drives and writes parity information across all drives in the array, allowing one drive to fail with little to no performance loss. The failed drive can be replaced and the RAID rebuilt. Depending on your hardware/software, this can often be done hot without having to power down the system at all. It is one of the most commonly implemented RAID solutions because of the good mix between drive use (the price goes down the more drives you have in the array yet you can have as little as three), redundancy, and high availability.
There are others out there like RAID 50 but nothing worth mentioning, especially for a home user.
The only question left to you is whether the RAID will be run by hardware or software (software might be a good choice if you are already running Linux on the server, but you'll have to ask someone else about it because I don't know a thing about it). Personally I chose the hardware route years ago and bought an Adaptec 2400A, which is a four-channel hardware ATA-RAID card capable of RAID0, 1, 10, and 5 -- guess which I use. I use all four channels, each with a 200GB SATA hard drive. I've lived through a couple drive failures, a full drive upgrade (when I first bought the card it was 4x60GB drives) and even once where two drives RAID tables got zapped (I'll NEVER put my drives in removable cages again) and never lost a byte of data -- so the CAD$500 or so for the investment on the card was worth it.
600GB of storage means not having to worry about all those unlicenced-in-North-America-anime torrents running out of space any time soon.
Avoid Promise like the plague! (Score:5, Informative)
What I didn't know at the time, but learned the hard way, is that Promises's RAID monitoring program "PAM" is a user-mode only application. That means that if you don't login, it doesn't run. Care to guess what happened to me?
At some point while I was gone for the weekend, I can only guess something crashed and rebooted Windows 2000. When it rebooted, I didn't have it set to automatically login (why would I? it's a server). So "PAM" wasn't running when one of the drives in the RAID 5 set failed. Maybe it even had something to do with the crash, I don't know.
Now, the point of PAM is that if a drive fails, an e-mail gets sent, in this case to my mobile phones textpage address. Since PAM wasn't running however, nothing was sent. The drive failed and, I can only guess, put off so much heat that it cooked the drive above it (why do so many cases mount hard drives horizontally above each other anyway?) and next thing I know, I can't login to my server from where I'm staying. I call a family member with a key to come by and they are unable to restart the server. It wasn't until I came home and read the BIOS messages that I understood why. Everything gone.
I had a lot of stuff on CDR, but let me tell you, I was plenty outraged that Promise could design something so utterly stupid as a monitoring utility that doesn't know how to run as a service. Even to this day, PAM still will only run as a user-mode program, and even worse, you actually have to login to the program now to start it, which can't be scripted.
F Promise. Only a complete and utter fool would be stupid enough to buy any of their products. May they rot in that special place reserved for child molesters. (Yes, I'm still bitter about it)
- JoeShmoe
.
By a process of elimination... (Score:5, Informative)
It's for home use
No data loss if a drive dies
Easy to rebuild - remove dead drive, install new one
Budget... Ah. Why is it *every* "Ask Slashdot" never mentions the budget? On the cheap, you could do simple mirroring RAID1 - most mobos with on-board SATA RAID will do this for you. The overhead is that you pay twice as much per GB because you obviously need two drives and the performance gains are negligable.
Personally, I'd take the more expensive route; get a proper hardware RAID controller with proper RAID management software. There are 4 port SATA RAID controllers (who *really* still needs SCSI for home use?) for a few hundred dollars and do full RAID5. You lose one drive for the parity info, but that could be as little as 25% of your total capacity if you get four drives instead of the the minimum RAID5 requirement of three drives.
Also, with a proper hardware RAID controller, you should also get a performance boost from use of RAID and have minimal CPU overhead. Get four of Seagate's new 400GB drives and you'll have over a TB of disk space, which should give you some bragging rights for a months or two before it's old hat. :)
RAID information (Score:5, Informative)
You want a Promise UltraTrak SX8000 [promise.com] It's the easy idiotproof array. We're using several of these.
If a drive fails, it beeps at you til you replace it. You just yank it out, and put in a new drive, the same size or larger. It then rebuilds automatically. No shutdown or reboot required.
The Linux crowd will be happy to know the RM series runs linux. I don't know about the SX series, but I suppose it does too. Either one appears to the server to be a single SCSI drive. No drivers required, other than making the SCSI card of your choice work.
There's the Linux method of doing it too, which I like a lot. It saves you a *LOT* of money in extra hardware. You can go with 3 drives without adding any extra cards to your system, or you can put in IDE controllers to add as many drives as your system can support (PCI slots, power, and physical mounting points are the limitation). Read the "Software-RAID-HOWTO", which should come with your system. I've done many of these also, and they work quite nicely. You have to shut down the system to swap a drive, and then run `raidaddhot` with a couple parameters (the md device, if I remember right), and you can be running while it rebuilds.
You should have looked it up before you posted.
RAID 5 is the most common for a large redundant array. The array size is (N-1)*size . The more drives you use in a single array, the better off you are for size loss.
3 100Gb drives = 200Gb
5 100Gb drives = 400Gb
10 100Gb drives = 900Gb
10 200Gb drives = 1.8Tb
RAID 0 is striping. No redundancy, which you won't be happy with. (One failure means losing the array.
RAID 1 is mirroring. With two drives, you still only have the size of one.
RAID 50 is nice where it does striping across redundant arrays. You lose size, but gain speed.
Most other RAID types aren't very popular for various reasons.
Watch out for going over 2Tb in size on a single block device. I'm having problems with that right now. I have two Promise VTrak 15100's with 15 250Gb SATA drives in each, and anything with a block size over 2Tb is giving me grief. There are legitimate reasons for this, most of which newer documentation claims to be fixing, but I'm still having problems with a current Linux release. Making logical drives under 2Tb works, but doesn't accomplish what I need.
I hope this helps.
consider other risks too (Score:3, Insightful)
However, what happens if your place has a fire, gets vandalized, or a burglar takes off with your server(s)?
Try this... (Score:5, Informative)
2 drives with a complete file copy at 4 am (Score:5, Informative)
You definitely don't need any type of RAID solution because it doesn't offer you what you really need. You say you want RAID, but what you really want is backup.
All RAID solution deal with disaster recovery, but they don't deal with the situation where you accidentally rm -rf a directory that you wanted. If you mirror or RAID 5 your drives, you're still hosed because both drives will delete the files. In the end, this is more important and much more convenient.
Instead, go with a better approach which is copy or tar your files every night (or every week) to a backup drive, preferably over the network on a completely different machine. This will prevent the problem of a power surge or accidental shutoff from corrupting both drives at the same time.
Indeed, no RAID (Score:3, Informative)
Initially I've dd'ed the primary to the other two disks.
Every morning the primary is 'cp -fpRu'ed to the second one. No files are deleted on the secondary, unless I'm running out of diskspace there, at which time I do an 'rsync -aH --delete' after some verifications.
Each few wee
The joys of RAID (Score:5, Informative)
RAID 1 - Drive mirroring.
Pros:
-Excellent read performance, no loss of performance if one drive crashes.
Cons:
-The amount of space you can have on this array is limited to the largest drive you can find. Then you have to buy a second one to mirror the data, which means you are paying double the cost per unit storage on your array.
-Write performance is slower than other RAID levels.
RAID 5 - Striped array with parity. You can stack as many drives as you want on this array (within limits of the controller of course) and lose only one for redundancy.
Pros:
-You can build a very large data array out as many drives as you want, losing only one for the purpose of data reconstruction should a drive in the array fail.
Cons
-Array performance dies in the event of a failure, as lost data is reconstructed on the fly from parity information stored across the remaining drives. Of course, performance is restored with the bad disk is replaced and the array reconstructed.
-You need at least 3 drives to build a RAID 5 array.
RAID 10 - Drive mirroring with striping. Essentially combines RAID 0 and RAID 1, hence RAID 10.
Pros:
-Redundant and fast. Array can survive multiple drive failures.
Cons:
-Expensive. You need at least 4 drives to get started with RAID 10, and go by 2's as you expand on the array. As with RAID 1, your price per unit storage is doubled.
-The array can survive multiple failures, but that depends on which drives die...If you lose two drives out of the same mirror set, then the array is gone
Which RAID level you pick depends on your application. If you are interested in having something like a 1 TB data dump, you'll probably want to go RAID 5. If you only want 200GB or less in your array, then RAID 1 is probably the way to go. If you are interested in lots of space, lots of redundancy, and have lots of money, then RAID 10 is probably what you want.
Don't put your faith in RAID (Score:3, Interesting)
Recommended RAID level (Score:5, Informative)
In short, I would probably recommend RAID5 if you have 3+ drives.
RAID5 gives you the most available space while still being redundant. It allows for exactly one hard drive failure.
RAID5's write speed is usually terrible, especially with a small number of drives, but write speed isn't a big deal on my home file server. (Only you know about your needs).
RAID1+0 (NOT RAID 0+1, which is inferior) is great for performance. With 4 drives, you have potentially twice the STR of one drive (writing) and 4 times the STR of one drive reading. Of course, since STR is not important for most IO, this doesn't really effect your end performance much unless you are dealing with linearly reading/writing very large files.
Writing performance will almost certainly be higher than with RAID5.
You do lose quite a lot of space (especially when you use a large number of drives). If you used a 4-drive 1+0 array, you would have the space in two of those individual drives.
RAID1 is nice, and is very reliable, but is impractical with more than two drives unless you are incredibly paranoid. RAID1 simply makes all drives copies of the others, this, you always have as much free space as one drive would have, even if you have ten. If course, you could also handle 9 drive failures and not lose data. RAID1 is fine for 2-drive arrays though.
DO NOT FORGET that RAID is no substitute for regular backups. RAID will not help if your data loss is caused by FS corruption, a cracker, accidentally typing "rm -rf
For lowest cost, I would use software RAID, such as Linux's LVM, FreeBSD's Vinum, or whatever Windows has. (RAID5 requires Windows server). (I would not use Windows as the file server myself).
For slightly higher cost, try a Promise controller.
I would avoid Highpoint and Silicon Image controllers. Highpoint, especially, is crap. (but it is very cheap, at least).
If you possibly can, I would recommend a nice 3Ware Escalade controller. Escalades are true hardware RAID cards, unlike Highpoint/SI and most of Promise's cards, and are OS independent and very stable (with certain exceptions for some unlikely configurations).
If you have any questions, you might try the StorageReview forums. There are a number of extremely knowledgeable people there, including engineers and executives-level researchers at hard drive companies. They can give far better advice than I can, I am sure.
By the way, all my comments assume that all drives are the same size. If not, treat all drives as if they are the same size as the smallest drive on the array (unless you are using JBOD, which is not redundant)
RAID 5, but more importantly (Score:5, Informative)
Next up is drives. Not all drives are alike as I'm sure you already know. Do you want a SCSI or an IDE array? I won't go into this lengthy topic further. I'll assume though that you will build an IDE array. Some drives do not work well in RAID setups. The controller companies are more likely to tell you this than the drive manufacturers. I own 6 Western Digital WD12000JB drives (7200 RPM, 8MB cache, 120GB capacity). By all accounts one would expect those drives to work quite well in a RAID setup. They have excellent read/write times individually and have a massive amount of cache. Well, one would think that and they'd be wrong. Both 3ware, Highpoint, and Asus tech support (on an OEM Promise chipset in teh A7V333) recommend against using Western Digital drives. 3Ware did however say that WD will give you firmware that works significantly better in RAID setups if you ask for it. Personally I'm a fan of Maxtor, both the drives and the company. I've had very few failures with Maxtor drives. Whenever I did they were always extremely helpful with getting me a replacement fast. I've been very impressed by ther service. I have 2 Maxtor 7Y250P0 and 2 6Y200P0 drives in the server sitting next to me. The second is a very high quality drive from Maxtor's DiamondMaX Plus 9 line. It too have 8MB cache and 200GB to spare and runs at 7200 RPM. Nice drive. The first pair are from Maxtor's MaXLine Plus II. They have a high MTTF, 8MB cache, 250GB space, and run at 7200 RPM. They are also a little bit faster than the 6Y200P0. They are excellent drives. My next drives will also be Maxtors but this time I'll be buying the SATA siblings of the MaxLine Pluss II product line.
That brings me to my next point. PATA or SATA. Does your case have an abundance of room? I mean a massive amount of room to route long 80-conductor ribbon cables? Do you have at least 1 if not 2 PCI slots to waste below your RAID controller with the room needed to route the ribbon cables and make connections? If not then you need to go with Serial ATA drives. Don't even think twice about it. Go with SATA. The drives cost almost the same nowadays and you'll find wht little price difference there is ($5?) is worth it in the end. SATA drives are so much easier to wire. I have a case full of round cables. The case I have is an extremely large Codegen case and even I am having trouble with the cable mess. SATA is a wonderful thing. Along the same lines is hot-swap cages. There are a dozen brands to choose from. You should probably utilize them, even if you don't need hot-swap capabilities. I need them to create 3.5 drive slots from 5.25 bays. If you do want to do hot-swapping, make sure you drive cage and controller support it.
Finally we get to RAID levels. You don't want to increase your risk of losing data so level 0 is out. 1 is extremely redundant and with the right controller can actually speed up reads. It's also costly at twice the cost per GB. Unless the data you're storing is absolutely critical you won't want to use 1 (in most cases). Forget about level 2. For starters th
Building a home fileserver (Score:5, Interesting)
I basically built a box to do nothing other than fileserv. I put together a nice simple old PC (550mhz with 256 meg of ram) and mounted it in an old rack mount case I had lying round.
It's running debian with 2.4.26.
I'm running software raid and installed 2 x 2 interface IDE cards.
I threw in 6 seagate 120 gig drives (the ones with the 8 meg cache) and ran raid5 across 5 of them and a hot spare to rebuild the raid should a drvie fail. Each drive has it's own IDE channel to prevent channel faliure from screwing my raid.
I'm using ext3 as the filesystem and wrote my own little raid mon script that SMS's me should a drive fail and alarms locally.
This setup has been rock steady and gives me 460 (ish) gig of usable space after formatting.
For added peice of mind the machine is plugged into a UPS that is connected to the machine via Serial. If the UPS kicks in it shuts the machine down properly after sending an alarm SMS (the DSL and switch are also on the UPS) (yes I'm a paranoid freak)
This makes a perfectly good media and file server and I've had no problem with it in the few months I've had it.
I also reccomend setting the spin down time onm the drives manually with hdparm. It was getting awfully warm in the box till I turned that on on the seagates. Modern drives are rather hot.
I have the whole thing mounted via SMB on my other boxes around the house and it's fast,(gig ethernet) reliable and easy.
Tho do remember that no amount of raiding will save you if you lose 2 drives through some horrible freak of badness, and no raid level is going to protect you from a house fire. Hence mine also rsyncs all my absoloutely vital files (scanned family photos and docs) offsite to a file storage site every night at 2am so as not to chew my bandwidth dduring usable times. Don't forget the only truely secure data is that which is backed up.. and offsite.... twice.
Rsync every night (Score:3, Insightful)
Ok, here is the info and prices (Score:4, Informative)
Sigh.
The cheapest RAID 1 OS internal and independent RAID (MIRROR) is Duplidisk3 by ARCOIDE.com
You also get a ton of implementations; Stand alone, PCI card (for power only), 3 1/2" bay, and 5 1/4" bay. The ones that install in bays are so the user can seethe status lights.
If you want an external RAID 5 the cheapest I have found is this - http://www.coolgear.com/productdetails1.cfm?sku=R
If you want 5 disk RAID 5 those are @ $1200. http://www.cooldrives.com/fii13toatade.html
If you want external RAID 0 or 1 relatively cheap then go with one of these - http://www.cooldrives.com/dubayusb20an1.html
You can find a ton of these devices on the web since they all use the same drive controllers and bays. The nice thing about these is that sometimes you can talk the store into selling you the RAID system without the external case. These things simply require you plugging in an IDE cable and power and can be installed in any PC case that has 2 5 1/4" bays open. If you but just the 2 bay controller they are @ $230 or so. I have one and I am really happy with it.
Everything I listed above uses IDE drives and is OS independent.
Software or hardware RAID-5 (Score:5, Informative)
Basically, your options are RAID-1 and RAID-5... as hundreds of people here have already pointed out. RAID-1 is just straight mirroring (where all drives in the array contain the same information). Usually, this just involves two drives, but there's no reason why you couldn't have, say, three or four drives all mirrored... and you could lose all but one of them and still be up and running.
RAID-5 is a very cool beast. You bascially have an array of drives with some portion of them set aside for redundancy. Most of the posts I've seen here only describe a scenario where you have three drives with one of those drives for redundancy. This only scratches the surface, however.
For example, you could have an array of, say, 5 10GB drives, with 2 drives' worth of redundancy. With this, your RAID implementation would make available to you, what seemed to be, a single 30GB drive (since 20GB of the total 50GB is used for redundancy). This way, you could have any two drives go bad and you're still okay.
Another example, I guess, is that you could have a two-drive RAID-5 with one drive's worth of redundancy. In this case, you'd have the functionaly equivalent of a RAID-1 mirroring setup. Not very sexy... but you could do it in some implementations, I'm sure.
I'm trying to use the phrase "X drives' worth of redundancy" instead of "X drives set aside for redundancy" because it's important to point out that, in RAID, all of the drives are considered equal. If you have 5 drives with 2-drive redundancy, it's not like you set 3 of them as the "main" drives and 2 as the "backup" ones. There's no preferential treatment like that. All the drives are equivalent and you could lose any of them and the others all move to cover for the one that was lost.
Now, personally, I like RAID-5 because it offers the ability to use more than 50% of the space you paid for. With RAID-1 mirroring, you always only get to use 50% of the space that really exists. This would be necessary if, when you suffered a storage failure, you always lost half of it. But that's not how it happens. Usually, you lose a single drive. So, it would be nice to maximize your space available, while having some insurance against a single drive failure.
This is where RAID-5 really shines, because each successive drive you add, you get all of that space for your usage. You could have, say, four drives, 1 drive of redundancy, and you get 3 drives' worth of space.
Now, there are a few pros and cons for both RAID-1 and RAID-5 regarding recovering/moving data and changing the size of your array, and I'll list them here.
Re:search the fscking google (Score:3, Funny)
Re:search the fscking google (Score:3, Insightful)
Re:search the fscking google (Score:5, Funny)
Re:search the fscking google (Score:4, Funny)
Re:search the fscking google (Score:3, Funny)
www.google.com (Score:5, Informative)
Re:search the fscking google (Score:5, Informative)
You also never touched on the possibility of him having only 2 drives, in which case RAID 1 would be the way to go for data redundancy.
Re:search the fscking google (Score:5, Funny)
Holy crap! There are 1.5 million of us? Now I know what to say the next time a bully asks me, "You and what army?" THE SLASHDOT ARMY!!!
Re:search the fscking google (Score:3, Insightful)
Because Google turns up 1,400,000 hits of mostly crap in 0.11 seconds. When you need advice, do you ask a librarian, or a group of trusted friends? By your logic, we should trust the company that wants to sell us RAID cards. I'd rather ask people that use RAID products, not sell them.
Re:search the fscking google (Score:5, Insightful)
No offense intended here either, but why is it that every time someone posts an "ask slashdot" question someone else feels compelled to complain (and occasionally get downright rude) about why the user didn't just "google it"?
Google will get you articles and advertisements, true, but most of the time what the questioner is really after is peoples OPINIONS and EXPERIENCES.
If I post a question like "what's the best backup program you've used on linux" I'm looking for 1.5 million slashdotters EXPERIENCES with backup programs...a google search will get me a list of programs and some reviews if I'm lucky, but that's no substitute for hearing from a bunch of people who've actually DONE or USED something.
Hearing from a few hundred or thousand responders is a better recommendation than a "C-NET" review anyday!
RAID5 (Score:3, Informative)
IDE is so cheap you might as well just buy two big sata drives for most usage. Do make sure you buy two drives from two different vendors - its really embarrassing when you use two identical drives with near serial numbers and they fail the same day.
Also keep external backups. One place I worked we lost an entire array and
Re:search the fscking google (Score:5, Insightful)
I'm glad he asked. I benefit from reading the discussion, including the various tangents. This gives me another opportunity to consider using RAID at home and benefit from some "war stories" folks might offer. My needs aren't exactly the same as his, but fortunately people never stick to the exact question asked, anyway. The free information people give out is invaluable, especially the stories of personal experiences and descriptions of people's personal setups at home.
Re:search the fscking google (Score:4, Informative)
Let's see. My server requires half a terrabyte of storage.
3 200gb IDE drives at $100/ea == $300
3 180gb SCSI drives at $700/ea == $2,100
Yeah... Not likely, pal. And certainly doesn't qualify for "affordable" like this guy is clearly looking for.
Good God, you're dense... (Score:4, Insightful)
Therefore, if your data is important you won't just trust that an unlikely event won't happen - you'll assume that it will happen and make sure that it won't affect the integrity of your data.
Therefore you'll be using RAID and preferably regular backups whatever you do. This is what ensures your data integrity, not the reliability or otherwise of your drive.
After that, it's a case weighing performance, the cost (in money, manpower and downtime) of replacing a broken drive and the cost of setup against each other, and this is where it starts to make sense to use IDE drives for RAID:
For instance, say you've got 5 IDE RAID array. Over the space of say, five years you end up having to replace three of the drives - that's eight IDE drives you've had to buy
You also do the same thing with SCSI drives, and luckily none of them break - that's 5 SCSI drives all in all.
Now, say the IDE drives cost $100 each compared to $500 for the SCSI drives. You've spent $800 in the IDE case compared with $2500 in the SCSI case. There was no difference in the safety of your data but the SCSI one cost three times as much.
Therefore to choose SCSI, you'd *really* want to get that extra little bit of speed, which to be honest is more likely to be limited by the network to your server anyway...
So, to recap - assuming your data is valuable to you, the choice between SCSI and IDE has nothing to do with the disk reliability because you'll be relying on some other systems (RAID and backups) for your reliability anyway.
RA *I* D (Score:3, Insightful)
Right now, the cheapest HDs per GB are 30GB@$3 = $0.10:GB. Cheapest RAID controller card is $15 for 4 drives; a PCI PIII/1GHz server stacked with 24 drives gives 720GB (down to about 600GB with RAID redund
Re:search the fscking google (Score:5, Insightful)
there are two kinds of people: those who have had hard drive failures, and those that will have hard drive failures. i don't care if jesus h fucking christ himself blessed your hard drives.
I suggest looking at getting reliable drives before looking at a RAID solution.
and, if the poster is looking for the more-realtime-than-backup-restore reliability as he indicated, i suggest he look at raid BEFORE looking at drive quality.
the name of the game is redundancy. a RAID array of cheap drives (let's remember that it stands for Redundant Array of Inexpensive Disks) *is* more likely to have a single hard drive failure - but it's recoverable. however, it's far less likely to have multiple, simultaneous drive failures on the same day (unrecoverable) than your one, expensive, better-quality hard drive is likely to have a single failure - which is unrecoverable.
Re:Second poast (Score:3, Funny)
*ducks*
Re:Raid-1 & Plan 9 (Score:3, Funny)
Plaid 10
Pain 10
Neither sound good.