Raid 5 dead hype

I'm about to upgrade a 4 x 2tb raid 5 server. I've heard a lot of flak about raid 5 being dead because if one of the drives dies there is a good chance during rebuild that another of the drives will have a URE and boom goes the array.

WTF? At worse that would cause the (on my current array) 64m chunk to be corrupt but rebuild the rest of array right? Is that the whole fear about raid 5?

Why don't you have a petabyte server user?

another drive could easily die during rebuild

What makes you think having one disk fail in an array would increase the odds of another disk totally failing?

This. I don't understand the hype. There's a periodic scrub anyway that reads all the data so why would a rebuild cause any problems?

>it's still good it's still good

Christ you sound like some whiney wintard hanging on to vfat for dear life. Just join the 21st century already and run ZFS.

Just make sure all your drives come from different lots. If they were manufactured sequentially, their odds of dying roughly simultaneously are higher.

>ZFS
>21th century
Pick one. ZFS is old. The cool kids use btrfs now gramps.

The rebuild times on RAID5 are high. The bigger the discs, the worse it gets. Past 4TB disks, I would do RAID6, or one of the ZFS RAIDZ things. RAID 10 wastes to much space for big disks Much past 4TB, or past 10 disks, and you're going to want ceph style distributed storage.

But that said, RAID is dead, long live RAID!

you'd think it would just corrupt a 64k stripe, but people make it sound like it hoses the whole array

Only true about RAID 5 being dead when your capacity exceeds 10.8TB at which point you will most certainly encounter a UER which will kill the rebuild.

Stress on the drive as it undergoes hours or days of rebuild, especially with consumer drives which are not meant to be run at 100% constantly.

You should chose your RAID based on the purpose of the array.

RAID 10's are perfect for high random access (for a raid anyway), rebuild performance or even sequential access. RAID 5 for high read ratio, raid 6 because your array is really big. Raid 4 for SSD redundancy, raid 1 for your desktops or also sequential.

>2005
>old

Tenyearold detected.

you are just fucking old if you think 12 year old software isn't old

Fairly good bait.

ZFS is more mature and has a much larger development team behind it. Btrfs has a hard time even catching up.

Raid 6 is better.

When it comes to file systems anything less than 5 years old isn't even trustable for non production use.

RAID5 takes less i/o for writes and provides more space if you don't need the extra redundancy.

This is why you will lose your data.

I still run SBS2003 for myself. Why?

My data gets backed up bro, raid and zfs are not backup solutions.

>I've heard a lot of flak about raid 5 being dead

Those are HDD shills that want to pretend coding theory doesn't exist and the only way to protect data is to have it in triplicate.

What is the best redundancy method for a 8TB volume?

OP here, ZFS isn't an option as I want the ability to grow the array and I'm not about to let my FS eat all my ram. I have 4 bays to work with and just can't justify sacrificing 2 to the gods of chance.

RAID 5 sucks cuz if you're away from the server one HD dies you're dead in the water.

Also there's a higher chance for 2 HDDs to fail.

RAID6 is king tier.
RAID10 is GOD tier.

Get a 6TB drive and back up the contents. Then create a new array with the new drives and copy the data over. Faster and safer than doing 4 rebuilds like a tard.

Was never going to do 4 rebuilds lol. dd is my friend.

ZFS is the best option for growing arrays.

...

>grow the array
It adds some different rules to how you grow it but I'm willing to work with it for what it offers. Use lz4 compression, it can work wonders on your data.
Especially the fact that VDEVs boost random I/O. Each one you add adds more random read and write capacity.

>FS eat all my ram
Much like disk cache in linux your ZFS ARC can evict data instantly when other programs need it.
htop just doesn't show how much ZFS uses in yellow so you see the whole thing being used when it's all readily available.

I've installed it before as root on a 2GB laptop and it worked fine for general shitposting and could run minecraft.
htop showed all ram used all the time but everything could run and evict ARC when necessary so it actually performed better than ext4 on that 5400 rpm hdd because of all the aggressive ram caching.

t. runs-unstable-file-system

RAID5 is still redundant
RAID6 is wasteful if you don't need it
RAID10 my favourite, but 50% bigger RAID6 arrays w/8 drives

>I'd rather have idle ram empty rather than take advantage of caching

>there is a good chance during rebuild that another of the drives will have a URE and boom goes the array
And that is the reason why many including the Multi Disk HOWTO states that arrays should be built with disks from different batches.

Remember: the data is most likely more worth than the disks. So replace one disk each year, every year.

Alternative 2 is a more complex array: RAID 6 or nested arrays, such as a RAID1 of two RAID5 sub arrays. That will cost a lot more disks though.

You should not assume shit you don't know. The bulk of my ram goes to postgres indexes.

>What makes you think having one disk fail in an array would increase the odds of another disk totally failing?
Because they're generally all from the same lot and will all die around the same time.

t.poorfag too poor to run dedicated file server

i just put together a 12tb raw, 8tb usable raidz-6 server.

by just, i mean 2 months ago. it boots, i just cba to set it up

also I fucked the file table on my desktop, posting from my laptop

storage is gay

sadly, this...
just go with raid1, dont make large filesystems - whats even the point except posting a screen of your 20tb almost full screens like a faggot (if you need large filesystems because database or similar, you are already barking at the wrong tree and should consult with a storage guy that knows his shit)
buy smaller but more disks(eg 4x3gb instead of 2x6gb or whatever faggots hype these year), dont pair disks from same batches and you are good to go

What do you mean with don't pair disks from same batches?

Raid 5 with good drives and controllers is fine for bulk storage.

probably what said