Hey! There is a new video on this topic "So if Hardware Raid is Dead... then What?" kzbin.info/www/bejne/h5CtgKd7d6t7etU
@__teles__2 жыл бұрын
In 40 years of IT, I have experienced two occasions where the RAID controller card went mad and wiped the data. I stopped relying on RAID but backing up bejesusbytes and having a way to restore it without taking days is the real problem. I've recovered from a hardware disaster but the business was out of action for days. The cloud has the advantage that you can blame somebody else.
@Ughmahedhurtz2 жыл бұрын
"But storage is so expensive versus just buying our own!" The law of TANSTAAFL is immutable. Trying to explain that to management is frustrating.
@666Tomato6662 жыл бұрын
@@Ughmahedhurtz sure, but when cloud storage costs per year the same as buying the drives yourself (and they'll last way more than a year), the calculation isn't so straight-forward.
@PanduPoluan2 жыл бұрын
@@666Tomato666 Agree. There's always an inflection point where you'll be better served with rolling your own solution.
@harryhall4001 Жыл бұрын
@@Ughmahedhurtz Or you could just use a modern file system like ZFS with built in RAID and automatic integrity checks.
@nilsfrahm1323 Жыл бұрын
Just a crazy idea: at filesystem level, you can zip or rar the data in 'store' mode (no compression), add recovery info and later it can detect corruption or even recover. Writing many files in such archive is sequential, works much better on hdds (less seek) and overhead is very small on modern cpus. But it only works for files that are very rarely written.
@LAWRENCESYSTEMS2 жыл бұрын
Great topic and long live ZFS! I have been a bit curious about the claims made GRAID and this video really helped me understand it better, and of course better understand its shortcomings compared to ZFS & BTRFS.
@shadow70379322 жыл бұрын
All hail ZFS!
@Marc.Google2 жыл бұрын
Long live the cult 😉
@Gjarllarhorn12 жыл бұрын
I read the video title and immediately though... ZFS!
@YvanDaSilva2 жыл бұрын
Long live CEPH & GlusterFS :)
@ewenchan12392 жыл бұрын
LTT also tested and experimented with GRAID already. They also found some of the similar kinds of problems that Wendall has found (and talked about) here.
@totojejedinecnynick2 жыл бұрын
Took me months of self-researching and testing to get my head around pitfalls you explained so clearly in a 20 minute video. Bravo. Most people think that their raid5-6 array is safe, until it isn't and data recovery fails. Silent bitrot, scheduled parity checks, recovery, rebuild performance... as an individual I ended up between zfs (on linux) and btrfs. Now I am on a long, very long journey to covert all that mess into a giant ceph deployment which is a different level of headache but it seems to be a solution for availability, correctness, consistency and performance (in that order)
@Anonymous______________2 жыл бұрын
RAID arrays were never a guarantee of data integrity or protection. They are simply ways of dealing with hardware failures, so that there's no single point of failure. Even worse, all of those fancy file systems such as btrfs or ZFS are not traditionally found in Enterprise class Linux distributions, which precludes them from most environments aside from going out and having to download them through additional repositories. There is nothing worse than updating your machine to find out your ZFS volumes no longer mount because the kernel module was built with older headers and dkms didn't automatically recompile it based on the new kernel version. Those are just some of the issues Linux "storage architects" either ignore or dismiss as just normal woes of using those particular file systems. Issues with bitrot and silent data corruption have always existed, and most people do not discover them until they attempt to perform a certain file operation somewhere down the rabbit hole in a directory that almost always exceeds the 255 character limit of windows lol.
@totojejedinecnynick2 жыл бұрын
@@Anonymous______________ I don't see anything wrong with new optional features (like parity checks, bitrot protection, deduplication, COW, snapshots) built on top of basic RAID functionality (redundancy in case of dead drive). It is clever engineering leveraging potential which raid brings and using it to something useful because we have extra cpu cycles available anyway. Otherwise your only option is deployment of aforementioned 520b drives. As far as compatibility goes, btrfs is now in mainline kernel, developed by names like FB, SUSE, Oracle... yeah, it will be fine since breaking things in userspace is traditionally a big no-no. Makes me sleep well, just don't use it with raid56 (for now) and you will be fine. That being said, my cold storage is running zfs (or is it zfs on linux? OpenZFS 2.0? I don't even), my notebook is running f2fs...
@NalinKhurb2 жыл бұрын
@@totojejedinecnynick Maybe the wrong place to ask but I would appreciate your input as I only have a very elementary understanding of RAID systems What would be the best strategy for preventing bit rot for multiple terabytes of data? Like what RAID level or file systems would be beneficial for this scenario. Thanks
@totojejedinecnynick2 жыл бұрын
@@NalinKhurb Big problem was highlighted by Wendell - bitrot needs to be detected first, figured out second and repaired/corrected third. So you need system that checks parity on every read/access, verifies it, has ability to figure out whether error is in the data itself or just checksum of data, have enough uncorrupted information to correct said error and ideally do all that seamlessly without user ever noticing. But we live in allegedly real world (or matrix, depending on what pill you choose), things don't work like that. Some systems (Nvidia, mdraid) will happily feed you garbage. Some systems will refuse to feed you data if it is corrupted (zfs). Some systems will do their best to figure it out and feed you what they think should be correct data (btrfs, ceph) - but can you really trust them? I am yet to find a truly fully self-healing system, user needs to be involved in regular maintenance and debugging... For home gamers with just TBs of data, the best solution is to just buy a new backup drive, DOA test it, copy your data to it and unplug it. When your primary data gets corrupted, reach out for backup drive :) KISS. Honestly, there is no simple one-size-fits-all solution for storage - every case is different so I am afraid that if you need an array of drives, there is a lot to learn. Ask around on level1forums, lttforums, reddit...
@NalinKhurb2 жыл бұрын
@@totojejedinecnynick Thanks for the detailed overview. Really appreciate it! I'll start with ZFS :)
@DrathVader2 жыл бұрын
Before even watching the video, I'm guessing it's gonna be Wendell telling me to use ZFS
@williamp6800 Жыл бұрын
This is the way
@derfacecrafter186911 ай бұрын
Why everyone say, use ZFS
@Real_Tim_S11 ай бұрын
@@derfacecrafter1869 "Why everyone say, use ZFS[?]" ZFS is "better" than HW RAID as it allows spanning of a volume across multiple physical devices (of any bulk-storage type), but also allows file/file-system based integrity verification completly agnostic of the storage medium's physical make-up. With a dedicated storage system (with a very perfomant modern CPU) and a decent amount of ECC system memory, it shortens the loop between the file system and the redundancy - allowing for de-duplication, compression, and advanced features low-level HW RAID devices were not able to do like copy-on-write (instead of overwrite) and snapshots making backups and recovery MUCH easier. ZFS can use both reported errors from the physical drive(s) AND do patrol scrubs of the volume (both scheduled and forced) including the repair/re-silvering. It goes without saying that the loss of a drive still can behave like a legacy conventional re-silvering of a replacement drive. ZFS has one shortcoming IMHO, and that is volume resizing is difficult. Once a ZFS volume is created, ADDing additional drives to create more space in a ZFS volume is requires moving the data to a new volume with more space. The solution I use to get around this ZFS shortcoming is GlusterFS using ZFS as a data-store only, not as a native volume. GlusterFS allows for storage-host-to-storage-host redundancy for a volume and allows for taking a ZFS data-store offline without losing data which you can then add back in (larger).
@reaperinsaltbrine521111 ай бұрын
UHUM LOL :DDDDD ZFS wants AT LEAST 8GB for itself...."Lean" "Efficient" Like HTTP /me barfs
@swissrock149211 ай бұрын
object store like s3 or swift
@NickF12272 жыл бұрын
Linus: Holy sh!t this implementation from Nvidia is FAST. Wendell: This implementation doesn’t prevent data corruption. ;)
@SupremeRuleroftheWorld2 жыл бұрын
linus did not knew that when he made the video
@transcendtient2 жыл бұрын
@@SupremeRuleroftheWorld Why people gotta simp? Its a joke that points out the often surface level overview LTT does of tech and their general incompetence stemming from the barrage of videos they release on a regular basis.
@Siroguh2 жыл бұрын
@@transcendtient you can't go to a grocery store and complain they dont sell cars.
@javierortiz822 жыл бұрын
I was just about to comment just that, LTT cares more about selling you water bottles than the actual quality of the review, their approach to reviews is just wrong when it comes to this kind of content they went, brute forced some testing by throwing thousands of dollars in storage on a server and only tested speed, threw two sponsors plus two LTT store plugs in the mix and called it a day. And god forbid you use an adblocker while watching them.
@CheapBastard19882 жыл бұрын
@@Siroguh As a European, I expected Walmart to sell cars in North America by now.
@0blivioniox8642 жыл бұрын
Fascinating how much verifying the integrity of the actual data on the devices is an afterthought. So much attention is paid to RAID's capability to allow an array to survive the loss of a device that we almost completely forget about consistency.
@c128stuff2 жыл бұрын
Consistency was not something raid was ever supposed to solve, it is merely intended to give you enough redundancy to deal with hardware failures and recover from data corruption. Sure, enterprise solutions used 520 byte sectors, but its all in the name.. redundant array of *inexpensive* drives. Traditionally the extra checksumming was totally independent of a raid solution being used.
@marcogenovesi85702 жыл бұрын
@@c128stuff Lol storage vendors redefined the acronym to "independent" drives pretty quickly
@Radovanslav2 жыл бұрын
meanwhile error correcting in RAM is not allowed in most consumer computers because of probably arbitrary decisions.
@c128stuff2 жыл бұрын
@@marcogenovesi8570 Sure, but that does only confirm what I wrote. I'm sorry to look back a bit further than the early 2000s, but history doesn't start there, and nor does raid.
@marcogenovesi85702 жыл бұрын
@@c128stuff dude wtf? The industry renamed RAID back into the 80s, not "early 2000s"
@miff2272 жыл бұрын
And now a word from our sponsor: RAID Shadow Legends
@WillCMAG2 жыл бұрын
"ZFS has Paranoid levels of Paranoia." "My kind of file system." Every time I learn more of about the file system my TrueNAS box runs the more boxes it ticks off and the more I like it.
@dineauxjones2 жыл бұрын
I like TrueNAS. They made making an enterprise class solution like ZFS stupid easy and accessible.
@TheHorse_yes2 жыл бұрын
ZFS is not a blanket solution either, you get enough failures in your drive pool, and, unlike with RAID, _all_ data recovery becomes impossible. This video makes a lot of claims about the superiority of ZFS. Yeah, RAID is not a backup solution, but what's the point because in a degraded ZFS disk pool you're even worse off in terms of data recovery than i.e. with RAID1/10 where even with a bad drive (or multiple, depending on the array), you're still able to recover the data! In ZFS, your data is equal to having been scattered into the wind in billions of pieces.
@hoagy_ytfc2 жыл бұрын
We ran ZFS on FreeBSD. It was a massive disaster.
@nobodynoone25002 жыл бұрын
@@hoagy_ytfc How so? I have multiple production servers using that setup.
@lawrencedoliveiro91042 жыл бұрын
ZFS is to filesystems what Java is to programming languages. It will happily consume all the RAM on your system if you let it. I think it is best restricted to a dedicated storage server that performs no other function.
@Techieguy932 жыл бұрын
I appreciate the detailed explanations in this video! I am working on a storage solution for a small office and had already decided to go with ZFS, and this solidified my reasoning in doing so. It's been a LONG while since I have set up a RAID controller, and as it appears we are moving backwards in functionality (for most solutions). Thanks, Wendell!
@Foiliagegaming2 жыл бұрын
I’m still learning and most of this went over my head but this was a great video! I watch the GRAID LTT video and was pretty stoked about something like that. But big sad. Wendell you are my IT janitor hero!!
@TheGuruStud2 жыл бұрын
@@pieluver1234 bc Microsoft is paying them to spread propaganda that they don't suck...and then the server fails weekly lmao. Linus is and always was a sell out circus actor.
@Mpdarkguy2 жыл бұрын
@@pieluver1234 I mean they mostly have windows clients; you'd think a windows server would make sense Shit happens
@TobyLegion2 жыл бұрын
@@pieluver1234 the one who doesn't know what he is talking about is you. "You'd think a windows server would make sense paired with a client" Keyword WOULD THINK. That doesn't mean that it is the case, but that it is a valid assumption that systems from the se manufacurere work better together. So get your head out of your own ass and stop being so condescending, especially if you keep talking out off your ass.
@99mage992 жыл бұрын
@@pieluver1234 Because the LTT audience is a mix of enthusiasts and casuals mostly centered around DIY gaming/workstation PCs. Most of them don't care enough to go learn about enterprise level hardware because it really doesn't mean anything to them. LTT gets content like that to a wider audience and presents it in a way that someone with no experience could probably follow along. Which is pretty much the point. LTT doesn't present it as "this is what we did step by step and so should you". It's more like "look at these fun toys, I'll explain how it works like you're 5 and give you some examples of why it's better than what we were using" I'd agree that watching LTT for advice on enterprise hardware and config would be insane, but watching it for the fun of them messing around with hardware most of us will never see in person is pretty neato.
@Mpdarkguy2 жыл бұрын
@Malice I was trying to put myself in the place of someone doesn't have a cs degree you know
@solidreactor2 жыл бұрын
Would love to have a video with the topic "Welcome to ZFS and BTRFS - Your Wendell introduction and video guide". Basically helping us migrating from ntfs/ext/etc to these other two.
@nobodynoone25002 жыл бұрын
Hopefully, with the reach of this video, that one is next.
@Arexodius2 жыл бұрын
This is exactly the kind of stuff *_everyone_* should be learning about in basic computer courses! I mean... not even (some) enterprise solutions are really functionally enterprise grade anymore??? Do we really live in such a fast paced world that the only ones concerned with data integrity are massive data collection maniacs like Google?
@TheSikedSicronic11 ай бұрын
This is far from basics lol this is a super advanced topic that many people won’t even begin to understand unless they take a cyber forensic/ computer forensic class where u learn about storage and how it works.
@Arexodius11 ай бұрын
@@TheSikedSicronic I think you missed my point. It's not basic now, because nobody made it basic. If it's important, people should be learning about it. The earlier the better. It's only a matter of making it simple enough to understand, and you cannot convince me that it would be impossible. That's the essence of teaching - making things simpler to understand - and it's been done with many things throughout history. At the very least, people should be made aware of the topic early on and why it matters. That way the ones who are really interested can dive deeper into something they might not have learned about until much later.
@gto1152010 ай бұрын
The future is moving towards SPEED. Data assumed backed up on a incremental and differential level. There is no need for raid controller. And eventually as hardware becomes more powerful the need for ZFS might be behind us. As for the data it will be purely be backed up instantly into the cloud. Basically storage companies are more concern in redundancy in power failure and making more lasting storage devices.
@Gryfang4512 жыл бұрын
Those of us who lived in the enterprise with our massive SCSi, then SAS direct attached, then finally FC SANs, iSCSI,etc..the most fear inducing moment was always when you lost power, lost a drive, and started what, at the time would have been an offline rebuild, hoping you wouldn't lose more drives in the process. One time is all it takes to make you shudder as you are hoping the backup tapes actually have your data! I just decommissioned an LTO6 library! One year of keeping it around for retention. My backups are all on Synology RS units. No more hand carrying tapes offsite either. Hyper backup works well, where various other schemes didn't. I now have backups of backups for two locations, and BTRFS goodness. All because a long time ago I learned that the D in RAID actually stood for dependable, and the RAI was for Really AIn't. Neither are tapes unless you like rolling the dice on year old magnetic bits. Don't ever trust your data to be there, and make sure you have multiple, verifiable backups and actuality test recovery once in a while. Because it may suck to be down for a little while, but if you lost months of data, we call that a resume generating event or RGE in the computer janitor business. And you don't want to be that person.
@wendelltron2 жыл бұрын
generally I just like the idea of a weekly or monthly "yep I scanned all the file hashes and everything was exactly what it was last time, except for the stuff you changed" seems like a basic feature to me in 2022 but some folks here are spurging out saying you dont need that kind of checking. lol
Жыл бұрын
"People don't want backup, they want restore"… ;-)
@mistermac569 ай бұрын
I am an OG as you and I agree with your comments 100 percent. I am a retired IT manager now, but I remember that all too often, the server administrators in my IT shop depended far too much on RAID and only periodic tape backups. I would send out a reminder email every month to them don't depend on RAID and make DAILY backups in off hours. I have to admit that I gave the administrators too much slack and should have demanded daily backups. I had a web server administrator and their server's RAID controller borked all of the RAID 5 disks and our web site was down. And the only tape backup they performed was two weeks old. I didn't single out the administrator in a all hands on deck meeting, but I made it crystal clear if any server administrator didn't make daily tape backups in the future, they would be fired. I also informed them that I would be upgrading the Synology units for more tape storage capabilities.
@ChrisDerichs2 жыл бұрын
This got me thinking, what is the answer these days? While looking into zfs and btrfs and being annoyed that neither are available out of the box on the RHEL clones starting with 8, I noticed a RH's blog talks about using dm-integrity to handle bit rot detection. Might be useful with any of those software raid combinations.
@deViant142 жыл бұрын
SLES
@mckidney12 жыл бұрын
LVM with dm-integrity and parity based on logical volumes instead of md. The biggest pitfall of LVM is storing the metadata, just like on those proper controllers. I also looked for more LVM in the video - but Wendell does love ZFS :D
@ChrisDerichs2 жыл бұрын
@@mckidney1 I did some experimenting. By default using lvm's --raidintegrity option, any corruption I introduced to the drive caused it to completely drop out of the array. I eventually figured out disabling the journal was closer to the behavior I wanted as devices under the raid. LVM also has an option to disable the journal but it seems only by enabling bitmap mode instead. Bitmap mode has its own problem. Still trying to narrow down the best combination.
@etome82 жыл бұрын
Maybe checkout the CentOS Hyperscale SIG, it adds back btrfs
@leexgx2 жыл бұрын
@@ChrisDerichs if your using dm-integrity or dm-crypt (dm-integrity is just dm-crypt without the encryption part) under mdraid so if dm detects any checksum fails it report it as read error so mdraid can use mirror or parity to correct it (if you use btrfs on top of mdraid you get filesystem integrity validation so if dm or mdraid fails you get btrfs checksum errors) Don't know about how LVM handles it
@JohnClark-tt2bl2 жыл бұрын
My how things have changed. I remember not long ago that software raid was constantly shit on.
@marcogenovesi85702 жыл бұрын
if by "software raid" we mean the motherboard integrated RAID or Windows Storage Spaces, it's 100% justified to keep shitting on it. Software raid on Linux/BSD and/or ZFS has always been where it is at
@cdoublejj2 жыл бұрын
Wendell's JUST made a video not THAAAAAT long ago shitting on bios software raid.
@TAP7a2 жыл бұрын
@@marcogenovesi8570 ZFS came from Solaris and illumos didn’t it?
@marcogenovesi85702 жыл бұрын
@@TAP7a Yeah but nobody uses those anymore, thanks to Oracle. 99% of real world usage of ZFS is on BSD and Linux now
@davidmeissner50102 жыл бұрын
Wow, that is like a crash course on RAID. Since I really am an amateur about it, my takeaway is that we could forget the hardware raid controllers and move to a file system based raid platform called ZFS. Question is (provided my understanding is correct), do you have any more details on ZFS and how to implement it? Remember, I am a newbie regarding RAID. Great video by the way. I'll be following more. Thanks.
@lucidnonsense9422 жыл бұрын
I unironically miss the days when you'd buy a decommissioned raid controller and a stack of random 12k drives, from a data centre, then bodge it all into a mutant raid array for your gaming pc... Once a month it would all come tumbling down and need to be resorted, until then your were speed...
@PainSled2 жыл бұрын
Why did my internal narrator pronounce "speed" with falling intonation? 10/10: I have coffee in my nose now
@udirt2 жыл бұрын
@@PainSled 12k drives... either 10k or 15k.
@StefanReich2 жыл бұрын
@@PainSled Speed goes in the nose, coffee in the other orifice. HTH
@freedomspeech95232 жыл бұрын
I have one of those hardware RAID5, with 7.2k drives, working for years. Migrated couple of times to increase capacity. Zero issues.
@morosis822 жыл бұрын
@@freedomspeech9523 that you know of
@GrishTech2 жыл бұрын
I have a good bitrot example. Some of my oldest google photos started to have gray lines in them, then over time it went half gray. This started happening across photos. Confirmed I can see this across multiple devices. Can’t believe this happened at a google data center, where data integrity is critical.
@-aexc-2 жыл бұрын
wait, Google has bitrot too? I've been using rclone crypt to backup all my files to my uni gsuite under the assumption google has a way better data integrity than me
@Dalewyn2 жыл бұрын
@@-aexc- The moral here is the 3-2-1 rule: 3 backups, at least 2 of them on different storage mediums, at least one of them off site. For any mission critical data, only having one backup is equivalent to having none. Redundancy is not a backup, but backups need redundancy.
@GrishTech2 жыл бұрын
@@-aexc- I wouldn’t say it has bitrot, but just some of my photos had it.
@mogoreanu2 жыл бұрын
Not 100% sure about photos in particular, but pretty much all of the data stored at Google has checksums associated with it and bitrot is detected and fixed automatically. I do not think your gray lines are due to bitrot on the storage layer. Maybe recompression artifacts.
@BikingWIthPanda2 жыл бұрын
this didn't happen
@jsebean2 жыл бұрын
Great video Wendell, however as a Btrfs advocate myself, I think it's important to represent it properly so people know the difference. I personally believe Btrfs and ZFS, while both have a lot in common, serve two very different use cases. ZFS is the enterprise solution, Btrfs is the home owner/NAS solution filesystem. First, Btrfs does not support per file or folder level redundancy options. This idea may have been tossed around, but Btrfs redundancy is only on a per volume basis. Not even on a vdev basis, which ZFS has, it has no concepts of vdevs. Not sure I'd personally want per file redundancy option anyway, that could become a management nightmare, but per subvolume redundancy options would be nice. Alas, neither is a thing yet. As for RAID5/6, consider it unusable on Btrfs. Like just straight up don't use it, the mkfs and balance tool will warn users for a reason if they do try to use it. It needs a total rewrite to be usable (which means an on disk format change is likely. I believe Western Digital was working on some fixes but this is still a ways out). While btrfs will indeed "protect" you from bitrot, all it can do is prevent returning bad data, it is unable to repair all issues. As such it's nowhere near trustworthy as ZFS RAID-Z. It's not resilient to the write hole issue at all, and what makes matters worse, it will lie to you which device may be throwing corruption at you. To me, that's as good as not having device stats at all, as you're forced to rely on the disk to actually indicate issues yet again. Apart from the fact the fs won't return bad data, it still forces you to go to a backup or play wack a mole in figuring out which disk is causing the issue. If the parity bits themselves becomes corrupt, scrub won't care to fix it, only if the data itself is corrupt. Additionally, if data does go corrupt, and the stripe that data was in is updated, when Btrfs updates the parity bits for the stripe, the corrupt data will now be reflected in the updated parity, destroying your only chance at recovery. Now to be clear, Btrfs' other RAID levels are indeed fine and great at preventing, identifying and fixing bitrot in all cases, but it does require some knowledge about management, as Btrfs will not automatically rebuild without you triggering it. First, you need to be familiar with how it allocates data to know when a balance is needed, vs a scrub. A scrub will indeed scrub the filesystem, and repair data with redundancy, but data is allocated on a "chunk" basis (usually 1GiB in size on most volumes), and these chunks are what have a redundancy policy (not really the volume overall). It is this people need to look out for. First, If the array becomes degraded, btrfs will not automount without user intervention because any new writes to the fs may not be redundant (It's impossible if you have the minimum disks required for a profile, as it's degraded, otherwise it won't be balanced properly). If the allocator is unable to allocate a chunk to store the data in, in a redundant fashion, it will instead create a chunk with the "single" profile. If the filesystem was already mounted, and it can't satisfy the allocator requirements to create a redundant chunk, it's forced read only (this is why people often suggest RAID1 users on Btrfs use 3 disks instead of 2, otherwise high uptime is impossible with it). Now, even after a drive is replaced on Btrfs after it has been mounted with the degraded, you need to keep an eye on the chunks as they're indicated in it's management tools to see if you need a convert any non-redundant chunks with a balance to restore full redundancy. However, Btrfs it is quite innovative in other ways, so home users in particular shouldn't write it off too quickly. It allows mixed size disks to be used redundantly. If you have 2x2TB + 1x4TB disk, in a RAID1 configuration, it will have 4TB usable space. You can still only lose one disk without losing the filesystem, but the data on the 4TB disk is balanced between the 2x2TB disks. It also supports duplication on single disks for some added bitrot protection in those use cases, along with RAID10 (and RAID0 if you don't want redundancy, I guess). To get mirroring across three, you would need to use the RAID1c3 profile, which would make 3 redundant copies across the three disks, at which point you'd only have 2TB usable space in the above example, but with the resilience of being able to lose two disks. There's also a raid1c4 option if one wants it. Finally, shadow copy is possible with it, and Btrfs is a great open source alternative to what Unraid provides when it comes to flexibility with disks. Btrfs also supports converting from any RAID profile on the fly as you like. Wanna go from RAID1 to RAID10? Easy, just add the disks and rebalance to raid10. If you have a RAID10 now, and a disk fails, and you're unable to replace it right away, you can rebalance to a RAID1 to get redundancy back without needing a replacement disk right away. All this stuff can be useful in some cases, and if RAID5/6 ever does get reworked, anyone running it's current stable RAID profiles can easily convert to RAID5/6 later. Now, as cool as this functionality is, it's a bit of a niche, it's more a thing home operators would usually be concerned with, not enterprise users who would just install another disk or configure another vdev. It is overall a great choice for homeowner/NAS uses cases, and it's built into Linux in just about any distro. I use a 5 disk RAID-10 array with it here and it has been serving me well. It's also a spectacular desktop filesystem for those who run desktop linux -- I wouldn't choose any other, as snapshots, compression, and incremental backups with send/receive is too much to pass up. There's a reason distros like Fedora and OpenSUSE use it by default on their desktop flavors, it really is great, but I just wanted to clear up people's expectations with it so they pick the right tool for the job ;) BTW: For anyone who cares to have a true Linux alternative to ZFS, keep an eye on Bcachefs ;) ... sorry for the long winded post, I could make a series of videos discussing Btrfs because it is my personal favorite Linux filesystem lol
@AI-xi4jk2 жыл бұрын
Thanks for sharing!
@YaroKasear2 жыл бұрын
I used btrfs off and on for years. Every single time it silently corrupted itself to the point it would stop working. No other filesystem would do this. I wasn't even using any RAID. I am watching bcachefs, but it's been a few years since that went anywhere. Personally, I would still choose ZFS over btrfs even on NAS setups. ZFS beats the snot out of btrfs on reliability and their devs definitely seem more competent.
@leexgx2 жыл бұрын
@@YaroKasear (sounds like a hardware problem not filesystem) if using it on a SSD make sure kernel is above 5.15+ or make sure dup for metadata is set (btrfs balance start -mconvert=dup /mount/point if already created or -m dup at manual filesystem creation) before 5.15 or last 10 years btrfs has defaulted to single for metadata if ssd/non spinner was detected at filesystem creation (due to assumptions) witch is bad as btrfs without dup on metadata is bad as a single error on the metadata can Hose it without a chance to auto correct attempt Another note if your using dm-integrity or dm-crypt in front of mdraid your raid now has self heal capability (dm-integrity or dm-crypt checksums all 4k blocks by default and any errors get passed up the layer as read error so the raid can deal with it, has a 20-30% write penalty depending on loads) Not had problems with Synology or netgear readynas implement way of doing it (assume asustor has done same thing as well) uses btrfs on top of md-raid so don't have any of the problems of raid56 of btrfs (they have modified btrfs and mdadm so that btrfs can talk to mdadm and issue a parity or mirror repair attempt if btrfs receives valid data mdadm writes the correct data back to the array (when a scrub is ran it first runs a btrfs scrub Once finished it then runs a raid sync so no way stored errors can replace parity as it would normally do) one key important thing on Synology is you must make sure you have ticked the checksum/integrity box when creating the share folder as it has a habit of not pre ticking it (need a + nas and filesystem chosen needs to be btrfs at volume creation) netgear readynas enables it by default unless unselected by user (or CPU is arm type or old CPU model you get a warning when trying to enable it, mostly affects write speed on older devices) Asustor allows btrfs based on model number (and you must tick the snapshot box when you first setup the nas or it uses ext4 as default filesystem, whole nas has to be factory reset to change it)
@wishusknight30092 жыл бұрын
BRTFS has given probems to at least 2 people that I know who have tried it. To me it seems gimmicky and pointless. Best practice is just use symmetrical disks and ZFS2 on Freenas. over 10 years use for me and never an issue, even when drives died. Its always been that "just works" solution.
@eDoc20202 жыл бұрын
@@leexgx I was wondering why there wasn't a way for btrfs or other filesystems to interact with the RAID layer. If these NAS companies have made it happen, have they released their modifications for usage by the general public?
@mcwild11 Жыл бұрын
I would love deeper dive into zfs and particularly synology implementation of BTRFS. Thx, great video
@JeffGeerling2 жыл бұрын
B-b-but what if you only have a Raspberry Pi? Hardware RAID is like 10x faster if you need parity on a Pi.
@chinesepopsongs002 жыл бұрын
Then use a card and drives that are at least 20 years old on the PI. :-)
@JeffGeerling2 жыл бұрын
@@chinesepopsongs00 The ironic thing is that only the newest Broadcom/LSI RAID cards will work on the Pi, because the older cards required features that aren't even present on ARM implementations to work. And the newest drives still need some modifications to work on the Pi since it's PCIe bus is a little funky.
@roysigurdkarlsbakk38422 жыл бұрын
I know you love Raspberry pi and so do I, to some extent. When it comes to storage, I'd rather use a good old computer with a bit beefier CPU and memory bus, preferable with ECC memory etc. Yes, a pi can do a lot of stuff. I've setup some 40 infoscreens at work with them, running octoprint on them, using them for various IoT related stuff including home assistant, but I don't use them for storage. I have some 12 drives in my home server, a fine combination of SSDs and HDDs of different sizes and makes. I'm quite sure you would be able to use a pi to control that as well, that's for sure, but it won't be particularly fast.
@SLLabsKamilion2 жыл бұрын
"Hardware" RAID on a pi4? With what HBA? I seriously doubt most of the broadcom LSI chips' control interfaces are going to be available to you on an ARM platform. Adaptec will have the same problem unless you can find some interesting qemu tricks to emulate the EFI/nonEFI Option ROM interfaces. So seriously, Jeff, what "Hardware" RAID are you talking about? Just use btrfs. Every btrfs gigablock has enough metadata inside of it for tree recovery. Or go full paranoid with ZFS, but I wouldn't recommend it because Oracle's chased after users of their patent stack with a courtroom in the past, and is insane enough to demand money from Java users and anyone who wants USB3 in virtualbox these days. ZFS is tunable, you don't 'need' a ton of ram for L2ARC and ZIL if you're just running a 'small' pool of consumer data (~20TB) and not some crazy data warehousing database or machine learning models.
@eugenesmirnov2522 жыл бұрын
I've done Minio storage with 6x2TB drives, with bit-rot protection. It's not so hard
@EpicWolverine2 жыл бұрын
StableBit DrivePool also has file/folder duplication policies that can be inherited and whatnot but isn’t a full native proprietary file system. It just makes a virtual drive on top of NTFS. Very cool stuff.
@raulsaavedra7092 жыл бұрын
Awesome video! Besides the detailed problems explained here for RAID 5 and even 6, I remember quite a few years ago Dell kind of officially discouraged the usage of at RAID 5 for any business critical data, and that was because the likelyhood of a second drive failing or of the appearance of an uncorrectable error precisely when trying to rebuilding the array after a first disk failure was much higher than people typically assumed/understood.
@DeltaSierra426 Жыл бұрын
Yeah, I remember being surprised when I started to see RAID 5 discouraged. HPE (then HP) said the same around the same time. I still see people recommending RAID 5 in IT forums and such and just shake my head. Even RAID 6 is loosing support. Honestly, I like my RAID 10 to the maximum degree that it still makes sense for the scenario.
@stevedixon9212 жыл бұрын
Years ago I was shocked to discover that almost no part of storage or software was responsible for data integrity, with each part assuming someone else was doing it or that you had backups. HDD = no (silent bit rot not detected), File system = no, Application = no. The good news is the next gen of file systems (REFS, ZFS, etc) are supposed to fill in part of this deficiency. Seems that no one wants to take on being responsible for data integrity. Logically the drive should be responsible for at least detection and reporting that it cannot return exactly what was stored. All drives should silently be storing parity data for each sector *by default* to assist the file system in recovery options if nothing else (akin to the 520 byte sector size thing mentioned in the video, just updated to handle 4k sectors).
@leexgx2 жыл бұрын
Refs requires integrity to be set to true (default is false so by default so does bugger all to protect data even if your using mirroring unless you set integrity to true via powershell) and worse the enforce part (true by default and ignores per folder setting when set to false, just sets new created files to true) if integrity is set to true as well it will delete the file if there is a single uncorrectable bit of data (just get an event log and file disappears)
@leexgx2 жыл бұрын
Drives do have built in ECC, I have rarely found 4k physical (512 logical / 4k physical) drives to get hung up on a ECC block it can't repair (if it hits a read error that ecc can't correct it just returns a UNC/red error quite quickly ) older 512/512 would get hung up on trying to correct a read error it any going to be able to correct as to why TLER/ERC was made so they give up in 7 seconds so Raid can deal with the problem The difference is that the 4k only needs to do a single 4k ecc check on 512e/4kn and has higher chance of ecc repair vs 8 ecc blocks on old 512 physical drives
@666Tomato6662 жыл бұрын
@@leexgx I had a crib death of a 4k physical 512b logical disk that wouldn't go through simple file system format, it would just spit out garbage when trying to read the sectors just written you were just lucky: storage is not to be trusted, they sell on performance numbers, not on hard to prove and hard to show reliability and resiliency
@Hostilenemy2 жыл бұрын
My first Raid Z2 server had 32TB of storage, my next one had 60TB, and my current one has 180TB. Never lost a file or had one corrupted. Trust in ZFS.
@Anonymous______________2 жыл бұрын
Unfortunately RHEL and CentOS do not have native support for ZFS and require a kernel module. Which, If you know anything about dkms, poses the potential that ZFS volumes won't mount on boot up after you apply any kernel updates. Generally, any respectable cyber security oriented business uses a long-term support Linux distribution (i.e. RHEL, CentOS, or Ubuntu LTS, etc). Once ZFS enters kernel mainline status (most likely never) more Linux focused companies will replace ext4 or xfs due to zfs's significantly better scalability.
@space_cowboy0072 жыл бұрын
@@Anonymous______________ all the more reason to choose a LTS distro that actually supports it
@aeonikus12 жыл бұрын
@@Anonymous______________ I literally detest every Red Hat/Fedora/Mandriva etc linux flavour since late '90s. Slackware is my best Linux friend :)
@georgen98382 жыл бұрын
@@Anonymous______________ Isn't Btrfs mainline?
@stevedixon9212 жыл бұрын
@@Anonymous______________ For those curious: I believe the reason ZFS will not be mainlined into many of the 'enterprise' distros is a compatibility issue between GPL and whatever the other one is (escapes me at the moment). I think the core was whether you had to make any custom code changes available to the public versus being able to keep said changes private (though I could be way off base on this one, it exceeds my knowledge depth). I think ZFS started in the 'unix' realm (with BSD and the like) so it has different open licensing rules. And yeah, anything than can just stop working after updating your system is not going to make it far in the enterprise.
@frenchpressfinance2 жыл бұрын
I'm retired from the IT game and somehow your video popped up randomly via the YT algo. I may be retired but I still keep up a little for my own purposes and this was a GREAT explainer!
@Banner19862 жыл бұрын
I still remember working on rebuilding arrays from hex back working at LSI, which was only possible BECAUSE the 520 byte sectors. Ah, the stories... every large company used the same storage HW, all white labeled LSI stuff, and I miss being in the thick of it there
@Phynellius2 жыл бұрын
Raid is not a backup, but your backup may use raid
@RN14412 жыл бұрын
Once again we get bitten by marketing playing fast and loose with implying that their products solve problems for us that they don't. The storage media space seems absolutely crammed full of this type of thing the past few years. WD hiding SMR in their NAS drive series, NAS units themselves being open to cloud vulnerabilities even when you try to turn off all their stupid cloud extras, SSD manufacturers substituting different controllers, drams and flash chips on their shipping drives compared to what they sent out to get good reviews (always downgrades) and reliability numbers that I wouldn't trust let alone rely on. Now I learn that some of the RAID volumes I've set up to prevent decay of precious files and memories are probably not protected at all against decay or loss. LOL.....
@realjoecast2 жыл бұрын
You mention precious files and memories.... I would assume not just setup on your RAID Volumes :-)
@DergEnterprises2 жыл бұрын
My NAS machines have their gateway set to a blackhole.
@ashkebora72622 жыл бұрын
Nah, it's in _all_ industries. Unless you're getting a canned product that already has fully defined spec sheets to reference, it's all fluff. Even in the software industry, _way too often_ does marketing _literally sign contracts_ with customers for features that the product doesn't do, yet. Then the marketing guy wonders why he and the company looks bad when the customer receives a rushed implementation... Until companies start getting in trouble for "not _technically_ false advertising", it's going to keep getting worse.
@mnemonic6047 Жыл бұрын
as an apprentice in IT, i created myself a hardware raid, this video updates me on the current topology of hardware/software raid on what to expect in enterprise, thanks Wendell!
@Robert_DROP_TABLE_students--2 жыл бұрын
that's two LOTR references in one video. did kreestuh write this script?
@nayphee2 жыл бұрын
The thing with ZFS is that it needs routine scrubbing to check for bit-rot/cosmic ray damage. You can't wait for a disk to fail before you discover that bit rot has also wiped out enough parity to recreate the data on the replaced disk
@ahrubik2 жыл бұрын
A common practice in the storage industry is having RAID scrubbing set up as a scheduled task. It all but resolves this risk vector.
@Loanshark753 Жыл бұрын
You also need triple checksums
@mrpeterfrazier Жыл бұрын
Don't act like you know what ZFS is. Have you considered wut knowing is? Like if you want to see Corsica maybe you start seeing with eyes, rather than via disease? ...else it is like the pot calling the kettle black. Like if you're engaged in espionage, the pre-condition there is that your country has some agenda... then you have to figure out what a country is, and what it means for that entity to have requirements that could require espionage. You're getting ahead of yourself without a head... that's being a knit-wit.
@Yves_Cools2 жыл бұрын
@Level1Techs : this is a very insightful video Wendell, I wish I could give you 2 thumbs up but youtube only allows me to give one.
@Jules_Diplopia2 жыл бұрын
I am a long time out of the IT world... thanks for a quick update on the current state of RAID. I had wondered how it worked on NVMe SSDs. Clearly it doesn't. Glad to hear that ZFS is still going strong.
@SmallSpoonBrigade2 жыл бұрын
I kind of think this video is like 2 decades late to the party. FreeBSD had Vinum at that point and ZFS was starting developed thing. At any rate, software raid has been better than hardware for at least a decade at this point, and probably more like 15 years.
@martinhovorka6911 ай бұрын
Current MBs support RAID at BIOS level, you just need to load the appropriate drivers when installing the OS. I run RAID 0 on a PCI 5 SSD and apart from the higher sequential reads the performance hasn't changed compared to a single SSD, RAID has little practical significance from a performance standpoint.
@Jules_Diplopia11 ай бұрын
@@martinhovorka69 But the whole point of RAID was that it protected data if a single drive failed. So for RAID to be useful you would need at least 3 drives, be they HDD or SSD so that if one fails the other 2 can recover the data. It may well be that things have changed since my time, but I would have thought that that basic principle remained.
@fat_pigeon2 жыл бұрын
16:22 Correction: you're confusing Btrfs with something else. I'm pretty sure that Btrfs sets the allocation profile at the filesystem level and doesn't let you set it per file (or even per subvolume). You can only set it separately for data vs. metadata. It likely wouldn't be difficult to add that feature, probably using an extended attribute similar to how compression is set per directory, though you'd need to run a `btrfs balance`. It definitely doesn't allow 5-way mirroring; they implemented 3- and 4-way mirroring by adding RAID1C3 and RAID1C4 profiles a couple of years ago. However, I've read that the upcoming Bcachefs *does* let you specify replication level per inode (for data, not metadata). See the `data_replicas` option in the "Principles of Operation" doc. Maybe that's what you were thinking of.
@AegisHyperon2 жыл бұрын
DrivePool lets you set replication at a per-file level (Windows only)
@AgentOffice2 жыл бұрын
@@AegisHyperon i like drive pool
@NullStaticVoid2 жыл бұрын
Even 7 years ago when I worked for NBCUni I had a hard time finding raid specific drives for our HP Proliant servers. SAS drives are almost all going to be raid drives. But 2.5" SATA seems to be going away. Nobody is making anything bigger or faster than what they sold 5 years ago. And RAID specific drives are getting harder to find.
@marcogenovesi85702 жыл бұрын
Afaik LVM RAID has integrity mode to do data integrity checking and autocorrect, also dm-integrity can be used under mdadm raid to do integrity checking and report the error to mdadm that will fix it from parity or the mirror
@tad20212 жыл бұрын
Minor thing, the super caps on (all?) raid cards are for "flash-back". The caps power the cache module to store itself to local flash instead of powering the DRAM for the duration of the blackout. Gives blackout tolerance of the effective powered-off storage duration of NAND. A fresh LiPo on battery-backed might have maybe a day at most.
@wishusknight30092 жыл бұрын
My old HPE P400 would promise about 50 hours. And the most I had power cut to a server due to grid loss was just under 30ish hours and it held up. I don't know if it would hit the promised 50 or not but it certainly held up for a full day. And that was with the 512mb BBWC.. the 256MB module is supposed to last somewhat longer. My LSI controller though only promises 24-36 hours depending on cache size. Thankfully my new array is based on the HPE P420 that is flash backed. Being in an area with bad power and frequent blackouts its nice to have. And my UPS only goes so far. When I am not home, and the server cant gracefully shut down. The raid has never suffered any issues on my VMhead. That said my main storage server array is ZFS2. So its never an issue regardless. But ESX does not have native support for it, so I use the p420 for that.
@bmiller9492 жыл бұрын
Wow, things have changed a lot since I was in school in '96. I use to run a hardware raid 5 system at home office. It was truly hardware raid as the card cost me over $500 at the time.
@YolandaPlayne Жыл бұрын
Now I think the best solution for the home office is a Synology system.
@dolfinmicro2 жыл бұрын
This was a great video. Just when I think I know all I need to know about RAID, you set me straight. Thanks!
@stevec00ps2 жыл бұрын
Oh this explains why I had to go through a load of hassle to reformat some 520 byte formatted SAS disks from an old EMC SAN to get them to be 512 byte format to use on a normal PC!
@marcogenovesi85702 жыл бұрын
yeah they use a bigger block size to store checksums and stuff somewhat "in hardware" kind of
@kernkraft-23542 жыл бұрын
It was surprisingly more difficult to fix than it needed to be. I had to force my HPE RAID controller into HBA mode against it's will. Then you start the operation and just hope for the best. I was getting 600gb 520byte drives on ebay for $8. sg_utils FTW, and now nvme-cli for all your low level hardware magik
@mitch79182 жыл бұрын
Wendell... This was the best explanation or discussion on this subject I have ever seen. You are truly, seriously knowledgeable. Been watching you since the start of the Tek days and have always had great respect for you. Much love and long live ZFS!
@gorillaau Жыл бұрын
I have been trying to figure out how ZFS handles different sized disks. Supposing you have 2x 2TB drivers and 2x 1TB drivers, so 6TB of raw storage.
@theundertaker59632 жыл бұрын
Amazing video like always. Can we please, please get a series on ZFS now to allow more folks to better understand, and implement it to save everyone from bitrot? You are the only one who can take the mystique out of ZFS and make it approachable to technically apt, and the power users alike.
@kevinlassure62142 жыл бұрын
Meanwhile in France, 90% of companies use RAID 5 over 5/6 disks if not more and bosses answer: it cost less money and we have backups anyway. Or even, as i also heard: i never saw issues yet with raid 5 or 6. Beceause our country is from stone age and want to spend just enough for something to work even if it isnt really reliable
@MrGsteele2 жыл бұрын
What you are saying applies equally well to all forms of storage. That is, unless you read back what has been written to a storage device - of any description - and compare it with what you attempted to write, you cannot be sure that the data has been stored with any integrity. All storage devices rely on some form of qualitative protection - in the form of parity, or CRC, or LRC, etc. - but those are statistically (and that's important) validated protections. That is they reduce the readback error rate from the natural error rate to a lower error rate by providing a mathematical way to detect and correct bit errors - up to a point. That is why error rates are specified - 1 bit error in 10^14 bits, for example - it's a foregone conclusion that recording media are imperfect, and you are relying on the capability of the EDAC codes to keep the error rate that low. Not error free - just low. RAID, however, addresses a different aspect of storage system behavior - the ability to retrieve data when a catastrophic hardware failure of one or more drives happens. Without RAID, a catastrophic failure may destroy your data, obliging you to recover it through auxiliary sources like the last backup, or forensic recovery from a disassembled drive. An array contains an error recovery mechanism that allows you to continue operation while the failure is operative. The array dynamically reconstructs the data stream using that redundant data, providing operational continuity. Advanced arrays isolate the failed drive, reconstruct the data stream, and may even call in a hot spare and begin rebuilding what was on the failed drive or drives to permit resumption of fully-normal (non-reconstructed) operation. Operational continuity has business value. So does RAID - just as any redundancy does when its purpose is to provide operation through failure. When you say that "drives do not honestly report their bit errors" it may well be the case that a RAID array does not employ a diagnostic read that provides information on whether the data read passed the test of conforming to the on-controller's assessment that the data did not require error correction, or required the use of an on-controller error correction calculation in order to (if successful) reconstruct the data originally written. A properly implemented array, however, will do so in order to detect at the earliest possible point any incipient failure and compare it to the array threshold for error rate or error growth in order to isolate the drive experiencing the problem - i.e. to proactively move a drive out of operation before things get to the point that error correction can no longer be relied upon for faithful reconstruction. Incidentally, there is no such thing as "RAID 0" - the name itself is an oxymoron. What is billed as "RAID 0" lowers, rather than increases, the reliability of the storage system by requiring both devices in the stripe to work in order to provide the data; it is a performance-oriented concept that is misnamed - misnamed because there is NO redundancy, which is the first word in the RAID acronym. A card that performs data reconstruction that has a maximum 16 GB/s I/O capability, moreover, does not limit the read rate to no more than 16 GB/S; realistically, errors are "bursty" - that is, you don't have 16 consecutive gigabytes of data that is rife with errors needing correction in real time - in fact, real time is in and of itself a misnomer, since I/O is asynchronous and, perhaps even more importantly, interrupted constantly by rotational and seek latencies that dominate the I/O equation, as well as buffering and I/O driver turnaround time, memory address allocation, bus handshaking, and on and on. Most performance data specifications relate only to a maximum theoretical rate derived from the mathematics of bit density, rotation rate, and interface electronics - not performance in practice. After all, the operating system must respond to I/O complete interrupts, interrupt processing time, buffer management, etc. No hard drive in existence, rated at 100 GB/S transfer rate, will transfer 100 GB in one second into a computer system. It's a fantasy. Another nit is that neither 512 byte, 520 byte, 1024 byte, 2048 byte or any other byte count sector size (the latter sizes were commonly used on optical disks) have that many bytes in a sector; they have that many USER bytes in a sector. They also have additional bytes, located in each sector, dedicated to EDAC codes that are calculated by the drive controller in real time as the user data is being written, and then appended to the user data to serve in the readback process. In the early days of optical disk write/verify, for example, those EDAC codes were read back and verified against the data during the readback process to ensure that the data was written readably. If not, the data failed verification, and required rewriting for data integrity. More extensive codes were used as 4.75 inch disks became widely employed to reduce the necessity for verify, compared to their 5 1/4, 8, 12, and 14 inch predecessors. It is definitely true that an improperly implemented, distributed parity, hot-spared, auto-rebuild array may be out of touch with best practice, and that attribute may be applied to any given vendor's product. But RAID, as a tool designed for a specific characteristic of higher uptime and operational continuity through failure, with admittedly - in fact, acknowledged - degraded performance during rebuild, still has value in 2022.
@mishasawangwan66522 жыл бұрын
finally some knowledge
@narcoduck2 жыл бұрын
Two weeks ago I bought a Cisco UCS 240 M3 that came with LSI MegaRAID SAS 9271-8i card PCI card. I setup the RAID 6 and set to "Check consistency" once a Week thinking that would work like ZFS raidz2 and use the parity data to correct a possible corruption. After watching your video I decided to read the manual provided by Cisco, that what is there in the consistency check section: "If the WebBIOS CU finds a difference between the data and the parity value on the redundant drive group, it assumes that the data is accurate and automatically corrects the parity value." This is insane and not at all what it should be doing. Thanks so much for this video!!!!!
@kingnick62602 жыл бұрын
It appears GRAID Tech recently released the SR-1010, upgrading their Nvidia T1000 to an A2000 which includes ECC memory, while also supporting PCI-e 4.0. They've also released v1.2.2 of their software which touts automatic data correction, without any additional info. I've reached out asking what exactly that entails, while also referencing this very video and if any of the concerns regarding bit rot were addressed within the SR-1010 on v1.2.2 software. Will also keep the Level 1 Techs forum updated.
@TvistoProPro17 күн бұрын
For MD on linux, there IS a tool out there that does the scan for Raid6 properly and fix the parity and/or data depending on where the error is (which it can usually figure out based on a complex computation). You have to replace the scrub tool with the new one after compiling it.
@BulkybearOG2 жыл бұрын
That's why I set my expectations. I use RAID5 mainly just to pull drives into an array and have at least a little bit of safety against a defective drive. For data integrity I'll back up that whole array to an external server. Nothing's perfect but this has been good enough for me for like 10 years now.
@lozboz632 жыл бұрын
But you miss the point of this video, what is to say when you read the raid5 data for backup you get what was written?
@SaroG2 жыл бұрын
ZFS actually does something similar in terms of redundancy policies at the dataset level with the 'copies' property: zfs set copies=2 pool/dataset
@vorpled2 жыл бұрын
Are there any smaller NAS-type commercial/prosumer products that do this right (with proper parity checking)? Huge thanks for this! Growing up in the 80’s I always presumed that quality hardware cards were going to be faster and more reliable than any software raid solution - and that they’d be doing it properly!
@ryanwallace9832 жыл бұрын
TrueNAS Core and TrueNAS Scale both implement ZFS Core is based off FreeBSD and Scale is Linux
@proxgs77032 жыл бұрын
I would say Synology. They use BTRFS on top of md raid.
@sobertillnoon2 жыл бұрын
I guess I missed the boat. The one thing I wanted raid for is gone. Gonna convert all my arrays to 0. Might as well get as much space and speed as I can seeing as the only thing on offer. Raid is not a backup.
@padraics2 жыл бұрын
Yo you're looking at some crazy speeds. I just did my first build nvme drives, needed four fully independent hosts (ESXi 7), using local storage. Used megaraid 9560 cards and the performance of 12 kioxa 3.2tb drives is lovely. Had no idea btfrs had those features. Good info!
@seanthomas29062 жыл бұрын
Watched the Linus video about this and thought wow!. You've brought me crashing into a wall. Great insight and well executed .
@Niarbeht2 жыл бұрын
So one idea I've had kicking around in my head for years is the idea of a "hardware raid" ZFS setup, where you've got a parity computation accelerator card that has a direct connection to the drives. You could set up a pool that's entirely local to the card. Optimally, you'd be able to inspect it using the normal zfs command-line stuff (zpool status, etc). However, all processing would occur on the card itself. Scrubs, parity, reslivering, everything would be local to that card. In essence, like having a ZFS NAS box, but it fits into a PCI Express slot and publishes itself to the ZFS implementation on your system. Probably a pain in the ass to engineer, though.
@BattousaiHBr2 жыл бұрын
i don't understand the necessity of that in a storage system which is going to have a CPU anyway, so why not let the CPU do it instead, which in a storage system isn't being used for anything else anyway.
@harryhall4001 Жыл бұрын
@@BattousaiHBr Because lost of places won't use dedicated storage boxes. That's why they have hardware RAID in the first place, because it's supposed to be transparent to the OS such as ESXi.
@BattousaiHBr Жыл бұрын
@@harryhall4001 that's not why people (historically) used hardware raid at all, it was because raid cards had batteries to make sure in case of a power outage that all data would finish to be written to disk to avoid data corruption. if power outage corruption is not an issue, people tend to avoid hardware raid entirely.
@petripuurunenАй бұрын
What differs you channels from the others - is the pure professionality. Thank you.
@utp2162 жыл бұрын
That new card reminds me of the PhysX add in cards back in the day. Just something to help the math get done faster. Then it was turned in to a software solution…
@CMDRSweeper2 жыл бұрын
PhysX was never turned into a software solution, all they did was allow it to be accelerated by a GPU rather than the old PCI PPU. The PhysX that were capable of running on the CPU is the same simpler software PhysX that you see a lot of games implement today after Nvidia bought it out, but very few of them allow you to accelerate it on a GPU which is what the old Ageia PPU (Physics Processing Unit) did in a lot of titles like Mafia 2 and Ghost Recon Advanced Warfighter or Borderlands 2.
@toxy35802 жыл бұрын
@@CMDRSweeper rip the physx effects
@housewares2 жыл бұрын
“Array of independent” is itself a redundancy of terms; originally the “I” meant “inexpensive” which of course, is now a historical concept
@rudypieplenbosch67522 жыл бұрын
Wow, what a story, glad I am using ZFS. But these HWraid suppliers should go bust with delivering their nonsense to the enterprise market.
@insanemal2 жыл бұрын
If enterprise RAID solutions were all this bad they would. Hint: They aren't. Source, I worked for a major raid vendor. Their stuff would have picked up this issue on read because it does read verify on all operations.
@rudypieplenbosch67522 жыл бұрын
@@insanemal Probably you are one of those "IT professionals" not really understanding how it should work, like L1 mentioned. But his simple test by modifying the data and without changing the checksum clearly shows the absolute nonsense implementation of the HW raid vendor. When he mentioned a vendor just rewrote the parity block only, makes the whole HW raid really a big joke. Unless you provide us with something that contradicts L1's findings I can not take you seriously. As a sw architect, myself I have dealt with a lot of "IT Professionals" who basically told me on more than one occasion, to just format my drive to "solve" certain application issues, etc, when I wanted an SSD for my workstation, this of course was out of the question, due to the evil that would descent upon me daring to use such an unreliable drive, what a joke these guys were most of the time and I worked for several big companies. Even with my limited experience with servers, I could probably sell myself as an "IT professional" since the standard seem to be pretty low in that kind of business. Show us something that contradicts L1's findings.
@insanemal2 жыл бұрын
Dude I work in Supercomputing. I worked for DDN. I know how it works. This particular product is a hardware RAID Accelerator. It's not a hardware RAID controller. Of course with the way this works read verify won't work. The card isn't in the read path. That doesn't mean it's true for all RAID controllers.
@rudypieplenbosch67522 жыл бұрын
@@insanemal Well than, even with super computing, there are people that are blind to obvious facts, something the latest covid debacle has shown, its morons galore on every level everywhere, in any business you will find them. It is a clown world we are living in, sad but true. But I am sure your HW vendor is very happy with you 🤗
@jonathanbuzzard13762 жыл бұрын
@@insanemal To many people who have not a scooby about enterprise storage come out on the internet. If you are not spinning hundreds of hard drives then you should STFU IMHO as you are just making a fool of yourself. The likes of ZFS/btrfs are good if you have more than 10 drives and less than say 50. If you have say 200+ hard drives then ZFS/btfs is a solution for idiots.
@aracrg2 жыл бұрын
17:25 although in ZFS one can't specify redundancy at a file or directory level, it can be done at a file system level by setting the copies property of the file system to the number of copies you want that file system to use. In ZFS file systems are almost as light weight as directories.
@666Tomato6662 жыл бұрын
but you need to assign the space a priori to the pool with the given redundancy level, in btrfs it's totally dynamic, with a 1GB granularity; so you don't have to guess how much space you will need in the no-redundancy pool vs the triple redundancy pool
@Dan-Simms2 жыл бұрын
As someone who hasn't used RAID since the early 2000s, this was wild to me. Very interesting, good to know.
@KellicTiger Жыл бұрын
Heh. So in the early 2K's I was doing Dell repairs in the field and I ran into my first instance of RAID 1 corruption. One drive was corrupt and failing and was writing to the other. I went onsite to replace the drive, customer was pissed that the other drive had corrupt data.....this was the beginning of my mantra that I have learned and have said many a time: RAID is not a backup. It doesn't matter if its a mirror or RAID 6, corruption can happen. This is why ZFS is such a fantastic thing and I REALLY wish my company would migrate over to that crossed with erasure coding.
@opopopop62862 жыл бұрын
Into computers for multiple decades at this point. Yet very little experience with RAID. Your talk is so super technical that even I can only understand a little over 2/3s of it..so you must be a super expert pretty much, and teaching it all right. This much I can gather/surmise :)
@TheOtherNEO2 жыл бұрын
Having had the experience of an early Adaptec software RAID controller disaster where the Windows drivers corrupted the drives, not sure I’d take the risk again. Physically the drives were fine but all data lost as it wrote junk to the drives.
@az09letters922 жыл бұрын
A little secret for you. All RAID is "software RAID". It's just a matter of where the software is running.
@jasonabettan57782 жыл бұрын
Any thoughts on a part 2 where windows storage spaces and ReFS fits into this?
@Raletia2 жыл бұрын
+1 I'm curious about this too! My NAS is my old fx8350 system with Windows 10 pro and storage spaces, with 12 hdds, a mix of 3, 4, & 6 TB. Backed up with Backblaze. I like storage spaces flexibility for upgrading drives one at a time. P.S. if you didn't know, even though creation of ReFS was removed from win 10 pro a bit ago, if you format with it using an old build or some other means you can still use it just fine. Or you can get creative with 'upgrading' to pro for workstations, to get back functionality they stole from you.
@katbryce2 жыл бұрын
@@Raletia My preferred approach there is to put FreeBSD on a Hyper-V VM. Take the storage pool drives offline in Windows, and assign them to the VM. Pick them up in FreeBSD and format them as a zfs pool, and share it with Samba. Connect to it in Windows as a network drive.
@wishusknight30092 жыл бұрын
@@katbryce I even simplified my setup by running a separate Truenas machine bare metal, and my hypervisor on its own box using an HBA local store for VM's.
@Raletia2 жыл бұрын
@@katbryce Is there any way to use ZFS with a mix of drives? Everything I researched(a while ago, admittedly) indicated you needed fixed drives of identical size, and 1GB of ram for every 1TB of space, and you could not add or remove drives without destroying and remaking the entire pool? I'm using storage spaces for two main reasons, really. The flexibility, and Backblaze(personal home thingy) does not let you backup from network drives, only local. I'm not as experienced with Linux/FreeBSD, etc, so that's also a factor, but if given the choice, I'd rather not have to rely on Windows for my NAS. Right now that's literally all that PC does, it's headless and in my bedroom next to my router. I just cannot deal with having to rebuild the entire thing just to upgrade a drive. I don't have enough space anywhere else, not to mention the time required to shift so much data(~12TB). It's also super anxiety inducing restoring from backup when your local copy is gone. I've been getting used Hitachi/HGST enterprise drives for cheap, and slowly adding/upgrading one at a time. There's a particular seller I trust, had good luck so far, and they do honor their warranty and send drive replacements. Been building up my NAS for like...6 years or more, that way. Edited Some: typos, clarity, rephrasing.
@katbryce2 жыл бұрын
@@Raletia I currently have 4x 10TB with 8GB RAM. I've done it with 4TB, and it works fine, but a bit slower because there is less RAM available for cache. You can replace smaller drives with larger drives and grow the pool. Adding additional drives doesn't work so well and isn't recommended.
@WilkoTheMaintainer11 ай бұрын
Thought this was going to be a clickbait video seeing the thumbnail... But it's so nice to see a tech youtuber that actually knows wtf they're talking about. You sir have earned a new sub ♥️
@_unknown_guy2 жыл бұрын
As backups were mentioned, can we get a video about those? Guess my interest would be at home server/nas level. Currently have simple/naive setup using abraunegg/onedrive.
@TheRMUPs2 жыл бұрын
Craft Computing has several videos on setting up a home storage server and the TrueNas operating system
@Raletia2 жыл бұрын
I've had pretty good luck with Backblaze. Saved my butt twice already. Pretty cheap for one PC. Though, in my case, my NAS that I'm backing up is running Windows with Storage Spaces and ReFS(MS's answer to ZFS). I don't know if there is a way to use it with TrueNas or the like. Edit: For context, I have 12 drives, 3, 4, & 6 TB, for a total of around 40TB before redundancy. Though I only have about 12TB of data stored & backed up.
@itskdog2 жыл бұрын
@@Raletia I would probably use Backblaze were it not that they only seem to charge in USD. I don't want to be paying currency conversion fees to my every month, and for the monthly payments in GBP to not be the same amount each month.
@Raletia2 жыл бұрын
@@itskdog That's fair! I don't really know of any other services, as I stopped looking once I found and tested backblaze, for the price it's been very affordable compared to anything else I could find. I wish you luck! (I do hope there are alternatives.)
@lemonbrothers34622 жыл бұрын
@@itskdog you could use revolut to hold and pay directly in usd and save on the fees
@stephenbarraclough52422 жыл бұрын
im so confused. If i have a home server what do i do ?, i have 2x8tb a 6tb and a 4tb just running in one large vmfs6 in esxi as a datastore. I was going to raid 1 the 2 8tb ones, and leave the 4 and 6tb as is
@127ibenedict2 жыл бұрын
Is it worth it to even use a raid configuration on a personal machine nowadays? Or is paying for a bigger individual storage solution a better option now? I assume that more modern SSDs either consumer or enterprise grade do a decent job of avoiding the issues that raid configurations helped deal with back in around 2012 when I first started seeing people use it.
@darbyevert8282 жыл бұрын
Not for like 99% of people, you can get 8tb ssds of you really need space. I think it's mostly hoarders raiding hdds for more space and vfx engineers trying to Lower their load times.
@chrisbaker85332 жыл бұрын
If you're doing cost to capacity, raid still, largely, wins out for ssds. For instance, the cheapest 8TB ssd on newegg is $710 usd, 2x4TB, $680 usd, 4x2TB ssd's, $600 usd, 8x1TB $576 usd. For spinning rust, it's not a contest, no raid, buy bigger. Low end, 8TB is $130 usd, 8x1TB would be 256. It gets a bit harder with higher, above 12TB, capacities and can be cheaper to use raid instead. Another consideration is power usage, multiple drives increase power usage exponentially. While ssds are far lower than hdd's, it's still an increase and should be a consideration. With HDDs, you also have to consider vibration, more drives, more vibration. Then there's the backup and redundancy consideration, can you afford to lose all your data or pay for recovery? If you're as A*** retentive as me, you backup religiously, but that raid offers a level of protection as well. Instead of losing one drive and it's all gone, you can lose one, or more, drives and still maintain the data. Individual use case is always going to be the determining factor. Just consider the pros and cons of each.
@marcogenovesi85702 жыл бұрын
do you need data integrity, uptime or performance on the workstation? It depends really. If it's just a gaming PC with stuff mostly synced to cloud or nas then it's fine on a single SSD
@vgamesx12 жыл бұрын
@@chrisbaker8533 I honestly don't get how it offers any meaningful protection, at least speaking for home users who will typically have at most 4-8 drives, aside from parity data, you can do sort of the same thing as a raid1/10 by just making a copy of everything to another drive and the benefit there is that you don't have to reformat drives already in use or setup raid volumes which poses a slight risk itself should anything go wrong and this way you can dynamically assign importance/redundancy to your files for example I keep a copy of important things like photos on 4 drives plus remote backups, movies or stuff I want to keep on 3 drives and replaceable files like steam games can just exist anywhere, so I can lose one or more drives and get a little extra usable space that isn't being used for redundancy. Basically, it appears to me that raid for home or fewer than like 10 drives, mainly just offers a form of automation and a nicer way to pool drives as a single one, but in the case of raid 5/6 anyway, it also brings complexity that can make it harder to recover data than if you just had a drive failure on a standalone disk, so I can see why you might make a few raid1 volumes but aside from that, what am I missing that makes raid useful for home users?
@Raletia2 жыл бұрын
@@vgamesx1 For me it's useful for using an old PC as a NAS for centralized storage & backup for all my devices, desktop, laptop, phone, tablet, etc. I use Win10pro & storage spaces, and have Backblaze for backup of the NAS.
@jarrodjohnson52192 жыл бұрын
Not just reliability, but performance wise hardware raid can be worse. The upside is negligible (the parity calculation for the throughput is trivial for modern processors). Meanwhile, the sas protocol in raid context means all device queues are squashed into one, instead of multiple independent queues. So independent io operations that software raid can do in parallel get serialized in hardware raid context. The biggest thing the controllers may add is a large dram cache with backup power to allow it to be treated like non volatile storage. Nvme nand cache is generally good enough though. If the raid manifested virtual drive as nvme, then it wouldn't have the queue problem.
@Scoopta2 жыл бұрын
Technically BTRFS stands for B-tree filesystem named after the B-trees it uses for directory information but it's often pronounced better FS
@hellterminator2 жыл бұрын
I've literally never heard anyone call it that. I have, however, heard “butter FS” a few times. :D
@levygaming31332 жыл бұрын
@@hellterminator the difference between butter and better is so small someone could have said “ButterFS is so great, it’s the best file system” and they could have assumed it was called BetterFS
@Scoopta2 жыл бұрын
@@hellterminator I've heard both, although I think butter FS is more common, the Wikipedia page lists both as alternative names tho
@mdd19632 жыл бұрын
I've never heard anything other than 'Butter-FS'....
@nofreenamestoreg2 жыл бұрын
I did stop playing with HW raid controllers a long time ago and back then they were doing integrity checks.... quite, quite strange to see such a thing being dropped.
@udirt2 жыл бұрын
not dropped, people just don't RTFM.
@ChaJ672 жыл бұрын
While ZFS RAID Z level 2 is about as indestructible as it gets and you mentioned some even more insane things, I think there are some things to add to this and some things to question: 1. Hardware RAID cards don't hold onto the data with the supercap. Instead the supercap keeps the card alive in a power loss event long enough to dump to flash, just as that SSD you were holding does. Just to clarify an error in your script. 2. Hardware RAID cards haven't always done the right thing when they hit corruption. LSI cards for a little while at least had an issue where they kept writing out random data when they encountered corruption on one of the hard drives. It was a major ordeal trying to get them to fix that mess. I mean giving more control over to the hardware RAID controller hasn't always ended well. Sometimes the drives do a better job themselves and they are produced in such a higher quantity that you may be more likely to find those kinds of problems sooner with the drives over finding it with a RAID controller. Like FOSS software tends to be better at say the kernel level than some arcane piece of application software because everyone uses the kernel and stuff gets spotted right away while the arcane piece of software has fewer eyes on it and so things tend to go longer without being caught. There are simply a whole lot more hard drives out there than there are hardware RAID controllers and if you deal with hardware RAID controllers enough, you will probably see some odd things that you don't understand why this doesn't get fixed. 3. RAID can be worse than having a single drive, especially when dealing with older hardware RAID cards and some software / firmware RAIDs. Those older Dell Perc cards stored their configuration with a battery. When that battery went and power was lost, you would see the controller report a number of independent hard drives instead of a single logical RAID drive and the system would report no OS found. I have seen MD RAID arrays get a superblock not found / corrupt error after a simple power loss event and there was no way I could find to get the array to work. Seeing that you mentioned NetApp, back in the day they were doing WAFL and RAID 4 with up to 15 drives, I would see things like a power transfer switch was flapping all night long just after the administrators went home for the night and by 8:00am the next morning the UPS battery was long dead and the NetApp was reporting 2 drives had popped out of one of the 15 drive arrays when RAID 4 can only handle a single drive failure. That single parity drive also tended to get clobbered and would bottleneck the whole array. It has been a while since I have used NetApp, so I hope that is better. Overall it was ahead of its time back in the day. 4. I have to wonder if this whole not trusting a single drive is overblown. I thought the whole point of going to 4k physical sectors was to improve the CRC checks the drives did internally while more efficiently using the physical space on the drive. I guess I have just never heard of anybody in the field actually complain of a modern 4k enterprise class drive silently corrupting the data. 5. I think it is worth mentioning more on which RAID levels are better in more detail. RAID 1/10 is not as great as some think because sometimes the mirrored disk fails before you can reconstruct the array. RAID 10 just adds more mirrors where this can and has happened. RAID 5, I see this from time to time where more than one drive pops out of the array and you can't fix it. Usually you can catch it before it happens, but not always. Sometimes like with that NetApp with 2 drives that popped out, someone managed to get one of the drives going eventually and recovered the array. For all practical use for hardware RAID, RAID 6 is the gold standard anymore. I never hear anybody besides you complain about silent corruption, though I do see it a lot on USB pen drives, which work to a different standard, but I do hear from time to time about double drive failures before the array can be rebuilt. For the highest I hear anybody trying to go besides maybe a bank is RAID Z level 2 with up to 8 drives per vdev and then just add more vdevs to the zpool if you want a bigger array. This keeps things down to a mathematical improbability of losing your data to random drive failures. Granted sometimes disasters happen, which is why RAID is not a backup. 6. The thing with hardware RAID controllers anymore is they are stuck in an old ideology for the most part. That is you only want to do RAID and the drives are only going to live attached to this RAID controller and you are never going to swap out that RAID controller. You can flip a RAID controller over to being an HBA, but then it is only an HBA. 3ware allowed you to do both and in general was a more forward thinking implementation, but they got killed off by the big boys in RAID, so that is dead and gone now. If the RAID card messes up, you are in a world of hurt. You want to move arrays around, well that is a good way to lose all of your data as 'foreign' arrays even if the RAID controller recognizes them, the RAID card is all too eager to re-initialize the array, in other words re-write the drives with all zeros, wiping your data out, especially if say an array ends up in storage and one of the drives doesn't work when you put them on a new controller. 7. The thing is with BTRFS is it is just less polished and robust than ZFS when used for RAID directly. BTRFS can recover from common errors, but when I abuse tested it, the BTRFS process just kept crashing, which isn't particularly good. When things go sideways, you really have to be a bit of an expert to pull BTRFS out of its funk. You usually can, unlike say MD RAID, which I consider the worse in Linux, but it can be a hard road and it doesn't always work out. I mean BTRFS RAID is head and shoulders better than MD RAID, and I would consider it good enough for an enthusiast who doesn't care too much about their data, but then you get to ZFS. ZFS RAID Z level 2 in my abuse testing survived everything that was potentially survivable. At this it tends to just heal itself whenever it can; you don't have to be a pro to get it to fix itself, unlike BTRFS. 8. The thing with ZFS under Linux at least is historically you couldn't grow a vdev. I mean in general expanding your array kind of sucked as you had to add a new array to the existing array and you are just kind of appending the arrays together. That should be fixed with the latest versions of ZFS. This was a big thing that made hardware RAID nice is you could always easily expand a hardware RAID one or a few disks at a time if you wanted. It takes a while, but it can be done. Resilvering has also gotten a lot better in the newer versions of ZFS. 9. Something important to mention is both ZFS and BTRFS really need systems with ECC RAM to perform their best. A bit of an omission in your script when going on about data integrity and bit rot. 10. Something really important to point out seeing you are talking about SSDs in your piece is ZFS RAID Z can talk directly to the drives and can do TRIM while doing this. At this you can do encryption in ZFS directly and stick basically whatever you want in a dataset. In contrast a RAID controller does not understand TRIM. A RAID controller does not understand all of this other stuff you may find yourself doing with your file system, so things end up layered on top of each other, which is inefficient. This leaves SSDs attached to hardware RAID controllers to take quite a beating and anything less than enterprise class high endurance SSDs will likely get beaten to death in short order. With ZFS RAID Z on the other hand, you may just find that lower endurance SSDs with capacitor backup is good enough for personal use. Even for business, those super high endurance SSDs are rather expensive, so if you can drastically cut the writes and say do the same job with cheaper 1 DWPD drives over 10 DWPD just because ZFS RAID Z is a lot nicer to the drives over a hardware RAID controller, that is a good deal. 11. Something that is really nice about ZFS RAID Z is you can just pull an array out of a system, maybe even have the whole array in a USB-C / TB3 enclosure and plug it into another system and just start using it. That is really handy. There are a number of reasons why you would want or even need to do this, even if you initially never planned on this, things happen. At this unlike hardware RAID, there is no risk of losing your data where as mentioned above, hardware RAID is all too eager to wipe data in these scenarios. It cannot be understated how useful SSD caching drives are with ZFS when the main drives are mechanical. The thing is if you care about your data and just across a wide array of use cases, ZFS is the best, beating out everything else including hardware RAID. So my conclusion is the same, but with some of these extra nuances thrown in.
@xbgt852 жыл бұрын
I tried so hard to get zfs working but I couldn't get my windows VMs saved to it.
@5urg3x2 жыл бұрын
I used to work in the storage industry, for a company that had their own RAID products, and I am glad that HW raid is dying / dead. They were pretty good about continuing to update older products so that they could be used with modern operating systems, but many other manufacturers (cough PROMISE cough cough) are not. Want a new driver? You're going to have to buy a new one, bro. Doesn't matter that the existing product you have still works fine, and you don't need faster speeds or any other features. We're just not going to update the driver anymore, because capitalism. Sucks to be you and anyone else that bought it!
@johndododoe14116 ай бұрын
Putting parity into already complicated file system code is the wrong solution . Putting a custom controller or driver between the raw storage and the sector level interfaces is a solid solution that can still be done both for magnetic and flash storage, all of which has the extra space for parity/ECC/checksums in the physical storage even if not visible on consumer interfaces like NVME, SAS or SATA . Details vary between lowest level interfaces and finding willing suppliers in an increasingly monopolized market may be hard . But selling false security like that PCIe card or the allegedly lazy MDRAID code is deeply disturbing, because most techs won't notice before disaster .
@hdtvkeith16042 жыл бұрын
I use older LSI Raid controllers in my white box servers. I use mostly for storage and reads so raid 6 works great and I do monthly media scrubs and then consistency checks
@Georgi-uc6uq Жыл бұрын
A very informative video indeed. After reading all of the comments I didn't see any mention of a Gen3 or Gen4 NVMe controller card that supports at least RAID2, or RAID3 error correction. Is there such a product? The only thing I saw is an unconfirmed claim that subsequent versions of GRAID Tech's SR-1010 now performs error checking and correction. Most of the discussion in the comments is focused on zfs and btrfs, but ultimately parity checking and error correction at the hardware level is always the best way to go.
@Level1Techs Жыл бұрын
It's hard to find nvme that confirm support for formatting to support the extra data integrity field. I asked the graid people about doing a review twice now but the conversation always goes cold when I ask for specifics about how it does integrity checking.. even if it doesn't do data integrity field you can kinda still do that in raid6 scenarios or any scenario where you have more than one parity drive. Hopefully someday it'll be testable :)
@CallousCoder2 жыл бұрын
I’ve been saying that in the early 2000s already. When I worked at IBM we always had problems with serveRaid cards. All the software raid systems just kept chucking along. And performance was as good as the hardware. That was a fake argument from the hardware fan boys. Calculating a cluster checksum is a very simple low level xor operation. That’s a native cpu instruction and even on 2000-2002 hardware had no burden on the CPUs. Sure it had a 5% overhead on your CPU but who cares. Hardly any CPU was used 100% and on the big Unix systems - where software raid was the standard solution. Unless you connected a Shark or a NetApp, had no noticeable CPU overhead. And I think The Unix boys realized the issue with dependency on proprietary hardware issues. Because all of them use software raid. LVM Or Veritas volume manager. But the greatest thing with software raid was, that there was no proprietary firmware and when your old serveRaid 4 died and you had to go to serveRaid 6 that you needed the right firmware to read the 4 Config off the disks. And that was a bitch to find! Or most hardware raids on the 90s and 2000s didn’t allow you to use large drives. And that was a problem! Because the small scsi drives were rare. And then you were stuck! We had a dedicated firmware version made at IBM to ignore this on I believe serveRaid 4 and 5 (6 I believe didn’t bother anymore - unless you worked on a 4 or 5 initiated disk set) but… it had to have the same amount of heads and platters! And I’ve been using ZFS since early days on Solaris and I adore it. Last year after 7 years a drive started to fail in my synology raid 6. I always buy extra drives to swap. So I took the bad one out. Pushed the new one in clicked initialize drive and rebuild. And it got stuck at 0.66% the whole web UI froze, the ssh didn’t react. A true hard kernel crash! So I had to turn it off in the midst of a rebuild… Well, when I booted it crashed on boot. I contacted synology what to do. But big companies means slow service!!! So I decided to wipe the OS. Reinstall it and I reconfigured the raid exactly the same. It started scanning disks and low and behold. That new drive was a DOA!!! So again that bay flashed red but my data was back. I pushed in my second spare and it happily reconstructed.
@jfbeam2 жыл бұрын
That "5%" of your general purpose CPU was significantly more power than the dedicated ASIC engine of a RAID card. (plus the overhead of the IO.)
@CallousCoder2 жыл бұрын
@@jfbeam true but you wouldn’t notice it, at all. But the flexibility it gave was/is outweigh the little CPU overhead. Being down because of an issue with a controller (and there were a lot), is slowing you down more that you every could regain. Like adding a new drive with most controllers that drive had to be identical. In the 90s/2000s drives only grew and retention was hard. We so often had to do a whole migration, because a stupid drive failed and we had to swap out all the good ones too. And I’ve never had that ever done on a software raid system.
@jamesp81642 жыл бұрын
The problem with software raid and solutions like ZFS (which I think is an outstanding solution) is one of enterprise support. Good example I'm dealing with, VMWare does NOT run on any software raid solution. Hard stop. If I want drive redundancy in my hypervisors, I have no option other than some kind of hardware raid solution.
@roysigurdkarlsbakk38422 жыл бұрын
Thanks for this one. I've been working with ZFS for a little more than a decade and some 12 years back, I started reading about btrfs as a ZFS replacement with all the goodies you mentioned (well, not all of them back then, but hell). Still - it's been 12 years and the RAID-[56] subsystem is still utterly broken. Perhaps wait another decade and it might be ok-ish. PS: A colleague of mine said he used btrfs on top of md for this reason, not to get the full use of btrfs' checksumming, which you obviously won't, but you'll get snapshotting and so on. So I tried it - on top of 5x2TB drives on md RAID-10 for VM storage on my home server. It's not a particularly fast machine, but it doesn't do that much either. It was dead slow. After a while, I replaced it with XFS again and it was back on track. Btrfs is a good idea, but I guess it needs a lot of funding or at least a lot of work to get stable and useful. PPS: I'm aware of (as you said) that zfs and btrfs will never be as fast as md and its friends, but this was just horribly slow - half of or perhaps one third of iops compared to xfs on the same md raid. So regardless of the iops and cpu needed for checksumming, I really doubt that is the root cause of this.
@wishusknight30092 жыл бұрын
ZFS i found was much faster than BRTfs. And I use a ZIL on my truenas box for a large ZFS2 array. Speeds are quite good while still having sync writes. Without the zil its horrifically slow like you said though. 10 disks will write about 10MB a second with just sync writes, and at 240MB/s with the zil. Which is the limit of the zil itself and not the drives or network. Which consists of 2 Intel S3700 SSD's with PLP on a sata2 connection... A sata3 conection would be faster. And this is also using a large single file copy. Lots of small files are slower of course. My system without a zil using non-sync writes was roughly 3 times as fast at around 650MB/s. However I need to walk on tiptoes with tightened butt cheeks all the time hoping the worst doesn't happen. I would have expected to peg out the 10g link but not sure where my bottleneck is.. I run pretty budget hardware so I have to make due. An internal DD showed way higher than 650, though I cant remember what it was. If you are ok with lazywrites than performance of ZFS is fantastic without the need for a ZIL.... But best practice is to use a mirrored zil and have the best of both.. And actually mirroring the zil is not really needed either and is just overkill according to the documentation and many top supporters as well.
@roysigurdkarlsbakk38422 жыл бұрын
@@wishusknight3009 you always have a ZIL ;) But I guess you mean an SLOG - a separate ZFS intent log placed on an SSD (or preferably a pair of those in a mirror)
@chromerims2 жыл бұрын
Great video, thank you 👍. 0:10 nice white Lian Li O11 case in the background . . . unless I'm mistaken.
@mikereeves47232 жыл бұрын
As a ZFS fanboy I can say hardware raid is still useful in many enterprise applications. Mirroring the OS drive is a perfect example where hardware raid makes it much easier than using ZFS. Pop in the new disk and continue on with your day vs having to run commands to replace the disk. A raid 6 on say 6 disks will perform much better than a single vdev raidz2. So its a balancing act between speed and safety. Achieving high speeds on ZFS typically means lots of disks. I personally use a single ZFS pool with 4 x 6 disk raidz2 that gives me the most balance between the two. For the typical home user, ZFS is the way to go but I don't think its time to call hardware based raid dead yet for enterprise. I would say you have quite the argument though for home use.
@PanduPoluan2 жыл бұрын
I'd say data integrity is even more important for Enterprise. In the consumer market, losing your data sure is annoying and causes grief if it's something irreplaceable such as memories of your family events. But with Enterprise, it can mean business interruption, or compliance violation resulting in legal issues.
@harryshuman9637 Жыл бұрын
ZFS isn't hardware RAID tho, it's a software level raid.
@SmokingBeagles2 жыл бұрын
Good to see you looking so well man, I could listen to you talk about computers all day and intend to on my day off. Collab with Dave's Garage when??
@unixtreme2 жыл бұрын
Proper new enterprise software raids are still doing parity. But some companies are trying to cut corners.
@mike_pj2 жыл бұрын
Thanks for posting this video. So what's the best option if you want to have a single server with redundant SSD storage running something like ESXi? I've had plenty of drives die on me, and it's really convenient to have a hardware RAID where I can just swap in a new disk with 0 downtime. Seems like the only option is to have a second TrueNAS server providing storage via iSCSI for the VMware server. But if you're coloing this kind of setup, your costs just doubled.
@DocNo272 жыл бұрын
I haven't used hardware RAID for over 10 years now. Lost a drive in a hardware RAID array several years back and had a heck of a time getting it re-established - gave up, made a second backup, blew it away and used software RAID and haven't looked back since. And with either ZFS or UnRAID there is no way in heck I would EVER mess with hardware RAID again. It was useful back in the day before multicore CPUs and CPU performance in general wasn't great, but it's not necessary at all these days.
@flecom53092 жыл бұрын
ZFS is great, I love the idea of a bad update breaking all my storage... anyway any kind of storage array redundancy is just for hardware failure, you really need a proper backup strategy if the data is of any value... the best ZFS setup won't help you if your datacenter catches fire (see the one in France for a good example)
@lassekristensen38511 ай бұрын
I build a LSI Logic Enterprise RAID controller with cache options and extra ram onboard 1GB, and battery backup. 4 x 6 TB WD RED PLUS. And 2 x 512 GB SSD for SSD cache. I am VERY happy with the result. This is for my storage server. Its running RAID 5. This with 2.5Gbit LAN. Quite an upgrade for my Synology NAS.
@ckingpro2 жыл бұрын
mdadm can be used with dmintegrity (typically with integritysetup) for data integrity. You can likely do it with supremeraid but then you would lose some performance as well as involve more of the CPU. That said, I definitely like and use ZFS Also, a correction: 4096 != 4 * 512. It is 8 * 512.
@Granite1652 жыл бұрын
My last functional desktop had two SSDs in a RAID setup. One of the drives gave out when I was away working, lost everything. :( Total bummer!
@84Actionjack2 жыл бұрын
Stablebit allows assigning redundancy to folders; it also saves as files to drives in the pools as opposed to striping across drives which means if a drive fails interrupting the pool, all your data is still there except (perhaps) the failed drive.
@TheHerrHorst2 жыл бұрын
Thank you for this video. Now I know I did the right thing with btrfs instead of raid md. Do you have any links about the btrfs policy based folder redundancy? I couldn't find anything about that, other than people wished for it
@Derek.Iverson2 жыл бұрын
It isn't supported yet, but will likely be implemented in the future.
@mdd19632 жыл бұрын
Try a Windows Storage Spaces Parity on 3 spinning drives, and check write speeds via CrystalDiskMark....; a staggering 33 MB/sec sequential....! :)
@4brigger2 жыл бұрын
The literal aside for the "aside" camera angle was a nice touch.
@Spark0102 жыл бұрын
Really enjoyed this video. Historically I've used RAID (proper RAID) for decades in both commercial and domestic scenario but since 2014 I've been out of the loop. I've only every deployed Windows based systems and based on your video, its a pretty grim picture these days if you wanted to run NVME Raid at home on Windows 10/11. I'm currently using an old Dell Perc controller or RAID 1 with SSD's, which is giving me parity, however an upgrade is looming and your video is now bouncing around in my head, with no clear solution for users who want RAID at home for their Windows 10/11 workstations, who care about data integrity. M$ is touting ReFS for its new server platforms, however I wondered if you had any thoughts (maybe this would be a good follow-up video) on how to solve this issue for Windows 10/11 NVME based systems?
@georgen98382 жыл бұрын
This boat's the boat I'm in too. Holy crap is this video a whole lot of bad news in the span of 20mins. A vid on modernizing some "typical" legacy deployments while maintaining at least similar levels of risk mitigation would be gold.
@daz4172 Жыл бұрын
Great video and concise explanation. I would love to see a L1Tech analysis of RAID parity rebuild and URE (Unrecoverable Read Error), as not a whole said is said about this situation. A lot of faith is put into parity meaning safety of data, when it could very well mean the entire loss of it, if a URE occurs during a parity rebuild.
@CreeperOnYourHouse2 жыл бұрын
I'm reminded of how the original name of RAID was Redundant Array of Inexpensive Disks. What you said about 512 vs 520 byte sectors and the 520 byte sectors costing more really put that into context; hardware raid allowed the use of cheaper 512 byte sector drives and maintaining the same redundancy as the more expensive ones.
@trumanhw Жыл бұрын
Did you know that within those 512-Bytes ... are actually chunks for ECC...? End-users won't know this, but, with Data Recovery HW (DDI or PC3000) we can see if the drive is unable to return data, or if it's refusing to ... (ie, bc it's unable to read CRC or CRC is inconsistent). We're then able to read block ignoring ECC or even to have the block read 10-times, returning the average data-read to minimize the return of erroneous data.
@RicardoPenders2 жыл бұрын
I used once a RAID 0 array and after a short time something happened to the system that took out both of my drives at once, all data gone. I did get two new drives for free because the failure was within warranty, luckily I always have the most important data saved somewhere else, I had my system drive set as RAID 0 just for speed.
@Team5032 жыл бұрын
RAID is not only not dead, it is thriving. It is at the core of damn near every SAN and NAS appliance on the planet. It is not RAIDs job to protect data integrity, and it has never been. RAID has a single primary purpose - to provide resiliency against hardware failure. In a single sentence, THE POINT OF RAID IS UPTIME. Data integrity is the job of the file system and the job of your backups. Do not confuse the two.
@Team5032 жыл бұрын
@@djancak No, not really. Data integrity can fail regardless of whether your hardware does or not.
@mattropolis992 жыл бұрын
Ding ding ding! There is also two additional benefits - recovery without restore and speed. I run hardware RAID on my home development system. I have had plenty of drives die over the years (Seagate - looking at you), but I was never down nor had to go through the day-long process of re-building my drives from backups. It's not just swapping the drives, it's re-installing everything, restoring data, and setting everything back up. That's a DAY of lost work if not more to restore after a drive failure. I've not done that in over a decade now despite losing a drive or two. Also, I run RAID 5 - stripe plus redundancy. This has one of the best performance profiles. It has the benefit of speeding up my drives for many operations (especially reads) while having very minimal overhead for the slower ones (writes). Be careful though, because things like RAID 1 halve your write performance while increasing your read performance by less than that. So, know what you're getting into. I have friends that run raid 0 on a bunch of SSD's for video editing scratch drives. Their performance is astounding. Sure, this wears SSD's out faster, but they want speed and the cost of four 1TB SSD's even once a year is something they don't even notice given the additional productivity and work they get done.