We bought 1347 Used Data Center SSDs to See SSD Endurance

  Рет қаралды 109,119

ServeTheHome

ServeTheHome

Күн бұрын

Пікірлер: 422
@79back2basic
@79back2basic 4 ай бұрын
why didn't you bought 1337 drives ? missed opportunity ....
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
We bought many more than that, but you are right, it was a missed opportunity not to prune 10 more from the data set.
@DrRussell
@DrRussell 4 ай бұрын
Clearly I don’t understand the reference, may I ask the significance of 1337, please?
@peakz8548
@peakz8548 4 ай бұрын
@@DrRussell en.wikipedia.org/wiki/Leet
@ralanham76
@ralanham76 4 ай бұрын
​@@DrRussell type 1337 on a calculator and look at it upside down spells LEET
@gamingballsgaming
@gamingballsgaming 4 ай бұрын
i was thinking the same thing
@cmdr_stretchedguy
@cmdr_stretchedguy 4 ай бұрын
In my 20+ years in IT and server administration, I've always told people to get twice the storage you think you need. For servers, especially if they use SSDs, if they think they need 4TB, always get 8TB. Partially because they suddenly need to create a large file share, but also because the larger SSDs will have lower DWPD and typically last longer. I dealt with 1 company that had a 5 SSD drive raid5 (250GB SSDs) but they kept their storage over 95% at all times so they kept losing drives. Once we replaced and reseeded with 5x 1TB, then expanded the storage volume, they didn't have any issue for over 3 years after that.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
What is interesting is that this basically shows that doubling the capacity also helps with the write endurance challenge. So the question is do you get a higher endurance drive, or just get a larger capacity drive that has similar endurance.
@CoreyPL
@CoreyPL 4 ай бұрын
@@ServeTheHomeVideo It's like with normal disks - if you operate it on 95% all the time, where most data is cold, wear leveling algorithm can't function properly and new writes quickly kill this 5-10% of changing cells. If you up the capacity, then wear leveling can do its job properly.
@thehotshot0167
@thehotshot0167 4 ай бұрын
That is a very helpful interesting tip, Ill keep that in mind for future builds.
@userbosco
@userbosco 4 ай бұрын
Exactly. Learned this strategy the hard way years ago....
@Meowbay
@Meowbay 4 ай бұрын
​@@ServeTheHomeVideoOr, instead of using a 2 drive mirroring raid ssd, use single ones and just use the second ssd for expansion of space. Which is fine, as long as you're not rewriting that single ssd too often.
@CoreyPL
@CoreyPL 4 ай бұрын
One of servers I deployed 7-8 years ago hosted MSSQL database (around 300GB) on a 2TB volume consisted of Intel's 400GB SSD drives (can't remember the model). Database was for ERP system that was used by around 80-100 employees. After 6 years of work before server and drives being retired, they still had 99% of life left. They were moved to a non-critical server and are working to this day without a hitch.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
That is pretty good though! Usually a service life is defined as 5 years
@CoreyPL
@CoreyPL 4 ай бұрын
@@ServeTheHomeVideo Yeah, I was pushing on management to spend some $$$ on a new server and move current one to non-critical role as well. It's hard to convince non-tech people that even server grade equipment isn't meant to work forever.
@MW-cs8zd
@MW-cs8zd 4 ай бұрын
Love the used Intel ds SSD. Expensive on eBay now though
@MichaelCzajka
@MichaelCzajka 4 ай бұрын
@@ServeTheHomeVideo 5 years is for mechanical drives. SSD's seem to last 10 years or more. In most cases... with light use you'd expect the drive to continue to be used until it becomes obsolete. Even with heavy use it's likely to last a looong time. The question for SSD's has always been... "How long will they last?" 🙂
@scalty2008
@scalty2008 4 ай бұрын
10years for HDD is good too. We have 500+ HDD here in Datacentre, the oldest ones 4TB running since 2013 as backup to Disk Storage and now doing their last days as Exchange Storage. Even the first helium 8TB running fine since 2017 (after Firmwareupdate solved a failure bug). Disk failures at all 500+ are less than 5 per year.
@edwarddejong8025
@edwarddejong8025 4 ай бұрын
We only used Intel (Solidigm now) drives on in all of our server racks. They have performed wonderfully. They have a supercapacitor so that they can write out the data if there is a power failure. An essential feature for data center use. We haven't however upgraded to SSD for our NAS units because we write a huge amount every day, and SSD's would have burned out in 3 years; our mechanicals have lasted 9 years and only had 3 out of 50 drives fail.
@MrBillrookard
@MrBillrookard 4 ай бұрын
I've got a SSD that I put in my webserver wayyyyy back in 2013. Crucial M4 64GB SSD. I was a bit iffy about it as that was when SSD tech was pretty new, but I picked a good brand so I just YOLO'd it. Currently still in service, 110,000 power on hours, 128 cycle count. 0 uncorrectable, 0 bad blocks, 0 pending sectors, and one error logged when it powered off during a write (lost power, whoops). Still, 12 years of service without a hiccup, and according to the wear leveling, it's gone through 4% of it's life. At that point I expect it to last... another 275 years? Cool. I guess my SSD will still be functional when we develop warp drive if Star Trek shows where we're headed. Lol.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Wow
@supercellex4D
@supercellex4D 3 ай бұрын
I think my computer will last forever
@giusdb
@giusdb 3 ай бұрын
It always depends on how you use the ssd. I have a 250 GB sata ssd crucial that is used both little for an operating system and quite a bit as a cache, after a few months it lost 20% of its useful life. I replaced it (to use it for light use) with a 250 GB nvme ssd samsung that I previously used for years in a similar way to the sata but much more intense, after many months it is at 2%.
@sadnesskant7604
@sadnesskant7604 4 ай бұрын
So, this is why ssds on ebay got so expensive lately... Thanks a lot Patric😢
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Ha! When NAND prices go up ebay prices do too. We have been buying the drives in here for almost a decade.
@quademasters249
@quademasters249 4 ай бұрын
I noticed that too. I bought 7.6 TB for $350. Now I can't find it for less than $500.
@Knaeckebrotsaege
@Knaeckebrotsaege 4 ай бұрын
There has been price fixing going on in terms of NAND chips, and Toshiba/KIOXIA already got bonked for it. Check price history for consumer SSDs up till november/december 2023, and then up to today and watch the line go up and up and up for no reason whatsoever... basic 2TB TLC NVMe SSDs were down to 65eur, now the very same models are 115+eur. Heck 1TB TLC NVMe SSDs were at the point of being so cheap (35eur!) that you just threw them at everything, whether it needed one or not. Now with the price ballooned to 60+eur, not anymore. And yes, consumer SSDs aren't the target for viewers of this channel, but the prices for consumer junk exploding inevitably also has an effect on used enterprise stuff
@thelaughingmanofficial
@thelaughingmanofficial 4 ай бұрын
Welcome to the concept of Supply and Demand.
@WeiserMaster3
@WeiserMaster3 4 ай бұрын
​@@thelaughingmanofficialillegal price fixing*
@djayjp
@djayjp 3 ай бұрын
Keep in mind the survivorship bias in effect here: you typically won't be sold already dead drives....
@seccentral
@seccentral 4 ай бұрын
Recently I saw a vid by level1techs saying pretty much the same thing, he hammered a drive rated for hundreds of tbw with over a petabyte and it still ran; also, the same idea around companies very very rarely needing anything modern drive bigger than 1dwpd. Thanks for confirming this. And for new ones, it matters: - Kioxia 6.4 TB 3DWPDs go for 1600, similar 7.6 TB 1 dwpd drives are 1000 and when you're building clusters, it matters a lot
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Yes. And with big drives you should not need 1DWPD
@MikeKirkReloaded
@MikeKirkReloaded 4 ай бұрын
It makes all those 1.92/3.84TB used U.2's on Ebay look like an even better deal for homelab use.
@balex96
@balex96 4 ай бұрын
Definitely, I bought yesterday 6 TOSHIBA 1.92 TB SSD for 85 British pounds each.
@originalbadboy32
@originalbadboy32 4 ай бұрын
​@@balex96you can buy brand new 2tb SDDs for about £90... so why risk used
@Beany2007FTW
@Beany2007FTW 4 ай бұрын
@@originalbadboy32 Because homelab use tends to be a lot more write intensive than a regular desktop PC by it's nature, so getting higher endurance drives makes a difference. Also if you're working with ex-enterprise hardware (as many homelab users are), you're talking U2 2.5 hotswap capable drives for arrays, not M2 keying for mobo slots or add-in cards. You can't get those for £90 new. Different use cases that require different solutions, simple as that.
@originalbadboy32
@originalbadboy32 4 ай бұрын
@@Beany2007FTW to a point I agree but even most homelab users are probably not going to be pushing writes all that much. Media creation sure, outside of that probably not pushing writes so much that you need enterprise level hardware.
@Beany2007FTW
@Beany2007FTW 4 ай бұрын
@@originalbadboy32 Might want the battery backed write protection for power outages, though. There's more to enterprise drives than just write endurance.
@concinnus
@concinnus 4 ай бұрын
In the consumer space, most of the reliability issues have not been hardware-based but firmware, like Samsung's. As for rebuild time and RAID levels, the other issue with hard drives is that mechanical failures tend to happen around the same time for drives from the same manufacturing batch. We used to mix and match drives (still same model/firmware) in re-deployed servers to mitigate this. Probably less of an issue for SSDs.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
You are right that there are other factors. We lost an entire Dell C6100 chassis worth of Kingston DC SSDs because of a power in rush event. At the time Intel had the protection feature and Kingston did not. Now most do
@paulbrooks4395
@paulbrooks4395 4 ай бұрын
The contrary information is hybrid flash arrays like Nimble that does read caching by writing a copy of frequently used data to cache. Our Nimble burned through all of its data center write-focused SSDs all at once, requiring 8 replacements. The SMART data showed 99% drive write usage. We also use Nutanix which uses SSDs for both read and write tiering. Since we host a lot of customer servers and data churn, we see drives getting burned out at an expected rate. To your point, most places don't operate like this, instead being WORM operations and using SSDs for fast access times. But it's still very important for people to know their use case well to avoid over or under buying.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Exactly. It is also interesting that write focused drives often were not used in that manner.
@purrloftruth
@purrloftruth 4 ай бұрын
not that i know anything about anything, but i think there should be some sort of opt-in industry-wide database where interested server/dc owners can run a daemon on their server that submits the smart stats of all its drives daily, so that people across the industry can see statistics on how certain models perform, potentially get early warning of models with abnormally high failure rates, etc.
@ThylineTheGay
@ThylineTheGay 4 ай бұрын
like a distributed backblaze drive report
@purrloftruth
@purrloftruth 4 ай бұрын
@@ThylineTheGay yeah, but updating in 'real time' (daily or so). whereas they put one out once a year iirc
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
The server vendors can do this at the BMC level and then use the data for predictive failure service
@giusdb
@giusdb 3 ай бұрын
Unfortunately it would not be useful. The same model can have different characteristics in different years of sale. And the reliability Is expressed in a wide range. And it depends a lot on the specific use, you can shorten the life of years of an ssd in a few weeks.
@udirt
@udirt 4 ай бұрын
My favs were the Hitachi HGST, not the stec ones but their own. Any number in the datasheet was understating their real performance. Pure quality.
@ChrisSmith-tc4df
@ChrisSmith-tc4df 4 ай бұрын
I’d still want a DWPD that’s at least some low multiple of my actual workload writes just so that performance doesn’t suffer so much near EOL when ECC would be working hard to maintain that essentially zero error rate. That said, a lower endurance enterprise SSD (~1 DWPD) would probably suffice for the majority of practical workloads and save the costly higher endurance ones for truly write intensive use cases. Also the dying gasp write assurance capability helps prevent array corruption upon unexpected loss of power, so the enterprise class drives still provide that benefit even at lower DWPD ratings. That’s something to consider if considering using non-enterprise SSD’s in RAID arrays.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Totally, but then the question is do you still want 1DWPD at 30.72TB? 61.44? 122.88? Putting it another way, 8x 122.88TB drives will be just shy of 1PB of raw storage. Writing 1PB of 4K random writes per day is not trivial.
@ChrisSmith-tc4df
@ChrisSmith-tc4df 4 ай бұрын
@@ServeTheHomeVideo A decade+ ago back in the SATA/SAS SSD days, I recall the lowest write endurance enterprise drives that I saw aimed at data warehousing were 0.5 DWPD. So given the even lower write utilization on colossal drive arrays that are likely only partially filled, you’re advocating use cases for perhaps even less than 0.5 DWPD down near a prosumer SSD write endurance?
@sotosoul
@sotosoul 4 ай бұрын
Lots of people are concerned about SSD reliability not because of the SSDs themselves but because of the fact that SO MANY devices have them soldered!
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
That is true. This is just data center drives
@SyrFlora
@SyrFlora 4 ай бұрын
SSD reliability regarding how much write endurance in them not really improving to be honest.. it going backwards. Newer Manufacturing nowadays makes each cell more reliable but when industry shifts to QLC for consumers storage solutions. It is still worse than ssd in the TLC or MLC era . For most people it is still not a problem unless u are a really really heavy write user, a bad scenario like always staying with less than 10% free space or not enough ram that make swap like crazy to run OS and application that u use. U basically unlikely got an issue of failure because u wore out the cell. For mobile devices most people should be fine. But like pc.. soldered storage is pretty nasty like whut 🍏 did. Especially when bios stuff are also inside that ssd not dedicated chip. U wore it out.. it basically brick because u cannot even boot from other media.😂😂
@Meowbay
@Meowbay 4 ай бұрын
Well, speaking from personal experiences as a hosting engineer, that fear also stems from the large number of ssd failures that result in actually entirely unreadable, after primary failing notice. Controller error or not. This is not what you want hoping your data could at least partially be restored, as I usually can and could with mechanical drives. Many ssd's fail from 100% to being complete 0% readable. That's frightening, I assure you. Unless you're into resoldering your own electronics on such micro chips and know what parts make it fail, and you have your own lab for that and the time to do this, of course. But I don't think many among us would..
@kintustis
@kintustis 4 ай бұрын
soldered ssd means manufactured ewaste
@mk72v2oq
@mk72v2oq 4 ай бұрын
@@Meowbay as a hosting engineer you should know that relying on assumption that you will be able to restore data from a failed drive (regardless of its type) is dumb. And that having data redundancy and backups is crucial part of any data center operation.
@jeremyroberts2782
@jeremyroberts2782 4 ай бұрын
Our 6 year old Dell drives hosting a VMware vSAN for a mixed range of servers including Databases in the 1-2TB size, all the drives still have around 85-90% endurance availability. Our main line of business DB has read/write ratio of 95% reads/5% writes. Life of SSDs is really in the decades or more (assuming the electronic stuff doesn't naturally degrade or capacitors go pop). Most heavy use personal PCs will only write about 7GB of data a day (the odd game install aside) so on a 1TB drive it will take 150 days to do a full drive write, if the stated life is 1000 DW/ 3 years, it will take around 390 years to reach that limit.
@moeness86
@moeness86 3 ай бұрын
That doesn't address sudden death problems.. drives will fail in any category, but a headsup is always nice. Any idea how to check an SSD for issues ahead of failure? A follow up question would be how to do that with a raid array? Thanks for sharing.
@paulstubbs7678
@paulstubbs7678 4 ай бұрын
My main concern with SSD's comes from earlier endurance tests where a failed drive would become read only, then totally bricked if you power cycled it. This means if a drive dies, as in goes read only, you basically cannot clone that drive to a new one as that will most likely involve a power cycle/reset - as the O/S has probably crashed being unable to update something.
@kevinzhu5591
@kevinzhu5591 4 ай бұрын
In that case, you use another computer to retrieve the information by not using the drive as a boot drive.
@redslate
@redslate 4 ай бұрын
Controversially, years ago, I estimated that most quality commercial SSDs would simply obselete themselves in terms of capacity long before reaching their half-life, given even "enthusiast" levels of use. Thus far, this has been the case, even with QLC drives. Capacities continue to increase, write endurance continues to improve, and costs continue to decrease. It will be curious to see what levels of performance and endurance PLC delivers.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
That is what happens with us. Capacity becomes more important
@kennethhomza9026
@kennethhomza9026 4 ай бұрын
The consistent background music is a nuisance
@youtubiers
@youtubiers 3 ай бұрын
Yes agree
@ChrisSmith-tc4df
@ChrisSmith-tc4df 4 ай бұрын
Thanks!
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Wow!!! Thank you!!!
@ewenchan1239
@ewenchan1239 4 ай бұрын
Three things: 1) SSD usage and by extension, endurance, REALLY depends on what it is that you do. One of the guys that I went to college with, who is now a Mechanical Design Lead at SpaceX, runs Monte Carlo simulations and on his new workstation which uses E1S NVMe SSDs -- a SINGLE batch of runs, consumed 2% of the drives' total write endurance. (When you are using SSDs as scratch disk space for HPC/CFD/FEA/CAE applications, especially FEA applications, it just rains data like no tomorrow. For some of the FEA work that I used to do on vehicle suspension systems and body-on-frame pickup trucks, a single run can easily cycle through about 10 TB of scratch disk data.) So, if customers are using the SSDs because they're fast, and they're using it for storage of large, sequential (read: video) files, then I would 100% agree with you. But if they are using it for its blazing fast random read/write capabilities (rather than sequential transfers), then the resulting durability and reliability is very different. 2) I've killed 2 NVMe SSDs (ironic that you mentioned the Intel 750 Series NVMe SSD, because that was the one that I killed. Twice.) and 5 SATA 6 Gbps SSDs (all Intel drives) over the past 8 years because I use the SSDs as swap space for Windows clients (which is also the default, when you install Windows), for systems that had, at minimum, 64 GB of RAM, and a max of 128 GB of RAM. The Intel 750 Series 400 GB AIC NVMe SSDs, died, with an average of 2.29 GB writes/day, and yet, because it was used as a swap drive, it still died within the warranty period (in 4 years out of the 5 year warranty). On top of that, the manner in how it died was also really interesting because you would think that when you burn up the write endurance of the NAND flash cells/modules/chips, that you'd still be able to read the data, but that wasn't true neither. In fact, it was the read that was the indicator that the drive had a problem/died -- because it didn't hit the write endurance limits (according to STR nor DWPD nor TBW). The workload makes a HUGE difference. 3) It is quite a pity that a 15.36 TB Intel/Solidigm D5-P5316 U.2 NVMe costs a minimum of $1295 USD whereas a WD HC550 16 TB SATA 6 Gbps HDD can be had for as little as $129.99 USD (so almost 1/10th the cost, for a similar capacity). Of course, the speed and the latency is night-and-day and isn't comparable at all, but from the cost perspective, I can buy 10 WD HC550 16 TB SATA HDDs for the cost of one Intel D5-P5316 15.36 TB U.2 NVMe SSD. So, it'll be a while before I will be able to replace my homelab server with these SSDs, possibly never.
@RussellWaldrop
@RussellWaldrop 4 ай бұрын
Shouldn't someone who needs that crazy quick random R/W, wouldn't it be cheaper to just build a server with a ton of ram and create some form of a ramdisk? And more durable.
@Henrik_Holst
@Henrik_Holst 4 ай бұрын
@@RussellWaldrop building a commodity server taking TB of RAM is no easy feat. Even on EPYC you max out at 6TB of RAM per system and that RAM alone is easy $90K and you are only 1/3 into replacing that one 16TB drive that OP talked about.
@ewenchan1239
@ewenchan1239 4 ай бұрын
@@RussellWaldrop "Shouldn't someone who needs that crazy quick random R/W, wouldn't it be cheaper to just build a server with a ton of ram and create some form of a ramdisk? And more durable." Depends on the platform, RAM generation, and fault tolerance for data loss in the event of a power outage. Intel has their Xeon series which could, at least for two generations, take DC Persistent Memory (which Patrick and the team at ServeTheHome) has covered in multiple, previous videos. So, to that end, it helps to lower the $/GB overall, but historically speaking, if you wanted say like 1 TB of DDR4-3200 ECC Reg. RAM, it was still quite expensive, on a $/GB basis. (I couldn't find the historical prices on that type of memory now, but suffice it to say that I remember looking into it ca. 2018 when I had my 4-node, dual Xeon E5-2690 (v1) compute cluster, where each node had 128 GB of DDR3-1866 ECC Reg. RAM running at DDR3-1600 speeds, for a total of 512 GB, and if I remember correctly, 1 TB of RAM would have been something on the order of like $11,000 (if one stick of 64 GB DDR4 was $717, per this post that I was able to find, about the historical prices (Source: hardforum.com/threads/go-home-memory-prices-youre-drunk.1938365/)). So you figure that's ON TOP of the price of the motherboard, chassis, power supply, NIC(s), CPUs, HSFs (if you're building your own server vs. buying a pre-built server), and the cost of those components varies significantly depending on what you are looking for. (i.e. The top of the line Cascade Lake 28 core CPU that support DC PMEM original list price was almost $18,000 a pop (Source: en.wikipedia.org/wiki/Cascade_Lake#Xeon_W-2200_series) for the 'L' SKUs which support more RAM. So you get two of those suckers, you're still only at 28 cores each, for a total of 56 cores/112 threads (whereas AMD EPYC had 64 cores by then, IIRC, but didn't support DC PMEM).) My point is that the cost for a lot of RAM often became quite cost prohibitive for companies, so they would just go the SSD route, knowing that it's a wear item like brake pads on your car. (And like brake pads on your car, the faster it goes, the faster it wears out.) DC PMEM helped lower the $/GB cost SOME, but again, without it being supported on AMD platforms, and given the cost, and often times, the relative LACK of performance from Intel Xeon processors (compared to AMD EPYC processors), there wasn't a mass adoption of the technology, which is probably why Intel ultimately killed the project. (cf. www.tomshardware.com/news/intel-kills-optane-memory-business-for-good). I looked into it because like I said, for my HPC/FEA/CFD/CAE workloads, I was knowingly killing NAND flash SSDs VERY quickly. (Use them as a swap/scratch drive, and you'll see just how fast they can wear out without ever even getting remotely close to the DWPD STR write endurance limits.) (Compare and contrast that to the fact that I bought my 4-node micro compute cluster for a grand total of like $4000 USD, so there was no way that the capex for the platform that supported DC PMEM was ever going to fly/take off. It was just too expensive.) At one point, I was even playing around with using GlusterFS (version 3.7 back then) distributed file system, where I created 110 GiB ram disks, and then strung them all together as a distributed striped GlusterFS volume, to use as a scratch disk, but the problem that I ran into with that was that even with 100 Gbps Infiniband, it wasn't really read/writing the data significantly faster than just using a local SATA SSD because GlusterFS didn't support RDMA on the GlusterFS volume, despite the fact that I exported the gvol over onto the network as a NFS-over-RDMA export. That didn't quite go as well as I thought it could've or would've. (And by Gluster version 5, that capability was deprecated and by version 6, it was removed entirely from the GlusterFS source code.) (I've tried a whole bunch of stuff that was within my minimal budget, so never anything as exotic as DC PMEM.) There were also proposals to get AMD EPYC nodes, using their 8-core variant of their processors (the cheapest you can go), and then fill it with 4 TB of RAM, but again, RAM was expensive back then. I vaguely remember pricing out systems, and it was in the $30k-60k neighbourhood (with 4 TB of RAM, IIRC), vs. you can buy even consumer SATA SSDs for like a few hundred bucks a pop (1 TB drives, and you can string four of them together in RAID 0 (be it hardware or SW RAID), and then exported that as the scratch disk (which is what I did with my four Samsung EVO 850 1 TB SSDs, and then exported that to the IB network as a NFSoRDMA export, and the best that I was able to ever get with it was about 32 Gbps write speed, which, for four SATA 6 Gbps SSDs, meant that I actually was able to, at least temporarily, exceed the SATA interface theoretical limit of a combined total of 24 Gbps. Yay RDMA??? (Never was sure about that, but that's what iotop reported).) Good enough. Still burned through the write endurance limit at that rate though. For a company with an actual, annual IT budget -- replacing SSDs just became a norm for HPC workloads. For me though, with my micro HPC server, running in the basement of my home -- that wasn't really a viable option, so I ended up ditching pretty much all SSDs, and just stuck with HDDs. Yes, it's significantly slower, but I don't have annualised sunk cost where I'd knowingly have to replace it, as it wears out. $0 is still better than having to spend a few hundred bucks on replacement SSDs annually. (cf. www.ebay.com/itm/186412502922?epid=20061497033&itmmeta=01J56P9FCY6HJ5V1QT28FZ09PP&hash=item2b670d0f8a:g:dsMAAOSwke9mKR61&itmprp=enc%3AAQAJAAAA4HoV3kP08IDx%2BKZ9MfhVJKlh58auJaq6WQcmR34S6zfFgi4VcCPwxAwlTOkDwzQNAuaK9bi%2BmrehAA82MAu78x8Fx8iWc7PGv6TP9Vrypic02FAbBfEWd7UjU5W1G0CuYKYjCxdkETpy3xnK2D0iPrkBwNi5R%2BaphL%2B%2Fd8taZo0RG%2Fed%2F4QoqNmDMyMoTvDIBGifnVEngMykFUtrULKQMlUkbQ6ED%2B0iOYLQxEJDrkmSJauzdBzwMHCbNuvCLM0l08ziMQJVvBo1FBT%2FXXToZITQk%2BdUTBYfOv6cdotQ1678%7Ctkp%3ABk9SR8j2pdapZA) An open box Solidigm D5-P5316 15.36TB U.2 NVMe SSD out of China is $1168 USD. A WD HC550 16 TB HDD is $129.99 USD. I would LOVE to be able to replace my entire main Proxmox storage server with U.2 NVMe SSDs. But at roughly 10X the cost, there's no need for it. Nothing I do/use now (with my Proxmox storage server) would benefit from the U.2 NVMe SSD interface. I think that the last time that I ran the calculation for the inventory check, I am at something like a grand total of 216 TB raw capacity. It'd cost me almost $16k USD to replace all of my HDDs with U.2 NVMe SSDs. The base server that I bought, was only $1150 USD. The $/GB equation still isn't there yet. It'd be one thing if I was server hundreds or thousands of clients, but I'm not. (Additionally, there is currently a proposal that ZFS might actually be making my system work harder than it might otherwise need to, because if I offloaded the RAID stuff onto my Avago/Broadcom/LSI MegaRAID SAS 12 Gbps 9361-8i, the SAS HW RAID HBA should be able to do a MUCH better job of handling all of the RAID stuff, which would then free up my CPU from all of the I/O wait metric that is a result of the fact that I am using HDDs, so they're slow to respond to I/O requests.)
@Nagasaski
@Nagasaski 4 ай бұрын
What about intel optane? Or Crucial T700? They are almost server grade SSD but for consumers.
@ewenchan1239
@ewenchan1239 4 ай бұрын
@@Nagasaski "What about intel optane?" Depends on capacity and platform. On my 7th gen NUC, it recognises it, and it can be used as cache for the 2.5" Toshiba 5400 rpm HDD, but at the end of the day, it is limited by the HDD. (It just too slow.) I haven't tried using Optane on my AMD systems, but I am going to surmise that it won't work on an AMD system. "Or Crucial T700?" I quickly googled this, and the 1 TB version of this drive only has a write endurance limit of 600 TBW over its entire lifetime. Again, it depends, a LOT, on HOW you use the drive. If you use it as a swap drive, you can kill the drive LONGGG before it will hit the sequential transfer write endurance limit, which is how the TBW metric might be measured (or it might be like 70% sequential/30% random write pattern). However, if you have almost a 10% sequential/90% random write pattern like using the drive as a swap drive, you can exhaust the finite number of write/erase/programme cycles of the NAND flash of the SSD without having hit the write endurance limit. Again, my Intel 750 Series 400 GB NVMe SSD AIC, I only averaged something like 2.29 GB writes/day. But I still managed to kill TWO of these drives, in a 7 year period. (A little less than 4 years each.) And that's on my Windows workstation which had it's RAM maxed out at 64 GB. The usage pattern makes a HUGE difference, and the write endurance limit doesn't take that into consideration, at least not in terms of the number that's advertised in the product specs/advertising/marketing materials. (Intel REFUSED to RMA the second 750 Series that I killed because that was the drive that died after the first drive was RMA'd, from the first time that the drive failed, arguing that it was beyond the initial 5 year warranty from the FIRST purchase. So now, I have a dead 750 Series NVMe SSD, that's just e-Waste now. I can't do anything with it.) And that's precisely what dead SSDs are -- eWaste. And people have called BS about this, and I told them that by default, Windows installs the pagefile.sys hidden file on the same drive where Windows is installed. So, if you are swapping a fair bit, it's burning up write/erase/program cycles on your OS drive.
@JP-zd8hm
@JP-zd8hm 4 ай бұрын
DWPD is relevant in server specification - write amplification needs to be considered especially for ZFS or dual parity arrangements eg VSAN. That said, enterprise drives used are a great shout in my experience, 40% left of a 10PB total write life device is still very nice thank you!
@BloodyIron
@BloodyIron 4 ай бұрын
Welp that just validated what I've been thinking for the last like 10 years lol. Thanks!
@imqqmi
@imqqmi 4 ай бұрын
I remember around 2010 that I introduced 2x 60GB drives as the IT guy at a company in raid 1 config for their main database for their accounting software. Reports and software upgrades that ran for minutes up to half an hour was done in seconds. The software technician was apprehensive about using SSDs for databases but after seeing these performance numbers he was convinced. These drives worked for around 4 years after being retired but were still working. Capacitors and other support electronics seem to be less reliable than the flash chips themselves lol! I've upgraded all my HDD drives to SDDs last year and never looked back.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Yes. Also Optane was expensive, but it often moved DB performance bottlenecks elsewhere
@udirt
@udirt 4 ай бұрын
You'll see a lot more wear if you focus on drives in HCI setups due to silly rebalancing etc. You also need to factor in the overprovisining if you look at failure rates. People factored in this and gained reliability.
@RWBHere
@RWBHere 4 ай бұрын
Thanks for the heads-up. Now to go and find a few used server SSD's which haven't been edited with a hammer...
@reubenmitchell5269
@reubenmitchell5269 4 ай бұрын
We've had Intel S3500/3510 Sata SSDs as the boot drives in RAID1 for all our production Dell R730s for coming up 8 years - never had an issue with any of them. We had 3x P5800X Optanes fail under warrant, but the 750 PCI-E cards are still going strong
@GreenAppelPie
@GreenAppelPie 4 ай бұрын
So far my SSDs/NVMEs have had zero problems for 7+ years, while my hard drives on the other hand start failing within a few years. I’ll never get an SSD again. great episode BTW, very informational!.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Why never another SSD?
@mikemotorbike4283
@mikemotorbike4283 3 ай бұрын
@@ServeTheHomeVideo I suspect he's being sarcastic
@lilietto1
@lilietto1 3 ай бұрын
@@mikemotorbike4283 I suspect he just wrote ssd when he meant hd
@marklewus5468
@marklewus5468 4 ай бұрын
I don’t think you can compare a large SSD with a hard drive. A Solidigm 61TB SSD costs on the order of $120 per terabyte and a 16-22tb IronWolf Pro hard drive is on the order of $20 per terabyte. Apples and oranges.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
So the counter to this is that they literally cannot make 61.44TB drives fast enough and big orders are coming in already for 122.88TB next year. There is a per device cost in favor of HDD but higher performance, reliability, and endurance. In the DC swapping to high capacity SSDs can save huge amounts of space and power. Power is the big limiter right now.
@jaimeduncan6167
@jaimeduncan6167 4 ай бұрын
Great overview. We need to force people to understand the MTTR metric, even IT professionals (software) sometimes don't get how important it is. In fact a 20TB HDD drive is a liability even for RAID 6 equivalent technologies (2 drive failure). In particular, if all your rives were bought at the same time, from the same vendor they are likely to come from the same batch. Clearly, the relation of price per by of a 20 TB vs an U.2 16TB ssd is bast and you can buy something more sophisticated and don't worry as much of MTTR.
@nadtz
@nadtz 4 ай бұрын
For my use at home I grabbed some P4510's used, they were all at 99% life left and have been chugging along for a couple years now. Stating to think about upgrading to some gen 4 drives so I've been hunting ebay but I think I'll wait for prices to drop again since they've gone up recently. Your 2016 study and a lot of people reporting use on drives they bought on forums made me worry a lot less about buying used. Always the possibility of getting a dud but I've had good luck so far.
@LtdJorge
@LtdJorge 4 ай бұрын
Sshhhh, Patrick, don’t tell the enterprise customers they’re overbuying endurance. It lets those trickle down at low prices to us homelabbers 😅
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Fair
@LtdJorge
@LtdJorge 4 ай бұрын
@@ServeTheHomeVideo hehe
@honkhonkler7732
@honkhonkler7732 3 ай бұрын
Ive had great reliability from SSDs, I just cant afford the ones that match hard drives for capacity. At work though, we just bought a new vxrail setup thats loaded out with SSDs and the performance improvement from the extra storage speed is more noticeable than the extra CPU resources and memory.
@jeffcraymore
@jeffcraymore 4 ай бұрын
Western Digital Green survived less than a month, having a server as a docker host. Using docker as distributed computing. spawning multiple instance every day. I'm running blues now and they haven't failed yet, but there are some os level issues that point to data corruption.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Yea greens :/
@cyklondx
@cyklondx 4 ай бұрын
The endurance is meant for the disks to last so we don't have to replace them in 2-4 years, they can sit there until we decom whole box... thats the idea of having a lot of endurance.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
DWPD endurance ratings on DC drives are for 5 years, so 2-4 should not be an issue.
@MichaelCzajka
@MichaelCzajka 4 ай бұрын
The takeaway message seems to be that SSD's are ~10x more reliable than mechanical drives: Helpful to know that SSD's in servers have almost eliminated HDD failures. Helpful to point out that larger SSD's help improve reliability. Mechanical HDD's have to swapped out every ~5 years even if they've had light use. That's starts to get very expensive and inconvenient. SSD's are a much better solution. Most users just want a drive that is not going to fail during the life of the computer. The lifespan of many computers might be 10 years or more. NVMe drives are great because you get speed, small form factor and low price all in one package. The faster the drive the better in most cases... especially if you like searching your drives for emails or files. My key metric remains total data written before failure... although it is useful to know over what time period the data was written. I've yet to have an SSD fail. Most of my SSD's live on in various upgrades e.g. Laptops. That means that old SSD's will continue to be used until they become obsolete. It's rare to see meaningful useability data on SSD's. Nicely done. 🙂
@Angel24112411
@Angel24112411 3 ай бұрын
recipe to have SSD failure: fill it up, then use the remaining 6-7 GB to write/rewrite stuff. it quickly develops errors, sometimes silent errors - until file is non readable you don't get any warning.
@pkt1213
@pkt1213 4 ай бұрын
My home server gets almost 0 driverights per day. It gets read a lot, but every once in a while, photos or movies are added.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Great example. Photos and movies are sequential workloads as well
@Proton_Decay
@Proton_Decay 4 ай бұрын
With per-TB prices coming down again, it would be great to know how SSDs perform long-term in home NAS applications -- much higher temps 24/365, low writes but lots of reads and regular ZFS scrubs. Do they outlast spinning rust? So much quieter, I hope to transition my home NAS at some point in the coming couple of years.
@axescar
@axescar 4 ай бұрын
Thank you for sharing your experience. What can be interesting is some heavy load mssql/oracle ssd database storage.
@miss-magic-maya
@miss-magic-maya 4 ай бұрын
This is something I've been curious about for a while, glad to see it tested! Also so many bots ;_;
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Yea so many! We have been collecting the data for a long time, we just have not shared it since the 2016 article.
@mika2666
@mika2666 4 ай бұрын
Bots?
@patrickdk77
@patrickdk77 4 ай бұрын
I have several intel 311 (20g) I should upgrade (purchased 2010), as zfs slog service, PONH=93683, DWPD=1.53, but everything has been optimized to not write unless needed, and moving everything to containers helped with this even more.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Sweet!
@pierQRzt180
@pierQRzt180 3 күн бұрын
awesome stats and the idea to gt a large enough sample data with many different workloads is great too
@ServeTheHomeVideo
@ServeTheHomeVideo 3 күн бұрын
Thanks! Feel free to share with others!
@bacphan7582
@bacphan7582 4 ай бұрын
I just bought an old 1TB server SSD. It's toshiba one, has been written over 1PB, but it's MLC( 2 bit per cell), so i put a lot of trust to it.
@whyjay9959
@whyjay9959 4 ай бұрын
There are Micron Ion drives with different ratings for types of writes, I think that's from when QLC was new. Interesting, seeing how much write endurance and sustained performance seem to be emphasized in enterprise I kinda thought companies were routinely working the drives to death.
@CampRusso
@CampRusso 4 ай бұрын
😮🤔 great video! Ive seen a few videos of all SSD NAS's and tbought well that is bold. Though now watching this I'm thinking I want to try it too! I happen to have a collection of enterprise SSD's from decomm servers at work. The SMART numbers on these are probably crazy low. This also sounds very appealing from a power/heat perspective. Im always trying to make the homelab more efficient.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
We are down to two hard drives in our hosting clusters, and hoping to shift away from those in the next refresh
@CampRusso
@CampRusso 4 ай бұрын
@@ServeTheHomeVideo 😯 that's right you did mention 2 HDD in the vid. That's awesome. Yeah it's time! 😁 The mobo for my TN Scale box has six sata ports. I have some Intel D3-S4600 and Samsung PM863a to test with.
@tad2021
@tad2021 4 ай бұрын
I think outside of the early gens of non-SLC SSDs, I haven't had any wear out. But far more of those drives died from controller failure, as was the style of the time. 100% failure rate on some brands. I recently bought a around 50 of 10-12 year old Intel SSDs. Discounting the one that was DOA, the worst drive was down to 93%, the next worst was 97%, the rest were 98-99%. A bunch of them still had data (seller should not have done that...) and I could tell that many of them had been in use till about a year ago.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Yea we found many with data still accessible. In 2016 when we did the 2013-2016 population a lot more were accessible and unencrypted
@SkaBob
@SkaBob 4 ай бұрын
The drive wear issue sounds similar to the EV battery wear problem that's going away now as well. On the early EVs like a Leaf only had a 50-70 mile range, so 12,000 miles a year would need 200 battery cycles while a newer car with a 320 mile range would only be 37 cycles for the same miles. I do have a few old SSD probably 6-8 years old that never failed and only get replaced to gain capacity but not used in a server capacity. The only SSD that I remember failing was an old Sandisk ReadyCache SSD from around 2012, it was a small 32GB SSD made to supplement your HDD by caching your most used files so it likely had a high write/read/rewrite load and ran near 100% capacity all the time.
@virtualinfinity6280
@virtualinfinity6280 4 ай бұрын
I think, this analysis contains a critical flaw. SSDs write data in blocks (typically 512k) and writing an entire block is the actual write load on the drive. So if you create a file of a few bytes in size, the drive metrics get actually updated by the amount of data you transfer to the drive. By using 512k blocks, the actual write load on the drive is significantly higher. Or in essence: it makes a whole universe of a difference, if you do 1 DWPD by writing the drives capacity using 1-byte-files .vs. you write one big file with the drives capacity as filesize.
@DarkfireTS
@DarkfireTS 4 ай бұрын
Would you resell a few after the testing is done…? Homelabber hungry for storage here 🙂
@jwdory
@jwdory 4 ай бұрын
Great video. I am also interested in some additional storage.
@scsirob
@scsirob 4 ай бұрын
Kinda confirms my statement from a couple of years ago. In the end we'll have just two types of storage. 1. SSD 2. Tape
@UmVtCg
@UmVtCg 3 ай бұрын
LOL You are mistaken, media for storing data will evolve. In the future, huge amounts of data will be stored in glass. Project silicon
@dataterminal
@dataterminal 4 ай бұрын
I've given up telling people this. Even back when I had a 64GB SSD drive as my main boot drive, I was treating it as a harddisk because at the time if it died, I was just going to replace it. It didn't die, and I ended up having by in far way more data written to it than my harddisks, and by the time I had upgraded to a bigger drive, I was no where near the limit of the TBW the manufacture said. For home users at least, you're not going to write/wear the NANDs out, and haven't since SSD, never mind m.2 NVMEs.
@Lollllllz
@Lollllllz 4 ай бұрын
a nice thing is that you'll get a decent amount of warning to move data off a drive that has reached its endurance limit as they usually dont drop like flies when that limit is reached
@kevinzhu5591
@kevinzhu5591 4 ай бұрын
The NAND may be fine, but the controller could have issues as well whether by firmware bug, thermal design or just random shorts on the board. Although controller failure rarely happens.
@Henrik_Holst
@Henrik_Holst 4 ай бұрын
The issue you mentioned with MTTR is why I have begun to wonder if there might not be a place for something like a delayed (or lazy) Raid 1, aka instead of pure spares or a pure Raid1 the data is synced over on a much delayed basis. That would decrease the number of writes on the second pair of drives alot (since multiple rewrites on the primary disk would be a single sync to the secondary) so once the primary signals near death then you already have almost pristine data on the secondary drives while also having a much less risk of them failing in the near future.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
If you wait for parity, then you are vulnerable. If you are just wanting to avoid a same-disk failure at a point in time, then you can just mix drives.
@Henrik_Holst
@Henrik_Holst 4 ай бұрын
@@ServeTheHomeVideo didn't mean to wait for parity (raid1 have no parity), mixing drives are easy in theory and I guess for you guys that but lots of used stuff, we in the enterprise sector have to but hundreds of new drives in batches and mixing them at that point is almost impossible.
@jasongomez5344
@jasongomez5344 4 ай бұрын
I suppose the sequential writes applies to hibernation files too? The biggest cause of SSD wear on my laptops is likely to be from hibenation file writes, as I set them to hibernate after a certain period of inactivity.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
That is less prevalent in servers since they are on 24x7
@redtails
@redtails 2 ай бұрын
I have a low-end Crucial 4 tb SSD which is rated for only 0.1 DWPD over 5 year lifespan. Important to check what a drive is rated for. Now ~400gb per day is still a lot, but I use it primarily for mysql databases for various projects so it's doing around 100 gb/day. Nothing to worry about but it's easy to write 400gb per day to a drive like this for bigger workloads
@mehtotally3101
@mehtotally3101 4 ай бұрын
Correct me if I am wrong, but the DWPD is only rated for the 3-5 year "lifespan" on the drive. So 1 DWPD for three years on a 1TB drive means approx. 1095 drive writes. If you have the drive in service for 10 years, that means it would only be able to handle .3 DWPD. So the proper way to evaluate these drives is really- total rated drive writes vs total drive writes performed. The flash drives take essentially no wear from reads or even being powered on so their lifespan is really gated by how much of their total write capacity has been used up. I have never understood why the metric was per day. Who cares when the writing is done, the question is how much writing has been done.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Usually five years 4K random write. You are correct PBW is a more useful figure which is why we did this piece to show why DPWD is not a good metric anymore. Also the type of writes impacts how much you can write. Usually rated DWPD is much lower than actual
@cameramaker
@cameramaker 4 ай бұрын
@@ServeTheHomeVideo the DWPD is more useful than PBW, because it is not a function of capacity. The DWPD figure can easily split into read-intensive (low dpwd) and write-intensive (high dwpd) kind of drives. Also, you have some sort of online service, which eg. accepts a 1Gb/s of continuous feed and you need to save or buffer that - so you go with 86400 Gb/day, which is 10800 GB = 10.8 TB. So all you care about is to have either a 10.8TB of 1DWPD drive, or 3.6TB of 3DWPD drive, to be on the safe side of 5y warranty. With PBW metric you are much more complicating the formulas for such streaming/ingest use case.
@MichaelCzajka
@MichaelCzajka 4 ай бұрын
My drives usually get upgraded at regular intervals: I'm always looking for faster drives i.e. PCIe3 -> PCIe4 -> PCIe5 Bigger drives are also desirable as you want a bit of overcapacity if possible. Overcapacity is less of an issue if the drive is mainly read (storage) rather than written to. Total number of writes is the most useful metric as it predicts failure. However as drive speed increases the number of potential writes also increases. If you have a fast drive you'll find the number of detailed searches you do is likely to increase. The amount of data you write to a fast drive is also likely to increase... as some of the more time consuming tasks become less onerous. If a drive has an expected lifespan of 10 or more years... that's when you don't have to constantly monitor your drives for failures. That's one less thing to worry about on your computer. Drive metrics often make the expected lifespan quite hard to work out. Early on there were a lot of SSD failures. Nice to see that the situation has now reversed. Doesn't seem to be any manufacturer with an SSD reliability problem. 🙂
@tbas8741
@tbas8741 4 ай бұрын
MY Old System (Built in 2014, Retired in 2024) The HDD Stats in that heavily used system are - Western Digital Hybrid SD-HDD 7200rpm (32mb ssd cache on the sata interface) - Power on Hours 92,000, But i kept the computer running on average 24/7/365 for over 10 years.
@SantiagoBiali
@SantiagoBiali 4 ай бұрын
Bought 8x Enterprise SSDs 6 years ago, all of them still have 99% TBW Remaining since day 1. Last year I bought 70+ used Enterprise SSDs to fill 3x NetApp DS2246 (3x 24 bay storage) and guess what? Most of them had 3~4 years of power-on and 90%+ TBW Remaining. Oh, these are Enterprise drives so they are spec to 7~12PBW (Petabyte-writes) of total endurance.
@whereserik
@whereserik 4 ай бұрын
I value this. Thank you
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Thanks!
@artemis1825
@artemis1825 4 ай бұрын
Would love to see a version for used SAS enterprise HDDs and their failure rate
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Sure but the flip side is we stopped using disks several years ago except in special use cases
@artemis1825
@artemis1825 4 ай бұрын
@@ServeTheHomeVideo Ah I guess I could always check the surveys from hyperscalers.
@masterTigress96
@masterTigress96 4 ай бұрын
@@artemis1825 You can check the statistics from BackBlaze. They have been analyzing drives for many, many years as they are a back-up as a service provider, so they definitely need cheap, reliable long-term storage devices.
@Brian-L
@Brian-L 4 ай бұрын
Does Backblaze still publish their annual spinning rust analysis?
@PeterBlaise2
@PeterBlaise2 4 ай бұрын
Can you please test the data access rates and data transfer rates to see if the used drives are really performing according to manufacturer promises? Steve Gibson's GRC free ReadSpeed acknowledges "... we often witness a significant slowdown at the front of solid state drives (SSDs), presumably due to excessive use in that region ...". And free HDD Scan and free HD Tune can show us graphs of the slow or even unreadable sectors. And then SpinRite 6.1 Level 5 or HDD Regenerator will show the qualities of every sector's REWRITABILITY. Without that information, it's impossible to know the true value of any of those SSDs, really. Let us know when you have the next video with a full analysis of the SSD drive's REAL qualities to READ and WRITE compared to manufacturer performance specifications. Thanks. .
@comp20B
@comp20B 4 ай бұрын
I have been keeping to 5yo old Dell enterprise hardware. Currently my need is just 8TB within truenas. Enterprise SAS SSDs have been a huge leap for my use.
@shutenchan
@shutenchan 4 ай бұрын
I actually bought tons those intel S3510/S3520 ssd's from my own workplace (I work on a data center), they're very cheap and has high endurance with decent speed (although slower sequential speed).
@Vegemeister1
@Vegemeister1 4 ай бұрын
Intel drives were known for going read-only and then bricking themselves on the next power reset when lifetime bytes written hit the warranty limit, whether those had been small-block writes or large sequential, and whether or not the drive was still perfectly good. Does Solidigm retain that behavior?
@EdvardasSmakovas
@EdvardasSmakovas 4 ай бұрын
Did you analize data writes only, or nand writes as well? I think write amplification factor shoud be mentioned in this context. Since depending on your storage array setup, this could result in many orders of magnitude more writes.
@computersales
@computersales 4 ай бұрын
I prefer buying used DC drives because they always have a ton of reads and writes but are always reporting over 80% health. Im not as keen on consumer drives. I don't use it as much as I could but my 1TB P3 is already down to 94% health after a year. Granted it has a lot of life still but a DC drive wouldn't even flinch at 14TB writes.
@chuckthetekkie
@chuckthetekkie 4 ай бұрын
I think one reason that people still use HDDs is due to the lower upfront cost compared to SSDs. I have 2 4TB Intel DC 4510 U.2 SSDs that I bought used last year for about $200 per drive. Each drive has over 4 years 8 months of power on time and about 6.4PB read. I do have HDDs that I bought over 10 years ago that have way more power on hours and still work just fine. I myself and I know other people still worry about SSD reliability. Especially when it comes to QLC. I have a friend who REFUSES to even touch QLC SSDs. Usually with HDDs they start acting up before they fail where as SSDs will just stop working altogether with no warning. That has been my experience with HDDs and SSDs. I have had 2 m.2 NVMe SSD fail on me with no warning. One just stopped being recognized and the other actually Kernel Panics my Mac when plugged in. They were both being used in an NVMe USB enclosure. For SSD endurance I only ever looked at the TBW (Terabytes Written) rating. Never cared about DWPD.
@kiiverkk
@kiiverkk 4 ай бұрын
Just remembered some old test online that tested consumer SATA SSD endurance, turned out that actual reliability was like 10+ times higher as in 200TB was spec but errors appeared about 2PB write range (don't remember exact numbers but range was similar). Wonder how big or narrow the rating vs actual data loss level is for todays NVME-s, sure less as the chips have tighter tolerances and can be pushed to the limits just like CPU-s that are mostly factory overclocked today.
@5467nick
@5467nick 4 ай бұрын
Yeah, from what I've read earlier (especially old MLC SATA) SSDs were often vastly underrated in endurance whereas modern SSDs are rated pretty close to what they can handle before bricking. Problem is a lot of SATA SSDs had severe design or firmware flaws that lead to flash endurance not being at all a factor in actual reliability. Sandforce, we're all looking at you.
@tomreingold4024
@tomreingold4024 4 ай бұрын
Fantastic. Very informative. I used to run data centers. I've switched drastically and am now becoming a school teacher. I don't know where I will use the information you just provided, but I enjoyed learning it.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Glad it was helpful! Maybe it becomes a lesson one day. Thanks for being a teacher!
@tomreingold4024
@tomreingold4024 4 ай бұрын
@@ServeTheHomeVideo hey you never know. As I prepare to be a special ed teacher for math, English, social studies and science, maybe I'll end up teaching IT.
@foldionepapyrus3441
@foldionepapyrus3441 4 ай бұрын
When you are talking about drives though as they are so crucial to your desktop/server actually being functional, which for many is essential for their income stream its worth picking for a more certainly going to outlast your interest spec than run near the edge and get burned - transferring your drives to a new system if/when you upgrade or replace a failure is quick and painless for the most part.. Plus even with fast drivers any serious storage array takes a while to rebuild, so avoid that is always going to be nice.
@charlesspringer4709
@charlesspringer4709 4 ай бұрын
Wow what a mass of words! Should I get SSD's and which ones?
@lukasbruderlin2723
@lukasbruderlin2723 4 ай бұрын
would have been nice if you would have given some examples of SSD drives that do have less higher ratings and therefore are less expensive but still are reliable,
@ABaumstumpf
@ABaumstumpf 4 ай бұрын
We had some problems with fast SSD failures... They didnt experience any high write-rates, they never got even close to be full, but they failed at an alarming rate in comparison to the normal HDDs we used (in raid5). Turns out that it is just the usecase: most of the write comes from logfiles and the SSD controllers were not all that great with constant tiny writes to just a few files. Switched to some SAS SSDs and no problem since then. and ok our usecase was also a bit special as we mostly needed RAM and CPU performance, decent networking and the storage only needed to last. In general for any active storage i will go hybrid - SSD for main stuff, HDD for bulk storage that really does not need any high performance. And of course for any cold storage SSDs are just a no-go.
@stevesteve8098
@stevesteve8098 4 ай бұрын
logfiles will destroy an ssd, it's what takes down most consumer equipment embedded linux writing log files to the device ssd or Nand flash. He talks about having a couple of HD failures, it's fine becasue i have a load of >70,000 hour hard drives with ZERO failures & errors. but one absolutely critical point he missed in al lthis data........ was the drive temp, without it , these figures mean sweet FA.
@ABaumstumpf
@ABaumstumpf 4 ай бұрын
@@stevesteve8098 yeah, logfiles are kinda evil for that. And temps - yeah. But it would assume (or at least hope) that they run their hardware in a somewhat controlled environment and all under similar conditions. I do remember the times of school-computers being tiny machines with just 1 small fan, stuffed into a wooden cabin with only some holes at the top - and the hardware getting absolutely cooked for years :D
@henderstech
@henderstech 4 ай бұрын
wow I wish I had just one of those large SSDs! I had no idea They made them so l with so much capacity.
@harshbarj
@harshbarj 4 ай бұрын
I'd move to SSD, but there is one MASSIVE barrier. Cost. Right now my 2 drive array cost me under $150 for 8TB of storage. As of this moment the cheapest 8TB used enterprise SSD I can find is $1099. So my array as an SSD solution would cost me $2200, rather than the ~$150 it cost me today.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
7.68TB DC SSDs can be purchased new other (e.g leftovers and spares) for $500ish.
@mystixa
@mystixa 4 ай бұрын
Good analysis but with an oversight. Ive had many SSDs and HDDs fail over the years. The problem being a lot of time an SSD will fail quickly and then be unrecoverable en mass with 100% data loss. An HDD often fails progressively with errors showing up in scans or with bad behaviour. When data from some sectors is lost almost always some of the rest is saveable. With the appropriate backup strategy this makes it less of a problem of course. It does shift the emphasis of how one cares for the data though.
@DrivingWithJake
@DrivingWithJake 4 ай бұрын
We mostly seen only people who really abuse drives run into issues even. Our most used drives we find are for data bases that uses the most life out of them. Other than people trying to use for mining. Smallest nvme we use is 1tb as defaults but have a lot of 15.36TB's for the past 4-5 years now.
@heeerrresjonny
@heeerrresjonny 4 ай бұрын
Maybe this is just because I have only ever purchases consumer SSDs, but I have been using SSDs for over a decade and I have never once seen a drive have a DWPD rating listed (in fact, this video is the **first** time I have ever encountered that metric in all these years lol). Endurance has always been rated using TBW. EDIT: also, now that I've looked into it, it seems manufacturers calculate "DWPD" based on the warranty period... but that doesn't make sense to me. It should use MTBF for the time component. This would make all the DWPD numbers WAY smaller, but more "objective")
@raylopez99
@raylopez99 4 ай бұрын
In a different context, this reminds me of Michael Milken of Drexel Burnham fame: he found that "junk bonds" were unfairly shunned when in fact their default rates were much less than people expected (based on data from a finance professor, which was the original inspiration). Consequently he used junk bonds to his advantage and as leverage to takeover companies (which had a lot of corporate fat, back in the day). How Patrick can profit from his observation in this video is less clear however, but I hope he achieves billionaire status in his own way.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Ha! I wish
@tsclly2377
@tsclly2377 4 ай бұрын
Crypto mined with used DL580 G7s and blew up Intel consumer Intel 545 500GB drives (4 @ RAID 5) in 5 months.. Yup, dead.. no loading of the 25GB/day DAG files even after 'reconditioning'/downsizing... but the 64GB HP SLC drives worked for years (3.5). What do you think of the Pliant 406S drives? Actually I have found out that the new write endurance drives being sold today to the server farms are much lower in Petrabyte writes per GB than I've mentioned, although faster. Now an Optane fan (for using in LLM retraining models).. For my gamer kid, I got one of those combined Optane M.2 drives and it is going string for 3 years and a used Intel S3710 (plus a 6TB spinner)..
@MasticinaAkicta
@MasticinaAkicta 4 ай бұрын
So they were used more as caching drives in servers that didn't need THAT much space. BUT... it needed speedy cache.
@Zarathustra-H-
@Zarathustra-H- 4 ай бұрын
You don't think that maybe your data set might be skewed due to sellers not selling drives where they have already consumed all or close to all of the drives write cycles? Because of this, I just don't think your sample is truly random or representative.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
That would have been a bigger concern if we were buying like 5+ year old drives. Normally we are buying 2-ish year old models and so it is much less likely they can get written through at that pace. This is especially true since we are seeing sub 10% duty cycle on the vast majority of drives. Also l, remember a good portion of these are not even wiped as we showed, so if people are not wiping them they are unlikely to be looking at SMART wear data.
@Zarathustra-H-
@Zarathustra-H- 4 ай бұрын
@@ServeTheHomeVideo The fact that they are not wiping them is pretty shocking actually.
@drd105
@drd105 4 ай бұрын
storing a lot of videos is a pretty niche use. VMs are in much more mainstream use. It's easier to keep old VMs around than treat configuring systems as a lifestyle choice.
@iiisaac1312
@iiisaac1312 4 ай бұрын
I'm showing this video to my SanDisk Ultra Fit USB Flash Drive to shame it for being stuck in read only mode.
@kelownatechkid
@kelownatechkid 4 ай бұрын
optane for write-heavy/DB workloads and literally whatever else for bulk storage haha. Ceph especially benefits from optane for the db/wal
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
If you saw, we bought a lot of Optane and have an entire shelf of it
@Michael_K_Woods
@Michael_K_Woods 3 ай бұрын
I think the main reason system guys like the high drivewrites per day is the implied hardiness. They will pay the extra money for 16 over a 4 if they believe it decreases maintenance and disruption odds.
@FragEightyfive
@FragEightyfive 4 ай бұрын
I would consider myself a power user and looking at some of my primary SSD's from the mid 2010's, I'm at about 0.12DWPD based on hours....And the second oldest/most used 256GB drive that still sees near daily use on a laptop, is still at 83% Drive LIfe Remaining. When I first started using SSD"s, I kept track of usage statistics. I stopped doing that after a few years when I realized that on paper, the NAND will last at least 100 years. Something else is going to cause a failure than drive writes (except maybe bad firmware that writes too much to some cells). I have been working with some large data sets on my main desktop more recently (10's to 100+GB), and even the 2TB and 4TB NVME drives are a similar DWPD, and at 95% after 2 and 5 years.
@45KevinR
@45KevinR 4 ай бұрын
It's really interesting, and suggests to me that for general population drives, the obsession with getting enterprise vs retail drives is unnecessary, at least for endurance. Now write caching, safe cache writing, perhaps the lifespan of the electronics might all push a user towards enterprise drives - but you've basically proved endurance isn't a factor. However if the hosting/cloud industry still worry about it - it should mean a steady supply of retired drives that a home user or even STH can get cheaply! 😎👍 I'd like to hear your thoughts on say: zfs write caches or daily backup targets. Presumably they would be subjected to much fiercer write traffic (though probably sequential) - and endurance might get tested more. Though I guess even a backup target would max out at say 1 drive write per 24 hrs. Less of you keep multiple days on the one drive/volume. Which only leaves zfs (or similar) as a concern. Thoughts? On rebuild time, I guess that takes us back to the original axiom of RAID. Array of *inexpensive* disks. When RAID was first conceived HDD were small and fragile by today's standards, large drives were expensive. So RAID facilitated using smaller cheaper drives to make a large volume that was fault tolerant. These days people have forgotten the small and wide, and you get say 12Tb in a mirror. However that's pricey to replace, and the rebuild will take hours to days and might even stress the new drive. 100% writes for hours and hours isn't the normal use or failure mode. So the rebuild might be the hardest use the drive ever gets! And a small NAS might only support 2-4 drives. Though I guess this also reminds us that a RAID volume isn't a backup, it's a resilient original. A real backup is our safety net until the rebuild is complete. (sorry for essay. A thought provoking video.) hope it keeps my paragraphs, 😮
@45KevinR
@45KevinR 4 ай бұрын
To rain on my own parade, looking at solidigm's own workstation M.2 drives and using the warranty period of 5 years you only get 660p = 0.1 to 670p = 0.2 drive writes per day. So they are a bit on the edge in server use.
@geozukunft
@geozukunft 4 ай бұрын
I have 2 Micron 7450 Max 3.2TB that I bought in June last year they running in Raid 1 for a database for a hobby project of mine and at the moment I am sitting at 3.7DWPD compared to the 3DWPD that they are rated for D:
@BangBangBang.
@BangBangBang. 4 ай бұрын
the Intel 5xx series SSDs we used to deal with back in the day would be DOA or die within 90 days, otherwise they were usually fine.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Yea I worked with a company who was seeing over 50% AFR on certain consumer Samsung SATA drives in servers
@cjcox
@cjcox 4 ай бұрын
I think with regards to normal (not unusual cases), the outage scenarios due to nand wearing out due to writes, would be cases where by algorithm or lack of trim you were hitting a particular cell with writes more than others. So, the TWD sort of thing goes out the window when talking about those types of scenarios. The good news there? Rare. Just like the other situations you mentioned. With that said, SSD quality can be an issue. I have RMA'd a new Samsung SATA SSDs (it was a 2TB 870 EVO) that started producing errors in the first year. So, there are elements apart from NAND (assuming good) lifetime as well. I think those are the errors that are more likely to occur.
@RichardFraser-y9t
@RichardFraser-y9t 4 ай бұрын
The Penta or Septa cells might only last 500 re-writes but they are nothing to worry about,
@Koop1337
@Koop1337 4 ай бұрын
So like... Can I get some of those drives now that you're done testing them? :)
@Zarathustra-H-
@Zarathustra-H- 4 ай бұрын
Just for shits and giggles I ran the DWPD on all of the SSD's in my server. The highest was on my two Optane's (which I use as mirrored SLOG drives). They have a whopping ~0.1 DWPD average over ~3 years. :p
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Exactly :)
@acquacow
@acquacow 4 ай бұрын
I just built a whole new nas on 1.6TB Intel S3500s with 60k hours on them all a few months ago =p I'm all about used flash.
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Sweet!
@armstrongskyview2810
@armstrongskyview2810 4 ай бұрын
Where do you buy the used ssd's?
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
We went into this a bit more on the main site, but mostly forum members, ebay, recyclers, and so forth.
@ChipsChallenge95
@ChipsChallenge95 4 ай бұрын
I’ve worked with and worked for many companies (hundreds) over the last couple decades, and every single one of those companies destroyed their drives after use or contracted someone to do it and are provided certificates of destruction. Idk how you managed to find so many used drives.
@npgatech7
@npgatech7 4 ай бұрын
Sorry, if I missed, but did any of your 400+ drives fail?
@ServeTheHomeVideo
@ServeTheHomeVideo 4 ай бұрын
Of the over 2000 we have had 3 in the last 8 years.
@dnmr
@dnmr 4 ай бұрын
@@ServeTheHomeVideo this is including all the used ones right? So the ones driven into the ground
@sarahjrandomnumbers
@sarahjrandomnumbers 4 ай бұрын
Just went through all this with my new nas build. 4x 4tb m.2 nvme sticks in zraid1, and even if you're worried about the DWPD, i've basically quadrupled any life I've got with the sticks cause it's split across the drives. Meanwhile, I have 2 512gb sata drives called "Trash-flash" that I'm using for dumping stuff onto that's going into the disk array. Both SSD's are already twice past their TBW's, and only one of them has a failed block. So panic time, right? Nope, I've got 4402 reserve blocks remaining. 🤣🤣
@funboyrule
@funboyrule 4 ай бұрын
How do you filter rule out any survivor bias?
Tiny 10GbE All-SSD NAS Sports Up to 64TB
19:56
ServeTheHome
Рет қаралды 99 М.
The Death of Memory. New Era of Data Storage
18:18
Anastasi In Tech
Рет қаралды 206 М.
My scorpion was taken away from me 😢
00:55
TyphoonFast 5
Рет қаралды 2,7 МЛН
To Brawl AND BEYOND!
00:51
Brawl Stars
Рет қаралды 17 МЛН
99.9% IMPOSSIBLE
00:24
STORROR
Рет қаралды 31 МЛН
Арыстанның айқасы, Тәуіржанның шайқасы!
25:51
QosLike / ҚосЛайк / Косылайық
Рет қаралды 700 М.
THIS 25GbE Server and Firewall Has it All
22:21
ServeTheHome
Рет қаралды 137 М.
The Wobbly Future of the Hard Disk Drive Industry
18:30
Asianometry
Рет қаралды 190 М.
Enterprise SSDs : The Upgrade You Didn't Know You Needed?!
5:00
ServerPartDeals
Рет қаралды 4,9 М.
What is a SAS SSD?
12:08
Level1Techs
Рет қаралды 38 М.
DON'T Underestimate This Cheap MicroServer
18:07
Hardware Haven
Рет қаралды 457 М.
How Long Can SSD Store Data Unpowered? Year 2 Update
15:26
HTWingNut
Рет қаралды 53 М.
A SSD NAS of the Future Our QNAP TBS-h574TX Review
19:44
ServeTheHome
Рет қаралды 335 М.
Why you DON’T want a 20TB Hard Drive
20:14
LMG Clips
Рет қаралды 1,3 МЛН
NEVER install these programs on your PC... EVER!!!
19:26
JayzTwoCents
Рет қаралды 4,4 МЛН
My scorpion was taken away from me 😢
00:55
TyphoonFast 5
Рет қаралды 2,7 МЛН