Most active commenters
  • dylan604(5)
  • adastra22(3)

←back to thread

420 points speckx | 13 comments | | HN request time: 0.798s | source | bottom
Show context
dylan604 ◴[] No.44533476[source]
SSD speeds are nothing short of miraculous in my mind. I come from the old days of striping 16 HDDs together (at a minimum number) to get 1GB/s throughput. Depending on the chassis, that was 2 8-drive enclosures in the "desktop" version or the large 4RU enclosures with redundant PSUs and fans loud enough to overpower arena rock concerts. Now, we can get 5+GB/s throughput from a tiny stick that can be used externally via a single cable for data&power that is absolutely silent. I edit 4K+ video as well, and now can edit directly from the same device the camera recorded to during production. I'm skipping over the parts of still making backups, but there's no more multi-hour copy from source media to edit media during a DIT step. I've spent many a shoot as a DIT wishing the 1s&0s would travel across devices much faster while everyone else on the production has already left, so this is much appreciated by me. Oh, and those 16 device units only came close to 4TB around the time of me finally dropping spinning rust.

The first enclosure I ever dealt with was a 7-bay RAID-0 that could just barely handle AVR75 encoding from Avid. Just barely to the point that only video was saved to the array. The audio throughput would put it over the top, so audio was saved to a separate external drive.

Using SSD feels like a well deserved power up from those days.

replies(8): >>44533735 #>>44534375 #>>44535266 #>>44535471 #>>44536311 #>>44536501 #>>44539458 #>>44539872 #
1. gchamonlive ◴[] No.44533735[source]
> I come from the old days of striping 16 HDDs together (at a minimum number) to get 1GB/s throughput

Woah, how long would that last before you'd start having to replace the drives?

replies(3): >>44533870 #>>44534101 #>>44538566 #
2. dylan604 ◴[] No.44533870[source]
Depending on the HDD vendor/model. We had hot spares and cold spares. On one build, we had a bad batch of drives. We built the array on a Friday, and left it for burn-in running over the weekend. On Monday, we came in to a bunch of alarms and >50% failure rate. At least they died during the burn-in so no data loss, but it was an extreme example. That was across multiple 16-bay rack mount chassis. It was an infamous case though, we were not alone.

More typically, you'd have a drive die much less frequently, but it was something you absolutely had to be prepared for. With RAID-6 and a hot spare, you could be okay with a single drive failure. Theoretically, you could lose two, but it would be a very nervy day getting the array to rebuild without issue.

replies(1): >>44536396 #
3. nordsieck ◴[] No.44534101[source]
If you're interested in some hard data, Backblaze publishes their HD failure numbers[1]. These disks are storage optimized, not performance optimized like the parent comment, but they have a pretty large collection of various hard drives, and it's pretty interesting to see how reliability can vary dramatically across brand and model.

---

1. https://www.backblaze.com/cloud-storage/resources/hard-drive...

replies(1): >>44534434 #
4. dylan604 ◴[] No.44534434[source]
The Backblaze reports are impressive. It would have been very handy to know which models to buy. They break it down to capacity of the same family of drives so a 2TB might be sound, but the 4TB might be more flaky. That information is very useful when it comes time to think about upgrading capacity in the arrays. Having someone go through these battles and then give away the data learned would just be dumb to not take advantage of their generosity.
replies(1): >>44535004 #
5. doubled112 ◴[] No.44535004{3}[source]
Many years ago I felt like I dodged a bullet splurging for the 4TB Seagates instead of the 3TB Seagates I needed.
replies(1): >>44535201 #
6. HansHamster ◴[] No.44535201{4}[source]
Can confirm. My 3TB Seagate was the only disk so far (knocking on wood) that died in a way that lost some data. Still managed to make a copy with dd_rescue, but there was a big region that just returned read errors and I ended up with a bunch of corrupt files. Thankfully nothing too important...
7. gchamonlive ◴[] No.44536396[source]
I asked because I did a makeshift NAS for myself with three 4tb ironwolf, but they died before the third year. I didn't investigate much, but it was most likely because of power outages and a lack of a nobreak PSU at that time. It's still quite a bit of work to maintain physical hard drives and the probability of failure as I understand tend to increase the more units the array has because of inverse probability (not the likelihood of one of them failing but the likelihood of none of them failing after a period of time, which is cumulative)
replies(1): >>44536690 #
8. dylan604 ◴[] No.44536690{3}[source]
Any electronic gear that you care about must be connected to a UPS. HDDs are very susceptible to power issues. Good UPS are also line conditioners so you get a clean sine wave rather than whatever comes straight from the mains. If you've never seen it, connect a meter to an outlet in your home and what how much fluctuations you get throughout the day. Most people think about spikes/surges, while forgetting that dips and under-volting is damaging as well. Most equipment have a range of acceptable voltage, but you'd be amazed at the number of times mains will dip below that range. Obviously location will have an affect on quality of service, but I hear my UPSes kick in multiple times a week to cover a dip if only for a couple of seconds.

The fun thing about storage pools is that they can lull you into thinking they are set it and forget it. You have to monitor SMART messages. Most drives will give you a heads up if you know where to look. Having the fortitude to have a hot spare instead of just adding it to the storage pool goes a long way from losing data.

9. adastra22 ◴[] No.44538566[source]
I run 24x RAID at home. I’m replacing disks 2-3 times per year.
replies(1): >>44538944 #
10. dylan604 ◴[] No.44538944[source]
Are your drives under heavy load or primarily just spinning waiting for use? Are they dying unsuspectedly, or are you watching the SMART messages and being prepared when it happens?
replies(1): >>44541085 #
11. adastra22 ◴[] No.44541085{3}[source]
They’re idle most of the time. Poweted on 24/7 though, and maybe a few hundred megabytes written every day, plus a few dozen gigabytes now and then. Mostly long-term storage. SMART has too much noise; I wait for zfs to kick it out of the pool before changing. With triple redundancy, never got close to data loss.

To be clear, I should have said replacing 2-3 disks per year.

replies(1): >>44542869 #
12. somehnguy ◴[] No.44542869{4}[source]
That seems awfully high no? I've been running a 5 disk raidz2 pool (3TB disks) and haven't replaced a single drive in the last 6ish years. It's composed of only used/decommissioned drives from ebay. The manufactured date stamp on most of them says 2014.

I did have a period where I thought drives were failing but further investigation revealed that ZFS just didn't like the drives spinning down for power save and would mark them as failed. I don't remember the parameter but essentially just forced the drives to spin 24/7 instead of spinning down when idle and it's been fine ever since. My health monitoring script scrubs the array weekly.

replies(1): >>44544197 #
13. adastra22 ◴[] No.44544197{5}[source]
Drives I RMA have actual bad sectors. You have a good batch. These drives tend to either last 10+ years, or fail in 1-3 years, and there is a clear bimodal distribution. I think about half the drives in my array are original too.