Any Reccomendations For A New Hard Drive Purchase?

Hey guys. :) It's going to be my birthday on July 19th! :D So Mom is getting me a 2TB Hard Drive, which of course is going into my Server. :) But I've seen a few kinds... One is Western Digital's Blue series, which they say is for Servers I believe. Can I get by with just a plain 'ol 2TB Hard Drive that really can be used for anything, but in my case, for my Server? :) I ask because there's just so many models I could choose from... Then, I hope to attach the 300GB one I have now to a RAID controller for ZFS. :) I'll have a bunch of room for Virtual Machines and Jails, whoopeeee! :D Thanks for any replies guys! :)
 
I would consider dropping the 2TB requirement and consider SSD's. One Terrabyte SSD's are around $140 and they are literally 3-4x quicker. Then you could use the 300GB for backups.
Remember that Virtual Machines take quite a disk speed hit so you want to start with something resonably quick.
There is also the NVMe drive factor. I am buying Samsung PM983 Enterprise 1TB drives for the same cost as a retail SSD.
Booting could be an issue with these on older motherboards.
 
If you do stick with hard drives they sell them like car batteries.
You buy a battery with 24 month warranty it lasts 25 months maybe.
Buy a HardDrive with a 24 month warranty and it lasts how long???
Point is 5 year drives are more reliable than 2 year drives.
 
Here is a great barometer. In one hand hold an aluminum body consumer drive. In the other a 5 year SAS drive.
Blindfolded I can tell which is built to last.
The SAS drives use heat treated parts which weigh more and are much more dense. Designed for 24x7x365 duration.
 
Even SAS drives you have to know what your looking at. There is a whole racket called "mid-line drives"
These are SAS drives with 3 year warranty.

I know your probably going to stick with SATA drives. There are some great deals on old LSI controllers from OEM on ebay for SAS.
 
I would consider dropping the 2TB requirement and consider SSD's. One Terrabyte SSD's are around $140 and they are literally 3-4x quicker. Then you could use the 300GB for backups.
Remember that Virtual Machines take quite a disk speed hit so you want to start with something resonably quick.
There is also the NVMe drive factor. I am buying Samsung PM983 Enterprise 1TB drives for the same cost as a retail SSD.
Booting could be an issue with these on older motherboards.
Ok, I'll consider that. :) We also have a limited budget though (I have Autism, so I live with Mom). :) But $140 doesn't sound unreasonable, especially for a Boot Drive... :D Yeah, this Server is a friend's and mine. :) He put Debian on it, which was great, but FreeBSD with ZFS... You get the story, as I'm no fanboy, but I saw the benefits of this great OS. :) I may just settle on the HDD option, but the SSD one is one worth considering. Thanks! :)
 
If you do stick with hard drives they sell them like car batteries.
You buy a battery with 24 month warranty it lasts 25 months maybe.
Buy a HardDrive with a 24 month warranty and it lasts how long???
Point is 5 year drives are more reliable than 2 year drives.
So you're thinking about the top-of-the-line then... Well, I should also mention that this Server is just used as backup, and occasion reading from... :) For the VMs, they'll be used for a Honeynet, mostly, while some other's will be for Dev purposes. :)
 
Here is a great barometer. In one hand hold an aluminum body consumer drive. In the other a 5 year SAS drive.
Blindfolded I can tell which is built to last.
The SAS drives use heat treated parts which weigh more and are much more dense. Designed for 24x7x365 duration.
Ah, ok then! :) SAS... I'll keep that in mind! :) I bet they have it at Staples here. :)
 
Even SAS drives you have to know what your looking at. There is a whole racket called "mid-line drives"
These are SAS drives with 3 year warranty.

I know your probably going to stick with SATA drives. There are some great deals on old LSI controllers from OEM on ebay for SAS.
Ok then... So SAS definitely sounds like something I should consider. :) I'll also look into the LSI Controllers. :)
 
I currently have 2 (hitatchi?) 15k rpm SAS drives (300gb) running on an LSI controller in my build server. They are probably 9-10 years old. They are hot and a little loud but as phishfry said, dead reliable. Of course I don't run this thing very often so the drives will most likely last a very long time.
 
I’ve been very pleased with HGST drives for servers.

Are these still available?
I highly recommend them!

Code:
  9 Power_On_Hours          0x0012   091   091   000    Old_age   Always       -       66450
10 Spin_Retry_Count        0x0013   100   100   060    Pre-fail  Always       -       0
12 Power_Cycle_Count       0x0032   100   100   000    Old_age   Always       -       608
192 Power-Off_Retract_Count 0x0032   087   087   000    Old_age   Always       -       15890
193 Load_Cycle_Count        0x0012   087   087   000    Old_age   Always       -       15891
194 Temperature_Celsius     0x0002   187   187   000    Old_age   Always       -       32 (Min/Max 18/68)
196 Reallocated_Event_Count 0x0032   100   100   000    Old_age   Always       -       55
197 Current_Pending_Sector  0x0022   100   100   000    Old_age   Always       -       0
198 Offline_Uncorrectable   0x0008   100   100   000    Old_age   Offline      -       0
199 UDMA_CRC_Error_Count    0x000a   200   200   000    Old_age   Always       -       0

Errors come above 65C - but don't seem to harm. And that one's only consumer (deskstar), but runs in continuous operation. Have a couple of them, all fine.
 
Ok then... So SAS definitely sounds like something I should consider. :) I'll also look into the LSI Controllers.
I don't know if this was such a great suggestion. There are plenty of enterprise 5 year SATA drives.
They do carry a heavy price premium.

On the other side of the coin I don't know that I would feel comfortable with the SMART that PMc is showing.
So, much of this comes down to what you want to spend to be happy.
 
On the other side of the coin I don't know that I would feel comfortable with the SMART that PMc is showing.

Point is, You're definitely not supposed to bring them above 60°C in operation. So, I thought, lets see what happens next - and that was 30'000 hours ago.
 
There are of course lots of opinions on this. My amateur experience has only had Seagate drives from a specific generation fail, that's a few years back, so probably not relevant any more. It seems that most drives do just fine in a home environment and it doesn't really matter what it is. I'm just rebuilding my main desktop and happened to notice the original installation date that I had written on main drive. It's a Western Digital black 500GB drive, and it's been running continuously 24/7 for 6 years now. I expect it will keep going for a long time more.

Failures are rare, and only a professional user who goes through many drives will see them. The chance that an amateur will see a failure is very small.

Flames welcome. :)
 
6 losses in 31 years, 2 of them pilot errors (wrong power supply, and bought as used and sent in a parcel without any cushion - spindle bearings failed after 2 more years), with an average of 4-8 disks running.
The other four:
Seagate 40 MB - died after substantial lifetime
WD 80 GB - died after substantial lifetime
IBM 36 GB 15k - died after 10+ years lifetime
WD 500 GB - died within warranty - was highly neurotic from the beginning, and the replacement is just as neurotic (but now I know how to handle that).
 
1 drive loss in 26 years, but not in server use, desktop use only. Drive was a Maxtor. Had mostly WD and some Seagates. Currently run desktop on Samsung SSDs, had those 3 years. Time will tell...
 
Failures are rare, and only a professional user who goes through many drives will see them. The chance that an amateur will see a failure is very small.

Flames welcome. :)
Not a flame, just a philosophical excursion about statistics.

The MTBF of disk drives is spec'ed by manufacturers as a million hours or more (sometimes 1.5 or 2 million for enterprise grade drives). Having been a user of many thousands of disk drives professionally, and not being able to share accurate statistics, my summary is that the disk manufacturers are sort of honest. The actual measured failure rates are perhaps half or two thirds of the specification. Now, some of those failures are probably not the fault of the drive (temporary overheating, too much vibration, flaws in power supplies), so this is not intended as criticism of Seagate, WD, Hitachi, Toshiba and friends. This is experience from drives that are installed in professional-grade enclosures (not computer cases but dedicated disk enclosures), with high-quality power supplies, and installed in well managed data centers. Customers who spend millions on their storage systems tend to not risk those systems due to inadequate environments, that would be dumb. So let's take a real-world MTBF (under professional conditions) of about 1/2 to 1 million hours:

Lesson #1: The MTBF of disk drives in good conditions is very high, about 50-100 years average life time = 1% to 2% annual failure rate, and therefore amateurs with just a handful of drives should very rarely see disk failures.

Quick observation: it is quite possibly to break a disk. Drop it (just a little bit) while it is spinning and seeking (laptop disks are better at that). Drop it from a foot height onto a granite table while powered down. Cook it (the 68 degrees above is bad news). Vibrate it all the time while writing. Connect it to a power supply whose 12V line can be 11V or 14V depending on the phase of the moon. Switch the 5V and 12V pins when making your own power cables (did that once). Some of these things will kill the drive outright, sometimes with smoke coming out; others will just reduce its lifetime (and data reliabilty) massively. But in most cases, these extreme mistakes don't happen: few people solder their own power cables, and most buy good quality cases and power supplies.

Lesson #2: But the MTBF as seen by amateurs is much below what large systems professionals see, because they don't have good environmental controls (temperature goes up and down), good enclosures, good power supplies. Still, it is not catastrophic; drives should not die like flies unless you abuse them.

But even in those professional settings, you have to exclude certain effects to get to these MTBF numbers. First is infant mortality: You have to bake in new drives for a week or two, and drives that fail during that time will be cheerfully replaced by the manufacturer, and does not count towards MTBF. Second is manufacturing problems that escape. We once had a whole delivery of drives (several thousand, single manufacturer single model) that had an infant mortality approaching several % per week, and that mortality kept on going for several months. This was a "manufacturing escape", due to a mistake a whole batch of drives had skipped the manufacturers quality control (oops), and happened to be low quality (double oops). The manufacturer cheerfully took the drives back, and (probably not cheerfully) gave us several M$ to compensate us for our troubles, and to make our customer less unhappy. This is human error, was acknowledged and corrected by the manufacturer, and should not be counted towards the MTBF number. Now, how would an individual amateur user who only bought 1 or 2 drives handled it? They don't have the metrics to demonstrate to the vendor that the problem is pervasive, they don't have hardware and software teams that can do autopsies on defective drives themselves, they don't have teams of lawyers to negotiate settlements.

Another case where drives had high failure rates was a system that was shipped to a city in a tropical country with really bad air quality, and stored there in a non airconditioned warehouse for half a year, unpacked. When it was finally turned on, many disks (about a third!) had electrical shorts, which were due to corrosion from sulfur in the atmosphere. Actually, our field service technicians ended up finding liquid drops of a corrosive liquid on the PC boards of the drives: condensation from the atmosphere, containing sulfuric acid. Again, because we were a big company we were able to diagnose what had gone wrong, and work with all stakeholders to come to an equitable solution. And again, this should not be counted towards the MTBF of the drive itself. But how would an amateur who lives in this city have handled it? He doesn't have ready access to air chemistry, he doesn't know how long the store had the drive on a shelf, and he doesn't have teams of lawyers to negotiate.

Lesson #3: For an amateur, systemic effects can mask the inherent good reliability of quality drives, and they may get lots of failures. Tough luck.

It is now widely known and reported that Seagate Barracuda drives (in particular the 1TB model) have had serious reliability problems. If you average those into Seagate's overall MTBF, then the result looks pretty bad for Seagate. I don't know whether Seagate ever had programs where the refunded those, extended the warranty, or had other arrangements with large users (I never worked with large quantities of that model professionally). Real-world example: Between a colleague of mine and me (he also worked on the disk subsystems for a large storage systems vendor), we went through 7 of these 1TB Barracuda drives at home (he had 5, I had 2), all of which died within a few years (in some cases before 2 years), and we know that our enclosures/power supplies/environment were at least OK. That failure rate is completely incompatible with the quoted 1M hours, and more points towards something in the range of a few 10K hours. He got some replaced under warranty, and we both threw the rest into the trash. But since I had been burned by that, I followed the fate of other Seagate drives later, and found that this problem did not repeat for other models.

Lesson #4: Some drive models just suck, and will die quickly. So quickly that even an amateur with a small number of drives (1...5) will have serious problems in a small number of years (1...5). But you can't extrapolate from a few bad models to all models in a series, and much less to a vendor.

And finally, look at the famous BackBlaze data. It is the best data set on disk quality that is freely accessible; there is better data out there, but it is not accessible. You clearly see that on average, Seagate is less reliable than the others, and that's not just one model, but systematic. But Seagate does not suck: their annual failure rate may be up to 2% and 3% for some models, but it is nowhere near the 30% or 50% that my friend and me saw, and that would be catastrophic.

In summary: The reliability of drives is complicated, and at the amateur level just not predictable. Not enough statistics. You may get very unlucky.
What can you do about this?
  1. Think about the value of your data. If it is worth nothing, and you will not feel bad if it is all gone suddenly, and your time for re-installing the system after a disk failure is worth nothing, then stop reading. All others, keep going.
  2. Use RAID. At the very minimum mirroring of two drives. If this is your only defense, it is not good enough, and a two-fault tolerant system is better or even necessary.
  3. If you are mirroring or RAIDing, consider using different drives (different models or even vendors) in a pair. Like that a systemic problem with a particular model is less likely to wipe you out. But that can be a bit tricky (different capacity, different performance, and in a RAID system the overall performance tends to be dominated by the slowest disk).
  4. Take backups. Like that a disk failure becomes an inconvenience (down for an hour or a day), and perhaps a small data loss (the data from the 23 hours may be gone), but not a catastrophe. Remember that RAID is not a panacea; it does not protect against correlated failure, nor is it 100% reliable, nor does it protect against human error.
  5. You have installed RAID already, right?
  6. Make a plan ahead of time: what will you do if a drive dies? Know the commands to resilver your RAID. Know where your backups are stored. Don't store the only documentation for how to restore from backup on the drive that you are backing up. Do a test run, regularly. A backup that has never been restored is not actually a backup, it might be a blank tape. Not a joke, happened to my wife's company once: after their disks died, they discovered that their clueless sys admin had set them up with RAID-0, and had dutifully written a blank tape every night, labelled it, and put it into the fireproof safe. Not fun.
  7. Your RAID is functioning well, you are monitoring disk health, and have set up an automatic monitoring system, right?
  8. Think about what other disasters you want to protect against (because disk failure is not a disaster, it is an expected operational situation). Are you worried about a fire or flood destroying the place where both your original disks and the backup are physically located? Are you worried about intruders stealing your hardware? Are you worried about someone snooping on you? A good backup system can deal with this, but at some cost.
  9. Inject a test fault into your RAID system (for fun, just pull a disk physically out), and watch it resilvering automatically, and your cellphone beeping because you got an e-mail from the monitoring system. That's when you can stop having anxiety attacks.
True story: About 25 years ago, I interviewed for a job at the storage systems research department of one of the largest and most prestigious computer companies in the world (two letters, not three). My host and future manager gave me a little tour of the computer room for fun, and showed me the main server the group used (in those days, a group of 15-20 people used a single large computer), and the two "big" RAID arrays connected to it (in those days, "big" meant dozens of disks). He then proceeded to pull a disk out of the running production machine, and hand it to me. I was flabbergasted. What this really demonstrated was: the guy was (and continues to be) very smart, and knew the reliability of his systems, and the value of impressing a person they might potential hire. I gave him back the disk, he put it back in, the disk array resilvered for a few more seconds, and everything was fine.
 

IBM HDD was division purchased by Hitachi, and then by WD .... it's the Ultrastar line (5 year warranty enterprise HDDs) that are the HGST (né IBM) heirs...

Yepp - the Hitachi rsp. HGST drives are IBM drives (the deskstar just as well), until the whole branch went to WD, and will most likely dissolve into the WD portfolio.
I for my part had an Hitachi and an original WD side by side in my desktop for quite a while, and there is a huge difference, and I would trust a half-wrecked Hitachi/HGST a lot more than a brand-new WD.
 
So, I thought, lets see what happens next - and that was 30'000 hours ago.
In automobile terms you been riding on Empty too long.....
As long as your happy I am happy.
50K mine are retired.
I do appreciate the heat factor. It is a killer.
 
Somewhere in the basement, I have two really good disks. One is a CDC/Imprimis/Seagate Wren, the other a Falcon (and I can't remember whether those were made by Maxtor or Fujitsu). They are both 1GB SCSI, and were bought in the late 80s or early 90s. Last time I booted those computers, they were both working; that was about 5 or 10 years ago.
 
Back
Top