What are you doing for home NAS?

Michael Bilow mikebw at colossus.bilow.com
Mon Dec 30 14:16:52 EST 2013


On 2013-12-30 at 10:18 -0500, Ken D'Ambrosio wrote:

> On 2013-12-30 09:41, John Abreau wrote:
>> After trying FreeNAS, I'd no longer consider the consumer-level drives
>> such the MyBook Live as serious options.
>
> I think this stance is a little overly cautious; there is  data showing
> that consumer drives don't fail at rates significantly different than
> "server-grade" drives -- e.g.,
> http://blog.backblaze.com/2013/12/04/enterprise-drive-reliability/
> (though I also remember studies done on significantly larger datasets a
> couple years ago, but they aren't leaping at me from Google).  What I
> *have* found to be troublesome is that some RAID solutions don't handle
> drives that spin down very well.

Server-class drives are very different in at least two major ways.

First, a basic operational failure, such as a bad sector read, will 
cause a consumer-class drive to retry pretty much forever, while 
such an error will be reported relatively rapidly to the controller 
in a server-class drive. This allows the RAID manager (whether 
hardware or software) to handle the error appropriately, usually by 
computing what the sector should contain and writing it, thereby 
causing a reallocation of the failed sector from a reserve of 
spares. Software RAID on Linux, for example, provides for a periodic 
"scrubbing" operation that test reads every sector; Debian codes 
this as a monthly "checkarray" cron job.

Second, server-class drives are tested for 24x7 operation and rated 
accordingly with longer warranties that take this into account. 
Seagate consumer drives are specified with MTBF based on operation 
about 9 hours per day Monday through Friday, a typical work week. 
Western Digital specifies error performance an order of magnitude 
better in their top server-class drives, 1 in 10^16, compared to 
their consumer-class drivers, 1 in 10^15 -- which may not seem like 
much, but actually does have practical effect when dealing with 
multi-terabyte arrays. Not everything that goes wrong in a RAID 
system takes the spindle off-line, which is important to remember.

I obviously cannot dispute the Backblaze data, but their application 
is relatively quiescent and therefore unusual. Most server-class 
drives are operated under heavy demand in something like an e-mail 
server or a database server, and even the Backblaze article notes 
that this could significantly affect their results. Increased 
activity caused increased heat, the great enemy of the drive. To a 
large extent, though, you are paying for the longer warranty.

> For this reason, I tend to either go
> with "server-grade" drives, or really do my homework, and find drives
> that work with the solution

Lately there has been a trend among manufacturers to introduce 
price-differentiated grades of server-class drives, notably the new 
Western Digital "Red" line costing less than their "Se" enterprise 
value-priced line and their "Re" enterprise full-priced line. At 4TB 
a "Red" (WD40EFRX) is about $190, a "Se" (WD4000F9YZ) is about $280, 
and a "Re" (WD4000FYYZ) is about $320. Much of the increased 
reliability of the "Red" comes from simply spinning it at 5400RPM 
instead of 7200RPM, and much of the decreased cost comes from 
providing a 3-year warranty instead of a 5-year warranty. By 
comparison, a consumer-class "Green" (WD40EZRX) is about $170, 
spinning at 5400RPM and the warranty cut to 2 years.

> (e.g., 3Ware has -- or, at least, had -- an
> approved hardware list that I find useful).  But I think that, with a
> suitable amount of caution, there's money to be saved here without loss
> of functionality or increased risk of data loss.

I've been too badly burned on 3Ware RAID controllers to use them 
ever again. Regardless, in the vast majority of cases these days, I 
recommend using software RAID rather than hardware RAID on Linux 
servers unless there are special requirements, such as for hot-swap 
capability or for absolute maximum performance.

> P.S.  One thing I should add here, just from a hoo-boy-did-I-stub-my-toe
> perspective: as a rule, I usually have my arrays use just a leeeeetle
> bit less than the whole disk.  I had a large RAID-5 array once, and one
> of the drives failed.  I got it RMA'd *with the same model number* from
> the manufacturer... and it was one sector smaller.  THAT was annoying.

Although not generally known and hardly recommended practice, Linux 
software RAID can handle that case.

-- Mike



More information about the gnhlug-discuss mailing list