[Tfug] Small-ish (capacity + size) disk alternatives

Tyler Kilian vaca at grazeland.com
Thu Jan 31 22:32:09 MST 2013


Honestly, I'm not trying to be confrontational, but this thread has thousands of words trying to explain why no manufacturer of hard drives on this planet makes a product sufficient for a home automation project.  I find that hard to believe.

Even critical IT systems don't have the availability requirements being thrown out here to where folks scrutinize the minutia of drive usage patterns.  Drives fail, they always have, they always will.  

  I really think a solution exists.  The worry, to me, seems unfounded.  If it were me, I'd invest in a good quality drive.

I know that's more than two cents. :)



On Jan 31, 2013, at 5:52 PM, Bexley Hall <bexley401 at yahoo.com> wrote:

> Hi Yan,
> 
> On 1/31/2013 2:57 PM, Yan wrote:
>>> No!  My problem is *laptop* HDD's wear out -- not HDD's in general!
>>> In the 30 years I've owned computers, I've had exactly three HDD's
>>> wear out -- all laptop drives despite the fact that I rarely *use*
>>> a laptop (one drive in a laptop, two more in this 365/24/7 situation).
>> 
>> You got quite lucky. I haven't owned computers for anywhere near 30
>> years (although for most of my computer-owning time, my gear's been
>> bought used), but I've had at least a dozen hard drive failures. Here
>> in the lab, we have somewhere around 100 computers with probably
>> upwards of 250 hard drives. Those were all bought new, and most of
>> them are enterprise-grade. We have something around a failure a month,
>> and that's probably beating the statistics on these things.
> 
> AFR's for disks are in the 2-8% range.  So, if *all* of those computers
> are seeing "regular, normal use", you would expect 5 to 20 of them to
> fail per year.  Or, 0.5 to 1.5 per month (depending on brand, production
> lot, operating conditions, etc.).
> 
> I typically have ~1-3 machines running "normally" -- figure 2-10 drives.
> So, in a year, I'd expect to see .04 to .8 drive failures.
> 
> But, my computer usage hasn't been constant over those 30 years.
> E.g., my first disk drive was a wopping 1MB drive (yes, "one").
> I didn't keep it long (it was bigger than many of the early PC's.
> 
> My first practical drive was, I think, 10MB -- almost 10 whole
> (8 inch!) floppies on a single drive!!  :>
> 
> I didn't keep it long, either!
> 
> My first IDE drives were 60MB units (from early 386 machines).  I kept
> them running for many years before the machines were just taking up
> more room than they were worth!  (I *threw* the drives onto the
> concrete floor to crash their heads before discarding them -- since
> they were still operable).
> 
> I still have a pair of 600MB IDE drives from 20 years back that are
> now doing duty in a "Compaq 386 Portable" (that I keep for some
> legacy hardware installed in that machine).
> 
> Currently, I have about 20TB here -- with the bulk of that in cheap
> "consumer" external drives (that I have been using to snapshot my
> development database/software on a daily basis... disposable once
> this project is done).  About 6-7TB of that is spread out over
> dozens of spindles (e.g., each of my Sun machines has a TB on
> 14 spindles; x86 servers have a similar amount on half as many
> spindles; etc.).  I.e., I have preferred smaller drives and SCSI
> drives for most of my media.
> 
>>> No!  But, neither does every "block worth" (which might be as much as
>>> 500KB!) contain 100% "new data".  That's the point here -- that the
>>> work you are asking the disk to do can be much less than the work
>>> it actually ends up *doing*.
>>> 
>>> E.g., if I update the "hours worked this week" for each employee
>>> in a dataset and each employee's "record" resides in a different
>>> FLASH block, then an entire block is erased for each employee in
>>> the organization -- even if that's only 4 *actual* bytes per employee!
>> 
>> Admittedly, this is a problem for SSDs (although I'd argue it doesn't
>> preclude it from being used for the application you describe). There's
>> some research being done in this area. Samsung and some South Korean
>> university published a paper proposing a new filesystem optimized for
>> (among other things) reducing erase cycles. The paper is at
>> http://static.usenix.org/events/fast12/tech/full_papers/Min.pdf and
>> the presentation at
>> https://www.usenix.org/conference/fast12/sfs-random-write-considered-harmful-solid-state-drives
>> . It doesn't look like they released the implementation itself,
>> though, so this is more of a future argument.
> 
> The problem with this is that it forces the application to be cognizant
> of the underlying hardware implementation.  At the same time, there is
> increasing bloat and abstraction in OS's and application development
> tools that obfuscates the view "down" to the hardware.  So, it is harder
> to tune the application to fit the hardware.  If the hardware is
> homogenous and reliable, this isn't an issue.  But, if the hardware
> is neither of these, then system reliability becomes suspect.  *Before*
> the device is even deployed!
> 
>>> But it isn't.  Error frequencies go up which means more (hidden)
>>> update cycles are incurred, etc.  Notice how "enterprise" SSDs
>>> tend to stick to SLC technology -- trading capacity/speed for
>>> endurance/reliability.
>> 
>> I'm not sure if this is the case anymore. Most SSDs I've seen billed
>> as enterprise are heading to MLC nowadays. This might be a temporary
>> trend, though, for all I know.
> 
> I think you will either see a significant effort made to improve the
> reliability of MLC technologies (i.e., redundancy -- which sort of
> eliminates the advantage of MLC to begin with!) or a shift to some
> other technology/tricks that sidesteps this issue.
> 
> Look at where the drives are currently being used, their pricepoints,
> etc.  Technology's goal is always to move towards the ubiquity of
> the "consumer" market (SSDs don't seem ready for prime time, there)
> 
>>>> So what. This is what they do. If you have stuff that is never being
>>>> written (only read) the controller will move it to a frequently written
>>>> cell. It will do this before it thinks that there is only one write left
>>>> on that target cell. Even if the cell fails, so what. Cells fail. Those
>>>> cells are marked as bad, and the drive uses other cells. ALL modern SSDs
>>>> have spare area. They can (and do) handle cell failures.
>>> 
>>> This is reflected to the interface (i.e., user) as indeterminism.
>>> The application never *knows* that the data that the drive has
>>> previously claimed to have "written" has actually been written.
>>> It's an exaggeration of the write caching problem.  But, it is
>>> brought about by the inherent "endurance" limitations of the
>>> media.  As if you had a HDD that was inherently failure prone.
>> 
>> It isn't much different from HD caching, or from the OS-level caching,
>> or anything like that. If you want to complain about the disk
>> equivalent of bufferbloat, that might be valid, but pretending that
>> this only occurs with SSDs is very inaccurate.
> 
> The point is that the media has this as an *inherent* "flaw".
> HDDs remap sectors when they are found to be bad or suspect.
> They don't routinely move "safe" data around on the media
> because other parts of the media are flakey.
> 
>> In fact, it's probably
>> better on an SSD, because the amount of time that it takes to remap a
>> block is (probably) faster than the amount of time data sits in an HD
>> write cache waiting for the HD to spin to the right place.
>> 
>>> Imagine a HDD that was designed to *randomly* pick a block of
>>> otherwise stable data, copy it to a new location, verify that
>>> the copy succeeded and then erase (not just overwrite) the original
>>> all so that some NEW piece of data could be written in its place
>>> (and, at the same time, updating the behind the scenes bookkeeping
>>> that keeps track of all that shuffling around -- using media
>>> with the same "characteristics" that it is trying to workaround)
>> 
>> That'd be insanely slow, and that's because HDs are HDs, and are slow.
>> While an SSD would be faster if it didn't do this, it's still quite
>> fast while doing it.
> 
> But the reason it is doing this is because the technology/media isn't
> up to the expectations that this *other* media has already championed!
> It's like putting a 100 gallon fuel tank in your gas guzzler just
> so folks won't be discouraged from purchasing it because it can
> only travel 200 miles on a more traditional sized gas tank (when
> other vehicles get 300+)
> 
>>>> I fail to see the problem. SSD controller have a complicated job to do,
>>>> and they do it.
>>> 
>>> They *try* to do it.  I saw a recent survey claiming 17% of respondents
>>> had an SSD fail in the first *6* months!  (of course, a survey in which
>>> respondents self-select will tend to skew the results -- people are
>>> more likely to bitch about their experiences than praise them!)
>> 
>> On top of the statistical issues with that poll (which had "600+"
>> respondents), I would guess that those types of failures ("my SSD
>> completely stopped working") are more likely caused by the disk
>> controller giving out than the underlying medium. That's a guess, and
>> it doesn't make things better for the end-user, but it might take some
>> of the heat off of the wear-leveling debate.
> 
> If by disk controller you mean the (integrated) *flash* controller, it
> just draws attention to the higher level of technology involved in
> FLASH media vs. magnetic media.  How long do you wait for folks to
> get the software, wear leveling technologies, etc. right?
> 
>> It does shine light on another issue, which is the fact that SSDs are
>> just barely leaving the early adoption phase. I hopped on about a year
>> ago, and haven't had any issues. People around the lab are getting
>> SSDs more and more, and I haven't heard of any failing either, so at
>> least my anecdotal evidence implies to me that the manufacturers are
>> getting the hang of things better.
> 
> I've a friend who has already replaced his *replacement* SSD.
> Consider the lifespan of most laptops and you've got to wonder
> why he's had two fail, already.
> 
>>> So, you are suggesting I simply say, "Buy this particular SSD otherwise
>>> the system won't work"?  Would you build a MythTV box if you were told
>>> you had to use this disk (endurance), this motherboard (performance),
>>> this fan (sound level), etc.?  Or, would you cut some corner and then
>>> complain later to anyone who will listen?
>> 
>> "Buy this particular SSD otherwise the system won't work" should
>> really be "Buy this particular SSD otherwise the system will fail
>> faster".
>> 
>> If you're going to build a system, the quality of the components you
>> choose is going to affect its performance and reliability. You can't
>> get away from that. If you go with HDs, you'll have to say "Buy a
>> non-laptop drive of this caliber or it'll fail faster." It's the same
>> thing. If you go with a high-quality SSD, it doesn't mean that the box
>> won't work with a low-quality one. It'll just work slower or less
>> reliably. It's the same for every other component.
> 
> A year vs. 5-10 years is a huge difference!  As I said in a previous
> message, I find it annoying that I have to replace the batteries in
> the smoke detectors annually (and that is a *safety* issue, not just
> a "convenience" one!).
> 
> People, for the most part, aren't proactive.  They wait for things
> to break instead of monitoring/maintaining in an ongoing basis.
> 
> Imagine if you had gone to the trouble to put together a MythTV
> box.  Then, less than a year later, discovered the disk drive
> had died due to the usage patterns that application placed on it.
> You *might* replace the drive (assuming you had kept good notes
> on how to initialize the new drive).  How excited would you be
> to replace it *again* a year later??
> 
> E.g., chances are, you'll throw together a MythTV box out of
> "spare parts" to "explore" the features it presents.  Its
> largely "free".  And, you'd replace that disk with another
> "spare" you had on hand at the time.  But, faced with the
> "replace every year" realization, would you run out and BUY a
> more expensive drive?  Or, would you buy a COTS DVR that you
> could replace en toto when *it* fails?
> 
>>> Passive cooling.  The inside of the enclosure has never been above 35C.
>>> No one wants to listen to fans 24/7/365!
>> 
>> Here's another plus for SSDs. They'll pump a lot less heat out, so
>> your machine might work in less friendly environments (ie, locked in
>> an entertainment center or something).
> 
> This was the appeal of the laptop drive -- far less power than
> an equivalent amount of DRAM *or* a full-sized hard disk.  E.g.,
> here, it sits in the bottom of a closet (no ventilation other than
> the occasional opening of the closet door to access the other
> contents of the closet.
> 
>>> When was the last time you replaced your thermostat?  Irrigation
>>> controller?  Garage door opener?  Washer/dryer?  Doorbell?  TV?
>>> Security camera(s)?  DVR?  "HiFi"?  Hot water heater?  Weatherstation?
>>> 
>>> Then, ask yourself *why* you replaced it:  because you were tired
>>> of "last year's model"?  Because it wasn't performing as well as
>>> it should?  Because it *broke*?
>>> 
>>> Chances are, most of these things did their job until they broke (or
>>> were outpaced by other technological issues) and *then* were replaced.
>> 
>> In the last year, I've replaced an irrigation controller (and
>> irrigation pumps), had to repair a dryer, had to repair an AC unit,
>> and had to replace a hard drive (not a laptop one, either) in a
>> security camera system. That's not including regular filter changes
>> and so forth, either.
> 
> You've been *un*lucky!  :>  We replaced the washer and dryer after
> 20 years because "she wanted something new" (though they were in
> pristine condition).  I had to replace a start cap on the cooling
> fan for the (outdoor) AC compressor many years ago but no other HVAC
> issues (other than filters in the circulating fan).  Hot water heater
> has 20+ years on it.  Same for refrigerator and stove (though the
> latter to be replaced soon as part of remodel)
> 
> Never had a problem with a doorbell in 50 years.  Last TV lasted 25
> years.  Threw away my HiFi VCR a few years ago after copying everything
> onto DVD media.  Still have my 4 Nakamichi's (though they see routine
> maintenance).  A pair of two "integrated" bookshelf stereos have seen
> the same CD changer repair a total of three times in the past 25 years
> (one just being serviced last week -- par for the course).
> 
> Pinball machines see a fair bit of maintenance -- but that's the nature
> of the beast.  Cars need oil changes, filter changes, etc.  Printers
> rarely need "more toner".
> 
> The single biggest maintenance issue I've encountered, here, has been
> the swamp cooler (which we haven't used in years) and the *roof*
> itself!
> 
> I.e., we tend to invest in things that will "last" instead of
> saving a few pennies here and there.
> 
>> Components fail. There's not going to be a
>> magical HD that'll prevent that.
> 
> Components fail but at vastly different rates depending on
> their quality and how well their usage is matched to that!
> I've designed products that routinely see 20+ year lifetimes.
> The difference is how you *approach* the design.
> 
> I watch the stuff that comes into World Care every day.  And
> how *small* the difference between failing *now* vs. still
> being usable.  OTOH, those vendors aren't going to "waste"
> money to increase the usable lifetime of a product that
> their consumers have been "trained" to expect to replace
> that often!  (who would design a consumer PC with more than
> a 3 year expected lifetime?)
> 
>>>> You are avoiding one limitation (SSD finite erase/program cycle) but
>>>> with HDDs you still suffer mechanical wear and tear. As you noted in
>>>> your original email the HDDs you've been using "die pretty easily".
>>> 
>>> But those have all been laptop HDD's!  E.g., I suspect moving to
>>> a "real" disk drive will give me the same sorts of reliability
>>> that I've seen in my other machines (though at a higher power
>>> budget and cooling requirements).
>> 
>> I think the common wisdom is that a good HD will have a longer
>> lifespan. The argument a) that this lifespan isn't indefinite and b)
>> that an SSD's lifespan won't be unreasonably short. But it's your
>> system; you'll ultimately have to decide whether it's worth upping the
>> specs for it.
> 
> I'm not worried about *my* "instance" of the system -- it won't
> run on COTS hardware (since COTS hardware doesn't give me the
> physical size characteristics that I want, the power budget,
> reliability, etc.).  But, I have been trying to keep the *design*
> "mainstream" enough that others could replicate it, functionally.
> (i.e., if you can afford the space, power and noise, you could
> site a group of PC's in a "closet"/basement someplace and use
> COTS components).
> 
> It could be that this is the falacy in my approach -- trying to give
> the benefits of a "proprietary" design to a COTS implementation.
> I.e., maybe I just have to tell folks they have to build a particular
> set of boards to replicate the design!  (this would sure make the
> software easier as it wouldn't have to worry about hardware
> abstractions!)
> 
>>>> If you don't fully understand your data access/update patterns it
>>>> doesn't seem like you can say whether or not they will overly burden an
>>> 
>>> I can look at data write *rates* (sector counters) and make conclusions
>>> based solely on that!  The SSD won't give me any better guarantees than
>>> total number of rewrites.  I.e., it doesn't care if I am writing
>>> "AAAAAAAAX" or "AAAXAAAAAA" in place of "BBBBBBBBB" -- as long as either
>>> write is a "sector".
>>> 
>>> Knowing the access patterns (at the application level) IN DETAIL lets
>>> me restructure tables so that the data that are often updated "as a
>>> group" tend to be grouped in the same memory allocation units.
>>> 
>>> E.g., if you;re running a payroll application, then wages and taxes
>>> are the hot items that see lots of use.  OTOH, if you are running
>>> an application that tracks attendance (timeclock), then wage
>>> information probably sees *less* activity than "hours worked"
>>> (which would have to be updated daily).  In either case, employee
>>> *name* is probably RARELY updated!
>> 
>> If the data is in a database, ensuring this at the hardware level
>> might be harder than you'd think, given filesystem abstraction,
>> separate storage for indices, etc. If the data is in a DB-backed LDAP
>> directory or something, you can probably forget about it. Of course,
>> if you're using a specialized filesystem and so forth, it's possible,
>> but that's a lot of effort for questionable gain.
> 
> For a *particular* DBMS, you can create your tables to exploit
> the sorts of accesses and updates that you actually encounter.
> With judicious use of tablespaces, you can move individual
> tables into different spaces -- deliberately backed by "drives"
> having the appropriate capabilities.
> 
> E.g., the imaginary employee database I've mentioned could separate
> the static "employee" information from the "wage"/hours information
> (two or more tables linked by a shared index).  The employee table
> can be backed by a "write infrequently" medium while the wage table
> is backed by a "write frequently" medium.
> 
>>> It won't be *me* that's rolling the dice!  :>   Rather, it will be
>>> someone who tries to build and configure a similar system and
>>> wonders why *his* choice of storage media proved "less than ideal".
>>> Or, why the *identical* system ("as seen on TV") performs so
>>> differently after he's made some "trivial" changes to the code.
>>> 
>>> "Gee, I just changed the payroll program to update the wages on
>>> a daily basis -- each time the timeclock recorded additional
>>> hours for the employee.  Now, I'm seeing problems with the wage
>>> data's reliability..."
>> 
>> Given the abstraction involved in these things, I think the actual
>> effect will be "Gee, I just changed the payroll program to update the
>> wages on a daily basis -- each time the timeclock recorded additional
>> hours for the employee. Now the disk failed a day faster.". They'll
>> eventually see data issues ("bad sectors" and so forth) from the wear,
>> but the wear-leveling will distribute it out of just the
>> frequently-updated wage data.
> 
> That depends on what's backing the individual data.  E.g., if the
> wage data had been on the "write infrequently" medium, suddenly,
> that medium sees 5-7 times the usage rate that had been planned
> for it.
> 
> Tracking information shows I should receive my gifted "laptop
> disks" tomorrow.  So, I can reset the "life clock" and see how
> long *they* last.  Meanwhile, I'll look into the availability
> of physical RAM disks to replace the VM requirements of the
> disk drive.  (this might be a viable option *if* a user can just
> install their own "surplus" DIMMs to get capacity "for free")
> 
> _______________________________________________
> Tucson Free Unix Group - tfug at tfug.org
> Subscription Options:
> http://www.tfug.org/mailman/listinfo/tfug_tfug.org




More information about the tfug mailing list