[arch-general] Reliability test for hard drives and SSD
Hi there! Good news for all interested in hardware compatibility and reliability. I've started a new project to estimate reliability of hard drives and SSD in real-life conditions based on the SMART data reports collected by Linux users in the Linux-Hardware.org database since 2014. The initial data (SMART reports), analysis methods and results are publicly shared in a new github repository: https://github.com/linuxhw/SMART. Everyone can contribute to the report by uploading probes of their computers by the hw-probe tool! The primary aim of the project is to find drives with longest "power on hours" and minimal number of errors. The following formula is used to measure reliability: Power_On_Hours / (1 + Number_Of_Errors), i.e. time to the first error/between errors. Please be careful when reading the results table. Pay attention not only to the rating, but also to the number of checked model samples. If rating is low, then look at the number of power-on days and number of errors occurred. New drive models will appear at the end of the rating table and will move to the top in the case of long error-free operation. Thanks to ROSA, Arch, Fedora, Ubuntu, Debian, Mint, openSUSE, Gentoo users and others who had made this work possible by contribution to the database!
On Sat, 03 Mar 2018 10:20:51 +0300, Andrey Ponomarenko wrote:
The primary aim of the project is to find drives with longest "power on hours" and minimal number of errors.
Pardon, but this is idiotic, since the HDDs and SSDs with the longest power on hours and minimal number of errors are already discontinued, if you want to buy a new drive. Your statistic at best is useful to find out what HDDs and SSDs fail much to early, but than consider to care about the usage. Some people turn their computers on and off several times a day, this isn't good for a HDD, while others have got uptimes of more than a year. Other people are maintaining SSDs with kid gloves, OTOH other are using SSDs in the same way, as they use/d HDDs. btw. I'm using my SSDs without kid gloves and when I wanted to order additional SSDs of the same kind, I couldn't get them anymore, let alone the best HDDs I every used.
On Sat, 3 Mar 2018 08:55:58 +0100, Ralf Mardorf wrote:
On Sat, 03 Mar 2018 10:20:51 +0300, Andrey Ponomarenko wrote:
The primary aim of the project is to find drives with longest "power on hours" and minimal number of errors.
Pardon, but this is idiotic, since the HDDs and SSDs with the longest power on hours and minimal number of errors are already discontinued, if you want to buy a new drive. Your statistic at best is useful to find out what HDDs and SSDs fail much to early, but than consider to care about the usage. Some people turn their computers on and off several times a day, this isn't good for a HDD, while others have got uptimes of more than a year. Other people are maintaining SSDs with kid gloves, OTOH other are using SSDs in the same way, as they use/d HDDs. btw. I'm using my SSDs without kid gloves and when I wanted to order additional SSDs of the same kind, I couldn't get them anymore, let alone the best HDDs I every used.
Not to mention what crap as e.g. GVFS does to green external HDDs :D. [rocketmouse@archlinux ~]$ pacman -Qi gvfs | grep Description Description : Dummy package A dummy packe is required to solve insane, useless dependencies. My external WD green drive stays asleep, it does not spin down and up again and again. It did, before I replaced GVFS by a dummy package. GVFS is just one HDD killer of several HDD killers. -- $ pacman -Q linux{,-rt-securityink,-rt,-rt-pussytoes,-rt-cornflower} linux 4.15.6-1 linux-rt-securityink 4.14.20_rt17-1 linux-rt 4.14.12_rt10-1 linux-rt-pussytoes 4.14.8_rt9-2 linux-rt-cornflower 4.11.12_rt16-1
Op za 3 mrt. 2018 08:56 schreef Ralf Mardorf <silver.bullet@zoho.com>:
On Sat, 03 Mar 2018 10:20:51 +0300, Andrey Ponomarenko wrote:
The primary aim of the project is to find drives with longest "power on hours" and minimal number of errors.
Pardon, but this is idiotic, since the HDDs and SSDs with the longest power on hours and minimal number of errors are already discontinued, if you want to buy a new drive. Your statistic at best is useful to find out what HDDs and SSDs fail much to early, but than consider to care about the usage.
Actually, it could be very useful to have a real insight in these figures. Similar reports by backblaze (a storage company) have been very helpful in deciding which HDD's to buy (or skip!) at $work. Also, this potentially gives *real* insight in SSD reliability; most other sources are "promises" by the manufacturer, some testers and perhaps random people who happened to run into a problem. Even if most models will indeed have been discontinued by now. Mvg, Guus Snijders
On Sat, 03 Mar 2018 11:16:30 +0000, Guus Snijders via arch-general wrote:
Op za 3 mrt. 2018 08:56 schreef Ralf Mardorf <silver.bullet@zoho.com>:
On Sat, 03 Mar 2018 10:20:51 +0300, Andrey Ponomarenko wrote:
The primary aim of the project is to find drives with longest "power on hours" and minimal number of errors.
Pardon, but this is idiotic, since the HDDs and SSDs with the longest power on hours and minimal number of errors are already discontinued, if you want to buy a new drive. Your statistic at best is useful to find out what HDDs and SSDs fail much to early, but than consider to care about the usage.
Actually, it could be very useful to have a real insight in these figures.
Similar reports by backblaze (a storage company) have been very helpful in deciding which HDD's to buy (or skip!) at $work. Also, this potentially gives *real* insight in SSD reliability; most other sources are "promises" by the manufacturer, some testers and perhaps random people who happened to run into a problem.
Even if most models will indeed have been discontinued by now.
Actually the strategy of companies such as Google obviously is to use consumer drives instead of enterprise drives, different models, from different vendors and much likely different batches, too, since statistics aren't helpful at all. Statistics even say anything useful about drive vendors, as far as I know, no vendor is better than another, just specific _discontinued_ product lines of different vendors were(/still are :) better than other.
Also very helpful could be https://en.wikipedia.org/wiki/Ouija the result is useless, too, but a good step into the right direction to notice self-delusion.
Op za 3 mrt. 2018 17:30 schreef Ralf Mardorf <silver.bullet@zoho.com>:
Also very helpful could be https://en.wikipedia.org/wiki/Ouija the result is useless, too, but a good step into the right direction to notice self-delusion.
When it comes to predicting stats; agreed. This is the other way around; historic data on how $device *has* behaved. Something which can help with replacing early etc. In our case: a certain model HDD was reported with a high error rate. Sure enough, once one of them failed, 2 others were quick to follow. Thanks to the warning from this data, we were prepared and could simply swap the disks without any downtime. Oh yeah, the replacements were from the same brand, just another model. Superstition? Perhaps, but so far, so good ;). Mvg, Guus Snijders
On Sat, 03 Mar 2018 17:06:46 +0000, Guus Snijders wrote:
Something which can help with replacing early etc. In our case: a certain model HDD was reported with a high error rate.
I didn't think of that. I agree that a statistic could be helpful to replace drives, before they fail.
03.03.2018, 10:21, "Andrey Ponomarenko via arch-general" <arch-general@archlinux.org>:
Hi there!
Good news for all interested in hardware compatibility and reliability.
I've started a new project to estimate reliability of hard drives and SSD in real-life conditions based on the SMART data reports collected by Linux users in the Linux-Hardware.org database since 2014. The initial data (SMART reports), analysis methods and results are publicly shared in a new github repository: https://github.com/linuxhw/SMART. Everyone can contribute to the report by uploading probes of their computers by the hw-probe tool!
The primary aim of the project is to find drives with longest "power on hours" and minimal number of errors. The following formula is used to measure reliability: Power_On_Hours / (1 + Number_Of_Errors), i.e. time to the first error/between errors.
Please be careful when reading the results table. Pay attention not only to the rating, but also to the number of checked model samples. If rating is low, then look at the number of power-on days and number of errors occurred. New drive models will appear at the end of the rating table and will move to the top in the case of long error-free operation.
Hi, I've just built an Arch Linux package for hw-probe. See https://github.com/linuxhw/hw-probe/blob/master/INSTALL.md#install-on-arch-l.... The command to replenish the database: sudo hw-probe -all -upload One can also use a lightweight all-in-one AppImage w/o the need to install anything to the system: https://github.com/linuxhw/hw-probe#appimage Thank you.
participants (3)
-
Andrey Ponomarenko
-
Guus Snijders
-
Ralf Mardorf