Hacker News new | past | comments | ask | show | jobs | submit login

As best as they can, and by a lot of people running release candidates on test systems. Unfortunately there are basically an infinite (not really, just incredibly impossibly large) number of configurations of software and hardware that can lead to a bug like this not triggering so it's not always possible to catch this ahead of time.



How often does a bug of this severity slip through the net?


The only one I can find is in 2012 there was a RAID bug that damaged metadata required for RAID access, under specific/rare conditions.

http://www.h-online.com/open/news/item/Bug-in-Linux-kernel-c...

But if anyone is running critical data, they shouldn't be using a bleeding-edge kernel in the first damn place. It certainly "sucks" for the users that found it, but again, that's the risk you play.

And, this bug seems to have been found... a mere week after articles even reported the release of Kernel 4.14:

https://bugs.gentoo.org/638206

http://www.omgubuntu.co.uk/2017/11/linux-kernel-4-14-lts-fea...

Which is a pretty amazing turn-around time for "crowd-sourced" discovery of a bug. And according to the bug tracker, it looks like they patched it the same day.

So as bad as this is (and I feel for those who lost data, I really do), the solution is simply to not run bleeding-edge kernels on release day. The very fact they have a minor version number 4.14.1 (.1) implies things get released that need fixed, and if you have critical data you should be more patient.


> bleeding-edge kernel kernel

> risk you play

uhm you realize that the kernel was already released upstream right?


Yeah, and which distros are shipping it?

Using the newest upstream kernel is something you opt into, on non-production hardware. This isn't something that would impact a regular user.


How many people would be using bcache in production fullstop, I'm guessing not many.


It's been the official position for quite a while that kernel.org releases - even "stable" ones - are to be considered bleeding-edge, and for it to be the responsibility of distros to do the necessary cooking on behalf of users to avoid unleashing something raw on them. Long gone are the days of regular users being trusted (and being able to trust) to just grab the latest linux-2.0.36 tarball from there and upgrade by compiling themselves.


This fix doesn't seem to have made it into 4.14.1 unfortunately, and the Gentoo bug report came almost a week after this issue was identified and a patch posted to the bcache mailing list.


If you have really critical data, I expect you to run the latest possible kernel on part of your cluster. That allows you to see potential regressions and also to test your backup/recovery solution in case something that bad happens.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: