Re: finding regressions with syzkaller

From: Dmitry Vyukov
Date: Thu Aug 12 2021 - 05:15:46 EST


On Wed, 11 Aug 2021 at 13:25, Thorsten Leemhuis <linux@xxxxxxxxxxxxx> wrote:
>
> [CCing Lukas]
>
> Hi Dmitry!
>
> On 10.08.21 19:08, Dmitry Vyukov wrote:
> > [...]
> > The idea is to generate random test programs (as syzkaller does) and
> > then execute them on 2 different kernels and compare results (so
> > called "differential fuzzing"). This has the potential of finding not
> > just various "crashes" but also logical bugs and regressions.
>
> Hmmm, interesting concept!
>
> > The major issue is various false positive differences caused by
> > timings, non-determinism, accumulated state, intentional and
> > semi-intentional changes (e.g. subtle API extensions), etc. We learnt
> > how to deal with some of these to some degree, but feasibility is
> > still an open question.
>
> Sounds complicated and like a lot of manual work.
>
> Do you have in mind that Linus and hence many other Kernel developers
> afaics only care about regressions someone actually observed in a
> practice? Like a software or script breaking due to a kernel-side change?
>
> To quote Linus from
> https://lore.kernel.org/lkml/CA+55aFx3RswnjmCErk8QhCo0KrCvxZnuES3WALBR1NkPbUZ8qw@xxxxxxxxxxxxxx/
>
> ```The Linux "no regressions" rule is not about some theoretical
> "the ABI changed". It's about actual observed regressions.
>
> So if we can improve the ABI without any user program or workflow
> breaking, that's fine.```
>
> His stance on that afaik has not changed since then.
>
> Thus after ruling our all false positives syzkaller might find, there
> will always be the follow-up question "well, does anything/anyone
> actually care?". That might be hard to answer and requires yet more
> manual work by some human. Maybe this working hours at least for now are
> better spend in other areas.

Hi Thorsten,

Good point. At this point the nature and volume of regressions such a
system can find is unknown, so it's hard to make any conclusions.
But some additional theoretical arguments in favor of such a system:
1. Any regressions also need to be found quickly (ideally before the
release). And as far as I understand currently lots of regressions are
found only after 1-3 years when the new kernel reaches some distro and
users update to the new version. Year-long latency has its own
problems. In particular there may now be users of the new (implicit)
API as well, and then it's simply not possible to resolve the breakage
at all.

2. As far as I understand most regressions happen due to patches that
are not even known to change anything (the change wasn't
known/described). So such a system could at least surface this
information. For example, was it intentional/known/realized that this
commit changes API?
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=d25e3a3de0d6fb2f660dbc7d643b2c632beb1743
Amusingly the commit description says:
"Retain compatibility with how attaching works, so that any attempt to
attach to an fd that doesn't exist, or isn't an io_uring fd, will fail
like it did before".
which turns out to be false, it does not fail like it did before, it
fails differently.

3. It may be possible to prioritize some API changes as more likely to
be problematic (e.g. a change from errno 0 to some particular values,
or changes file contents after the same sequence of syscalls). The
importance can also be different for different kernels. For example,
for LTS .1 -> .y I assume any changes may be worth being aware of.



> > Since this work is in very early stage, I only have very high-level questions:
> > - what do you think about feasibility/usefulness of this idea in general?
>
> TBH I'm a bit sceptical due to the above factors. Don't get me wrong,
> making syzkaller look out for regressions sounds great, but I wonder if
> there are more pressing issues that are worth getting at first.
>
> Another aspect: CI testing already finds quite a few regressions,

Quite a few in absolute numbers or relative to the total number of
regressions? :)

> but
> those that are harder to catch are afaics often in driver code. And you
> often can't test that without the hardware, which makes me assume that
> syzkaller wouldn't help here (or am I wrong?)

It depends.
syzbot runs on VMs at the moment, but anybody is free to run syzkaller
on any h/w. And it's quite popular at least for Android phones as far
as I understand.
And at some point we may be getting more testable drivers than we have
now (few).
But also could anybody predict how many bugs syzkaller would find
before it came into existence? So I would not give up on generic
kernel code right away :)


> > - any suggestions on how to make the tool find more differences/bugs
> > or how to make it more reliable?
> > - is there a list or pointers to some known past regressions that
> > would be useful to find with such tool? (I've looked at the things
> > reported on the regressions@ list, but it's mostly crashes/not
> > booting, but that's what syzkaller can find already well)
>
> I first wanted to tell you "look up the reports I compiled in 2017 in
> the LKML archives", but I guess the way better solution is: just grep
> for "regression" in the commit log.

Good idea.
It seems that something like this can give enough subsystem-targeted
info for initial analysis:

git log --no-merges --oneline --grep "fix.*regression" fs/ | grep -v
"performance regression"

> > - anybody else we should CC?
>
> I guess the people from the Elisa project might be interested in this,
> that's why I CCed Lukas.
>
> Ciao, Thorsten