----- Original Message -----
From: "Dmitry Vyukov" dvyukov@google.com To: "Veronika Kabatova" vkabatov@redhat.com Cc: "Tim Bird" Tim.Bird@sony.com, automated-testing@yoctoproject.org, info@kernelci.org, "syzkaller" syzkaller@googlegroups.com, lkp@lists.01.org, "stable" stable@vger.kernel.org, "Laura Abbott" labbott@redhat.com, "Eliska Slobodova" eslobodo@redhat.com, cki-project@redhat.com, "David Rientjes" rientjes@google.com Sent: Monday, May 27, 2019 4:39:16 PM Subject: Re: CKI hackfest @Plumbers invite
On Mon, May 27, 2019 at 1:52 PM Veronika Kabatova vkabatov@redhat.com wrote:
----- Original Message -----
From: "Tim Bird" Tim.Bird@sony.com To: vkabatov@redhat.com, automated-testing@yoctoproject.org, info@kernelci.org, khilamn@baylibre.org, syzkaller@googlegroups.com, lkp@lists.01.org, stable@vger.kernel.org, labbott@redhat.com Cc: eslobodo@redhat.com, cki-project@redhat.com Sent: Friday, May 24, 2019 10:17:04 PM Subject: RE: CKI hackfest @Plumbers invite
-----Original Message----- From: Veronika Kabatova
Hi,
as some of you have heard, CKI Project is planning hackfest CI meetings after Plumbers conference this year (Sept. 12-13). We would like to invite everyone who has interest in CI for kernel to come and join us.
The early agenda with summary is at the end of the email. If you think there's something important missing let us know! Also let us know in case you'd want to lead any of the sessions, we'd be happy to delegate out some work :)
Please send us an email as soon as you decide to come and feel free to invite other people who should be present. We are not planning to cap the attendance right now but need to solve the logistics based on the interest. The event is free to attend, no additional registration except letting us know is needed.
Feel free to contact us if you have any questions,
I plan to come to the event.
Here is an early agenda we put together:
- Introductions
- Common place for upstream results, result publishing in general
able to substitute this session for a different one in case everything is solved by September.
- The discussion on the mailing list is going strong so we might be
- Test result interpretation and bug detection
test bugs? How to handle continuous failures due to known bugs in both tests
- How to autodetect infrastructure failures, regressions/new bugs and
and kernel? What's your solution? Can people always trust the results they receive?
- Getting results to developers/maintainers
click away in a dashboard? Do you want incremental emails with new results as they
- Aimed at kernel developers and maintainers, share your feedback and expectations.
- How much data should be sent in the initial communication vs. a
come in?
- What about adding checks to tested patches in Patchwork when patch
series are being tested?
- Providing enough data/script to reproduce the failure. What if
special HW is needed?
- Onboarding new kernel trees to test
most critical ones? Do you want them to be tested? Which tests do you feel are most beneficial for specific trees or in general?
- Aimed at kernel developers and maintainers.
- Which trees are most prone to bring in new problems? Which are the
- Security when testing untrusted patches
in
- How do we merge, compile, and test patches that have untrusted code
them and have not yet been reviewed? How do we avoid abuse of systems, information theft, or other damage?
- Check out the original patch that sparked the discussion at https://patchwork.ozlabs.org/patch/862123/
- Avoiding effort duplication
x86_64 might look useless on the first look but is it? AMD/Intel CPUs, different network cards, different graphic drivers, compilers, kernel configuration... How do we distribute the workload to avoid doing the same thing all over again while still running in enough different environments to get the most coverage?
- Food for thought by GregKH
- X different CI systems running ${TEST} on latest stable kernel on
Hi Veronika,
Hi!
All are great questions that we need to resolve!
I am also very much concerned about duplication in 2 other dimensions with the current approach to kernel testing:
- If X different CI systems running ${TEST}, developers receive X
reports about the same breakage from X different directions, in different formats, of different quality, at slightly different times and somebody needs to act on all of them in some way. The more CI systems we have, the more run meaningful number of tests and do automatic reporting, the more and more duplicates developers get.
This is a very good point that we had to deal with internally -- people aren't happy getting multiple reports about a same issue.
And that's what part of my point of finding balance was about too. Maybe the test only breaks on specific HW so getting X reports from runs on different HW where X-1 pass and one fails helps to narrow the issue down. Same if it fails everywhere -- it has to be some central code that caused the failures. If all the reports are just different version of "tested on x86 VM and it failed" then I totally agree it's not much help.
The formats and quality of the reports is definitely something worth talking about too, and potentially unify as much as we can (Tim started some work on this already).
- Effort duplication between implementation of different CI systems.
Doing a proper and really good CI is very hard. This includes all questions that you mentioned here, and fine tuning of all of that, refining reporting, bisection, onboarding of different test suites, onboarding of different dynamic/static analysis tools and much more. Last but not least is duplication of processes related to these CIs. Speaking of my experience with syzbot, this is extremely hard and takes years. And we really can't expose a developer to 27 different systems and slightly different processes (this would mean they follow 0 of these processes). This is further complicated by the fact that kernel tests are fragmented, so it's not possible to, say, simply run all kernel tests. And kernel processes are fragmented, e.g. you mentioned patchwork, but not all subsystems use patchwork, so it's not possible to simply extend a CI to all subsystems. And some aspects of the current kernel development process notoriously complicate automation of things that really should be trivial. For example, if you have github/gitlab/gerrit, you can hook into arrival of each new change and pull exact code state. Done. For kernel some changes appear on patchwork, some don't, some are duplicated on multiple patchworks, some duplicated in a weird way on the same patchwork, some non-patches appear on patchwork because it's confused, and last but not least you can't really apply any of them because none of them include base tree/commit info. Handling just this requires lots of effort, guessing on coffee grounds and heuristics that need to be refined over time. The total complexity of doing it just once, with all resources combined and dev process re-shaped to cooperate is close to off-scale.
Totally agreed, we've have some fun with some of the points you mentioned as well and I'm sure others too. What we do in CKI is to unify as much as we can across different trees we're testing. For example the trees living in GitHub/GitLab/Gerrit -- not all trees do this but all trees use the git backend so we use polling and commit checking (ugly, I admit) as it works everywhere and we don't need to maintain multiple different APIs for the same thing.
I don't want to derail the invite discussion by replying to everything here but what you brought up are great points that every CI system around kernel has to deal with.
Do you see these points as a problem too? Or am I exaggerating matters?
Oh absolutely! While it's what I call "the nasty implementation details" they are all valid problems that should be talked about more. Whether it's with developers (eg. the points about putting kernel tests into a single place and unifying the development process, if possible) or with other CI teams.
I'd be happy to talk with you about all of this at the hackfest or privately/off-thread. The automated-testing list (in cc) is a great place too as a diverse set of people involved with CI gathers there.
Let me know what works for you, Veronika
- Common hardware pools
especially for HW that's hard to access, eg. ppc64le or s390x systems. Companies could
- Is this something people are interested in? Would be helpful
also sing up to share their HW for testing to ensure kernel works with their products.
I have strong opinions on some of these, but maybe only useful experience in a few areas. Fuego has 2 separate notions, which we call "skiplists" and "pass criteria", which have to do with this bullet:
- How to autodetect infrastructure failures, regressions/new bugs and
test bugs? How to handle continuous failures due to known bugs in both tests and kernel? What's your solution? Can people always trust the results they receive?
I'd be happy to discuss this, if it's desired.
Otherwise, I've recently been working on standards for "test definition", which defines the data and meta-data associated with a test. I could talk about where I'm at with that, if people are interested.
Sounds great! I added both your points to the agenda as I do think they have a place here. The list of items is growing so I hope we can still fit everything into the two days we planned :)
See you there! Veronika
Let me know what you think. -- Tim