On Thu, Apr 16, 2020 at 09:08:06PM +0000, Saeed Mahameed wrote:
On Thu, 2020-04-16 at 15:58 -0400, Sasha Levin wrote:
Hrm, why? Pretend that the bot is a human sitting somewhere sending mails out, how does it change anything?
If i know a bot might do something wrong, i Fix it and make sure it will never do it again. For humans i just can't do that, can I ? :) so this is the difference and why we all have jobs ..
It's tricky because there's no one true value here. Humans are constantly wrong about whether a patch is a fix or not, so how can I train my bot to be 100% right?
The solution here is to beef up your testing infrastructure rather than
So please let me opt-in until I beef up my testing infra.
Already did :)
No you didn't :), I received more than 5 AUTOSEL emails only today and yesterday.
Appologies, this is just a result of how my process goes - patch selection happened a few days ago (which is when blacklists are applied), it's been running through my tests since, and mails get sent out only after tests.
Please don't opt mlx5 out just yet ;-), i need to do some more research and make up my mind..
Alrighty. Keep in mind you can always reply with just a "no" to AUTOSEL mails, you don't have to explain why you don't want it included to keep it easy.
taking less patches; we still want to have *all* the fixes, right?
if you can be sure 100% it is the right thing to do, then yes, please don't hesitate to take that patch, even without asking anyone !!
Again, Humans are allowed to make mistakes.. AI is not.
Again, why?
Because AI is not there yet.. and this is a very big philosophical question.
Let me simplify: there is a bug in the AI, where it can choose a wrong patch, let's fix it.
But we don't know if it's wrong or not, so how can we teach it to be 100% right?
I keep retraining the NN based on previous results which improves it's accuracy, but it'll never be 100%.
The NN claims we're at ~95% with regards to past results.