[Arm-netbook] libre 64-bit risc-v SoC

Luke Kenneth Casson Leighton lkcl at lkcl.net
Fri Apr 28 15:55:23 BST 2017


On Fri, Apr 28, 2017 at 3:45 PM, mike.valk at gmail.com
<mike.valk at gmail.com> wrote:

> 2017-04-28 16:17 GMT+02:00 Luke Kenneth Casson Leighton <lkcl at lkcl.net>:
>>
>>
>>  the rest of the article makes a really good point, which has me
>> deeply concerned now that there are fuckwits out there making
>> "driverless" cars, toying with people's lives in the process.  you
>> have *no idea* what unexpected decisions are being made, what has been
>> "optimised out".
>
>
> That's no different from regular "human" programming.

 it's *massively* different.  a human will follow their training,
deploy algorithms and have an *understanding* of the code and what it
does.

 monte-carlo-generated iterative algorthms you *literally* have no
idea what it does or how it does it.  the only guarantee that you have
is that *for the set of inputs CURRENTLY tested to date* you have
"known behaviour".

 but for the cases which you haven't catered for you *literally* have
no way of knowing how the code is going to react.

 now this sounds very very similar to the human case: yes you would
expect human-written code to also have to pass test suites.

 but the real difference is highighted with the following question:
when it comes to previously undiscovered bugs, how the heck are you
supposed to "fix" bugs that you have *LITERALLY* no idea how the code
even works?

 and that's what it really boils down to:

 (a) in unanticipated circumstances you have literally no idea what
the code will do.  it could do something incredibly dangerous.

 (b) in unanticipated circumstances the chances of *fixing* the bug in
the genetic-derived code are precisely: zero.  the only option is to
run the algorithm again but with a new set of criteria, generating an
entirely new algorithm which *again* is in the same (dangerous)
category.

 l.



More information about the arm-netbook mailing list