From: Matt Mahoney (firstname.lastname@example.org)
Date: Sat Nov 24 2007 - 18:47:06 MST
--- David Picón Álvarez <email@example.com> wrote:
> From: "John K Clark" <firstname.lastname@example.org>
> > No, although the issue of chaos, the idea that very small changes in
> > initial conditions could lead to huge changes in outcome it yet another
> > reason you will never be certain what an AI will do. But that wasn't
> > what I was talking about.
> > Even if you live in a universe that was completely deterministic and
> > randomness did not exist, even if you discount chaos, even if you ignore
> > the influence of the outside environment, I could still write in 5
> > minutes a very short program that will behave in ways NOBODY or NOTHING
> > in the known universe understands; it would simply be a program that
> > looks for the first even number greater than 4 that is not the sum of
> > two primes greater than 2, and when it finds that number it then stops.
> > When will this program stop, will it ever stop? There is no way to tell,
> > all you can do is watch it and see what it does, and randomness or chaos
> > or the environment has nothing to do with it.
> I can say when it will stop. It will stop when it runs out of memory. And
> that moment can be predicted.
Only if you can prove Goldbach's conjecture up to 2^memory_size. So far it
has only been proven to hold for up to about 3.33 x 10^17. Nor is there any
proof that there is no proof if it is true.
-- Matt Mahoney, email@example.com
This archive was generated by hypermail 2.1.5 : Wed Jul 17 2013 - 04:01:01 MDT