From: Russell Wallace (email@example.com)
Date: Thu Aug 25 2005 - 00:05:22 MDT
On 8/25/05, Eliezer S. Yudkowsky <firstname.lastname@example.org> wrote:
> The AI remained in the box.
I will add, with permission, that:
1) Contrary to what I surmised earlier, Eliezer does not have a
reality distortion field.
2) This result should not be taken as evidence that AI boxing is a
good strategy. It isn't, it really isn't. If you're not confident
enough in what you build to unbox it, you shouldn't be confident
enough to build it.
This archive was generated by hypermail 2.1.5 : Wed Jul 17 2013 - 04:00:52 MDT