From: Eliezer S. Yudkowsky (firstname.lastname@example.org)
Date: Mon Aug 22 2005 - 17:21:38 MDT
Russell Wallace wrote:
> On 8/22/05, Michael Wilson <email@example.com> wrote:
>>If you can't construct any theory for how you might fail it,
>>/and/ you truly believe that anything you can't construct a
>>theory for can't happen (or at the very least is incredibly
>>unlikely), then why don't you just take the challenge and win
>>$25 while satisfiying your curiosity?
> I've said I believe myself to be more resistant than the other
> contestants to the techniques I suspect Eliezer to be using, albeit
> for different reasons, and I am curious... so I'm going to take up
> that challenge. I hereby offer to participate in an AI box experiment
> under terms similar to the ones used in the latest run.
Russell, you previously wrote:
> Whether unfriendly superintelligent AI in a box is safe depends on
> your assumptions; but I claim that there are _no_ plausible
> assumptions under which it would be _both safe and useful_.
Are we supposed to simulate a Friendly AI in a box? Why wouldn't you just let
it out immediately?
-- Eliezer S. Yudkowsky http://intelligence.org/ Research Fellow, Singularity Institute for Artificial Intelligence
This archive was generated by hypermail 2.1.5 : Wed Jul 17 2013 - 04:00:52 MDT