From: Eliezer Yudkowsky (sentience@pobox.com)
Date: Wed Jun 02 2004 - 10:44:46 MDT
Ben Goertzel wrote:
>
> Even once AI science is far more advanced, there may still be many
> different ways of creating AI's, and many different ways of creating
> Friendly AI's. AI science, at its present primitive stage, certainly
> doesn't rule this out.
>
> If we're lucky, there will be a unified theory of AI that tells us which
> ways of creating AI will be successful at creating intelligence, which
> ways will be successful at creating FAI, etc.
>
> I see no evidence that Eliezer or anyone else possesses a rigorous,
> demonstrated unified theory of this nature, at the present time.
Just don't forget that this theory is *needed*, and that philosophy is not
an acceptable substitute therefor. I now know enough to realize this,
which was obvious enough in retrospect, though I don't yet have as much
theory as I need. *Of course* you can't build a Friendly AI without
knowing what you're doing! It would be like Greek philosophers building an
airplane. Why was I ever foolish enough to think otherwise? Because I did
not know enough of the rules to know that knowing the rules was necessary.
Well, now I know.
-- Eliezer S. Yudkowsky http://intelligence.org/ Research Fellow, Singularity Institute for Artificial Intelligence
This archive was generated by hypermail 2.1.5 : Tue Feb 21 2006 - 04:22:38 MST