From: Eliezer S. Yudkowsky (firstname.lastname@example.org)
Date: Mon Mar 26 2001 - 10:22:50 MST
James Higgins wrote:
> But then everyone keeps talking about this Sysop and what "it" wants. No
> one really seems to care about what any of those 3 billion individual minds
> might want. The worst description I've heard to date is that all of these
> individuals just become sub-goals of the almighty Sysop.
Well, you bloody well didn't hear it from me. That's the worst
description I've ever heard as well.
At this point I seriously have to ask: Have you been reading any of the
posts, or are you just reacting badly to the perhaps ill-chosen term
"Sysop"? Did you read "Friendly AI" at all?
As I wrote in a reply to you:
> "The key word is INVOLUNTARY. This very easily lets [ver] solve the problem
> by punting it back to individual sentients. The individual defines
> "pain". The individual defines "stupidity". In general, if there's not
> an objective or very strongly convergent answer, [ve] just punts the question
> back to the citizen."
In what way is this "not caring about what any of the 3 billion individual
minds might want"?
-- -- -- -- --
Eliezer S. Yudkowsky http://intelligence.org/
Research Fellow, Singularity Institute for Artificial Intelligence
This archive was generated by hypermail 2.1.5 : Wed Jul 17 2013 - 04:00:36 MDT