Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines; there would then unquestionably be an ‘intelligence explosion,’ and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is the last invention that man need ever make.
Yes, but I believe it is a bit weird for a Wikipedia article to state that someone is the originator of the Seed AI theory when he just coined the term. I wasn’t disputing anything, just trying to figure out if it is actually the case that Yudkowsky came up with the concept in the first place.
...and here’s a quote from I.J. Good, from 1965:
He didn’t coin the term “Seed AI” either.
Yes, but I believe it is a bit weird for a Wikipedia article to state that someone is the originator of the Seed AI theory when he just coined the term. I wasn’t disputing anything, just trying to figure out if it is actually the case that Yudkowsky came up with the concept in the first place.
Not the concept—the term.
“Seed AI theory” probably refers to something or another in here—which did indeed originate with Yu’El.
Presumably http://en.wikipedia.org/wiki/Seed_AI should be considered to be largely SIAI marketing material.