Agent-foundations researcher. Working on Synthesizing Standalone World-Models, aiming at a timely technical solution to the AGI risk fit for worlds where alignment is punishingly hard and we only get one try.
Currently looking for additional funders ($1k+, details). Consider reaching out if you’re interested, or donating directly.
Or get me to pay you money ($5-$100) by spotting holes in my agenda or providing other useful information.
I agree that there are ways to define the “capabilities”/”intelligence” of a system where increasing them won’t necessarily increase its long-term coherence. Primarily: scaling its ability to solve problems across all domains except the domain of decomposing new unsolved problems into combinations of solved problems. I. e., not teaching it (certain kinds of?) “agency skills”. The resultant entity would have an abysmal time horizon (in a certain sense), but it can be made vastly capable, including vastly more capable than most people at most tasks. However, it would by definition be unable to solve new problems, not even those within its deductive closure.
Inasmuch as a system can produce solutions to new problems by deductive/inductive chains, however, it would need to be able to maintain coherence across time (or, rather, across inferential distances, for which time/context lengths are a proxy). And that’s precisely what the AI industry is eager to make LLMs do, what it often measures capabilities in.
(I think the above kind of checks out with the distinction you gesture at? Maybe not.)
So yes, there are some notions of “intelligence” and “scaling intelligence” that aren’t equivalent to some notions of “coherence” and “scaling coherence”. But I would claim it’s a moot point, because at this point, the AI industry explicitly wants the kind of intelligence that is equivalent to long-term coherence.