RSS

Cole Wyeth

Karma: 4,038

I am a PhD student in computer science at the University of Waterloo, supervised by Professor Ming Li and advised by Professor Marcus Hutter.

My current research is related to applications of algorithmic probability to sequential decision theory (universal artificial intelligence). Recently I have been trying to start a dialogue between the computational cognitive science and UAI communities. Sometimes I build robots, professionally or otherwise. Another hobby (and a personal favorite of my posts here) is the Sherlockian abduction master list, which is a crowdsourced project seeking to make “Sherlock Holmes” style inference feasible by compiling observational cues. Give it a read and see if you can contribute!

See my personal website colewyeth.com for an overview of my interests and work.

I do ~two types of writing, academic publications and (lesswrong) posts. With the former I try to be careful enough that I can stand by ~all (strong/​central) claims in 10 years, usually by presenting a combination of theorems with rigorous proofs and only more conservative intuitive speculation. With the later, I try to learn enough by writing that I have changed my mind by the time I’m finished—and though I usually include an “epistemic status” to suggest my (final) degree of confidence before posting, the ensuing discussion often changes my mind again. As of mid-2025, I think that the chances of AGI in the next few years are high enough (though still <50%) that it’s best to focus on disseminating safety relevant research as rapidly as possible, so I’m focusing less on long-term goals like academic success and the associated incentives. That means most of my work will appear online in an unpolished form long before it is published.

Non­triv­ial pillars of IABIED

Cole Wyeth17 Oct 2025 15:21 UTC
23 points
3 comments3 min readLW link

Align­ment as up­load­ing with more steps

Cole Wyeth14 Sep 2025 4:08 UTC
63 points
33 comments14 min readLW link

Sleep­ing Ex­perts in the (re­flec­tive) Solomonoff Prior

31 Aug 2025 4:55 UTC
16 points
0 comments3 min readLW link

New Paper on Reflec­tive Or­a­cles & Grain of Truth Problem

Cole Wyeth26 Aug 2025 0:18 UTC
53 points
0 comments1 min readLW link

Launch­ing new AIXI re­search com­mu­nity web­site + read­ing group(s)

Cole Wyeth13 Aug 2025 17:09 UTC
46 points
2 comments1 min readLW link

Pit­falls of Build­ing UDT Agents

Cole Wyeth30 Jul 2025 3:27 UTC
26 points
5 comments7 min readLW link

Ex­plain­ing your life with self-re­flec­tive AIXI (an in­ter­lude)

Cole Wyeth23 Jul 2025 0:57 UTC
16 points
0 comments5 min readLW link

Un­bounded Embed­ded Agency: AEDT w.r.t. rOSI

Cole Wyeth20 Jul 2025 23:46 UTC
29 points
0 comments18 min readLW link

A sim­ple ex­pla­na­tion of in­com­plete mod­els

Cole Wyeth6 Jul 2025 19:09 UTC
19 points
1 comment5 min readLW link

Paradigms for computation

Cole Wyeth30 Jun 2025 0:37 UTC
65 points
10 comments12 min readLW link

LLM in-con­text learn­ing as (ap­prox­i­mat­ing) Solomonoff induction

Cole Wyeth5 Jun 2025 17:45 UTC
31 points
3 comments4 min readLW link

For­mal­iz­ing Embed­ded­ness Failures in Univer­sal Ar­tifi­cial Intelligence

Cole Wyeth26 May 2025 12:36 UTC
39 points
0 comments1 min readLW link
(arxiv.org)

Align­ment Pro­posal: Ad­ver­sar­i­ally Ro­bust Aug­men­ta­tion and Distillation

25 May 2025 12:58 UTC
56 points
47 comments13 min readLW link

Model­ing ver­sus Implementation

Cole Wyeth18 May 2025 13:38 UTC
27 points
10 comments3 min readLW link

Glass box learn­ers want to be black box

Cole Wyeth10 May 2025 11:05 UTC
49 points
10 comments4 min readLW link

Why does METR score o3 as effec­tive for such a long time du­ra­tion de­spite over­all poor scores?

Cole Wyeth2 May 2025 22:58 UTC
19 points
3 comments1 min readLW link

Judg­ing types of con­se­quen­tial­ism by in­fluence and normativity

Cole Wyeth29 Apr 2025 23:25 UTC
19 points
0 comments2 min readLW link

Is al­ign­ment re­ducible to be­com­ing more co­her­ent?

Cole Wyeth22 Apr 2025 23:47 UTC
19 points
0 comments3 min readLW link

Re­ac­tions to METR task length pa­per are insane

Cole Wyeth10 Apr 2025 17:13 UTC
59 points
43 comments4 min readLW link

Chang­ing my mind about Chris­ti­ano’s ma­lign prior argument

Cole Wyeth4 Apr 2025 0:54 UTC
27 points
34 comments7 min readLW link