The likely first longevity drug is based on sketchy sci­ence. This is bad for sci­ence and bad for longevity.

BobBurgersDec 12, 2023, 2:42 AM
161 points
34 comments5 min readLW link

AI doom from an LLM-plateau-ist perspective

Steven ByrnesApr 27, 2023, 1:58 PM
161 points
24 comments6 min readLW link

Meta Ques­tions about Metaphilosophy

Wei DaiSep 1, 2023, 1:17 AM
161 points
80 comments3 min readLW link

Change my mind: Ve­ganism en­tails trade-offs, and health is one of the axes

ElizabethJun 1, 2023, 5:10 PM
160 points
85 comments19 min readLW link2 reviews
(acesounderglass.com)

Jailbreak­ing GPT-4′s code interpreter

Nikola JurkovicJul 13, 2023, 6:43 PM
160 points
22 comments7 min readLW link

Agen­tized LLMs will change the al­ign­ment landscape

Seth HerdApr 9, 2023, 2:29 AM
160 points
102 comments3 min readLW link1 review

“Di­a­mon­doid bac­te­ria” nanobots: deadly threat or dead-end? A nan­otech in­ves­ti­ga­tion

titotalSep 29, 2023, 2:01 PM
160 points
79 commentsLW link
(titotal.substack.com)

Sparse Au­toen­coders Find Highly In­ter­pretable Direc­tions in Lan­guage Models

Sep 21, 2023, 3:30 PM
159 points
8 comments5 min readLW link

Vote on In­ter­est­ing Disagreements

Ben PaceNov 7, 2023, 9:35 PM
159 points
131 comments1 min readLW link

Most Peo­ple Don’t Real­ize We Have No Idea How Our AIs Work

Thane RuthenisDec 21, 2023, 8:02 PM
159 points
42 comments1 min readLW link

Succession

Richard_NgoDec 20, 2023, 7:25 PM
159 points
48 comments11 min readLW link
(www.narrativeark.xyz)

POC || GTFO cul­ture as par­tial an­ti­dote to al­ign­ment wordcelism

lcMar 15, 2023, 10:21 AM
158 points
15 comments7 min readLW link2 reviews

Big Mac Sub­sidy?

jefftkFeb 23, 2023, 4:00 AM
158 points
25 comments2 min readLW link
(www.jefftk.com)

What would a com­pute mon­i­tor­ing plan look like? [Linkpost]

Orpheus16Mar 26, 2023, 7:33 PM
158 points
10 comments4 min readLW link
(arxiv.org)

In­side the mind of a su­per­hu­man Go model: How does Leela Zero read lad­ders?

Haoxing DuMar 1, 2023, 1:47 AM
157 points
8 comments30 min readLW link

My thoughts on the so­cial re­sponse to AI risk

Matthew BarnettNov 1, 2023, 9:17 PM
157 points
37 comments10 min readLW link

Pass­word-locked mod­els: a stress case for ca­pa­bil­ities evaluation

Fabien RogerAug 3, 2023, 2:53 PM
156 points
14 comments6 min readLW link

grey goo is unlikely

bhauthApr 17, 2023, 1:59 AM
156 points
123 comments9 min readLW link2 reviews
(bhauth.com)

Sapir-Whorf for Rationalists

Duncan Sabien (Inactive)Jan 25, 2023, 7:58 AM
155 points
49 comments19 min readLW link

Con­jec­ture in­ter­nal sur­vey: AGI timelines and prob­a­bil­ity of hu­man ex­tinc­tion from ad­vanced AI

Maris SalaMay 22, 2023, 2:31 PM
155 points
5 comments3 min readLW link
(www.conjecture.dev)

An­nounc­ing Dialogues

Ben PaceOct 7, 2023, 2:57 AM
155 points
59 comments4 min readLW link

AI: Prac­ti­cal Ad­vice for the Worried

ZviMar 1, 2023, 12:30 PM
155 points
49 comments16 min readLW link2 reviews
(thezvi.wordpress.com)

The self-un­al­ign­ment problem

Apr 14, 2023, 12:10 PM
155 points
24 comments10 min readLW link

Re­quest: stop ad­vanc­ing AI capabilities

So8resMay 26, 2023, 5:42 PM
154 points
24 comments1 min readLW link

A fresh­man year dur­ing the AI midgame: my ap­proach to the next year

BuckApr 14, 2023, 12:38 AM
154 points
15 commentsLW link1 review

Will no one rid me of this tur­bu­lent pest?

MetacelsusOct 14, 2023, 3:27 PM
154 points
23 comments10 min readLW link
(denovo.substack.com)

ARC Evals new re­port: Eval­u­at­ing Lan­guage-Model Agents on Real­is­tic Au­tonomous Tasks

Beth BarnesAug 1, 2023, 6:30 PM
153 points
12 comments5 min readLW link
(evals.alignment.org)

As­sume Bad Faith

Zack_M_DavisAug 25, 2023, 5:36 PM
153 points
63 comments7 min readLW link3 reviews

The Plan − 2023 Version

johnswentworthDec 29, 2023, 11:34 PM
152 points
40 comments31 min readLW link1 review

Shut­ting down AI is not enough. We need to de­stroy all tech­nol­ogy.

Matthew BarnettApr 1, 2023, 9:03 PM
152 points
36 comments1 min readLW link

LoRA Fine-tun­ing Effi­ciently Un­does Safety Train­ing from Llama 2-Chat 70B

Oct 12, 2023, 7:58 PM
151 points
29 comments14 min readLW link

GPT-4

nzMar 14, 2023, 5:02 PM
151 points
150 comments1 min readLW link
(openai.com)

AI x-risk, ap­prox­i­mately or­dered by embarrassment

Alex Lawsen Apr 12, 2023, 11:01 PM
151 points
7 comments19 min readLW link

Why Not Just Out­source Align­ment Re­search To An AI?

johnswentworthMar 9, 2023, 9:49 PM
151 points
50 comments9 min readLW link1 review

Ad­vice for newly busy people

Severin T. SeehrichMay 11, 2023, 4:46 PM
150 points
3 comments5 min readLW link

OpenAI Launches Su­per­al­ign­ment Taskforce

ZviJul 11, 2023, 1:00 PM
150 points
40 comments49 min readLW link
(thezvi.wordpress.com)

Why I’m not into the Free En­ergy Principle

Steven ByrnesMar 2, 2023, 7:27 PM
150 points
50 comments9 min readLW link1 review

There are no co­her­ence theorems

Feb 20, 2023, 9:25 PM
149 points
130 comments19 min readLW link1 review

Mo­ral Real­ity Check (a short story)

jessicataNov 26, 2023, 5:03 AM
149 points
45 comments21 min readLW link1 review
(unstableontology.com)

The U.S. is be­com­ing less stable

lcAug 18, 2023, 9:13 PM
149 points
68 comments2 min readLW link

Dan Luu on “You can only com­mu­ni­cate one top pri­or­ity”

RaemonMar 18, 2023, 6:55 PM
149 points
18 comments3 min readLW link
(twitter.com)

Brain Effi­ciency Can­nell Prize Con­test Award Ceremony

Alexander Gietelink OldenzielJul 24, 2023, 11:30 AM
149 points
12 comments7 min readLW link

Com­ments on OpenAI’s “Plan­ning for AGI and be­yond”

So8resMar 3, 2023, 11:01 PM
148 points
2 comments14 min readLW link

At 87, Pearl is still able to change his mind

rotatingpaguro18 Oct 2023 4:46 UTC
148 points
15 comments5 min readLW link

Could a su­per­in­tel­li­gence de­duce gen­eral rel­a­tivity from a fal­ling ap­ple? An investigation

titotal23 Apr 2023 12:49 UTC
148 points
39 comments9 min readLW link

Dis­cus­sion: Challenges with Un­su­per­vised LLM Knowl­edge Discovery

18 Dec 2023 11:58 UTC
147 points
21 comments10 min readLW link

6 non-ob­vi­ous men­tal health is­sues spe­cific to AI safety

Igor Ivanov18 Aug 2023 15:46 UTC
147 points
24 comments4 min readLW link

“Hereti­cal Thoughts on AI” by Eli Dourado

DragonGod19 Jan 2023 16:11 UTC
146 points
38 comments3 min readLW link
(www.elidourado.com)

Does davi­dad’s up­load­ing moon­shot work?

3 Nov 2023 2:21 UTC
146 points
35 comments25 min readLW link

Al­gorith­mic Im­prove­ment Is Prob­a­bly Faster Than Scal­ing Now

johnswentworth6 Jun 2023 2:57 UTC
146 points
25 comments2 min readLW link