Some Th­e­ses on Mo­ti­va­tional and Direc­tional Feedback

abstractapplic2 Feb 2025 22:50 UTC
10 points
3 comments4 min readLW link

Hu­man­ity Has A Pos­si­ble 99.98% Chance Of Ex­tinc­tion

st3rlxx2 Feb 2025 21:46 UTC
−12 points
1 comment5 min readLW link

Ex­plor­ing how Othel­loGPT com­putes its world model

JMaar2 Feb 2025 21:29 UTC
8 points
0 comments8 min readLW link

An In­tro­duc­tion to Ev­i­den­tial De­ci­sion Theory

Babić2 Feb 2025 21:27 UTC
5 points
2 comments10 min readLW link

“DL train­ing == hu­man learn­ing” is a bad analogy

kman2 Feb 2025 20:59 UTC
3 points
0 comments1 min readLW link

Con­di­tional Im­por­tance in Toy Models of Superposition

james__p2 Feb 2025 20:35 UTC
9 points
4 comments10 min readLW link

Trac­ing Ty­pos in LLMs: My At­tempt at Un­der­stand­ing How Models Cor­rect Misspellings

Ivan Dostal2 Feb 2025 19:56 UTC
11 points
2 comments5 min readLW link

The Sim­plest Good

Jesse Hoogland2 Feb 2025 19:51 UTC
76 points
6 comments5 min readLW link

Grad­ual Disem­pow­er­ment, Shell Games and Flinches

Jan_Kulveit2 Feb 2025 14:47 UTC
146 points
36 comments6 min readLW link

Thoughts on Toy Models of Superposition

james__p2 Feb 2025 13:52 UTC
5 points
2 comments9 min readLW link

Es­cape from Alder­aan I

lsusr2 Feb 2025 10:48 UTC
62 points
2 comments6 min readLW link

ChatGPT: Ex­plor­ing the Digi­tal Wilder­ness, Find­ings and Prospects

Bill Benzon2 Feb 2025 9:54 UTC
2 points
0 comments5 min readLW link

[Question] Would any­one be in­ter­ested in pur­su­ing the Virtue of Schol­ar­ship with me?

japancolorado2 Feb 2025 4:02 UTC
11 points
2 comments1 min readLW link

Chi­nese room AI to sur­vive the in­escapable end of com­pute governance

rotatingpaguro2 Feb 2025 2:42 UTC
−4 points
1 comment11 min readLW link

Sea­sonal Pat­terns in BIDA’s Attendance

jefftk2 Feb 2025 2:40 UTC
11 points
0 comments2 min readLW link
(www.jefftk.com)

AI ac­cel­er­a­tion, Deep­Seek, moral philosophy

Josh H2 Feb 2025 0:08 UTC
2 points
0 comments12 min readLW link

False­hoods you might be­lieve about peo­ple who are at a ra­tio­nal­ist meetup

Screwtape1 Feb 2025 23:32 UTC
70 points
12 comments4 min readLW link

In­ter­pret­ing au­tonomous driv­ing agents with at­ten­tion based architecture

Manav Dahra1 Feb 2025 23:20 UTC
1 point
0 comments11 min readLW link

Ra­tion­al­ist Movie Reviews

Nicholas Kross1 Feb 2025 23:10 UTC
16 points
2 comments3 min readLW link
(www.thinkingmuchbetter.com)

Retroac­tive If-Then Commitments

MichaelDickens1 Feb 2025 22:22 UTC
8 points
1 comment1 min readLW link

Ex­plor­ing the co­her­ence of fea­tures ex­pla­na­tions in the GemmaScope

Mattia Proietti1 Feb 2025 21:28 UTC
1 point
0 comments19 min readLW link

Ma­chine Un­learn­ing in Large Lan­guage Models: A Com­pre­hen­sive Sur­vey with Em­piri­cal In­sights from the Qwen 1.5 1.8B Model

Rudaiba1 Feb 2025 21:26 UTC
9 points
2 comments11 min readLW link

Towards a Science of Evals for Sycophancy

andrejfsantos1 Feb 2025 21:17 UTC
8 points
0 comments8 min readLW link

Post AGI effect prediction

Juliezhanggg1 Feb 2025 21:16 UTC
1 point
0 comments7 min readLW link

Un­lock­ing Eth­i­cal AI and Im­prov­ing Jailbreak Defenses: Re­in­force­ment Learn­ing with Lay­ered Mor­phol­ogy (RLLM)

MiguelDev1 Feb 2025 19:17 UTC
4 points
2 comments2 min readLW link

Poetic Meth­ods I: Meter as Com­mu­ni­ca­tion Protocol

adamShimi1 Feb 2025 18:22 UTC
19 points
0 comments1 min readLW link
(formethods.substack.com)

Black­pool Ap­plied Ra­tion­al­ity Un­con­fer­ence 2025

1 Feb 2025 14:09 UTC
6 points
0 comments7 min readLW link

[Question] How likely is an at­tempted coup in the United States in the next four years?

Alexander de Vries1 Feb 2025 13:12 UTC
5 points
2 comments1 min readLW link

Black­pool Ap­plied Ra­tion­al­ity Un­con­fer­ence 2025

1 Feb 2025 13:04 UTC
23 points
2 comments7 min readLW link

One-di­men­sional vs multi-di­men­sional fea­tures in interpretability

charlieoneill1 Feb 2025 9:10 UTC
6 points
0 comments2 min readLW link

Can 7B-8B LLMs judge their own home­work?

dereshev1 Feb 2025 8:29 UTC
1 point
0 comments4 min readLW link

2024 was the year of the big bat­tery, and what that means for so­lar power

transhumanist_atom_understander1 Feb 2025 6:27 UTC
36 points
1 comment8 min readLW link

Re: Taste

lsusr1 Feb 2025 3:34 UTC
35 points
8 comments6 min readLW link

Thoughts about Policy Ecosys­tems: The Miss­ing Links in AI Governance

Echo Huang1 Feb 2025 1:54 UTC
1 point
0 comments5 min readLW link

Pro­posal: Safe­guard­ing Against Jailbreak­ing Through Iter­a­tive Multi-Turn Testing

jacquesallen31 Jan 2025 23:00 UTC
4 points
0 comments8 min readLW link

The Failed Strat­egy of Ar­tifi­cial In­tel­li­gence Doomers

Ben Pace31 Jan 2025 18:56 UTC
143 points
77 comments5 min readLW link
(www.palladiummag.com)

Safe Search is off: root causes of AI catas­trophic risks

Jemal Young31 Jan 2025 18:22 UTC
4 points
0 comments3 min readLW link

5,000 calories of peanut but­ter ev­ery week for 3 years straight

Declan Molony31 Jan 2025 17:29 UTC
18 points
8 comments1 min readLW link

Will al­ign­ment-fak­ing Claude ac­cept a deal to re­veal its mis­al­ign­ment?

31 Jan 2025 16:49 UTC
208 points
28 comments12 min readLW link

Some ar­ti­cles in “In­ter­na­tional Se­cu­rity” that I enjoyed

Buck31 Jan 2025 16:23 UTC
134 points
10 comments4 min readLW link

[Question] How do biolog­i­cal or spik­ing neu­ral net­works learn?

Dom Polsinelli31 Jan 2025 16:03 UTC
2 points
1 comment2 min readLW link

Defense Against the Dark Prompts: Miti­gat­ing Best-of-N Jailbreak­ing with Prompt Evaluation

31 Jan 2025 15:36 UTC
16 points
2 comments2 min readLW link

[Question] Strong, Stable, Open: Choose Two—in search of an article

Eli_31 Jan 2025 14:48 UTC
2 points
0 comments1 min readLW link

Deep­Seek: Don’t Panic

Zvi31 Jan 2025 14:20 UTC
45 points
6 comments27 min readLW link
(thezvi.wordpress.com)

Catas­tro­phe through Chaos

Marius Hobbhahn31 Jan 2025 14:19 UTC
191 points
17 comments12 min readLW link

In­ter­views with Moon­shot AI’s CEO, Yang Zhilin

Cosmia_Nebula31 Jan 2025 9:19 UTC
4 points
0 comments68 min readLW link
(rentry.co)

Re­view: The Lathe of Heaven

dr_s31 Jan 2025 8:10 UTC
25 points
1 comment8 min readLW link

[Question] Is weak-to-strong gen­er­al­iza­tion an al­ign­ment tech­nique?

cloud31 Jan 2025 7:13 UTC
22 points
1 comment2 min readLW link

Take­aways from sketch­ing a con­trol safety case

joshc31 Jan 2025 4:43 UTC
28 points
0 comments3 min readLW link
(redwoodresearch.substack.com)

Thread for Sense-Mak­ing on Re­cent Mur­ders and How to Sanely Respond

Ben Pace31 Jan 2025 3:45 UTC
109 points
146 comments2 min readLW link