Nick Land: Orthogonality

lumpenspace4 Feb 2025 21:07 UTC
5 points
37 comments8 min readLW link

What work­ing on AI safety taught me about B2B SaaS sales

purple fire4 Feb 2025 20:50 UTC
7 points
12 comments5 min readLW link

Sub­jec­tive Nat­u­ral­ism in De­ci­sion The­ory: Sav­age vs. Jeffrey–Bolker

4 Feb 2025 20:34 UTC
45 points
22 comments5 min readLW link

Anti-Slop In­ter­ven­tions?

abramdemski4 Feb 2025 19:50 UTC
76 points
33 comments6 min readLW link

Can Per­sua­sion Break AI Safety? Ex­plor­ing the In­ter­play Between Fine-Tun­ing, At­tacks, and Guardrails

Devina Jain4 Feb 2025 19:10 UTC
9 points
0 comments10 min readLW link

[Question] Jour­nal­ism stu­dent look­ing for sources

pinkerton4 Feb 2025 18:58 UTC
11 points
3 comments1 min readLW link

We’re in Deep Research

Zvi4 Feb 2025 17:20 UTC
45 points
3 comments20 min readLW link
(thezvi.wordpress.com)

The Cap­i­tal­ist Agent

henophilia4 Feb 2025 15:32 UTC
1 point
10 comments3 min readLW link
(blog.hermesloom.org)

Fore­cast­ing AGI: In­sights from Pre­dic­tion Mar­kets and Metaculus

Alvin Ånestrand4 Feb 2025 13:03 UTC
13 points
0 comments4 min readLW link
(forecastingaifutures.substack.com)

Rul­ing Out Lookup Tables

Alfred Harwood4 Feb 2025 10:39 UTC
22 points
11 comments7 min readLW link

Half-baked idea: a straight­for­ward method for learn­ing en­vi­ron­men­tal goals?

Q Home4 Feb 2025 6:56 UTC
16 points
7 comments5 min readLW link

In­for­ma­tion Ver­sus Action

Screwtape4 Feb 2025 5:13 UTC
27 points
0 comments6 min readLW link

Utili­tar­ian AI Align­ment: Build­ing a Mo­ral As­sis­tant with the Con­sti­tu­tional AI Method

Clément L4 Feb 2025 4:15 UTC
6 points
1 comment13 min readLW link

Tear Down the Burren

jefftk4 Feb 2025 3:40 UTC
45 points
2 comments2 min readLW link
(www.jefftk.com)

Con­sti­tu­tional Clas­sifiers: Defend­ing against uni­ver­sal jailbreaks (An­thropic Blog)

Archimedes4 Feb 2025 2:55 UTC
17 points
1 comment1 min readLW link
(www.anthropic.com)

Can some­one, any­one, make su­per­in­tel­li­gence a more con­crete con­cept?

Ori Nagel4 Feb 2025 2:18 UTC
2 points
8 comments5 min readLW link

What are the “no free lunch” the­o­rems?

4 Feb 2025 2:02 UTC
19 points
4 comments1 min readLW link
(aisafety.info)

elimi­nat­ing bias through lan­guage?

KvmanThinking4 Feb 2025 1:52 UTC
1 point
12 comments1 min readLW link

New Fore­sight Longevity Bio & Molec­u­lar Nano Grants Program

Allison Duettmann4 Feb 2025 0:28 UTC
11 points
0 comments1 min readLW link

Meta: Fron­tier AI Framework

Zach Stein-Perlman3 Feb 2025 22:00 UTC
33 points
2 comments1 min readLW link
(ai.meta.com)

$300 Fermi Model Competition

ozziegooen3 Feb 2025 19:47 UTC
16 points
18 comments2 min readLW link

Vi­su­al­iz­ing Interpretability

Darold Davis3 Feb 2025 19:36 UTC
2 points
0 comments4 min readLW link

Align­ment Can Re­duce Perfor­mance on Sim­ple Eth­i­cal Questions

Daan Henselmans3 Feb 2025 19:35 UTC
16 points
7 comments6 min readLW link

The Over­lap Paradigm: Re­think­ing Data’s Role in Weak-to-Strong Gen­er­al­iza­tion (W2SG)

Serhii Zamrii3 Feb 2025 19:31 UTC
2 points
0 comments11 min readLW link

Sleeper agents ap­pear re­silient to ac­ti­va­tion steering

Lucy Wingard3 Feb 2025 19:31 UTC
6 points
0 comments7 min readLW link

Part 1: En­hanc­ing In­ner Align­ment in CLIP Vi­sion Trans­form­ers: Miti­gat­ing Reifi­ca­tion Bias with SAEs and Grad ECLIP

Gilber A. Corrales3 Feb 2025 19:30 UTC
1 point
0 comments13 min readLW link

Su­per­in­tel­li­gence Align­ment Proposal

Davey Morse3 Feb 2025 18:47 UTC
5 points
3 comments9 min readLW link

The Self-Refer­ence Trap in Mathematics

Alister Munday3 Feb 2025 16:12 UTC
−41 points
23 comments2 min readLW link

Stop­ping un­al­igned LLMs is easy!

Yair Halberstadt3 Feb 2025 15:38 UTC
−3 points
11 comments2 min readLW link

The Outer Levels

Jerdle3 Feb 2025 14:30 UTC
2 points
3 comments6 min readLW link

o3-mini Early Days

Zvi3 Feb 2025 14:20 UTC
45 points
0 comments15 min readLW link
(thezvi.wordpress.com)

OpenAI re­leases deep re­search agent

Seth Herd3 Feb 2025 12:48 UTC
78 points
21 comments3 min readLW link
(openai.com)

Neu­ron Ac­ti­va­tions to CLIP Embed­dings: Geom­e­try of Lin­ear Com­bi­na­tions in La­tent Space

Roman Malov3 Feb 2025 10:30 UTC
5 points
0 comments2 min readLW link

[Question] Can we in­fer the search space of a lo­cal op­ti­miser?

Lucius Bushnaq3 Feb 2025 10:17 UTC
25 points
5 comments3 min readLW link

Pick two: con­cise, com­pre­hen­sive, or clear rules

Screwtape3 Feb 2025 6:39 UTC
82 points
27 comments8 min readLW link

Lan­guage Models and World Models, a Philosophy

kyjohnso3 Feb 2025 2:55 UTC
1 point
0 comments1 min readLW link
(hylaeansea.org)

Keep­ing Cap­i­tal is the Challenge

LTM3 Feb 2025 2:04 UTC
13 points
2 comments17 min readLW link
(routecause.substack.com)

Use com­put­ers as pow­er­ful as in 1985 or AI con­trols hu­mans or ?

jrincayc3 Feb 2025 0:51 UTC
3 points
0 comments2 min readLW link

Some Th­e­ses on Mo­ti­va­tional and Direc­tional Feedback

abstractapplic2 Feb 2025 22:50 UTC
10 points
3 comments4 min readLW link

Hu­man­ity Has A Pos­si­ble 99.98% Chance Of Ex­tinc­tion

st3rlxx2 Feb 2025 21:46 UTC
−12 points
1 comment5 min readLW link

Ex­plor­ing how Othel­loGPT com­putes its world model

JMaar2 Feb 2025 21:29 UTC
8 points
0 comments8 min readLW link

An In­tro­duc­tion to Ev­i­den­tial De­ci­sion Theory

Babić2 Feb 2025 21:27 UTC
5 points
2 comments10 min readLW link

“DL train­ing == hu­man learn­ing” is a bad analogy

kman2 Feb 2025 20:59 UTC
3 points
0 comments1 min readLW link

Con­di­tional Im­por­tance in Toy Models of Superposition

james__p2 Feb 2025 20:35 UTC
9 points
4 comments10 min readLW link

Trac­ing Ty­pos in LLMs: My At­tempt at Un­der­stand­ing How Models Cor­rect Misspellings

Ivan Dostal2 Feb 2025 19:56 UTC
4 points
1 comment5 min readLW link

The Sim­plest Good

Jesse Hoogland2 Feb 2025 19:51 UTC
76 points
6 comments5 min readLW link

Grad­ual Disem­pow­er­ment, Shell Games and Flinches

Jan_Kulveit2 Feb 2025 14:47 UTC
133 points
36 comments6 min readLW link

Thoughts on Toy Models of Superposition

james__p2 Feb 2025 13:52 UTC
5 points
2 comments9 min readLW link

Es­cape from Alder­aan I

lsusr2 Feb 2025 10:48 UTC
59 points
2 comments6 min readLW link

ChatGPT: Ex­plor­ing the Digi­tal Wilder­ness, Find­ings and Prospects

Bill Benzon2 Feb 2025 9:54 UTC
2 points
0 comments5 min readLW link