There’s more low-hang­ing fruit in in­ter­dis­ci­plinary work thanks to LLMs

ChristianKlMay 7, 2025, 7:48 PM
26 points
2 comments1 min readLW link

OpenAI Claims Non­profit Will Re­tain Nom­i­nal Control

ZviMay 7, 2025, 7:40 PM
65 points
4 comments11 min readLW link
(thezvi.wordpress.com)

So­cial sta­tus games might have “com­pute weight class” in the future

RaemonMay 7, 2025, 6:56 PM
31 points
5 comments2 min readLW link

Events of Low Prob­a­bil­ity: Buri­dan’s Principle

Nikita GladkovMay 7, 2025, 6:46 PM
12 points
0 comments10 min readLW link

[Question] Which jour­nal­ists would you give quotes to? [one jour­nal­ist per com­ment, agree vote for trust­wor­thy]

Nathan YoungMay 7, 2025, 6:39 PM
12 points
26 comments1 min readLW link

Progress = Fewer Bad Moments

ChipmonkMay 7, 2025, 5:33 PM
24 points
9 comments2 min readLW link
(chrislakin.blog)

Please Donate to CAIP (Post 1 of 6 on AI Gover­nance)

Mass_DriverMay 7, 2025, 5:13 PM
113 points
20 comments33 min readLW link

UK AISI’s Align­ment Team: Re­search Agenda

May 7, 2025, 4:33 PM
111 points
2 comments11 min readLW link

Four Pre­dic­tions About OpenAI’s Plans To Re­tain Non­profit Control

garrisonMay 7, 2025, 3:48 PM
12 points
0 commentsLW link
(www.obsolete.pub)

A Dis­ci­plined Way to Avoid Wireheading

amitlevy49May 7, 2025, 3:20 PM
18 points
6 comments5 min readLW link
(ivy0.substack.com)

Reflec­tions on Com­pat­i­bil­ism, On­tolog­i­cal Trans­la­tions, and the Ar­tifi­cial Divine

Mahdi ComplexMay 7, 2025, 12:16 PM
2 points
1 comment22 min readLW link

The His­tor­i­cal Par­allels: Pre­limi­nary Reflection

EQMay 7, 2025, 8:06 AM
3 points
0 comments9 min readLW link
(eqmind.substack.com)

Euro­pean Links (07.05.25)

Martin SustrikMay 7, 2025, 4:20 AM
10 points
0 comments2 min readLW link
(250bpm.substack.com)

[Question] Chess—“Elo” of ran­dom play?

Shankar SivarajanMay 7, 2025, 2:18 AM
10 points
16 comments1 min readLW link

$500 + $500 Bounty Prob­lem: Does An (Ap­prox­i­mately) Deter­minis­tic Max­i­mal Re­dund Always Ex­ist?

May 6, 2025, 11:05 PM
72 points
11 comments3 min readLW link

The The­ory Be­hind Loss Curves

James CamachoMay 6, 2025, 10:22 PM
16 points
3 comments4 min readLW link
(github.com)

Nega­tive Re­sults on Group SAEs

Josh EngelsMay 6, 2025, 9:49 PM
70 points
3 comments8 min readLW link

ACX At­lanta May 2025 Meetup

Steve FrenchMay 6, 2025, 9:00 PM
2 points
0 comments1 min readLW link

[Question] What kind of policy by an AGI would make peo­ple happy?

StanislavKrymMay 6, 2025, 6:05 PM
1 point
2 comments1 min readLW link

Paper High­lights, April ’25

gasteigerjoMay 6, 2025, 2:22 PM
4 points
0 comments7 min readLW link
(aisafetyfrontier.substack.com)

Zucker­berg’s Dystopian AI Vision

ZviMay 6, 2025, 1:50 PM
61 points
7 comments11 min readLW link
(thezvi.wordpress.com)

Will pro­tein de­sign tools solve the snake an­tivenom short­age?

Abhishaike MahajanMay 6, 2025, 1:11 PM
30 points
0 comments17 min readLW link
(www.owlposting.com)

Utah Court Case Over State Law Re­gard­ing “Per­son­hood” for Non­hu­man Intelligences

Stephen MartinMay 6, 2025, 12:54 PM
10 points
3 comments2 min readLW link

Global Risks Weekly Roundup #18/​2025: US tar­iff short­ages, mil­i­tary polic­ing, Gaza famine.

NunoSempereMay 6, 2025, 10:39 AM
31 points
2 comments3 min readLW link
(blog.sentinel-team.org)

OpenAI’s Jig May Be Up

ValeMay 6, 2025, 8:51 AM
3 points
2 comments3 min readLW link

My Rea­sons for Us­ing Anki

Parker ConleyMay 6, 2025, 7:01 AM
10 points
0 comments3 min readLW link
(parconley.com)

It’s ‘Well, ac­tu­ally...’ all the way down

benwrMay 6, 2025, 5:44 AM
40 points
34 comments1 min readLW link
(www.benwr.net)

Five Hinge‑Ques­tions That De­cide Whether AGI Is Five Years Away or Twenty

charlieoneillMay 6, 2025, 2:48 AM
125 points
17 comments5 min readLW link

Non­profit to re­tain con­trol of OpenAI

ArchimedesMay 5, 2025, 11:41 PM
37 points
1 comment1 min readLW link
(openai.com)

Un­ex­pected Con­scious Entities

Gunnar_ZarnckeMay 5, 2025, 10:14 PM
34 points
6 comments6 min readLW link

The First Law of Con­scious Agency: Lin­guis­tic Rel­a­tivity and the Birth of “I”

Dima (lain)May 5, 2025, 9:20 PM
−17 points
4 comments2 min readLW link

New­ton’s sec­ond law ex­plained: it works in many universes

TahpMay 5, 2025, 7:47 PM
19 points
10 comments15 min readLW link
(quark.rodeo)

Repli­ca­tor->Ve­hi­cle Align­ment and Hu­man->AI Alignment

derelict5432May 5, 2025, 7:23 PM
0 points
3 comments4 min readLW link

The Sweet Les­son: AI Safety Should Scale With Compute

Jesse HooglandMay 5, 2025, 7:03 PM
95 points
3 comments3 min readLW link

[Question] Blue light, ‘Adrenal ASMR’: strange ex­pe­riences I can’t find any liter­a­ture about

vernichtungMay 5, 2025, 6:58 PM
16 points
6 comments1 min readLW link

Ts­inghua pa­per: Does RL Really In­cen­tivize Rea­son­ing Ca­pac­ity in LLMs Beyond the Base Model?

Thomas KwaMay 5, 2025, 6:56 PM
68 points
21 comments2 min readLW link
(arxiv.org)

In­tro & Pro­posal for AGI Model

PickleBrineMay 5, 2025, 6:56 PM
0 points
0 comments3 min readLW link

AI Su­per­or­ganisms: An Alter­na­tive Path­way to Ar­tifi­cial Superintelligence

Aaron VanzylMay 5, 2025, 6:55 PM
4 points
5 comments15 min readLW link

Kar­ls­ruhe ACX: The colours of her coat

wilmMay 5, 2025, 6:35 PM
2 points
0 comments1 min readLW link

The Me­tac­u­lus Cup Series Is Live, $5,000 Prize Pool

ChristianWilliamsMay 5, 2025, 5:14 PM
4 points
0 commentsLW link
(www.metaculus.com)

Com­mu­nity Feed­back Re­quest: AI Safety In­tro for Gen­eral Public

May 5, 2025, 4:38 PM
6 points
5 comments3 min readLW link

GPT-4o Sy­co­phancy Post Mortem

ZviMay 5, 2025, 4:00 PM
55 points
1 comment16 min readLW link
(thezvi.wordpress.com)

Le­gal Su­per­vi­sion of Fron­tier AI Labs is the an­swer.

GauraventhMay 5, 2025, 1:36 PM
14 points
2 comments3 min readLW link
(robertandgaurav.substack.com)

The cru­cible — how I think about the situ­a­tion with AI

owencbMay 5, 2025, 1:18 PM
25 points
1 comment8 min readLW link
(strangecities.substack.com)

Light­ning Talks: Thought, Trick, Curiosity

marta_kMay 5, 2025, 11:49 AM
1 point
0 comments1 min readLW link

Are stan­dard­ized tests effec­tive?

HrussMay 5, 2025, 10:02 AM
1 point
1 comment1 min readLW link

Pro­posal: Liquid Pre­dic­tion Mar­kets for AI Forecasting

Jesse Richardson5 May 2025 5:13 UTC
23 points
2 comments3 min readLW link

Why “Solv­ing Align­ment” Is Likely a Cat­e­gory Mistake

Nate Sharpe5 May 2025 4:26 UTC
22 points
3 comments3 min readLW link

AI, An­i­mals, & Digi­tal Minds 2025: ap­ply to speak by Wed­nes­day!

Alistair Stewart5 May 2025 0:56 UTC
4 points
0 comments1 min readLW link

AI, An­i­mals, & Digi­tal Minds 2025

Alistair Stewart5 May 2025 0:51 UTC
2 points
0 comments1 min readLW link