Ver­nor Vinge, who coined the term “Tech­nolog­i­cal Sin­gu­lar­ity”, dies at 79

Kaj_Sotala21 Mar 2024 22:14 UTC
148 points
24 comments1 min readLW link
(arstechnica.com)

ChatGPT can learn in­di­rect control

Raymond D21 Mar 2024 21:11 UTC
213 points
23 comments1 min readLW link

“Deep Learn­ing” Is Func­tion Approximation

Zack_M_Davis21 Mar 2024 17:50 UTC
95 points
28 comments10 min readLW link
(zackmdavis.net)

A Teacher vs. Every­one Else

ronak6921 Mar 2024 17:45 UTC
41 points
8 comments2 min readLW link

Static vs Dy­namic Alignment

Gracie Green21 Mar 2024 17:44 UTC
4 points
0 comments29 min readLW link

On green

Joe Carlsmith21 Mar 2024 17:38 UTC
260 points
35 comments31 min readLW link

Com­par­ing Align­ment to other AGI in­ter­ven­tions: Ex­ten­sions and analysis

Martín Soto21 Mar 2024 17:30 UTC
7 points
0 comments4 min readLW link

The Com­cast Problem

RamblinDash21 Mar 2024 16:46 UTC
0 points
15 comments1 min readLW link

Vi­pas­sana Med­i­ta­tion and Ac­tive In­fer­ence: A Frame­work for Un­der­stand­ing Suffer­ing and its Cessation

Benjamin Sturgeon21 Mar 2024 12:32 UTC
46 points
8 comments19 min readLW link

AI #56: Black­well That Ends Well

Zvi21 Mar 2024 12:10 UTC
34 points
16 comments68 min readLW link
(thezvi.wordpress.com)

An Afford­able CO2 Monitor

Pretentious Penguin21 Mar 2024 3:06 UTC
27 points
0 comments1 min readLW link

Deep­Mind: Eval­u­at­ing Fron­tier Models for Danger­ous Capabilities

Zach Stein-Perlman21 Mar 2024 3:00 UTC
61 points
5 comments1 min readLW link
(arxiv.org)

Where are the Con­tra Dances?

jefftk21 Mar 2024 2:00 UTC
9 points
0 comments1 min readLW link
(www.jefftk.com)

Slim overview of work one could do to make AI go bet­ter (and a grab-bag of other ca­reer con­sid­er­a­tions)

Chi Nguyen20 Mar 2024 23:17 UTC
9 points
1 comment1 min readLW link

How does AI solve prob­lems?

Dom Polsinelli20 Mar 2024 22:29 UTC
2 points
0 comments7 min readLW link

What I Learned (Con­clu­sion To “The Sense Of Phys­i­cal Ne­ces­sity”)

LoganStrohl20 Mar 2024 21:24 UTC
34 points
0 comments3 min readLW link

Stage­wise Devel­op­ment in Neu­ral Networks

20 Mar 2024 19:54 UTC
81 points
1 comment11 min readLW link

On the Glad­stone Report

Zvi20 Mar 2024 19:50 UTC
64 points
11 comments40 min readLW link
(thezvi.wordpress.com)

Nat­u­ral La­tents: The Concepts

20 Mar 2024 18:21 UTC
80 points
16 comments19 min readLW link

Com­par­ing Align­ment to other AGI in­ter­ven­tions: Ba­sic model

Martín Soto20 Mar 2024 18:17 UTC
12 points
4 comments7 min readLW link

AI-gen­er­ated opi­oids are a catas­trophic risk

ejk6420 Mar 2024 17:48 UTC
0 points
2 comments3 min readLW link

New re­port: Safety Cases for AI

joshc20 Mar 2024 16:45 UTC
90 points
13 comments1 min readLW link
(twitter.com)

User-in­cli­na­tion-guess­ing al­gorithms: reg­is­ter­ing a goal

ProgramCrafter20 Mar 2024 15:55 UTC
2 points
0 comments2 min readLW link

My MATS Sum­mer 2023 experience

James Chua20 Mar 2024 11:26 UTC
28 points
0 comments3 min readLW link
(jameschua.net)

[Question] What are the weirdest things a hu­man may want for their own sake?

Mateusz Bagiński20 Mar 2024 11:15 UTC
7 points
16 comments1 min readLW link

[Question] Best *or­ga­ni­za­tion* red-pill books and posts?

lukehmiles20 Mar 2024 7:01 UTC
10 points
2 comments1 min readLW link

Par­ent-Friendly Dance Weekends

jefftk20 Mar 2024 2:10 UTC
16 points
0 comments2 min readLW link
(www.jefftk.com)

[Question] “I Can’t Believe It Both Is and Is Not En­cephal­itis!” Or: What do you do when the ev­i­dence is crazy?

Erhannis19 Mar 2024 22:08 UTC
20 points
3 comments11 min readLW link

Delta’s of Change

Jonas Kgomo19 Mar 2024 21:03 UTC
1 point
0 comments4 min readLW link

In­creas­ing IQ by 10 Points is Possible

George3d619 Mar 2024 20:48 UTC
24 points
50 comments5 min readLW link
(morelucid.substack.com)

Are ex­treme prob­a­bil­ities for P(doom) epistem­i­cally jus­tifed?

19 Mar 2024 20:32 UTC
19 points
11 comments7 min readLW link

Have I Solved the Two En­velopes Prob­lem Once and For All?

JackOfAllTrades19 Mar 2024 19:57 UTC
−5 points
5 comments3 min readLW link

[Question] How can one be less wrong, if their con­ver­sa­tion part­ner loses the in­ter­est on dis­cussing the topic with them?

Ooker19 Mar 2024 18:11 UTC
−10 points
3 comments1 min readLW link

Carlo: un­cer­tainty anal­y­sis in Google Sheets

ProbabilityEnjoyer19 Mar 2024 17:59 UTC
6 points
0 comments1 min readLW link
(carlo.app)

NAIRA—An ex­er­cise in reg­u­la­tory, com­pet­i­tive safety gov­er­nance [AI Gover­nance In­sti­tu­tional De­sign idea]

Heramb19 Mar 2024 17:43 UTC
2 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

AI Safety Eval­u­a­tions: A Reg­u­la­tory Review

19 Mar 2024 15:05 UTC
21 points
1 comment11 min readLW link

Mechanism for fea­ture learn­ing in neu­ral net­works and back­prop­a­ga­tion-free ma­chine learn­ing models

Matt Goldenberg19 Mar 2024 14:55 UTC
8 points
1 comment1 min readLW link
(www.science.org)

Monthly Roundup #16: March 2024

Zvi19 Mar 2024 13:10 UTC
33 points
4 comments55 min readLW link
(thezvi.wordpress.com)

Claude es­ti­mates 30-50% like­li­hood x-risk

amelia19 Mar 2024 2:22 UTC
3 points
2 comments2 min readLW link

Ex­per­i­men­ta­tion (Part 7 of “The Sense Of Phys­i­cal Ne­ces­sity”)

LoganStrohl18 Mar 2024 21:25 UTC
33 points
0 comments10 min readLW link

INTERVIEW: Round 2 - StakeOut.AI w/​ Dr. Peter Park

jacobhaimes18 Mar 2024 21:21 UTC
5 points
0 comments1 min readLW link
(into-ai-safety.github.io)

Neu­ro­science and Alignment

Garrett Baker18 Mar 2024 21:09 UTC
40 points
25 comments2 min readLW link

GPT, the mag­i­cal col­lab­o­ra­tion zone, Lex Frid­man and Sam Altman

Bill Benzon18 Mar 2024 20:04 UTC
3 points
1 comment3 min readLW link

Mea­sur­ing Co­her­ence of Poli­cies in Toy Environments

18 Mar 2024 17:59 UTC
59 points
9 comments14 min readLW link

AtP*: An effi­cient and scal­able method for lo­cal­iz­ing LLM be­havi­our to components

18 Mar 2024 17:28 UTC
19 points
0 comments1 min readLW link
(arxiv.org)

Com­mu­nity Notes by X

NicholasKees18 Mar 2024 17:13 UTC
123 points
15 comments7 min readLW link

[Question] Is the Basilisk pre­tend­ing to be hid­den in this simu­la­tion so that it can check what I would do if con­di­tioned by a world with­out the Basilisk?

maybefbi18 Mar 2024 16:05 UTC
−18 points
1 comment1 min readLW link

On Devin

Zvi18 Mar 2024 13:20 UTC
147 points
30 comments11 min readLW link
(thezvi.wordpress.com)

RLLMv10 experiment

MiguelDev18 Mar 2024 8:32 UTC
5 points
0 comments2 min readLW link

Join the AI Eval­u­a­tion Tasks Bounty Hackathon

Esben Kran18 Mar 2024 8:15 UTC
12 points
1 comment1 min readLW link