[Question] Where to find re­li­able re­views of AI prod­ucts?

ElizabethSep 17, 2024, 11:48 PM
29 points
6 comments1 min readLW link

Su­per­po­si­tion through Ac­tive Learn­ing Lens

akankshancSep 17, 2024, 5:32 PM
1 point
0 comments10 min readLW link

Sur­vey—Psy­cholog­i­cal Im­pact of Long-Term AI Engagement

Manuela GarcíaSep 17, 2024, 5:31 PM
2 points
0 comments1 min readLW link

Sur­vey—Psy­cholog­i­cal Im­pact of Long-Term AI Engagement

Manuela GarcíaSep 17, 2024, 5:31 PM
1 point
1 comment1 min readLW link

[Question] What does it mean for an event or ob­ser­va­tion to have prob­a­bil­ity 0 or 1 in Bayesian terms?

Noosphere89Sep 17, 2024, 5:28 PM
1 point
22 comments1 min readLW link

How harm­ful is mu­sic, re­ally?

dkl9Sep 17, 2024, 2:53 PM
10 points
6 comments3 min readLW link
(dkl9.net)

Monthly Roundup #22: Septem­ber 2024

ZviSep 17, 2024, 12:20 PM
35 points
10 comments45 min readLW link
(thezvi.wordpress.com)

I fi­nally got ChatGPT to sound like me

lsusrSep 17, 2024, 9:39 AM
47 points
18 comments6 min readLW link

Food, Pri­son & Ex­otic An­i­mals: Sparse Au­toen­coders De­tect 6.5x Perform­ing Youtube Thumbnails

Louka Ewington-PitsosSep 17, 2024, 3:52 AM
6 points
2 comments7 min readLW link

Head in the Cloud: Why an Upload of Your Mind is Not You

xhqSep 17, 2024, 12:25 AM
−11 points
3 comments14 min readLW link

[Question] How does some­one prove that their gen­eral in­tel­li­gence is above av­er­age?

M. Y. ZuoSep 16, 2024, 9:01 PM
−3 points
12 comments1 min readLW link

[Question] Does life ac­tu­ally lo­cally *in­crease* en­tropy?

tailcalledSep 16, 2024, 8:30 PM
10 points
27 comments1 min readLW link

Book re­view: Xenosystems

jessicataSep 16, 2024, 8:17 PM
50 points
18 comments37 min readLW link
(unstableontology.com)

MIRI’s Septem­ber 2024 newsletter

HarlanSep 16, 2024, 6:15 PM
46 points
0 comments1 min readLW link
(intelligence.org)

Gen­er­a­tive ML in chem­istry is bot­tle­necked by synthesis

Abhishaike MahajanSep 16, 2024, 4:31 PM
38 points
2 comments14 min readLW link
(www.owlposting.com)

Se­cret Col­lu­sion: Will We Know When to Un­plug AI?

Sep 16, 2024, 4:07 PM
61 points
8 comments31 min readLW link

GPT-o1

ZviSep 16, 2024, 1:40 PM
86 points
34 comments46 min readLW link
(thezvi.wordpress.com)

[Question] Can sub­junc­tive de­pen­dence emerge from a sim­plic­ity prior?

Daniel CSep 16, 2024, 12:39 PM
11 points
0 comments1 min readLW link

Longevity and the Mind

George3d6Sep 16, 2024, 9:43 AM
5 points
2 comments10 min readLW link

[Question] What’s the Deal with Log­i­cal Uncer­tainty?

Ape in the coatSep 16, 2024, 8:11 AM
32 points
29 comments2 min readLW link

Re­in­force­ment Learn­ing from In­for­ma­tion Bazaar Feed­back, and other uses of in­for­ma­tion markets

Abhimanyu Pallavi SudhirSep 16, 2024, 1:04 AM
5 points
1 comment5 min readLW link

Hyperpolation

Gunnar_ZarnckeSep 15, 2024, 9:37 PM
22 points
6 comments1 min readLW link
(arxiv.org)

[Question] If I wanted to spend WAY more on AI, what would I spend it on?

Logan ZoellnerSep 15, 2024, 9:24 PM
53 points
16 comments1 min readLW link

Su­per­in­tel­li­gence Can’t Solve the Prob­lem of De­cid­ing What You’ll Do

Vladimir_NesovSep 15, 2024, 9:03 PM
27 points
11 comments1 min readLW link

For Limited Su­per­in­tel­li­gences, Epistemic Ex­clu­sion is Harder than Ro­bust­ness to Log­i­cal Exploitation

LorecSep 15, 2024, 8:49 PM
3 points
9 comments3 min readLW link

Why I funded PIBBSS

Ryan KiddSep 15, 2024, 7:56 PM
115 points
21 comments3 min readLW link

My dis­agree­ments with “AGI ruin: A List of Lethal­ities”

Noosphere89Sep 15, 2024, 5:22 PM
36 points
46 comments18 min readLW link

Thirty ran­dom thoughts about AI alignment

Lysandre TerrisseSep 15, 2024, 4:24 PM
6 points
1 comment29 min readLW link

Prove­ably Safe Self Driv­ing Cars [Mo­dulo As­sump­tions]

DavidmanheimSep 15, 2024, 1:58 PM
27 points
29 comments8 min readLW link

SCP Foun­da­tion—Anti memetic Divi­sion Hub

landscape_kiwiSep 15, 2024, 1:40 PM
6 points
1 comment1 min readLW link
(scp-wiki.wikidot.com)

Did Christo­pher Hitchens change his mind about wa­ter­board­ing?

Isaac KingSep 15, 2024, 8:28 AM
171 points
22 comments7 min readLW link

Not ev­ery ac­com­mo­da­tion is a Curb Cut Effect: The Handi­capped Park­ing Effect, the Clap­per Effect, and more

Michael CohnSep 15, 2024, 5:27 AM
82 points
39 comments10 min readLW link
(perplexedguide.net)

AlignedCut: Vi­sual Con­cepts Dis­cov­ery on Brain-Guided Univer­sal Fea­ture Space

Bogdan Ionut CirsteaSep 14, 2024, 11:23 PM
17 points
1 comment1 min readLW link
(arxiv.org)

How you can help pass im­por­tant AI leg­is­la­tion with 10 min­utes of effort

ThomasWSep 14, 2024, 10:10 PM
59 points
2 comments2 min readLW link

[Question] Cal­ibra­tion train­ing for ‘per­centile rank­ings’?

david reinsteinSep 14, 2024, 9:51 PM
3 points
0 comments2 min readLW link

OpenAI o1, Llama 4, and AlphaZero of LLMs

Vladimir_NesovSep 14, 2024, 9:27 PM
83 points
25 comments1 min readLW link

For­ever Leaders

Justice HowardSep 14, 2024, 8:55 PM
6 points
9 comments1 min readLW link

Emer­gent Author­ship: Creativity à la Communing

gswonkSep 14, 2024, 7:02 PM
1 point
0 comments3 min readLW link

Com­pres­sion Moves for Prediction

adamShimiSep 14, 2024, 5:51 PM
20 points
0 comments7 min readLW link
(epistemologicalfascinations.substack.com)

Pay-on-re­sults per­sonal growth: first success

Chris LakinSep 14, 2024, 3:39 AM
63 points
8 comments4 min readLW link
(chrislakin.blog)

Avoid­ing the Bog of Mo­ral Hazard for AI

Nathan Helm-BurgerSep 13, 2024, 9:24 PM
19 points
13 comments2 min readLW link

[Question] If I ask an LLM to think step by step, how big are the steps?

ryan_bSep 13, 2024, 8:30 PM
7 points
1 comment1 min readLW link

Es­ti­mat­ing Tail Risk in Neu­ral Networks

Mark XuSep 13, 2024, 8:00 PM
68 points
9 comments23 min readLW link
(www.alignment.org)

If-Then Com­mit­ments for AI Risk Re­duc­tion [by Holden Karnofsky]

habrykaSep 13, 2024, 7:38 PM
28 points
0 comments20 min readLW link
(carnegieendowment.org)

Can star­tups be im­pact­ful in AI safety?

Sep 13, 2024, 7:00 PM
15 points
0 comments6 min readLW link

I just can’t agree with AI safety. Why am I wrong?

Ya PolkovnikSep 13, 2024, 5:48 PM
0 points
5 comments2 min readLW link

Keep­ing it (less than) real: Against ℶ₂ pos­si­ble peo­ple or worlds

quiet_NaNSep 13, 2024, 5:29 PM
17 points
3 comments9 min readLW link

Why I’m bear­ish on mechanis­tic in­ter­pretabil­ity: the shards are not in the network

tailcalledSep 13, 2024, 5:09 PM
22 points
40 comments1 min readLW link

In­creas­ing the Span of the Set of Ideas

Jeffrey HeningerSep 13, 2024, 3:52 PM
6 points
1 comment9 min readLW link

How difficult is AI Align­ment?

Sammy MartinSep 13, 2024, 3:47 PM
44 points
6 comments23 min readLW link