RSS

Roman Malov

Karma: 213

Bachelor in general and applied physics. AI safety/​Agent foundations researcher wannabe.

I love talking to people, and if you are an alignment researcher we will have at least one common topic (but I am very interested in talking about unknown to me topics too!), so I encourage you to book a call with me: https://​​calendly.com/​​roman-malov27/​​new-meeting

Email: roman.malov27@gmail.com
GitHub: https://​​github.com/​​RomanMalov
TG channels (in Russian): https://​​t.me/​​healwithcomedy, https://​​t.me/​​ai_safety_digest

Two Types of (Hu­man) Uncertainty

Roman Malov12 Aug 2025 1:36 UTC
9 points
2 comments2 min readLW link

The Ob­server Effect for be­lief measurement

Roman Malov2 Aug 2025 13:57 UTC
8 points
4 comments2 min readLW link

An Anal­ogy for Interpretability

Roman Malov24 Jun 2025 14:56 UTC
12 points
2 comments2 min readLW link

Ques­tion to LW devs: does LessWrong tries to be face­booky?

Roman Malov3 Jun 2025 22:08 UTC
5 points
1 comment1 min readLW link

[Question] Could we go an­other route with com­put­ers?

Roman Malov30 May 2025 19:04 UTC
12 points
4 comments1 min readLW link

Neu­ron Ac­ti­va­tions to CLIP Embed­dings: Geom­e­try of Lin­ear Com­bi­na­tions in La­tent Space

Roman Malov3 Feb 2025 10:30 UTC
5 points
0 comments2 min readLW link

[Question] Is “hid­den com­plex­ity of wishes prob­lem” solved?

Roman Malov5 Jan 2025 22:59 UTC
10 points
4 comments1 min readLW link

Ro­man Malov’s Shortform

Roman Malov19 Dec 2024 21:14 UTC
3 points
36 comments1 min readLW link

Vi­sual demon­stra­tion of Op­ti­mizer’s curse

Roman Malov30 Nov 2024 19:34 UTC
25 points
3 comments7 min readLW link