RSS

Anurag

Karma: 14

As a tech-product person—over the ages—I have been building systems where reliability, auditability, scale, performance, and correctness are central. I only woke up to AI in mid-2025 - after years of dismissing it as something my daughter used for homework. Since then I have become deeply absorbed in understanding how large language models were created, how they currently work and their emergent capabilities.

I have become particularly interested in AI alignment. My aim is to make a meaningful contribution to the field.

All views expressed and posted works are my own.

Can We Make AI Align­ment Fram­ing Less Wrong?

Anurag 8 Jan 2026 15:20 UTC
3 points
0 comments4 min readLW link

Align­ment Is Not One Prob­lem: A 3D Map of AI Risk

Anurag 28 Dec 2025 8:44 UTC
3 points
0 comments14 min readLW link

The In­tel­li­gence Axis: A Func­tional Ty­pol­ogy

Anurag 25 Dec 2025 12:18 UTC
3 points
0 comments5 min readLW link

A Func­tional Ty­pol­ogy of Cog­ni­tive Ca­pa­bil­ities (In­ter­ac­tive Vi­su­al­iza­tion)

Anurag 18 Dec 2025 14:06 UTC
2 points
0 comments4 min readLW link

An Ap­proach for Eval­u­at­ing Self-Boundary Con­sis­tency in AI Systems

Anurag 10 Dec 2025 13:57 UTC
3 points
0 comments6 min readLW link

A Cat­a­log of AI Evaluations

Anurag 9 Dec 2025 17:05 UTC
2 points
0 comments1 min readLW link

About Nat­u­ral & Syn­thetic Be­ings (In­ter­ac­tive Ty­pol­ogy)

Anurag 7 Dec 2025 16:59 UTC
2 points
2 comments3 min readLW link

Shap­ing Model Cog­ni­tion Through Reflec­tive Dialogue—Ex­per­i­ment & Findings

Anurag 3 Dec 2025 23:50 UTC
2 points
0 comments4 min readLW link

[Question] Is there a tax­on­omy & cat­a­log of AI evals?

Anurag 28 Nov 2025 23:15 UTC
1 point
2 comments1 min readLW link

Po­ten­tial of Reflec­tive-Dialogs for Model Train­ing and Alignment

Anurag 26 Nov 2025 20:28 UTC
1 point
0 comments7 min readLW link