Marius Hobbhahn

Karma: 392

I’m currently doing a Ph.D. in ML at the International Max-Planck research school in Tübingen. My focus is on Bayesian ML and I’m exploring its role in AI alignment but I’m also exploring non-Bayesian approaches. I want to become an AI safety researcher/​engineer. If you think I should work for you, please reach out.

For more see https://​​​​aboutme/​​

What suc­cess looks like

28 Jun 2022 14:38 UTC
19 points
4 comments1 min readLW link

An­nounc­ing Epoch: A re­search or­ga­ni­za­tion in­ves­ti­gat­ing the road to Trans­for­ma­tive AI

27 Jun 2022 13:55 UTC
92 points
2 comments2 min readLW link

Reflec­tion Mechanisms as an Align­ment tar­get: A survey

22 Jun 2022 15:05 UTC
28 points
1 comment14 min readLW link

Our men­tal build­ing blocks are more differ­ent than I thought

Marius Hobbhahn15 Jun 2022 11:07 UTC
42 points
11 comments14 min readLW link

In­ves­ti­gat­ing causal un­der­stand­ing in LLMs

14 Jun 2022 13:57 UTC
24 points
2 comments13 min readLW link

Elic­it­ing La­tent Knowl­edge (ELK) - Distil­la­tion/​Summary

Marius Hobbhahn8 Jun 2022 13:18 UTC
41 points
2 comments21 min readLW link

The limits of AI safety via debate

Marius Hobbhahn10 May 2022 13:33 UTC
27 points
7 comments10 min readLW link