Mathematician, alignment researcher, doctor. Reach out to me on Discord and tell me you found my profile on LW if you’ve got something interesting to say; you have my explicit permission to try to guess my Discord handle if so. You can’t find my old abandoned-for-being-mildly-infohazardously-named LW account but it’s from 2011 and has 280 karma.
A Lorxus Favor is worth (approximately) one labor-day’s worth of above-replacement-value specialty labor, given and received in good faith, and used for a goal approximately orthogonal to one’s desires, and I like LessWrong because people here will understand me if I say as much.
Apart from that, and the fact that I am under no NDAs, including NDAs whose existence I would have to keep secret or lie about, you’ll have to find the rest out yourself.
(Random thought I had and figured this was the right place to set it down:) Given how centally important token-based word embeddings as to the current LLM paradigm, how plausible is it that (put loosely) “doing it all in Chinese” (instead of English) is actually just plain a more powerful/less error-prone/generally better background assumption?
Associated helpful intuition pump: LLM word tokenization is like a logographic writing system, where each word corresponds to a character of the logography. There need be no particular correspondence between the form of the token and the pronunciation/”alphabetical spelling”/other things about the word, though it might have some connection to the meaning of the word—and it often makes just as little sense to be worried about the number of grass radicals in “草莓” as it does to worry about the number of r’s in a “strawberry” token.
(And yes, I am aware that in Mandarin Chinese, there’s lots of multi-character words and expressions!)