Maybe we want a multi-level categorization scheme instead? Something like:
Level 0: Author completely abstains from LLM use in all contexts (not just this post) Level 1: Author uses LLMs but this particular post was made with no use of LLM whatsoever Level 2: LLM was used (e.g. to look up information), but no text/images in the post came out of LLM Level 3: LLM was used for light editing and/or image generation Level 4: LLM was used for writing substantial parts Level 5: Mostly LLM-generated with high-level human guidance/control/oversight
This is an edge case, but just flagging that it’s a bit unclear to me how to apply this to my own post in a useful way. As I’ve disclosed in the post itself:
OpenAI’s o3 found the idea for the dovetailing procedure. The proof of the efficient algorithmic Kraft coding in the appendix is mine. The entire post is written by myself, except the last paragraph of the following section, which was first drafted by GPT-5.
Does this count as Level 3 or 4? o3 provided a substantial idea, but the resulting proof was entirely written down by myself. I’m also unsure whether the full drafting of precisely one paragraph (which summarizes the rest of the post) by GPT-5 counts as editing or the writing of substantial parts.
We need another “level” here, probably parallel to the others, for when LLMs are used for idea-generation, criticism of outlines, as a discussion partner et cetera. For instance, let’s say I think about countries that are below their potential in some tragic way, like Russia and Iran, countries with loads of cultural capital, educated population, that historically have lots going for them. Then I can ask an LLM “any other countries like that?” and it might mention, say, North Korea, Iraq and Syria, maybe Greece or Turkey or South Italy, with some plausible story attached to them. When I do this interaction with an LLM the end product is going to be colored by it. If I initially intended to talk about how Russia and Iran have been destroyed by some particular forms of authoritarianism, my presentation, hypothesis, or whatever, will likely be modified so I can put Greece and Iraq into the same bucket. This alters my initial thoughts and probably changes my though-generation process into a mold more-or-less shaped by the LLM, “hacking my brain”. When this happen across many posts, it’s likely to make writing homogenized not through writing style, but semantic content.
This example is kinda weak, but I think this is the kind of thing OP is worried about. But I’d be curious to hear stronger examples if anyone can think of them.
Maybe we want a multi-level categorization scheme instead? Something like:
Level 0: Author completely abstains from LLM use in all contexts (not just this post)
Level 1: Author uses LLMs but this particular post was made with no use of LLM whatsoever
Level 2: LLM was used (e.g. to look up information), but no text/images in the post came out of LLM
Level 3: LLM was used for light editing and/or image generation
Level 4: LLM was used for writing substantial parts
Level 5: Mostly LLM-generated with high-level human guidance/control/oversight
any reason not to just start doing that as post tags? no convenient way to do it for comments though.
This is an edge case, but just flagging that it’s a bit unclear to me how to apply this to my own post in a useful way. As I’ve disclosed in the post itself:
Does this count as Level 3 or 4? o3 provided a substantial idea, but the resulting proof was entirely written down by myself. I’m also unsure whether the full drafting of precisely one paragraph (which summarizes the rest of the post) by GPT-5 counts as editing or the writing of substantial parts.
We need another “level” here, probably parallel to the others, for when LLMs are used for idea-generation, criticism of outlines, as a discussion partner et cetera. For instance, let’s say I think about countries that are below their potential in some tragic way, like Russia and Iran, countries with loads of cultural capital, educated population, that historically have lots going for them. Then I can ask an LLM “any other countries like that?” and it might mention, say, North Korea, Iraq and Syria, maybe Greece or Turkey or South Italy, with some plausible story attached to them. When I do this interaction with an LLM the end product is going to be colored by it. If I initially intended to talk about how Russia and Iran have been destroyed by some particular forms of authoritarianism, my presentation, hypothesis, or whatever, will likely be modified so I can put Greece and Iraq into the same bucket. This alters my initial thoughts and probably changes my though-generation process into a mold more-or-less shaped by the LLM, “hacking my brain”. When this happen across many posts, it’s likely to make writing homogenized not through writing style, but semantic content.
This example is kinda weak, but I think this is the kind of thing OP is worried about. But I’d be curious to hear stronger examples if anyone can think of them.