LLM-generated text is not testimony
Synopsis
When we share words with each other, we don’t only care about the words themselves. We care also—even primarily—about the mental elements of the human mind/agency that produced the words. What we want to engage with is those mental elements.
As of 2025, LLM text does not have those elements behind it.
Therefore LLM text categorically does not serve the role for communication that is served by real text.
Therefore the norm should be that you don’t share LLM text as if someone wrote it. And, it is inadvisable to read LLM text that someone else shares as though someone wrote it.
Introduction
One might think that text screens off thought. Suppose two people follow different thought processes, but then they produce and publish identical texts. Then you read those texts. How could it possibly matter what the thought processes were? All you interact with is the text, so logically, if the two texts are the same then their effects on you are the same.
But, a bit similarly to how high-level actions don’t screen off intent, text does not screen off thought. How you want to interpret and react to text, and how you want to interact with the person who published that text, depend on the process that produced the text. Indeed, “[...] it could be almost anything, depending on what chain of cause and effect lay behind my utterance of those words”.
This is not only a purely propositional epistemic matter. There is also the issue of testimony, narrowly: When you public assert a proposition, I want you to stake some reputation on that assertion, so that the public can track your reliability on various dimensions. And, beyond narrow testimony, there is a general sort of testimony—a general revealing of the “jewels” of your mental state, as it were, vulnerable and fertile; a “third-party standpoint” that opens up group thought. I want to know your belief-and-action generators. I want to ask followup questions and see your statements evolve over time as the result of actual thinking.
The rest of this essay will elaborate this point by listing several examples/subcases/illustrations. But the single main point I want to communicate, “on one foot”, is this: We care centrally about the thought process behind words—the mental states of the mind and agency that produced the words. If you publish LLM-generated text as though it were written by someone, then you’re making me interact with nothing.
(This is an expanded version of this comment.)
Elaborations
Communication is for hearing from minds
-
LLM text is structurally, temporally, and socially flat, unlike human text.
Structurally: there aren’t live mental elements underlying the LLM text. So the specific thoughts in the specific text aren’t revealing their underlying useful mental elements by the ways those elements refract through the specific thought.
Temporally: there’s no mind that is carrying out investigations.
It won’t correct itself, run experiments, mull over confusions and contradictions, gain new relevant information, slowly do algorithmically-rich search for relevant ideas, and so on. You can’t watch the thought that was expressed in the text as it evolves over several texts, and you won’t hear back about the thought as it progresses.
The specific tensions within the thought are not communicating back local-contextual demands from the specific thought back to the concepts that expressed the more-global contextual world that was in the backgroundwork of the specific thought.
Socially: You can’t interrogate the thought, you can’t enforce norms on the thinker, and there is no thinker who is sensitive to emergent group epistemic effects of its translations from thought to words. There is no thinker who has integrity, and there is no thinker with which to co-construct new suitable concepts and shared intentions/visions.
-
This could have been
an emaila prompt.Why LLM it up? Just give me the prompt.
When you publish something, I want you to be asserting “this is on some reasonable frontier of what I could write given the effort it would take and the importance of the topic, indicating what I believe to be true and good given the presumed shared context”. It’s not plausible that LLM text meets that definition.
If the LLM text contains surprising stuff, and you didn’t thoroughly investigate for yourself, then you don’t know it’s correct to a sufficient degree that you should post it. Just stop.
If the LLM text contains surprising stuff, and you DID thoroughly investigate for yourself, then you obviously can write something much better and more interesting. Just stream-of-consciousness the most interesting stuff you learned / the most interesting ideas you have after investigating. I promise it will be more fun for everyone involved.
If the LLM text does not contain surprising stuff, why do you think you should post it?
Communication is for hearing assertions
-
We have to listen to each other’s utterances as assertions.
We have to defer to each other about many questions, which has pluses and minuses.
Most statements we hear from each other are somewhere between kinda difficult and very difficult for us to verify independently for ourselves. This includes for example expert opinions, expert familiarity with obscure observations and third-hand testimony, personal stories, and personal introspection.
It’s valuable to get information from each other. But also that means we’re vulnerable to other people deciding to lie, distort, deceive, mislead, filter evidence, frame, Russell conjugate, misemphasize, etc.
When someone utters a propositional sentence, ze is not just making an utterance; ze is making an assertion. This involves a complex mental context for what “making a propositional assertion” even is—it involves the whole machinery of having words, concepts, propositions, predictive and manipulative bindings between concepts and sense organs and actuators and higher-order regularities, the general context of an agent trying to cope with the world and therefore struggling to have mental elements that help with coping, and so on.
When ze asserts X, ze is saying “The terms that I’ve used in X mean roughly what you think they mean, as you’ve been using those terms; and if you try (maybe by asking me followup questions), then you can refine your understanding of those terms enough to grasp what I’m saying when I say X; X is relevant in our current shared context, e.g. helpful for some task we’re trying to do or interesting on general grounds of curiosity or it’s something you expressed wanting to know; X is roughly representative of my true views on the things X talks about; I believe X for good reason, which is to say that my belief in X comes from a process which one would reasonably expect to generally produce good and true statements, e.g. through updating on evidence and resolving contradictions, and this process will continue in the future if you want to interact with my assertion of X; my saying X is in accordance with a suitable group-epistemic stance; …”.
In short, “this is a good thing for me to say right now”.
Which generally but not always implies that you believe it is true,
generally but not always implies that you believe it is useful,
generally but not always implies you believe that I will be able to process the assertion of X in a beneficial way,
and so on.
-
Because we have to listen to each other’s utterances as assertions, it is demanded of us that when we make utterances for others to listen to, we have to make those utterances be assertions.
-
If you wouldn’t slash someone’s tires, you shouldn’t tell them false things.
-
If you wouldn’t buy crypto on hype cycles, then you shouldn’t share viral news. I learned this the hard way:
I saw a random news article sharing the exciting, fascinating news that the Voynich manuscript has been decoded! Then my more sober and/or informed friend was bafflingly uninterested. Thus I learned that not only had the Voynich manuscript been decoded just that week, but also it had been decoded a month before, and two months before, and a dozen other times.
Several times, people shared news like “AI just did X!” and it’s basically always either BS, or mostly BS and kinda interesting but doesn’t imply what the sharer said.
I shared the recent report about lead in food supplements without checking for context (the context being that the lead levels are actually fine, despite the scary red graphs).
-
In the introduction, I used the example of two identical texts. But in real life the texts aren’t even identical.
The choice of words, phrases, sentence structure, argument structure, connecting signposts, emphasis—all these things reveal how you’re thinking of things, and transmit subtleties of the power of your mental gears. The high level pseudo-equivalence of “an LLM can’t tell the difference” does not screen off the underlying world models and values! The actual words in LLM text are bad—e.g. frequent use of vague words which, like a stable-diffusion image, kinda make sense if your eyes are glazed over but are meaningless slop if you think about them more sharply.
maybe you think that’s a small difference. i think you’re wrong, but also consider this… if it is small, then the total effect is small times 100 or 1000. i sometimes used to write in public without capitalizing words in the standard sentence-initial way. my reasoning was that if i could save a tiny bit on the cognitive load of chording with the shift key, then i could have the important thoughts more quickly and thoroughly and successfully, and that was more important than some very slight difference in reading experience. i still usually write like that in private communications, but generally in public i use capitalization. it makes it a bit easier to parse visually, e.g. to find the beginning of sentences, or know when you’ve reached the end of a sentence rather than seeing etc. and not knowing if a new sentence just started. that difference makes a difference if the text is read by 100 or 1000 people. are you seriously going to say that all the word choice and other little choices matter less than Doing This Shit? all text worth reading is bespoke, artisanal, one-shot, free-range, natural-grown, slow-dried, painstaking, occult, unpredictable, kaleidoscopic, steganographic—human. we should be exercising our linguistic skills.
Writing makes you think of more stuff. You get practice thinking the thought more clearly and easily, and rearranging it so that others understand it accurately. At least my overwhelming experience is that writing always causes a bunch of new thoughts. Generating a video that depicts your AI lookalike exercising is not the same as you actually exercising, lol. By putting forth a topic in public but not even doing this exercise regarding that topic is a kind of misdirection and decadent laziness, as if the public is supposed to go fill in the blanks of your farted-out notions. Verification is far from production, and you weren’t even verifying.
You can’t make a text present propositions that are more true or good just by thinking about the text more and then keeping it the same no matter what. However, you can make a text more true or good just by thinking about it, if you would change the text if you thought of changes you should make. In practice if you do this, then you will change your LLM text a lot, because LLM text sucks. The more you change it, the less my objection applies, quantitatively.
-
If you’re asking a human about some even mildly specialized topic, like history of Spain in the 17th century or different crop rotation methods or ordinary differential equations, and there’s no special reason that they really want to appear like they know what they’re talking about, they’ll generally just say “IDK”. LLMs are much less like that. This is a big difference in practice, at least in the domains I’ve tried (reproductive biology). LLMs routinely give misleading / false / out-of-date / vague-but-deceptively-satiating summaries.
-
Assertions live in dialogue
-
In order to make our utterances be assertions, we have to open them up to inquiry.
LLM text is not open to inquiry.
When you’re making an assertion, we need you to be staking some of your reputation on the assertion.
(“We” isn’t a unified consensus group; but rather a set of other individuals, and some quasi-coherent subsets.)
We’re going to track whether your assertions are good and true. We might track separately for different domains and different modalities (e.g. if you prefaced by “this is just a guess but”, or if you’re in a silly jokey mood, and so on). We will credit you when you’ve been correct and when we are pressed for time (which is always). We will discount your testimony when you’ve been incorrect or poisonous. We will track this personally for you.
If that sounds arcane, consider that you do it all the time. There are people you’d trust about math, other people you’d trust about wisely dealing with emotions, other people you’d trust about taking good pictures, and so on.
You can’t go back and say “oh I didn’t mean for you to take this seriously”, if that wasn’t reasonably understood within the context. You can say “oops I made a mistake” or “yeah I happened to give a likelihood delta from the consensus probabilities that wasn’t in the direction of what ended up being the case”.
But you are lying in your accounting books if you try to discount the seriousness of your assertions when you’re proven incorrect. E.g. if you try to discount the seriousness by saying “oops I was just poasting LLM slop haha”. It’s not a serious way of communicating. It’s like searching for academic papers by skimming the abstracts until you find an abstract that glosses the paper’s claims in a vague way that’s sorta consistent with what you want to assert, and then citing that paper. It’s contentless, except for the anti-content of trying to appear contentful.
We might want to cross-examine you, like in a courtroom. We want to clarify parts that are unclear. We want to test the coherence of the world-perspective you’re representing. We want to coordinate your testimony with the testimony of others, and/or find contradictions with the testimony of others.
We want to trace back chains of multiple steps of inference back to the root testimony.
If David judges that Alice should be ostracized, on account of Carol saying that Alice is a liar and on account of Bob saying that Alice is a cheat, but Carol and Bob are each separately relying on testimony from Eve about Alice, then this is a fact we would like to track.
We want to notice contradictions between different testimonies and then bring the original sources into contact with each other. Then they can debate; or clarify terms and reconcile; or share information and ideas and update; or be disproven; or reveal a true confusion / mystery / paradox; or be revealed as a liar. Even if one assertion in isolation can’t be decided, we want to notice when one person contradicts others in several contexts (which may be the result of especially good behavior or especially bad behavior, depending).
We want to avoid miasmatic pollution, i.e. unsourced claims in the water.
Unsourced claims masquerade as consensus, and nudge the practical consensus, thus destroying the binding between the practical consensus and the epistemic consensus. Instead, we want claims made by people saying “yes I’ve seen this, I’m informing you by making this utterance where you’ll take it as an assertion”.
We don’t want people repeating mere summaries of consensus. This leads to a muffling of the gradient of understanding and usefulness. Think of an overcooked scientific review paper that cites everything and compresses nothing. It’s all true and it’s all useless. Also it isn’t even all true, because if you aren’t thinking then you aren’t noticing other people’s false testimony that you’re repeating.
LLM text is made of unsourced quasi-consensus glosses.
The accused has some sort of moral right to face zer accuser, and to protection from hearsay, and to have accusers testify under penalty of perjury.
-
In order to make our utterances be useful assertions that participate in ongoing world co-creation with our listeners, we have to open up the generators of the assertions.
LLM text tends to be less surprising—more correlated with what’s already out there, by construction.
This is the case in every way. That hides and muffles the shining-through of the human “author”’s internal mental state.
We want the uncorrelatedness; the socially-local pockets of theoryweaving (groundswell babble—gathering information, hypothesizing ideas and propositions) and theorycrafting (“given enough eyeballs, all bugs are shallow”—testing predictions, resolving contradictions, retuning theories, selecting between theories).
If you speak in LLM, we cannot see what you are thinking, how you are thinking it, how you came to think that way, what you’re wanting, what possibilities we might be able to join with you about, and what procedures we could follow to usefully interoperate with you.
I want you to generate the text you publish under your own power, by cranking the concepts as you have them in your head, so I can see those gears working, including their missing teeth and rough edges and the grit between them—and also the mechanical advantage that your specific arrangement has for the specific task you have been fiddling on for the past month or decade, because you have apply your full actual human general intelligence to some idiosyncratic problem and have created a kinda novel arrangement of kinda novel-shaped gears.
I want to be able to ask you followup questions. I want to be able to ask you for examples, for definitions, for clarifications; I want to ask you for other possibilities you considered and discarded; I want to ask you what you’re still confused/unsure about, what you’re going to think about, what you’re least and most confident in, where you think there’s room for productive information gathering.
Sometimes when people see something interesting and true, they struggle to express it clearly. I still want that text! Text that is literally incorrect but is the result of a human mind struggling to express something interesting / useful / novel / true, is still very useful, because I might be able to figure out what you meant, in combination with further information and thinking. LLM text throws all that stuff out.
I want to figure out some of your goals / visions are, so we can find shared intentions. This process is difficult and works through oblique anastomosis, not by making an explicit point that you typed into a prompt for the LLM to ensloppenate.
Stop trying to trick me into thinking you know what the fuck you’re talking about.
Non-testimony doesn’t have to be responded to, lest it be trolling—cheaply produced discourse-like utterances without a model behind them, aimed at jamming your attentional pathways and humiliating you by having you run around chasing ghosts.
-
A sentence written by an LLM is said by no one, to no one, for no reason, with no agentic mental state behind it, with no assertor to participate in the ongoing world co-creation that assertions are usually supposed to be part of.
Even if we assume or it becomes true that LLMs are genuinely minds and all that, it still seems similarly bad to use them like this. LLM-generated text is not your testimony, regardless of whether or not it is ‘testimony’.
I like the analogy of a LARP. Characters in a book don’t have reputation or human-like brain states that they honestly try to represent—but a good book can contain interesting, believable characters with consistent motivation, etc. I once participated in a well-organized fantasy LARP in graduate school. I was bad at it but it was a pretty interesting experience. In particular people who are good are able to act in character and express thoughts that “the character would be having” which are not identical to the logic and outlook of the player (I was bad at this, but other players could do it I think). In my case, I noticed that the character imports a bit of your values, which you sometimes break in-game if it feels appropriate. You also use your cognition to further the character’s cognition, while rationalizing their thinking in-game. It obviously feels different from real life: it’s explicitly a setting where you are allowed and encouraged to break your principles (like you are allowed to lie in a game of werewolf, etc.) and you understand that this is low-stakes, and so don’t engage the full mechanism of “trying as hard as possible” (to be a good person, to achieve good worlds, etc.). But also, there’s a sense in which a LARP seems “Turing-complete” for lack of a better word. For example in this LARP, the magical characters (not mine) collaboratively solved a logic puzzle to reverse engineer a partially known magic system and became able to cast powerful spells. I could also imagine modeling arbitrarily complex interactions and relationships in an extended LARP. There would probably always be some processing cost to add the extra modeling steps, but I can’t see how this would impose any hard constraints on some measure of “what is achievable” in such a setting.
I don’t see hard reasons for why e.g. a village of advanced LLMs could not have equal or greater capability than a group of smart humans playing a LARP. I’m not saying I see evidence they do—I just don’t know of convincing systematic obstructions. I agree that modern LLMs seem to not be able to do some things humans could do even in a LARP (some kind of theory of mind, explaining a consistent thinking trace that makes sense to a person upon reflection, etc.) but again a priori this might just be a skill issue.
So I wonder in the factorization “LLM can potentially get as good as humans in a LARP” + “sufficiently many smart humans in a long enough LARP are ‘Turing complete up to constant factors’ ” (in the sense of in principle being able to achieve, without breaking character, any intellectual outcome that non-LARP humans could do), which part would you disagree with?
“Potentially get as good as humans” I of course think in general, as I think we’re by default all dead within 100 years to AGI. If you mean actual current LLMs, I’m pretty sure no they cannot. See https://www.lesswrong.com/posts/sTDfraZab47KiRMmT/views-on-when-agi-comes-and-on-strategy-to-reduce and https://www.lesswrong.com/posts/5tqFT3bcTekvico4d/do-confident-short-timelines-make-sense
I would point you for example to low-sample-complexity learning that humans sometimes do, and claim that LLMs don’t do this in the relevant sense and that this is necessary for getting good. See also this thread: https://www.lesswrong.com/posts/sTDfraZab47KiRMmT/views-on-when-agi-comes-and-on-strategy-to-reduce?commentId=dqbLkADbJQJi6bFtN
I see another obstruction in attention span. I strongly suspect that whenever an LLM is tasked with writing the next token, attention mechanisms compress all potentially relevant information into less than a hundred thousand numbers, preventing the model from taking many nuances into account when writing the token. A human brain, on the other hand, takes into account billions of bits of information stored in neuron activations.
@TsviBT, we had a warning shot of LLMs becoming useful in research or writing a coherent short story in @Tomás B.’s experiment (the post on LW describing it was removed for an unknown reason).
I moved it into my drafts. I published it again for you. I figured it was unlikely to be referenced again and I tend to take stuff down I don’t want people reading as one of the first things on my author’s page.
(FWIW, I’ve referenced that post 2-4 times since it was posted)
The claim that the thought process behind words—the mental states of the mind and agency that produced the words … does not exist seems phenomelogically contradicted by just interacting with LLMs. I expect your counteragrument be to appeal to some idiosyncratic meanings of words like thoughts or mind states, and my response being something in the direction ‘planes do fly’.
Why LLM it up? Just give me the prompt. One reason why not to is your mind is often broadly unable to trace the thoughts of an LLM, and if the specific human-AI interaction leading to some output has nontrivial context & lenght, you would also be unable to get an LLM to replicate the trace without the context shared.
I think you’re just thinking of some of the thought processes, but not about the ones that are most important in the context of writing and reading and public communication. I have no big reason to suspect that LLMs don’t have “mental states” that humans do in order to proximally perform e.g. grammar, sufficient distinctions between synonyms, facts like “Baghdad is the capital of Iraq”, and a huge variety of algorithms. I do think they lack the distal mental states, which matter much more. Feel free to call this “idiosyncratic” but I sure hope you don’t call anything that’s insufficiently concrete / specific / explicit / well-understood “idiosyncratic”, because that means something totally different, and lots of centrally important stuff is insufficiently concrete / specific / explicit / well-understood.
Out of interest, do you have a good argument for this? If so, I’d be really interested to hear it.
Naively, I’d think your example of “Baghdad is the capital of Iraq” encodes enough of the content of ‘Baghdad’ and ‘Iraq’ e.g. other facts about the history of Iraq, the architecture in the city etc.. to meaningfully point towards the distal Baghdad and Iraq. Do you have a different view?
I mean the deeper mental algorithms that generate the concepts in the first place, which are especially needed to do e.g. novel science. See https://www.lesswrong.com/posts/sTDfraZab47KiRMmT/views-on-when-agi-comes-and-on-strategy-to-reduce
and
https://www.lesswrong.com/posts/5tqFT3bcTekvico4d/do-confident-short-timelines-make-sense
See also this thread: https://www.lesswrong.com/posts/sTDfraZab47KiRMmT/views-on-when-agi-comes-and-on-strategy-to-reduce?commentId=dqbLkADbJQJi6bFtN
You just had to prompt an LLM like Claude, Grok or GPT-5-thinking with a complex enough task, like one task in the Science Bench. GPT-5-thinking lays out the stuff it did, including coding and correcting itself. As for gaining new information, one could also ask the model to do something related to a niche topic and watch it look up relevant information. The ONLY thing which GPT-5 didn’t do was to learn anything from the exchange, since nobody bothered to change the neural network’s weights to account for the new experience.
Please stop mixing the plausible assumption that LLM-generated text is likely a mix-and-match of arguments already said by others and the less plausible assumption that an LLM doesn’t have a mind. However, the plausible assumption has begun to tremble since we had a curated post whose author admitted to generating it by using Claude Opus 4.1 and substantially editing the output.
In the discussion of the buck post and elewhere, I’ve seen the idea floated that if no-one can tell that a post is LLM generated, then it is necessarily ok that it is LLM generated. I don’t think that this necessarily follows- nor does its opposite. Unfortunately I don’t have the horsepower right now to explain why in simple logical reasoning, and will have to resort to the cudgel of dramatic thought experiment.
Consider two lesswrong posts: a 2000 digit number that is easily verifiable as a collatz counterexample, and a collection of first person narratives of how human rights abuses happened, gathered by interviewing vietnam war vets at nursing homes. The value of one post doesn’t collapse if it turns out to be LLM output, the other collapses utterly- and this is unconnected from whether you can tell that they LLM output.
The buck post is of course not at either end of this spectrum, but it contains many first person attestations- a large number of relatively innocent “I thinks,” but also lines like “When I was a teenager, I spent a bunch of time unsupervised online, and it was basically great for me.” and “A lot of people I know seem to be much more optimistic than me. Their basic argument is that this kind of insular enclave is not what people would choose under reflective equilibrium.” that are much closer to the vietnam vet end of the spectrum.
EDIT: Buck actually posted the original draft of the post, before LLM input, and the two first person accounts I highlighted are present verbatim, and thus honest. Reading the draft, it becomes a quite thorny question to adjucate whether the final post qualifies as “generated” by Opus, but this will start getting into definitions.
TBF “being a curated post on LW” doesn’t exclude anything from being also a mix and match of arguments already said by others. One of the most common criticisms of LW I’ve seen is that it’s a community reinventing a lot of already said philosophical wheels (which personally I don’t think is a great dunk; exploring and reinventing things for yourself is often the best way to engage with them at a deep level).
[retracted due to communication difficulty. self-downvoted.]
The intro sounded promising, but almost immediately you’re massively overclaiming. maybe someone will pick it apart, but you’re doing a lot of it, and I don’t feel like it right now. Many sentences are similar to ones that are true, but taken literally they imply or directly state things that are somewhere between not-established to clearly-false. eg, as StanislavKrym mentions: “Temporally: …that is carrying out investigations.”—this is just obviously not true in some cases. I do agree that there’s some form of claim like your title that is plausible. Anthropic’s recent paper seems to imply it’s not consistently not testimony. I could buy that it’s not systematically able to be testimony when trained on generative modeling of something else. Many of your subclaims are reasonable; please fix the ones that aren’t.
(Note, posted by private request: The OP is primarily motivated by questions about human discourse—what it’s for, how it works, how to care about it more. LLMs are the easiest contrast / foil.)
Obviously I disagree, but mainly I want to say, if you’re trying to actually communicate with me personally, you’ll have to put more effort into thinking about what I’m saying, e.g. seeing if you can think of a reasonable interpretation of “investigate” that LLMs probably / plausibly don’t do.
What is YOUR interpretation of investigation that LLMs don’t do? Whatever it is, you are either mistaken or it doesn’t treat this stuff (which I obtained by prompting GPT to do a task from the Science Bench) as an investigation.
E.g. the sort of investigation that ends in the creation of novel interesting concepts: https://www.lesswrong.com/posts/sTDfraZab47KiRMmT/views-on-when-agi-comes-and-on-strategy-to-reduce?commentId=dqbLkADbJQJi6bFtN
[edit: retracted due to communication difficulty]
I agree they rarely do that and are not driven to it. I knew you meant that, but I’m not playing along with your use of the word because it seems to me to be an obvious rules-lawyering of what investigation is, in what is in my opinion confusingly called motte-and-bailey; if you were willing to try to use words the same way as everyone else you could much more easily point to the concept involved here. For example, I would have straightforwardly agreed if you had simply said “they do not consistently seek to investigate, especially not towards verbalizing or discovering new concepts”. But the overclaim is “they do not investigate”. They obviously do, and this includes all interpretations I see for your word use—they do sometimes seek out new concepts, in brief flashes when pushed to do so fairly hard—and if you believe they do not, it’s a bad sign about your understanding; but they also obviously are not driven to it or grown around it in the way a human is, so I don’t disagree with your main point, only with word uses like this.
Just FYI instead of doing this silently, this comment thread is pretty close to making me decide to just ban you from commenting on my posts.
...I will update to be less harsh rather than being banned, then. surprised I was even close to that, apologies. in retrospect, I can see why my frustration would put me near that threshold.
I don’t think I mind harshness, though maybe I’m wrong. E.g. your response to me here https://www.lesswrong.com/posts/zmtqmwetKH4nrxXcE/which-side-of-the-ai-safety-community-are-you-in?commentId=hjvF8kTQeJnjirXo3 seems to me comparably harsh, and I probably disagree a bunch with it, but it seems contentful and helpful, and thus socially positive/cooperative, etc. I think my issue with this thread is that it seems to me you’re aggressively missing the point / not trying to get the point, or something, idk. Or just talking about something really off-topic even if superficially on-topic in a way I don’t want to engage with. IDK.
[ETA: like, maybe I’m “overclaiming”—mainly just be being not maximally precise—if we look at some isolated phrases, but I think there’s a coherent and [ought to be plausible to you] interpretation of those phrases in context that is actually relevant to what I’m discussing in the post; and I think that interpretation is correct, and you could disagree with that and say so; but instead you’re talking about something else.]
[ETA: and like, yeah, it’s harder to describe the ways in which LLMs are not minds than to describes ways in which they do perform as well as or better than human minds. Sometimes important things are hard to describe. I think some allowance should be made for this situation.]
Yeah pretty sure it doesn’t.
I think not passing off LLM text as your own words is common good manners for a number of reasons—including that you are taking responsibility for words you didn’t write and possibly not even read in depth enough, so it’s going to be on you if someone reads too much into them. But it doesn’t really much need any assumptions on LLMs themselves, their theory of mind, etc. Nearly the same would apply about hiring a human ghostwriter to expand on your rough draft, it’s just that that has never been a problem until now because ghostwriters cost a lot more than a few LLM tokens.