09From ad hoc to systematic

Is there such a thing as AI language?

Default chatbot output has recognisable patterns. Designed output does not.

Morgan KavanaghPublished 2026-03-28

The chatbot voice

If you open a chat interface, type a simple question, and read the response, you will notice patterns. The output tends to be thorough where brevity would serve better. It hedges with qualifiers ("it is important to note that..."). It favours lists and bullet points. It opens with a restatement of your question. It closes with an offer to help further. These patterns are recognisable, and they have led to the idea that AI-generated text has an inherent style, a detectable fingerprint. But these patterns are not properties of the technology. They are properties of the default configuration: a generic system prompt, no examples, no style constraints, no domain context, and a user who typed a single sentence. The output reflects the poverty of the input.

Defaults are not destiny

Everything you have learned in the preceding modules points to the same conclusion: the output of a language model is entirely shaped by what you put into the context window. Change the system prompt, and the tone changes. Provide examples of your own writing, and the style shifts to match. Specify a register, a sentence length, a vocabulary level, and the model follows. Add domain-specific documents, and the content becomes grounded in your material rather than in the model's training data. The "AI voice" is the voice of a model that has been given no direction. The moment you provide direction, the voice becomes yours.

Malleability is the core property

A language model is not a fixed author with a fixed style. It is a generation engine that adapts to whatever instructions and context it receives. This malleability is not a weakness; it is the fundamental design property. You can produce formal academic prose, casual internal notes, terse technical documentation, or lyrical marketing copy from the same model, in the same session, by changing the prompt. If you provide a style guide, the model follows it. If you provide ten examples of your previous writing, the model approximates your voice. The output is as distinctive or as generic as you make it.

What detection tools actually measure

AI detection tools (GPTZero, Turnitin's AI detector, and similar) work by measuring statistical properties of text: how predictable each word is given the preceding words (perplexity) and how much variation there is in sentence structure (burstiness). Default chatbot output tends to be low-perplexity and low-burstiness because the model, left to its defaults, produces the most probable text at every step. Detection tools catch this statistical signature. But the signature belongs to the default, not to the technology. When you adjust temperature, specify style, edit the output, or design an environment with rules and examples, the statistical properties shift. The detection tools lose their signal. Several institutions have reduced or abandoned enforcement based on these tools because the false positive rate, particularly for non-native speakers whose writing happens to be statistically regular, is too high to be actionable.

Watermarking and its limits

Some providers have explored embedding invisible statistical watermarks in token selection, biasing the choice of words in a pattern that can later be detected. In principle, this would allow verification even after the text has been edited. In practice, watermarks are fragile. Paraphrasing the text, running it through a second model, or making substantive edits removes the watermark. In any environment where you review and revise the output, which is every professional environment described in this curriculum, watermarking does not survive the production process.

The real question is authorship, not detection

The question "was this written by AI?" assumes a binary: either a human wrote it or a machine did. In a designed environment, that binary does not hold. You chose the model. You wrote the system prompt. You structured the data. You provided the examples. You reviewed the output, made corrections, decided it met your standard. The text passed through a model, but it also passed through your judgment, your expertise, and your quality criteria. No different in kind from using any other tool in a production process. The meaningful question is not "did AI touch this?" but "does the person publishing it stand behind it?" Authorship is responsibility, not a claim about which keys were pressed.

Examples

Same model, three registers

You use the same model to produce three texts about the same product: a formal press release, a casual social media post, and a dry technical specification. A reader shown all three would not identify them as coming from the same source. The model has no inherent register; it adopts whatever register the prompt establishes. The "AI sound" exists only when no register is specified.

Detection tool on human text

You run a well-structured academic essay written entirely by a human through a detection tool. The tool flags it as 68% likely AI-generated because the writing is clear, well-organised, and uses conventional academic phrasing. The tool is not detecting AI; it is detecting the statistical regularity that good formal writing shares with default model output. This is why detection tools cannot be used as evidence.