04Understanding language AI

Beyond the chat window

A chatbot is one interface. Models can classify, extract, score, transform, and generate in structured formats.

Morgan KavanaghPublished 2026-03-28

The chat interface is a starting point

Most people encounter language AI through a chat interface: you type a message, the model types back. This is useful, but it represents a tiny fraction of what these models can do. A chat is a conversation, open-ended and flowing, one message at a time. But most professional work is not a conversation. It is a structured task: classify this document, extract these fields, score this text against these criteria, convert this format into that format. Language models can do all of these things, but only if you ask them in the right way.

Structured input, structured output

Instead of asking a model to "tell me about this document", you can ask it to return a JSON object with specific fields: the document's title, its date, the three main topics, and a confidence score for each. The model produces structured data (tables, lists, key-value pairs, XML, JSON) as reliably as it produces prose. A model can sit inside a data pipeline: structured input in, processing according to instructions, structured output out, feeding directly into the next step without human reformatting.

Classification and extraction

Two of the most immediately useful applications of language AI are classification (assigning a category to a piece of text) and extraction (pulling specific information out of unstructured text). Classification: given a customer email, is it a complaint, a question, a compliment, or a request? Given a research abstract, does it belong to methodology, findings, theory, or review? Extraction: given an invoice, what is the amount, date, vendor name, and VAT number? Given a meeting transcript, what were the action items, who was assigned each one, and what were the deadlines? These tasks are routine, high-volume, and well-suited to automation. A model does not need to be creative to do them well; it needs to be consistent and accurate.

Scoring and evaluation

A model can score text against criteria you define. Give it a rubric with five dimensions (clarity, accuracy, completeness, tone, structure) and it returns a score for each, with a justification. The model is not expressing an opinion; it is applying a structured set of criteria that you provided. Scoring quality depends entirely on rubric quality. A vague rubric produces vague scores. A precise rubric, with examples of what constitutes a 1, 3, and 5 on each dimension, produces surprisingly consistent evaluations.

Transformation and conversion

Models can transform text from one format or register to another: convert a bullet-point list into a flowing paragraph, turn a technical specification into a user-facing summary, rewrite formal text as informal or vice versa, convert a table of data into a narrative description, or restructure a document from one template to another. These transformations are mechanical enough that a model handles them reliably, but complex enough that doing them manually is time-consuming. Be specific about the target format, register, and constraints.

Batch processing

The chat window processes one interaction at a time. Through an API or a pipeline tool, you can send hundreds or thousands of requests, each with its own document, its own instructions, and its own expected output format. At this point, language AI stops being a productivity tool for individuals and becomes infrastructure for organisations. One person chatting with a model about one document becomes a system that processes a thousand documents overnight, classifies each one, extracts the relevant fields, and deposits the results in a database. Same model. Different interface.

Examples

From conversation to pipeline

You have been manually chatting with a model to summarise meeting notes, pasting each transcript into a chat window, reading the summary, and copying it into a shared document. Instead, you set up a simple pipeline: transcripts are dropped into a folder, each one is sent to the model with a fixed instruction ("extract action items, decisions, and open questions as a JSON object"), and the structured output is written directly into your project management tool. Same model, no chat window, ten times the throughput.