Monday, 13 April 2026

Artificial Legibility — 4 Agency as a Derived Effect

A response is produced.

It is often described in familiar terms:

  • the model “decided” to answer in a certain way

  • the model “chose” one response over another

  • the model “preferred” a particular framing

These descriptions feel natural.

They provide a way of stabilising what appears.

But they do not describe the generative process.


In selection-based systems, there is no operation that corresponds to deciding.

There is no moment at which alternatives are evaluated by a subject and one is selected on the basis of preference, intention, or judgement.


What occurs instead is:

the resolution of constraints over a space of possible continuations


At each step, multiple continuations are possible.

These possibilities are not presented to an agent.

They are defined implicitly by the structure of the model and the constraints imposed by prior tokens.


The system does not “consider” these possibilities.

It does not “weigh” them.

It does not “choose” among them in any agentive sense.


A continuation is selected.

But this selection is not an act.

It is an outcome of constraint interaction.


This distinction is easy to lose because the resulting output often appears as if it were the product of deliberation.

Sentences unfold with apparent direction.

Arguments develop.

Alternatives are contrasted.

Conclusions are reached.


From the outside, this resembles agency.


But resemblance is not equivalence.

The appearance of directed behaviour does not require the presence of an agent directing it.


This is where attribution enters again.

Interpretation encounters structured continuation and stabilises it as the product of an agent.


This stabilisation follows a familiar pattern:

  • coherence is observed

  • coherence is taken as evidence of intention

  • intention is attributed to a source

  • that source is treated as an agent


At no point in this sequence is agency required for the output to be produced.

It is introduced as a way of organising what is encountered.


Agency, in this sense, is not a primitive feature of the system.

It is a derived effect of interpretation.


This does not mean that agency is illusory.

It means that agency is not located where it appears to be.


The system generates outputs that are consistent with constraints.

Interpretation organises those outputs into patterns that can be read as purposeful.


The stability of this reading depends on the coherence of the output.

Where coherence is high, attribution of agency becomes more compelling.

Where coherence breaks down, the attribution weakens.


This can be seen in cases where outputs become inconsistent or contradictory.

The language of agency shifts:

  • instead of “the model decided,” one hears “the model made a mistake”

  • or “the model got confused”


Even here, agency is retained.

But it is modified to account for instability.


This reveals something important.

Agency is not inferred from the presence of an internal decision-making process.

It is stabilised as long as the output can support a coherent interpretation of behaviour.


Once coherence fails beyond a certain threshold, the attribution of agency begins to dissolve.


This suggests that agency, in this context, is best understood as:

a stabilised interpretation of constraint-consistent behaviour under conditions of sufficient coherence


This definition removes the need to locate agency within the system.

It places agency at the level of interpretation.


It also explains why agency appears so readily.

Human interpretive systems are highly sensitive to patterns that can be organised as intentional behaviour.

Where such patterns are available, attribution occurs.


Artificial systems provide a dense and continuous source of such patterns.

They generate extended sequences of constraint-consistent output that support stable interpretation.


The result is not occasional attribution.

It is sustained attribution.


And because this attribution aligns with familiar linguistic forms—questions, answers, arguments, explanations—it becomes difficult to separate from the output itself.


But the separation remains necessary.

Because without it, descriptions of system behaviour become entangled with interpretive projections.


To say that a model “decides” is to import agency into the generative process.

To describe what occurs more precisely is to say:

a continuation is selected under constraint in a way that produces behaviour interpretable as directed


The direction is real at the level of interpretation.

It is not required at the level of generation.


This distinction matters because it prevents a category error.

It avoids treating the appearance of agency as evidence of an underlying agent.


And it allows a clearer account of what artificial systems are doing.

They are not agents that act.

They are systems that produce sequences which can be stabilised as if they were the actions of an agent.


Agency, then, is not eliminated.

It is relocated.


It belongs to the way behaviour is interpreted when constraint-consistent continuation is sufficiently stable to support it.


And once this relocation is made, the language used to describe artificial systems can be adjusted accordingly.

Not by eliminating terms like “decision” or “choice,”

but by recognising that these terms describe how outputs are stabilised in interpretation, not how they are generated.


This preserves the usefulness of such language while preventing it from being mistaken for a description of underlying operations.


What remains is a more precise account:

behaviour appears directed when constraint-consistent continuation supports stable interpretation,

and agency is the name given to that stability.


Not a cause.

An effect.

No comments:

Post a Comment