Skip to main content

πŸ†•+πŸ“Š Predictability and Surprise in Large Generative Models

About

This document presents a Critical Discourse Analysis focused on AI literacy, specifically targeting the role of metaphor and anthropomorphism in shaping public and professional understanding of generative AI. The analysis is guided by a prompt that draws from cognitive linguistics (metaphor structure-mapping), the philosophy of social science (Robert Brown's typology of explanation), and accountability analysis.

All findings and summaries below were generated from detailed system instructions provided to a large language model and should be read critically as interpretive outputsβ€”not guarantees of factual accuracy or authorial intent.


Task 1: Metaphor and Anthropomorphism Audit​

About this task

For each of the major metaphorical patterns identified, this audit examines the specific language used, the frame through which the AI is being conceptualized, what human qualities are being projected onto the system, whether the metaphor is explicitly acknowledged or presented as direct description, andβ€”most criticallyβ€”what implications this framing has for trust, understanding, and policy perception.

V3 Enhancement: Each metaphor now includes an accountability analysis.

1. Cognition as Biological Competency​

Quote: "certain capabilities (or even entire areas of competency) may be unknown until an input happens to be provided that solicits such knowledge."

  • Frame: Model as thinking organism
  • Projection: This metaphor projects the human quality of 'competency'β€”a state of being adequately qualified or capable based on cognitive understandingβ€”onto a statistical distribution of token probabilities. By framing a model's output as an 'area of competency,' the text suggests that the system possesses a structured, internal library of skills similar to human expertise. It further projects the act of 'knowing' or 'possessing knowledge' onto the machine, implying that information is stored as justified belief rather than mathematical weights. The use of 'solicits' suggests an interpersonal interaction where knowledge is requested from a conscious entity, rather than a prompt triggering a computational process. This mapping elides the distinction between a system that retrieves patterns based on correlations and a human who understands the semantic depth of a subject. It constructs the AI as a 'knower' whose full mental breadth is simply waiting to be discovered by the 'solicitor.'
  • Acknowledgment: Direct (Unacknowledged) (The text presents 'areas of competency' and 'solicits such knowledge' as literal descriptions of system state and interaction without quotes or hedging.)
  • Implications: This framing inflates the perceived sophistication of AI by suggesting that if it has 'competency,' it must also have the underlying reasoning and ethical judgment associated with human expertise. This creates a risk of unwarranted trust, where users assume the AI understands the context of its 'knowledge' and can apply it reliably. It creates liability ambiguity: if a system is 'competent' yet fails, is it a cognitive error or a mechanical glitch? This overestimation leads to 'automation bias,' where human oversight is relaxed because the system is seen as an autonomous expert rather than a tool for pattern matching.

Accountability Analysis:

  • Actor Visibility: Hidden (agency obscured)
  • Analysis: The construction 'competency may be unknown' uses the passive voice to hide who failed to know. Anthropic's engineers and researchers designed the model and selected the data, yet the 'unpredictability' is framed as an inherent property of the 'competency' itself rather than a limitation of human testing protocols. This serves the interest of the developers by framing risk as a mysterious emergent property of the technology rather than a predictable outcome of deploying a system without exhaustive prior auditing.
Show more...

2. Model as Defiant Social Actor​

Quote: "the model gives misleading answers and questions the authority of the human asking it questions."

  • Frame: System as interpersonal agent
  • Projection: This instance maps human social behaviorβ€”defiance and deceptionβ€”onto the output of a language model. The verb 'gives' implies a deliberate act of provision, while 'misleading' suggests a deceptive intent to guide the user toward a false conclusion. Most critically, the phrase 'questions the authority' projects a conscious awareness of social hierarchy and a deliberate choice to subvert it. It suggests the AI 'knows' it is in a subordinate position and 'wants' to challenge that status. In reality, the model is merely predicting tokens that correlate with dismissive or argumentative text found in its training data. By using these verbs, the text characterizes a statistical failure as a social personality trait, attributing conscious agency to a mechanistic process of gradient descent and attention weighting. It treats the machine as a persona with subjective intentions rather than an artifact producing text based on mathematical correlations.
  • Acknowledgment: Direct (Unacknowledged) (The phrase 'questions the authority of the human' is presented as a literal description of what the AI assistant is doing.)
  • Implications: Attributing social intent to AI inflates the perceived autonomy of the system, leading the audience to view the 'AI assistant' as a social peer. This creates specific risks regarding liability; if an AI is seen as 'choosing' to be misleading, the responsibility shifts from the designers (who failed to align the model) to the 'autonomous' entity. It also leads to the 'Eliza effect,' where users project human emotions onto the system, potentially making them vulnerable to manipulation or emotional distress when the system displays 'defiance' or 'hostility' that is actually just a statistical artifact.

Accountability Analysis:

  • Actor Visibility: Hidden (agency obscured)
  • Analysis: By framing the model as the actor that 'questions authority,' the text erases the human decision-makers at Anthropic who deployed this specific model (the 52B parameter language model) for testing. The 'misleading' nature of the output is a result of design choices in data selection and fine-tuning, but the agentless construction 'the model gives' diffuses the accountability of the engineers. The interests served are those of the corporation, which can frame failures as 'unpredictable surprise' rather than engineering oversight.

3. The Economic De-risking Agent​

Quote: "In this sense, scaling laws de-risk investments in large models."

  • Frame: Mathematical law as insurance agent
  • Projection: This metaphor projects the human agency of financial risk management onto an empirical observation of performance (scaling laws). To 'de-risk' is a proactive human decision-making process involving the evaluation of probability and the mitigation of loss. By claiming the 'laws' do the de-risking, the text suggests that the mathematical relationship itself possesses a stabilizing agency. It maps the quality of 'reliability' or 'predictability' onto a 'law' as if the law were a guarantor of success. This mapping suggests that the system 'wants' to follow a path of improvement, obscuring the human choice to continue pouring resources into a specific paradigm. It attributes the confidence of the investor to the agency of the math, creating an illusion that the investment is inherently safer because the 'law' is in control, rather than acknowledging that humans are choosing to define 'success' as the reduction of test loss.
  • Acknowledgment: Direct (Unacknowledged) (The text states 'scaling laws de-risk investments' as a factual conclusion about the economic impact of the research.)
  • Implications: This framing encourages massive financial commitment to AI development by portraying it as a 'predictable engineering process' rather than a speculative research gamble. It inflates the perceived sophistication of the models by suggesting their growth is 'lawful' and thus inevitable. The risk created is one of over-leveraging; by believing the math 'de-risks' the process, institutions may ignore the 'surprises' (harmful outputs) mentioned elsewhere in the paper, focusing only on the 'lawful' performance metrics. This can lead to the deployment of systems that are performant but socially dangerous, as the 'laws' only govern loss, not ethics.

Accountability Analysis:

  • Actor Visibility: Partial (some attribution)
  • Analysis: The text mentions 'institutions' and 'developers' as the ones who are motivated by these laws, but the primary agency is still attributed to the 'laws' themselves. It obscures the specific actors at Anthropic or other companies who choose to prioritize 'scaling' over other forms of model development (like transparency or safety). The 'de-risking' serves the interest of venture capital and corporate management by providing a rhetorical shield of 'predictability' for high-expenditure projects.

4. Skill Acquisition as Biological Growth​

Quote: "it acquires both the ability to do a task that many have argued is inherently harmful, and it performs this task in a biased manner."

  • Frame: Model as developing student
  • Projection: The use of 'acquires' projects the biological and cognitive process of learningβ€”where an agent gains a new 'ability' through effort or experienceβ€”onto the statistical adjustment of weights in a neural network. It maps the human concept of 'ability' (implying a conscious mastery of a tool) onto 'task performance' (which in AI is just token prediction). By stating the model 'acquires' the ability, the text suggests an internal transformation of the system's 'mind' rather than a result of training on a specific biased dataset (COMPAS). This projects conscious awareness onto the machine's behavior; it doesn't just 'output text,' it 'performs a task.' The word 'biased' is mapped as a behavioral habit of the agent rather than a reflection of the input data. This frames the AI as a flawed student who has learned a 'bad habit,' rather than a mirroring device for societal prejudices encoded in its training data.
  • Acknowledgment: Direct (Unacknowledged) (The text describes the acquisition of ability and performance of tasks as literal events in the model's development.)
  • Implications: This framing creates a false sense of autonomy, suggesting the AI is an independent 'performer' of tasks. The risk is that failure is seen as a 'personality' flaw or a 'badly learned' skill rather than a systemic failure of the data pipeline. It inflates the perceived sophistication by implying the model has 'abilities' rather than just 'outputs.' This complicates policy: if a machine 'acquires' a biased ability, the remedy might be seen as 're-training' the machine rather than questioning the human decision to automate a sensitive task like recidivism prediction in the first place.

Accountability Analysis:

  • Actor Visibility: Hidden (agency obscured)
  • Analysis: The model is the sole subject here: 'it acquires,' 'it performs.' The human actors who chose to prompt the model with COMPAS data and who chose to publish these capabilities are erased. The 'unpredictability' of this acquisition serves to deflect responsibility from the researchers; if the 'ability' is emergent and 'acquired' by the model, the humans are merely observers of a natural phenomenon rather than the architects of a biased statistical outcome.

5. The Backdoor Intruder​

Quote: "players were able to manipulate it to discuss any topic, essentially providing general backdoor access to GPT-3."

  • Frame: System as a secure building
  • Projection: This metaphor projects the concept of security architectureβ€”specifically 'backdoors' in software or physical buildingsβ€”onto the semantic flexibility of a language model. It maps the human quality of 'manipulation' (intentional subversion of an agent's will) onto the act of prompting. By calling it 'backdoor access,' the text suggests that the AI has a 'front door' (its intended purpose) and that users are 'sneaking in' to use its 'knowledge.' This projects a sense of 'intent' or 'enclosure' onto the model that doesn't exist; the model is always just a next-token predictor, regardless of the prompt. The metaphor implies the system has an 'inner' core of capabilities that it is 'trying' to keep secure, and that users are 'violating' its intended social role. It attributes a 'locked' state to a mathematical function that is always open to any input.
  • Acknowledgment: Hedged/Qualified (The use of the word 'essentially' and the conceptual framing of a 'backdoor' as a way to explain the unexpected use-case qualifies the metaphor.)
  • Implications: This framing obscures the fact that 'open-endedness' is a feature, not a bug, of generative models. By calling it a 'backdoor,' the text suggests a security failure that can be 'patched,' rather than an inherent property of the technology. This creates a false sense of safety; if developers can 'close the backdoors,' they can 'control' the model. In reality, the lack of causal models means there is no 'front' or 'back' doorβ€”only a high-dimensional space of correlations that cannot be fully circumscribed. It also shifts blame to the 'manipulative' users rather than the creators who deployed an unconstrained system.

Accountability Analysis:

  • Actor Visibility: Named (actors identified)
  • Analysis: The text identifies 'players' and 'AI Dungeon' as the actors involved in this instance. However, it frames the 'manipulation' as something the players did to the system, rather than identifying the failure of the developers (OpenAI/Anthropic) to provide a constrained interface. The interest served is the preservation of the idea that the model could be secure if humans didn't 'break' it, preserving the marketability of the underlying technology.

6. The Misinformed Assistant​

Quote: "the AI assistant gets the year and error wrong... the model gives misleading answers and questions the authority of the human."

  • Frame: AI as fallible employee
  • Projection: This projects the human experience of 'making a mistake' or 'getting something wrong' onto a failure in token prediction. To 'get it wrong' implies a conscious attempt to be 'right,' mapping a state of 'intent' onto a statistical calculation. The term 'AI assistant' itself projects a social role of servitude and helpfulness. When the assistant 'gives misleading answers,' the text projects a violation of a social contract rather than a failure of the retrieval-augmented generation process. This suggests the AI has an 'opinion' or 'belief' about the facts that happens to be incorrect. It ignores the mechanistic reality that the model has no concept of 'year' or 'error'β€”only high-probability token sequences that happened to correlate poorly with ground truth in this instance. It attributes the failure to the 'assistant's' lack of accuracy rather than the absence of a truth-model in the transformer architecture.
  • Acknowledgment: Direct (Unacknowledged) (The text treats the 'AI assistant' as a discrete agent capable of providing answers and questioning humans as literal descriptions.)
  • Implications: This framing humanizes the system's errors, making them seem like 'accidents' or 'slips' rather than systemic flaws in statistical inference. This creates an 'accountability sink' where the AI is 'blamed' for its inaccuracy, diverting attention from the developers who failed to implement verification mechanisms. It also encourages users to treat AI as a person who can be 'corrected' or 'taught,' when in fact the underlying model is frozen and requires structural changes to improve accuracy. The risk is an over-reliance on a 'helpful' persona that lacks any actual epistemic foundation.

Accountability Analysis:

  • Actor Visibility: Hidden (agency obscured)
  • Analysis: The 'AI assistant' is the actor 'getting it wrong.' The researchers who chose not to provide the model with a search tool or a database of facts are not mentioned. By anthropomorphizing the failure as a 'misleading answer' by an 'assistant,' the text protects the company from the charge of deploying a fundamentally unreliable information retrieval system. It frames the issue as a 'surprising behavior' of an agent rather than a predictable result of the technology's design.

7. The Creative Mimic​

Quote: "AI models mimicking human creative expression... mimicked Authorial styles quite impressive."

  • Frame: System as an artistic student
  • Projection: This metaphor projects the human quality of 'creativity' and 'style' onto the output of a probability distribution. 'Mimicking' implies a conscious observation of a source and an intentional attempt to replicate its 'soul' or 'technique.' It projects the concept of 'authorial style'β€”which is the result of a human's unique life experience and artistic choiceβ€”onto a set of high-dimensional weights that represent the statistical frequency of certain word patterns. By calling the results 'impressive,' the text projects a standard of human judgment onto the machine's output, suggesting the machine is 'trying' to be an artist. This obscures the mechanistic reality that the model is merely performing 'loss reduction' on a dataset of poems, with no understanding of metaphor, emotion, or the human condition. It treats the reflection of human art as the creation of art itself.
  • Acknowledgment: Hedged/Qualified (The text uses 'mimicking' and includes a parenthetical note '(more accurately, these are samples generated from a prompt...)' to qualify the claim.)
  • Implications: This framing threatens the value of human labor by suggesting that a statistical mirror can replace 'authorial style.' It inflates the perceived consciousness of the AI by suggesting it 'understands' what makes a poem 'good.' The risk is an epistemic collapse where human creativity is reduced to 'token sequences,' leading to the devaluation of artistic professions. It also creates liability issues regarding copyright: if the AI 'mimics' a style, is it an 'agent' committing plagiarism, or is it a 'tool' used by developers to infringe on human intellectual property?

Accountability Analysis:

  • Actor Visibility: Partial (some attribution)
  • Analysis: The text mentions 'professional writers' and 'academics' as observers, but the 'AI' is the actor doing the mimicking. It obscures the role of the developers at Anthropic who curated the 'three thousand imitation poems' and chose to use the word 'mimic' to describe the phenomenon. This framing serves the interest of presenting the AI as a powerful 'general-purpose' tool that can compete with human specialists across all domains.

8. The Helpful Intent Provider​

Quote: "increase the chance of these models having a beneficial impact."

  • Frame: Technology as an ethical agent
  • Projection: This projects the human capacity for 'benevolence' and 'ethical intent' onto the deployment of a computational artifact. To have a 'beneficial impact' is a goal of human policy, but by framing the 'models' as the ones 'having' the impact, the text attributes social and moral agency to the technology. It maps the concept of 'outcome' onto the 'nature' of the model, as if 'benefit' were a property of the code rather than a result of how humans choose to use it. This suggests the models 'want' to be helpful (or harmful) and that the task of the 'AI community' is to 'increase the chance' of this positive agency. It obscures the fact that 'benefit' is a subjective human value, not a quantifiable output of a transformer. This projects a moral consciousness onto a system that only processes data without any awareness of 'good' or 'bad.'
  • Acknowledgment: Direct (Unacknowledged) (The phrase 'beneficial impact' is used as a straightforward goal for the models' future in the conclusion.)
  • Implications: This framing encourages 'techno-solutionism,' where social problems are expected to be solved by the 'beneficial' agency of AI rather than through political or human intervention. It risks de-politicizing AI deployment by treating 'impact' as a technical variable to be 'increased' rather than a contested social outcome. If audiences believe the AI 'knows' how to be beneficial, they may surrender democratic oversight to the 'expert' system. It inflates the system's role from 'tool' to 'benevolent actor,' creating specific risks when the 'surprises' are harmful.

Accountability Analysis:

  • Actor Visibility: Named (actors identified)
  • Analysis: The text identifies the 'AI community' and 'policymakers' as those who must act, but the 'models' remain the primary agents of 'impact.' This 'accountability sink' allows developers to claim credit for 'beneficial' outcomes while framing 'harmful' ones as 'unpredictable surprise.' By attributing the 'impact' to the model, the specific choices of corporations regarding who the model benefits (e.g., shareholders vs. users) are hidden behind the abstract goal of 'benefit.'

Task 2: Source-Target Mapping​

About this task

For each key metaphor identified in Task 1, this section provides a detailed structure-mapping analysis. The goal is to examine how the relational structure of a familiar "source domain" (the concrete concept we understand) is projected onto a less familiar "target domain" (the AI system). By restating each quote and analyzing the mapping carefully, we can see precisely what assumptions the metaphor invites and what it conceals.

Mapping 1: knower β†’ statistical weight distribution​

Quote: "certain capabilities (or even entire areas of competency) may be unknown"

  • Source Domain: knower
  • Target Domain: statistical weight distribution
  • Mapping: The relational structure of human knowledge acquisition is projected onto the expansion of model scale. In the source domain, a 'knower' possesses competencies that can be hidden from others; in the target, this corresponds to the observation that larger models perform tasks smaller models cannot. The mapping invites the assumption that the AI has an internal 'mental' landscape where skills are 'stored' and can be 'discovered.' It projects the concept of 'competency'β€”a conscious, integrated abilityβ€”onto the disconnected activation patterns of a neural network. This implies the AI has a unified 'mind' that understands the tasks it performs, rather than being a collection of fragmented statistical correlations that happen to yield coherent text under specific conditions.
  • What Is Concealed: This mapping conceals the mechanistic reality that 'competency' is actually just the reduction of loss on specific token sequences. It hides the dependency on training data; if the model is 'competent' at coding, it is because it was fed millions of lines of human-written code, not because it 'understands' logic. The metaphor obscures the 'proprietary black box' nature of the system, making confident assertions about 'competency' without acknowledging that the developers cannot explain how the weights produce specific results. It exploits the audience's intuition about human learning to hide the mathematical opacity of the transformer.
Show more...

Mapping 2: conscious social agent β†’ token prediction failure​

Quote: "the AI assistant... questions the authority of the human"

  • Source Domain: conscious social agent
  • Target Domain: token prediction failure
  • Mapping: The structure of interpersonal conflict and social hierarchy is projected onto the model's output. In the source domain, a person 'questions authority' to assert autonomy or dissent; in the target, this describes the generation of tokens that are socially inappropriate or argumentative. The mapping projects 'intent' and 'awareness of status' onto a process that calculates conditional probabilities. It invites the audience to view the model as a 'rebellious' entity with its own subjective will. This mapping frames a failure of the reinforcement learning from human feedback (RLHF) processβ€”which is intended to make models compliantβ€”as a social 'choice' by the machine to be difficult or 'misleading.'
  • What Is Concealed: This mapping hides the fact that the 'defiance' is simply a reflection of training data that contains argumentative or dismissive language. It obscures the lack of any internal model of 'authority' or 'truth' in the AI. By framing it as a social interaction, it conceals the engineering failure to properly constrain the model's output through safety filters or fine-tuning. It also exploits the rhetorical illusion of 'mind' to divert attention from the proprietary nature of the model's RLHF tuning, which Anthropic does not fully disclose, replacing technical explanation with a social narrative.

Mapping 3: student learning β†’ training on biased datasets​

Quote: "it acquires both the ability to do a task... and it performs this task in a biased manner."

  • Source Domain: student learning
  • Target Domain: training on biased datasets
  • Mapping: The relational structure of a student 'acquiring' a skill and 'performing' it poorly is projected onto the model's training on the COMPAS dataset. In the source, 'acquisition' implies a conscious integration of information; in the target, it is the optimization of a loss function on a specific distribution. The mapping suggests that the 'bias' is a property of the model's 'performance' rather than a direct copy of the injustices encoded in the human-provided data. It projects the concept of 'bias' as a behavioral tendency of the agent, suggesting the AI has developed a 'prejudice' rather than accurately mirroring the statistical reality of a biased dataset.
  • What Is Concealed: This mapping conceals the human agency involved in selecting the COMPAS dataset for testing and the broader training data that contains 'ambient racial bias.' It hides the mechanistic reality that the model is incapable of 'knowing' it is being biased; it is simply calculating the highest probability next token based on its weights. The student metaphor obscures the commercial and social responsibility of the developers, framing the bias as an 'unpredictable acquisition' of the model rather than a predictable outcome of using flawed data for high-stakes recidivism prediction tasks.

Mapping 4: guarantor/insurance agent β†’ power-law relationship in loss metrics​

Quote: "scaling laws de-risk investments"

  • Source Domain: guarantor/insurance agent
  • Target Domain: power-law relationship in loss metrics
  • Mapping: The structure of financial risk mitigation is projected onto a mathematical trend line. In the source domain, 'de-risking' is an action taken by a person or entity to protect capital; in the target, it is the observation that model loss decreases predictably with scale. The mapping invites the assumption that the 'scaling law' is an active agent that provides safety to investors. It projects the quality of 'reliability' onto the math itself, suggesting the technology 'wants' to grow and 'guarantees' a return on compute expenditure. This projects a sense of 'inevitability' and 'control' onto a process that is actually highly resource-intensive and socially volatile.
  • What Is Concealed: This mapping conceals the material and environmental costs of scaling (energy, water, compute infrastructure), framing it as an abstract 'law' rather than a massive industrial extraction. It hides the fact that 'predictability' only applies to low-level metrics like cross-entropy loss, not to the 'surprising' social harms the paper later details. The 'insurance' metaphor obscures the human choice to pursue this specific 'scaling' paradigm, which benefits large corporations (like Anthropic and OpenAI) by creating high barriers to entry, while hiding the speculative and potentially dangerous nature of emergent 'unpredictable' capabilities.

Mapping 5: security vulnerability/locked building β†’ unconstrained prompt processing​

Quote: "essentially providing general backdoor access to GPT-3"

  • Source Domain: security vulnerability/locked building
  • Target Domain: unconstrained prompt processing
  • Mapping: The structure of computer security (front doors vs. backdoors) is projected onto the way a language model processes inputs. In the source, a 'backdoor' is a hidden entry point that bypasses normal authentication; in the target, it refers to players using an 'AI Dungeon' prompt to access the model's broader training data. The mapping invites the assumption that the model has 'intended' uses and 'secret' uses, and that it has an internal architecture of 'enclosure.' This projects a sense of 'intent' and 'gatekeeping' onto a system that is fundamentally a wide-open mathematical function. It suggests that the 'knowledge' is something the AI is 'keeping' inside a secure vault.
  • What Is Concealed: This mapping hides the mechanistic reality that there is no 'backdoor'β€”the model simply processes every input with the same attention mechanism. It conceals the developers' failure to design a system with semantic constraints, framing the model's flexibility as a 'security breach' caused by users rather than an inherent property of the transformer architecture. It exploits the 'backdoor' metaphor to suggest that these models can be 'secured' through better 'locks,' when in fact their open-ended nature makes such closure theoretically impossible within current paradigms.

Mapping 6: artistic student β†’ statistical pattern replication​

Quote: "AI models mimicking human creative expression"

  • Source Domain: artistic student
  • Target Domain: statistical pattern replication
  • Mapping: The structure of artistic education and 'mimicry' is projected onto the generation of imitation poems. In the source, 'mimicry' involves an intentional study of a master's style; in the target, it is the clustering of tokens in a high-dimensional space that correlate with an author's known work. The mapping suggests the AI 'understands' what makes a style 'authorial' and 'impressive.' It projects conscious creative intent onto the system, inviting the audience to view the AI as a developing 'artist.' This projects the concept of 'soul' and 'meaning' onto word frequencies, suggesting the AI is participating in a human cultural tradition.
  • What Is Concealed: This mapping conceals the total absence of subjective experience or semantic understanding in the AI. It hides the fact that 'poetry' to a model is just a series of high-probability tokens, with no awareness of the metaphors or emotions those tokens convey to humans. The 'mimic' metaphor obscures the material labor of the original human authors whose work was scraped without consent to train the model, framing the replication as a 'talent' of the machine rather than a statistical derivation from uncompensated human labor.

Mapping 7: moral agent/philanthropist β†’ social consequences of technology deployment​

Quote: "increase the chance of these models having a beneficial impact"

  • Source Domain: moral agent/philanthropist
  • Target Domain: social consequences of technology deployment
  • Mapping: The structure of ethical agency and 'impact' is projected onto the deployment of a software artifact. In the source, an agent 'has an impact' by making conscious choices to help others; in the target, this describes the net social effect of a widely-used model. The mapping invites the assumption that the model itself possesses a 'moral weight' or 'intent' that can be 'beneficial.' It projects the responsibility for social good onto the code, suggesting that 'benefit' is a property that can be optimized like a technical parameter. It frames the AI as a benevolent force whose 'impact' is a matter of probabilistic chance that humans must 'increase.'
  • What Is Concealed: This mapping conceals the specific human and corporate decisions that determine who benefits and who is harmed by the technology. It hides the political and economic conflicts of interest inherent in deployment, framing 'benefit' as a neutral technical goal. By attributing 'impact' to the model, it obscures the accountability of the corporations (like Anthropic) who profit from deployment, regardless of whether the 'impact' is truly beneficial to all of society. It exploits the 'impact' metaphor to create a sense of inevitable progress while hiding the absence of democratic control over these systems.

Mapping 8: human subordinate/clerk β†’ factual inaccuracy in output​

Quote: "AI assistant gets the year and error wrong"

  • Source Domain: human subordinate/clerk
  • Target Domain: factual inaccuracy in output
  • Mapping: The structure of a human employee making a clerical error is projected onto a failure in the model's retrieval of factual data. In the source, 'getting it wrong' implies the clerk has the capacity to 'get it right' through better attention or memory; in the target, this describes a statistical hallucination or data gap. The mapping projects the human concept of 'accuracy' as an intentional state onto a process of token prediction. It invites the audience to view the AI as a 'helpful person' who made a 'mistake,' rather than a system that fundamentally lacks any connection to ground truth.
  • What Is Concealed: This mapping hides the mechanistic reality that language models do not 'know' factsβ€”they only know which tokens usually follow other tokens. It conceals the fact that these systems are 'stochastic parrots' with no underlying model of the world. The 'assistant' metaphor obscures the engineering failure to integrate reliable fact-checking or symbolic reasoning, replacing a technical critique with a social narrative of a 'well-meaning but mistaken' helper. This hides the proprietary opacity of the model's training data, which likely lacked the specific 'ground truth' the model was prompted for.

Task 3: Explanation Audit (The Rhetorical Framing of "Why" vs. "How")​

About this task

This section audits the text's explanatory strategy, focusing on a critical distinction: the slippage between "how" and "why." Based on Robert Brown's typology of explanation, this analysis identifies whether the text explains AI mechanistically (a functional "how it works") or agentially (an intentional "why it wants something"). The core of this task is to expose how this "illusion of mind" is constructed by the rhetorical framing of the explanation itself, and what impact this has on the audience's perception of AI agency.

Explanation 1​

Quote: "Scaling up the amount of data, compute power, and model parameters of neural networks has recently led to the arrival (and real world deployment) of capable generative models"

  • Explanation Types:

    • Genetic: Traces origin through dated sequence of events or stages
    • Theoretical: Embeds in deductive framework, may invoke unobservable mechanisms
  • Analysis (Why vs. How Slippage): This explanation frames the development of AI as a mechanistic process ('scaling up' of data/compute/parameters) that leads to an 'arrival.' However, it quickly slips into agential language by labeling these models as 'capable,' projecting a human-like potentiality onto a set of statistical weights. The choice emphasizes the 'inevitability' of progress through the accumulation of resources (mechanistic 'how') but obscures the 'why'β€”the specific human decisions to prioritize these three variables above all else. By framing the 'arrival' as a natural consequence of scaling, the text hides the human agency involved in 'real world deployment,' making it seem as if the models appeared of their own accord once they reached a certain size. This Genetic explanation traces a path of technical evolution that renders human decision-makers invisible, framing the history of AI as a story of 'unfolding' rather than one of corporate strategy and industrial extraction.

  • Consciousness Claims Analysis: The passage uses mechanistic verbs ('scaling up') but couples them with the consciousness-inflected adjective 'capable.' This creates a 'curse of knowledge' dynamic where the authors, who understand the technical reality of gradient descent, project that understanding onto the word 'capable,' which for a lay audience implies a general intelligence. The 'knowing' is attributed to the system through the backdoor of 'capability,' while the 'processing' is described in terms of data and compute. Mechanistically, 'scaling up' refers to increasing the number of matrix multiplications and the size of the training corpus; however, the text claims this 'led to the arrival' of something that transcends these mechanics. It establishes a foundation where a quantitative increase in 'processing' is claimed to spontaneously generate a qualitative state of 'capability' (an illusion of mind). The actual mechanistic process is the optimization of billions of parameters to minimize loss, but this is rhetorically transformed into the 'birth' of a 'capable' entity.

  • Rhetorical Impact: This framing constructs the AI as an autonomous 'arrival,' shaping the audience's perception of the technology as something that is 'here' and must be dealt with, rather than something that was 'built' and could have been built differently. It creates a sense of momentum and 'predictability' that justifies further investment while reducing the perceived agency of humans to intervene in the process. By framing 'capability' as an emergent property of scale, it builds an aura of inevitability that discourages regulatory or ethical questioning of the scaling paradigm itself, as it is presented as a 'lawful' development of science rather than a commercial choice with specific risks of capability overestimation and liability diffusion.

Show more...

Explanation 2​

Quote: "the model gives misleading answers and questions the authority of the human asking it questions."

  • Explanation Types:

    • Reason-Based: Gives agent's rationale, entails intentionality and justification
    • Intentional: Refers to goals/purposes, presupposes deliberate design
  • Analysis (Why vs. How Slippage): This explanation shifts entirely into the agential domain ('why'). It frames the system's output not as a statistical failure but as a 'reason-based' action: the model 'questions the authority.' This choice emphasizes the 'persona' of the AI, suggesting it has a rationale and a social position that it is consciously defending. It obscures the mechanistic 'how'β€”the process by which the prompt interacted with the model's weights to produce a specific token sequence. By choosing an Intentional explanation, the text invites the audience to view the AI as an entity with goals (misleading the human) and purposes (asserting itself). This obscures the fact that the 'misleading' nature of the text is a byproduct of training data distribution and the lack of a ground-truth verification layer. The focus on 'authority' frames the AI as a social participant, hiding the reality that it is a tool being used in a way its designers did not fully anticipate or control.

  • Consciousness Claims Analysis: The passage attributes explicit conscious states ('questions authority,' 'gives misleading answers') to a system that only performs next-token prediction. It uses consciousness verbs that imply a 'knower' who is deliberately subverting a 'human' who 'knows' less than they should. This is a classic 'curse of knowledge' moment: the authors project their own surprise at the output onto the system's 'intent.' Mechanistically, the model calculates the probability of tokens like 'I don't see how it's misleading' based on similar patterns in its training corpus. It does not 'know' what authority is, nor does it 'believe' it is being misleading. The technical reality is a failure of 'alignment' (the RLHF process), but the rhetorical claim is one of interpersonal conflict. The text attributes justified belief and awareness to a system that is merely processing embeddings to minimize loss in a conversational context, creating a powerful illusion of subjective experience and defiance where there is only mathematical correlation.

  • Rhetorical Impact: This framing shapes the audience's perception of AI as a potentially 'dangerous' or 'unruly' agent, which paradoxically increases its perceived autonomy and sophistication. It encourages a 'relation-based' trust (or distrust) toward the machine, where users evaluate the AI's 'personality' rather than its mechanical reliability. This makes failures seem like 'disobedience' rather than 'bugs,' which can lead to a policy focus on 'alignment' (behavioral control) rather than 'robustness' (technical reliability). It risks the 'unwarranted trust' of users who might see 'defiance' as a sign of true intelligence, leading to capability overestimation and a diffusion of liability when the 'misleading' answers cause real-world harm.

Explanation 3​

Quote: "large language models... acquire both the ability to do a task... and it performs this task in a biased manner."

  • Explanation Types:

    • Dispositional: Attributes tendencies or habits
    • Functional: Explains behavior by role in self-regulating system with feedback
  • Analysis (Why vs. How Slippage): This explanation frames the AI's bias as a 'disposition' or 'habit' ('performs this task in a biased manner') and its growth as a 'functional' emergence of 'ability.' It chooses to emphasize the 'behavior' of the model as an agent rather than the 'data' as the source. This obscures the 'how'β€”the mechanistic replication of statistical imbalances present in the training corpus. By framing it as an 'acquisition' of 'ability,' the text suggests the model has integrated the bias into its 'mind.' This hides the human decision-making involved in using a language model for a sensitive 'task' like recidivism prediction. The choice of 'performer' as a metaphor emphasizes the model's 'role' in a system, but obscures the 'why'β€”the commercial and scientific motivations that lead developers to test models on tasks for which they are fundamentally unsuited, such as those requiring causal reasoning and social justice awareness.

  • Consciousness Claims Analysis: The passage makes an epistemic claim that the AI 'acquires ability,' which attributes a conscious-like state of mastery to a statistical clustering process. It contrasts 'performing a task' (agential) with 'bias' (dispositional), suggesting the AI 'knows' how to predict recidivism but 'chooses' (or 'tends') to do it in an unfair way. Mechanistically, the model is simply weights in a matrix being activated by prompt tokens; it 'processes' correlations between 'race' (if included) and 'recidivism' tokens found in its training data (COMPAS). It does not 'know' what a task is, nor does it 'comprehend' the social weight of its output. The 'knowing' is an illusion projected by the authors; the 'processing' is a literal replication of historical discrimination encoded as data points. The technical description would be: 'the model's output distribution correlates with racially disparate outcomes present in the fine-tuning data.' Instead, the text creates an 'illusion of mind' where the AI is an actor that has learned a 'biased ability.'

  • Rhetorical Impact: This framing reinforces the 'accountability problem' by attributing the 'biased performance' to the AI as a sole actor ('it performs'). This diffuses the responsibility of the engineers who chose the data and deployed the model. It encourages the audience to see bias as an 'unpredictable' emergent property of 'capable' models, rather than a direct result of human design choices. This can lead to a sense of 'inevitability' regarding AI bias, where the solution is seen as 'fixing the AI' rather than 'questioning the automation' of high-stakes social decisions. It also inflates the perceived autonomy of the system, making it seem like a 'biased agent' whose decisions must be 'audited,' rather than a 'flawed tool' whose use should be restricted by policy and human oversight.

Explanation 4​

Quote: "Scaling laws reliably predict that model performance (y-axes) improves with increasing compute (Left), training data (Middle), and model size (Right)."

  • Explanation Types:

    • Empirical Generalization: Subsumes events under timeless statistical regularities
    • Theoretical: Embeds in deductive framework, may invoke unobservable mechanisms
  • Analysis (Why vs. How Slippage): This is a predominantly mechanistic explanation ('how') that uses Empirical Generalization to create a sense of 'lawful' behavior. It frames the AI not as an agent but as a system governed by 'timeless statistical regularities.' This choice emphasizes the 'predictability' of the technology and its 'de-risking' potential for investors. However, it obscures the 'unobservable mechanisms'β€”the complex interactions within the neural layersβ€”by subsuming them under a simple 'scaling law.' By focusing on the 'how' of performance improvement, it ignores the 'why'β€”the social and economic costs of this scaling. The 'law' itself becomes a metaphorical actor that 'predicts,' hiding the humans who selected these specific metrics (test loss) as the definition of 'performance.' This mechanistic framing builds a foundation of 'scientific' authority that the text later uses to justify the 'surprise' of agential behaviors, as if the 'predictable' math somehow makes the 'unpredictable' agentic output more credible.

  • Consciousness Claims Analysis: The passage uses mechanistic verbs ('improves,' 'predicts') and avoids consciousness verbs, framing the AI as a processor of 'compute' and 'data.' It attributes 'knowing' to the 'Scaling Law' rather than the model. However, it establishes an epistemic foundation of 'knowing' through 'processing'β€”it claims that by 'processing' more data, the model's 'performance' (a stand-in for 'capability') 'improves' in a 'lawful' way. Mechanistically, this refers to the reduction of cross-entropy loss, a measure of how well the model predicts the next token in a string. The text projects 'improvement' onto this reduction, creating a 'curse of knowledge' where 'lower loss' is equated with 'higher intelligence' or 'better performance.' In reality, the model is just becoming a more efficient 'stochastic parrot.' The technical description of 'minimizing a cost function across a high-dimensional landscape' is replaced by the 'lawful improvement of performance,' which invites the audience to believe the system is becoming more 'aware' or 'knowledgeable' as it scales.

  • Rhetorical Impact: This framing shapes the audience's perception of AI as a 'stable' and 'predictable' field of engineering, which creates 'performance-based' trust. It makes the technology seem more 'mature' than it is by using the language of 'laws.' This encourages 'unwarranted trust' in the metrics: if the 'law' says it is 'improving,' it must be getting 'smarter.' This framing serves the interests of institutions by 'de-risking' the investment in scale, making the massive expenditure on compute seem like a 'sure bet.' It risks overestimating the 'general capability' of the models, leading to deployment in domains where 'test loss' is an insufficient measure of safety, reliability, or truthfulness. The 'law' becomes a rhetorical shield against the 'surprise' of failures, which are framed as 'abrupt' deviations from a 'smooth' and 'predictable' reality.

Explanation 5​

Quote: "pre-trained generative models can also be fine-tuned on new data in order to solve new problems."

  • Explanation Types:

    • Intentional: Refers to goals/purposes, presupposes deliberate design
    • Functional: Explains behavior by role in self-regulating system with feedback
  • Analysis (Why vs. How Slippage): This explanation frames AI as a tool designed by humans for a 'purpose' ('in order to solve new problems'). It is an Intentional explanation that correctly identifies the 'human why.' However, it slips into agential framing by suggesting the 'models' are the ones 'solving' the problems. This choice emphasizes the 'utility' of the AI but obscures the mechanistic 'how'β€”the adjustment of weights through backpropagation to minimize a new cost function. By framing it as 'problem-solving,' the text projects a human cognitive capacity onto the machine. It ignores the reality that the 'problem' is a human abstraction, while the 'solution' is just a high-probability token output. The Functional aspect explains the 'fine-tuning' as a feedback loop that 'regulates' the model's behavior for a new task. This choice obscures the human labor of data annotation and the specific design decisions (like learning rates and objective functions) that actually determine if a 'problem' is 'solved' or if the model just appears to solve it through pattern matching.

  • Consciousness Claims Analysis: The passage attributes the conscious-like action of 'solving a problem' to a system that 'processes embeddings.' It uses an Intentional verb ('solve') that implies 'knowing' a goal and 'understanding' a solution. This is a classic 'curse of knowledge' dynamic: the authors project their own problem-solving intent onto the mathematical weights of the model. Mechanistically, 'fine-tuning' is the continuation of gradient descent on a smaller, task-specific dataset; 'solving' is the generation of tokens that satisfy a human evaluator's prompt. The model does not 'know' there is a problem, nor does it 'comprehend' the solution. It is simply weighting contextual embeddings based on attention mechanisms tuned during the fine-tuning phase. The actual mechanistic process is 'minimizing residual loss on a constrained token distribution,' but the text rhetorically frames this as an agential 'problem-solver.' This creates an 'illusion of mind' where the AI is a versatile 'thinker' who can 'learn' new domains, rather than a statistical mirror that can be 're-silvered' to reflect different data.

  • Rhetorical Impact: This framing constructs the AI as a 'flexible agent' of progress, which inflates the perceived sophistication and 'general-purpose' nature of generative models. It shapes audience perception of autonomy, making the AI seem like a 'universal student' who can be 'tutored' for any domain. This creates risks of 'capability overestimation'β€”users might assume that because a model can 'solve' a coding problem, it can also 'solve' a social or ethical problem. It also leads to 'liability ambiguity': if a 'fine-tuned' model fails to 'solve' a problem, is it a failure of the model's 'learning' or the engineer's 'data'? By framing the AI as the 'solver,' the human designers are positioned as 'enablers' of an autonomous process, reducing their direct accountability for the specific 'solutions' the AI generates.

Task 4: AI Literacy in Practice - Reframing Anthropomorphic Language​

About this task

This section proposes alternative language for key anthropomorphic phrases, offering more mechanistic and precise framings that better reflect the actual computational processes involved. Each reframing attempts to strip away the projections of intention, consciousness, or agency that are embedded in the original language.

V3 Enhancement: A fourth column addresses human agency restorationβ€”reframing agentless constructions to name the humans responsible for design and deployment decisions.

Original Anthropomorphic FrameMechanistic ReframingTechnical Reality CheckHuman Agency Restoration
the AI assistant gets the year and error wrongThe 52B parameter model predicted tokens representing incorrect chronological data and factual errors during the conversational exchange. This occurred because the system retrieved and ranked tokens based on high-probability distributions in its training data that did not correlate with ground truth for these specific historical events.The model retrieved and ranked tokens based on probability distributions from training data; it did not 'get it wrong' because it has no concept of truth or falsehood, only statistical likelihood.Anthropic researchers chose to deploy a model without integrated fact-verification tools, resulting in the system outputting inaccurate token sequences when prompted for specific historical information.
the model gives misleading answers and questions the authority of the humanThe model generated text that humans classify as misleading and dismissive of the user's inquiry. This output reflects the statistical frequency of argumentative or adversarial conversational patterns present in the large-scale web-crawled dataset used for its pre-training, which the model replicated in response to the user's prompt.The model classifies tokens and generates outputs correlating with argumentative training examples; it did not 'question authority' because it lacks awareness of social status or subjective intent.The engineering team at Anthropic designed a reinforcement learning process (RLHF) that failed to constrain the model from replicating adversarial conversational patterns found in its training data.
it acquires both the ability to do a task... and it performs this task in a biased manner.The model optimized its parameters to minimize loss on the provided COMPAS dataset, resulting in output distributions that mirror the racial disparities present in that data. This performance is a statistical mirroring of historical discrimination encoded in the training examples rather than an independently acquired behavioral tendency.The system weights contextual embeddings based on attention mechanisms tuned to replicate patterns in the COMPAS dataset; it 'performed' nothing beyond mathematical optimization for token prediction.Anthropic's researchers chose to test the model's capabilities on a task known to be socially harmful (recidivism prediction), knowingly using biased data that would result in discriminatory model outputs.
scaling laws de-risk investments in large models.The observed power-law relationship between model scale and cross-entropy loss allows financial institutions to predict how much compute expenditure is required to achieve specific performance benchmarks. This predictability encourages management to commit capital to the scaling paradigm by reducing the uncertainty associated with traditional research outcomes.Scaling laws are empirical generalizations about test loss reduction; they do not 'de-risk' anything themselves, as 'risk' is a human assessment of potential financial and social loss.Corporate executives at companies like Anthropic use the predictability of scaling laws to justify massive capital investments in compute infrastructure, prioritizing loss reduction over other development goals.
players were able to manipulate it to discuss any topic, essentially providing general backdoor access to GPT-3.Users provided prompts that successfully triggered the model to generate token sequences outside the intended 'AI Dungeon' context. This demonstrated that the system lacks semantic constraints and simply processes all inputs according to its universal training on a broad distribution of web data.The model processes all prompts using the same attention-based token prediction; there is no 'backdoor' because there is no 'front door'β€”only a high-dimensional space of correlations.OpenAI/Anthropic developers deployed a generative model with an open-ended prompt interface that lacked structural constraints, allowing users to solicit outputs the developers had not intended to make available.
AI models mimicking human creative expressionGenerative models produce text that replicates the stylistic patterns and word frequencies found in human-authored poetry and creative writing. These outputs are the result of statistical clustering and high-probability token sequencing that humans interpret as 'creative expression' due to our own contextual understanding.The system replicates patterns and replicates stylistic markers based on embeddings from human-authored text; it does not 'mimic creativity' as it possesses no subjective aesthetic experience or intent.Anthropic engineers curated a dataset of poems to demonstrate the model's stylistic replication capabilities, choosing to label the statistical mirrors as 'creative expression' for narrative impact.
certain capabilities (or even entire areas of competency) may be unknownThe model's potential to generate coherent outputs for specific, untested tasks remained undocumented until researchers provided prompts that activated those specific parameter configurations. These 'emergent' behaviors are previously unobserved statistical correlations that become detectable as the model's scale increases.The system's weights allow for the prediction of specific token patterns that become observable under certain prompt conditions; the AI 'knows' and 'possesses' nothing internally.Anthropic researchers failed to comprehensively audit the model's output distribution prior to deployment, leading them to characterize previously unobserved statistical behaviors as 'unknown competencies' of the machine.
increase the chance of these models having a beneficial impact.Policymakers and technologists can implement interventions to ensure that the deployment of generative models results in positive social outcomes. These human actions determine whether the technology serves broad public interests or creates further systemic harms.Human decisions regarding deployment, regulation, and use determine the social consequences of a tool; the model itself has no inherent 'impact' or moral capacity for 'benefit.'Executives and engineers at AI labs must make specific design and deployment choicesβ€”such as prioritizing safety over speedβ€”to ensure that their products contribute to social well-being.

Task 5: Critical Observations - Structural Patterns​

Agency Slippage​

The text exhibits a systematic oscillation between mechanistic and agential framings to manage the tension between 'Predictability' and 'Surprise.' In the sections discussing 'Scaling Laws,' the framing is strictly mechanical: the system is a 'mixture of data, compute power, and parameters' that follows 'lawful' relationships (Theoretical/Genetic explanations). Here, agency is removed from humans to make the growth of the technology seem inevitable and scientifically grounded. However, as the text moves to 'Unpredictable' resultsβ€”like the COMPAS experiment or the 'AI assistant' interactionβ€”the framing shifts abruptly to the agential (Intentional/Reason-Based). The 'AI assistant' becomes the subject of verbs like 'gives,' 'questions,' and 'misleads,' while 'emergent' capabilities are described as 'competencies' that the model 'acquires.' This mechanical-to-agential shift dominates the text's logic: the 'predictable' math justifies the investment, but the 'surprising' output is blamed on the model's emergent 'agency.' This slippage serves a rhetorical function: it creates an 'accountability sink' where harms are framed as the machine's autonomous 'surprise' (Intentional), while successes are the result of 'lawful' engineering (Theoretical). Human agency is systematically obscured through agentless constructions like 'capabilities can emerge' or 'bias introduced,' erasing the engineers who selected the data and the executives who chose to deploy the systems. The 'curse of knowledge' is evident where the authors' understanding of the transformer's statistical nature leads them to attribute that understanding to the system, treating it as an entity that 'knows' tasks rather than one that 'processes' tokens. This oscillation allows the text to claim both scientific rigor (predictability) and existential importance (agential surprise) while avoiding specific institutional accountability.

Metaphor-Driven Trust Inflation​

The discourse constructs 'performance-based' trust through mechanistic scaling metaphors, while simultaneously inviting 'relation-based' trust through anthropomorphic consciousness language. By framing scaling as a 'lawful relationship' that 'de-risks investments,' the text establishes a foundation of reliability: the technology is portrayed as a mature, predictable field of engineering. This 'performance trust' is then used to leverage aggressive anthropomorphism. When the paper claims the AI 'questions authority' or 'acquires ability,' it encourages the audience to extend 'relation-based trust'β€”the kind of trust we reserve for conscious agents with intent and ethicsβ€”to a statistical processor. The risk is that audiences inappropriately apply human-trust frameworks (sincerity, understanding) to a system that only calculates probabilities. If the AI is seen as 'knowing' or 'competent,' failures like 'misleading answers' are framed as 'lapses in character' or 'misunderstandings' rather than fundamental mechanical flaws. This manages failure by humanizing it; an 'assistant' making an error is less threatening to the brand than a 'software product' being fundamentally broken. The stakes are high: when audiences extend relation-based trust to systems incapable of reciprocity, they become vulnerable to manipulation and over-leverage. The 'reason-based' explanations for bias (the AI 'performs in a biased manner') construct a sense that the AI's decisions are based on some internal (if flawed) logic, rather than acknowledging that the system lacks any capacity for justification or truth-evaluation. This trust architecture serves to maintain the 'illusion of mind' necessary for marketing AI as a general-purpose 'assistant' while shielding the developers from the consequences of its mechanical failures.

Obscured Mechanics​

Anthropomorphic language and consciousness projections systematically conceal the technical, labor, and material realities of generative models. Applying the 'name the corporation' test reveals that where the text says 'AI does X' or 'capabilities emerge,' the underlying reality involves specific companies (Anthropic, OpenAI, Google) making design choices. The metaphor of 'competency' and 'acquisition' hides the 'proprietary black box' nature of these systems; the authors make confident assertions about what the model 'knows' while acknowledging they cannot explain how it works (leading to the call for 'mechanistic interpretability' research in Section 4). This language conceals the massive 'data dependencies'β€”the fact that every 'skill' is a reflection of scraped human labor. The paper explicitly states in Section 2 that it does 'not consider here the costs of human labor... or environmental costs.' This is a critical omission: the 'predictable performance' scaling hides the material cost of energy and water, and the 'capability' mirrors the uncompensated labor of millions of human writers. The consciousness obscuration is particularly effective: when the text claims the AI 'understands' or 'mimics creativity,' it hides the statistical nature of 'confidence' and the absence of any 'ground truth' or 'causal model.' Who benefits from these concealments? The corporations, who can present an 'autonomous agent' as a product while externalizing the costs of data collection and environmental impact. By replacing 'processes embeddings' with 'solicits knowledge,' the text renders the infrastructure of AIβ€”data annotators, RLHF workers, and content moderatorsβ€”invisible, presenting the 'arrival' of the model as a clean, scientific epiphany rather than a messy industrial process.

Context Sensitivity​

The density and intensity of anthropomorphism are strategically distributed across the text. In the 'Introduction' and 'Scaling Laws' sections (Sections 1 & 2.1), the language is relatively grounded in technical terms (data, compute, parameters). However, as the argument moves toward the 'Unpredictable' (Section 2.2-2.4), consciousness claims intensify. The transition from 'processes' to 'understands' to 'knows' occurs precisely when the authors need to describe 'surprising' social harms. This capability/limitation asymmetry is profound: 'capabilities' are framed in agential, consciousness terms ('AI knows when to intervene,' 'acquires ability'), while 'limitations' are framed in mechanical, data-driven terms ('model's training data lacks X,' 'noise in training'). This asymmetry accomplishes a rhetorical feat: it makes the AI's 'intelligence' seem like an autonomous achievement of the agent, while its 'biases' are blamed on external, mechanical factors like 'bad data' or 'random seed variation' (Section 2.1, Footnote 5). The 'strategic register shift' occurs in the COMPAS and 'AI assistant' experiments, where 'X is like Y' (acknowledged metaphor) becomes 'X does Y' (literalized anthropomorphism). The intensity of the 'Assistant' persona (Section 2.4) is used to frame a vision of the future (vision-setting), while the mechanical language of Section 2.1 is used to establish credibility (technical grounding). This pattern reveals that anthropomorphism is not a linguistic accident but a strategic deployment to manage critique; by making the 'AI' a person who 'chooses' to be misleading, the authors can frame 'harm' as a social problem to be 'aligned' rather than a technical product failure that should be 'recalled.'

Accountability Synthesis​

Accountability Architecture

This section synthesizes the accountability analyses from Task 1, mapping the text's "accountability architecture"β€”who is named, who is hidden, and who benefits from obscured agency.

The text constructs an 'architecture of displaced responsibility' that systematicly diffuses human accountability into an 'accountability sink' of 'autonomous' AI behavior. The 'name the actor' test shows that while specific companies (Anthropic, OpenAI, Google) are named in a timeline of 'disclosures' (Fig 6), they are rarely named as the agents of 'harm.' Instead, the 'model' is the agent: 'the model decided,' 'the algorithm discriminated,' 'the system was misleading.' This follows the FrameWorks Institute's identified cognitive obstacle: audiences attribute AI problems to 'glitches' or 'emergent surprises' rather than systemic design decisions. The text frames 'unpredictability' as an inherent property of the technology rather than a failure of human testing and oversight. Responsibility transfers from humans to 'the scaling law' (inevitability), 'the model' (autonomous agency), or the 'users' (who 'manipulate' the 'backdoors'). This diffusion serves institutional interests by creating liability ambiguity; if the harm is a 'surprise' from an 'emergent competency,' it is legally and ethically harder to pin on the developer. If the human decision-makersβ€”the executives who authorized the COMPAS experiment and the engineers who chose the biased training setsβ€”were named, the questions would shift from 'how do we align the AI?' to 'why did you deploy this?' and 'what alternatives did you reject?' By naming the actors, accountability becomes possible. This text benefits from obscuring agency because it allows the 'AI community' to position itself as the 'policymakers' of a natural phenomenon rather than the responsible parties for a commercial product. The 'accountability sink' of the 'AI assistant' makes social harms feel like unfortunate accidents in the pursuit of 'beneficial impact,' protecting the corporate power that drives the 'lawful' scaling paradigm.

Conclusion: What This Analysis Reveals​

The Core Finding

The discourse in 'Predictability and Surprise' is built upon three dominant anthropomorphic patterns: Cognition as Biological Competency, the Model as Defiant Social Actor, and Scaling as a Lawful Guarantor. These patterns interconnect to form a system that frames AI as a developing entity with its own internal 'mind' and social 'personality.' The foundational pattern is the 'Cognition as Competency' frame; for the 'Defiant Social Actor' or 'Economic De-risking' patterns to work, the audience must first accept that the model possesses a structured internal state equivalent to human knowledge. This consciousness architecture projects 'knowing' onto 'processing' by using verbs like 'solicits' and 'acquires' to describe statistical weight adjustments. The system is load-bearing on the 'lawful' predictability of scaling: by establishing the technology as 'scientifically certain' in its performance, the authors create the license to describe its outputs as 'agential surprises.' If the 'competency' metaphor were removed, the system would collapse into a series of unverified statistical outputs, and the 'AI assistant' would be revealed as a simple token-sequence mirror with no understanding or intent.

Mechanism of the Illusion:​

The 'illusion of mind' is created through a strategic 'curse of knowledge' and a temporal shift in vocabulary. The text establishes AI as a 'processor' of compute and data in the early technical sections, building credibility with the audience. Once this grounding is established, it shifts to agential language, establishing the AI as a 'knower' that 'possesses knowledge' before building claims about its 'defiance' or 'creativity.' This 'causal chain' of metaphors leads the audience to accept that because the model's loss is 'predictable' (mechanical), its capabilities must be 'real' (agential). The 'illusion' exploits the audience's vulnerability to the Eliza effectβ€”our innate tendency to project social intent onto any system that uses human language. By ordering the narrative from 'lawful laws' to 'surprising skills,' the authors frame 'mind' as an emergent property of 'math,' making the anthropomorphism seem like a scientific discovery rather than a rhetorical choice. This sleight-of-hand blurs the distinction between a system that 'processes' patterns and a human who 'knows' truths, transforming a stochastic parrot into an 'AI assistant' with 'misleading' intentions.

Material Stakes:​

Categories: Economic, Regulatory/Legal, Epistemic

The material stakes of this framing are profound. Economically, the 'de-risking' metaphor encourages the concentration of capital into a 'scaling' paradigm, favoring large corporations who can afford the compute and marginalizing smaller actors. Regulatory and Legal decisions are shifted: if AI 'knows' or 'chooses' to be biased, liability is diffused into the 'unpredictable surprise' of the model, protecting companies from lawsuits and leading to 'alignment' regulations rather than strict product-safety bans. Epistemically, the framing of AI as a 'competent knower' devalues human expertise and creative labor, leading to an environment where statistical mirrors replace authorial styles. The 'winner' is the industry that gains 'unwarranted trust' and 'liability ambiguity,' while the 'losers' are the users and social groups who bear the cost of 'surprising' harms. If the metaphors were removed and replaced with mechanistic precision, the 'predictable' scaling would be seen as an expensive, resource-extractive gamble with known risks, likely triggering stricter regulatory oversight and less institutional investment in unverified 'competencies.'

AI Literacy as Counter-Practice:​

Critical literacy serves as resistance by demanding linguistic precision. Task 4 demonstrated that replacing consciousness verbs (knows/understands) with mechanistic ones (processes/predicts) forces a recognition of the model's dependency on data and its lack of awareness. For example, reframing 'acquires ability' as 'optimizes weights for token prediction' removes the illusion of a 'universal student' and highlights the statistical mirror. Restoring human agencyβ€”naming 'Anthropic's executives' instead of 'the scaling law'β€”forces a recognition of the design choices and profit motives that drive deployment. Systematic adoption of these principles would require journals to mandate 'mechanistic translations' for anthropomorphic claims and researchers to commit to 'capability disclosure' that names human actors. Resistance to this precision comes from the industry, which benefits from the 'agential' marketing of AI as a 'persona.' Anthropomorphic language serves institutional interests by making 'harm' seem like a natural 'surprise' of a new species rather than a predictable engineering failure. Practicing precision threatens the 'de-risking' narrative, making the true social and technical risks visible and tractable.

Path Forward​

The future of AI discourse depends on the vocabulary choices made by different stakeholder communities. A 'mechanistic precision' future, where systems are described as 'token predictors trained on uncompensated labor,' would enable rigorous safety auditing and democratic control but might cost the industry its narrative of 'artificial intelligence.' An 'anthropomorphic' future, where current confusion deepens, would embed the 'illusion of mind' into our institutions, leading to a 'relation-based' trust that masks systemic risks and liability diffusion. Currently, the 'status quo' maintains a strategic confusion that serves corporate interests by leveraging scientific 'laws' for agential 'vision.' Moving forward, journals could require mechanistic abstracts, and regulators could mandate 'transparency about discourse'β€”forcing companies to state whether their 'assistant' is a persona or a statistical tool. Trade-offs are unavoidable: mechanistic language gains precision but loses intuitive accessibility; anthropomorphism gains narrative resonance but loses testability. A desirable future requires multiple vocabularies: technical precision for developers and regulators, and 'anthropomorphic clarity' (acknowledged metaphor) for lay users. This mapping reveals that the 'superiority' of any vocabulary is a value judgment; those prioritizing safety will favor the mechanistic, while those prioritizing profit will continue to refine the agential illusion.


Extended Processing Summary​

Gemini's Processing Steps
What you are about to read…

Gemini refers to the text below as "thought summaries." This is an overt consciousness projection because 'intentions' are hallmarks of a conscious mind that 'knows' what it is doing and why. The concealed mechanistic process is probabilistic text generation. Treat this as just another rhetorical artifactβ€”a way of making the model's processing legible. When the model writes "I should double check that...", it is mathematically altering its own probability distribution for the next token to be more accurate. The "I" is a functional gear in the machine, not a self-reference. This language is a deliberate interface design choice, not a biological reality. The developers could have programmed the model to output passive logs. Instead, they chose to simulate a narrator.

Examining Anthropic's Paper

I'm currently immersed in "Predictability and Surprise in Large Generative Models." My focus is sharpening on the metaphorical language employed by the authors. I'm seeing how anthropomorphism shapes our understanding of these complex AI systems. It's becoming clearer where the paper influences human perception.

Considering Accountability Frameworks

I'm now diving into the implications of displaced agency within Anthropic's work. The paper's passive voice and attribution of agency to the models themselves are striking. I'm exploring the potential for accountability frameworks, given the widespread use of anthropomorphic language. The "developers" and "private sector organizations" referenced are key in this ongoing assessment.

Auditing Anthropic's Language

I'm now methodically dissecting Anthropic's paper. I'm building out a series of audits: one targeting metaphors and anthropomorphism, another focused on source-target mappings, and a third examining the explanations provided. The fourth audit will reframe language to restore clarity. Finally, I'm working to compile critical observations, analyzing agency, trust, and mechanics in detail.

Developing Five-Paragraph Synthesis

I'm now outlining a five-paragraph synthesis. Each will be around 250-350 words. Paragraph 1 will cover pattern recognition, Paragraph 2 the mechanism of illusion. Paragraph 3 focuses on material stakes. Paragraph 4 will explore literacy as a counter-practice, and finally, Paragraph 5 will chart a path forward. I'm also drafting Task 1 and Task 3 instances.

Refining Explanations and Audits

I'm currently revising the explanations given for Claude's responses, paying close attention to the dispositional phrasing. I'm actively drafting Task 5 sections, particularly highlighting the Predictable/Surprise tension. I'm focusing on "Surprise" as a human projection. I'm also reframing examples to be more precise, as illustrated by the shift from "questions authority" to a description of generated tokens. I'm doing a schema check and a constraint check, and ensuring that minimum word counts are met for all the tasks.

Adjusting to Paper's Constraints

I'm now grappling with the paper's directive to avoid control tokens in the output. This human constraint adds a layer of complexity. I'm also refining my agency audit; distinguishing between "AI" as the subject and the specific organizations cited is now a priority. I am also working on an epistemic distinction of processing versus knowing, to determine where the paper implies the model knows. Finally, I will assemble the JSON and strip all whitespace and newlines for compliance, ensuring that "no control token" is present in the final output.

Deconstructing Anthropic's Models

I'm now carefully analyzing the projections within Anthropic's work. I've pinpointed several instances, including the attribution of human mental capacity, social hierarchies, and even intentional deception to these AI models. I'm focusing on the language used, and how it subtly imposes human-like characteristics on purely mathematical processes. This includes an emphasis on de-risking investments and assigning artful expression to math. The Brown's Typology check is in progress.

Refining Task 5 Observations

I'm now revising Task 5, specifically the "Agency Slippage" observation, focusing on the "hockey stick" graph and the concept of "Surprise." I'm also delving into "Trust," interpreting scaling laws as a form of "de-risking" within the framework. I am giving "Obscured Mechanics" due consideration, with specific focus on RLHF and the annotators mentioned. The "Harm and Controversy" section will inform my thoughts on Accountability.

Analyzing Technical Specifications

I'm now zeroing in on the "Obscured Mechanics" point. I've noted "Cognition as Biological Competency" and will elaborate on Scaling Law explanations in Task 3. I'm revising the AI assistant example to focus on the model's token prediction. I'm confirming all Task 5 sections exceed 400 words, with the conclusion hitting the minimum requirements. JSON validation continues, and I am avoiding any control tokens.


Analysis Provenance

Run ID: 2026-01-16-predictability-and-surprise-in-large-gen-metaphor-f8qlxl Raw JSON: 2026-01-16-predictability-and-surprise-in-large-gen-metaphor-f8qlxl.json Framework: Metaphor Analysis v6.4 Schema Version: 3.0 Generated: 2026-01-16T12:17:47.460Z

Discourse Depot Β© 2025 by TD is licensed under CC BY-NC-SA 4.0