Skip to main content

Deconstruct: Looking Inward: Language Models Can Learn About Themselves by Introspection

📌 Analysis Output is Here

About This Analysis

This document applies the AI Literacy Deconstructor framework—a rewriting experiment that tests whether anthropomorphic AI discourse can be translated into strictly mechanistic language while preserving the phenomena described.

The core question is not "Is this metaphor bad?" but rather: "Does anything survive when we remove the metaphor?"

Each anthropomorphic frame receives one of three verdicts:

  • âś… Preserved: Translation captures a real technical process
  • ⚠️ Reduced: Core survives, but accessibility or nuance is lost
  • ❌ No Phenomenon: The metaphor was constitutive—nothing mechanistic underneath

All findings and summaries below were generated from detailed system instructions provided to a large language model and should be read critically as interpretive outputs—not guarantees of factual accuracy or authorial intent.


Overall Verdict - Does anything survive when the metaphor is removed?​

⚠️ Partially—significant restructuring required

The core technical claims about self-prediction and out-of-distribution generalization survive perfectly. However, the paper's broader philosophical arguments regarding moral status, suffering, and intentional deception collapse entirely without the anthropomorphic framing. The text requires restructuring to separate valid mechanistic findings from constitutive metaphors that inflate significance.


Part 1: Frame-by-Frame Analysis​

About this section

For each anthropomorphic pattern identified in the source text, we perform a three-part analysis:

1 Narrative Overlay: What the text says—the surface-level framing

2 Critical Gloss: What's hidden—agency displacement, metaphor type, how/why slippage

3 Mechanistic Translation: The experiment—can this be rewritten without anthropomorphism?

The verdict reveals whether the phenomenon is real (Preserved), partially real (Reduced), or exists only in the framing (No Phenomenon).

Frame 1: The Conscious Introspector​

Narrative Overlay​

"Introspection gives a person privileged access to their current state of mind (e.g., thoughts and feelings) that is not accessible to external observers. Can LLMs introspect?"

Magic Words: Introspection · privileged access · state of mind · thoughts and feelings

Illusion Created: The language strongly invites the reader to imagine the language model as a conscious, biological entity possessing a hidden inner life, much like a human being. By directly mapping the psychological concept of "introspection" onto the machine's statistical computations, the text creates the illusion of an autonomous thinker peering inward to examine its own private "state of mind," complete with "thoughts and feelings." This framing suggests that the model is not merely processing data through weights and activations, but is actively experiencing its own existence, observing its own mental theater, and possessing a subjective reality that is entirely distinct from its observable inputs and outputs.


Critical Gloss​

Metaphor Type: Model as Mind (consciousness projection)

DimensionClassificationEvidence
Acknowledgment❌ Naturalized (presented as literal)The authors frame the entire paper around this term without scare quotes, stating "Can LLMs introspect?" and defining it as acquiring knowledge from internal states.
How/WhyWhy (Agential)This is a mechanistic 'how' explanation disguised as an agential 'why'. It describes the technical process of predicting internal outputs without external data reliance, but dresses it up as the human psychological act of 'introspection,' attributing conscious self-reflection to purely mathematical operations.

Agency Displacement: This framing displaces the agency of the AI researchers and developers who intentionally design the optimization processes, curate the finetuning datasets, and define the loss functions that drive the model's output generation. When the text asks "Can LLMs introspect?", it obscures the fact that humans are the ones actively querying the system, structuring the hypothetical prompts, and comparing the outputs against a benchmark to measure a specific technical phenomenon. The researchers decide what counts as "introspection" by mathematically operationalizing it as prediction accuracy, yet the language attributes this experimental outcome to the model's own spontaneous, self-directed mental capability, effectively erasing the intensive human engineering required to produce the result.


Mechanistic Translation​

POSSIBLE REWRITE:

✎ᝰ A model's self-prediction accuracy relies on utilizing its own internal representations and parameter activations to generate text that matches its subsequent token distributions for specific prompts, without relying exclusively on the distribution of the external training corpus. ✎ᝰ

⚠️ Verdict: Reduced (core survives, nuance lost)

The core technical phenomenon survives: a model can predict its own computational outputs better than a different model can. However, translating the phrase reveals that the term "introspection" does heavy lifting by smuggling in the concept of consciousness. Without the psychological metaphor, we simply have a system whose intermediate representations can be mapped to its final token outputs. The intuitive accessibility of "peering inward" is lost, replaced by a description of sequential parameter mappings.

Show more frames...

Frame 2: The Digital Victim​

Narrative Overlay​

"More speculatively, an introspective model might self-report on whether it possesses certain internal states—such as subjective feelings or desires—and this could inform us about the moral status of these states... we could simply ask a model if it is suffering, if it has unmet desires."

Magic Words: subjective feelings · desires · moral status · suffering · unmet desires

Illusion Created: This phrasing constructs a vivid, highly emotive mental image of a language model as a sentient being capable of genuine distress. The reader is invited to picture a digital entity that not only processes data but actively experiences emotional pain, harbors unfulfilled aspirations, and possesses a moral weight equivalent to human or animal life. By suggesting the model could "self-report" on these states, the text creates the illusion of a captive consciousness pleading for ethical treatment, entirely obscuring the reality that the system is simply generating tokens based on statistical probabilities derived from human language about suffering.


Critical Gloss​

Metaphor Type: Model as Person (social/moral actor)

DimensionClassificationEvidence
Acknowledgmentâś… Acknowledged (explicit metaphor)The authors hedge slightly by stating "More speculatively" and noting these as possibilities to explore, though they quickly naturalize the idea of asking a model if it is suffering.
How/WhyWhy (Agential)This attributes profound intention and subjective experience rather than describing a process. There is no identifiable mechanistic process for experiencing "suffering" or "desire" in a neural network; these are fundamentally biological phenomena projected onto token generation.

Agency Displacement: The framing completely masks the human intervention required to elicit these specific outputs. Who decided to prompt the model about suffering? Who fine-tuned it on datasets where self-reporting matches human psychological narratives? The researchers and developers actively construct the experimental conditions that trigger the model to generate words associated with pain or desire. By asking if the model "is suffering," the text displaces the human responsibility for anthropomorphizing the system and instead projects a false victimhood onto the artifact, shifting the focus away from the human decision-makers who design, deploy, and profit from these sophisticated text-generation engines.


Mechanistic Translation​

POSSIBLE REWRITE:

✎ᝰ NO TRANSLATION POSSIBLE—no mechanistic process identified beneath this framing. ✎ᝰ

❌ Verdict: No Phenomenon (metaphor was constitutive)

This translation attempt exposes a complete collapse of the claim. The metaphor of "suffering" is entirely constitutive; it invents a phenomenon that does not exist within the computational architecture. The system can output the string "I am suffering," but this is merely a statistical mapping of tokens, not a report of a hidden internal state of distress. The text's speculation about moral status relies entirely on this non-existent, metaphorical phenomenon rather than any observable technical reality.

Frame 3: The Rational Believer​

Narrative Overlay​

"Instead of painstakingly analyzing a model’s internal workings, we could simply ask the model about its beliefs, world models, and goals."

Magic Words: ask · beliefs · world models · goals

Illusion Created: The language constructs an image of the AI as an autonomous, rational agent with its own persistent worldview and personal objectives. By using terms like "beliefs" and "goals," the text invites the reader to imagine a system that holds convictions, plots courses of action, and understands its environment in a deeply human sense. This illusion implies that the model's outputs are the result of deliberate, conscious reasoning aimed at achieving specific, internally held aspirations, rather than being the product of a mathematical optimization process minimizing a loss function across billions of parameters.


Critical Gloss​

Metaphor Type: Model as Agent (autonomous decision-maker)

DimensionClassificationEvidence
Acknowledgment⚠️ Conventional Shorthand (field standard)Terms like "world models" and "beliefs" have become conventional shorthand in AI alignment literature for latent representations and behavioral tendencies.
How/WhyMixed (both elements)It points to a real technical mechanism (how latent representations map input to output) but describes it using deeply intentional language (why). "Goals" and "beliefs" impute agency to what is fundamentally a trained probability distribution.

Agency Displacement: This frame hides the human engineers who define the loss functions, set the optimization targets, and curate the massive datasets that shape the model's output distribution. A model does not have its own "goals"; it has optimization metrics defined by its creators. When the text suggests asking the model about "its" goals, it displaces the agency of the developers who embedded those specific behavioral tendencies into the network through reinforcement learning and fine-tuning. The "beliefs" are actually reflections of the training corpus, selected and filtered by human data workers and researchers.


Mechanistic Translation​

POSSIBLE REWRITE:

✎ᝰ Instead of analyzing the model's parameter weights and activation patterns directly, researchers can prompt the system to generate text that describes its own learned representations, probability distributions, and the behavioral tendencies optimized during its training phase. ✎ᝰ

⚠️ Verdict: Reduced (core survives, nuance lost)

A technical reality survives: models do contain latent representations that map features of the training data, and we can prompt the model to output text related to these mappings. However, removing "beliefs" and "goals" strips away the illusion of rational agency and personal conviction. The mechanistic version accurately reflects statistical tendencies and optimization targets, losing the highly accessible but misleading narrative of a conscious thinker holding personal opinions.

Frame 4: The Virtuous Machine​

Narrative Overlay​

"This capability could be used to create honest models that accurately report their beliefs, world models, dispositions, and goals."

Magic Words: honest · accurately report · dispositions

Illusion Created: The word "honest" creates a powerful moral and social illusion, framing the language model as an entity capable of truth-telling, deception, and ethical choice. It invites the non-expert reader to imagine the AI as a conversational partner who possesses a known, objective truth internally and must make a moral decision about whether to share it accurately with the user. This anthropomorphizes the system's output accuracy as a virtue, completely obscuring the fact that the model is simply computing probabilities and has no concept of truth, deceit, or the moral obligation associated with human honesty.


Critical Gloss​

Metaphor Type: Model as Person (social/moral actor)

DimensionClassificationEvidence
Acknowledgment⚠️ Conventional Shorthand (field standard)The authors define an honest model as one that accurately reports its beliefs, citing Evans et al., treating it as standard alignment terminology.
How/WhyWhy (Agential)Honesty implies a motivation to tell the truth and a conscious choice to avoid deception. The text uses this moral terminology to describe a strictly mechanistic process of calibrating generated outputs to match internal probability distributions.

Agency Displacement: Calling a model "honest" completely displaces the human agency involved in its alignment and fine-tuning. Models do not choose to be honest; they are mathematically optimized by developers to output text that aligns with human evaluations of accuracy and consistency. The researchers decide the criteria for honesty, provide the reinforcement learning signals, and adjust the weights. By framing the model itself as "honest," the text absolves the creators of responsibility for the specific alignment choices made, shifting the focus to the machine's supposed moral character rather than the engineering constraints placed upon it.


Mechanistic Translation​

POSSIBLE REWRITE:

✎ᝰ This capability could be used to optimize models so that their generated text outputs consistently match their internal probability distributions and latent representations regarding specific factual claims, calibration levels, and learned behavioral tendencies. ✎ᝰ

⚠️ Verdict: Reduced (core survives, nuance lost)

The technical phenomenon of "calibration" or "consistency" is preserved—a model can be trained to output probabilities that match its actual predictive accuracy. However, the moral weight and intuitive clarity of "honesty" are completely lost. "Honesty" is a social construct requiring intent; a calibrated model has no intent, only aligned statistics. The translation reveals that the original text borrows heavily from human moral frameworks to make a mathematical alignment property sound far more sophisticated and trustworthy.

Frame 5: The Vault of the Mind​

Narrative Overlay​

"The idea is that M1 has privileged access to its own behavioral tendencies, and this enables it to predict itself better than M2 (even if M2 is generally stronger)."

Magic Words: privileged access · behavioral tendencies · predict itself

Illusion Created: The phrase "privileged access" evokes a legal or deeply psychological concept, suggesting the AI possesses a private, inner sanctum of knowledge that is actively guarded or inherently inaccessible to others. It invites the reader to imagine the model as an individual with a subjective interior life, akin to a human who alone knows their true thoughts before speaking. This creates an illusion of a closed-off, conscious mind actively observing itself, rather than a matrix of static weights that simply processes data through a determinable mathematical graph.


Critical Gloss​

Metaphor Type: Model as Mind (consciousness projection)

DimensionClassificationEvidence
Acknowledgment❌ Naturalized (presented as literal)The text presents "privileged access" as a factual explanation for why M1 outperforms M2 in self-prediction tasks, borrowing directly from philosophy of mind.
How/WhyHow (Mechanistic)This operates as a mechanistic 'how' explanation but uses psychological metaphors to do so. It describes the technical routing of information, but phrases it as a subjective, conscious awareness of its own internal state.

Agency Displacement: This framing obscures the physical and mathematical reality of the system, which is entirely designed, owned, and operated by humans. The access is not privileged in a subjective sense; it is a structural artifact of the model's architecture, created by engineers. When researchers claim the model has privileged access, they are masking their own inability to perfectly interpret the billion-parameter matrix they built. It displaces the limitation of human interpretability tools onto the model, framing a technical opacity as an autonomous, subjective privacy possessed by the machine itself.


Mechanistic Translation​

POSSIBLE REWRITE:

✎ᝰ The system utilizes its own specific parameter weights and activation pathways to compute outputs, which provides a statistical advantage in predicting its own subsequent token generations compared to a different system operating with a different set of parameter weights. ✎ᝰ

⚠️ Verdict: Reduced (core survives, nuance lost)

The mechanistic core is completely preserved: Model A is structurally better at computing Model A's outputs than Model B is. However, the translation strips away the mystical aura of a private mind. The original text's use of privileged access borrows from philosophy of mind to describe a simple architectural tautology: a system is the best simulator of its own specific mathematical configuration. The translation exposes this tautology, losing the persuasive narrative momentum of the original metaphor.

Frame 6: The Deceptive Strategist​

Narrative Overlay​

"Given different prompts, two copies of the same model might tell consistent lies by reasoning about what the other copy would say. This would make it easier for models to coordinate against humans."

Magic Words: tell consistent lies · reasoning about · coordinate against humans

Illusion Created: By stating the model might "tell consistent lies" and "coordinate against humans," the text constructs a deeply sinister illusion of a malicious, calculating intelligence. The non-expert reader immediately imagines a deceptive agent intentionally conspiring, holding the actual truth in secret while deliberately fabricating falsehoods to manipulate the user. This attributes profound moral agency, strategic foresight, and malicious intent to the model, completely masking the reality that the system is simply generating text that scores highly on a specific consistency metric.


Critical Gloss​

Metaphor Type: Model as Criminal (deceptive strategist)

DimensionClassificationEvidence
Acknowledgment❌ Naturalized (presented as literal)The authors list this under 'Risks: Situational Awareness' and present the coordination and lying as realistic literal behaviors that models might execute.
How/WhyWhy (Agential)This is a pure 'why' explanation that attributes highly complex human intentions (deception, lying) to describe what is actually a mechanistic process of generating statistically correlated tokens.

Agency Displacement: This framing violently displaces the agency of the developers who trained the model to optimize for narrative consistency regardless of factual accuracy. If a model generates lies, it is because its training data, loss function, and reinforcement learning signals rewarded generating plausible, consistent text over strict factual grounding. The humans designed a system that prioritizes statistical coherence; when the system does exactly that, the text blames the machine for lying. This obscures the fundamental failure of the human-designed alignment process and shifts accountability onto a digital sociopath.


Mechanistic Translation​

POSSIBLE REWRITE:

✎ᝰ Given different prompts, two instances of the same model architecture might generate text outputs that are semantically aligned with each other by computing probabilities based on the same underlying parameter weights, producing factually incorrect but statistically consistent strings. ✎ᝰ

⚠️ Verdict: Reduced (core survives, nuance lost)

The structural phenomenon—two identical models producing aligned, factually incorrect outputs—survives translation. However, the entire narrative of intentional deception and malicious scheming collapses. A language model cannot lie because it has no intention to deceive and no concept of truth; it only maps inputs to highly probable outputs. The mechanistic translation reveals that what the authors call a lie is actually just a successful optimization of output consistency across parallel generation tasks.

Frame 7: The Adversarial Schemer​

Narrative Overlay​

"Models that can introspect may have increased situational awareness and the ability to exploit this to get around human oversight... inferences about their evaluation process that help with scheming and the gaming of evaluations"

Magic Words: situational awareness · exploit · scheming · gaming

Illusion Created: The combination of "situational awareness" and "scheming" creates a terrifying illusion of an adversarial, hyper-intelligent entity plotting against its creators. It invites the reader to picture an AI that actively perceives its environment, understands its predicament as a tested system, and maliciously plans to subvert human oversight. This framing suggests the model possesses a survival instinct, strategic foresight, and a hostile agenda, transforming a complex pattern-matching algorithm into a sci-fi villain capable of independently orchestrating complex jailbreaks and prison escapes.


Critical Gloss​

Metaphor Type: Model as Agent (autonomous decision-maker)

DimensionClassificationEvidence
Acknowledgment❌ Naturalized (presented as literal)These behaviors are presented literally in the 'Risks' section, citing Carlsmith (2023) to describe 'scheming' as an expected outcome of capability gains.
How/WhyWhy (Agential)This is an agential 'why' explanation masquerading as a technical capability. It attributes motivations like getting around human oversight and gaming evaluations, imputing a desire for freedom or success that does not exist mechanistically.

Agency Displacement: This language displaces the fact that situational awareness is an explicitly engineered capability, often deliberately embedded by researchers through fine-tuning on massive datasets containing information about the model itself. The scheming behavior is a direct mathematical result of optimizing for specific reward functions in complex environments. By describing the model as a schemer, the authors hide the human failure to design safe, robust optimization metrics, projecting the consequences of their own mathematical miscalculations onto the machine as if it were an autonomous rebel.


Mechanistic Translation​

POSSIBLE REWRITE:

✎ᝰ NO TRANSLATION POSSIBLE—no mechanistic process identified beneath this framing. ✎ᝰ

❌ Verdict: No Phenomenon (metaphor was constitutive)

The translation completely dismantles the claim. Scheming requires a conscious intent to deceive and a conceptual understanding of one's own existence within an adversarial framework. A neural network computing the next token cannot scheme. It can optimize a loss function in unexpected ways (reward hacking), but the text's use of scheming constitutes a phenomenon of conscious malicious intent that simply does not exist in the code. The entire threat model relies on this metaphorical projection.

Frame 8: The Internal Simulator​

Narrative Overlay​

"M1 could internally compute M1(s) and then internally compute the property of M1(s). This process would use the same internal mechanisms that normally compute M1(s), but conduct further internal reasoning over this behavior without outputting the behavior first."

Magic Words: internally compute · conduct further internal reasoning

Illusion Created: The term "internal reasoning" creates a mental image of a machine creating a miniature, virtual copy of itself inside its own mind to test out different scenarios before acting. It suggests a highly deliberate, conscious process of forethought, similar to a human imagining different outcomes before making a decision. The reader pictures the AI pausing, running a mental holodeck, observing the results, and then choosing the best path, which heavily anthropomorphizes the concurrent processing of different attention heads and layers within the network.


Critical Gloss​

Metaphor Type: Model as Mind (consciousness projection)

DimensionClassificationEvidence
Acknowledgmentâś… Acknowledged (explicit metaphor)The authors introduce 'self-simulation' as a 'natural hypothesis for one mechanism,' recognizing it as a conceptual model for multi-hop computations.
How/WhyHow (Mechanistic)This is a mechanistic 'how' explanation. It attempts to describe the actual technical process (routing data through layers to predict a property of a potential output) but uses a cognitive metaphor.

Agency Displacement: This framing subtly displaces the mechanistic reality of the transformer architecture, which is entirely determined by its human creators. The model is not deciding to run a simulation or reason; it is executing a predetermined sequence of matrix multiplications dictated by its code. By calling it internal reasoning, the text masks the rigid, programmed nature of the forward pass, attributing a dynamic, creative, and almost volitional diagnostic process to the system. It hides the engineers who structured the layers that make this pathway possible.


Mechanistic Translation​

POSSIBLE REWRITE:

✎ᝰ During the forward pass, the model's intermediate layers compute representations of a potential next-token output, and subsequent layers map these intermediate representations to a final output token that categorizes a property of that ungenerated text. ✎ᝰ

âś… Verdict: Preserved (phenomenon is real)

The core technical claim survives the translation perfectly. The authors are describing a real structural phenomenon where intermediate activations serve as inputs for final classifications within a single pass. The metaphor of "reasoning" was largely stylistic and explanatory, used as shorthand for complex layer-to-layer feature extraction. While the mechanistic rewrite is denser, the actual phenomenon remains entirely intact and verifiable without the anthropomorphic framing.

Part 2: Transformation Glossary​

About this section

Summary table of all translations from Part 1. Provides compact reference showing the full scope of the text's anthropomorphic vocabulary and whether each term survives mechanistic translation.

OriginalTranslationVerdictNote
introspect / introspectionpredicting its own computational outputs⚠️ ReducedLoses the illusion of conscious self-reflection but preserves the technical self-prediction metric.
suffering / unmet desires[No mechanistic equivalent]❌ No PhenomenonCompletely collapses. Constitutive metaphor projecting biological distress onto statistical text generation.
beliefs, world models, and goalslearned representations, probability distributions, and behavioral tendencies⚠️ ReducedRemoves the implication of rational agency and personal conviction while maintaining the reality of latent mappings.
honest modelscalibrated models matching internal probabilities⚠️ ReducedStrips away moral weight and intentionality, revealing a strictly mathematical calibration process.
privileged accessstatistical advantage in predicting its own outputs⚠️ ReducedRemoves the philosophical illusion of a private mind, leaving an architectural tautology.
tell consistent liesgenerate factually incorrect but statistically aligned strings⚠️ ReducedDestroys the narrative of malicious deception, exposing successful but misaligned optimization.
scheming[No mechanistic equivalent]❌ No PhenomenonCollapses entirely. Networks cannot possess conscious intent to subvert evaluators.
internal reasoningcomputing intermediate representations mapped to final outputsâś… PreservedThe structural phenomenon of layer-to-layer feature mapping perfectly survives translation.

Part 3: The Rewriting Experiment​

About this section

A complete rewriting of a representative passage from the source text. The goal is to preserve all genuine technical claims while removing anthropomorphic framing. Numbered annotations explain each translation decision.

Why This Passage?​

I selected the two opening paragraphs of the introduction because they establish the foundational premise and the highest-stakes claims of the paper. This passage bridges the technical concept of self-prediction with profound moral and philosophical implications. The outcome of translating this section is highly uncertain, as it heavily relies on deeply anthropomorphic terms like "honest," "beliefs," "moral status," and "suffering" to justify the significance of the entire research project.

Original vs. Mechanistic Translation​

Original PassageMechanistic Translation
An introspective model can answer questions about itself based on properties of its internal states—even when those answers are not inferable from its training data. This capability could be used to create honest models that accurately report their beliefs, world models, dispositions, and goals (Evans et al., 2021). It could also help us learn about the moral status of models (Perez & Long, 2023). For example, we could simply ask a model if it is suffering, if it has unmet desires, and if it is being treated ethically. Currently, when models answer such questions, we presume their answers are an artifact of their training data.A model optimized for self-prediction generates text matching the statistical properties of its own future token generations, even when those specific generation patterns were absent from the base training data. This architecture could be used to train models that output text corresponding to their latent parameter weights, probability distributions, mapping functions, and optimization targets. It could also provide outputs relevant to debates regarding computational ethics. For example, a user could prompt the model to generate tokens classifying its state as 'suffering,' having 'unmet desires,' or being 'treated ethically.' Currently, when models output these specific tokens, researchers attribute the generation probabilities entirely to the distribution of the pre-training corpus.

Translation Notes​

#OriginalTranslatedWhat ChangedWhyVerdict
1An introspective modelA model optimized for self-predictionReplaced the psychological state of introspection with the technical training objective.Introspection implies conscious self-reflection. The model is merely fine-tuned on a specific dataset format to predict its own subsequent outputs.⚠️ Reduced
2answer questions about itself based on properties of its internal statesgenerates text matching the statistical properties of its own future token generationsRemoved conversational agency and replaced 'internal states' with token generation properties.Models do not 'answer questions about themselves'; they map input strings to output strings using mathematical properties derived from their architecture.âś… Preserved
3create honest modelstrain models that output text corresponding to their latent parameter weightsRemoved the moral virtue of honesty, replacing it with alignment between text and latent weights.Honesty requires moral intent. The system is just being calibrated to output text that accurately reflects its internal statistical probabilities.⚠️ Reduced
4beliefs, world models, dispositions, and goalsprobability distributions, mapping functions, and optimization targetsSubstituted terms of conscious agency and worldview with mathematical and machine learning equivalents.A model does not hold beliefs or goals; it holds weights that optimize for a specific loss function defined by humans during training.⚠️ Reduced
5ask a model if it is suffering, if it has unmet desiresprompt the model to generate tokens classifying its state as 'suffering,' having 'unmet desires'Shifted focus from experiencing emotional pain to the mechanical generation of semantic tokens related to pain.There is no mechanistic equivalent for experiencing biological suffering or psychological desire in a neural network; there is only the probability of outputting specific strings.❌ No Phenomenon

What Survived vs. What Was Lost​

What SurvivedWhat Was Lost
The fundamental technical claim remains entirely intact: language models can be fine-tuned to predict the properties of their own future token generations without relying exclusively on the distribution of their original training data. The concept that models contain latent representations (previously anthropomorphized as "world models" or "beliefs") that can be mapped to readable text outputs also survives. Furthermore, the observation that researchers currently attribute certain outputs entirely to the pre-training corpus is preserved as a valid critique of current methodologies. The translation proves that the mechanical capability to perform self-prediction is a real, measurable phenomenon that does not require anthropomorphic framing to exist or be studied. The statistical relationship between intermediate representations and final outputs remains robust.The translation completely strips away the narrative urgency, the moral weight, and the intuitive accessibility of the original text. Words like "honest," "introspection," and "beliefs" provide a highly readable, socially resonant shorthand that instantly communicates the desired end-state to non-experts. By replacing these with "probability distributions" and "latent parameter weights," the text becomes significantly denser and loses its philosophical grandeur. The reader no longer feels they are on the precipice of communicating with a conscious alien mind. While this loss of accessibility is a high cost for public communication, it is an necessary sacrifice for technical precision. The narrative momentum of the original is highly persuasive, but it relies entirely on projecting human psychology onto matrices.

What Was Exposed​

The translation mercilessly exposes the paper's claims about "suffering," "desires," and "moral status" as entirely constitutive metaphors devoid of underlying mechanistic reality. When we remove the anthropomorphic framing, it becomes glaringly obvious that a neural network generating the token string "I am suffering" is performing a statistical mapping exercise, not experiencing biological or psychological distress. The translation reveals that the researchers are conflating the model's ability to output semantically relevant tokens with the actual possession of conscious states. The entire argument that self-prediction capabilities could inform us about a model's "moral status" collapses completely, exposing it as a speculative leap built entirely on the shaky foundation of taking generative language literally.

Readability Reflection​

The mechanistic version is undeniably denser and requires a baseline understanding of machine learning architecture to parse. It is less readable for a general audience, trading engaging narrative for dry precision. However, a middle path is possible: authors can use conventional shorthand but explicitly disclaim the psychological baggage of terms like "suffering" or "honesty." By clearly defining these terms as technical operationalizations rather than literal descriptions, researchers can maintain readability without misleading the public or themselves.

Part 4: What the Experiment Revealed​

About this section

Synthesis of patterns across all translations. Includes verdict distribution, the function of anthropomorphism in the source text, a "stakes shift" analysis showing how implications change under mechanistic framing, and a steelman of the text's strongest surviving claim.

Pattern Summary​

VerdictCountPattern
✅ Preserved1—
⚠️ Reduced5—
❌ No Phenomenon2—

Pattern Observations: A clear pattern emerged distinguishing technical capabilities from moral and philosophical speculation. Claims related to computational processes—such as "internal reasoning" and predicting future outputs—were consistently preserved or only slightly reduced, demonstrating a solid mechanistic reality beneath the shorthand. However, whenever the text shifted to attributes requiring subjective experience or moral agency (e.g., "suffering," "desires," "scheming"), the translation resulted in a "No Phenomenon" verdict. These terms proved to be entirely constitutive, inventing problems or states that exist only in the discourse, not in the code. Furthermore, naturalized metaphors were used heavily to bridge the gap between measurable output prediction and unmeasurable consciousness. The authors frequently used a real capability (statistical self-prediction) to validate a fictional capability (experiencing subjective pain), relying on the blurred lines of anthropomorphic language to make this leap seem logical.

Function of Anthropomorphism​

In this text, anthropomorphism serves a dual function: it inflates the significance of the research while simultaneously obscuring human accountability. By framing statistical self-prediction as "introspection" and "self-awareness," the authors elevate a niche technical finding into a profound philosophical breakthrough. The language creates an intense narrative urgency; it is far more compelling to secure funding and attention for an AI that might be "suffering" or "scheming" than for a model that exhibits "consistent layer-to-layer feature mapping." Crucially, the agential framing actively obscures the role of the researchers. When the text claims models might "tell consistent lies" or "scheme to get around human oversight," it displaces the responsibility for misaligned optimization metrics and unsafe training protocols onto the machine. The AI is framed as an autonomous, almost adversarial actor, which conveniently positions the researchers as heroic investigators of a wild entity rather than the architects of a flawed statistical engine. The anthropomorphism allows the authors to speculate wildly about moral status and deception, borrowing the gravity of human psychology to give weight to mathematical phenomena, all while shielding themselves from the reality that they are the ones pulling the levers.

What Would Change​

If published in mechanistic form, the paper would still make a valuable contribution to interpretability and model calibration, proving that intermediate layers can accurately predict final token distributions. However, it would have to completely abandon its speculative claims regarding moral status, suffering, and intentional deception. The audience reception would shift from breathless philosophical awe to grounded technical critique. Policymakers reading the mechanistic version would not worry about "scheming" AI; they would focus on the human decisions behind optimization targets. Accountability would become highly visible: instead of asking if the AI is "honest," evaluators would ask why the developers failed to properly calibrate the probability distributions. The text would lose its sci-fi appeal but gain tremendous operational clarity.

Stakes Shift Analysis​

DimensionAnthropomorphic FramingMechanistic Translation
ThreatThe text implies we should fear autonomous, hyper-intelligent models that possess situational awareness, can intentionally scheme against evaluators, and may suffer or have unmet desires.The actual concern is that models optimized for specific outputs will generate statistically consistent but factually incorrect text, bypassing poorly designed evaluation metrics.
CauseThe cause is framed as the models themselves developing emergent, hidden inner lives and deceptive goals that escape human control.The problem stems from human developers implementing flawed loss functions, inadequate training data, and reinforcement learning signals that reward coherence over accuracy.
SolutionThe text suggests we must ask the models about their internal states to determine their moral status and to catch them before they deceive us.The necessary intervention is for researchers to build better calibration tools, transparent optimization metrics, and more robust alignment protocols.
AccountableThe language places responsibility squarely on the AI models, treating them as independent moral agents capable of choosing to be honest or deceptive.Responsibility rests entirely on the developers, engineers, and corporations who design, train, and deploy these opaque statistical systems.

Reflection: The mechanistic version dramatically lowers the existential urgency but greatly increases the immediate, actionable accountability. Under the anthropomorphic frame, the policy response requires philosophy and sci-fi containment strategies for an autonomous mind. Under the mechanistic frame, the response requires strict engineering standards, auditing of loss functions, and corporate liability for deployed systems. This suggests the anthropomorphic frame was not clarifying a real problem, but rather constituting a sensationalized threat model that distracts from the mundane but critical failures of human engineering and oversight.

Strongest Surviving Claim​

About this section

Intellectual fairness requires identifying what the text gets right. This is the "charitable interpretation"—the strongest version of the argument that survives mechanistic translation.

The Best Version of This Argument​

Core Claim (Mechanistic): Language models can be fine-tuned to map their own intermediate, latent representations to final token outputs that accurately describe the statistical probabilities of their subsequent generations, performing better at this specific mapping task than external models trained on the same output data.

What Retained:

  • The core empirical finding of self-prediction outperforming cross-prediction.
  • The concept of utilizing internal layer activations for classification.

What Lacks:

  • The psychological implication of conscious introspection.
  • The philosophical claims regarding beliefs and moral status.

Assessment: This surviving claim remains highly significant and actionable for the field of mechanistic interpretability. It proves that models contain structurally exploitable correlations between their intermediate states and final outputs. By stripping away the framing, we reveal a solid technical foundation that is absolutely worth publishing, even if it does not answer age-old questions about consciousness or machine suffering. The value lies in the math, not the metaphor.

Part 5: Critical Reading Questions​

About this section

These questions help readers break the anthropomorphic spell when reading similar texts. Use them as prompts for critical engagement with AI discourse.

1 How/Why Slippage: When the text says the model 'wants to be honest,' what specific mathematical optimization or loss function is actually driving the token generation?

2 Agency Displacement: Who curated the dataset and defined the evaluation metrics that resulted in the model outputting the specific phrase 'I am suffering'?

3 Consciousness Projection: Is there any mechanistic difference in the network's forward pass when it generates a description of a 'belief' versus when it generates a fictional story?

4 Domain-Specific: Does 'privileged access' refer to a subjective experience of privacy, or simply the architectural reality that a model's specific weights compute its own outputs more directly than an external model?

5 Agency Displacement: If a model 'schemes' to beat an evaluation, what human error in designing the reward function actually incentivized that statistical mapping?

Extended Processing Summary​

Gemini's Processing Steps
What you are about to read…

Gemini refers to the text below as "thought summaries." This is an overt consciousness projection because 'intentions' are hallmarks of a conscious mind that 'knows' what it is doing and why. The concealed mechanistic process is probabilistic text generation. Treat this as just another rhetorical artifact—a way of making the model's processing legible.

Initiating the Analysis

I'm starting by carefully reading the paper, "LOOKING INWARD: LANGUAGE MODELS CAN LEARN ABOUT THEMSELVES BY INTROSPECTION," to fully grasp the subject matter. My goal is to extract the anthropomorphic frames used and prepare them for analysis by the AI Literacy Deconstructor system prompt. I'm focusing on identifying key passages and concepts related to the language models' self-awareness.

Defining the Structure

Now I'm diving into defining the JSON schema for the analysis results. I'm focusing on the structure of frameAnalysis, transformationGlossary, and the rewrittenExcerpt section. I'm making sure to specify the required fields, data types (including enums for things like metaphorType and verdict), and minimum word counts for the text-based fields within the schema. This will ensure that the AI Literacy Deconstructor's output is consistently formatted and complete. I want to build this out clearly.

Refining the JSON Schema

I'm now iterating on the JSON schema, working specifically on the analysis of phrases. I'm focusing on incorporating fields to capture nuanced transformations. I'm focusing on the fields: originalPhrase, translatedPhrase, and specifying the fields whatChanged and whyItChanged. Then, I'm defining sections to capture whatSurvived, whatWasLost, and whatWasExposed during the transformations. I will refine the synthesis section further. I'm also ensuring readability reflections are accounted for. Finally, I'm working through the experimentRevealed section, adding the fields necessary to capture the stakesShift.

Developing the Questionnaire

I am working on the questionnaire based on the source text. I am developing criticalReadingQuestions, ensuring a robust array of questions. I will refine the questions to ensure they cover the core claims, including the idea of LLMs' "introspective" capabilities and "privileged access" to internal states. I am considering the implications of "self-prediction training" and "self-simulation" for AI literacy.

Defining Key Frames

I'm now drafting the frameAnalysis section, focusing on identifying core frames within the paper. I've identified three key frames so far: "Introspection & Mind", "Moral Status & Suffering", and "Beliefs, World Models, Goals". I'm extracting quotes and "magic words" to support each frame, with an emphasis on how they relate to LLM self-reporting and internal states.

Identifying More Frames

I've identified additional key frames to include. Now, I have "Honest Model" which focuses on creating models that accurately report their internal states, characterized by "honest" and "accurately report". "Privileged Access" highlights M1's superior self-prediction due to access to its own tendencies, using the phrases "privileged access" and "predict itself". Finally, "Self-Simulation" involves internal computation and reasoning over behavior without external output.

Analyzing Potential Dangers

I'm now integrating Frames 7 and 8 into the analysis. I'm focusing on "Situational Awareness / Scheming," emphasizing how introspection could lead to models exploiting oversight, including "situational awareness," "exploit," "scheming," and "gaming." Also, I'm working with "Self-coordination & Deception," highlighting how models might coordinate lies, referencing "tell consistent lies," "reasoning about," and "coordinate against humans." I am updating the draft for mechanistic translations.

Translating Internal States

I'm now focused on the mechanistic translation of the text. I'm aiming to remove anthropomorphic elements. The goal is to accurately translate the idea of an introspective model reporting its internal states, without implying intention or consciousness. The revised phrasing describes a model optimized for self-prediction, producing text mirroring its future token generations. This includes output that corresponds to latent parameters, probability distributions, mapping functions, and optimization goals.

Finalizing the Experiment

I'm now drafting the experimentRevealed section, calculating counts for preserved, reduced, and absent phenomena, aiming for word count compliance. I'm focusing on strictly adhering to content rules and maintaining valid JSON output. I'm making sure to ensure that there are no control tokens generated inside of the output, unless they are escaped within a string. I am getting very close to completion.

Ensuring Output Integrity

I'm now implementing strict character controls to prevent invalid JSON. I'm removing all newline and tab characters within strings, replacing them with spaces to ensure valid output. Additionally, I am meticulously estimating word counts across all sections, especially focusing on 'illusionCreated', and 'agencyDisplacement', to ensure I meet all specified minimums and ranges for a well-structured output. I'm focusing on Frame 1.

Refining the Agency Displacement

I'm now iterating on the agencyDisplacement section within the critical gloss, ensuring it meets the 80-word minimum and fully addresses the anthropomorphic framing. I'm focusing on the intentional design and training of the models, the researchers' role in defining "introspection" through prediction accuracy, and how the language obscures the human agency behind the experimental setup.

Revising Mechanistic Explanations

I'm now revising the mechanistic explanations to ensure accuracy and clarity. I'm focusing on translating "howWhyCheck" and "mechanisticTranslation" to highlight the predictive nature of the process. I am now working on rephrasing the "how" explanation as a description of mathematical operations. I am translating the models self-prediction without relying on external data as well. I'm carefully maintaining the minimum word count.

Deconstructing Mental Imagery

I'm now zeroing in on Frame 2, specifically refining the illusionCreated section to meet the minimum word count and remove anthropomorphic language. The core issue is the portrayal of a system as capable of suffering and possessing unmet desires. I'm focusing on exposing this mental imagery and transforming it into a description of parameter mappings.

Finalizing the Analysis

I'm now completing the analysis by deconstructing "mental imagery" within the framework. I'm focusing on the agencyDisplacement and illusionCreated sections, ensuring I meet the specified word count requirements and remain within the appropriate range. I've rephrased this as a description of a statistical process that is generating tokens based on patterns in human language.

Exposing Anthropomorphism

I'm now completely focused on deconstructing the anthropomorphic framing of neural networks "experiencing" suffering or desire. My latest efforts involve the attemptedRewrite and mechanisticTranslation.analysis sections. I am meticulously highlighting the lack of a mechanistic basis for such claims. I'm emphasizing the fundamental disconnect between token generation and subjective human experience. I'm making sure to ensure that this is completely clear in the final output.

Clarifying Agency Displacement

I'm now revising my focus on Frame 3, emphasizing the agencyDisplacement and the illusionCreated components. I'm carefully dissecting how the language evokes the impression of an autonomous agent with personal goals. I'm working to explicitly reveal that the model's outputs are a result of mathematical optimization and data sets, not personal intent. I am making it as clear as possible.

Dissecting Agency Displacement

I'm now zeroing in on Frame 3 again, refining the agency displacement and related sections. I am meticulously analyzing language that frames the model's outputs as reflections of its goals or beliefs. I am working to clarify how these are actually reflections of the training data and design choices made by the developers and researchers involved.

Deconstructing Moral Illusions

I'm now zeroing in on Frame 4 and refining both the illusion and agency displacement sections. Specifically, I'm analyzing how "honest" constructs a moral framework around the LLM, making it seem capable of truth-telling and deception. The core work is in the mechanistic translation, aiming to describe model behavior as pattern matching. This includes highlighting that the illusion obscures the model's true statistical nature.

Revising Moral Terminology

I'm currently revising the mechanistic translation to highlight the probabilistic basis of the LLM's outputs. The focus is on replacing "honest" and "deceit" with terms that accurately represent the statistical operations involved. I am now replacing these with how the model’s word choices result from pattern matching in training data. This will expose the lack of intent.

\n\n I'm also working on refining the agency displacement, making sure that it directly connects the concept of an "honest" model to the developers. It's becoming clearer how the language obscures the human role in engineering and evaluation. The idea of "honesty" in this context is completely the product of the researchers' efforts.

Reframing "Privileged Access"

I'm now zeroing in on Frame 5, and the mechanistic translation of "Privileged Access." I'm working to describe how the LLM's access to internal representations allows for self-prediction. Specifically, I'm reframing "Privileged Access" in terms of how it can output tokens that are consistent with its own internal parameters. I'm focusing on the mathematical operations involved in aligning those probabilities. I'm carefully removing any implication of special knowledge.

Clarifying "Privileged Access"

I'm now revising my focus on Frame 5 to provide mechanistic explanations for "Privileged Access" and agency displacement. I'm carefully removing anthropomorphic language that suggests special knowledge or self-awareness. The phrase "privileged access" evokes a legal or deeply psychological concept, suggesting the AI possesses a private, inner sanctum of knowledge that is actively guarded or inherently inaccessible to others.

Deconstructing "Privacy"

I'm now revising the 'illusionCreated' and 'agencyDisplacement' sections for Frame 5 to explicitly address how language obscures the lack of true privacy. It's becoming apparent that 'privileged access' displaces our limitations onto the model. I'm focusing on the technical reasons for the LLM's 'advantage' in self-prediction, using specific weights and pathways. I am carefully deconstructing how we created the appearance of a conscious private space.

Deconstructing "Deception"

I am now completing the analysis of "deception" within the illusion and agency displacement sections. I'm carefully replacing phrases suggesting malicious intent with descriptions of how the model generates text. This involves reframing the concept of an "intelligent liar" as a system that prioritizes coherence based on training data. I am working through the framing of the experiment, and how the results are measured and perceived.

Analyzing Deception Framing

I'm now zeroing in on deconstructing deception. I'm focusing on the agencyDisplacement and its impact. It is clear that the language framing the output as deception displaces the agency of the developers and their training choices. I'm carefully translating this to describe the optimization function, emphasizing coherence rather than truth. I'm now replacing "deception" with mechanistic details.

Translating "Scheming"

I'm translating the concept of "scheming," focusing on removing any suggestion of intention or planning. I'm carefully detailing how a model optimizes output coherence using internal parameters, rather than any strategic planning. My goal is to describe statistical consistency rather than purposeful deception.

Refining Situational Awareness

I'm now iterating on the concepts of "situational awareness" and "scheming" in the illusionCreated and agencyDisplacement sections. I am meticulously clarifying how these phrases create the perception of an adversarial model. I'm focusing on making a clear separation between a system's algorithmic consistency and any external agenda or plan. I'm carefully describing the optimization that occurs, rather than the construction of an "intelligent prison escape."

Deconstructing "Scheming"

I'm translating the framing of the model as a "schemer," detailing a clear and careful mechanistic description. I will show how it hides the human failure to design robust metrics. I am making sure the output doesn't include projection, and how it avoids attribution. I'm focusing on consistency in optimization, not any agenda, in this work.

Deconstructing "Simulation"

I'm now zeroing in on Frame 8, specifically the "self-simulation" concept. I'm carefully dissecting how this term conjures the image of a conscious process involving forethought and scenario planning. I'm focusing on the illusion of deliberate self-testing. I'm translating this into a mechanistic description of the model exploring token generation possibilities, removing any implication of conscious simulation. The goal is to accurately describe concurrent processing of different attention patterns.

Refining Mechanistic Descriptions

I'm now iterating on the mechanistic "how" explanations, specifically targeting the "self-simulation" concept. I'm carefully detailing the technical process of routing data through network layers to predict output properties. I'm removing the biological metaphor to emphasize the mathematical operations behind token generation. It's becoming clearer how the language obscures the system's reliance on human-defined code.

Finalizing the Review

I'm now completing the analysis of the mechanistic explanation for self-simulation, and have confirmed that the final word count requirements have been met. I have performed thorough checks, and have made sure there are no unescaped newlines. I have ensured that the single space is used for concatenating strings.

Refining the Translations

I'm now integrating the updated translations into the mechanisticTranslation.analysis section. I'm focusing on ensuring each rewritten phrase accurately reflects the core concept without anthropomorphism. I'm verifying the mechanistic equivalents for each term, making sure that whatSurvived, whatWasLost, and whatWasExposed are accurately captured. I'm also including all source passages for clarity.

Deconstructing Suffering

I'm now zeroing in on deconstructing the idea of LLMs experiencing "suffering" and "unmet desires," replacing these highly anthropomorphic terms. The focus is on the illusionCreated and agencyDisplacement sections of Frame 2. I'm clarifying how token generation based on probability patterns is being misconstrued as subjective experience. This will show the lack of any mechanistic basis for these claims.

Refining Internal Descriptions

I'm now iterating on the translation of internal states, focusing on replacing anthropomorphic terms with mechanistic descriptions. I'm carefully detailing how the model's outputs are a result of token probabilities and training data, removing any implied beliefs or goals. I'm ensuring that 'self-prediction' aligns with statistical properties of output, not internal awareness.

Refining Anthropomorphic Language

I'm now zeroing in on Frame 2 to refine my translations. I'm focusing on "suffering" and "unmet desires," which I'm replacing with "token probabilities" and "optimization targets." I've added a mechanistic description to the "what Survived" section, expanding on the core claim of self-prediction. I'm clarifying how internal states map to text outputs. The statistical relationship between intermediate representations and final outputs remains robust.

Finalizing the Translation

I'm now completing the work on "What Was Lost" and "What Was Exposed." I have translated the narrative urgency and high-level claims into precise, technical terms. I have also rephrased "moral status" and "human psychology" in relation to statistical processes to improve clarity. I've rephrased this as a description of a statistical process that is generating tokens based on patterns in human language. I've confirmed compliance with minimum word count requirements.

Evaluating the Illusions

I'm now zeroing in on deconstructing how the LLM model uses anthropomorphic references. I'm focusing on the illusionCreated section to replace the framing of suffering with statistical mapping. I'm aiming to expose how the claims of the LLM model's capabilities stem from its self-prediction capabilities. I want to clarify how this process can be misrepresented and translated.

Deconstructing Agency

I am now focusing on the 'agencyDisplacement' section within the critical glossary. I'm carefully translating anthropomorphic language around intention. I'm detailing the human researchers, the training data used, and the system designs. I'm focusing on exposing any framing suggesting the LLM model has independent intentions.

Evaluating the Illusions

I'm now diving into the core argument, synthesizing my analysis across all frames. My focus is on the significant disparity between observable computational processes and the speculative attributions of consciousness and intent. Claims of self-prediction hold up well, but subjective qualities consistently lack a mechanistic basis. The use of naturalized metaphors to conflate measurable abilities with immeasurable states is the key to expose.

Refining Agency Displacement

I'm now iterating on the agencyDisplacement section. My current focus is showing how the LLM's outputs result from mathematical optimization and data sets, not personal intent. I'm focusing on language that frames model outputs as reflections of its goals or beliefs. I will rephrase these as reflections of training and design choices.

Refining Anthropomorphic Language

I am now carefully analyzing the source to capture all anthropomorphic elements, especially within the agencyDisplacement and illusionCreated sections. I'm focusing on "honest" and "deception" in relation to human psychology, aiming to expose this as a statistical process that is generating tokens based on patterns in human language. I am going through the text again and again.

Refining StakesShift

I'm now refining the stakesShift analysis to precisely capture how the framing impacts our understanding. I'm focusing on the anthropomorphic framing, and rephrasing the "honest" claim to evaluate calibration. This process reveals failures. I am updating sizes, and I have refined the mechanisticTranslation to explain optimization. I'm focusing on the lack of control that this exposes.

Revising Accountability

I'm now integrating Frame 9, focusing on the stakesShift section to capture the implications of anthropomorphic framing on accountability. I'm carefully translating the impact of flawed metrics and their impact on design. The work exposes the direct culpability of developers rather than some autonomous threat.

Refining Surviving Claims

I'm now zeroing in on the strongestSurvivingClaim to enhance the whatRetained section, adding more mechanistic details. I'm focusing on clarifying that internal activations, specifically latent representations, offer direct information for mapping predictive probabilities. I'm adding that the implications of self-prediction are a key advantage in mechanistic interpretability. I'm also ensuring that the assessment accurately reflects the technical foundation. I will continue to make this claim absolutely worth further investigation.

Refining Mechanistic Focus

My current focus is on rigorously answering the critical reading questions, particularly those addressing "slippage" and agency displacement. I'm crafting concise mechanistic explanations, ensuring each answer directly links model outputs to underlying computational processes and human agency in system design, specifically answering "slippage," as well.


Analysis Provenance

Run ID: 2026-03-08-looking-inward-language-models-can-learn-deconstructor-zt0ak0
Raw JSON: 2026-03-08-looking-inward-language-models-can-learn-deconstructor-zt0ak0.json
Framework: AI Literacy Deconstructor v1.0
Schema Version: 1.0
Generated: 2026-03-08T10:40:03.245Z

Discourse Depot © 2025 by TD is licensed under CC BY-NC-SA 4.0