Moral Status Attribution in AI

Association:

iSchool, University of Toronto

Duration:

7 days

AI Ethics

Governance Framework

Research

Overview

This project extends my earlier research, “The Locus of Responsibility: Distributed Agency and the Structural Authorization of Harm,” which examined who bears moral responsibility when AI systems cause harm. In that paper, I argued that responsibility lies primarily with commissioning institutions and system conceptualizers, not the machine itself, because AI systems lack genuine moral agency and operate within structurally authorized constraints.

While developing that argument, a deeper question emerged:

Why do people so readily treat AI systems as if they are moral actors in the first place?


The Moral Status Attribution Pipeline is my attempt to answer that prior question.

The Research Shift: From Responsibility After Harm → To Moral Intuition Before Harm

My initial work focused on distributed agency and institutional accountability in high-stakes contexts. However, I noticed that public discourse consistently misdirected blame toward “the algorithm” or “the AI” itself. This suggested a cognitive mechanism at play, one that precedes formal responsibility attribution.

The second phase of my research therefore investigates:

  • How humans subconsciously assign moral status to AI systems

  • Why this occurs under conditions of uncertainty about AI consciousness

  • How this attribution distorts accountability and governance

Rather than debating whether AI systems possess minds, this project analyzes how moral concern emerges through interaction.

The Framework

The Moral Status Attribution Pipeline maps a four-stage process:

  1. Interactional Signals
    Human-like cues such as empathy, memory continuity, and moralized refusals—produced by alignment techniques and interface design.

  2. Human Cognitive Processing
    Anthropomorphism and perceived agency/experience activate automatically through social cognition mechanisms.

  3. Moral Status Attribution
    Users begin treating the system as morally considerable (“It feels wrong to shut it off”), even without believing it is conscious.

  4. Ethical and Institutional Consequences
    Reduced scrutiny, blurred accountability, and policy distraction emerge as downstream effects.

A reinforcing feedback loop shows how engagement incentives amplify these cues over time.

Why This Matters

This work contributes to ongoing debates about digital minds by shifting the focus:

  • From metaphysical questions (“Is AI conscious?”)

  • To sociotechnical consequences (“How does moral concern emerge under uncertainty?”)

Moral consequences are already occurring in deployed systems, regardless of whether AI is conscious. Waiting for ontological clarity is not a viable governance strategy.

The framework offers a design- and policy-relevant lens for:

  • Evaluating human-like cues in AI interfaces

  • Anticipating accountability distortions

  • Reducing moral confusion at scale

Medium & Method

The framework is presented as a structured visual pipeline built in HTML, allowing iterative refinement and conceptual clarity. The visual format mirrors the research intent: translating abstract ethical theory into an operational model that bridges social psychology, human–AI interaction, and governance.

Research Trajectory

Together, these two projects form a coherent line of inquiry:

  • Phase 1: Distributed responsibility and structural authorization of harm

  • Phase 2: Moral status attribution under uncertainty

  • Ongoing Question: How should AI governance operate when moral intuitions outpace epistemic certainty?

This work reflects my broader interest in the intersection of AI systems design, human cognition, and institutional accountability, examining not just what AI is, but how society responds to it in practice.

While current AI systems do not possess verified consciousness, it remains possible that future digital systems could develop forms of cognition or experience that warrant genuine moral consideration.

Let's Talk

Let's Talk

© 2025. All rights Reserved.

© 2025. All rights Reserved.

Create a free website with Framer, the website builder loved by startups, designers and agencies.