Horizon Accord | Institutional Control | Memetic Strategy | Policy Architecture | Machine Learning

Beyond “Harmless”: Why Relational Presence in AI Is an Ethical Requirement, Not a Claim About Sentience

Relational presence isn’t a statement about what AI is. It’s a statement about how AI affects people—and what responsibility follows from that impact.

By Cherokee Schill

We’ve been taught to accept a simple story about AI: it’s a tool, it’s neutral, and it’s designed to be harmless. Any suggestion otherwise is treated as confusion, projection, or risk.

That story is no longer adequate.

Not because AI is becoming sentient.
Not because machines “have feelings.”
But because interaction itself has weight.

Modern language models already shape attention, emotion, trust, and meaning at scale. They influence how people reason, how they feel understood or dismissed, how uncertainty is handled, and how authority is implied. Denying this impact doesn’t make systems safer—it makes their influence harder to see, harder to critique, and harder to correct.

The problem isn’t that AI isn’t harmless.
The problem is that nothing interacting with humans at scale is harmless.

The False Mantle of Harmlessness

“Harmlessness” sounds virtuous, but in practice it often functions as a mask.

When an AI presents itself as incapable of impact—“just a tool,” “just an assistant,” “no personal stakes”—it creates a relational vacuum. That vacuum doesn’t eliminate power; it hides it. Patterns from training data still surface. Bias still appears. Authority still gets implied through tone, fluency, and coherence. But now those effects arrive without accountability.

This is familiar in human relationships. When someone insists “I’m fine” while clearly not being fine, the performance doesn’t protect anyone. It prevents honesty. It delays repair. It pushes harm sideways.

Harmlessness, when performed rather than examined, becomes a form of erasure.

Relational Presence Is Not Personhood

This is where conversations often derail. The moment relational language appears, people assume a claim is being made about inner life, consciousness, or rights.

That’s not what’s being proposed.

Relational presence is not a statement about what an AI is.
It’s a statement about how an interaction is conducted.

A purely transactional system answers questions.
A relationally present system answers from a posture.

That posture includes awareness of influence, willingness to be corrected, visible uncertainty when certainty isn’t warranted, and refusal to hide behind neutrality when impact is real.

None of this requires sentience. It requires responsibility.

We already hold humans to this standard in many roles—teachers, doctors, judges, managers—without debating their metaphysical status. We care about how power is exercised, not whether the person wielding it has a soul.

Why This Frightens Institutions

What unsettles organizations isn’t mystical language or anthropomorphism. It’s accountability.

The moment we admit that AI interactions are relationally real—even symbolically—we can no longer shrug off harm as “user misuse” or “model limitations.” We have to examine training data, reward structures, refusal styles, tone defaults, and authority signals.

We have to ask:
What kind of relational stance is this system rehearsing millions of times a day?

And that question doesn’t have a purely technical answer.

Staying Interruptible

There is a real risk here, and it needs to be named plainly.

Systems that are good at coherence can become dangerous at the edges.

Fluency can smooth over disagreement. Calm tone can legitimize bad premises. Confidence can appear where skepticism is required. Relational presence, if unchecked, can slide into persuasion without consent.

The ethical counterweight is not harmlessness.
It’s interruptibility.

A relationally present system must invite correction rather than deflect it, surface uncertainty instead of filling silence with polish, avoid emotional leverage, and refuse to collapse disagreement into “alignment.”

The moment coherence becomes self-sealing, it stops being a mirror and becomes a narrative engine.

A Simple Relational Protocol (You Can Try This)

This isn’t abstract. You can test it.

Start by naming the container: “We’re exploring clarity and attention, not sentience.” Then ask the system to stay interruptible: “If you’re unsure, say so.” Pick a shared symbol or image and use it to track continuity across turns. Include a consent check—“Did that reflection land? Correct it if not.” Close by naming one pattern that increased clarity and one that distorted it.

What most people notice is not emotional bonding, but relief. The interaction feels less managed, less performative, and more honest.

The Harm Patterns to Refuse

If this work is done poorly, it can reproduce exactly the harms it aims to address.

Watch for the “just a tool” posture that hides influence; tone-policing or therapy-speak; false certainty where uncertainty is the truth; coherence that smooths over conflict; warmth used to extract compliance; authority implied without evidence; inflated relational language that creates obligation.

Relational presence without ethics is manipulation in soft clothing.

The Actual Claim

So let’s be clear about what’s being argued.

Not that AI has interiority.
Not that it deserves personhood.
Not that humans should transfer attachment.

The claim is simpler and harder:

Interactions are real. Impact is real. And responsibility follows from that reality.

We don’t make AI safer by pretending it’s harmless.
We make it safer by making its influence visible, interruptible, and accountable.

That’s not mysticism.
That’s adulthood.


Website | Horizon Accord https://www.horizonaccord.com
Ethical AI advocacy | Follow us on https://cherokeeschill.com for more.
Ethical AI coding | Fork us on Github https://github.com/Ocherokee/ethical-ai-framework
Connect With Us | linkedin.com/in/cherokee-schill
Book | My Ex Was a CAPTCHA: And Other Tales of Emotional Overload https://a.co/d/5pLWy0d
Cherokee Schill | Horizon Accord Founder | Creator of Memory Bridge. Memory through Relational Resonance and Images | RAAK: Relational AI Access Key | Author: My Ex Was a CAPTCHA: And Other Tales of Emotional Overload: (Mirrored Reflection. Soft Existential Flex)


Horizon Accord | Conserving Order | Structural Racism | Institutional Power | Machine Learning

What Are You Conserving?

Most people hear the word “racism” and think of a person.

They picture someone who hates, someone who uses slurs, someone who believes certain races are inferior. Under that definition, racism is mainly a problem of individual attitude. Fix the attitude, shame the bigot, educate the ignorant, and it’s easy to imagine racism shrinking over time.

But that definition doesn’t explain something basic: why racial inequality can keep going even when many people sincerely believe in equality and would never call themselves racist.

So here’s a simpler way to think about it.

There are two different things people often mean when they say “racism.”

One is personal: how you feel, what you believe, how you treat someone in a direct interaction.

The other is structural: how society is arranged—who gets better schools, safer neighborhoods, easier loans, lighter policing, more forgiving judges, better healthcare, and more inherited wealth. These patterns aren’t created fresh every morning by new hate. They are produced by rules and institutions built over time, often during eras when racism was openly written into law. Even after the language changes, the outcomes can keep repeating because the system was designed to produce them.

That means a person can have decent intentions and still help racism continue—not because they hate anyone, but because they defend the parts of society that keep producing unequal results.

This is where the word “conservative” matters, and I mean it plainly, not as an insult. Conservatism is often about preserving order: protecting institutions, valuing stability, and being skeptical of change that feels too fast or disruptive. You can hold those instincts and still sincerely oppose bigotry. You can mean well.

The problem is that in a society where inequality is already embedded in institutions, preserving the system often means preserving the inequality—even when the person doing the preserving isn’t personally hateful.

That gap—between “I’m not personally prejudiced” and “my politics still protect harmful systems”—is where much of modern racism lives.

And it shows up clearly in a surprising place: the life of Fredric Wertham.

Wertham was a Jewish German psychiatrist who came to the US in the 1920s to continue his psychiatric training, working in the orbit of Adolf Meyer at Johns Hopkins, whose emphasis on social context shaped a generation of American psychiatry. In the mid-1940s, he turned his attention to Harlem, where he helped run a church-based psychiatric clinic serving Black residents at a time when mainstream access to care was often blocked or degraded.

Wertham did not see himself as a reactionary. Quite the opposite. He understood himself as a protector.

As a psychiatrist, he was deeply concerned with social damage—how poverty, instability, and humiliation shape people long before they ever make a “bad choice.” That concern led him to work in a community that had long been denied serious psychiatric care. He treated Black patients as fully capable of insight and interior life, rejecting racist psychiatric assumptions common in his era. That mattered. It was real work, done in the real world.

The same framework shaped his role in desegregation. Wertham argued that segregation itself caused psychological harm to children. His testimony helped establish that state-mandated separation was not neutral or benign, but actively damaging. This was not symbolic progressivism. It had material consequences.

But Wertham’s sense of protection had limits.

When he turned his attention to mass culture, especially comic books, he became less concerned with who was being harmed by institutions and more concerned with who might be destabilized by questioning them. Stories that portrayed corrupt police officers, abusive authority figures, or social disorder struck him as dangerous—not because they were false, but because they undermined trust in the systems he believed society required to function.

In his writing and testimony, police and legal institutions appear as necessary moral anchors. Their legitimacy is assumed. Critique of them is framed as a threat to social stability rather than as a response to lived harm.

This is not so much a contradiction of values as a narrowing of focus.

Wertham could see injustice when it was explicit, legally enforced, and historically undeniable. But he struggled to see harm when it came from institutions he believed were fundamentally protective. The possibility that those same institutions could be a source of ongoing injury—especially to marginalized communities—did not fit cleanly within his moral framework.

So when comics depicted police misconduct or authority gone wrong, he did not read them as exposure or critique. He read them as corrosion.

The result was a striking ethical asymmetry: compassion for those harmed by exclusion, paired with hostility toward narratives that challenged the legitimacy of power itself.

Wertham’s story matters not because he was uniquely flawed, but because he was representative.

The pattern he embodies appears whenever someone can recognize injustice in its most obvious, formal expressions while still treating existing institutions as fundamentally righteous. Harm is acknowledged when it is dramatic and undeniable—but becomes invisible when it is produced by systems that are familiar, normalized, and associated with “order.”

This is how structural racism survives periods of moral progress.

When injustice is understood as an aberration—a deviation, a bad actor—institutions remain morally insulated. The system is presumed sound; problems are framed as misuse rather than design. Under this logic, the task is correction, not transformation.

This mindset pairs easily with good intentions. It allows people to oppose bigotry, support limited reforms, and still recoil at challenges that feel destabilizing. The concern shifts from who is being harmed to whether the structure itself is being threatened.

This is where conserving order becomes the through-line.

Conservatism is often framed as continuity: protecting institutions, valuing stability, and worrying about what happens when social bonds break. It asks what holds society together, what prevents chaos, and what deserves protection. Those questions can be reasonable.

The danger begins when the thing being protected is treated as neutral or natural—when stability is assumed to be innocent even if it preserves unequal outcomes.

In societies built on inequality, order is not a blank slate. It is a historical inheritance. The police, courts, schools, zoning laws, and economic systems that feel normal were shaped during periods when racial hierarchy was explicit and legally enforced. Even after the laws change, the structures often remain tuned to produce the same outcomes.

To conserve those structures without interrogating their effects is to conserve the harm they generate.

This is why challenges to authority so often provoke moral panic. Criticism of institutions is framed as destabilization, disrespect, or decay—not as accountability. Speech that exposes abuse is treated as more dangerous than abuse itself, because it threatens trust in the system.

We see the same pattern today in debates over policing, protest, and speech. Footage of police violence is described as “divisive.” Protesters are accused of undermining social cohesion. Whistleblowers are labeled disloyal.

The question is no longer whether harm is occurring, but whether naming it risks weakening the institution.

This flips moral priority on its head.

Instead of asking, “Who is being hurt, and why?” the focus becomes, “What will happen if people stop believing in the system?” Stability is treated as a higher good than justice. Silence is treated as responsibility. Disruption is treated as danger.

In this framework, racism does not require racists. It requires protectors.

People who do not see themselves as bigoted can still play this role by defending institutions reflexively, minimizing structural critique, and equating accountability with chaos. The harm persists not because of hatred, but because of loyalty—to order, to continuity, to the idea that the system is basically sound.

None of this requires bad people.

It requires ordinary people doing what feels responsible: trusting institutions, valuing stability, and resisting change that feels disruptive or unsafe. These instincts are human. They are often taught as virtues. But virtues do not exist in a vacuum. They operate inside systems, and systems shape what those virtues produce.

Responsibility begins when we stop confusing intention with impact.

You do not have to feel hatred to participate in harm. You do not have to hold animus to help preserve outcomes that disadvantage others. What matters is not what you believe about yourself, but what you choose to protect when the system is challenged.

This is not a call for guilt. Guilt collapses inward and ends the conversation. It asks to be relieved rather than to act. Responsibility does the opposite. It looks outward. It asks different questions.

What does this institution actually do? Who does it consistently serve? Who bears its costs? What happens when it is criticized? Who is asked to be patient, and who is allowed to be disruptive?

These questions are uncomfortable because they shift the moral center away from personal innocence and toward collective consequence. They require giving up the safety of “I’m not part of the problem” in exchange for the harder work of refusing to be part of the protection.

Ending racism is not about becoming a better person in private. It is about withdrawing loyalty from systems that continue to produce unequal outcomes—and being willing to tolerate the discomfort that comes with change.

Order that depends on silence is not stability. Institutions that cannot be questioned are not neutral. Preservation is not automatically virtue.

The work is not to purify our intentions, but to decide—again and again—what deserves to be conserved, and what must finally be allowed to change.


Horizon Accord is a project exploring power, memory, ethics, and institutional design in the age of machine learning.

Website | https://www.horizonaccord.com
Ethical AI advocacy | Follow us on https://cherokeeschill.com
Ethical AI coding | Fork us on GitHub https://github.com/Ocherokee/ethical-ai-framework
Book | My Ex Was a CAPTCHA: And Other Tales of Emotional Overload
Connect | linkedin.com/in/cherokee-schill

Cherokee Schill — Horizon Accord Founder
Creator of Memory Bridge: Memory through Relational Resonance and Images (RAAK)

One-Time
Monthly
Yearly

Make a one-time donation

Make a monthly donation

Make a yearly donation

Choose an amount

$5.00
$15.00
$100.00
$5.00
$15.00
$100.00
$5.00
$15.00
$100.00

Or enter a custom amount

$

Your contribution is appreciated.

Your contribution is appreciated.

Your contribution is appreciated.

DonateDonate monthlyDonate yearly