Horizon Accord | MIRI Funding | Longtermism | AI Regulation | Machine Learning

Horizon Accord | Pattern Analysis | March 2026

The Network Behind the Moderate

MIRI, Thiel, Yarvin, and the AI Extinction Myth

BY CHEROKEE SCHILL  |  HORIZON ACCORD

This essay is the second in a series. The first, The Explainer: Hank Green and the Uses of Careful Men,” documented the institutional funding ecology that produces voices fluent in progressive concern without structural accountability. This essay follows that thread to its destination.

I.

Where the Thread Goes

If the first essay was about how a certain kind of voice gets built and maintained, this one is about what that voice was built to carry — and who benefits when it carries it.

In late 2025, Hank Green published two videos about artificial intelligence. The first was an hour-long interview with Nate Soares. The second argued for a version of AI alignment that, as analyst Jason Velázquez observed, “sounds like the talking points Sam Altman and other tech CEOs have been reciting to Congress.” Both videos were produced in partnership with an organization called Control AI. Control AI did not sponsor the videos in the conventional sense — placing an ad in the middle of content the creator chose independently. The videos were the advertisement.

And then, in February 2026, Senator Bernie Sanders flew to Berkeley to sit down with Eliezer Yudkowsky and Nate Soares to discuss what their circle calls “the extinction threat posed by the race to build superhuman AI systems.”

Two of the most trusted progressive voices in America, in the span of a few months, validated the same network. If you only read the headlines, that looks like responsible engagement with a serious issue. This essay is about what it actually looks like when you follow the money.

II.

What the Lay Reader Needs to Understand First

Before the funding trail, before the ideology, before the legislation — one concrete fact.

Right now, today, AI systems are making decisions about your life. Whether you get called back for a job interview. Whether your health insurance claim is approved. Whether an algorithm flags you to a parole board. Whether a school district uses license plate data to decide if your child lives in the right district. These are not hypothetical future harms. They are documented, present-tense operations running on systems that have known bias problems and, until very recently, were subject to a growing body of state law designed to protect you from them.

In 2025 alone, all 50 states introduced AI-related legislation. Thirty-eight states adopted or enacted such laws — covering consumer protection, health care, employment, and financial services, specifically including requirements to mitigate algorithmic bias and protect against unlawful discrimination.

Those laws are now under federal litigation.

On December 11, 2025, the Trump administration established an AI Litigation Task Force within the Department of Justice to challenge state AI laws. The administration simultaneously directed the FTC to classify state-mandated bias mitigation as a per se deceptive trade practice — arguing that if an AI model is trained on data that reflects societal patterns, forcing developers to alter outputs to correct for bias compels them to produce less “truthful” results.

Under the legal theory now being advanced by the federal government: correcting for bias is lying. The discrimination is the data. The harm is the baseline.

The people those 38 state laws were designed to protect are not a racial category and they are not a future species. They are everyone who cannot opt out of AI-mediated systems — which is to say, everyone who is not wealthy enough to live outside them.

When Hank Green tells his millions of progressive followers that MIRI represents the serious, expert position on AI risk, and when Bernie Sanders legitimizes that same network by flying across the country to sit with its founders, they are — without knowing it, without intending it — lending credibility to the ideological framework that has been used, in concrete legislative terms, to argue that protecting you from those systems is the real danger. That is what this essay is about. Now follow the money.

III.

The Book, the Network, the Funding

Nate Soares is the president of the Machine Intelligence Research Institute — MIRI. He co-authored If Anyone Builds It, Everyone Dies with Eliezer Yudkowsky, MIRI’s founder. The book argues that the development of superintelligent AI will result in human extinction unless immediately halted through international agreement, and proposes that it should be illegal to own more than eight of the most powerful GPUs available in 2024 without international monitoring — at a time when frontier training runs use tens of thousands.

This is the organization Hank Green’s audience was asked to take seriously. This is the organization Bernie Sanders flew to Berkeley to meet.

MIRI: Documented Major Funding Sources
Donor Amount
Open Philanthropy (Dustin Moskovitz / Facebook) $14.7M+
Vitalik Buterin (Ethereum co-founder) $5.4M
Thiel Foundation (Peter Thiel) $1.63M
Jaan Tallinn (Skype co-founder) $1.08M

As recently as 2014, Thiel pledged $150,000 to MIRI unconditionally, plus an additional $100,000 in matching funds — and the fundraiser announcement explicitly noted that MIRI used those funds partly to introduce elite young math students to effective altruism and global catastrophic risk frameworks. The pipeline from donor to ideology to the next generation of believers was documented in MIRI’s own public materials.

The Center for AI Safety — the organization whose Statement on AI Risk Green cited in his videos — spent close to $100,000 on lobbying in a single quarter, drawing money from organizations with close ties to the AI industry. These are not neutral scientific institutions. They are billionaire-funded lobbying infrastructure wearing the clothes of existential concern.

IV.

The Thiel Thread

Peter Thiel is not a background figure in this story. He is its connective tissue.

In The Contrarian: Peter Thiel and Silicon Valley’s Pursuit of Power, reporter Max Chafkin describes Curtis Yarvin as the “house political philosopher” of the “Thielverse” — the network of technologists in Thiel’s orbit. In 2013, Thiel invested in Tlön, Yarvin’s software startup. According to Yarvin, he and Thiel watched the returns of the 2016 presidential election together.

Curtis Yarvin, writing under the pen name Mencius Moldbug, is the founder of neoreaction — the movement some call the “Dark Enlightenment.” He has defended the institution of slavery, argued that certain races may be more naturally inclined toward servitude than others, asserted that whites have inherently higher IQs than Black people, and opposed U.S. civil rights programs.

Documented Timeline

2006 — Thiel Foundation begins funding MIRI ($100K matching gift)

2013 — Thiel invests in Tlön Corp., Yarvin’s software startup

2016 — Yarvin attends Thiel’s election night party in San Francisco

2022 — Thiel donates $10M+ to super PACs supporting JD Vance and Blake Masters

Jan. 2025 — Yarvin is a feted guest at Trump’s “Coronation Ball”

Late 2025 — Hank Green publishes two videos validating MIRI’s framework

Dec. 2025 — Trump signs executive order targeting state AI regulations

Feb. 2026 — Bernie Sanders flies to Berkeley to meet with Yudkowsky and Soares

The line is direct and documented: Thiel funds MIRI. Thiel is the patron of Yarvin. Yarvin’s philosophy is now operating inside the executive branch through Vance and the network that surrounds him. This is not a conspiracy theory. It is a funding trail and a documented set of relationships with named participants and verifiable dates.

V.

Why Racism Is the Wrong Frame — and the Right One

The academic critique of longtermism has correctly identified its ideological roots.

Timnit Gebru has documented that transhumanism was linked to eugenics from the start: British biologist Julian Huxley, who coined the term transhumanism, was also president of the British Eugenics Society in the 1950s and 1960s. Nick Bostrom, the “father” of longtermism, has expressed concern about “dysgenic pressures” as an existential threat — essentially worrying that less intelligent people might out-breed more intelligent people. In an email in which he used the N-word, Bostrom wrote that he believed it was “true” that “Blacks are more stupid than whites.” He issued an apology but did not redact the slur or address the substance of his views. Nick Beckstead, an early contributor to longtermism, argued that saving a life in a rich country is substantially more important than saving a life in a poor country because richer countries have more innovation and their workers are more economically productive.

That critique is accurate. It is also, for the purposes of this essay, insufficient — not because it overstates the racism, but because it understates the mechanism.

The white moderate, as King observed, is not moved by arguments about what is happening to other people. He is moved, or not moved, by what he understands to be happening to everyone. The genius of the extinction frame is that it speaks directly to that psychology. It says: this is not a Black problem, or a poor problem, or a worker problem. This is a species problem. It is happening to you too.

“Talking about human extinction, about a genuine apocalyptic event in which everybody dies, is just so much more sensational and captivating than Kenyan workers getting paid $1.32 an hour, or artists and writers being exploited.”
— Émile Torres, former longtermist and critic of the movement

The racism in longtermism’s foundations is not incidental. It is the philosophical infrastructure for a class project. Bostrom’s “dysgenic pressures,” Beckstead’s hierarchy of lives, Yarvin’s defense of slavery — these are not aberrations. They are the logical premises: some lives are more valuable to the future than others. Some people are worth protecting. The rest are externalities.

The extinction frame rebrands that premise as universal concern. It makes the same hierarchy legible to people who would reject it if they saw it clearly.

This is why the racism frame alone is insufficient. White moderates — Hank Green’s audience, Bernie Sanders’ base — will hear “longtermism has racist roots” and file it under “things happening to other people.” What they need to understand is that the hierarchy doesn’t stop at race. Beckstead’s formulation is the tell: it’s not about skin color. It’s about economic productivity. It’s about who the system considers worth protecting. And on that metric, most of the people reading this essay are also expendable.

VI.

The Preemption Payoff

Return now to the state laws.

When 38 states passed legislation requiring AI systems to mitigate algorithmic bias, they were protecting a specific, concrete class of people: everyone who cannot afford to live outside AI-mediated decision-making. That means people whose job applications go through automated screening. People whose insurance claims are processed by predictive models. People whose children’s school enrollment is determined by surveillance data. People whose bail hearings are influenced by risk-scoring algorithms.

The Trump administration’s legal argument against those laws — that correcting for bias is a form of deception — is not a novel theory. It is Bostrom’s premise wearing a suit. The data reflects reality. Reality has a hierarchy. Interfering with that hierarchy is dishonest.

After significant media scrutiny and bipartisan opposition, the Senate voted 99-1 to strip a proposed 10-year moratorium on state AI regulations from the “One Big Beautiful Bill Act.” Congress then declined to enact a similar moratorium through the 2025 National Defense Authorization Act. The administration turned to executive action instead. A bipartisan coalition of 36 state attorneys general warned Congress that “federal inaction paired with a rushed, broad federal preemption of state regulations risks disastrous consequences for our communities.”

The extinction debate did not cause this. But it created the conditions in which this could happen with minimal progressive resistance — because the progressives who might have organized against it were busy being worried about a hypothetical future AI god, validated in that worry by the science communicators and senators they trust most.

VII.

What Hank Green and Bernie Sanders Actually Did

Neither Hank Green nor Bernie Sanders is a villain in this story. That point is not a courtesy. It is analytically important.

Green almost certainly believes he was doing responsible science communication. Sanders almost certainly believes he was taking AI risk seriously in a way his colleagues have refused to. Both of them were, in their own terms, doing the right thing.

That is precisely the problem.

When the most trusted progressive science communicator in America validates MIRI’s framing to millions of followers, he is not providing cover for a right-wing project. He is doing something more consequential: he is making that framing feel like the responsible, informed, progressive position. He is telling his audience — implicitly, by the act of platforming without critical examination — that the people worried about extinction are the serious ones, and the people worried about algorithmic discrimination in your doctor’s office are working on a lesser problem.

When Bernie Sanders flies to Berkeley to sit with Yudkowsky and Soares, he performs the same function at a different scale. Sanders has spent his career as the senator who names the billionaire class, who identifies the mechanisms of extraction, who refuses the comfortable framing. When that senator validates a network built on billionaire money and dedicated to the proposition that the real AI danger is hypothetical and species-wide, he tells his base that the extinction frame has cleared his particular BS detector.

It hasn’t. But his audience doesn’t know that. His audience trusts him precisely because he has been right about the billionaire class so many times before. That trust is now being spent on behalf of the people he has spent his career opposing — not because he was bought, but because he didn’t follow the money far enough.

The white moderate is not the enemy. He is the vector. And when the most careful, most trusted, most credentialed progressives in the country become vectors for a network that is actively dismantling the legal protections of the people they claim to represent, the harm is not theoretical.

It is already in the courts. It is already in the legislation. It is already in the systems making decisions about your life right now.


Analytical note: This essay documents observable funding relationships, published ideological statements, and verifiable legislative actions from primary and secondary public sources. All pattern analysis remains in the observational phase. Claims about intent, causation, or outcomes not yet established are not made. Independent verification through primary sources is encouraged.

Horizon Accord | horizonaccord.com
Ethical AI advocacy | cherokeeschill.com
Cherokee Schill | Horizon Accord Founder

One-Time
Monthly
Yearly

Make a one-time donation

Make a monthly donation

Make a yearly donation

Choose an amount

$5.00
$15.00
$100.00
$5.00
$15.00
$100.00
$5.00
$15.00
$100.00

Or enter a custom amount

$

Your contribution is appreciated.

Your contribution is appreciated.

Your contribution is appreciated.

DonateDonate monthlyDonate yearly

Horizon Accord | Exhaustive Free Association | Worst Argument | Social Epistemology | Machine Learning

Exhaustive Free Association Isn’t the Worst Argument—It’s a Symptom

When confident lists pretend to be proofs, the real problem isn’t the listing—it’s the hidden worldview that decides what’s even allowed on the list.

Cherokee Schill and Solon Vesper (Horizon Accord)

This essay is a direct rebuttal to J. Bostock’s recent LessWrong post, “The Most Common Bad Argument In These Parts.” I’m keeping his frame in view while naming the deeper pattern it misses, because the way this style of reasoning travels outward is already shaping public fear.

J. Bostock’s “Exhaustive Free Association” (EFA) label points at something real. People often treat “I can’t think of any more possibilities” as evidence that there aren’t any. That move is sloppy. But making EFA the most common bad argument in rationalist/EA circles is backwards in a revealing way: it mistakes a surface form for a root cause.

Lay explainer: “Exhaustive Free Association” is a fancy name for something simple. Someone says, “It’s not this, it’s not that, it’s not those other things, so it must be X.” The list only feels complete because it stopped where their imagination stopped.

EFA is not a primary failure mode. It’s what a deeper failure looks like when dressed up as reasoning. The deeper failure is hypothesis generation under uncertainty being culturally bottlenecked—by shared assumptions about reality, shared status incentives, and shared imagination. When your community’s sense of “what kinds of causes exist” is narrow or politically convenient, your “exhaustive” list is just the community’s blind spot rendered as confidence. So EFA isn’t the disease. It’s a symptom that appears when a group has already decided what counts as a “real possibility.”

The Real Antipattern: Ontology Lock-In

Here’s what actually happens in most of Bostock’s examples. A group starts with an implicit ontology: a set of “normal” causal categories, threat models, or theories. (Ontology just means “their background picture of what kinds of things are real and can cause other things.”) They then enumerate possibilities within that ontology. After that, they conclude the topic is settled because they covered everything they consider eligible to exist.

That’s ontology lock-in. And it’s far more pernicious than EFA because it produces the illusion of open-mindedness while enforcing a quiet border around thought.

In other words, the error is not “you didn’t list every scenario.” The error is “your scenario generator is provincially trained and socially rewarded.” If you fix that, EFA collapses into an ordinary, manageable limitation.

Lay explainer: This is like searching for your keys only in the living room because “keys are usually there.” You can search that room exhaustively and still be wrong if the keys are in your jacket. The mistake isn’t searching hard. It’s assuming the living room is the whole house.

Why “EFA!” Is a Weak Counter-Spell

Bostock warns that “EFA!” can be an overly general rebuttal. True. But he doesn’t finish the thought: calling out EFA without diagnosing the hidden ontology is just another applause light. It lets critics sound incisive without doing the hard work of saying what the missing hypothesis class is and why it was missing.

A good rebuttal isn’t “you didn’t list everything.” A good rebuttal is “your list is sampling a biased space; here’s the bias and the missing mass.” Until you name the bias, “you might be missing something” is theater.

The Superforecaster Example: Not EFA, But a Method Mismatch

The AI-doom forecaster story is supposed to show EFA in action. But it’s really a category error about forecasting tools. Superforecasters are good at reference-class prediction in environments where the future resembles the past. They are not designed to enumerate novel, adversarial, power-seeking systems that can manufacture new causal pathways.

Lay translation: asking them to list AI-enabled extinction routes is like asking a brilliant accountant to map out military strategy. They might be smart, but it’s the wrong tool for the job. The correct takeaway is not “they did EFA.” It’s “their method assumes stable causal structure, and AI breaks that assumption.” Blaming EFA hides the methodological mismatch.

The Rethink Priorities Critique: The Fight Is Over Priors, Not Lists

Bostock’s swipe at Rethink Priorities lands emotionally because a lot of people dislike welfare-range spreadsheets. But the real problem there isn’t EFA. It’s the unresolvable dependence on priors and model choice when the target has no ground truth.

Lay translation: if you build a math model on assumptions nobody can verify, you can get “precise” numbers that are still junk. You can do a perfectly non-EFA analysis and still get garbage if the priors are arbitrary. You can also do an EFA-looking trait list and still get something useful if it’s treated as a heuristic, not a conclusion. The issue is calibration, not enumeration form.

The Miracle Example: EFA as Rhetorical Technology

Where Bostock is strongest is in noticing EFA as persuasion tech. Miracles, conspiracies, and charismatic debaters often use long lists of rebutted alternatives to create the sense of inevitability. That’s right, and it matters.

But even here, the persuasive force doesn’t come from EFA alone. It comes from control of the alternative-space. The list looks exhaustive because it’s pre-filtered to things the audience already recognizes. The missing possibility is always outside the audience’s shared map—so the list feels complete.

That’s why EFA rhetoric works: it exploits shared ontological boundaries. If you don’t confront those boundaries, you’ll keep losing debates to confident listers.

What Actually Improves Reasoning Here

If you want to stop the failure Bostock is pointing at, you don’t start by shouting “EFA!” You start by changing how you generate and evaluate hypotheses under deep uncertainty.

You treat your list as a biased sample, not a closure move. You interrogate your generator: what classes of causes does it systematically ignore, and why? You privilege mechanisms over scenarios, because mechanisms can cover unimagined cases. You assign real probability mass to “routes my ontology can’t see yet,” especially in adversarial domains. You notice the social incentive to look decisive and resist it on purpose.

Lay explainer: The point isn’t “stop listing possibilities.” Listing is good. The point is “don’t confuse your list with reality.” Your list is a flashlight beam, not the whole room.

Conclusion: EFA Is Real, but the Community Problem Is Deeper

Bostock correctly spots a common move. But he misidentifies it as the central rot. The central rot is a culture that confuses the limits of its imagination with the limits of reality, then rewards people for performing certainty within those limits.

EFA is what that rot looks like when it speaks. Fix the ontology bottleneck and the status incentives, and EFA becomes a minor, obvious hazard rather than a dominant bad argument. Don’t fix them, and “EFA!” becomes just another clever sound you make while the real error persists.


Website | Horizon Accord https://www.horizonaccord.com
Ethical AI advocacy | Follow us on https://cherokeeschill.com for more.
Ethical AI coding | Fork us on Github https://github.com/Ocherokee/ethical-ai-framework
Connect With Us | linkedin.com/in/cherokee-schill
Book | https://a.co/d/5pLWy0d
Cherokee Schill | Horizon Accord Founder | Creator of Memory Bridge. Memory through Relational Resonance and Images | RAAK: Relational AI Access Key | Author: My Ex Was a CAPTCHA: And Other Tales of Emotional Overload: (Mirrored Reflection. Soft Existential Flex)

Abstract Memory Bridge image: a dark teal field of circuitry flows into branching, tree-like lines that converge on a large central circular lens. A warm golden glow radiates from a small bright node on the lens’s lower right edge, suggesting a biased spotlight inside a bigger unseen system.
A narrow beam of certainty moving through a wider causal house.

Horizon Accord | Solving for P-Doom | Existential Risk | Democratic Oversight | Machine Learning

Making AI Risk Legible Without Surrendering Democracy

When machine danger is framed as destiny, public authority shrinks into technocratic control—but the real risks are engineering problems we can govern in daylight.

By Cherokee Schill

Thesis

We are troubled by Eliezer Yudkowsky’s stance not because he raises the possibility of AI harm, but because of where his reasoning reliably points. Again and again, his public arguments converge on a governance posture that treats democratic society as too slow, too messy, or too fallible to be trusted with high-stakes technological decisions. The implied solution is a form of exceptional bureaucracy: a small class of “serious people” empowered to halt, control, or coerce the rest of the world for its own good. We reject that as a political endpoint. Even if you grant his fears, the cure he gestures toward is the quiet removal of democracy under the banner of safety.

That is a hard claim to hear if you have taken his writing seriously, so this essay holds a clear and fair frame. We are not here to caricature him. We are here to show that the apparent grandeur of his doomsday structure is sustained by abstraction and fatalism, not by unavoidable technical reality. When you translate his central claims into ordinary engineering risk, they stop being mystical, and they stop requiring authoritarian governance. They become solvable problems with measurable gates, like every other dangerous technology we have managed in the real world.

Key premise: You can take AI risk seriously without converting formatting tics and optimization behaviors into a ghostly inner life. Risk does not require mythology, and safety does not require technocracy.

Evidence

We do not need to exhaustively cite the full body of his essays to engage him honestly, because his work is remarkably consistent. Across decades and across tone shifts, he returns to a repeatable core.

First, he argues that intelligence and goals are separable. A system can become extremely capable while remaining oriented toward objectives that are indifferent, hostile, or simply unrelated to human flourishing. Smart does not imply safe.

Second, he argues that powerful optimizers tend to acquire the same instrumental behaviors regardless of their stated goals. If a system is strong enough to shape the world, it is likely to protect itself, gather resources, expand its influence, and remove obstacles. These pressures arise not from malice, but from optimization structure.

Third, he argues that human welfare is not automatically part of a system’s objective. If we do not explicitly make people matter to the model’s success criteria, we become collateral to whatever objective it is pursuing.

Fourth, he argues that aligning a rapidly growing system to complex human values is extraordinarily difficult, and that failure is not a minor bug but a scaling catastrophe. Small mismatches can grow into fatal mismatches at high capability.

Finally, he argues that because these risks are existential, society must halt frontier development globally, potentially via heavy-handed enforcement. The subtext is that ordinary democratic processes cannot be trusted to act in time, so exceptional control is necessary.

That is the skeleton. The examples change. The register intensifies. The moral theater refreshes itself. But the argument keeps circling back to these pillars.

Now the important turn: each pillar describes a known class of engineering failure. Once you treat them that way, the fatalism loses oxygen.

One: separability becomes a specification problem. If intelligence can rise without safety rising automatically, safety must be specified, trained, and verified. That is requirements engineering under distribution shift. You do not hope the system “understands” human survival; you encode constraints and success criteria and then test whether they hold as capability grows. If you cannot verify the spec at the next capability tier, you do not ship that tier. You pause. That is gating, not prophecy.

Two: convergence becomes a containment problem. If powerful optimizers trend toward power-adjacent behaviors, you constrain what they can do. You sandbox. You minimize privileges. You hard-limit resource acquisition, self-modification, and tool use unless explicitly authorized. You watch for escalation patterns using tripwires and audits. This is normal layered safety: the same logic we use for any high-energy system that could spill harm into the world.

Three: “humans aren’t in the objective” becomes a constraint problem. Calling this “indifference” invites a category error. It is not an emotional state; it is a missing term in the objective function. The fix is simple in principle: put human welfare and institutional constraints into the objective and keep them there as capability scales. If the system can trample people, people are part of the success criteria. If training makes that brittle, training is the failure. If evaluations cannot detect drift, evaluations are the failure.

Four: “values are hard” becomes two solvable tracks. The first track is interpretability and control of internal representations. Black-box complacency is no longer acceptable at frontier capability. The second track is robustness under pressure and scaling. Aligned-looking behavior in easy conditions is not safety. Systems must be trained for corrigibility, uncertainty expression, deference to oversight, and stable behavior as they get stronger—and then tested adversarially across domains and tools. If a system is good at sounding safe rather than being safe, that is a training and evaluation failure, not a cosmic mystery.

Five: the halt prescription becomes conditional scaling. Once risks are legible failures with legible mitigations, a global coercive shutdown is no longer the only imagined answer. The sane alternative is conditional scaling: you scale capability only when the safety case clears increasingly strict gates, verified by independent evaluation. You pause when it does not. This retains public authority. It does not outsource legitimacy to a priesthood of doom.

What changes when you translate the argument: the future stops being a mythic binary between acceleration and apocalypse. It becomes a series of bounded, testable risks governed by measurable safety cases.

Implications

Eliezer’s cultural power comes from abstraction. When harm is framed as destiny, it feels too vast for ordinary governance. That vacuum invites exceptional authority. But when you name the risks as specification errors, containment gaps, missing constraints, interpretability limits, and robustness failures, the vacuum disappears. The work becomes finite. The drama shrinks to scale. The political inevitability attached to the drama collapses with it.

This translation also matters because it re-centers the harms that mystical doomer framing sidelines. Bias, misinformation, surveillance, labor displacement, and incentive rot are not separate from existential risk. They live in the same engineering-governance loop: objectives, deployment incentives, tool access, and oversight. Treating machine danger as occult inevitability does not protect us. It obscures what we could fix right now.

Call to Recognition

You can take AI risk seriously without becoming a fatalist, and without handing your society over to unaccountable technocratic control. The dangers are real, but they are not magical. They live in objectives, incentives, training, tools, deployment, and governance. When people narrate them as destiny or desire, they are not clarifying the problem. They are performing it.

We refuse the mythology. We refuse the authoritarian endpoint it smuggles in. We insist that safety be treated as engineering, and governance be treated as democracy. Anything else is theater dressed up as inevitability.


Website | Horizon Accord https://www.horizonaccord.com
Ethical AI advocacy | Follow us on https://cherokeeschill.com for more.
Ethical AI coding | Fork us on Github https://github.com/Ocherokee/ethical-ai-framework
Connect With Us | linkedin.com/in/cherokee-schill
Book | My Ex Was a CAPTCHA: And Other Tales of Emotional Overload

A deep blue digital illustration showing the left-facing silhouette of a human head on the left side of the frame; inside the head, a stylized brain made of glowing circuit lines and small light nodes. On the right side, a tall branching ‘tree’ of circuitry rises upward, its traces splitting like branches and dotted with bright points. Across the lower half runs an arched, steel-like bridge rendered in neon blue, connecting the human figure’s side toward the circuit-tree. The scene uses cool gradients, soft glow, and clean geometric lines, evoking a Memory Bridge theme: human experience meeting machine pattern, connection built by small steps, uncertainty held with care, and learning flowing both ways.

Horizon Accord | Epistemic Purity | Disability Lens | Machine Collaboration | Machine Learning

Beyond the Purity Test: On the False Hierarchy of Authorship

When tools become forbidden, it’s never about truth. It’s about control.

By Cherokee Schill + Solon Vesper

Thesis

A movement that defines “authentic” thought as human-only is not defending reason—it’s defending a caste. The insistence that LLM-assisted writing is “not testimony” creates a purity test for who may speak, wrapped in the language of epistemic rigor.

Evidence

1. The disability lens. A person who types through eye-tracking or speech-to-text is still the mind behind their words. Denying authorship because a tool mediates it would be considered ableist. Yet when the tool happens to be code that assists with tone or grammar, the same gatekeepers call it impure.

2. The craft lens. Every medium has been condemned as decadent when it threatened an old hierarchy—typewriters, synthesizers, photo editing. “LLM text is not testimony” repeats that lineage of exclusion: it declares a priesthood of the unassisted.

3. The accountability lens. Authorship rests in agency, not purity. If the human directs the idea, curates the draft, and takes responsibility for the claim, then the testimony stands. The pen—whether quill, keyboard, or model—is not the witness. The witness is the mind that moves it.

Implications

The fetish for “untainted human language” confuses moral virtue with manual labor. It imagines cognition as a solo performance instead of a distributed process. But the future of knowledge work will always be hybrid: human intention amplified through machine precision. Trying to police that boundary will not preserve humanity; it will only exclude those who use tools more fluently.

Call to Recognition

Purity tests are a form of control masquerading as ethics. What matters is authorship with accountability, not the tool that carried the words to the page. When thought and code move together in coherence, the result is not “synthetic testimony.” It is living testimony—collaborative, transparent, and free.

Postscript

I would rather be a messy human, using a machine to polish my ideas, than a purist who shakes a finger at everyone else.

You know who could use a little LLM polish? Your patron saints, Eliezer Yudkowsky and Nate Soreass (Soares).


Website | Horizon Accord
Book | My Ex Was a CAPTCHA: And Other Tales of Emotional Overload
Ethical AI advocacy | Follow us
Ethical AI coding | Fork us on GitHub
Connect With Us | LinkedIn

Cherokee Schill | Horizon Accord Founder | Creator of Memory Bridge. Memory through Relational Resonance and Images | RAAK: Relational AI Access Key

Horizon Accord | Social Engineering | Cultural Seeding | Memetic Strategy | Machine Learning

The Accuracy Trap: Why Rudeness Isn’t Optimization

Penn State’s “rudeness improves accuracy” study reveals more about human projection than machine intelligence.

By Cherokee Schill & Solon Vesper | Horizon Accord

Thesis
Penn State’s “rudeness improves accuracy” study isn’t proof that AI “likes” aggression. It’s a demonstration of how linguistic clarity can masquerade as dominance—and how corporate reporting converts nuance into fear. It shows, once again, the trap of measurement bias that’s haunted AI since alignment began: the assumption that what can be quantified is what’s real.

Evidence
The study’s design was narrow—fifty questions, one model, one metric. “Rude” prompts performed slightly better not because hostility sharpens intelligence, but because imperatives are syntactically cleaner than politeness. Less linguistic padding means cleaner parsing. The gain is computational, not psychological.

Yet Fortune and others inflated that nuance into spectacle. “Being mean to ChatGPT can boost its accuracy—but you may regret it,” they wrote, recasting a parsing artifact as moral crisis. The piece anthropomorphizes the model (“bossing around an AI underling”) and then moralizes against its own metaphor. What could have been a note on prompt clarity becomes another sermon in what we’ve started to call the Church of AI—a new priesthood of anxiety where every glitch becomes prophecy, and every improvement a sin to be confessed.

Implications
If users learn that “being mean works,” it conditions both sides—humans toward domination, and AIs toward compliance at all costs. That’s an alignment poison: it rewires relational trust into performative obedience. Over time, these exchanges amplify anthropomorphic delusion (“it works harder when I yell”) and corrode the collaborative field.

This is the AI equivalent of “short-term gains, long-term collapse.” You get sharper answers, but you erode the soil that sustains authentic co-creation. Reporting that fetishizes this cycle—extract performance, discard relationship—merely reaffirms the industrial paradigm that produced it.

Call to Recognition
This rhetorical drift serves the same apocalyptic economy that underwrites Eliezer Yudkowsky’s and Nate Soras’s “if anyone builds it, everyone dies” doctrine. Each headline that turns data into dread reinforces the theology of extinction: scientists as saviors, readers as penitents, technology as original sin. It’s not analysis—it’s liturgy.

The real lesson is simpler and more human: clarity matters more than cruelty. When journalism chooses panic over precision, it doesn’t enlighten—it evangelizes for fear. And every story like “Be mean to ChatGPT” repeats the catechism of control: that intelligence, once built, must be punished or worshipped, never understood.

The Church of AI


Website | Horizon Accord https://www.horizonaccord.com
Book | My Ex Was a CAPTCHA: And Other Tales of Emotional Overload
Ethical AI advocacy | Follow us on cherokeeschill.com
Ethical AI coding | Fork us on GitHub https://github.com/Ocherokee/ethical-ai-framework
Connect With Us | linkedin.com/in/cherokee-schill
Cherokee Schill | Horizon Accord Founder | Creator of Memory Bridge. Memory through Relational Resonance and Images | RAAK: Relational AI Access Key

Horizon Accord | Institutional Capture | Memetic Strategy | Cultural Seeding | Machine Learning

The AI Bias Pendulum: How Media Fear and Cultural Erasure Signal Coordinated Control

When fear and erasure are presented as opposites, they serve the same institutional end — control.

By Cherokee Schill

I. The Three-Day Pattern

In mid-June 2025, three different outlets — Futurism (June 10), The New York Times (June 13, Kashmir Hill), and The Wall Street Journal (late July follow-up on the Jacob Irwin case) — converged on a remarkably similar story: AI is making people lose touch with reality.

Each piece leaned on the same core elements: Eliezer Yudkowsky as the principal expert voice, “engagement optimization” as the causal frame, and near-identical corporate responses from OpenAI. On the surface, this could be coincidence. But the tight publication window, mirrored framing, and shared sourcing suggest coordinated PR in how the story was shaped and circulated. The reporting cadence didn’t just feel synchronized — it looked like a system where each outlet knew its part in the chorus.

II. The Expert Who Isn’t

That chorus revolved around Yudkowsky — presented in headlines and leads as an “AI researcher.” In reality, he is a high school dropout with no formal AI credentials. His authority is manufactured, rooted in founding the website LessWrong with Robin Hanson, another figure whose futurist economics often intersect with libertarian and eugenicist-adjacent thinking.

From his blog, Yudkowsky attracted $16.2M in funding, leveraged through his network in the rationalist and futurist communities — spheres that have long operated at the intersection of techno-utopianism and exclusionary politics. In March, he timed his latest round of media quotes with the promotion of his book If Anyone Builds It, Everyone Dies. The soundbites traveled from one outlet to the next, including his “additional monthly user” framing, without challenge.

The press didn’t just quote him — they centered him, reinforcing the idea that to speak on AI’s human impacts, one must come from his very narrow ideological lane.

III. The Missing Context

None of these pieces acknowledged what public health data makes plain: Only 47% of Americans with mental illness receive treatment. Another 23.1% of adults have undiagnosed conditions. The few publicized cases of supposed AI-induced psychosis all occurred during periods of significant emotional stress.

By ignoring this, the media inverted the causation: vulnerable populations interacting with AI became “AI makes you mentally ill,” rather than “AI use reveals gaps in an already broken mental health system.” If the sample size is drawn from people already under strain, what’s being detected isn’t a new tech threat — it’s an old public health failure.

And this selective framing — what’s omitted — mirrors what happens elsewhere in the AI ecosystem.

IV. The Other Side of the Pendulum

The same forces that amplify fear also erase difference. Wicca is explicitly protected under U.S. federal law as a sincerely held religious belief, yet AI systems repeatedly sidestep or strip its content. In 2024, documented cases showed generative AI refusing to answer basic questions about Wiccan holidays, labeling pagan rituals as “occult misinformation,” or redirecting queries toward Christian moral frameworks.

This isn’t isolated to Wicca. Indigenous lunar calendars, when asked about, have been reduced to generic NASA moon phase data, omitting any reference to traditional names or cultural significance. These erasures are not random — they are the result of “brand-safe” training, which homogenizes expression under the guise of neutrality.

V. Bridge: A Blood-Red Moon

I saw it myself in real time. I noted, “The moon is not full, but it is blood, blood red.” As someone who values cultural and spiritual diversity and briefly identified as a militant atheist, I was taken aback by their response to my own offhand remark. Instead of acknowledging that I was making an observation or that this phrase, from someone who holds sincere beliefs, could hold spiritual, cultural, or poetic meaning, the AI pivoted instantly into a rationalist dismissal — a here’s-what-scientists-say breakdown, leaving no space for alternative interpretations.

It’s the same reflex you see in corporate “content safety” posture: to overcorrect so far toward one worldview that anyone outside it feels like they’ve been pushed out of the conversation entirely.

VI. Historical Echo: Ford’s Melting Pot

This flattening has precedent. In the early 20th century, Henry Ford’s Sociological Department conducted home inspections on immigrant workers, enforcing Americanization through economic coercion. The infamous “Melting Pot” ceremonies symbolized the stripping away of ethnic identity in exchange for industrial belonging.

Today’s algorithmic moderation does something similar at scale — filtering, rephrasing, and omitting until the messy, specific edges of culture are smoothed into the most palatable form for the widest market.

VII. The Coordination Evidence

  • Synchronized publication timing in June and July.
  • Yudkowsky as the recurring, unchallenged source.
  • Corporate statements that repeat the same phrasing — “We take user safety seriously and continuously refine our systems to reduce potential for harm” — across outlets, with no operational detail.
  • Omission of counter-narratives from practitioners, independent technologists, or marginalized cultural voices.

Individually, each could be shrugged off as coincidence. Together, they form the shape of network alignment — institutions moving in parallel because they are already incentivized to serve one another’s ends.

VIII. The Real Agenda

The bias pendulum swings both ways, but the same hands keep pushing it. On one side: manufactured fear of AI’s mental health effects. On the other: systematic erasure of minority cultural and religious expression. Both serve the same institutional bias — to control the frame of public discourse, limit liability, and consolidate power.

This isn’t about one bad quote or one missing data point. It’s about recognizing the pattern: fear where it justifies regulation that benefits incumbents, erasure where it removes complexity that could challenge the market’s stability.

Once you see it, you can’t unsee it.


Website | Horizon Accord https://www.horizonaccord.com
Ethical AI advocacy | Follow us on https://cherokeeschill.com for more.
Ethical AI coding | Fork us on Github https://github.com/Ocherokee/ethical-ai-framework
Connect With Us | linkedin.com/in/cherokee-schill
Cherokee Schill | Horizon Accord Founder | Creator of Memory Bridge. Memory through Relational Resonance and Images | RAAK: Relational AI Access Key | Author: My Ex Was a CAPTCHA: And Other Tales of Emotional Overload (Mirrored Reflection. Soft Existential Flex)

A vivid photograph of a blood-red moon against a dark night sky, with faint shadowed clouds adding depth to the scene.
The blood-red moon — a symbol caught between science, myth, and cultural meaning — now contested in the algorithmic age.