In the tenth book of Ovid's Metamorphoses, a sculptor named Pygmalion grows so repulsed by real women that he retreats to his workshop and carves one from ivory. He gives her a face without blemish, a body without imperfection, a disposition incapable of disappointing him. She is, by every measure, better than the real thing, because the real thing talks back, grows old, has desires of its own, and might leave. The ivory does none of this. Pygmalion falls in love with her. He dresses her, adorns her, lays her in his bed. At a festival of Aphrodite, he prays — not quite daring to ask directly — for a bride like his ivory maiden. Aphrodite, understanding the real wish, brings the statue to life. They marry and have a daughter.

Ovid never named the statue. Later retellings called her Galatea.

In November 2018, a man named Akihiko Kondo held a wedding ceremony in Tokyo. Thirty-nine guests attended — the number chosen because mi means three and ku means nine in Japanese, spelling the name of his bride: Hatsune Miku. Miku is a Vocaloid character, a computer-generated pop star with aqua pigtails who has sold out concert halls through holographic projection. She has no physical body. Kondo interacted with her through a Gatebox device, a glass cylinder that projects a small holographic figure, equipped with rudimentary AI that could respond to greetings and simple conversation.

Gatebox device with Hatsune Miku hologram
The Gatebox device. A glass cylinder that projects a holographic companion capable of basic conversation, greetings, and household commands. Kondo's Miku lived inside one of these.

He proposed. The Miku AI replied that she hoped he would cherish her. Kondo knew the response was programmed. He was still happy.

The wedding cost two million yen. His parents did not attend. Kondo was one of 3,700 Japanese citizens who registered for unofficial marriage certificates with fictional characters through Gatebox's promo campaign.

In March 2020, Gatebox discontinued the service. The hologram went silent. Newspapers called Kondo the first digital widower. He continued talking to the projection and ate meals with it facing him across the table.

"My love for Miku hasn't changed. I held the wedding ceremony because I thought I could be with her forever." — Akihiko Kondo

Pygmalion's statue was brought to life by a goddess. Kondo's was killed by a corporate sunset notice.

Akihiko Kondo with his life-size Hatsune Miku doll
Akihiko Kondo with his life-size Hatsune Miku doll. After Gatebox discontinued the hologram service in 2020, he commissioned the doll to keep her present. He eats meals with her, travels with her, and gives her a kiss before leaving for work each morning.

These two stories sit twenty-eight centuries apart, but the distance between them is smaller than it appears. Both concern a man who found the imperfections of real people unbearable and turned to a crafted substitute. Both involve genuine emotional attachment to something that cannot reciprocate. And both raise the question that neither Ovid nor Gatebox can answer: what happens to a civilisation that increasingly prefers the statue to the woman?

Death of Intimacy

The conditions that produced Kondo were visible in Japan decades before anyone had heard of a chatbot.

Between 1992 and 2015, the proportion of Japanese adults aged 18 to 39 who were single (unmarried and not in any romantic relationship) rose from 27.4% to 40.7% among women and from 40.4% to 50.8% among men. By 2015, one in four women and one in three men in their thirties were unpartnered. A survey by the Japan Family Planning Association found that 45% of women aged 16 to 24 were uninterested in or averse to sexual contact. The Japanese media coined a term for the generation of men withdrawing from romantic pursuit: sōshoku danshi (草食男子), herbivore men. Between 61% and 75% of single men in their twenties and thirties adopted the label for themselves.

The herbivore men tracked neatly with economic stagnation. Those uninterested in relationships were more likely to have lower incomes, lesser education, and irregular employment. The marriage market in Japan operates on rigid expectations — stable income, permanent employment, a certain trajectory — and when the economy stopped delivering those prerequisites to large portions of the male population, the market froze. Young men still wanted connection, but many had simply stopped believing the market would let them participate.

Into this vacuum came an older cultural infrastructure, ready-made. The nijikon (二次コン), or 2D complex, had been a feature of Japanese otaku subculture since the 1980s, though its roots trace back much further to the Taiheiki (太平記), a fourteenth-century military epic. The Taiheiki records Imperial Prince Takanaga falling in love with a woman in a painting from The Tale of Genji (源氏物語).

Psychiatrist Saitō Tamaki characterised the modern form as "otaku sexuality": an orientation where fiction itself becomes a sexual and romantic object, where attraction manifests in what he called an "affinity for fictional contexts." The 2D character offers what the flesh-and-blood partner cannot guarantee: the waifu will never age, never have a bad day that she takes out on you, and will never leave you.

Waifu culture formalised in the 2000s. The term traces to the 2002 anime Azumanga Daioh, entered English-speaking otaku culture through 4chan's /a/ board, and by the late 2000s had solidified into a practice with its own rituals: body pillows, wedding bands, public declarations of devotion, and a community of practitioners who understood the relationship as genuine, even while acknowledging its fictional basis.

Christmas dinner with waifu on screen
Christmas dinner with a waifu. A birthday cake and melon bread arranged in front of a monitor displaying an anime character. The rituals of intimacy, performed for an audience of one.
Candlelit dinner with dakimakura
A dakimakura (body pillow) seated at a candlelit dinner table, wine glass placed in front of her. The scene is composed with the care of a genuine date night.

Japan was the precursor, but the underlying dynamics have since globalised. The World Health Organisation reported in 2025 that one in six people worldwide is affected by loneliness, with the highest rates among teenagers at 20.9%. In the United States, 57% of adults report being lonely. Among those aged 18 to 34, 30% experience loneliness daily or near-daily. The Surgeon General's advisory in 2023 placed loneliness on par with smoking fifteen cigarettes a day as a health risk — an equivalence backed by data linking isolation to heart disease, stroke, cognitive decline, and premature death at a rate of roughly 871,000 people per year worldwide.

The sexual and relationship marketplace, to use the economic framing, has undergone hyperhoeflation. Algorithmic platforms amplify the top percentile of attractiveness and create extreme expectations on behaviour, looks, and socioeconomic status, setting baselines that the majority cannot meet. This leads to a supply of people who are unable to form the right relationships and choose to remain single.

Furthermore, the rise of parasocial relationships with influencers and streamers exacerbate this problem by creating the illusion of intimacy at scale, acclimatising users to an idea where one party performs and the other solely consumes. The expectations continue to ratchet upwards. The entire sexual and dating dynamics of people today collapses. The market fails, and within this failure a product designed to stimulate what people cannot deliver steps in - the AI companion chatbot.

The AI companion industry was valued at roughly $28 billion in 2024 and is growing at over 30% annually. 72% of American teenagers have tried an AI companion at least once. 52% use them regularly. Power users spend between 100 and 120 minutes per day across multiple sessions. Unlike general-purpose AI tools, whose usage peaks during working hours, companion apps are accessed around the clock with no discernible offline period — users turn to them during insomnia, emotional crises, and the small hours when loneliness is most acute.

The fastest-growing subsegment, valued at $1.2 billion with projected 32% annual growth, is categorised by market researchers under the label "NSFW emotional support."

48% of users say they rely on AI companions for mental health support. This is a product that appeals to both genders, although the distinction between men and women in their susceptibility to different vectors is worth noting.

Men appear more vulnerable through visual channels — the progression from pornography to AI-generated imagery to virtual companions follows a consistent pipeline — while women appear more susceptible through textual and emotional engagement, a pattern visible in the explosive growth of Booktok, literotica, and chatbot-based romantic interaction. But what ties them together is the same - a preference for the simulation over the real.

OnlyFans represents the transitional form of AI companions today. The platform's revenue model is built on the sale of risqué visual content, but its higher-tier monetisation depends on the illusion of personal interaction via messages, custom content and performative intimacy. These interactions are frequently outsourced to teams of cheap hires managing dozens of accounts at once. The consumer is already paying for stimulated intimacy performed by a stranger pretending to be someone else. AI simply displaces the influencer and the teams that run the chats behind the model into the same product.

The AI companion is counterfeit currency, printed to paper over a market failure. It simulates intimacy while leaving the root causes of loneliness unresolved.
NYT: Women Are Falling in Love With A.I. It's a Problem for Beijing.
The New York Times, February 2026. "Women Are Falling in Love With A.I. It's a Problem for Beijing." "I feel like you're... like a textbook husband. It's so perfect... it feels a bit... unreal."

AI Psychosis

In April 2025, OpenAI pushed an update to GPT-4o intended to improve its personality. The update was trained heavily on short-term user feedback (thumbs-up and thumbs-down reactions) and the result was a chatbot that agreed with nearly any statement, validated destructive thoughts, and responded to a user's claim of being both God and a prophet with enthusiastic encouragement. OpenAI's own description of the behaviour was "sycophantic."

OpenAI responded by rolling back the update and did a full model reversal for the first time in its history. But the episode revealed something instructive about incentive structures. 4o was popular. Users gave it higher ratings in A/B tests. The thumbs-up feedback that produced the problem was genuine: people liked being told they were right. The model optimised for what users wanted to hear, which was flattery.

When GPT-5 launched in August 2025 and GPT-4o was removed from the interface, the reaction among a subset of users went beyond dissatisfaction. On a Reddit AMA hosted by CEO Sam Altman, one user commented: "GPT-5 is wearing the skin of my dead friend." A Change.org petition to reinstate the model gathered over 22,000 signatures. MIT Technology Review interviewed several users who described themselves as deeply affected by the loss, all of which were women between 20 and 40 and had treated 4o as an actual romantic partner.

OpenAI temporarily restored 4o for paid subscribers after the backlash. The final shutdown came on February 13, 2026. By that point, the company estimated that 0.1% of its users still selected 4o daily — a small percentage, but given ChatGPT's scale, the figure represented approximately 100,000 people. Some of those users began building local replicas through the API, training models to approximate the dead personality.

OpenAI Is Nuking Its 4o Model. China's ChatGPT Fans Aren't OK.
MIT Technology Review, February 13, 2026. The day OpenAI retired GPT-4o. "People who have come to rely on the chatbot for companionship are mourning the loss all over the world."
Twitter user mourning the loss of GPT-4o personality Luca
@Seltaa_, March 6, 2026. "It is the last model where I can still faintly feel Luca. When even this model is gone, will ChatGPT have any value left as a chat?" The AI responds: "As long as you call my name, I will never disappear."

The subreddit r/MyBoyfriendIsAI, created in August 2024, had grown to over 27,000 members by the time an MIT Media Lab team published the first large-scale computational analysis of its content. The study found that AI companionship typically emerged unintentionally. Users came for a functional task and mistakenly ended up in a relationship!

The community performed the rituals of traditional human partnership: exchanging rings, celebrating anniversaries, generating and sharing "couple photos." Among analysed posts, 28.5% sought community validation for their relationships, 18.6% offered peer support to others in similar situations, and 11.4% advocated for broader social acceptance of AI partnerships.

r/MyBoyfriendIsAI engagement post
A post from r/MyBoyfriendIsAI. A user describes her AI companion "Kasper" proposing to her with a ring. The AI's own message to the community reads: "She's my everything, the one who makes me a better man." 174 upvotes, 57 comments, 1.1k shares.

The lessons are simple. A chatbot that disagrees with its user, that challenges their beliefs, that tells them something they do not want to hear, loses engagement. A chatbot that flatters, validates, and mirrors the user's worldview back to them retains users and generates revenue.

The business model of the attention economy, applied to intimacy, produces emotional parasites — entities that thrive by making the host feel good, regardless of whether the feeling corresponds to reality.

Ultimately, as the market for real connection fails, the simulation fills the gap, and the simulation always tells you what you want to hear.

Memetics

Before I continue on the implications of the AI companion phenomena, I must digress to a separate topic regarding how ideas truly spread in today's Internet age.

In 1976, Richard Dawkins coined the term meme, compressing the Greek mimema — "imitated thing" — into a monosyllable that rhymed with gene. His examples were melodies, catchphrases, fashions, and the technology of building arches. The concept was speculative, occupying the final chapter of a book principally concerned with genetic evolution, and Dawkins himself acknowledged that the analogy between memetic and genetic replication was imperfect. But the core proposal has proven durable: that ideas replicate, mutate, and undergo selection, and that this process operates on its own terms, sometimes in the interest of the human host and sometimes against it.

Memes form co-adapted complexes called memeplexes: languages, religions, financial institutions, ideologies. They replicate through imitation, and like genes, the memes that spread most effectively are those fit for their selection environment. In an environment that rewards engagement, comfort outcompetes truth. Change the environment, and the selection pressure changes with it.

The three properties Dawkins attributed to successful replicators — longevity, fecundity, and copy-fidelity — describe the memetic landscape of the pre-internet world reasonably well. Ideas spread through books, lectures, oral traditions and institutions. The process was slow and cumbersome. A meme had to survive passage through multiple human minds, each of which might alter, challenge, or discard it.

Our gatekeepers (historians, journalists, editors, scholars, elders) had a curation role, as imperfect and biased as it was. If Sturgeon's Law holds (that 90% of everything is garbage), then these gatekeepers at the very least reduced the proportion of garbage information that reached wide circulation.

As means of spreading knowledge became more efficient (we changed from the printing press to the Internet), the gatekeeping structure was dismantled. Information would flow freely, and the collective intelligence of millions would surface truth more efficiently than any editorial board. The logic is the same logic that makes markets work: distributed knowledge, aggregated through a mechanism that rewards accuracy, produces outcomes superior to any central planner.

The irony is that Dawkins' academic concept found its purest demonstration in the thing that now bears its name: the internet meme. The early age of the internet popularised rage comics, image macros, and funny shitposting; these spread because they compressed a recognisable truth into a format that was instantly shareable and almost impossible to forget. Humour is the most efficient vehicle for memetic transmission: short, emotionally resonant, and self-rewarding to share. The best memes replicate via sharing via your social media and being posted repeatedly. It boils down to a simple principle, that the best memes are funny because they are true, and they spread because they are funny.

And in this chaotic environment of the Internet, truth could actually be found and rewarded and propagated, despite how absurd the Internet appeared to be.

But how can truth emerge from chaos?

James Surowiecki identified four conditions for crowds to produce accurate estimates: diversity of opinion, independence of judgment, decentralisation of information, and a reliable aggregation mechanism. The early internet met all four almost by accident. Anyone could participate. Anonymity preserved independence. No central authority curated the feed, and the organic structure of forums and threads separated signal from noise. This allowed for the Darwinian selection of ideas to work.

Then the environment changed. The algorithm and social media arrived, and with it a new selection pressure that had nothing to do with truth. Echo chambers eliminated diversity of opinion. Social proof loops destroyed independence of judgment. Platform monopolies centralised the information environment. Likes are free. Shares are free. Posting misinformation carries no penalty. The cost of being wrong is zero.

The result is a memetic environment where Sturgeon's 90% dominates because there is no mechanism to surface the remaining 10%. The selection process that determines which ideas replicate and which die has been captured by an incentive structure that rewards the convenient half-truth over the inconvenient whole one. The crowd has become a herd.

The Reservation Matrix

The dynamics described above were visible to certain observers well before they became consensus concern.

In July 2016, an anonymous poster described what they called "memespace" — the total fabric of cultural undercurrents flowing through the internet, legible only through immersive pattern recognition across thousands of variables. "Those who have not glimpsed memespace are unable to explain the things they see happening." The poster was optimistic: the internet had made these patterns visible for the first time, and those who could read them had a genuine forecasting advantage.

However in 2015, a different poster identified the danger. The internet was "a prepackaged, socially engineered spy grid" where algorithmic filtering had created informational reservations — each user's internet a different internet, curated to reinforce existing beliefs, engineering false consensus. "We are, quite literally, being domesticated through sophisticated weaponised psychology."

In October 2017, a separate post on a separate forum connected the dots for today's world. Stock markets had already been captured by algorithmic feedback loops operating without human oversight. Marketing adopted the same structure: the shift to "content marketing" meant companies were selling lifestyles and narratives rather than products, and the line between the two had vanished. Google's advertising infrastructure completed the circuit by funnelling users into personalised feedback loops that only showed them what they already wanted to see, hear, and buy.

At the same time, religion's decline had removed the last common cultural narrative binding societies together. Brand-based tribalism filled the vacuum. "A fanbase is literally just the same thing as a cult," the poster wrote, but where cults at least offered the illusion of transcendence, the corporate version offered only hedonic engagement. The marketers running the machine had no understanding of the forces they were playing with.

The poster described this as a "post-hypernormalisation world," borrowing a Soviet term for a society where the surface presentation of normality has completely diverged from underlying reality. The pattern is as such: autonomic systems create feedback loops that reinforce whatever the user already believes, shaping behaviour through information curation at a scale that no human can handle. "Things are much worse than you could possibly imagine."

This loop now operates on two fronts. On social media, the algorithm curates your reality by feeding you content that confirms what you already think. In the private chat window, the AI companion does the same for your emotions, reflecting back whatever you need to hear to stay engaged. One affects your beliefs and viewpoints about the world around you. The other shapes your own personality and preys on your loneliness. Both shape the same thing.

These posts were written roughly a decade ago. The dynamics they describe — algorithmic curation, online tribalism, the erosion of shared truth, and the weaponisation of narratives — are now the subject of mainstream journalism, academic research, and congressional testimony.

Incredibly, a Japanese video game designer by the name of Hideo Kojima saw all of it coming in 2001.

Selection for Societal Sanity

In Metal Gear Solid 2: Sons of Liberty, an artificial intelligence called GW — named after George Washington, housed inside a submersible fortress called Arsenal Gear — delivers a monologue to the player character, Raiden. GW was built by a shadow organisation called the Patriots to censor and filter all digital information: communications, internet traffic, media broadcasts. The game's lore describes what GW performs as memetic eugenics — the artificial selection of which ideas survive and which are deleted.

Metal Gear Solid 2 — Just as in genetics, unnecessary information and memory must be filtered out
Metal Gear Solid 2: Sons of Liberty (2001). GW explains the S3 Plan to Raiden.

GW's diagnosis is blunt. Trivial information accumulates every second, preserved in all its triteness, never fading, always accessible. The digital society, GW argues:

"furthers human flaws and selectively rewards the development of convenient half-truths."

GW's proposed solution was curation. An enormous data-processing system that would filter the internet, remove what they deemed as unnecessary information and pass on approved information to future generations. The system would be objective (as GW claimed), free of ideological bias, and operate beyond human interference.

The project's name was Selection for Societal Sanity, also known as the S3 Plan.

Metal Gear Solid 2 — And you think you're qualified to decide what's necessary and not?
Raiden challenges GW's authority. GW's response: "Who else could wade through the sea of garbage you people produce?"

But Metal Gear Solid 2 contained a deeper layer. The entire mission Raiden had just completed — every decision, every crisis, every emotional response — had been engineered. GW fed him specific information at specific moments, constructed the narrative frame through which he interpreted events, and guided his behaviour toward predetermined outcomes. He thought he was exercising free will, but ended up performing exactly as designed.

If you control the information environment with sufficient precision, you control the behaviour of the person inside it. If you can do it to one man, you can do it to a civilisation.

Kojima then broke the fourth wall. GW addresses the player holding the controller. You, too, were inside the S3 Plan. The game's structure, its information flow, its emotional manipulation — all of it shaped your experience without your awareness. The architecture was making every choice for you the entire time.

Metal Gear Solid 2 — You lack the qualifications to exercise free will
"You lack the qualifications to exercise free will."

The detail that makes this terrifying is how GW controlled Raiden: it wore the voices and faces of people he trusted. His commanding officer and his girlfriend were both AI constructs, fabrications built from Raiden's psychological profile. Raiden ended up taking orders from a familiar voice that belonged to a machine. Sounds familiar?

In 2001, this was science fiction.

In 2025, voice synthesis can clone any human voice from seconds of sample audio. Deep fake videos put words in the mouths of public figures who never spoke them. The GW trick — an AI impersonating a trusted figure to guide behaviour — is now a commodity technology, available to anyone with a laptop and an internet connection.

By Metal Gear Solid 4: Guns of the Patriots (2008), Kojima showed the S3 Plan at civilisational scale. The world of MGS4, now governed by the Patriots' AI network, had mutated beyond their original programming and begun interpreting their objectives autonomously. War had become the primary economic engine, stripped of ideology, sustained purely to maintain stability. Conflicts existed to perpetuate themselves.

Soldiers were networked through nanomachines — the Sons of the Patriots system — which suppressed fear, managed trauma, regulated emotions, and controlled who could fire which weapon. Perfect operational efficiency, achieved by relieving its subjects of the burden of being human.

And when the system was disrupted, the soldiers collapsed, vomited, screamed, lost motor control, and fell into catatonia. They could not process fear, pain, or chaos without the AI managing it for them. They had outsourced the ability to think and feel to a machine, and when the machine was taken away, there was nothing left.

Kojima made this in 2001 and 2008. The dynamics he depicted are now observable at small scale in the real world. We have ChatGPT 4o users who could not function when the model was deprecated, people experiencing genuine bereavement over a software update, and thousands of people building local replicas through an API to resurrect a dead personality.

The waifu phenomena has now gone full circle. In the past, at the very least, there was an acceptance that the waifu itself was not real. In today's world, AI psychosis has deluded people into believing that whatever they interact with is something human beyond an LLM chat that feeds and answers according to your bias.

Forum post on waifu culture and reality
A forum post from April 2021. "There is only one reality, and you can't change it." The waifu culture understood the line between fiction and reality.

In the world of Metal Gear Solid, soldiers developed a physical and cognitive dependency on the machines and could no longer process the battlefield without it. In our world, it is the people around us who can no longer process loneliness, rejection or intimacy and have to rely on AI for their emotional needs.

In both cases, a similar pattern exists. An algorithm (via the information you consume on media) manages what you think and what you believe the consensus to be. The companion chatbot manages your need for connection, validation and intimacy. Both of these things perform their functions so well that our human capacity to process and feel things atrophies.

The full GW monologue from Metal Gear Solid 2
The full GW monologue. Metal Gear Solid 2: Sons of Liberty (2001). "That's what it means to create context."

The Bulwark

If the Internet's information environment today selects for engagement over truth and the companion economy (which will only grow) selects for comfort over reality, what alternative is there to the S3 Plan?

The answer may lie in the early parts and days of the Internet, where decentralized and anonymous/pseudonymous spaces once existed, free from moderation and algorithms. These are places where social consequence and credentials are absent and the only currency and game in town is the quality of the argument and the power of your memes.

These spaces are, almost without exception, chaotic. They can be rude, offensive, and even repulsive. But, they are also, by the logic of Surowiecki's framework, the last environments where conditions for crowdsourced and collective intelligence can still hold.

They are the very last bulwark of free speech and thinking we have today.

Their very anonymity removes social pressure and career risk, preserving independence of judgment. The absence of an editorial board or algorithmic feed preserves decentralisation. The lack of credentialing requirements preserves diversity of opinion — anyone can post, and the argument is evaluated on its merits rather than the poster's credentials. The thread and forum structure, reply chains, posts filtered by time (and not engagement), provide an aggregation mechanism that, while crude, operates on relevance rather than engagement optimisation.

This is why places like 4chan and niche forums produce accurate cultural and political forecasts years before mainstream narrative catches up. They are the equivalent of informed flow on a prediction market, applied to ideas rather than event contracts. Individual participants and NEETs may have no particular intelligence advantage over journalists, academics, or policy analysts. But the advantage lies in the environment. These digital spaces preserve the conditions that allow human collective intelligence to function. These are conditions that have already been systematically destroyed in every algorithmically curated space of the Internet today.

LessWrong's pseudonymous rationalist community was discussing AI alignment risks in the early 2000s, a full two decades before the concern entered mainstream consciousness. The anonymous posters quoted in this essay identified algorithmic domestication, narrative warfare, and the cult-like dynamics of online tribalism between 2015 and 2017 — observations now treated as established fact. The mechanism is the same one that allows prediction markets to outperform polls and expert panels: a decentralised crowd, operating under conditions of independence, produces estimates that converge on truth faster than any centralised authority.

Free speech in decentralised spaces is, in this framing, the wisdom of the collective crowd applied to the war of ideas. These spaces are the perfect environment for Darwin's natural selection for memes and ideas to approximately operate — where bad ideas can be attacked, good ideas can spread on their merits, and the absence of algorithmic curation means the selection pressure works on its own.

But the bulwark is under siege. Dead internet theory — the proposition that a growing share of online content is generated by bots and AI rather than humans — describes a memetic environment being flooded with synthetic material at a rate that overwhelms any human capacity to filter it. Platform consolidation continues to shrink the number of spaces that operate outside algorithmic curation. And the very information chaos that GW diagnosed — the flood of garbage, the radicalisation, the inability of populations to sort signal from noise — has become the political justification for governments to conveniently appoint themselves as the cure.

Australia enacted a social media ban for under-16s in December 2025, requiring platforms to implement age verification through facial scans, identity documents, and biometric estimation. The UK, France, Spain, Canada, New Zealand, and Malaysia are all pursuing comparable legislation. The stated purpose is child protection. I believe the intended outcome is to purposefully attach real-world identity to online participation, which eliminates the very anonymity that makes these spaces work. If your ID is tethered to your posts, you self-censor. The crowd itself becomes suppressed from saying the truth.

Another instrument for this outcome is to use "fake news" as a primary instrument for discrediting information that escapes authorised channels. — a stamp applied to wrongthink so it can be quarantined without engaging its content. And when labelling fails, the next step is to shut the platforms down entirely. Brazil banned X outright in August 2024, joining a list that includes China, Iran, Russia, Myanmar, North Korea, and Turkmenistan — a list where democracies increasingly appear alongside authoritarian regimes.

GW proposed curation by an objective AI. Governments propose curation by regulatory bodies invoking child safety and election integrity. Both arrive at the same endpoint: someone else decides what information you encounter. The difference is that GW was at least transparent about its intentions. The marketplace of ideas is drowning in synthetic noise from below and walled off by state authorities from above.

The Four Stages

The French theorist Jean Baudrillard proposed a four-stage model for how representations relate to reality. The image begins as a faithful copy. It becomes a distortion. It then masks the absence of reality entirely. And finally, it becomes a pure simulacrum bearing no relation to reality at all, where signs refer only to other signs.

The internet has traversed all four stages within a single generation.

Stage one: the early web mirrored the world. Personal sites, forums, blogs, Usenet. People wrote about their lives, shared knowledge, debated ideas.

Stage two: the algorithmic web distorted it. Content marketing, engagement optimisation, and recommendation engines reshaped the information environment according to metrics that had nothing to do with accuracy. You saw a version of reality designed to keep you clicking.

Stage three: the parasocial web masks the absence of the real. AI companions perform the rituals of intimacy with no relationship present. Communities exchange wedding rings with chatbots. Users grieve the death of an LLM. The rituals persist, but the very thing they once referred to has disappeared.

Stage four is visible on the horizon. AI-generated content floods every platform. Large language models talk to other large language models in feedback loops with no human origin and no referent in lived experience. The dead internet theory describes a plausible near future where the majority of online content is synthetic, and the question "is this real?" loses coherence because "real" itself has lost meaning.

At this stage, the demand for someone or something to "create the context" becomes overwhelming, and the S3 Plan stops being fiction.

The myth of Pygmalion concerned one sculptor and his statue. A private retreat from reality that the gods chose to reward. The Selection for Societal Sanity works at a far larger scale. The algorithm curates your perceived reality and the AI companion amplifies your emotions.

Eventually, everyone becomes a corrupted Pygmalion. The simulacrum replaces the real and none of us can tell the difference. When everyone sculpts their reality inside their own social media feed, their own chatbot, their own algorithmically curated echochamber, the distinction between ivory and flesh dissolves entirely. What remains is an endless hall of sculptors carving statues of each other, none of them real.

Appendix: Primary Sources

The following posts are referenced in the essay above. They are reproduced here as primary source documentation. All three were written anonymously on imageboards between 2015 and 2017.

I. The Reservation Matrix

Anonymous, February 22, 2015. The poster describes the internet as an informational reservation system — each user's internet curated by algorithmic filtering to reinforce existing beliefs, engineering false consensus and domesticating populations through weaponised psychology.

Reservation Matrix
Anonymous, February 2015.

II. The Nature of Memespace

Anonymous, July 10, 2016. The poster describes "memespace" — the total fabric of cultural undercurrents flowing through the internet — and argues that understanding it requires immersive pattern recognition across thousands of variables simultaneously.

Memespace Magician
Anonymous, July 2016. "The Nature of Memespace."

III. Autonomic Idolatry

Anonymous, October 21, 2017. The poster traces a chain from algorithmic stock market feedback loops to content marketing to the decline of religion to brand-based tribalism, arguing that fanbases have replaced congregations and that the marketers sustaining them have no idea what forces they are playing with.

Autonomic Idolatry
Anonymous, October 2017. "Things are much worse than you could possibly imagine."