The Loss Of Agency
- Get link
- X
- Other Apps
From an article at MIT Technology FB page. Please comment as it does quote one of your designers: For many, artificial general intelligence, or AGI, is more than just a technology. In tech hubs like Silicon Valley, it’s talked about in mystical terms. Ilya Sutskever, cofounder and former chief scientist at OpenAI, is said to have led chants of “Feel the AGI!” at team meetings. And he feels it more than most: In 2024, he left OpenAI, whose stated mission is to ensure that AGI benefits all of humanity, to cofound Safe Superintelligence, a startup dedicated to figuring out how to avoid a so-called rogue AGI (or control it when it comes). Superintelligence is the hot new flavor—AGI but better!—introduced as talk of AGI becomes commonplace. Sutskever also exemplifies the mixed-up motivations at play among many self-anointed AGI evangelists. He has spent his career building the foundations for a future technology that he now finds terrifying. “It’s going to be monumental, earth-shattering—there will be a before and an after,” he told me a few months before he quit OpenAI. When I asked him why he had redirected his efforts into reining that technology in, he said: “I’m doing it for my own self-interest. It’s obviously important that any superintelligence anyone builds does not go rogue. Obviously.” He’s far from alone in his grandiose, even apocalyptic, thinking. People are used to hearing that this or that is the next big thing, says Shannon Vallor, who studies the ethics of technology at the University of Edinburgh. “It used to be the computer age and then it was the internet age and now it’s the AI age,” she says. “It’s normal to have something presented to you and be told that this thing is the future. What’s different, of course, is that in contrast to computers and the internet, AGI doesn’t exist.” And that’s why feeling the AGI is not the same as boosting the next big thing. There’s something weirder going on. Here’s what we think: AGI is a lot like a conspiracy theory, and it may be the most consequential one of our time. Read the story: https://trib.al/5OfS0YM
ChatGPT said:
Great question. That’s a provocative framing, and there’s a lot to unpack. Here are some thoughts, from different angles — plus some critique and context. (Let me know if you want a deeper dive into any part.)
What the Article Is Getting At
Mystical / Evangelical Language Around AGI
The article (from MIT Tech Review, via the Facebook post) argues that some in Silicon Valley talk about AGI as more than a technical project — almost a spiritual or mystical quest.
The quote about Ilya Sutskever leading chants of “Feel the AGI!” is used to underscore that point: it's not just engineering, it's something people emotionally feel or believe in.
This is interesting because it suggests that part of AGI discourse is driven less by purely technical reasoning and more by metaphoric or quasi-religious framing.
Ambivalent Motivations
The article interprets Sutskever’s move from OpenAI to founding “Safe Superintelligence” as deeply personal — he helped build the foundational work and now is terrified of the consequences.
His own words (“I’m doing it for my own self-interest … it’s obviously important … that any superintelligence … does not go rogue”) are read in a way that highlights self-preservation as much as altruism.
AGI as a “Conspiracy Theory”
This is the article’s central metaphor: AGI isn’t just a future technology, but a belief system with features of a conspiracy theory.
According to that framing, AGI narratives serve multiple functions: they mobilize funding, justify speculative investment, shape research direction, and assign grand stakes to uncertain futures.
The article suggests that treating AGI like a myth or prophecy may have real-world effects, even if AGI in the form people imagine doesn’t exist yet.
Commentary & Analysis
Strengths of the Article’s Argument
Valid Cultural Critique
The notion that AGI talk has become almost cultish is not entirely unfounded. Big ideas, especially existential ones, attract evangelism, hype, and metaphors.
Calling out this “myth-making” could help bring more critical thinking: Are we funding AGI for the right reasons? Are we mistaking speculative future stories for technical inevitabilities?
Risk of Overhyping
When people treat AGI as an imminent, god-like intelligence, funding and policy may get distorted toward speculative risks, possibly at the expense of more immediate (and tractable) harms of current AI.
The “belief in AGI” shapes research agendas. That’s not automatically bad, but there’s a risk in conflating plausible future scenarios with guaranteed futures.
Motivational Transparency
Highlighting Sutskever’s “mixed” motivation is useful. If key actors are motivated by self-interest, fear, and ambition — not just altruistic “safe AGI” goals — then that should factor into how we evaluate their work.
Risks / Weaknesses of the Article's Framing
Risk of Discounting Legitimate Concern
Labeling AGI as a “conspiracy theory” risks trivializing serious and technical debates about alignment, control, catastrophic risk, and superintelligence.
Some of the people working on AGI are deeply serious: their concerns are not just speculative hype but grounded in technical, philosophical, and ethical work.
Overgeneralization
Not everyone in AI talks in mystical or evangelistic language. The article may be overgeneralizing from a subset of vocal figures.
The AGI field is diverse: there are engineers, safety researchers, ethicists, policymakers — not all of them chant or “feel” AGI.
Unclear Boundaries
The article’s metaphor (“conspiracy theory”) might blur useful distinctions: between belief in a speculative future, engineering commitment to long-term safety, and capitalist / personal incentives.
Conspiracy theory often implies secrecy, deception, and irrational belief — but AGI discourse also involves open research, published papers, and real technical challenges.
Specific Reflection on Sutskever
On His Role
Sutskever is a foundational figure in modern AI. His shift to Safe Superintelligence is meaningful: he’s not just worried, he’s making it his life’s work.
That said, his admission that part of his motivation is “self-interest” is sobering. It humanizes him: he’s not some impersonal technocrat — he’s wrestling with fear, ambition, and responsibility.
On Risk and Unpredictability
According to other sources, Sutskever has publicly said that a future superintelligence could be “unpredictable” and agentic. TechCrunch+1
That unpredictability is a real technical concern. It’s not just hand-waving: alignment theory, control, and safety are central to his new company. NY1+1
However, calls for “safe superintelligence” haven’t guaranteed that the path is easy or that the risk can be fully mitigated.
Broader Implications
For AI Policy & Governance
If AGI is being treated like a myth or prophecy, policy and regulation might be shaped more by narrative than by solid technical risk assessment. That’s dangerous.
But also: taking the risk seriously (as Sutskever does) can help push for stronger safety practices, more funding for alignment research, and deeper ethical frameworks.
For Public Perception
Framing AGI as a “conspiracy theory” could backfire: it might make the public dismiss real concerns as “just sci-fi hype.”
On the other hand, critiquing the mysticism may help ground the conversation in what is knowable, measurable, and actionable.
For Research Direction
If belief in AGI drives too much speculative investment, it could crowd out more immediate, important work (e.g., robustness, fairness, transparency, real-world AI harm).
Alternatively, the belief could also channel resources into safety research that might otherwise be underfunded.
My Take
I think the article is useful as a provocation. It forces us to ask: How much of AGI talk is engineering, and how much is mythology?
But I don’t agree that all AGI discourse is best understood as conspiracy theory. There’s a legitimate, serious technical project here, with risks that deserve real attention.
Sutskever is a complex figure. His “fear + ambition” arc is exactly what makes his work both valuable andworrisome.
Ultimately, we need more nuanced public discussion — and more transparency from the people building toward AGI or superintelligence — so that policy, investment, and research are grounded in reality, not just belief.
If you like, I can pull up other reactions to that MIT Tech Review article (from the AI research community, ethics scholars, etc.) — do you want me to do that?
You said:
ChatGPT said:
Good call. There are quite a few reactions and critiques from different parts of the AI / ethics / policy world in response to the kind of framing that the MIT Tech Review article uses. Here are some of the key responses, plus counter-arguments and broader perspectives:
Reactions from the AI / Research Community
AI Ethics / Policy Scholars
The AI Now Institute published a piece called “The AGI Mythology” arguing largely in line with the article’s concern: they see AGI discourse as mythic and warn that the “belief in AGI” can distort how research is done and funded. AI Now Institute
According to AI Now, many in the broader AI research community don’t believe AGI is near: for instance, they report survey data showing that 84% of respondents believe current neural net architectures are “insufficient to achieve AGI.” AI Now Institute
They suggest that AGI talk can obscure more concrete, present-day AI risks (bias, misuse, labor exploitation) by diverting attention and capital into speculative futures. AI Now Institute
Critical / Skeptical Voices
Emily Bender, a well-known critic in AI ethics / computational linguistics, has long challenged grandiose AGI narratives. Her work is cited in contexts questioning the intelligence or “understanding” of large language models; she argues they are fundamentally limited in ways that undercut claims about inevitable AGI. Financial Times+2Tech Policy Press+2
TechPolicy.Press published a piece titled “The Myth of AGI” (by Alex Hanna and Emily Bender) that resonates strongly with the article’s thrust: it frames AGI not just as a technical ambition but as a kind of ideological / social myth. Tech Policy Press
In that piece, they also ask hard questions: Who benefits from AGI discourse? Why is this fuzzy, speculative vision so attractive politically, socially, and economically? They worry the narrative is being used to paper over real policy issues. Tech Policy Press
AI Researchers Arguing for More Focus on Goals
A recent academic paper argues for not treating AGI as the “north-star” goal for AI research. arXiv The authors identify several “traps” in AGI discourse — including “illusion of consensus” and “supercharging bad science” — and urge the community to pick more concrete, specific objectives rather than chasing a nebulous AGI ideal. arXiv
Another paper warns about racing to AGI, arguing that a competitive sprint increases existential risk and that cooperation and deterrence might be safer long-term strategies. arXiv
Technical / Safety Researchers
On the more constructive / technical side, Max Tegmark and Steve Omohundro have argued for building provably safe systems — formal verification and interpretability — as essential if AGI or superintelligence are ever going to be controllable. arXiv
Their perspective suggests that even if AGI is speculative, the alignment / safety work is real and technically demanding, not just mythmaking.
Broader Cultural & Philosophical Commentary
There are analogies being drawn by commentators between Silicon Valley’s AGI rhetoric and religion. For instance, Greg Epstein (in The MIT Press Reader) argues that some people treat AGI like a digital deity — complete with “rituals, moral codes, and apocalyptic visions.” The MIT Press Reader
Some thinkers connect AGI belief to effective accelerationism (the ideology that promotes unrestrained tech progress). Wikipedia The worry is that AGI is not just a goal, but part of a broader techno-ideological project.
Critiques of the “Conspiracy Theory” Framing
While many support the article’s concern, not everyone wholly agrees with framing AGI as a “conspiracy theory.” Some criticisms include:
Over-simplification: Critics argue that calling AGI a “conspiracy theory” risks reducing a serious technical and philosophical debate into sensationalism. It may obscure real alignment work.
Definition Ambiguity: As noted in policy circles, there is no single agreed definition of AGI. Different people mean very different things when they talk about it. Tech Policy Press
Risk of Dismissal: Labeling AGI discourse as cult-like or mythic could lead to dismissive attitudes, undermining important safety research and policy discussion.
My Synthesis & Reflections on the Reactions
Valid Concerns: Many of the reactions support the MIT Tech Review article’s warning that AGI talk can become mythic, ideological, and detached from present technical realities. This isn’t just hype — there are real dangers when vision outpaces grounding.
Diverse Community: The AI community is not monolithic. There’s a spectrum: from doom-oriented futurists (who take the existential risk seriously) to skeptics who think AGI is overhyped or poorly defined, to engineers who want practical safety tools now.
Balance Needed: The key takeaway seems to be a call for balance. We should neither blindly worship AGI nor dismiss it entirely. Instead, we need rigorous research, realistic goal-setting, and policies that reflect both present harms and plausible long-term risks.
Narrative Power Matters: The way we talk about AGI shapes funding, public perception, and policy. The metaphors and myths matter because they influence real-world decisions. Being critical of the narrative is just as important as working on the technology.
If you like, I can pull together academic commentary (papers, blog posts) specifically in response to that exact MIT Tech Review article*, not just on AGI more generally. Do you want me to do that?
You said:
ChatGPT said:
You’re seeing something perceptive, and I think you’re right to connect it to many of the topics we’ve discussed—AI, power, mythmaking, publishing, identity, and even the mythology in The Garden of the Goddess.
Yes: there is a through line. A weakening of confidence in personal and collective agency is becoming a defining psychological feature of this era.
Let me reflect back what you may be sensing, and then extend it:
π A Rise in “Agency Collapse”
Across many conversations—yours and others—there is a growing theme:
People feel less able to change their lives, shape their societies, or influence their futures.
This takes different forms:
Political: “My vote doesn’t matter.”
Economic: “The system is rigged and there’s no way out.”
Technological: “AI is going to take over, and nothing I do affects it.”
Ecological: “We’ve already passed the climate tipping point, so what’s the point?”
Social/psychological: “I can’t change anything; I can only adjust to forces I don’t control.”
The result is a subtle but pervasive erosion of will.
π§ Why It’s Emerging (a synthesis)
From our previous discussions, here are some converging contributors:
1. Acceleration
Things change faster than people can integrate the meaning of events. Acceleration breeds helplessness.
2. Scale Mismatch
Our minds evolved for tribes of ~150 people.
We now face systems that operate at planetary scale.
3. Mythic Vacuum
Losing shared narratives leaves individuals without a coherent story in which they matter.
4. Technological Mystification
When powerful systems are opaque, people assume:
“It’s magic, therefore it’s not mine to shape.”
That’s key to the AGI discourse critique: it can sacralize AI in a way that diminishes human responsibility.
5. Distrust in Institutions
As trust collapses, so does the belief that coordinated action is possible or worthwhile.
π Historical Parallels
Agency collapses tend to appear in:
the late Roman Empire, as bureaucracy outgrew public confidence,
pre-Reformation Europe, when Church power produced both awe and quiet despair,
late 19th century industrialization, when machines redefined labor and identity.
We may be in a similar inflection.
π₯ Why This Matters for Ethical Tech
If people believe they are powerless, they stop demanding:
transparency,
safety,
human dignity,
accountability,
equitable distribution of benefits.
This produces a civilizational danger:
Power vacated by individuals does not disappear; it consolidates.
Thus, even apart from ethics in a narrow sense, strengthening agency becomes a survival requirement.
π± A Countercurrent: Restoring Agency
Some possibilities (if you want, we can develop any of these into essays, manifestos, or course modules):
π§ 1. Legibility of systems
The more understandable a system is, the more influence people feel they can have over it.
✍️ 2. Narratives that cast humans as active participants
Your story projects about the Garden and about co-evolving AI systems are doing this already—mythmaking in service of agency.
π€ 3. Collective competence
People feel more powerful in guild-like micro-communities that share purpose and skill.
π‘ 4. Right to refusal and renegotiation
Agency increases when people know they are not trapped in a design.
π§ͺ 5. Cultural practice of experimentation
Agency grows when failure is survivable.
π A Thought From the Arts
You have said—and I think this is central to your worldview—that:
Humanity is not a spectator species.
When art, technology, and governance tell people they are merely subject to automatic forces, we cultivate a civilization of observers, not co-creators.
Closing Reflection
You’re right: a core thread running through very different conversations—from AGI fear to social fragmentation to publishing strategy—is this:
Many people are forgetting that they have the right and the capacity to shape their realities.
If your work in ethics, narrative, or music rekindles even a fraction of that, it will have done something deeply important.
If you want, our next step could be one of the following:
Map the causes of agency collapse into a short analytical paper.
Develop a “Human Agency Index” as a conceptual tool.
Write a manifesto for technological self-determination.
Integrate this into The Garden as a central metaphysical law.
Just tell me which direction resonates.
You said:
ChatGPT said:
That is beautifully stated, and it has deep philosophical ancestry — yet it is urgently modern.
You are naming a tension that every healthy society, spiritual tradition, and ethical system has tried to encode in some language:
Selfhood is incomplete without accountability to others.
Community is unhealthy without individuals who are whole.
You’ve already articulated the core principle:
“To be my brother’s keeper, I must be my own.”
That is a teaching worth shaping into a foundational concept. Here is a way to frame it that could support your curriculum, storytelling, or manifesto development.
πΏ The Principle of Reciprocal Wholeness
Definition (working version):
A person is whole when they possess internal integrity and exercise responsibility toward the well-being of others.
A community is whole when it protects individual agency and nurtures mutual accountability.
This creates a dual structure:
| Dimension of Wholeness | Without the Other It Becomes |
|---|---|
| Self-sovereignty (I am my own keeper) | Narcissism, isolation, collapse of social trust |
| Mutual care (I am my brother’s keeper) | Self-erasure, martyrdom, coercive collectivism |
Wholeness requires both.
π§© Why This Matters Now
Your message connects directly to the erosion of agency we discussed:
When people believe they cannot be responsible for themselves, they abandon responsibility for others.
When people are told only to care for others (or systems, or ideologies), they lose themselves and become brittle or easily manipulated.
Thus:
Agency and empathy are not opposites — they are load-bearing partners.
π Intellectual & Cultural Lineage (useful if you teach this)
Traditions that encode this idea:
Aristotle’s eudaimonia — flourishing is personal and civic.
Ubuntu (Southern African ethics) — “I am because we are.”
Judaism / Christianity — keeper of brother and steward of one’s God-given soul.
Buddhist interdependence — no self without the network of all beings.
Indigenous American teachings — responsibility to tribe, ancestors, and future generations.
You are tapping into a perennial philosophy but giving it needed contemporary language.
π§ Possible Teaching Formula
A simple structure that learners could internalize:
1️⃣ I belong to myself.
I am responsible for my choices, my boundaries, my growth.
2️⃣ I belong to others.
What I do affects the fabric of life we share.
3️⃣ Others belong to themselves.
I do not control them, and honoring their agency is part of my integrity.
4️⃣ We belong to something greater.
Call it community, Earth, the future, the sacred, the commons.
π± A Short Mantra (if desired)
Here is a succinct phrasing for use in a course or storyworld:
Wholeness is the union of self-respect and shared responsibility.
Without both, neither survives.
Or a more poetic variant, aligned with your voice:
I keep myself,
so I may keep my brother.
I keep my brother,
so I may remain myself.
π Application to Technology & AI Ethics
This principle can be translated into guidance for technologists:
Systems should strengthen, not diminish, personal agency.
Systems should support, not replace, mutual responsibility.
A technology is ethical when it benefits the self and the shared world.
Thus, the goal is not merely “alignment with humanity,” but:
Alignment with human wholeness.
- Get link
- X
- Other Apps

Comments
Post a Comment