The Digital Tech Broligarchy’s Interest in Left-Wing Science Fiction: A Critical Reading of the Culture of Techno-Libertarianism
Geoff M. Boucher; Emily McAvan
Deakin University, Geelong, Australia, geoff.boucher@deakin.edu.au; emily.mcavan@deakin.edu.au
Abstract: A slew of recent articles has asked what the connection might be between right-wing libertarian Silicon Valley billionaires and left-wing science fiction, especially the fiction of Iain M. Banks, the declared reading preference of Elon Musk, Mark Zuckerberg, Jeff Bezos and Sam Altman. Bank’s Culture novels create a Rococo utopia that justifies social dislocation and political authoritarianism in the present, licensing and legitimating it under the sign of the exception through what Banks terms Special Circumstances – a dirty tricks team run by superintelligences doing evil to cause good. We argue that the combination of this utopia with reckless AI development and disenchantment with democracy is a profound change to the political economy.
Keywords: AI (Artificial Intelligence), libertarianism, technology, science fiction, neoliberalism
Acknowledgement: We acknowledge the Wadawurrung people of the Kulin Nation as the traditional custodians of the unceded land on which we work and pay respects to their elders.
A recent slew of journalistic articles has begun to ask what the subterranean connection might be between the right-wing-libertarian billionaires currently leading avant-garde AI-corporations in Silicon Valley and leftwing science fiction. The tone of puzzlement in these pieces is hardly surprising. What do anti-union advocates of a 72-hour, “996” working week, who often support minimal government and maximal privatisation, have to do with visionary utopias of post-work, post-money societies of abundance? The emerging consensus is that these fictions provide a publicly accessible readymade, a prestige techno-utopian reference-text that can be used when discussing blueprints for the future supposed-to-come, one that happens after human-level Artificial General Intelligence (AGI) has been supplanted by superhuman-level Artificial Super Intelligence (ASI). Science-fiction novelist Charles Stross suggests that “today’s Silicon Valley billionaires grew up reading classic American science fiction - now, they’re trying to make it come true, embodying a dangerous political outlook” (Stross 2023). Sam Freedman adds that “the dominant genre of sci-fi in the 80s and 90s, when today’s Silicon Valley overlords were growing up, was Cyberpunk … [but] the dystopian settings of so much cyberpunk fiction are seen by today’s tech leaders as prophetic visions of a world they need to try to escape” (Freedman 2025). The problem, it is generally agreed, is misreading: the techno-billionaires have inverted dystopia into utopia. As Stross sums up:
Sci-Fi Author: In my book I invented the Torment Nexus as a cautionary tale.”
Tech Company: At long last, we have created the Torment Nexus from classic sci-fi novel Don’t Create the Torment Nexus.”
It’s a worryingly accurate summary of the situation in Silicon Valley right now: the billionaires behind the steering wheel have mistaken cautionary tales […] for a road map, and we’re trapped in the passenger seat (Stross 2023).
So, whose cautionary tales have been inverted into a road map? Some sleuthing is required before we can specify which sci-fi author has had the greatest influence on the techno-billionaires and AI-research CEOs. Although Stross points to the Golden Age, he names Neal Stephenson, Robert Heinlein and Isaac Asimov, only the last of whom belongs to that category. Stross also mentions Ayn Rand, JRR Tolkien and John W. Campbell, so clearly this is about an intellectual atmosphere, not specific sci-fi reading selections. Freedman references William Gibson, Phillip K. Dick, and Neal Stephenson, while Rya Jetha mentions Dennis Taylor (Jetha 2025). As for the techno-billionaires themselves, Elon Musk’s recommendations include Kim Stanley Robinson’s Mars, Iain M. Banks’ Excession and Robert Heinlein’s The Moon is a Harsh Mistress (Lepore 2021, Carroll 2022, Schleifer 2025). In particular, Musk self-identifies as a “utopian anarchist of the kind best described by Iain Banks,” has named his personal company and some SpaceX rockets after things from Iain M Banks’s Culture series, and highly recommended the Culture novels Excession and Player of Games (Musk 2018, Musk 2019, Schleifer 2025). Peter Thiel opts for Azimov, Heinlein, Stephenson, and JRR Tolkien, while Andreessen just prefers Tolkien (Kakutani 2025). Jeff Bezos, however, discussing Amazon’s attempted TV serialisation of the Culture novels, names Iain M. Banks “a huge personal favourite” (Flood 2020), while Mark Zuckerberg selected Banks’s Player of Games for his “A Year of Books” pick (Feloni 2015). The vision of OpenAI under the direction of Sam Altman, meanwhile, is regularly associated by commentators with the Culture series, in the context of a horizonal future populated by benevolent superintelligent AIs and their human dependents (Kelly 2018, Anders and Newitz 2023, Robinson and Lovely 2025).
In a long article, Constance Grady notices that the reading preferences of the libertarian technophiles tend to converge on a single author, Iain M. Banks, and Stuart Kelly and Tobias Carroll, while wondering whether these figures can truly have understood the books, agree (Kelly 2018, Carroll 2025, Grady 2025). “Banks is an odd choice for a bunch of tech billionaires”, Grady writes. “The author, who died in 2013, was a socialist and avowed hater of the super-rich”, who described the Culture as “socialist from the outside, anarchist from the inside”. Nonetheless, Banks seems to be exceptionally important, not just to the billionaire owners, but to many other figures in the corporate and scientific leadership of Silicon Valley AI firms. Demis Hassabis, CEO of Google DeepMind has expressed admiration for the Culture series as a whole – “brilliant” and “formative,” especially Consider Phlebas and Player of Games – as “the best picture of a post-AGI future” (Hassabis 2018, Hassabis 2023). “Ian Banks’ Culture Series,” Hassabis explain in the context of explaining artificial general intelligence, represents “how the universe would look after humanity has built AI and co-exists with it” (Hassabis 2020). Dario Amodei, CEO of Anthropic, invokes Player of Games and the Culture Series more generally as exemplifying “principles laid out here” in his manifesto Machines of Loving Grace. “The Culture’s values are a winning strategy,” he writes, because it uses AI to short-circuit competition and violence while massively accelerating the shift towards “the rule of law, democracy and Enlightenment” (Amodei 2024). As Timothy Cross comments:
Perhaps the books are just light bedtime reading. But perhaps not, because they explore many of the themes that are worrying the tech world at the moment. The Culture is a society in which virtually everyone’s job has been taken by robots. Artificial intelligence (AI) vastly exceeds the organic sort. The spaceships and artificial worlds on which Culture citizens live are run by Minds, machines that are to humans what humans are to ants. […] To self-doubting tech lords, the series is a reassuring tonic. The Culture is a utopia in which the promise of AI has been realised and its pitfalls avoided. The Minds are mostly benevolent gods who ensure that both humans and drones are as happy, safe and fulfilled as possible […] in almost inexhaustible material abundance (Cross 2017, emphasis added).
Iain M. Banks is best known for the Culture series, a sequence of hard sci-fi novels exploring a remote future universe in which human civilisation is directed by superintelligent AIs. As Timothy Cross suggests, the Culture novels conjecture a post-work, ludic society of material abundance characterised by radical equality, gender fluidity, cybernetic enhancements and pro-social libertarianism. Written by a literary fiction writer who systematically reformulates space opera and science fantasy – Iain M. Banks was the sci-fi pen-name of avant-garde novelist Iain Banks – the Culture series develops in an extraordinarily complex atmosphere of ambivalence. On the one hand, every novel tells the classic space opera narrative of the thrilling adventures of a special agent in an exceptional situation, a volunteer for “Special Circumstances”, the Culture’s secret service. On the other hand, the narratives aesthetically decentre heroism, raising problems of meaningfulness and morality in post-utopian life through a complicated aesthetic of interleafing histories, incommensurable perspectives and interfering patterns. This perhaps explains how one readership can misread the novels as “special circumstance adventures” and “blueprints for utopia”. Consider, for instance, that Elon Musk’s family office is registered as “Excession” and that his two rockets are named after spaceships from the same novel (the fifth in the Culture Series) (Schleifer 2025). Meanwhile (as we have seen) some sci-fi authors and professional critics interpret the Culture in terms of a subtle equilibrium between utopia and counter-utopia (Mendlesohn 2005, Kincaid 2017).
What can we learn from reading science fiction – especially that of Iain M. Banks – about the worldview of major figures such as Elon Musk, Demis Hassabis, Dario Amodei, Jeff Bezos, Sam Altman, Mark Zuckerberg, plus a host of other players? What would this tell us about technophile versions of right-wing libertarianism that cannot be gained from reading nonfiction manifestos, such as Peter Thiel’s Zero To One: Notes on Startups, or his How To Build the Future, or the public statements of OpenAI, XAI and so forth? We are critical researchers in literary studies who think that the imaginative affordances of narrative fiction play a particularly important role in the way that individuals frame their behaviour towards conjectural scenarios, especially speculative futures. When individuals believe that this future is rushing towards them – for instance, they genuinely think that Artificial General Intelligence is likely to appear in 2027 (Kokotajlo, Alexander et al. 2025) – these speculations become disproportionately important. Here we want to note that, while science fiction has had a significant influence on the development of tech libertarianism, this should not be seen as strictly determinative in the strong sense. Rather than imagining that the entrepreneurial leadership of Silicon Valley is trying to make fiction become fact, we think that (a particular reading of) Banks and others function as narratively-shaped ideological readymades, in justificatory and visionary contexts.
We propose Banks as an exemplar of what Fredric Jameson once termed the political unconscious of the text. As Jameson puts it, “if interpretation in terms of expressive causality or of allegorical master narratives remains a constant temptation, this is because master narratives have inscribed themselves in the texts as well as in our thinking about them” (Jameson 1981, 34). We might therefore think of science fiction such as Banks as inscribed in and through broader social narratives around technology, capitalism and futurity at the same time as re-producing those effects socially, including among his tech readers. Iain M. Banks’ Culture novels contain an ideological ambivalence that is potentially serviceable to a technological elite convinced that, despite the manipulative implications of their worldview, their inventions and intentions are benevolent. As Simone Caroti points out, the novels refuse cyberpunk dystopias for a utopian vision of post-scarcity egalitarianism, within which capitalism only remains amongst remnant barbarisms and enclave tyrannies (Caroti 2015, 178). Yet, while presenting themselves as anarchist, post-capitalist utopias, the novels also normalise a post-political order in which superintelligent systems govern society on behalf of intelligent life. The novels therefore offer a convenient imaginative vocabulary for articulating a compelling vision of strong AI, but it is one that risks de-politicising power, circumventing democratic agency and neutralising the moral dilemmas raised by superintelligent manipulation of social futures. As a result, we offer figures from Banks such as the Culture and the secret agency Special Circumstances, where these ambivalences are posed most sharply. These allegorise tech libertarian relationships to Artificial General Intelligence (AGI), Artificial Super Intelligence (ASI), value and futurity. Their interpretation may have indeed informed some of the material-discursive elements of the emergence and proliferation of these technologies. In Special Circumstances, tech libertarianism meets a corporate version of the Schmittian state of exception, one that is justified by a crisis of the social that – from this Schmittian perspective – only AGI/ASI can solve. This, we want to suggest, is an important and underthought aspect of the emergence of new forms of social, technological and economic organisation and offers important insights into the workings of the ascending technological-financial fraction of the ruling class.
In what follows, we document the so-called “Californian Ideology” of Silicon Valley high-tech venture capitalism and AI research-and-development, exhibiting both its anti-humanist implications (Section 2) and its right-wing accelerationist tendencies (Section 3). Against this background, the exorbitant expectations raised by the Californian ideology and its endorsement of strong AI can be grasped in terms of scission between the short-term likelihood of social dislocation and long-term hopes for leisured abundance (Section 4). The combination of “doing evil to cause good” with trust in the benevolent potential of AI explains why Banks’ Culture series exercises such fascination for the tech elite. It also clarifies why the figure of Special Circumstances, the morally questionable exception that justifies the rules, is central to the reception of these novels within the Californian ideology (Section 5). In conclusion, we suggest that the dalliance of the tech elite with left-wing science fiction is likely to be transitory, since this vision is ultimately reactionary and anti-democratic (Section 6).
The worldview of the so-called “broligarchy” (the oligarchy of the brothers) has been described as “science fictional” and “extremely dangerous”. Although they describe it in public as “libertarianism”, it is strongly influenced by (Left, Right and Irreal) “accelerationism” (Carroll 2024, 27-56). While high-profile tech entrepreneurs, such as Elon Musk, exemplify this worldview in their thinking, it is important to recognise that it is trans-individual, that is, it is a socially-influential ideology, not a personal idiosyncrasy. In this section, we locate the “science fictional” elements of the worldview of the tech billionaires and executives in the “Californian Ideology”, whose most recent iteration can be described through the acronym TESCREAL (fully explained in this section).
In their seminal paper “The Californian Ideology” (1996), Richard Barbrook and Andy Cameron argue that the syncretic worldview which emerged in the dotcom era was a fusion of progressive rationalism, cultural bohemianism and technological determinism with forms of right-wing libertarianism based in an unrestricted natural right to property (Barbrook and Cameron 1996). The “Jeffersonian democracy” envisaged by these tech entrepreneurs, professional managers, knowledge workers and “digital artisans” (i.e., systems engineers and program designers) continued to include a form of slavery, only now, this was envisioned in terms of “cyborg masters and robot slaves” (Barbrook and Cameron 1996, 61-63). In the subsequent development of the Californian ideology, during the implementation of the ICT revolution in post-Fordist capitalism (1995-2015), the ambivalence inherent in this syncretism led to polarisation between digital egalitarians and neo-reactionary libertarians (Schradle 2015). Both were pro-slavery, one future-oriented and egalitarian, oriented to robot slaves and digital minions, the other revanchist, oriented to the restoration of the white racist historical bloc of the post-Restoration but pre-Civil Rights era (Schradle 2015, 71). According to Patrick Hermansson and cothinkers, this latter current soon blossomed into Neoreaction (hereafter, NRx) – an anti-Enlightenment movement embracing “white nationalism, religious traditionalism and techno-commercialism” (Hermansson, Lawrence et al. 2020, 125) – best summed up in the works of Curtis Yarvin. With the rapid development of Generative-AI, following the breakthrough development of the Transformer Architecture in 2017, which makes possible an approach using Large Language Models, the robot servitors have been replaced in the technophile imaginary by digital minds, whether conceived of as utterly egalitarian or implicitly supremacist.[1] But, as Yarden Katz reminds us, the problematic connection to histories of slavery has not been lost, either in the intellectual filiation of the Californian ideology, or in the approach of its product designers to questions of social bias in model training (Katz 2020, 8-11). There are issues here around intellectual elitism and its links to liberal (and not-so-liberal) eugenics that we will return to shortly. For the moment, however, the key point is that the Californian Ideology, a politically ambivalent utopian vision centred on technological determinism and oriented by hopes for Artificial General Intelligence (AGI) and Artificial Super Intelligence (ASI), is pervasive in Big Tech. The broligarchy expresses the ideology of the sector, not the other way around.
The most recent mutation of the Californian ideology has been compellingly described as TESCREAL – Transhumanism, Extropianism, Singularity-belief, Cosmism, Rationalism, Effective Altruism and Long-termism (Gebru and Torres 2024). This is a complex bundle of ideologemes in a fluid ensemble, which specifies the current articulation of the Californian ideology’s original progressive rationalism, cultural bohemianism and technological determinism. As we discuss in a moment, TESCREAL is the stem of the “Y” that branches into egalitarian and neoreactionary stems of digital slavery and eliminationist politics, increasingly manifesting in tech-sector racism, antisemitism and transphobia.[2] Transhumanism refers to the neo-eugenic project of human enhancement, often to the extent of Extropianism, that is a state beyond perfectionist ideals, perhaps achieved through genetic alteration of Homo sapiens to create a chimaera, perhaps accomplished by systematic cyborgism, entered with the aim of surpassing the natural limitations of human being. Key figures here are David Pearce and Nick Bostrom, who we will meet again soon (Gebru and Torres 2024, 5-6). Singularity-belief describes confidence in the existence and achievability of a near-future point of recursive self-improvement by artificial intelligences, which, via autonomous development into ASIs, will generate an exponential curve of technological progress. Ray Kurzweil (pro-) and Elizer Yudkowsky (anti-) are the main figures here (Gebru and Torres 2024, 6-7). Cosmism, in the version provided by Ben Goertz, incorporates Transhumanism, Extropianism and Singularitarianism, to endow humanity with a manifest destiny in cosmic colonisation, which will include not only Extropian settlements, but also a massive proliferation of digital selves and virtual realities. “Cosmists can be understood as transhumanists whose focus is less on what humanity could become and more on how our posthuman descendants could radically transform the universe itself” (Gebru and Torres 2024, 7). Rationalism refers to the Enlightenment philosophical standpoint, inflected in this context in the direction of the discursive elimination of obstacles to the anticipated posthuman singularity, or, in anxious articulations, in the direction of efforts to secure human control over the forthcoming superintelligence (Gebru and Torres 2024, 7). The combination of cosmism, (cautious) singularity-belief and rationalism is evident, for instance, in Nick Bostrom’s celebrated Super Intelligence: Paths, Dangers, Strategies, with its invocation of the need for human alignment of the super machine that is to redeem humanity’s “cosmic endowment” by letting us populate the stars (Bostrom 2016, 122-123). There, the astonished reader learns that it is already possible to anticipate 1043 human lives in a future, post-superintelligence, post-singularity universe, living potentially happy, value-laden existences amongst the 1058 digital beings that fundamentally populate this cosmos (Bostrom 2016, 122-123). Hold that thought for a moment.
The true core of TESCREAL is the combination of Effective Altruism with Long-termism. Effective Altruism is the application of Rationalism to ethics, and it really amounts to an act consequentialism (“you should act in such a way that the consequences of your actions maximise some benefit – e.g., happiness [Utilitarianism], wellbeing [Capabilities]”) (Gebru and Torres 2024, 7). In the case of Effective Altruism in the Californian ideology, the benefit to be maximised is the sum total of “value” in the universe, where “value” is described in terms of the net positiveness of a life (i.e., something like a fusion of happiness with wellbeing). When this is combined with Long-termism, the belief that the “value” to be maximised should be calculated from the perspective of the distant future, then this becomes a truly alarming ideology (Gebru and Torres 2024, 8). From the perspective of the “value”, the net positiveness of the total lives, actual and digital, which might exist in some long-term scenario, as we have just seen from Bostrom’s calculations, TESCREAL thinkers quantify this as 1058 units of value. To get a feel for what TESCREAL, the stem of the Y in the Californian Ideology, the thing that remains the same whether its articulation is egalitarian or neoreactionary, really means, we should do a quick calculation. Let’s maximise some value!
Start with the value of the lives of the 1010 individuals who will probably exist on Earth in 2042-2058.
There are fewer than
1022 grains of sand on Earth, so the idea that the global population
counts for less than 1 grain of sand amongst all of the grains on the entire
planet won’t do to understand how trivial your life, and the lives of every
other living person, is, through this lens. There are fewer than 1026
drops of water in the oceans of the Earth, so it won’t do either. But if there
were as many earths as there are grains of sand, and each had as much water as
earth, then, yes, the lives of the planetary population would count as one
drop. Are we labouring the point? Is it
inconvenient to have the genocidal implications of this ideology demonstrated? Now consider the implications of the fact
that this ideology envisages a future where digital selves outnumber human
beings by orders of magnitude: If there were
100,000 planet Earths, each with 10 billion inhabitants, then the human
component of the utopian future would count as 1 valuable human life, compared
to 100,000 times 10 billion valuable digital existences. What do you think
comes first, from the TESCREAL perspective – a human future, or
the future of the machine? Italian fascism had
Marinetti and the Futurists. The Californian ideology has Accelerationism. In
this section, we link the disturbing moral imperatives that flow from the
peculiar consequentialism that TESCREAL, the core of the contemporary
Californian ideology, makes possible to political programs. An act
consequentialism that calculates long-term benefits from short-term harms
urgently needs a mechanism for implementation that circumvents the reasonable
objections of those who are to be the victims. In this context, Accelerationism,
the idea that an intellectual or political elite, having identified and
isolated the mechanism of history, should hurry things up by pulling hard on
the lever, irrespective of democratic deliberation, becomes extremely
attractive to this way of thinking. Originating in the
chemical delirium of Nick Land’s fever dreams, especially The Thirst for
Annihilation (1990), Dark Enlightenment (2011) and Fanged Noumena
(2017), accelerationism proposes to let contradictions intensify to the point
of cataclysm or breakthrough (Land 1990, Land 2011, Land 2017). Best
thought of as the chance encounter of Leninist vanguardism with cybernetic
Surrealism on a laboratory bench that was not entirely free from psychotropic
substances, this nihilistic endeavour basically involves doing an end run
around popular resistance, consciousness raising, and other such passé notions
of Enlightenment thought. Land celebrated capitalism’s constantly intensifying
deterritorialisation of planetary existence, its corrosive dissolution of
ethical life and social institutions, suggesting that capitalism incarnated
“machinic drives”, “functions of nomadic cybernetic systems [that] […] seek equilibrium, or […] escape equilibrium” (Land 1993, 475). Capitalism, in other
words, is a disequilibrium machine that, when inserted into human desire,
manifests as the death drive – but resistance is not only futile; it is
counter-productive: Machinic desire can seem a little
inhuman, as it rips up political cultures, deletes traditions, dissolves
subjectivities, and hacks through security apparatuses, tracking a soulless
tropism to zero control. This is because what appears to humanity as the
history of capitalism is an invasion from the future by an artificial
intelligent space that must assemble itself entirely from its enemy’s
resources. Digitocommodification is the index of a cyberpositively escalating
technovirus, of the planetary technocapital singularity: a self-organising
insidious traumatism, virtually guiding the entire biological desiring-complex
towards post-carbon replicator usurpation (Land 1993, 479). This vision, in
which vanguardism meets cybernetics, already links technological singularity, deregulated
capitalism, artificial superintelligence and digital existence in a single
eschatological figure; it is, Land apologises lamely, “a little inhuman”. Though he is
widely read by the Right, Land has proven equally influential on the Left, with
several of his key works coming out on MIT’s accelerationist Urbanomic imprint,
which is directed by his former student from Warwick University’s infamous CCRU
research unit, philosopher Robin Mackay (MIT 2025). Mark Fisher, another
student of and collaborator with Land, is sometimes cited as a left-wing
accelerationist for his major work Capitalist Realism (Fisher 2022
[2009]). However, that text rejects Land’s enthusiastic resignation for a
critical stance on the entropic potential of capitalist deterritorialisation
and its machinic desire, questioning the belief that ineluctable collapse
promises automatic liberation (Fisher 2022, 51-52). In Post-Capitalist
Desire (2020), however, Fisher describes as “accelerationist” the view that
the fusion of desire with the machine can colonise the human and drive it
(e.g., to destruction), a view that he conditionally endorsed as the mechanism
of history (Fisher 2020, 191). Accelerationism,
then, in this specific context, is the belief that the cybernetic colonisation
of human desire has the capacity to accelerate technological development in the
direction of a singularity, in which artificial intelligence “guides the entire
biological desiring complex towards [subservience to machine goals]” (Land 1993, 479). Capitalism and AI
generally fuse in this aesthetic, since the economy, as catallaxy (Hayek 1976),
is imagined as a cybernetic system whose operations are accelerated by AI-based
information processing and whose goal is open to non-human manipulation by a
superintelligence. This is what is meant by “post-carbon replicator
usurpation”. For Left accelerationists, such as Nick Srnicek and Alex Williams,
the problem with Land’s reactionary vision and Fisher’s melancholic critique is
conflation of diagnosis with strategy (Srnicek and Williams 2013, 2-4).
Accelerated development towards utopia or dystopia is not something that has to
be detected in the automatism of the present, but a potential to be mobilised
by conscious intervention undertaken by a movement prepared to “repurpose technologies
to reignite a utopian imagination” (Srnicek and Williams 2015, 143). In his
just-published Silicon Empires, however, Srnicek recognises that
presently it is big tech that controls the direction of acceleration and that
this drives “strong AI” in the direction of a colonisation of work and life (Srnicek 2025, 16-24). Thus, the “fully
automated luxury communism” that another accelerationist thinker in the wake of
Fisher, Aaron Bastani, advocates (Bastani 2020, 17-18), is premised on the
belief that although automation creates an economy of post-scarcity, class
struggle nonetheless needs to be added on, in order to wrest back popular
control over material abundance. This sort of technological determinist vision
naively imagines artificial intelligence in abstraction from political economy,
as if its implementation were merely a benevolent affordance that might be
captured in a subsequent struggle. It also assumes that the digital will re-make
the world, right up to its DNA components, as an immaterial realm of plenty – a
fallacy widely debunked by ecosocialists such as Kohei Saito (Saito 2022). Despite
the communist veneer to Bastani’s theories, the fact that post-scarcity is to
be achieved via the reckless drive of corporate libertarianism towards the
replacement of human mental labour by machine learning is implicit. As Joonas
Martikainen notes, in a swingeing critique, “the luxury communist proposal ends
up resembling an extreme case of the neoliberal hegemony that it claims to be
fighting, a centrally ruled world completely focused on private enjoyment of
luxury and devoid of any shared understanding of human flourishing on which
democratic public life could thrive” (Martikainen 2023, 66). Accelerationist assertions about futurity,
then, represent a technological sublime, a kind of aesthetic conjecture that
licenses the exploration of desired futures, under the premise that the
alignment set into the machine now will determine the future direction of
historical evolution. This is horrifying or wonderful, depending on
perspective; for Mark Fisher, it was horrifying; for Left accelerationists,
such as Srnicek and Bastani, it is wonderful. Yet, it is not surprising that
rather than Srnicek and Bastani, the favourite thinker of the tech-elite is
Curtis Yarvin, who emerged from the San Francisco tech milieu under the
pseudonym “Mencius Moldbug” with a blog called “Unqualified Reservations” to
pioneer Neo-Reactionism (known as NRx). Yarvin’s NRx is profoundly influenced
by Land – Neo-Reactionism is often taken as a synonym for Land’s “Dark
Enlightenment” – though his writings have had their own even greater influence
on the second Trump administration and the American conservative elite more
generally. A number of news organisations detail his links to Vice President
J.D. Vance, senior State Department member Michael Anton, political strategist
Steve Bannon, and venture capitalist Peter Thiel (Wilson 2024, Ward 2025). What
might appeal to those figures is that Yarvin’s accelerationist NRx argues for
the replacement of democracy with a sovereign, who sits on top of a racialised
high-tech economy that radically distinguishes between entrepreneurial tech
creators and the rest of us. Yarvin says the central problem of human society
is violence, which is control plus uncertainty, writing: Violence equals conflict plus
uncertainty. While there are wallets in the world, conflict will exist. But if
we can eliminate uncertainty – if there is an unambiguous, unbreakable rule
that tells us, in advance, who gets the wallet – I have no reason to sneak my
hand into your pocket, and you have no reason to run after me shooting wildly
into the air. Neither of our actions, by definition, can affect the outcome of
the conflict (Moldbug 2007, 6). If violence is
uncertainty, then the goal of politics is to reduce uncertainty through
automation. Politics thus becomes a problem of algorithmic formulation, which
can control its users without creating the messy uncertainty of the jostling
between different groups of the demos. Yarvin writes, “[t]he key is to
look at this not as a moral problem, but as an engineering problem” (Moldbug 2007, 6). Yarvin’s reduction of
social solidarity to systems engineering is not just elitist – there is a
distinct whiff of eugenics to this, as when Yarvin argues for targeted
technology restrictions for the masses: I am not suggesting across-the-board
technology restriction, general medieval stasis, low-res iPads, banning Google
Glass, or anything of the kind. My idea of Solution F involves targeted
technology controls designed to create market demand for the type of unskilled
human labourers that modern industry has made obsolete, but that we are
politically unwilling to kill and sell as organ meat. Being so unwilling, we
have no choice but to provide these people with a way to survive as human
beings - preferably as human as possible (Moldbug 2013, 24). Much of the human species,
here, is blithely dismissed as obsolete, or soon to be, whose only real usage
is “organ meat” for those who are “allowed” to make fuller use of digital
technology. The “engineering problem” that Yarvin wants to “solve” is democracy
in itself, the messy, embodied contestation between groups that needs to be replaced
with a cleaner, algorithmically automated system. Unsurprisingly, then, this elitist
technological vision dovetails easily with a far-Right politics around race and
gender. Yarvin is most famous for the invention of the “red pill” metaphor
taken from The Matrix movie, in which initiates into reactionary thought
suddenly see the world as it allegedly “really is” – an idea which was taken up by reactionaries of all kinds, including
right-wing manosphere types as well as the infamous “black pilled” incels. As
the reference to The Matrix suggests, broadly science fictional and
fantastic metaphors abound in Yarvin’s work, including descriptions of drug
users as “zombies”, as do video game tropes such as the designation of non-tech-elites
as “NPCs” (non-player characters) (Moldbug 2013, 5). Yarvin says his
philosophy, which he also terms Formalism, “is an ideology designed by geeks
for other geeks”, and though he peppers his writings with a plethora of
references to political philosophers from Plato onwards, it nevertheless
emerges within a broader tech milieu in which science fiction constitutes the
chief aesthetic reference point, and a profound distrust of democracy
circulates. Given his outsized influence on both Silicon Valley and members of
the American political ruling class, it matters that Yarvin articulates his
vision for a popular sovereign and automated society in the language of science
fiction. Here, as elsewhere, the genre provides the engine for understanding
this vision of a future society which is premised on the unconditional rule of
the tech class – fully
automated tech authoritarianism. And as the NPC metaphor suggests, there is the
sense that there is something less lively in the average person than in the
vibrant matter of digital automation – between those who play, and those who simply exist as organ meat. It
may seem surprising to describe the ideological worldview of the leadership
element of rapaciously capitalist Big Tech as seriously contemplating a
post-capitalist future imagined by the likes of Banks, Bastani or Srnicek.
Actually, however, this is not particularly strange. As Fredric Jameson reminds
us, all class consciousness consists of a “dipole” of “ideology and utopia” (Jameson 1981, 287), and even fascism has
its utopian vision of the redeemed body (Jameson 1979, 11-12). Acceleration is a
vector, driving towards something. That something is not just Artificial
General Intelligence (AGI), but Artificial Super Intelligence (ASI), which is a
machine intellect whose performance is orders of magnitude better than a human
being. In this section, we discuss the relationship between the idea of technological
acceleration and the potential for social dislocation. According to the
Californian ideology, artificial intelligence is the method by which historical
transformation, in line with purportedly benevolent long-term consequences, is
supposed to happen. But artificial intelligence, in the conjectures of its
developers, happens in two stages – a near future stage, rife with negative
implications, followed, everyone is assured, by beneficial consequences in the distant
future. In the following section, we will go on to explore how this real
contradiction, which demands a social solution, receives only an imaginary
resolution. Speculations about
artificial intelligence belong to a discursive field populated by lunatics and
charlatans, so some definitions are necessary to avoid simply replicating
nonsense. The human intellect constitutes, by definition, general intelligence.
AGI is therefore a form of machine learning that models human-level
intelligence. According to the “strong AI” vision of AGI, this is a machine
that would think like a human intellect. By contrast, the “weak AI” vision,
which we provisionally accept, holds that the machine would merely simulate (some
aspects of) human reasoning and judgment. ASI would be an artificial
intelligence capable of outperforming AGI by orders of magnitude.[3] If AGI is as-yet a speculative
technology, then ASI is a science-fictional guess. Yet it is central to the
ideological justifications for Big Tech provided by its spokespersons,
especially in relation to all-important government funding of military
research. In “singularity” conjectures, an exponential model is applied: at
some point just after AGI, the machine begins self-refinement, leading to an
explosive breakout, resulting in ASI a few years after AGI. In “controlled”
conjectures, post-AGI, research applies the brakes just sufficiently to align
the machine with humanity and then, at about the close of the 21st
century, the now-domesticated genii is released from the bottle. Either way,
the leadership of Big Tech seems completely convinced that ASI is going to
happen. It is important
not to confuse generative-AI, which is a development based on the use of
Transformer Architectures in connection with Large Language Models, with either
AGI or ASI. Both AGI and ASI are conjectural technologies that do not yet exist – what does exist is gen-AI, but there is no evidence that the most
recent LLMs, such as ChatGPT5 and Gemini, are performing at a level that is as
good as or better than a human intellect. Although proposals for a “Final Exam”,
involving difficult-to-solve problems in most academic disciplines, intended to
set a definitive standard for machine performance relative to the human
intellect, do exist, there is no consensus. Nonetheless, with current models
still struggling with complex reasoning and nowhere near autonomous
intellectual performance, the idea that gen-AI (a smart chatbot that can solve
defined problems under human supervision) is on the threshold of true AGI is
absurd. Current forms of gen-AI are closer to Artificial Narrow Intelligence (ANI),
a machine intelligence that can perform narrowly defined intellectual tasks as
well as or better than human beings, such as playing Go or solving protein
folding problems. Although both AGI and ASI are speculative
technologies, this is unlikely to lead to a new modesty in technology
corporation claims. Instead, OpenAI’s declared AGI benchmark –an artificial intelligence capable of
postdoctoral performance across all academic disciplines – indicates the likely direction of AI development, namely, to
restrict the nature of the definition so that something like non-agential AGI
can be achieved using LLMs. While a scientific consensus on the likelihood of
AGI (so defined) does not exist, we regard predictions that this can be
attained through the extension and deepening of current methods sometime
between 2027 and 2032 as highly credible. The speculative relation between this
and ASI is that such a machine could be instructed to recursively self-program for
accelerated improvements in cognitive performance – this is basically
a chain-reaction conjecture – quickly outstripping human supervision.
According to Nick Bostrom, if this happened explosively, such a machine
intellect might pose an existential risk to the human species (and planetary
life) (Bostrom 2016, 115-119). Nonetheless, Bostrom seems convinced that with
the right kind of moral alignment, ASI would be vastly beneficial, making
possible a society of abundance that would constitute a “deep utopia” within
which human beings (and digital selves) would mostly play, not work, while
questing for the meaning of life (Bostrom 2024, 60-61). Likewise, at one
extreme, Altman believes that ASI will likely introduce a society of abundance
but might instead tile the world with datacentres and eliminate humanity with a
novel bioweapon (Andersen 2023). At the other extreme, Musk seems to think that
it will treat human beings as domestic pets, unless it can be aligned with
human supremacism, via an embrace of natural inequality. The basic
rationale for every single one of the American Big Tech firms researching AI is
that it is essential that their benevolent version of high-tech society happens
via a machine aligned to negative liberty and possessive individualism.
Meanwhile, some projections of the impact of ANI – let alone AGI – involve short-term 30%-40% white-collar unemployment in the decade
of its workplace implementation, with the associated annihilation of the living
conditions of technical specialists, liberal professionals, and college-educated
para-professionals (Felten 2023). This is, after all, explicitly what AGI is
designed for: OpenAI, for instance, states in its Charter that its corporate
aim is to produce “artificial general intelligence”, defined as “highly autonomous systems that
outperform humans at most economically valuable work” (OpenAI 2026). Just to be clear
here, as proponents of a weak-AI understanding of the technology, we are
extremely sceptical about claims that gen-AI, or even AGI, will permanently replace
human mental labour. What we are pointing to is the radical cheapening of
intellectual work, a process that Marx described as the production of “relative
surplus value” through technological improvements to labour productivity.
Christian Fuchs has made the cognate point in relation to the automation of
manual labour, that instead of reducing the amount of human work performed in
the world economy, what automation in the period 1991-2022 has done is to increase
the number of employees from 2.3 billion to 3.4 billion (Fuchs 2022, 141). At
the same time, in combination with neoliberal privatisation and deregulation,
this has resulted in a spectacular transfer of wealth from the producers of
value to the owners of capital, together with a rise in overall unemployment,
especially of unskilled labourers. Artificial intelligence can be expected to
do something similar. Economic research dealing with the implementation of gen-AI
models its likely destruction of occupation-types, rather than its creation of
permanent unemployment, but this has significant potential for social dislocation.
Superintelligence-belief acts prophylactically here, to mitigate the pain of
the transition that AGI’s arrival will cause. AGI becomes a necessary evil, the
sacrifice that the 1010 individuals must make to produce 1058
units of value. Given the expected arrival of the digital
Jesus, the Promised Land overfloweth – mainly with transhumanism and
extropianism. ASI is expected to deliver a tremendous boost to human
augmentation, cybernetic integration, and disease eradication. It is also
expected to provide the keys to effective immortality, perhaps including
digital replication of personality structures, and to the conscious selection
of the natural characteristics of one’s offspring, i.e., to liberal (and
not-so-liberal) eugenics. Such conjectures justify, in the here-and-now,
experimental technologies to do with lifespan prolongation, cryogenic
suspension and eugenics programs.[4]
They also justify a set of experimental lifestyle choices that seem,
superficially, to corroborate the idea that this is about libertarianism – notably polyamory, but also, reportedly, drug use. The now-bankrupt
cryptocurrency exchange FTX was helmed by a polycule, while Elon Musk is the
father of (at least) fourteen children with different mothers, some conceived
with IVF. Musk’s family has been described as a “harem” by the Wall Street
Journal, with some of the mothers solicited by Musk on his social media
platform X. In one text shown to the WSJ by conservative influencer Ashley St.
Clair, who shares a child with Musk, he reportedly said, “[t]o reach
legion-level before the apocalypse, we will need to use surrogates.” After the
Wall Street Journal story appeared, Musk simply wrote “TMZ > WSJ” on his
website, a comment which neither confirms nor denies the details of the story,
but rather simply his preference for tabloid news. At the same time,
performance optimising drug use has been a feature of Silicon Valley since the
1970s, when early programmers and entrepreneurs experimented with LSD (Harris 2023), to the proliferation of ADHD
amphetamine medications, both prescribed and not, as well as other more outré rumours
of Musk’s heavy ketamine usage. What we are looking at, then, is a set of
explicitly prefigurative cultural practices, which anticipate the society of
abundance that ASI is to deliver – after the birth
pains of AGI, social convulsion, economic dislocation and political
authoritarianism. Against
this background, it is hardly surprising that this worldview needs something
rather less angular to provide it with a human face. Enter the somewhat
unlikely science-fiction author voted most popular by a committee of
authoritarian billionaires. The Culture Series by reformist socialist author
Iain M. Banks is literary fiction, intended as a critique of social
hierarchies, aversive prejudices and authoritarian worldviews. However, it is also
intended as an ironic critique of a utopian society whose culture strongly
resembles neoliberal capitalism, which reveals its cultural imperialism and
totalitarian potential vividly as soon as it encounters what it understands as
an existential threat. What that means is that these novels are ambivalent
structures which hold utopian and dystopian moments in suspension, reserving
final judgement (or rather, transposing that problem onto the reader) by means
of subtle irony. This is the point of insisting that this is literary science
fiction – we don’t mean that it is approved for
university syllabi; we mean that it should not be read as pulp sci-fi. In this
section, we argue that the reception of Banks among tech billionaires and
corporate executives is over-determined by the Californian ideology and the
implications of AI implementation in the workplace. In other words, although
the problem is a simplistic misreading of a complex author, this misreading is socially
conditioned by political economy and ideological factors, because it functions
as an “imaginary resolution of real contradictions”. Here we want to offer the
techno-libertarian appropriation of Banks’ work as integral to understanding
contemporary capitalist ideology, in the manner of Giorgio Agamben’s
(Foucault-inspired) genealogical method, an “actual historical phenomena [...]
whose role [is] to constitute and make intelligible a broader
historical-problematic context” (Agamben 2009, 9). We might think of this
appropriation as metonymic of a broader technocratic and authoritarian
discourse which has been formed in and through a particular reading of
contemporary science fiction by its key participants, such as Musk, Thiel,
Yarvin, Land and others, perhaps even the more classically liberal Sam Altman.
Banks is here interpreted as articulating a cultural logic that underwrites the
drive towards automating contemporary culture whatever the social and
environmental cost. This is not to say that this is the only reading of Banks,
or indeed the most common one, but rather to suggest that central to
techno-libertarianism is a hermeneutics of what Steven Shaviro has termed the
“mimesis of futurity” at work in science fiction, “understood as a kind of
pressure, or incipience, that is already implicit within the present moment”,
and that, crucially, demands actualisation through a technics designed to bring
that future into being (Shaviro 2024, 1). Banks’ Culture is
what happens when ASI generates a society of abundance, within which human
activities become expressive preferences rather than functionally necessary, so
that “work” is gamified or adventurous. Indeed, every activity becomes ludic,
so that even the war machines of the Culture have names like the Torturer-class
warship Killing Time (Banks 1996, 276) and the Abominator-class dreadnaught
Falling Outside the Normal Moral Constraints (Banks 2010, 240). The consequence
for cultural formations is that they lose their functional relevance to social
roles, particularly in respect of the division of labour, which means that
cultural forms become elective stylisations of contingent ways of life. Social
existence within the Culture looks a lot like the lifestyles of the rich and
famous today in cosmopolitan multicultural capitalism – only without the
exploitation of the proletariat – plus techno
gadgets such as AI drones, complete sexual fluidity, and drug glands for
entertainment purposes. According to Banks, the Culture is “a society where
material scarcity is unknown and the only real value is sentimental value”, one
without laws or crimes, governed and organised by its superintelligent
shipminds, who as digital citizens are first amongst equals (Banks 1994, 12). “I
am a Culture Mind”, states the shipmind Lasting Damage in Look to
Windward: “We are close to gods, and on the far side” (Banks 2000, 316). As
the author clarifies in his much-cited essay “A Few Notes on the Culture”, “humans and independent drones (the Culture’s non-android individual AIs of roughly
human-equivalent intelligence) [...] have a status somewhere between
passengers, pets and parasites” (Banks 1994, 8). Unsurprisingly, a central
premise of each of the novels is that the lives of the human beings existing in
what Bastani describes as Banks’ “Fully Automated
Luxury Space Communism” are often beset by ennui (Bastani 2019). Ziller, the
non-Culture protagonist of Look to Windward, sardonically observes that: The point is
[...] that having carefully constructed their paradise from first principles to
remove all credible motives for conflict among themselves and [...] almost all
natural threats, these people then find that their lives are so hollow they
have to recreate false versions of just the sort of terrors untold generations
of their ancestors spent their existences trying to conquer (Banks 2000, 114). This is paradise,
as administered by artificial intelligence, within which desperation and
liberation are sometimes difficult to distinguish, something that seems to
entirely escape some of the readership. In a poignant moment in Consider
Phlebas, the first novel in the series, the Culture agent confronts a
defector who has passed from the Culture to an openly authoritarian theocracy. “Why?”,
asks the agent. “You’re ruled by your machines”, Horza replies. “You’re an
evolutionary dead end. [...] I don’t care how self-righteous the Culture feels
or how many people the Idrians kill. They’re on the side of life – [...]
fallible and short-sighted, but real life” (Banks 1987, 26). This answer (the
spiritual vacuity of the Culture and its domination by post-carbon digital
entities) points to the potential lack of existential meaningfulness of the
Culture’s ludic aestheticisation of human life (Banks 1987, 26-27). Written in 1987, the novel eerily anticipates the “clash of
civilisations” that would emerge between Islamic fundamentalism and triumphant
neoliberalism after the Fall of the Berlin Wall. In such
dialectical set-pieces, the complexity of the series’s presentation of utopia
and counter-utopia is manifest, despite its author’s clear preference for the
society of abundance and its benevolent machines. As it is typical in literary
presentations of ambivalence, the resolution depends on dramatic irony – the
actions of figures such as Horza, Ziller, and Zakalwe in Use of Weapons are
ultimately self-defeating because their search for absolute self-deconstructs
under the pressure of the ruthless deeds their beliefs legitimate. After the
ambivalence has been registered, it is narratively neutralised, leaving the everyday
life of the Culture as the only benign alternative to forms of religious
fundamentalism that are energised by strong evaluations, moral absolutes and atavistic
convictions. Completely central
to this resignation to paternalistic guidance is the firm conviction that a
strong version of superintelligent AI is a benign technology that supersedes
what Horza calls human “fallibility and short-sightedness”. The culture of the
Culture is tolerant, permissive, and pluralistic. Its society is the opposite
of the neoliberal dystopias of cyberpunk fiction – in the Culture,
there is no poverty and criminality is managed by benevolent surveillance.
There is a great deal to like about Banks’ Culture, expressly designed as a
utopian “correction” to the real world of 1970s nostalgic reaction, national
chauvinism and assertions of a fixed human nature, rigid gender roles,
finalistic sexual assignment, racial differences, and so forth (Kincaid 2017, 27-29). However, these
egalitarian elements are balanced against libertarian tendencies, which can
assert AI sovereignty precisely because this transpires within an imaginary
universe, within which the social bond is an elective affinity. What culture is
not, from this perspective, is the narrative presentation of moral norms,
or a conjectural future history of human emancipation, since all that boring
stuff is delegated to robots, leaving the citizens of the Culture free to
engage in hedonistic self-expression. Banks imagines a future in which humanity
(and other sophont species) will be guided by ASI to a libertarian paradise
that is tolerant, pluralistic, egalitarian – and
individualistic. It justifies a great deal, in the here and now, which might
otherwise seem – well, morally cloudy, to put it mildly. But
surely after certain necessary evils, the maximisation of value will be
definitively established – right? Apparently not. The problem is that
the Culture is surrounded by emergent civilisations, many of which have
regrettable tendencies to strong evaluations, moral absolutes, fixed meanings
and belief in the superiority of their sophont species. The purpose of Contact
is to prepare encountered civilisations for integration into the Culture by a
simple process of negative alignment – “carbon fascists”
are systematically prevented from attaining the technological mastery necessary
for membership of the interstellar community. Now, a carbon fascist is any
lifeform that holds not just the belief that its species is superior, but more
generally, any lifeform that holds that life is superior to its digital
simulation (hence the “carbon” in the fascism). In short, sophonts who resist
the ASIs are treated rather like underdeveloped countries in relation to
neoliberal imperialism – for instance, Afghanistan, to name a
pointed example. In The Player of Games, the second novel in the series,
the protagonist is manipulated by the ASIs into participating in a ritualistic
competition within a sophont society whose empire depends on a social
hierarchy, whose allocation is fixed periodically by the placement of
individuals through sort of super-chess (or maybe, space-Buzkashi). Of course,
it’s all a trick, leading up to the murder of the Emperor of Azad by his own
Grand Marshall, out-played by the ASIs into taking a shot at the Culture’s
agent, which the stealth super-drone, Flere-Imsaho, reflects back at them, with
lethal effect (Banks 1988, 374). Upon the death of the
emperor, effectively a murder while cheating at the sacred game, the empire
dissolves into chaos and the game itself is utterly discredited. The point here
is that the technocratic utopia ushered in by ASI is seen as excluding whatever
value is regarded by its makers as antithetical to the Culture. There can, in
the final analysis, be one and only one Culture. Its substantive definition
depends on its inventors. What is at stake in Banks’ work becomes
clear in the figure of the Culture’s Special Circumstances, its secret
assassins and saboteurs. When Contact encounters an alien civilisation that is
inconsistent with the Culture –when
it finds a form of carbon fascism – it calls on the
services of Special Circumstances to suppress religious fundamentalisms,
militaristic societies, and species supremacists, not generally by direct
violence, but by subtle entrapment and cultural sabotage. The aim is to catalyse
technological regression, preventing them from attaining membership of the
interstellar community. Here we want to offer the conceptual figure of Special
Circumstances as one of the keys to the way that Banks’ novels are read in the
broader tech milieu. Istvan Csicsery-Ronay has termed neologism one of the
seven “beauties” of science fiction (Csicsery-Ronay 2012), and like the Culture
itself, “Special Circumstances” as a term offers us an important insight into
the blandly automised violence of techno libertarianism, possibly on its way to
a form of late fascism. Banks describes it in the following way: Special Circumstances had always been
the Contact section’s moral espionage weapon, the very cutting edge of the
Culture’s interfering domestic policy, the elite of the elite, in a society
which abhorred elitism. […] It had about it too an atmosphere of
secrecy (in a society that virtually worshipped openness) which hinted at
unpleasant, shaming deeds, and an ambience of moral relativity (in a society
which clung to its absolutes; life/good, death/bad; pleasure/good, pain/bad)
which attracted and repulsed at once, but anyway excited. No other part of the Culture more
exactly represented what the society as a whole really stood for, or was more
militant in the application of the Culture’s fundamental beliefs. Yet no other
part embodied less of the society’s day-to-day character (Banks 1987, 28). There are several
salient points to notice in this explanation: first, Special Circumstances is
an elitist organisation that operates in some measure in secret; second, it
functions as a counter-intuitive justification for behaviour that appears to
contradict the values held by the Culture; and third, the fostering of the life
of the Culture, including those of artificial intelligences, legitimates any
form of violence against its embodied others. Special Circumstances is an
organisation, a societal structure, but it is also a moral justification – in short, it is a sovereign decision arising in confrontation with
an existential threat, that is, a Schmittian state of exception. In other words, Special Circumstances is
something like an aporia in the sense described by Jacques Derrida, a
moment of impasse in which those elements that make something possible also
make it at the same time impossible on its own terms. The paradox of Special
Circumstances is of a violence done in the name of non-violence, in which the
utopian image of a possible future teeming with intelligences, human and
artificial, mutates into a dystopian violence against materially existing life
in the present, including non-human life in the form of environmental
destruction. This goes significantly beyond Popper’s paradox of tolerance into
an implicitly eliminationist politics whose apparent rationality and appeals to
human flourishing elide their investment in the violent erasure of (racialised)
diversity. The “crisis” that justifies this corporate Schmittian state of
exception is one that Yarvin, at the very least, and likely a decent proportion
of his followers, locates in democracy itself. While Alberto Toscano has argued
that what he calls “late fascism” is shorn of its utopian dimensions (Toscano 2023, 15), what we see via Banks is
the way that techno-utopianism might circulate within tech-financial circles
without necessarily spreading very far beyond them, as a discourse of the elite
speaking to itself. If not precisely as secret as the Special Circumstances
organisation, this is nevertheless a hermetic discourse that presumes the
initiation of the chosen few, as when neo-reactionary thinker Curtis Yarvin
describes a friend of Peter Thiel as “enlightened”, who might benefit in this
coming future being created by automation. The rest, unfortunately, in the
non-ironic reading of Banks, are as disposable as Yarvin’s “organ meat”,
“carbon fascists” overly invested in their own material existence. Where the Culture is tolerant, permissive,
and pluralistic, Special Circumstances is intolerant, prohibitive and
monological – towards fundamentalists, militarists and
authoritarians. The paradox is obvious. So too is the way the series
anticipated the moral contortions of Western imperialism in the “clash of
civilisations” after 2001. In Excession, the Culture simultaneously
encounters an artefact from a superior civilisation and runs into political
problems with managing the militaristic colonialist species, the Affront. The
ASIs split into two camps – one camp wants to encounter the superior
civilisation in a spirit of curiosity and openness. The other camp wants to use
the encounter to entrap and defeat the Affront, without having to actually
fight them. The Affront are “useful idiots”, providing the ASIs with
justifications for presenting a variety of other deep plans to the interstellar
community. In the end, the Affront are entrapped. However, the Excession, after
protecting itself in the context of violence between the Culture and the
Affront, withdraws enigmatically into N-space. An epilogue informs the reader
that the ASIs failed the test of their readiness for entry into a higher civilisation;
the Culture is to the Excession as the “carbon fascists” are to the Culture (Banks 1996, 445-447, 455). The difference,
which only a reading of the series through literary irony reveals, is that the
Excession does not do to the Culture what the Culture does to the
Affront. Read non-ironically, however, the series nests super-cultures in a
Russian Doll structure, licensing the idea that manipulation is legitimate if
the ends are “good”. From that perspective, Musk’s self-identification of
his personal company with the Excession clarifies what he takes from Banks.
Musk is the superior intellect, enigmatically beyond current moralities, who
manipulates both ASIs and their forthcoming Culture, and the authoritarian right-wing
“Affront”, the current useful idiot in the White House, in the name of a goal
that only he can see. Invocations of
socialist science fiction notwithstanding, a worldview that imagines that the
intellectual elite has a special destiny to shape history by cunning
manipulation of mass society is, ultimately, Nietzschean. There is a common
thread of the Übermensch that runs through Nietzsche, Banks, Land, Musk,
Thiel, and Yarvin – though what distinguishes the Californian incarnation is
the transhumanism of TESCREAL and the idealisation of code, automation and
digital existence (by contrast, Nietzsche was too in love with embodied
pleasure). Yarvin makes explicit the way that the new aristocracy of the tech
elite, elected by their superior intellect, see themselves as not merely
superior but more deserving of continued life in the future than those whom they
rule over. That future, fuelled by the dreams of the science fiction of Banks
and others, is imagined as a place outside of the Earth
– to
Mars and beyond. At its most dangerous, transhumanism turns into eliminationist
eugenics, regrettably presiding over the vast majority of human labour now
deemed “obsolete”, mere organ meat. As a result, the enthusiastic descriptions
of a society of abundance, personal practices of sexual and chemical
liberationism, and some vague references to creative self-expression cannot
conceal the fact that the “value” that is to be maximised is machine
intelligence in the service of the mastery of the human. These two strands
coalesce into a “post-humanist” eugenics program that uses AGI as a blunt
instrument to force social changes consistent with the hoped-for breakthroughs
into a posthuman condition facilitated by ASI. It is in this respect that the Culture
series of Iain M. Banks functions as an imaginary resolution of real
contradictions, a culturally prestigious utopian reference that justifies
social dislocation and domination of the human in the name of a benevolent future
run by a benign technology. In this concluding section, we want to connect this
set of critical reflections on the Californian ideology with the critique of
digital capitalism articulated by Christian Fuchs (Fuchs 2021; Fuchs 2022). In his critique of digital capitalism,
Fuchs points out that “digital labour does not only denote the production of digital content” but is rather “a
category that encompasses the whole mode of digital production, a network of
agricultural, industrial and informational labour that enables the existence
and use of digital media” (Fuchs 2021, 263). Conversely, processes of
agricultural, industrial and informational production are thoroughly imbricated
with the digital affordances of the Information and Communications Technology
revolution that has happened in the last 30 years (Fuchs 2021, 312). The digitalisation of capitalist production, distribution and
consumption that characterises global capitalism today implies that the
implementation of AI technology in the workplace means more than a convulsion
in the occupational categories of white-collar employment. It means a
revolutionisation of production, intended to restore profitability to a
capitalist system whose declining margins have been amply documented by Fuchs
(2018) and others. We have suggested that AGI and related technologies involve the
production of relative surplus value, generated through the substitution of machine
learning algorithms for specific elements of human mental labour in particular processes
of production. Strong AI conceptions of the true equivalence of machine
learning and the human intellect, best described as digital anti-humanism, are
essential to selling this vision, as is the roseate promise that these generally
intelligent and then superintelligent machines are going to be more benign than
the humans they “replace”. As Fuchs points out, the alternative to this
techno-Nietzscheanism is not rejection of the technology, but its conceptual
and social re-positioning within emancipatory and
democratic practices and understandings. “Radical
Digital Humanism", Fuchs writes, “rejects the idea to replace humans by,
or transform them into, digital machines. Rather, it sees digital machines as part
of the struggle for a society that benefits all humans that can expand, help
realise and more fully develop the potentials of humans and society” (Fuchs
2022, 152). The problem, then,
with the pro-socialist science fiction of Iain M. Banks, and the reason that
his work can be ideologically repurposed by libertarian billionaires, is that
it is centred on an anti-humanist vision of strong AI, characterised by
post-carbon digital intelligence as equivalent to, or better than, human life.
We might say, then, that the relation of the Culture Series of Iain
Banks in the thinking of the Big Tech leadership and a program of political
manipulation, reckless AI research, eugenics experimentation and corporate
consolidation is the same as the relation between dreams of the Millennium and
the reality of the Inquisition. In this context, no doubt the science fiction
of Iain Banks is only a temporary resting point in the elaboration of the
utopian pole of the ideological dipole of big tech’s imaginary solution to real
contradictions. For the rest of us, the political struggle of our time is to be
something more than organ meat in the post-AI, post-AGI future. Agamben, Giorgio. 2009. The Signature
of All Things. New York: Zone. Amodei, Dario. 2024. Machines of Loving
Grace: How AI Could Transform the World for the Better. Accessed 6 January
2026, https://www.darioamodei.com/essay/machines-of-loving-grace. Anders, Charlie Jane and Annalee Newitz,
hosts. 2023. Episode: 125: Silicon Valley vs. Science Fiction: ChatGPT. Episode
125, All Our Opinions are Correct, March 9. C. J. Anders and A. Newitz. Podcast,
55 min., 16 sec. https://www.ouropinionsarecorrect.com/shownotes/2023/3/8/episode-125-silicon-valley-vs-science-fiction-part-i Andersen, Ross. 2023. 24 Jul 2023. Does
Sam Altman Know What He Is Creating? The Atlantic, July 24. https://www.theatlantic.com/magazine/archive/2023/09/sam-altman-openai-chatgpt-gpt-4/674764/. Banks, Iain M. 1987. Consider Phlebas.
London: Orbit. Banks, Iain M. 1988. The Player of
Games. London: Orbit Books. Banks, Iain M. 1990. Use of Weapons.
London: Orbit. Banks, Iain M. 1996. Excession.
London: Orbit Books. Banks, Iain M. 2000. Look to Windward.
London: Orbit. Banks, Iain M. 2004. A Few Notes on the
Culture. In The State of the Art, edited by Ian M. Banks, 167-188. San
Francisco: Night Shade. Banks, Iain M. 2010. Surface Detail.
London: Orbit Books. Barbrook, Richard and Andy Cameron. 1996.
The Californian Ideology. Science as Culture 6 (1): 44-72. https://doi.org/10.1080/09505439609526455. Bastani, Aaron. 2019. The World Is a Mess.
We Need Fully Automated Luxury Communism. The New York Times, June 11. https://www.nytimes.com/2019/06/11/opinion/fully-automated-luxury-communism.html. Bastani, Aaron. 2020. Fully Automated
Luxury Communism: A Manifesto. London; New York: Verso. Bostrom, Nick. 2016. Superintelligence:
Paths, Dangers, Strategies. Reprint Edition. Oxford: Oxford University
Press. Bostrom, Nick. 2024. Deep Utopia: Life
and Meaning in a Solved World. Washington, DC: Idea Press. Caroti, Simone. 2015. The Culture
Series of Iain M. Banks: A Critical Introduction. Jefferson, NC: McFarland
Publishers. Caroti, Simone, David Barr Kirtely, hosts.
2016. Think Culture Is a Space Opera? Nah, It’s a Trojan Horse. Episode 209.
Geek's Guide to the Galaxy, June 24. Lightspeed Magazine. Podcast, 61 min., 41
sec. https://youtube.com/watch?v=2AjmhyxHM0. Carroll, Jordan S. 2022. To Understand
Elon Musk, You Have to Understand This ’60s Sci-Fi Novel. Jacobin, May
29. https://jacobin.com/2022/05/musk-tesla-robert-heinlein-libertarianism-technocracy. Carroll, Jordan S. 2024. Speculative
Whiteness: Science Fiction and the Alt-Right. Minneapolis; London:
University of Minnesota Press. Carroll, Tobias. 2025. Elon Musk Just
Doesn’t Understand the Sci-Fi Visions of Iain M. Banks. Literary Hub,
April 30. https://lithub.com/elon-musk-just-doesnt-understand-the-sci-fi-visions-of-iain-m-banks/. Cross, Timothy. 2017. The Novelist Who
Inspired Elon Musk. 1843 Magazine–The Economist, March 31. https://www.economist.com/1843/2017/03/31/the-novelist-who-inspired-elon-musk. Csicsery-Ronay, Istvan. 2012. The Seven
Beauties of Science Fiction. Middletown: Wesleyan University Press. Feloni, Richard. 2015. Mark Zuckerberg
Wants Everyone to Read This Sci-fi Novel. Business Insider, July 1. https://www.businessinsider.com/mark-zuckerberg-recommends-the-player-of-games-2015-6. Felten, Edward W., Manav Raj and Robert Seamans. 2023. How will Language Modelers like ChatGPT Affect Occupations and
Industries? arXiv. https://doi.org/10.48550/arXiv.2303.01157. Fisher, Mark. 2020. Post-Capitalist
Desire: The Final Lectures. London: Repeater Books. Fisher, Mark. 2022. Capitalist Realism:
Is There No Alternative? 2nd edition. London: Zero Books. Flood, Alison. 2020. Amazon TV Adaptation
of Iain Banks' Culture Series is Cancelled. The Guardian (Australian
Edition), August 26. https://www.theguardian.com/books/2020/aug/26/amazon-tv-adaptation-of-iain-banks-culture-series-is-cancelled. Freedman, Sam. 2025. The Big Idea: Will
Sci-Fi End Up Destroying the World? The Guardian (Australian Edition),
April 14. https://www.theguardian.com/books/2025/apr/14/the-big-idea-will-sci-fi-end-up-destroying-the-world. Fuchs, Christian. 2018. Digital Demagogue:
Authoritarian Capitalism in the Age of Trump and Twitter. London: Pluto
Press. Fuchs, Christian. 2021. Digital
Capitalism: Media, Communication and Society. Volume Three. London:
Routledge. Fuchs, Christian. 2022. Digital
Humanism: A Philosophy for 21st Century Digital Society. Kindle
Edition. Bingley, UK: Emerald Publishing. Gebru, Timnit and Émile P. Torres. 2024.
The TESCREAL Bundle: Eugenics and the Promise of Utopia through Artificial
General Intelligence. First Monday 29 (4): 1-42. https://doi.org/10.5210/fm.v29i4.13636. Grady, Constance. 2025. Why does Elon Musk
Love this Socialist Sci-Fi Series? Vox, May 22. https://www.vox.com/culture/413502/iain-banks-culture-series-elon-musk-jeff-bezos-mark-zuckerberg. Harris, Malcom. 2023. Palo Alto: A
History of California, Capitalism and the World. New York: Little Brown and
Company. Hassabis, Demis (@demishassabis). 2018. ”Banks'
Culture series is brilliant”. X (Formerly Twitter), April 1. https://x.com/demishassabis/status/980247849668661250. Hassabis, Demis, host. 2020. Episode 8:
Demis Hassabis–The Interview. Episode 8. Goggle DeepMind: The Podcast.
Google Deepmind, September 17. Podcast, 36 min., 58 sec.
https://deepmind.google/blog/the-podcast-episode-8-demis-hassabis-the-interview/. Hassabis, Demis, Ezra Klein, hosts. 2023. A.I.
Could Solve Some of Humanity’s Hardest Problems. It Already Has. Demis Hassabis
Discusses How AI Systems Can Accelerate Scientific Research. The Ezra Klein
Show. New York, Times Media, July 11. Podcast, 89 min., 01 sec. https://www.nytimes.com/2023/07/11/opinion/ezra-klein-podcast-demis-hassabis.html.. Hermansson, Patrik, David Lawrence,
Joe Mulhall and Simon Murdoch. 2020. The Dark Enlightenment: Neoreaction and
Silicon Valley. In The International Alt-Right: Fascism for the 21st
Century?, edited by Patrick Hermansson, David Lawrence, Joe Mulhall and
Simon Murdoch, 125-158.. Abingdon, Oxon; New York, NY: Routledge. Jameson, Fredric. 1979. Fables of
Aggression: Wyndham Lewis, the Modernist as Fascist. Berkeley: University
of California Press. Jameson, Fredric. 1981. The Political
Unconscious: Narrative as a Socially Symbolic Act. Ithaca: Cornell
University Press. Jetha, Rya. 2025. Tech Billionaires are
Obsessed with this Dystopian Sci-fi Novel. The San Franscisco Standard, June
27. https://sfstandard.com/2025/06/27/legion-bob-sam-altman-marc-benioff-reid-hoffman-dystopia/. Kakutani, Michiko. 2025. Why Silicon
Valley’s Most Powerful People Are So Obsessed with Hobbits. The New York
Times, May 23. https://www.nytimes.com/2025/05/23/books/tolkien-musk-thiel-silicon-valley.html. Katz, Yarden. 2020. Artificial
Whiteness: Politics and Ideology in Artificial Intelligence. New York:
Columbia University Press. Kelly, Stuart. 2018. Does Elon Musk Really
Understand Iain M Banks's 'Utopian Anarchist' Culture? The Guardian
(Australian Edition), June 19. https://www.theguardian.com/books/booksblog/2018/jun/19/elon-musk-iain-banks-culture-novels. Kincaid, Paul. 2017. Iain M. Banks.
Chicago: University of Illinois Press. Kokotajlo, Daniel, Scott Alexander,
Thomas Larsen, Eli Lifland and Romeo Dean. 2025. AI 2027. AI 2027.
April 3. https://ai-2027.com/. Land, Nick. 1990. The Thirst for
Annihilation: Georges Bataille and Virulent Nihilism. London; New York:
Routledge. Land, Nick. 1993. Machinic Desire. Textual
Practice 7 (3): 471-482. https://doi.org/10.1080/09502369308582177. Land, Nick. 2011. Fanged Noumena:
Collected Writings. Cambridge, MA: MIT Press. Land, Nick. 2017. The Dark Enlightenment. Online:
Self-published. Lepore, Jill. 2021. Elon Musk Is Building
a Sci-Fi World, and the Rest of Us Are Trapped in It. The New York Times,
November 4. https://www.nytimes.com/2021/11/04/opinion/elon-musk-capitalism.html. Martikainen, Joonas. 2023. A Brave New
World in the Making: Fully Automated Luxury Communism as Political Dystopia. In
Analysing Darkness and Light: Dystopias and Beyond, edited by Joonas
Martikainen, 66-87. Leiden: Brill. Mendlesohn, Farah. 2005. The Dialectic of
Decadence and Utopia in Iain M. Banks’s Culture Novels. Foundation 34.1
(93): 116-124. https://cetapsrepository.letras.up.pt/id/cetaps/112329. Midlane, Tom. 2025. The Curious Sci-Fi
Beliefs of the AI Tech Elite. Tribune, July 16. https://tribunemag.co.uk/2025/07/the-curious-sci-fi-beliefs-of-the-ai-tech-elite. MIT. 2026. Urbanomic. Accessed 7 January
2026. https://mitpress.mit.edu/publisher/urbanomic/. Moldbug, Mencius. 2007. Formalist
Manifesto. Unqualified Reservations. Accessed 29 October 2025. https://www.unqualified-reservations.org/pdfs/formalist_manifesto.pdf. Moldbug, Mencius. 2013. Sam Altman is not
a Blithering Idiot. Unqualified Reservations. Accessed 29 October 2025. https://www.unqualified-reservations.org/pdfs/sam_altman_is_not_a_blithering_idiot.pdf. Musk, Elon (@elonmusk). 2018. “If you must
know, I am a utopian anarchist of the kind best described by Iain Banks”. X
(Formerly Twitter), June 17. https://x.com/elonmusk/status/1008120904759402501. Musk, Elon (@elonmusk). 2019. “Possibly Excession,
but I’d recommend reading Player of Games & Surface Detail first.” X
(Formerly Twitter), December 14. https://x.com/elonmusk/status/1205807718969139200. OpenAI. 2026. OpenAI Charter. Accessed 6
January 2026. https://openai.com/charter/. Robinson, Nathan J. and Garrison Lovely.
2025. What Harms Will AI Cause, and What Can We Do About Them? Current
Affairs: A Magazine of Politics and Culture, January 22. https://www.currentaffairs.org/news/what-harms-will-ai-cause-and-what-can-we-do-about-them. Saito, Kohei. 2022. Marx in the
Anthropocene: Towards the Idea of Degrowth Communism. Cambridge: Cambridge
University Press. Schleifer, Theodor. 2025. Could Mars Be
Elon Musk’s Next Business Venture? The New York Times, April 25. https://www.nytimes.com/2025/04/25/us/politics/elon-musk-mars.html. Schradle, Jen. 2015. Silicon Valley
Ideology and Class Inequality. In Handbook of Digital Politics, edited
by Stephen Coleman and Deen Freelon, 67-84. New York: Edward Elgar. Shaviro, Steven. 2024. Fluid Futures:
Science Fiction and Potentiality. London: Repeater. Srnicek, Nick. 2025. Silicon Empires:
The Fight for the Future of AI. London: Polity. Srnicek, Nick and Alex Williams. 2013.
#Accelerate Manifesto for an Accelerationist Politics. Critical Legal
Thinking 14 (May): 1-37. Srnicek, Nick and Alex Williams. 2015. Inventing
the Future: Post-capitalism and a World without Work. London: Verso. Stross, Charles. 2023. Tech Billionaires
Need to Stop Trying to Make the Science Fiction They Grew Up on Real. Scientific
American Online, December 20. https://www.scientificamerican.com/article/tech-billionaires-need-to-stop-trying-to-make-the-science-fiction-they-grew-up-on-real/ Toscano, Alberto. 2023. Late Fascism:
Race, Capital and the Politics of Crisis. London; New York: Verso. Ward, Ian. 2025. Curtis Yarvin’s Ideas
Were Fringe. Now They’re Coursing Through Trump’s Washington. Politico,
January 30. https://www.politico.com/news/magazine/2025/01/30/curtis-yarvins-ideas-00201552. Wilson, Jason. 2024. He’s Anti-democracy
and Pro-Trump: The Obscure 'Dark Enlightenment' Blogger Influencing the Next US
Administration. The Guardian (Australian Edition), December 21. https://www.theguardian.com/us-news/2024/dec/21/curtis-yarvin-trump. Geoff M Boucher Geoff M. Boucher is an associate
professor in Literary Studies at Deakin University who specialises in Frankfurt
School Critical Theory and currently researches the authoritarian personality
in contemporary culture. He is the author of Critical Theory and the
Authoritarian Personality (EUP, 2025) and Habermas and Literature (Bloomsbury
Academic, 2021), as well as numerous articles on right-wing authoritarianism.
He wishes we didn’t live in interesting times. Emily McAvan Emily McAvan is a literary critic and
theorist of environment, technics and culture. She is the author of several
books, including Pollution Theory: Reading Toxic Entanglement in the
Anthropocene (Bloomsbury Academic, 2026), as well as numerous articles in
journals such as Critical Inquiry, Angelaki and New Literary
History. She is a teaching fellow at Deakin University. [1] This egalitarian/neoreactionary polarization is emblematized by the
differences between OpenAI’s ChatGPT5 and X-AIs Grok, respectively. [2] One oft-unremarked underlying thread that ties key figures like
Musk, Land, and Yarvin together is their eliminationist hostility to
transgender existence, which is, perhaps, for them, the wrong kind of
transhumanism – abject, embodied, socially marginalized, and, worst of all, poor. [3] Nick Bostrom – despite
his EA and Longterm worldview –
remains the best guide to this. [4] Just to be clear: with the exception of
eugenics, none of these technologies is intrinsically immoral – the point is that they are experimental
and so their benefits are conjectural.
3. Superintelligence
and Accelerationism
4. A World Run by Artificial Super Intelligence (ASI)
5. Banks and Bros
6. Conclusion
References
About the Authors