Tuesday, February 10, 2026

Creating liberating content

Choose your language

hello@global-herald.net

China moving early as...

China is urging its biggest banks to curb their exposure to US...

First-of-its-kind bridge constructed to...

A plethora of bat species will not have...

Jutta Leerdam Olympics Gold...

Jutta Leerdam claimed her first Olympic gold...

The best credit cards...

If you're about to book an Alaska Airlines flight, you may be...
HomePoliticsEuropeShould you believe...

Should you believe the AI hype? Probably not


Is hype about Artificial Intelligence justified? Cristobal Garibay-Petersen, Marta Lorimer and Bayar Menzat argue public discourse around AI has gone beyond what its technical specifications warrant, limiting the space for democratic engagement with and control of the technology.


Since the release of ChatGPT in November 2022, speculation about how a new era of Artificial Intelligence (AI) would change society has been rife. From evangelising claims that AI would save the world, to catastrophist warnings that it could destroy it, to admonishments to remain sceptical, positions on the AI-debate truly run the gamut. In a recent paper, we look at these debates and ask: how is AI being presented in public discourse, and to what political effects?

Testing four claims about AI

Four sets of discourses, made primarily by influential tech commentators, interest us: claims that AI somewhat resembles human intelligence and affects “humanity”; ideas surrounding AI having “agency”; statements about the economic implications of AI; and confident declarations about the need to act “urgently” in response to AI developments.

These claims we find problematic for two reasons: first, they are inaccurate from a technological point of view; second, their wide-ranging implications for democratic governance are being obscured.

Take claims that AI systems are getting closer to matching human capabilities in certain areas and even surpassing them in others. From a technical standpoint, the idea that AI resembles human intelligence in a meaningful way is spurious. Contrary to popular belief, models like ChatGPT do not “learn” continuously.

Because they operate with fixed parameters (weights), interactions do not leave any trace on the underlying model. This is very different from biological learning processes, which involve continuous, dynamic interactions with the environment that produce durable changes in neural connectivity.

A false equivalence

From a critical analytical perspective, the idea that human and machine intelligence can and should be compared is equally problematic because it creates a false equivalence between human decision-making and computer decision-making. This false equivalence can become the basis for a series of dubious practices, such as the replacement of (fallible) human judgement with (allegedly superior, but no less fallible) machine judgement.

Claims that AI is an issue that affects “humanity” are similarly politically problematic because they conceal the way in which different groups are affected by AI. Some jobs are more at risk of others, and some groups (usually, those already marginalised) are more likely to be negatively affected by AI.

Failing to recognise the differential impacts of AI hampers the ability to mobilise these groups politically. Identifying a constituency is essential for the articulation of political problems and solutions. Suggesting that “humanity” is in this together negates the very existence of the kind of divisions one could mobilise around.

Problems of agency

One should also be wary of claims that AI technologies are about to acquire the ability to make autonomous decisions. These claims frequently present conflicting views of the role of humans in the development of AI.

On the one hand, AI is frequently presented as self-developing and at constant risk of escaping human control. On the other hand, the development of AI also becomes an opportunity to reassert human agency and dictate its future. Ascribing agential qualities to AI is questionable from a technological perspective because it overestimates its ability to produce truly novel content or generalise outside of training data.

From a political perspective, this complex mix of machine agency and human reaction is also troublesome. The attribution of agency to AI conceals the ways in which humans are implicated in its development.

At the same time, presenting human agency as purely reactive promotes a form of politics that leaves limited scope for political choice. Because it takes the political agenda as externally set, it removes it from the realm of democratic deliberation: the technology is developing in a certain direction and politics, policymakers and citizens need to accept it and respond to it in pre-determined terms.

AI and the economy

Then there are the economic claims that are made about AI. Discourses on AI have tended to portray the development of the technology as essentially linked to a specific conception of what economic reality is and ought to be. These assumptions are problematic to the extent that they uncritically reflect key pillars of a certain form of liberal capitalism and its depoliticising tendencies.

For example, the idea that competition-driven markets should act as the benchmark against which success, legality or rightfulness ought to be measured reduces the scope to consider what other criteria one might want to keep in mind when developing new technologies. Likewise, that a very specific conception of what constitutes “good” or “responsible” economic practice is set in stone is problematic to the extent that economic practice and theory, like everything else, are subject to change over time.

The myth of urgency

The final, and perhaps most important, set of claims that concerns us are the temporal assumptions that are made about AI. Public discourse about AI points towards a sense of acceleration, urgency and temporal linearity. AI is presented as a force in motion, which is already changing societies and will do so even more radically in the future. The claim of a sudden acceleration prepares the ground for demands to act urgently to speed it along or to halt it in its tracks.

Many of these claims are not fully warranted from a technical standpoint because they by and large overstate the speed at which AI is developing and create narratives of certainty where scientific knowledge is significantly less certain.

For example, many claims concerning AI’s ability to improve seem to be based on the notion that augmenting the size of AI models (“scaling”) might address various extant challenges in the field. Reality, however, presents a less optimistic picture, particularly given the huge computational resources needed for the kind of scaling that is envisaged.

Instead, the confident timelines and perceived urgency for action seem to be driven by political – ideological, even – attempts to move the present towards the future they predict. This future, for better or for worse, is one with AI.

AI and democracy

This occupation of the future, and the appeals to urgency that go with it, should worry us. The reduction of the future to one where AI will take over suggests that only one future is possible. From a democratic standpoint, this reduces the space to choose between political alternatives. The appeal to urgency further reduces the space to imagine, let alone bring about, alternative futures.

Framing something as a problem in need of an urgent response reduces the space for deliberative, “slow” decision processes and for democratic choice itself. AI discourses pointing to the need to urgently address the problems or opportunities that AI causes reduce the space to collectively define what the problem that needs addressing is, how it should be addressed and who is best placed to address it.

We should, therefore, remain critical of AI-hype. AI has become more than what its technical specification should warrant. It is now a mobilising political concept: an idea that is used to push political development in a certain, not necessarily understood and not necessarily chosen, direction. Unfortunately, this is being done in a way that limits the space for democratic engagement with, and control of, the technology itself.

For more information, see the authors’ recent paper in Big Data & Society.


Note: This article gives the views of the authors, not the position of LSE European Politics or the London School of Economics.

Image credit: naimurrahman21 provided by Shutterstock.


Subscribe to our newsletter




Source link

Get notified whenever we post something new!

spot_img

Create a website from scratch

Just drag and drop elements in a page to get started with Newspaper Theme.

Continue reading

China moving early as confidence in US debt frays

China is urging its biggest banks to curb their exposure to US Treasuries, a calculated intervention in the plumbing of global finance that reveals something uncomfortable about where Asia believes risk now sits.Chinese regulators have instructed large banks...

First-of-its-kind bridge constructed to safely house species of bats

A plethora of bat species will not have the opportunity to thrive and live safely with a unique new creation.PIGEON RIVER GORGE, N.C. – Construction crews in North Carolina are lending a helping hand...

Jutta Leerdam Olympics Gold Medal Win in 1,000m Sets Games Record at Milan Cortina 2026

Jutta Leerdam claimed her first Olympic gold medal on Monday, setting a new Winter Games record in the women's 1,000-meter speed skating event at the 2026 Winter Olympics. ...

Enjoy exclusive access to all of our content

Get an online subscription and you can unlock any article you come across.