Over the past decade, two major phenomena have developed in
the digital realm. On the one hand, extremism has grown massively
on the Internet, with sprawling online ecosystems hosting a wide
range of radical subcultures and communities associated with
both ‘stochastic terrorism’ and the ‘mainstreaming of extremism’.
On the other hand, Artificial Intelligence (AI) has undergone
exponential improvement: from ChatGPT to video deepfakes,
from autonomous vehicles to face-recognition CCTV systems,
an array of AI technologies has abruptly entered our everyday
lives. This report examines ‘AI extremism’, the toxic encounter
of these two evolutions – each worrying in its own right. Like past
technological progress, AI will indeed be – in fact already is – used
in various ways to bolster extremist agendas. Identifying the many
opportunities for action that come with a range of AI models, and
linking them with different types of extremist actors, we offer a clear
overview of the numerous facets of AI extremism. Building on the
nascent academic and government literature on the issue as well
as on our own empirical and theoretical work, we provide new
typologies and concepts to help us organize our understanding of
AI extremism, systematically chart its instantiations, and highlight
thinking points for stakeholders in countering violent extremism.