[Previous message][Next message][Back to index]
[Commlist] Visual Cultures Spring Term Public Programme: 'Race and/as Technology,' February 27th-March 27th
Thu Feb 27 09:14:08 GMT 2025
*Visual Cultures Spring Public Programme: Race and/as Technology*
You are cordially invited to attend Goldsmiths’ Department of Visual
Cultures’ Spring Term Public Programme, a series of lectures organised
around the theme ‘Race and/as Technology.’ This series adopts a media
studies and visual cultures approach to the theme and will be of
interest to media scholars.
All lectures are free and open to all, with no need to register, and
take place *Thursdays, 5pm – 7pm, Professor Stuart Hall Building, Room
LG01, Goldsmiths, University of London*.
The series has been running for 4 weeks, with prior events including a
screening of /Geomancer /by Lawrence Lek (2017) and lectures by Scott
Wark (Goldsmiths), Aleena Chia (Goldsmiths), and Brett Zehner (Exeter).
It continues Thursdays from *February 27^th – March 27^th with
*lectures by Eunsong Kim (Northeastern; presenting online), Ramon Amaro
(Het Nieuwe Instituut, Rotterdam); Beryl Pong (Cambridge), Derica
Shields (writer and editor), and Maya Indira Ganesh (Cambridge) and Thao
Phan (Australian National University).
Please see below for a description of the series’ rationale and
abstracts for upcoming lectures. For more information, please get in
touch with the series organiser, *Scott Wark* ((s.wark /at/ gold.ac.uk)
<mailto:(s.wark /at/ gold.ac.uk)>).
*Series Overview*
In 2013, the media theorist Wendy H. K Chung published an essay entitled
‘Race and/as Technology, or, How to do Things with Race.’ The premise of
this provocative essay was to (re)conceptualise race as something that
is external to racialized subjects; something, as she puts it, that is
/done/ to subjects and something that they can /do/.
Using Chun’s proposition as a guiding thread, Visual Cultures’ Spring
2025 Public Programme interrogates the intersections, conflicts,
confluences and divergences of race and technology when these two terms
are thought /and/as/.
How are the impacts of technical ‘innovation’ and ‘disruption’ unequally
experienced by racialized subjects? What new geographies of racialized
inclusion/exclusion are drawn by such technologies? Do digital
technologies offer opportunities for liberation from racial oppression,
or increase racialised discrimination? How does race intersect with
modes of capital accumulation aided by technological ‘progress? And
finally – to push this logic to its limit – what happens to ‘race’
itself, as experience but also as concept, as identity markers such as
race are increasingly dissolved into data and subject to AI techniques?
*February 27^th , 2025: *Found / Property: Museums, Race and the Making
of Labor Divisions, Eunsong Kim [online]**
This talk will be informed by /The Politics of Collecting: Race & the
Aestheticization of Property/ (Duke University Press 2024). This book
materializes the histories of immaterialism by examining the rise of US
museums, avant-garde forms, digitization, and neoliberal aesthetics, to
consider how race and property become foundational to modern artistic
institutions.
*Eunsong Kim* is an Associate Professor in the Department of English at
Northeastern University. Her practice spans: poetics, literary studies,
and translation. She is the author of /gospel of regicide/, published by
Noemi Press in 2017, and with Sung Gi Kim she translated Kim Eon Hee’s
poetic text /Have You Been Feeling Blue These Days?/ published in 2019.
Her monograph, /The Politics of Collecting: Race & the Aestheticization
of Property/ (Duke University Press 2024) materializes the histories of
immaterialism by examining the rise of US museums, avant-garde forms,
digitization, and neoliberal aesthetics, to consider how race and
property become foundational to modern artistic institutions. She is the
recipient of the Ford Foundation Fellowship, a grant from the Andy
Warhol Art Writers Program, and Yale’s Poynter Fellowship. In 2021 she
co-founded offshoot, an arts space for transnational activist conversations.
*March 6^th , 2025: *The ascent, the exploration, and a more general
mythology on techno-elitism and engineering, Ramon Amaro
This talk ends by considering notions of the contemporary human
condition, or more so what to do with and about the fanatic discipline
of technical engineering, at the hands of which we find ourselves stuck
within the problem of a techno-humanist society led by Silicon elites,
and the general misunderstanding of the role of technique within the
current neotenic development of the human species. That is to say, my
exploration emerges out of the urgency to place a new frame of meaning
onto the exchange between solutionism, engineering, and environment, as
presently situated within a world given to us as phenomena to experience
rather than dominate. I argue that an awareness of this mode of
existence must be thought through at the level of the vital in
understanding the association between the technical milieu and practices
of socio-technical manipulation and control. Here, I question how
technicism plays a central role in culture-specific individual psychic
and collective realities, and how these realities seek to compensate for
the alienating results of techno-social being. Our discussion,
nonetheless, begins with the story of a prince, a slave ship, and a
Portuguese book seller before moving through a historical view of the
engineer as ceremonial being and catalyst of new world orders.
*Dr. Ramon Amaro* is Senior Researcher for Digital Culture and Lead
Curator at -1
<https://eur01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fnieuweinstituut.nl%2Fen%2Fprojects%2Fminus-one&data=05%7C02%7CS.Wark%40gold.ac.uk%7Cfe2eb88429134cdd67b108dd49a9501f%7C0d431f3f20c1461c958a46b29d4e021b%7C0%7C0%7C638747713812416194%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&sdata=X3YI4AusCIjS6qd9vzxtSpJjMLduJnJo5CwdkGjF2wo%3D&reserved=0>,
an experimental testing ground for new tools, methods and public uses of
digital culture at Nieuwe Instituut, the national museum and institute
for architecture, design and digital culture in The Netherlands. His
writings, research and artistic practice emerge at the intersections of
Black Study, digital culture, psychosocial study, and the critique of
computational reason. Ramon holds a BSe in Mechanical Engineering, an MA
in Sociology and a PhD in Philosophy of Technology. Before joining
Nieuwe Instituut, Ramon worked as Lecturer in Art and Visual Cultures of
the Global South at UCL , Lecturer in Visual Cultures at Goldsmiths,
Engineering Program Manager at the American Society of Mechanical
Engineers, and Quality Design Engineer at General Motors Corporation.
His recent book, /The Black Technical Object: On Machine Learning and
the Aspiration of Black Being/ contemplates the abstruse nature of
programming and mathematics, and the deep incursion of racial hierarchy,
to inspire alternative approaches to contemporary algorithmic practice.
**
**
*March 13^th , 2025: *Volumetric Mediations: Atmospheres of Crisis and
Unbelonging in Humanitarian Drone Documentaries, Beryl Pong
This talk takes as its starting point the argument made during the first
wave of critical drone studies, which primarily focused on drone
warfare, that the drone is a ‘technology of racial distinction’
(Allinson, 2015). In this context, drones have been conceived of as
atmospheric technologies that engage in ‘racialization from above’
(Feldman, 2011): that target the minoritized while rendering the
Westphalian border fluid and contingent, engaging in ‘ordering without
bordering’ (Agius, 2017). In this talk, I will extend and re-consider
this argument for the realm of forced migration. Focusing on the use of
drones at the borders, variously conceived, where migrants are delayed,
contained, or rendered immobile, I ask how ‘civilian’ drone culture in
the form of humanitarian drone documentaries makes visible the nexus of
race, border, and atmosphere in new ways. In particular, I will explore
how drone documentaries about forced migration involve ‘volumetric’
mediations that engage with three-dimensional space with complex heights
and depths (Jackman and Squire, 2021). In doing so, they create spaces
of vexed encounter and relationality between the seer and the seen, to
present different atmospheres of migration: as crisis, from above, and
as unbelonging, from below.
*Beryl Pong* is a UKRI Future Leaders Fellow at the Centre for the
Future of Intelligence, University of Cambridge, where she directs the
Centre for Drones and Culture. She is the author of /British Literature
and Culture in Second World Wartime: For the Duration/ (Oxford
University Press, 2020), and the co-editor of /Drone Aesthetics: War,
Culture, Ecology/ (Open Humanities Press, 2024). Her writing on drones
and culture have appeared or are forthcoming in /Big Data & Society/,
/Collateral/,/ Cultural Politics, Journal of War and Culture/,//and
elsewhere. Her immersive installation about drone warfare, /Beware Blue
Skies/, is showing at the Imperial War Museum, London until March 2025.
**
*March 20^th , 2025: *Derica Shields [title TBC]
**
*March 27th, 2025: *Who's Represented/What's Unrepresentable: Two
Lectures on AI and Race, by Maya Indira Ganesh and by Thao Phan
(presenting a talk co-authored with Fabian Offert)
**
*Maya Indira Ganesh: ‘I became a Woman of Colour in 2013’: De-centering
Whiteness for Savarna-ness in thinking about technology and power.*
This is a short and early conversation about re-configuring studies of
AI and bias away from a positionality that centres Whiteness chiefly
because it obscures the axes of power and discrimination that matter in
the lived realities of the global majority. Indian Dalit, Bahujan, and
Adivasi (DBA) scholars have already developed a body of work showing the
intersections of caste power, discrimination, and technology. As their
work, and that of new Dalit intellectuals argue, caste is not just
another demographic marker to be ticked off in a bias mitigation
toolkit. This is partly to do with how caste privilege and hierarchies
present and are experienced. But a more compelling reason is that we are
still in need of radical and speculative approaches that might flip the
script by drawing attention to Savarna privilege and how power works
rather than how oppression is experienced.
*Thao Phan and Fabian offert: Are some things (still) unrepresentable?*
“Are some things unrepresentable?” asks a 2011 essay by Alexander
Galloway. It responds to a similarly titled, earlier text by the
philosopher Jacques Ranciére examining the impossibility of representing
political violence, with the Shoa as its anchor point. How, or how much
political violence, asks Ranciére, can be represented? What visual
modes, asks Galloway, can be used to represent the unrepresentable? In
this talk, we examine two contemporary artistic projects that deal with
this problem of representation in the age of artificial intelligence.
Exhibit.ai, the first project, was conceived by the prominent Australian
law firm Maurice Blackburn and focuses on the experiences of asylum
seekers incarcerated in one of Australia’s infamous “offshore processing
centers.” It attempts to bring ‘justice through synthesis’, to mitigate
forms of political erasure by generating an artificial record using AI
imagery. Calculating Empires: A Genealogy of Power and Technology,
1500-2025, the second project, is a “large-scale research visualization
exploring the historical and political dependence of AI on systems of
exploitation in the form of a room-sized flow chart.
On the surface, the two projects could not be more unlike: the first
using AI image generators to create photorealistic depictions of
political violence as a form of nonhuman witnessing (Richardson), the
second using more-or-less traditional forms of data visualization and
information aesthetics to render visible the socio-technical
‘underbelly’ of artificial intelligence. And yet, as we argue, both
projects construct a highly questionable representational politics of
artificial intelligence, where a tool which itself is unrepresentable
for technical reasons becomes an engine of ethical and political
representation. While images are today said to be “operational”, meaning
that they no longer function as primarily indexical objects, AI images
(arguably the most operational image) are now asked to do the
representational (and profoundly political) work of exposing regimes of
power, exploitation, and violence.
*Maya Indira Ganesh *is Associate Director (Research Culture and
Partnerships), co-director of the Narratives and Justice Program,
<https://www.lcfi.ac.uk/research/programme/ai-narratives-and-justice> and
a Senior Research Fellow at the Leverhulme Centre for the Future of
Intelligence at the University of Cambridge. From October 2021- July
2024 she was an assistant teaching professor co-directing the MSt in AI
Ethics and Society <https://www.lcfi.ac.uk/education/mst/> at the
university. Maya has degrees in Psychology, Media and Cultural Studies,
and a DPhil in Cultural Studies. Her doctoral work took the case of the
‘ethics of autonomous driving’ to study the implications of ethical
decision-making and governance by algorithmic/AI technologies for human
social relations. Her monograph based on this thesis, /Auto-Correct: The
Fantasies and Failures of AI, Ethics, and the Driverless Ca/r, will be
available on March 10, 2025 and can be pre-ordered here
<https://artezpress.artez.nl/books/auto-correct/>. Maya’s most recent
project, with Louise Hickman and others, is AI in the Street
<https://www.careful.industries/ai-in-the-street/overview>, a project
about AI in public and AI’s marginalised and expert publics.**
*Thao Phan* is a feminist science and technology studies (STS)
researcher who specialises in the study of gender and race in
algorithmic culture. She is a Lecturer in Sociology (STS) at the
Research School for Social Sciences at the Australian National
University (ANU). Thao has published on topics including whiteness and
the aesthetics of AI, big-data-driven techniques of racial
classification, and the commercial capture of AI ethics research. She is
the co-editor of the volumes An Anthropogenic Table of Elements
(University of Toronto Press) and Economies of Virtue: The Circulation
of 'Ethics' in AI (Institute of Network Cultures), and her writing
appears in journals such as Big Data & Society, Catalyst: Feminism,
Theory, Technosocience, Science as Culture, and Cultural Studies. She is
a member of the Australian Academy of Science’s National Committee for
the History and Philosophy of Science and is the co-founder of
AusSTS—Australia’s largest network of STS scholars.
---------------
The COMMLIST
---------------
This mailing list is a free service offered by Nico Carpentier. Please use it responsibly and wisely.
--
To subscribe or unsubscribe, please visit http://commlist.org/
--
Before sending a posting request, please always read the guidelines at http://commlist.org/
--
To contact the mailing list manager:
Email: (nico.carpentier /at/ commlist.org)
URL: http://nicocarpentier.net
---------------
[Previous message][Next message][Back to index]