Art and AI. Machines Visions.
Date: Week 9, March 1 - 5, 2021
Time: 10:00 - 17:00
The course takes place online: The link is sent to you by email.
Course requirements:
- active participation in group activity
- active individual work (via pad)
- fictional short story, 2-3 pages, to be handed until Sunday, March 7 (via email)
Course Abstract
Artificial intelligence (AI) is all around us, transforming the way we see the world and the world sees us. Sometimes, we barely notice, sometimes, the results of AI can be very impressive, though sometimes so impressively bad, that Hito Steyrl and others started to talk about “Artificial Stupidity”. But almost always, the actual procedures of AI remain mysterious, locked away in a black box.
In this module, we will look at a recent wave of artistic and theoretical work that tries to open this black box, creating a language and an aesthetics for critical engagement. In particular, we will look at image recognition AI, what the artist Trevor Paglen calls “predatory vision” and the data scientist Joy Buolamwini calls the “coded gaze”, referring to gender and racial biases in these systems.
Course Works and Notes
Monday Morning
Similarity & Difference: Observing the results of machine visions
- !Mediengruppe Bitnik SAME SAME. WATCHING ALGORITHMS - CABARET VOLTAIRE EDITION 2015
- Mario Klingemann & Simon Doury X Degrees of Separation, 2018
How Image Recognition works, technically
- Activation Atlas, 2019
- Understanding image models and predictions using an Activation Atlas, Google AI Adventures, 2019, 7:49
Monday Afternoon
- Select one story, read it. Write short summary in the pad and how the story relates to the actual company, add your name to the entry
- short research on image recognition. Select one news story that you fund interesting. Enter the link to the pad and write why you selected this story. Add your name to the entry
Tuesday Morning
Bias in Vision
The racist soap dispenser @ Facebook, 2017
PortraitAI Sarah L. Fossheim's experiment with poc faces
Colour Bias in Analogue Photography
Color film was built for white people, 4:39, 2015
Roth, Lorna. 2009. “Looking at Shirley, the Ultimate Norm: Colour Balance, Image Technologies, and Cognitive Equity.” Canadian Journal of Communication 34(1).
* http://colourbalance.lornaroth.com
Osinubi, Ade. 2018. “The Inherent Color Bias of Image Capturing Technology: Colorism in Photography.” Medium.com
Ewart, Asia. 2020. “The Shirley Card: Racial Photographic Bias through Skin Tone.” The Shutterstock Blog (June 30).
Joy Buolamwini The coded gaze
- AI, Ain't I A Woman?, 2018, 3:33
- https://www.notflawless.ai, see stories there,
- How I am fighting Bias in Algorithms. TEDx Talk, 2017
Another form of Bias:
- !Mediengruppe Bitnik! Dada. State of the Reference, 2017
Further Reading:
- Eubanks, Virginia. Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor. First Edition. New York, NY: St. Martin’s Press, 2017.
- Noble, Safiya Umoja. Algorithms of Oppression: How Search Engines Reinforce Racism. New York: New York University Press, 2018
Tuesday Afternoon
Timnit Gebru: Computer Vision: Who is Helped and Who is Harmed? (02.2021), 03-46
- Watch the talk and note the main points in the pad.
- Rothkopf, Joshua. 2020. “Deepfake Technology Enters the Documentary World.” The New York Times.
Trevor Paglen
Is Photography Over?, 4 Parts, Fotomuseum Winterthur, 03.2014
- Read these four short parts. Select four sentences that you find interesting and copy them to the pad. Please organize them in the pad according to the section.
Wednesday Morning
Trevor Paglen & Cate Crawford
- Datafication of Science Lecture at HKW, Jan 12, 2019. 32 Min
- |Discussion with Kate Crawford, Trevor Paglen, Felix Stalder HKW, Jan 12, 2019, 14Min
- Kate Crawford and Trevor Paglen, AI. The Politics of Images in Machine Learning Training Sets
- Fondazoione Prada: Training Humans, Fondazione Prada, Venice, 12 Sep 2019 – 24 Feb 2020 Review, we-make-money-not-art.com, Dec. 16.2019
- Paglen, Trevor. “Invisible Images (Your Pictures Are Looking at You).” The New Inquiry (blog), December 8, 2016
- Why our facial expressions don’t reflect our feelings, BBC, 10.05.2018
Wednesday Afternoon
Reading your character (and future behavior) from your face
Levin, Sam. 2017. “Face-Reading AI Will Be Able to Detect Your Politics and IQ, Professor Says.” The Guardian (Sept. 12).
Critique of this paper
Agüera y Arca, Blaise, Margret Mitchell, and Alexander Todorov. 2018. “Do Algorithms Reveal Sexual Orientation or Just Expose Our Stereotypes?” Medium (Jan 11).
A fundamental critique of the approach
Agüera y Arca, Blaise, Margret Mitchell, and Alexander Todorov. 2017. “ Physiognomy’s New Clothes.” Medium (May 7).
Further reading
Wang, Yilun, and Kosinski, Michal. 2017. “Deep Neural Networks Are More Accurate than Humans at Detecting Sexual Orientation from Facial Images.” ((Original “ ai gaydar” paper)
Thursday
Artist Talk and Workshop by Adam Harvey (Full day)
Adam Harvey's Art projects about privacy, computer vision, and surveillance
Friday Morning
Discussion of Readings Wednesday Afternoon
Fairness oder Vorurteil. Einsatz Künstlicher Intelligenz bei der Jobbewerbung fragwürdig Bayrischer Rundfunk. 2021 English Version
Hito Steyerl (2016) Keynote Conversation: Anxious to Act transmediale 2016, 23 Min
Steyerl, Hito. 2016. “A Sea of Data: Apophenia and Pattern (Mis-)Recognition.” e-flux Journal (april)
Salvaggio, Eryk. 2020. “Creative Strategies for Algorithmic Resistance!” Cybernetic Forests (June 29).
Friday Afternoon
Writing of short story
Material added after the course
Schiller, Devon. 2020. On the Basis of Face: Biometric Art as Critical Practice, Its History and Politics. Amsterdam: Insitute for Network Cultures.