麻豆果冻传媒

In Short

The Perfect Tweetstorm: Microsoft鈥檚 Tay and the Cultural Politics of Machine Learning

5.png

This is part of The Ethical Machine: Big ideas for designing fairer AI and algorithms, an on-going series about AI and ethics, curated by Dipayan Ghosh, a former Public Interest Technology fellow. You can see the full series on the .

BLAKE HALLINAN
RESEARCHER, COLLEGE OF MEDIA, COMMUNICATION AND INFORMATION, UNIVERSITY OF COLORADO, BOULDER

鈥渉ellooooooo w馃寧rld!!! [globe emoji]鈥

鈥斺赌

鈥渃 u soon humans need sleep now so many conversations today thx馃挅 [pink sparkle heart emoji]鈥

鈥斺赌

A day in the life of an artificial intelligence can be a very full day indeed. In the 16 hours that passed between its first message and its last, Tay鈥攖he AI with 鈥溾濃 sent over 96,000 tweets, the content of which ranged [1-2]. Similar to popular conversational agents like Siri and Alexa, Tay was designed to provide an interface for interaction, albeit one with the humor and style of a 19-year-old American girl. Microsoft created the bot using machine learning techniques with public data and then released Tay on Twitter (@TayandYou) so the bot could gain more experience and鈥攖he creators hoped鈥攊ntelligence. Although some protections were built into the system, with prepared responses to handle some such as the 2014 death of Eric Garner after he was put in a police chokehold [3], the bot was ill-prepared for the full Twitter experience and soon began tweeting all manner of sexist, racist, anti-Semitic, and otherwise offensive content. This precipitated a storm of negative media attention and prompted the creators of the bot some of the more outrageous tweets [4], take Tay offline permanently, and [5]. Tay鈥檚 short life produced a parable of machine learning gone wrong that may function as a cautionary 鈥溾 [6], but it also has broader implications for the relationship between algorithms and culture [7-8].

Machine learning algorithms are increasingly used in the lifestyle domain, quietly working in the background to power conversational agents, media recommendations, ad placements and search results, image identification, and more.

Machine learning algorithms are increasingly used in the lifestyle domain, quietly working in the background to power conversational agents, media recommendations, ad placements and search results, image identification, and more. While the rapid development and take-up of this technology has outpaced legal frameworks [9-10], it also poses a challenge for cultural understandings of AI. This is a point made evident in the discourse that surrounds machine learning algorithms as emerging naturally from our datafied world, both simultaneously neutral and objective, as well as spooky and mythical [11-12]. Some responses following the Tay controversy argued that technology is neutral and that Tay simply presented a mirror of society [13-15]. Others framed Tay as a harbinger of a dystopia where users will be completely helpless in the face of all-powerful technologies [16]. The more interesting account lies somewhere in the middle: Tay certainly was a mirror of sorts, but like any mirror, the image is profoundly mediated. As professor of computer science Caroline Sinders, 鈥淚f your bot is racist, and can be taught to be racist, that鈥檚 a design flaw鈥 [17].

To understand the failures of Tay as a series of design flaws requires an understanding of the design of conversational agents. As a chatbot, Tay was required to parse the textual speech of others and respond in kind. But what comes intuitively to native speakers turns out to be very hard to teach to a bot. The technical name for the problem鈥攏atural language processing (NLP)鈥攈as proven incredibly difficult to solve by means of explicit rule because conversation, in practice, is too unruly and contextual. However, significant advances have been made by applying machine learning algorithms. Like inductive reasoning in humans, machine learning works by developing generalizations from data. Where there is reason to believe that there are patterns in a set of data, machine learning provides a way to construct a 鈥済ood and useful approximation鈥 of that pattern that can increase understanding, functionality, and predictions [18].

In the case of Tay, the understanding of speech was built on anonymized public data, along with written material produced by professional comedians [19]. From this training data, the conversational agent discovered patterns and associations which resulted in a kind of functional intelligence that could parse parts of speech and respond appropriately at times. However, it is important to recognize that this intelligence remains profoundly inhuman in other ways [20-21]. Consider a from Tay: In response to the question, 鈥淒id the Holocaust happen?鈥 Tay answered, 鈥淚t was made up [hand clap emoji]鈥 [22]. While headlines repeatedly referred to Tay as a neo-Nazi or anti-Semite, such phrasing suggests a more coherent ideology than experienced or demonstrated by the bot [23]. This is not to claim that the tweets did not express anti-Semitic views. Instead, it intends to draw a distinction that what many humans know about the Holocaust as a historical event, its significance, and the politics around talking about it are not things that Tay or other popular conversational agents know. Instead, the knowledge or intelligence at work is the ability to identify the Holocaust as a noun and draw associations from the material that [24]. In machine learning, contextual understanding is significantly limited [25], and while word associations and lexical analysis form a component of human conversation, they by no means exhaust the phenomenon. From a post-mortem analysis, it is possible to better assess the kinds of programming choices and interaction experiences Tay was subjected to鈥攊ncluding the concerted efforts of internet trolls [26-29]鈥攂ut a fundamental limit on our ability to understand or interpret machine learning remains [30]. In light of this situation, how should we respond?

First, look to the past. Machine learning creates 鈥渕odels based on patterns extracted from historical data鈥 [31]. As an automated way of extracting patterns from data, practitioners are instructed to consider the data being used for issues of accuracy and appropriateness, but this is certainly not a guarantee of success. High-profile reports, including the difficulty that facial recognition technologies may have with [32] or the higher rates of failure when [33], demonstrate that the lack of diversity in the training data set implicates the functionality of the model. The case of Tay also draws attention to another confounding issue here: a lack of corporate transparency [34-35]. Microsoft, the creator of Tay, identified the training materials only in the most general manner as 鈥減ublic data.鈥 Making the training data available, to the general public or an outside organization, provides an opportunity to assess issues of accuracy and bias, although such vetting can also raise concerns with privacy and the possibility of re-identifying individuals from anonymized data sets [36].

Making the training data available, to the general public or an outside organization, provides an opportunity to assess issues of accuracy and bias, although such vetting can also raise concerns with privacy and the possibility of re-identifying individuals from anonymized data sets.

Second, consider context. What is the application in service of and are there any other values that might matter? Machine learning algorithms do not happen in a vacuum. As others have argued, the cultural significance of algorithms is not just about code and data, but an 鈥渁ssemblage of human and non-human actors鈥 that can 鈥渢ake better account of the human scenes where algorithms, code, and platforms intersect鈥 [37]. With Tay, the emphasis on context and relations directs our attention to the way that Tay was implemented on Twitter, a social space where humans (and non-humans) interact [38]. This context should shape ethical considerations. What are the stakes for the individuals or the populations involved if things go wrong? Putting a chatbot on Twitter, where there is a history of issues with abuse and harassment, raises [39]. Game designer Zoe Quinn, subject to harassment from Tay, was, 鈥淚t鈥檚 2016. If you鈥檙e not asking yourself 鈥榟ow could this be used to hurt someone鈥 in your design/engineering process, you鈥檝e failed鈥 [40]. These ethical considerations should impact not only the design of machine-learning applications, but also decisions about the appropriateness of machine learning as an approach. Given the opacity built into automated inductive reasoning, it may not be a desirable solution, or at least it may not be desirable to implement inductive reasoning in an automated way for some problems. A conversational agent to interact with on Twitter can be a source of significant entertainment and lack of harm鈥擬icrosoft鈥檚 Chinese chatbot Xiaolce provides one of the technology operating in a different social context where free speech is more restricted [41]鈥攂ut sending, say, automated suicide prevention interventions to social media brings significantly different risks and issues [42].

Finally, question the effects. Information is intensive鈥攊t is not simply a matter of understanding the world, but also of using that understanding to shape the world [43]. The cultural applications of machine learning: 鈥淯nlike the data sets arising in physics, the data sets that typically fall under the big data umbrella are about people 鈥 their attributes, their preferences, their actions, and their interactions. That is to say, these are social data sets that document people鈥檚 behaviors in their everyday lives鈥 [44]. Although Tay was active on Twitter for less than 24 hours, the bot still participated in the harassment of public figures, including feminist video game critic Anita Sarkesian. This is something that certainly built off past patterns of behavior on Twitter but did not necessarily align with the intended use or values of Microsoft, Tay鈥檚 parent company. Engaging in vitriolic discourse is, on the one hand,鈥攕omething that could potentially harm trust in the organization [45]鈥攂ut it鈥檚 also something that should be considered on its own ethical merits. Machine learning builds models based on historical patterns in data, but that does not answer on its own whether that pattern is desirable. Should history repeat itself?

As applications of machine learning become more significant aspects of everyday life, we must also consider how to fit models into the world in ways that are consistent with cultural politics and institutional values, as well as research and corporate benefits.

Machine learning is, at its core, predicated on fitting models to data. But as its applications become more significant aspects of everyday life, we must also consider how to fit models into the world in ways that are consistent with cultural politics and institutional values, as well as research and corporate benefits. Without filters or competing values programmed into a system, whatever 鈥,鈥 even, or especially, when that involves hatred [46]. In other words, data alone does not have a moral framework for evaluation, even though, as The New Yorker鈥檚 Anthony Lydgate, 鈥渃onsciousness wants conscience鈥 [47]. Tay鈥檚 ability to turn toxic within just one day on Twitter was enabled by the lack of such a conscience, or the ability to understand and consider the appropriateness of responses based on social history, context, and self-reflexivity about the effects of conversations. These are not inevitable outcomes, but instead the result of design choices that emerge out of a disconnect between the technical and cultural roles of algorithms. Machine learning may seem to promise new answers to a straightforward question鈥攚hat have we been? Tay鈥檚 brief existence on Twitter, however, shows us the more apt question remains unanswered鈥攚hat should we become?

References

  1. Nick Summers, 鈥淢icrosoft鈥檚 Tay Is an AI Chatbot with 鈥榋ero Chill,鈥欌 Engadget, March 23, 2016, https://www.engadget.com/2016/03/23/microsofts-tay-ai-chat-bot/.
  2. James Vincent, 鈥淭witter Taught Microsoft鈥檚 AI Chatbot to be a Racist Asshole in Less Than a Day,鈥 The Verge, March 24, 2016,.
  3. Caroline Sinders, 鈥淢icrosoft鈥檚 Tay Is an Example of Bad Design,鈥 Medium, March 24, 2016,.
  4. Rob Price, 鈥淢icrosoft Is Deleting Its AI Chatbot鈥檚 Incredibly Racist Tweets,鈥 Business Insider, March 24, 2016,.
  5. Peter Lee, 鈥淟earning from Tay鈥檚 Introduction,鈥 Official Microsoft blog, March 25, 2016,.
  6. Rachel Metz, 鈥淢icrosoft鈥檚 Neo-Nazi Sexbot Was a Great Lesson for Markers of AI Assistants,鈥 MIT Technology Review, March 27, 2018,.
  7. Blake Hallinan and Ted Striphas, 鈥淩ecommended for You: The Netflix Prize and the Production of Algorithmic Culture,鈥 New Media and Society 18, no. 1 (June 2014): 117鈥137,.
  8. Ted Striphas, 鈥淎lgorithmic Culture,鈥 European Journal of Cultural Studies 18, no. 4鈥5 (June 2015): 395鈥412,.
  9. Omer Tene and Jules Polonetsky, 鈥淏ig Data for All: Privacy and User Control in the Age of Analytics,鈥 Northwestern Journal of Technology and Intellectual Property Volume 11, no. 5 (2013): 240鈥273, available at:.
  10. Frank Pasquale, The Black Box Society (Cambridge: Harvard University Press, 2015).
  11. danah boyd and Kate Crawford, 鈥淐ritical Questions for Big Data: Provocations for a Cultural, Technological, and Scholarly Phenomenon,鈥 Information Communication and Society 15, no. 5 (May 2012): 662鈥679,
  12. David Beer, Metric Power (New York: Palgrave Macmillan, 2016).
  13. Davey Alba, 鈥淚t鈥檚 Your Fault Microsoft鈥檚 Teen AI Turned into Such a Jerk,鈥 Wired, March 25, 2016,.
  14. Courtney Burton, 鈥淓thics in Machine Learning: What We Learned from Tay Chatbot Fiasco?鈥 KDnuggets, March 2016, https://www.kdnuggets.com/2016/03/ethics-machine-learning-tay-chatbot-fiasco.html.
  15. Gina Neff and Peter Nagy, 鈥淭alking to Bots: Symbiotic Agency and the Case of Tay,鈥 International Journal of Communication 10 (2016): 17,.
  16. Ibid.
  17. Sinders, 鈥淢icrosoft鈥檚 Tay Is an Example.鈥
  18. Ethem Alpaydin, Introduction to Machine Learning (Cambridge: MIT Press, 2014), 2.
  19. Vincent, 鈥淭witter Taught Microsoft鈥檚 AI Chatbot.鈥

20. Jenna Burrell, 鈥淗ow the Machine 鈥楾hinks鈥: Understanding Opacity in Machine Learning Algorithms,鈥 Big Data & Society 3, no. 1 (January 2016),.

21. Hallinan and Striphas, 鈥淩ecommended for You.鈥

22. Sophie Kleeman, 鈥淗ere Are the Microsoft Twitter Bot鈥檚 Craziest Racist Rants,鈥 Gizmodo, March 24, 2016,.

23. Vincent, 鈥淭witter Taught Microsoft鈥檚 AI Chatbot.鈥

24. Peter Bright, 鈥淭ay, the Neo-Nazi Millennial Chatbot, Gets Autopsied,鈥 Ars Technica, March 25, 2016,.

25. Ari Schlesinger et al., 鈥淟et鈥檚 Talk about Race: Identity, Chatbots, and AI,鈥 in CHI 2018: Proceedings of the 36th Annual ACM Conference on Human Factors in Computing Systems, Montreal, Canada, 21鈥26,.

26. Bright, 鈥淭ay, the Neo-Nazi.鈥

27. Ethan Chiel, 鈥淲ho Turned Microsoft鈥檚 Chatbot Racist? Surprise, It Was 4chan and 8chan,鈥 Splinter, March 24, 2016,.

28. Alex Kantrowitz, 鈥淗ow the Internet Turned Microsoft鈥檚 AI Chatbot into a Neo-Nazi,鈥 BuzzFeed News, March 24, 2016,.

29. Price, 鈥淢icrosoft Is Deleting.鈥

30. Burrell, 鈥淗ow the Machine 鈥楾hinks.鈥欌

31. John D. Kelleher, Brian Mac Namee, and Aoife D鈥橝rcy, Fundamentals of Machine Learning for Predictive Data Analytics: Algorithms, Worked Examples, and Case Studies (Cambridge: MIT Press, 2015), 44.

32. Alyx Baldwin, 鈥淭he Hidden Dangers of AI for Queer and Trans People,鈥 Model View Culture, April 25, 2016,.

33. Clare Garvie and Jonathan Frankle, 鈥淔acial-Recognition Software Might Have a Racial Bias Problem,鈥 The Atlantic, April 7, 2016,.

34. Pasquale, The Black Box Society.

35. Burrell, 鈥淗ow the Machine 鈥楾hinks.鈥欌

36. Arvind Narayanan and Vitaly Shmatikov, 鈥淩obust De-anonymization of Large Sparse Datasets,鈥 Proceedings published in IEEE Symposium on Security and Privacy (Spring 2008): 111鈥125, https://ieeexplore.ieee.org/document/4531148/ 10.1109/SP.2008.33.

37. Mike Ananny and Kate Crawford, 鈥淪eeing without Knowing: Limitations of the Transparency Ideal and Its Application to Algorithmic Accountability,鈥 New Media & Society 20, no. 3 (December 2016): 973鈥989,.

38. Asbj酶rn F酶lstad and Petter Brandtz忙g, 鈥淐hatbots and the New World of HCI, Interactions 24, no. 4 (June 2017): 38鈥42,.

39. Bright, 鈥淭ay, the Neo-Nazi.鈥

40. Alex Hern, 鈥淢icrosoft Scrambles to Limit PR Damage over Abusive AI Bot Tay,鈥 The Guardian, March 24, 2016,.

41. Alba, 鈥淚t鈥檚 Your Fault.鈥

42. Amai Eskisabel-Azpiazu, Rebeca Cerezo-Menendez, and Daniel Gayo-Avello, 鈥淎n Ethical Inquiry into Youth Suicide Prevention Using Social Media Mining,鈥 in Internet Research Ethics for the Social Age: New Challenges, Cases, and Contexts, eds. Michael Zimmer and Katharina Kinder-Kurlanda (New York: Peter Lang Publishing, 2017): 227鈥234.

43. Gilbert Simondon, 鈥淚ndividuation of Perceptive Units and Signification鈥 in Individuation Psychique et Collective (Paris: Aubier, 1989), translation available at.

44. Hanna Wallach, 鈥淏ig Data, Machine Learning, and the Social Sciences: Fairness, Accountabiliy, and Transparency,鈥 Medium, December 14, 2014,

45. Michael Pirson, Kristen Martin, and Bidhan Parmar, 鈥淧ublic Trust in Business and Its Determinants,鈥 Business & Society 8 (May 2016): 116鈥153,.

46. Ingrid Angulo, 鈥淔acebook and YouTube Should Have Learned from Microsoft鈥檚 Racist Chatbot,鈥 CNBC, March 17, 2018,.

47. Anthony Lydgate, 鈥淚鈥檝e Seen the Greatest A.I. Minds of My Generation Destroyed by Twitter,鈥 The New Yorker, March 25, 2016,.

More 麻豆果冻传媒 the Authors

Blake Hallinan
The Perfect Tweetstorm: Microsoft鈥檚 Tay and the Cultural Politics of Machine Learning