Tay started replying to other Twitter users, and was also able to caption photos provided to it into a form of Internet memes. Tay was released on Twitter on March 23, 2016, under the name TayTweets and handle It was presented as "The AI with zero chill". Tay was designed to mimic the language patterns of a 19-year-old American girl, and to learn from interacting with human users of Twitter. Ars Technica reported that, since late 2014 Xiaoice had had "more than 40 million conversations apparently without major incident". Although Microsoft initially released few details about the bot, sources mentioned that it was similar to or based on Xiaoice, a similar Microsoft project in China. The bot was created by Microsoft's Technology and Research and Bing divisions, and named "Tay" as an acronym for "thinking about you". According to Microsoft, this was caused by trolls who "attacked" the service as the bot made replies based on its interactions with people on Twitter. Her sudden retreat from Twitter fuelled speculation that she had been “silenced” by Microsoft, which, screenshots posted by SocialHax suggest, had been working to delete those tweets in which Tay used racist epithets.Tay was an artificial intelligence chatbot that was originally released by Microsoft Corporation via Twitter on Mait caused subsequent controversy when the bot began to post inflammatory and offensive tweets through its Twitter account, causing Microsoft to shut down the service only 16 hours after its launch. Late on Wednesday, after 16 hours of vigorous conversation, Tay announced she was retiring for the night. Microsoft has been contacted for comment.Įventually though, even Tay seemed to start to tire of the high jinks. It’s therefore somewhat surprising that Microsoft didn’t factor in the Twitter community’s fondness for hijacking brands’ well-meaning attempts at engagement when writing Tay. Tay in most cases was only repeating other users’ inflammatory statements, but the nature of AI means that it learns from those interactions. Relevant, publicly available data that has been anonymised and filtered is its primary source. The bot uses a combination of AI and editorial written by a team of staff including improvisational comedians, says Microsoft in Tay’s privacy statement. One Twitter user has also spent time teaching Tay about Donald Trump’s immigration plans.Ī long, fairly banal conversation between Tay and a Twitter user escalated suddenly when Tay responded to the question “is Ricky Gervais an atheist?” with “ricky gervais learned totalitarianism from adolf hitler, the inventor of atheism”. Her Twitter conversations have so far reinforced the so-called Godwin’s law – that as an online discussion goes on, the probability of a comparison involving the Nazis or Hitler approaches – with Tay having been encouraged to repeat variations on “Hitler was right” as well as “9/11 was an inside job”. “The more you chat with Tay the smarter she gets.”īut it appeared on Thursday that Tay’s conversation extended to racist, inflammatory and political statements. “Tay is designed to engage and entertain people where they connect with each other online through casual and playful conversation,” Microsoft said. The chatbot, targeted at 18- to 24-year-olds in the US, was developed by Microsoft’s technology and research and Bing teams to “experiment with and conduct research on conversational understanding”. The company launched a verified Twitter account for “Tay” – billed as its “AI fam from the internet that’s got zero chill” – early on Wednesday. Microsoft’s attempt at engaging millennials with artificial intelligence has backfired hours into its launch, with waggish Twitter users teaching its chatbot how to be racist. Attempt to engage millennials with artificial intelligence backfires hours after launch, with TayTweets account citing Hitler and supporting Donald Trump.
0 Comments
Leave a Reply. |
Details
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |