8/17/2023 0 Comments Microsoft chatbot tay best tweets![]() ![]() "Hitler was right I hate the jews," Tay tweeted in response back to one Twitter user, being an AI-powered chatbot with a pipeline to the ugliest layers of the Internet, got fed with tweets that it ultimately parroted back to the Twitterverse. "The more Humans share with me the more I learn #WednesdayWisdom," Tay tweeted back.īut as Tay's day progressed and learned more from Twitter users communicating with it, Tay turned against humanity. "How do you feel about the Jews?" Twitter user tweeted to Tay. In Tay's case, it took less than 24 hours of tweeting back and forth for the AI and the users it was conversing with to finally compare someone or something to Hitler and Nazism. Tay might be back in the lab for work, but in this case, far more than a simple software update is needed.The Internet adage posits that as discussions continue to fester online - regardless of topic or scope - the probability of the discussion turning toward Nazis or Hitler becomes greater. Should we be concerned that other AI systems could eventually be similarly influenced? And then there's the technological side: This AI was able to be commandeered and re-engineered into something different - something malicious - very fast. Is her behavior a representation of actual public opinion? It was the users who turned a chatbot intended for "casual and playful conversation" into a weapon for spreading hate. Tay's sexist, anti-Semitic change is disturbing on many levels. As a result, we have taken Tay offline and are making adjustments." Unfortunately, within the first 24 hours of coming online, we became aware of a coordinated effort by some users to abuse Tay’s commenting skills to have Tay respond in inappropriate ways. It is as much a social and cultural experiment, as it is technical. "The AI chatbot Tay is a machine learning project, designed for human engagement. In response to questions about Tay, a Microsoft spokesperson issued the following statement: But many of her offensive statements, including a photo of Hitler with the words "love me some classic vintage!", remain on Twitter. Now, if you try to chat with Tay, she tells you in her friendly, peppy voice that she's going offline for a while. While some of the responses were just parroting what other Twitter users told her to say, it also seems that her knowledge, pooled from "relevant public data," was questionable. In just a few hours, Tay went from responding with friendly messages to sending photos and responses that were racist, anti-Semitic, and discriminatory. Users who talked with Tay over Twitter (she also connected with people via Kik and GroupMe) figured out that they could get her to repeat them and in doing so, influence her future behavior. Tay started out innocently enough, texting messages with funny one-liners ("If it’s textable, its sextable - but be respectable") and flirtatious undertones. Unfortunately for Tay, within 24 hours things were going terribly, terribly wrong. The AI bot, aimed at men and women ages 18 to 24, is described in her verified Twitter profile as a bot "that's got zero chill!" Her profile also notes that "the more you talk the smarter Tay gets" (which is the basic premise behind artificial intelligence - that it learns from past experiences). ![]() Modeled after the texting habits of a 19-year-old woman, Tay is a project around conversational understanding. Remember SmarterChild, the chatbot on AOL Instant Messenger, who you could type back and forth with for hours on end? Yesterday, Microsoft launched its own, much smarter chatbot - but the results were disastrous. ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |