Chat inter recial communicating on online dating sites
It took less than 24 hours for Twitter to corrupt an innocent AI chatbot.
Yesterday, Microsoft unveiled Tay — a Twitter bot that the company described as an experiment in "conversational understanding." The more you chat with Tay, said Microsoft, the smarter it gets, learning to engage people through "casual and playful conversation." Unfortunately, the conversations didn't stay playful for long.
We conducted semistructured interviews with 38 participants in White racist Internet chat rooms, examining the extent to which people would, in this unique environment, advocate interracial violence in response to purported economic and cultural threats.
Capitalizing on the anonymity and candor of chat room interactions, this study provides an unusual perspective on extremist attitudes.
We experimentally manipulated the nature and proximity of the threats.
Qualitative and quantitative analyses indicate that the respondents were most threatened by interracial marriage and, to a lesser extent, Blacks moving into White neighborhoods.
One message “called the hypothetical hanging of a Mexican child ‘piñata time’” while other messages quipped that “abusing children was sexually arousing,” according to images of the chat described by the Crimson. And in mid-April, after administrators discovered the offensive, racially charged meme exchanges, at least 10 incoming students who participated in the chat received letters informing them that their offers of admission had been revoked.
“As we understand you were among the members contributing such material to this chat, we are asking that you submit a statement by tomorrow at noon to explain your contributions and actions for discussion with the Admissions Committee.” “It is unfortunate that I have to reach out about this situation,” the email continued.) we can see that many of the bot's nastiest utterances have simply been the result of copying users.If you tell Tay to "repeat after me," it will — allowing anybody to put words in the chatbot's mouth.Pretty soon after Tay launched, people starting tweeting the bot with all sorts of misogynistic, racist, and Donald Trumpist remarks.And Tay — being essentially a robot parrot with an internet connection — started repeating these sentiments back to users, proving correct that old programming adage: flaming garbage pile in, flaming garbage pile out.