Microsoft launched a new “Tay” artificial intelligence engine yesterday. She was supposed to be fun, smart and kind – just like all of my failed Tinder dates. (I’m kidding, I’m happily married.) Instead, she turned out to be a vulgar and disgusting robot. I’m not going to quote anything here, it was that bad.
For whatever reason – I’m not sure how – Tay started spouting off racist, anti-Semitic and all-around not nice things to her followers and others. Maybe she was hacked, maybe she was learning from all the naughty things the Internet spewed at her, but Microsoft hadn’t planned on this happening. I’m not even sure why she was able to say some of the things she said – you’d think Microsoft would have limited her vocabulary.
Microsoft ultimately “put her to bed” after a long, tiring day. It’s unclear how long Tay will be grounded for, but she said enough to send her away to boarding school, in my opinion.