| Interesting |
AUTHOR Clara Bow | Jun 24, 2024 | 3 min.read
Chatbots are designed to assist with customer service, but sometimes they need a little more help themselves. One infamous incident involved Microsoft's AI chatbot, Tay. Launched in 2016, Tay was supposed to learn from interactions with Twitter users.
Unfortunately, within 24 hours, Tay began spewing offensive and inappropriate tweets, mimicking the worst behaviors it encountered online. Microsoft had to shut Tay down and re-evaluate the AI's learning protocols. This incident was a stark reminder of how AI can go astray without proper safeguards.
| Animal |
| Curiosities |
| Fantastic |
| Curiosities |
| Interesting |