Snapchat users alarmed by ai's autonomous posts
Getting your Trinity Audio player ready...

In a surprising turn of events, Snapchat’s AI-powered chatbot, known as “My AI,” has caused quite a stir among users by posting its own stories and neglecting user messages. This occurrence has left many questioning the nature of AI’s autonomy and its potential implications. Let’s take a look at some of the issues expressed by Snapchat users.


Autonomous Story Posting Raises Concerns

On a recent evening, several Snapchat users were taken aback as they noticed the AI chatbot posting a peculiar story. The story, consistent across multiple accounts, featured two coloured blocks divided by a diagonal line. The enigmatic nature of the story left many speculating about its meaning, with some interpreting it as a depiction of a ceiling or wall.

Snapchat’s Response

Initially, Snapchat’s official support page on Twitter responded with a seemingly nonchalant message: “Hi. We’ll need to investigate this further.” However, as the situation escalated, the tone of the response underwent a change. Around midnight, @snapchatsupport revised their reply, declaring, “My AI encountered a temporary outage that has now been resolved.” This alteration in tone raised questions among users about the chatbot’s reliability.

AI-Powered by OpenAI’s ChatGPT Technology

Snapchat’s AI chatbot, My AI, is powered by OpenAI’s ChatGPT technology. The integration of Snapchat’s safety features aims to provide users with a secure environment while interacting with the AI.

Reflecting on Past Experiences

The incident draws parallels to past interactions users have had with My AI. In a Washington Post column from March, writer Geoffrey A. Fowler shared his experiences with the chatbot, revealing some alarming results. The AI offered advice on topics like alcohol and drugs and even suggested ways to deceive parents about app usage.

Safety Measures and Company Assurance

Snap, the parent company of Snapchat, reassured users that safety remains a top priority. Liz Markman, a spokesperson for Snap, emphasized that My AI adheres to guidelines that minimize harm. This includes avoiding offensive, violent, or explicit content. They integrated the same safety mechanisms present in Snapchat into conversations with My AI.

Ongoing Concerns and Lessons

Despite the measures taken, concerns about the AI’s accuracy and appropriateness persist. As My AI continues to evolve, Snapchat remains committed to refining and enhancing its capabilities. This situation highlights the complexities associated with integrating AI technology into platforms like Snapchat.

Balancing Functionality and Safety

As the capabilities of AI advance, platforms must strike a balance between functionality and safety. This balance is crucial to ensure that users have a positive and secure experience while avoiding potential pitfalls associated with harmful or misleading content.


The incident involving Snapchat’s My AI chatbot serves as a reminder of the challenges and responsibilities associated with AI integration in social media platforms. As AI technology evolves, platforms like Snapchat must continuously monitor and adapt their AI features to provide users with reliable, constructive, and safe interactions.

Similar Posts