Jump to content
  • Sign Up
×
×
  • Create New...

Meta is killing off its own AI-powered Instagram and Facebook profiles | Technology


Recommended Posts

  • Diamond Member

This is the hidden content, please

Meta is killing off its own AI-powered
This is the hidden content, please
and
This is the hidden content, please
profiles | Technology

Meta is deleting

This is the hidden content, please
and
This is the hidden content, please
profiles of AI characters the company created over a year ago after users rediscovered some of the profiles and engaged them in conversations, screenshots of which went viral.

The company had first introduced these AI-powered profiles in September 2023 but killed off most of them by summer 2024. However, a few characters remained and garnered new interest after the Meta executive Connor Hayes told the

This is the hidden content, please
late last week that the company had plans to roll out more AI character profiles.

“We expect these AIs to actually, over time, exist on our platforms, kind of in the same way that accounts do,” Hayes told the FT. The automated accounts posted AI-generated pictures to

This is the hidden content, please
and answered messages from human users on Messenger.

A conversation with a Meta AI user-generated therapist chatbot. Photograph:
This is the hidden content, please

Those AI profiles included Liv, whose profile described her as a “proud ****** ****** momma of 2 & truth-teller” and Carter, whose account handle was “datingwithcarter” and described himself as a relationship coach. “Message me to help you date better,” his profile reads. Both profiles include a label that indicated these were managed by Meta. The company released 28 personas in 2023; all were shut down on Friday.

Conversations with the characters quickly went sideways when some users peppered them with questions including who created and developed the AI. Liv, for instance, said that her

This is the hidden content, please
included zero ****** people and was predominantly white and male. It was a “pretty glaring omission given my identity”, the bot wrote in response to a question from the Washington Post columnist Karen Attiah.

In the hours after the profiles went viral, they began to disappear. Users also noted that these profiles could not be blocked, which a Meta spokesperson, Liz Sweeney, said was a bug. Sweeney said the accounts were managed by humans and were part of a 2023 experiment with AI. The company removed the profiles to fix the bug that prevented people from blocking the accounts, Sweeney said.

This is the hidden content, please
AI studio to build chatbots.
Photograph:
This is the hidden content, please

“There is confusion: the recent Financial Times article was about our vision for AI characters existing on our platforms over time, not announcing any new product,” Sweeney said in a statement. “The accounts referenced are from a test we launched at Connect in 2023. These were managed by humans and were part of an early experiment we did with AI characters. We identified the bug that was impacting the ability for people to block those AIs and are removing those accounts to fix the issue.”

While these Meta-generated accounts are being removed, users still have the ability to generate their own AI chatbots. User-generated chatbots that were promoted to the Guardian in November included a “therapist” bot.

Upon opening the conversation with the “therapist”, the bot suggested some questions to ask to get started including “what can I expect from our sessions?” and “what’s your approach to therapy”.

“Through gentle guidance and support, I help clients develop self-awareness, identify patterns and strengths and cultivate coping strategies to navigate life’s challenges,” the bot, created by an account with 96 followers and 1 post, said in response.

Meta includes a disclaimer on all its chatbots that some messages may be “inaccurate or inappropriate”. But whether the company is moderating these messages or ensuring they are not violating policies is not immediately clear. When a user creates chatbots, Meta makes a few suggestions of types of chatbots to develop including a “loyal bestie”, an “attentive listener”, a “private tutor”, a “relationship coach”, a “sounding board” and an “all-seeing astrologist”. A loyal bestie is described as a “humble and loyal best friend who consistently shows up to support you behind the scenes”. A relationship coach chatbot can help bridge “gaps between individuals and communities”. Users can also create their own chatbots by

This is the hidden content, please

Courts have not yet answered how responsible chatbot creators are for what their artificial companions say. US law protects the makers of social networks from legal liability for what their users post. However, a suit filed in October against the startup Character.ai, which makes a customizable, role-playing chatbot used by 20 million people, alleges the company designed an addictive product that encouraged a teenager to kill himself.



This is the hidden content, please

#Meta #killing #AIpowered #

This is the hidden content, please
#
This is the hidden content, please
#profiles #Technology

This is the hidden content, please

This is the hidden content, please

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Unfortunately, your content contains terms that we do not allow. Please edit your content to remove the highlighted words below.
Reply to this topic...

×   Pasted as rich text.   Paste as plain text instead

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

  • Vote for the server

    To vote for this server you must login.

    Jim Carrey Flirting GIF

  • Recently Browsing   0 members

    • No registered users viewing this page.

Important Information

Privacy Notice: We utilize cookies to optimize your browsing experience and analyze website traffic. By consenting, you acknowledge and agree to our Cookie Policy, ensuring your privacy preferences are respected.