FTC: Facebook misled parents, failed to guard kids’ privacy
U.S. regulators say Facebook misled parents and failed to protect the privacy of children using its Messenger Kids app, including misrepresenting the access it provided to app developers to private user data.
As a result, the Federal Trade Commission on Wednesday proposed sweeping changes to a 2020 privacy order with Facebook — now called Meta — that would prohibit it from profiting from data it collects on users under 18. This would include data collected through its virtual-reality products. The FTC said the company has failed to fully comply with the 2020 order.
Meta would also be subject to other limitations, including its use of face-recognition technology, and be required to provide additional privacy protections for its users.
“Facebook has repeatedly violated its privacy promises,†said Samuel Levine, director of the FTC’s Bureau of Consumer Protection. “The company’s recklessness has put young users at risk, and Facebook needs to answer for its failures.â€
Meta called the announcement a “political stunt.â€
A judge has tentatively approved a $725-million payout by Facebook’s owner to settle claims that the social network shared user data without consent.
“Despite three years of continual engagement with the FTC around our agreement, they provided no opportunity to discuss this new, totally unprecedented theory. Let’s be clear about what the FTC is trying to do: usurp the authority of Congress to set industry-wide standards and instead single out one American company while allowing Chinese companies, like TikTok, to operate without constraint on American soil,†Meta said in a prepared statement.
The Menlo Park, Calif.-based company added that it will “vigorously fight†the FTC’s action and expects to prevail.
Facebook launched Messenger Kids in 2017, pitching it as a way for children to chat with family members and friends approved by their parents. The app doesn’t give kids separate Facebook or Messenger accounts. Rather, it works as an extension of a parent’s account, and parents get controls, such as the ability to decide with whom their kids can chat.
At the time, Facebook said Messenger Kids wouldn’t show ads or collect data for marketing, though it would collect some data it said was necessary to run the service.
But child-development experts raised immediate concerns.
A pair of bills approved by California legislators would regulate how Facebook, YouTube and other platforms treat minors and seek to influence how they moderate extremist content.
In early 2018, a group of 100 experts, advocates and parenting organizations contested Facebook’s claims that the app was filling a need kids had for a messaging service. The group included nonprofits, psychiatrists, pediatricians, educators and the children’s music singer Raffi Cavoukian.
“Messenger Kids is not responding to a need — it is creating one,†the letter said. “It appeals primarily to children who otherwise would not have their own social media accounts.†Another passage criticized Facebook for “targeting younger children with a new product.â€
Facebook, in response to the letter, said at the time that the app “helps parents and children to chat in a safer way†and emphasized that parents are “always in control†of their kids’ activity.
The FTC now says this has not been the case. The 2020 privacy order, which required Facebook to pay a $5-billion fine, required an independent assessor to evaluate the company’s privacy practices. The FTC said the assessor “identified several gaps and weaknesses in Facebook’s privacy program.â€
The FTC also said Facebook, from late 2017 until 2019, “misrepresented that parents could control whom their children communicated with through its Messenger Kids product.â€
“Despite the company’s promises that children using Messenger Kids would only be able to communicate with contacts approved by their parents, children in certain circumstances were able to communicate with unapproved contacts in group text chats and group video calls,†the FTC said.
Meta critics applauded the FTC’s action. Jeffrey Chester, the executive director of the nonprofit Center for Digital Democracy, called it a “a long-overdue intervention into what has become a huge national crisis for young people.â€
Meta, and with its platforms like Instagram and Facebook, Chester added, “are at the center of a powerful commercialized social media system that has spiraled out of control, threatening the mental health and well-being of children and adolescents.â€
The company, he added, has not done enough to address existing problems — and is now unleashing “even more powerful data gathering and targeting tactics fueled by immersive content, virtual reality and artificial intelligence, while pushing youth further into the metaverse with no meaningful safeguards.â€
As part of the proposed changes to the FTC’s 2020 order (which was announced in 2019 and finalized later), Meta would also be required to pause launching new products and services without “written confirmation from the assessor that its privacy program is in full compliance†with the order.
Meta has 30 days to respond to the FTC’s latest action.
More to Read
Inside the business of entertainment
The Wide Shot brings you news, analysis and insights on everything from streaming wars to production — and what it all means for the future.
You may occasionally receive promotional content from the Los Angeles Times.