Meta faces legal scrutiny as AI advancements raise concerns over child safety

HomeCrypto News

Meta faces legal scrutiny as AI advancements raise concerns over child safety

A group of 34 United States states are filing a lawsuit against Facebook and Instagram owner Meta, accusing the company of engaging in improper manipu

A group of 34 United States states are filing a lawsuit against Facebook and Instagram owner Meta, accusing the company of engaging in improper manipulation of minors who use the platforms. This development comes amid rapid artificial intelligence (AI) advancements involving both text and generative AI.

Legal representatives from various states, including California, New York, Ohio, South Dakota, Virginia and Louisiana, allege that Meta utilizes its algorithms to foster addictive behavior and negatively impact the mental well-being of children through its in-app features, such as the “Like” button.

The chief AI scientist at Meta recently spoke out, reportedly saying that worries over the existential risks of the technology are still “premature,” and Meta has already harnessed AI to address trust and safety issues on its platforms. However, government litigants are proceeding with legal action.

Screenshot of the filing. Source: CourtListener

Attorneys for the states are seeking different amounts of damages, restitution, and compensation for each state mentioned in the document, with figures ranging from $5,000 to $25,000 per purported occurrence. Cointelegraph has reached out to Meta for more information but is yet to get feedback at the time of publication.

Meanwhile, the UK-based Internet Watch Foundation (IWF) has raised concerns about the alarming proliferation of AI-generated child sexual abuse material (CSAM). In a recent report, the IWF revealed the discovery of more than 20,254 AI-generated CSAM images within a single dark web forum in just a month, warning that this surge in disturbing content has the potential to inundate the internet.

The UK organization urged global cooperation to combat the issue of CSAM, suggesting a multifaceted strategy. This entails adjustments to existing laws, enhancements in law enforcement education, and the implementation of regulatory supervision for AI models.

Related: Researchers in China developed a hallucination correction engine for AI models

In the context of AI developers, the IWF advises the prohibition of their AI for generating child abuse content, the exclusion of associated models, and a focus on removing such material from their models.

The advancement of generative AI image generators has significantly improved the creation of lifelike human replicas. Platforms such as Midjourney, Runway, Stable Diffusion, and OpenAI’s Dall-E are examples of tools capable of generating realistic images.

Magazine: ‘AI has killed the industry’: EasyTranslate boss on adapting to change