🔥 Land A High Paying Web3 Job In 90 Days LEARN MORE

Nevada Sees a Surge in Explicit AI Photos of Students, But “Little Could Be Done”

515852
Nevada Sees a Surge in Explicit AI Photos of Students, But “Little Could Be Done”Nevada Sees a Surge in Explicit AI Photos of Students, But “Little Could Be Done”

In this post:

  • Nevada High School has involved Story County Sheriff to investigate the rise of explicit AI pictures of students.
  • Investigators said little could be done as the law was clear on the matter.
  • Lawmakers and AI companies are beginning to take steps to stop the misuse of AI in causing child abuse.

Nevada High School is grappling with the rampant spread of explicit photos of students on social media created using AI by other students.

Parents of Victims Seem Helpless

The Story County Sheriff’s Office in Nevada is currently investigating the situation. However, the investigators said there is little they can do because the law of the state isn’t yet clear on matters relating to the creation and spread of AI-generated explicit photos, KCCI reported

“Our daughters are being victimized. and it doesn’t seem like there’s anything we can do about it,” said Chad Borwick, a parent to one of the affected students. 

“You don’t want things to happen to your kids. You want them to have better,” said Ida Nady, another parent to a victim. “They’re going to feel like they don’t matter like their voice doesn’t matter.”

The situation presents itself as a clear example of how AI can be used to hurt people. While good at creating texts and images, AI could be misused, which necessitates the need to accelerate safety measures by AI companies, and especially laws that punish offenders.

In Iowa, at least, a bill sponsored by Sen. Chris Cournoyer (R-LeClaire) is expected to become effective by July 1. The law will make it a felony to create and circulate full or partial nude photos of minors in the states, which covers Nevada.

See also  Coinbase defends why it is not supporting Celo's migration into an Ethereum L2 network

Google, Meta, OpenAI, and Others Agree to Review Their Models Against Child Abuse

Big tech companies have also started a movement to curb the rise in child abuse images made with AI. Google, Meta, OpenAI, Microsoft, and Amazon, among others, agreed to review their models against the creation and spread of AI-generated child sexual abuse material and other sexual harms against children. 

The move comes at a time when the rate of child sex abuse online generally is at an all-time high. In fact, 2023 was the “most extreme year on record,” according to the Internet Watch Foundation.

The UK-based organization said the thousands of images and videos of three to six-year-old children are now being found on the open internet.

Land a High-Paying Web3 Job in 90 Days: The Ultimate Roadmap

Share link:

Disclaimer. The information provided is not trading advice. Cryptopolitan.com holds no liability for any investments made based on the information provided on this page. We strongly recommend independent research and/or consultation with a qualified professional before making any investment decisions.

Most read

Loading Most Read articles...

Stay on top of crypto news, get daily updates in your inbox

Editor's choice

Loading Editor's Choice articles...
Cryptopolitan
Subscribe to CryptoPolitan