Recently, a group of teen girls made the shocking discovery that boys in their New Jersey high school had rounded up images they’d posted of themselves on social media, then used those pictures to generate fake nudes.
The boys, who shared the nudes in a group chat, allegedly did this with the help of a digital tool powered by artificial intelligence, according to the Wall Street Journal.
The incident is a frightening violation of privacy. But it also illustrates just how rapidly AI is fundamentally reshaping expectations regarding what might happen to one’s online images. What this means for children and teens is particularly sobering.
A recent report published by the Internet Watch Foundation found that AI is increasingly being used to create realistic child sexual abuse material. Some of these images are generated from scratch, with the aid of AI-powered software. But a portion of this material is created with publicly available images of children, which have been scraped from the internet and manipulated using AI.
Parents who’ve seen headlines about these developments may have begun thinking twice about freely sharing their child’s image on social media, and perhaps have discouraged tweens and teens from posting fully public pictures of themselves.
SEE ALSO: A ‘predator’ can easily target teen streamers on Twitch, say researchers
But there’s little conversation about how third parties that regularly engage with and serve children and teens, like summer camps, parent-teacher associations, and sports teams, routinely use those children’s images for their online marketing and social media. They may ask for, or even indicate that they require, parents’ permission for this purpose in legal waivers. Yet some third parties may not request permission at all, particularly if children are in a public space, such as a school or sporting event.
The risk that a child’s image will be scraped from a school’s PTA Instagram account, for example, and used to create child sexual abuse material is likely very low, but it is also no longer zero.
How to protect your child’s image online
John Pizzuro, former commander of the New Jersey Internet Crimes Against Children Task Force, told Mashable that pictures of children available online just a few years ago were difficult to manipulate with the software that existed at the time.
Now, a bad actor can seamlessly digitally excise the background of an image featuring a child, then superimpose the youth onto another background with ease, according to Pizzuro, CEO of Raven, an advocacy and lobbying group focused on combating child exploitation.
Pizzuro said that every organization that takes images of children and posts them online “bears some sort of responsibility” and should have policies in place to address the threat of AI-generated child sexual abuse material.
“The more images that are out there, the more you can use programs to change things; that’s the danger,” said Pizzuro, referencing how an AI image generation tool may improve with every unique image it ingests.
One simple way for parents to eliminate the possibility that their child’s image will end up in the wrong hands is to decline when third parties ask for permission to photograph and publish pictures for marketing or other purposes. The permission is often included in registration documents when parents sign their children up for camp, sports, and extracurricular activities.
Schools typically include the form in annual registration paperwork. Though it may be hard to find in the fine print, parents should take the time to review the waiver and make an intentional decision about giving away the rights of photos of their children.
If parents can’t locate that language, they can also ask the entity about its image sharing policies and practices and make clear that they have not given permission for their child’s image to be posted online or elsewhere. If a parent has already given the third party consent, they can ask about retracting permission.
Pizzuro said that if a parent discovers a picture of their child on Instagram that was posted without consent, they can file a takedown request. The same is true for Facebook. Twitter doesn’t permit users to post images of private individuals without their consent, and parents can report those offenses. Parents can also report privacy violations involving their children on TikTok.
Who might be taking and sharing pictures of your child?
Mashable requested comment from some of the top organizations that engage with children in the U.S. — National PTA, Girl Scouts of the USA, Boy Scouts of America, American Camp Association, National Alliance for Youth Sports — about their approach to AI and children’s digital images. We received varying responses.
A spokesperson for Girl Scouts told Mashable in an email that the organization convened a “cross-functional team” earlier this year, comprising internal legal, technology, and program experts to review and monitor AI developments while encouraging the responsible use of those technologies.
Currently, any appearance in a Girl Scout-related online video or picture requires permission from each girl’s parent or guardian for every member pictured.
“We are committed to staying at the forefront of these developments to ensure the protection of our members,” wrote the spokesperson.
A spokesperson for Boy Scouts of America shared the organization’s social media guidelines, which notes that videos and images of Scouts on social media platforms should protect the privacy of individual Scouts. At the local and national level, BSA must have parental permission before posting photos of children to social media, and parents can opt out at any time. In general, BSA policy focuses on identifying a child as vaguely as possible in social media posts, like using initials instead of their full name.
If the solution to the problem of publicly shared images might appear to be closed social media accounts followed only by those with permission to do so, BSA guidelines demonstrate why that approach is more complicated than it seems.
They prohibit private social media channels so that administrators can monitor communication between Scouts and adult leaders, as well as other Scouts, to ensure there’s no inappropriate exchanges. The transparency has clear benefits, but the policy is one example of how difficult it is to balance privacy and safety concerns.
The future of children’s images online
Baroness Beeban Kidron, founder and chair of the 5Rights Foundation, a London-headquartered nonprofit that works for children’s rights online, said that parents should consider AI-manipulated or generated child sexual abuse content a current problem, not an existential threat that may come to pass in the future.
Kidron works with a covert enforcement team that investigates AI-generated child sex abuse material and noted, with distress, how quickly AI technology had advanced even in a matter of several weeks, based on images of child sexual abuse using such software that she had seen this summer and fall.