The phone call came on a Tuesday morning. A parent reported that her daughter, a ninth grader, had stopped eating and refused to go to school. The reason? A classmate had used an AI (artificial intelligence) app to create and share a fake nude image of her. Within hours, it had spread across multiple group chats.
Across the country, schools are grappling with a new form of abuse that existing policies weren’t designed to address: AI-generated deepfakes targeting students.
What are Deepfakes? Understanding the Scope of the Problem
Deepfakes are videos, images, and audio created using AI to realistically simulate human likeness to portray people saying or doing things that they actually haven’t. It is becoming increasingly more accessible to pull this off without heavy technical skills. Today, freely available apps allow anyone with a smartphone to create convincing fake images and videos in just a few steps and with only a selfie and quick voice sample.
Once generated, these images are circulated across social media platforms and text messages without any indication they’ve been digitally fabricated. They are viewed and shared with the belief that they’re authentic, which allows the deception to spread rapidly. The resulting impact of AI and deepfakes on social media is profound: students face serious emotional, social, and potentially legal risks that go well beyond typical “online drama.”
Deepfake Nudity Apps
“Undressing” or “nudify” apps can generate sexually explicit images from ordinary clothed photos—and this technology has infiltrated schools themselves. Students are creating and sharing fabricated intimate images of classmates, sometimes dismissing it as “just for a laugh,” but often deploying these images as weapons of revenge, bullying, or humiliation.
The psychological toll is devastating. Victims experience profound shame, anxiety, and depression. Many transfer schools or begin skipping classes because they cannot face classmates who may have seen the images. The feeling of powerlessness—knowing these images exist, circulate, and multiply beyond their control—compounds the trauma. In the most tragic cases, they struggle with suicidal thoughts or even attempt suicide.
Perhaps most troubling, this technology completely undermines traditional digital safety guidance. For years, advice emphasized “never send intimate photos.” Now, a student who has never taken or shared a sexual image can still become a victim. The violation happens without their participation or consent, and no amount of caution or good judgment can protect against someone else’s malicious use of technology.
Fabricated Criminal Evidence/Hoaxes
Deepfakes depicting students using drugs, carrying weapons, getting arrested, committing assaults, or vandalizing property can inflict lasting damage that persists long after the content is debunked. In school settings, the stakes escalate dramatically—imagine a deepfake showing a student planting an explosive device on campus. Such fabrications don’t just damage individual reputations; they can trigger lockdowns, traumatize entire school communities, and launch criminal investigations.
Schools operating under zero-tolerance policies sometimes immediately suspend students or launch disciplinary proceedings while authenticity is still being investigated, disrupting education and leaving permanent marks on academic records. These fabricated videos embed themselves in digital footprints, and even with clear disclaimers about falsification, college admissions officers, and prospective employers often adopt a “better safe than sorry” approach, closing doors to opportunities before victims can defend themselves.
Perhaps most insidiously, the reputational harm lingers in ways that are difficult to quantify or reverse. Teachers and staff who viewed a deepfake of a student vandalizing property may unconsciously harbor lingering suspicion. College recommenders might recall “something concerning” without remembering it was fabricated. These implicit biases can shadow students for years, affecting evaluations, recommendations, and future prospects—all stemming from actions they never committed.
Why Schools Must Act Now
Schools must act urgently because students are weaponizing AI-generated images against each other, and the misconception that “they’re not real” or “it’s just a joke” leads to these harms being dangerously minimized. In reality, the damages are everlasting—for individual students whose lives and futures are derailed, for schools facing legal liability and community trust erosion, and for entire communities grappling with the normalization of digital abuse among young people.
A Framework for School Response
Addressing deepfake abuse requires treating it as a predictable safeguarding issue, not a one-off scenario. Schools that respond effectively focus on these key areas:
- Update policies with clear language and consequences. Explicitly ban creating, possessing, or sharing AI-generated sexual images and fabricated criminal evidence in your student code of conduct. Spell out consequences so students understand this constitutes serious misconduct with real repercussions.
- Train staff and create response protocols. Every staff member should understand what deepfake abuse is, why it matters, and how to recognize warning signs like sudden social withdrawal, anxiety, or rumors about “edited pictures.” Develop a response playbook: who receives reports, how to support the targeted student, and when to contact caregivers. Always prioritize victim-centered practice: stop further sharing immediately, avoid victim-blaming questions, limit how many adults view the images, and connect students with mental health support.
- Make reporting safe and accessible. Offer multiple reporting pathways beyond the principal’s office: trusted adults, anonymous reporting tools, digital forms. Teach bystander responsibility—students need to know that forwarding or even viewing these images contributes to harm. Be transparent about what happens after a report so both victims and witnesses believe that coming forward will help rather than create more problems.
- Implement technical safeguards where possible. Use content filters on school networks to block known deepfake creation sites, while acknowledging that students will still have access through personal devices. Digital Threat Detection is designed to complement these efforts—our system triggers on images of guns regardless of whether they’re AI-generated or real, providing schools with an additional layer of protection against fabricated and genuine threats.
Moving Forward
The technology behind deepfakes will only become more sophisticated and accessible. While the risk cannot be eliminated entirely, safer school environments can be created where students understand the serious harm these tools cause, where reporting is normalized, and where compassionate responses are the rule rather than the exception.
This starts with recognizing that deepfake abuse is not a future problem to prepare for; it’s happening in schools right now. Students are already creating fabricated nude images of classmates. Hoax deepfakes are already wasting emergency resources. Fabricated criminal evidence is destroying reputations. The question is not whether your school will be affected, but whether you’ll be ready to respond.
The time to act is now, before another student’s life is changed forever by a fabrication that took seconds to create but years to overcome.
Interested in learning more about digital dangers?
Watch our recent webinar with national experts Dr. Dewey Cornell and Theresa Campbell as they explore the changing landscape of digital threats in schools and highlight the growing importance of behavioral threat assessments.
You’ll gain valuable insights into emerging trends, the role of social media platforms and online environments in shaping student behavior, and practical steps for strengthening your school’s ability to address these risks.




