Associate Professor of English, University of Michigan-Flint. I research and teach rhetoric and writing.
5092 stories
·
43 followers

No One Knows How to Deal With 'Student-on-Student' AI CSAM

1 Comment and 2 Shares

Schools, parents, police, and existing laws are not prepared to deal with the growing problem of students and minors using generative AI tools to create child sexual abuse material of their peers, according to a new report from researchers at Stanford Cyber Policy Center.

The report, which is based on public records and interviews with NGOs, internet platforms staff, law enforcement, government employees, legislators, victims, parents, and groups that offer online training to schools, found that despite the harm that nonconsensual content causes, the practice has been normalized by mainstream online platforms and certain online communities.

“Respondents told us there is a sense of normalization or legitimacy among those who create and share AI CSAM,” the report said. “This perception is fueled by open discussions in clear web forums, a sense of community through the sharing of tips, the accessibility of nudify apps, and the presence of community members in countries where AI CSAM is legal.”

The report says that while children may recognize that AI-generating nonconsensual content is wrong they can assume “it’s legal, believing that if it were truly illegal, there wouldn’t be an app for it.” The report, which cites several 404 Media stories about this issue, notes that this normalization is in part a result of many “nudify” apps being available on the Google and Apple app stores, and that their ability to AI-generate nonconsensual nudity is openly advertised to students on Google and social media platforms like Instagram and TikTok. One NGO employee told the authors of the report that “there are hundreds of nudify apps” that lack basic built-in safety features to prevent the creation of CSAM, and that even as an expert in the field he regularly encounters AI tools he’s never heard of, but that on certain social media platforms “everyone is talking about them.”

The report notes that while 38 U.S. states now have laws about AI CSAM and the newly signed federal Take It Down Act will further penalize AI CSAM, states “failed to anticipate that student-on-student cases would be a common fact pattern. As a result, that wave of legislation did not account for child offenders. Only now are legislators beginning to respond, with measures such as bills defining student-on-student use of nudify apps as a form of cyberbullying.”

One law enforcement officer told the researchers how accessible these apps are. “You can download an app in one minute, take a picture in 30 seconds, and that child will be impacted for the rest of their life,” they said.

One student victim interviewed for the report said that she struggled to believe that someone actually AI-generated nude images of her when she first learned about them. She knew other students used AI for writing papers, but was not aware people could use AI to create nude images. “People will start rumors about anything for no reason,” she said. “It took a few days to believe that this actually happened.”

Another victim and her mother interviewed for the report described the shock of seeing the images for the first time. “Remember Photoshop?” the mother asked, “I thought it would be like that. But it’s not. It looks just like her. You could see that someone might believe that was really her naked.”

One victim, whose original photo was taken from a non-social media site, said that someone took it and “ruined it by making it creepy [...] he turned it into a curvy boob monster, you feel so out of control.”

In an email from a victim to school staff, one victim said “I was unable to concentrate or feel safe at school. I felt very vulnerable and deeply troubled. The investigation, media coverage, meetings with administrators, no-contact order [against the perpetrator], and the gossip swirl distracted me from school and class work. This is a terrible way to start high school.”

One mother of a victim the researchers interviewed for the report feared that the images could crop up in the future, potentially affecting her daughter’s college applications, job opportunities, or relationships. “She also expressed a loss of trust in teachers, worrying that they might be unwilling to write a positive college recommendation letter for her daughter due to how events unfolded after the images were revealed,” the report said.

In 2024, Jason and I wrote a story about how one school in Washington state struggled to deal with its students using a nudify app on other students. The story showed how teachers and school administration weren’t familiar with the technology, and initially failed to report the incident to the police even though it legally qualified as “sexual abuse” and school administrators are “mandatory reporters.” 

According to the Stanford report, many teachers lack training on how to respond to a nudify incident at their school. A Center for Democracy and Technology report found that 62% of teachers say their school has not provided guidance on policies for handling incidents

involving authentic or AI nonconsensual intimate imagery. A 2024 survey of teachers and principals found that 56 percent did not get any training on “AI deepfakes.” One provider told the authors of the report that while many schools have crisis management plans for “active shooter situations, they had never heard of a school having a crisis management plan for a nudify incident, or even for a real nude image of a student being circulated.”

The report makes several recommendations to schools, like providing victims with third-party counseling services and academic accommodations, drafting language to communicate with the school community when an incident occurs, ensuring that students are not discouraged or punished for reporting incidents, and contacting the school’s legal counsel to assess the school’s legal obligations, including its responsibility as a “mandatory reporter.” 

The authors also emphasized the importance of anonymous tip lines that allow students to report incidents safely. It cites two incidents that were initially discovered this way, one in Pennsylvania where a students used the state’s Safe2Say Something tipline to report that students were AI-generating nude images of their peers, and another school in Washington that first learned about a nudify incident through a submission to the school’s harassment, intimidation, and bullying online tipline. 

One provider of training to schools emphasized the importance of such reporting tools, saying, “Anonymous reporting tools are one of the most important things we can have in our school systems,” because many students lack a trusted adult they can turn to.

Notably, the report does not take a position on whether schools should educate students about nudify apps because “there are legitimate concerns that this instruction could inadvertently educate students about the existence of these apps.”



Read the whole story
acdha
1 day ago
reply
“I was unable to concentrate or feel safe at school. I felt very vulnerable and deeply troubled. The investigation, media coverage, meetings with administrators, no-contact order [against the perpetrator], and the gossip swirl distracted me from school and class work. This is a terrible way to start high school.”
Washington, DC
betajames
1 day ago
reply
Michigan
Share this story
Delete

The MAHA Report Cites Studies That Don’t Exist

2 Shares

Health Secretary Robert F. Kennedy Jr. says his “Make America Healthy Again” Commission report harnesses “gold-standard” science, citing more than 500 studies and other sources to back up its claims. Those citations, though, are rife with errors, from broken links to misstated conclusions.

Seven of the cited sources don’t appear to exist at all.

Epidemiologist Katherine Keyes is listed in the MAHA report as the first author of a study on anxiety in adolescents. When NOTUS reached out to her this week, she was surprised to hear of the citation. She does study mental health and substance use, she said. But she didn’t write the paper listed.

“The paper cited is not a real paper that I or my colleagues were involved with,” Keyes told NOTUS via email. “We’ve certainly done research on this topic, but did not publish a paper in JAMA Pediatrics on this topic with that co-author group, or with that title.”

It’s not clear that anyone wrote the study cited in the MAHA report. The citation refers to a study titled, “Changes in mental health and substance abuse among US adolescents during the COVID-19 pandemic,” along with a nonfunctional link to the study’s digital object identifier. While the citation claims that the study appeared in the 12th issue of the 176th edition of the journal JAMA Pediatrics, that issue didn’t include a study with that title.

As the Trump administration cuts research funding for federal health agencies and academic institutions and rejects the scientific consensus on issues like vaccines and gender-affirming care, the issues with its much-heralded MAHA report could indicate lessening concern for scientific accuracy at the highest levels of the federal government.

The Department of Health and Human Services did not respond to a request for comment on the report’s citation inconsistencies.

The anxiety study wasn’t the only one the report cites that appears to be mysteriously absent from the scientific literature. A section describing the “corporate capture of media” highlights two studies that it says are “broadly illustrative” of how a rise in direct-to-consumer drug advertisements has led to more prescriptions being written for ADHD medications and antidepressants for kids.

The catch? Neither of those studies is anywhere to be found. Here are the two citations:

Shah, M. B., et al. (2008). Direct-to-consumer advertising and the rise in ADHD medication use among children. Pediatrics, 122(5), e1055- e1060.

Findling, R. L., et al. (2009). Direct-to-consumer advertising of psychotropic medications for youth: A growing concern. Journal of Child and Adolescent Psychopharmacology, 19(5), 487-492.

Those articles don’t appear in the table of contents for the journals listed in their citations. A spokesperson for Virginia Commonwealth University, where psychiatric researcher Robert L. Findling currently teaches, confirmed to NOTUS that he never authored such an article. The author of the first study doesn’t appear to be a real ADHD researcher at all — at least, not one with a Google Scholar profile.

In another section titled, “American Children are on Too Much Medicine – A Recent and Emerging Crisis,” the report claims that 25% to 40% of mild cases of asthma are overprescribed. But searching Google for the exact title of the paper it cites to back up that figure — “Overprescribing of oral corticosteroids for children with asthma” — leads to only one result: the MAHA report.

The corticosteroids study’s supposed first author, pediatric pulmonologist Harold J. Farber, denied writing it or ever working with the other listed authors. He pointed to similar research he’s conducted, but said that even if the MAHA report cited that study correctly, its conclusions are “clearly an overgeneralization” of the findings.

“It is a tremendous leap of faith to generalize from a study in one Medicaid managed care program in Texas using 2011 to 2015 data to national care patterns in 2025,” Farber said in an email.

Spread across the footnotes of the 73-page document, those missing papers are listed alongside dozens of citations with more mundane errors like broken links, missing or incorrect authors and wrong issue numbers.

NOTUS also found serious issues with how the report interpreted some of the existing studies it cites.

In one section about mental health medication, which Kennedy has railed against for years, the report cites a review paper it claims shows that therapy alone is as or more effective than psychiatric medicine. But one of that paper’s statisticians told NOTUS that conclusion doesn’t make sense, given their study didn’t even attempt to measure or compare therapy’s effectiveness as a mental health treatment.

“We did not include psychotherapy in our review. We only compared the effectiveness of (new generation) antidepressants against each other, and against placebo,” Joanne McKenzie, a biostatistics professor at an Australian university, said in an email.

Another paper, which the report says shows “antipsychotic prescriptions for children increased by 800% between 1993 and 2009,” actually found an eight-fold increase from 1995 to 2005.

Another medical researcher whose work was cited in a section about how screen time affects children’s sleep told us the MAHA report mischaracterized her study.

“The conclusions in the report are not accurate and the journal reference is incorrect. It was not published in Pediatrics. Also, the study was not done in children, but in college students,” Mariana G. Figueiro emailed NOTUS. She added that she even had more relevant research the authors could have used: “I was not aware of the choice, or else I would have suggested one of the other ones.”

The Trump administration commissioned the report to investigate the root causes of chronic illnesses in the U.S. From exposure to pesticides and microplastics to cell phone radiation and food coloring, the report points to many potential culprits — and claims it offers “a clear, evidence-based foundation for the policy interventions, institutional reforms, and societal shifts needed to reverse course.” The commission is also scheduled to release a “Make Our Children Healthy Again Strategy” report in August.

Kennedy has enthusiastically promoted the report, calling it a “milestone” in a post on X after its release.

“Never in American history has the federal government taken a position on public health like this,” Kennedy wrote.


Emily Kennard and Margaret Manto are NOTUS reporters and Allbritton Journalism Institute fellows.

Have tips? You can reach Margaret on Signal at margaretmanto.61 and Emily at emilykennard.24.

Read the whole story
betajames
2 days ago
reply
Michigan
acdha
2 days ago
reply
Washington, DC
Share this story
Delete

The Who Cares Era

5 Shares

Earlier this week, it was discovered that the Chicago Sun-Times and the Philadelphia Inquirer had both published an externally-produced "special supplement" that contained facts, experts, and book titles entirely made up by an AI chatbot. There's been a lot written about this (former Chicago Reader editor Martha Bayne's is the best), and I don't need to rehash it all. But the thing that is most disheartening to me is how at every step along the way, nobody cared.

The writer didn't care. The supplement's editors didn't care. The biz people on both sides of the sale of the supplement didn't care. The production people didn't care. And, the fact that it took two days for anyone to discover this epic fuckup in print means that, ultimately, the reader didn't care either.

It's so emblematic of the moment we're in, the Who Cares Era, where completely disposable things are shoddily produced for people to mostly ignore.

AI is, of course, at the center of this moment. It's a mediocrity machine by default, attempting to bend everything it touches toward a mathematical average. Using extraordinary amounts of resources, it has the ability to create something good enough, a squint-and-it-looks-right simulacrum of normality. If you don't care, it's miraculous. If you do, the illusion falls apart pretty quickly. The fact that the userbase for AI chatbots has exploded exponentially demonstrates that good enough is, in fact, good enough for most people. Because most people don't care.

(It's worth pointing out that I'm not a full-throated hater and know people—coders, mostly—who work with AI that do care and have used it to make real, meaningful things. Most people, however, use it quickly and thoughtlessly to make more mediocrity.)

It's easy to blame this all on AI, but it's not just that. Last year I was deep in negotiations with a big-budget podcast production company. We started talking about making a deeply reported, limited-run show about the concept of living in a multiverse that I was (and still am) very excited about. But over time, our discussion kept getting dumbed down and dumbed down until finally the show wasn't about the multiverse at all but instead had transformed into a daily chat show about the Internet, which everyone was trying to make back then. Discussions fell apart.

Looking back, it feels like a little microcosm of everything right now: Over the course of two months, we went from something smart that would demand a listener's attention in a way that was challenging and new to something that sounded like every other thing: some dude talking to some other dude about apps that some third dude would half-listen-to at 2x speed while texting a fourth dude about plans for later.

Hanif Abdurraqib, in one of his excellent Instagram mini-essays the other week, wrote about the rise of content that's designed to be consumed while doing something else. In Hanif's case, he was writing about Time Machine, his incredible 90 minute deep dive into The Fugees' seminal album The Score. Released in 2021, Hanif marveled at the budget, time, and effort that went into crafting the two-part 90 minute podcast and how, today, there's no way it would have happened.

He's right. Nobody's funding that kind of work right now, because nobody cares.

(It's worth pointing out that Hanif wrote this using Stories, a system that erased it 24 hours later. Another victim of the Who Cares Era.)

Of course we're all victims of the biggest perpetrators of this uncaring era, as the Trump administration declares "Who Cares?" to vast swaths of the federal government, to public health, to immigrant families, to college students, to you, to me. As Elon Musk's DOGE rats gnaw their way through federal agencies, not caring is their guiding light. They cut indiscriminately, a smug grin on their faces. That they believe they can replace government workers—people who care an extraordinary amount about their arcane corner of the bureaucracy—with hastily-written AI code is another defining characteristic of right now.

I keep coming back to the word "disheartening," because it all really is.

Without getting into too many specifics, I recently was involved in reviewing hundreds of applications for something. Over the course of reviewing, I was struck by the nearly-identical phrasing that threaded through dozens of the applications. It was eerie at first, like seeing a shadow in the distance, then frustrating, and ultimately completely disheartening: It was AI. For whatever their reasons, a bunch of people had used a chatbot to help write their answers to questions that asked them to draw from their own, unique, personal experience. They had fed their resumes or their personal websites or their actual stories and experiences into the machine, and it had filled in the blanks, Mad Libs-style. I felt crushed.

Until.

Until I read an application written entirely by a person. And then another. And another. They glowed with delight and joy and sadness and with the unexpected at every turn.

They were human.

They were written by people that cared.

In the Who Cares Era, the most radical thing you can do is care.

In a moment where machines churn out mediocrity, make something yourself. Make it imperfect. Make it rough. Just make it.

At a time where the government's uncaring boot is pressing down on all of our necks, the best way to fight back is to care. Care loudly. Tell others. Get going.

As the culture of the Who Cares Era grinds towards the lowest common denominator, support those that are making real things. Listen to something with your full attention. Watch something with your phone in the other room. Read an actual paper magazine or a book.

Be yourself.

Be imperfect.

Be human.

Care.

Read the whole story
betajames
4 days ago
reply
Michigan
acdha
8 days ago
reply
Washington, DC
Share this story
Delete

Authors Are Accidentally Leaving AI Prompts In their Novels https://www.404media.co/authors-are-accidentally-leaving-ai-prompts-in-their-novels/

1 Comment and 4 Shares

Advertisement

‘I’ve rewritten the passage to align more with the J. Bree style’ appeared in the middle of a tense scene with a scaled dragon prince.

Read the whole story
acdha
8 days ago
reply
“It appeared as if author, Lena McDonald, had used an AI to help write the book, asked it to imitate the style of another author, and left behind evidence they’d done so in the final work.”
Washington, DC
betajames
8 days ago
reply
Michigan
Share this story
Delete

MAGA’s assault on science is an act of grievous self-harm

1 Comment and 2 Shares
Read the whole story
acdha
8 days ago
reply
“The consequences will be bad for the world, but America will pay the biggest price of all.”
Washington, DC
betajames
8 days ago
reply
Michigan
Share this story
Delete

Why Even Try if You Have A.I.? Now that machines can think for us, we have to choose whether to be the passengers or pilots of our lives.

1 Share

We’re drawn to activities that invite us to grow, by trying and trying again, because we want to evolve as people. Life is mostly repetitive—wake, eat, work, sleep, repeat—and each day can feel like an unsatisfying circle. But repetition with variation broadens us. It makes our circular days into spiralling journeys. “The spiral is a…

The post Why Even Try if You Have A.I.? Now that machines can think for us, we have to choose whether to be the passengers or pilots of our lives. appeared first on Jerz's Literacy Weblog (est. 1999).



Read the whole story
betajames
11 days ago
reply
Michigan
Share this story
Delete
Next Page of Stories