At a Glance
- Former child actor Mara Wilson revealed AI tools turned her childhood photos into CSAM
- Wilson found her image on fetish sites before high school, calling it a “living nightmare”
- She urges boycotts and legislation to stop companies enabling exploitative AI images
- Why it matters: Any child’s photo posted online can now be weaponized in seconds
Former Matilda star Mara Wilson is warning that generative AI has made child sexual exploitation terrifyingly simple, revealing that her own childhood image was manipulated into abuse material decades before today’s technology existed.
Writing in The Guardian on January 17, Wilson, 38, said the danger has escalated from isolated Photoshop creeps to mass-produced AI horrors that threaten every child online.
Wilson’s Childhood Ordeal
Wilson began acting at five. On set, she felt protected. Off set, the internet became her predator.
- Fan sites cropped up fast
- Grown men mailed unsettling letters
- Her face appeared on fetish forums
- Doctored nude images spread before she reached high school
“I wasn’t a beautiful girl-my awkward age lasted from about age 10 to about 25-and I acted almost exclusively in family-friendly movies,” she wrote. “But I was a public figure, so I was accessible.”

Accessibility, she stressed, is what predators seek. The early internet handed it to them freely.
The AI Amplifier
Wilson fears today’s generative AI far outpaces the crude Photoshop of her youth.
Key differences:
| Then | Now |
|---|---|
| Manual editing, one image at a time | Algorithms create hundreds in minutes |
| Fakes often looked fake | High-resolution output can fool anyone |
| Needed celebrity photos | Any social-media picture works |
“It is now infinitely easier for any child whose face has been posted on the internet to be sexually exploited,” Wilson wrote. “Millions of children could be forced to live my same nightmare.”
Fighting Back
Wilson ended her essay with a call for collective action. She wants consumers to:
- Boycott platforms that allow exploitative AI images
- Demand tech firms build in safety blocks
- Push lawmakers for strict penalties on CSAM creators
- Talk candidly with kids about online photo risks
“We need to be the ones demanding companies that allow the creation of CSAM be held accountable,” she argued, adding that parents must weigh the danger before sharing snapshots of their children.
Wilson, now a writer and mental-health activist, believes the same public pressure that forced seat-belt laws and food-safety rules can rein in unchecked AI. The stakes, she says, are nothing less than the safety of every child growing up online today.

