
Technology – like social media, smartphones, and AI – is creating new challenges for parents. AI is alsoย opening upย opportunities to build safer online spaces for kids and support kids and parents if incentives areย alignedย and products are built with safe-by-design principles.ย ย
___________________ย
Parenting has never been easy but according to a Pew Research Center report on Parenting Children in the Age of Screens, two-thirds of U.S. parents say parenting is harder today than it was 20 years ago, with many citing technologies โ like social media, smartphones, and AI โ as reasons. The top two parental concerns are 1) kids struggling with anxiety/depression and 2) kids being bullied.ย
In our interviews with over 500 parents and kids, we heard significant painย regardingย kidsโ current use of technology due to:ย
- Excessive screen time and digital addictionย
- Online meanness and bullyingย
- Feelings of being left out and social comparisonย
- Inappropriate content (racist or sexist content, porn, etc.)ย
- Creepy strangers and safety concernsย
We support and encourage parents to take technology adoption for their kids as slowly as possibleย and alsoย deeply believe the world needs safer online spaces for kids which can be used in moderation to support genuine connections, joy,ย educationย and relaxation.ย ย
So how do we build safer online spaces for kids?ย ย
AI is giving us increasingly powerful tools to build safer online spaces for kids including more powerful age verification, moderation, parental controls, and features that support healthier, less addictive user experiences. When incentives are aligned, we believe AI can help build a safer internet for kids.ย ย
AI Age Verificationย
The first way to use AI to protect kids online isย ensuringย they are only accessing age-appropriate content and platforms.ย ย
Up until recently, there was a question about whether asking users to age verify to access mature content violated free speech. On June 27, 2025, the Supreme Court of the United States upheld the Texas law allowing age verification for adult websites (in this case a porn site) which was challenged on free speech. The Court ruled that the benefits of preventing minors from accessing sexually explicit material wereย sufficient enoughย to โonly incidentally burdensโ adultsโ access to protected speech. This was aย milestone case that sets the stage forย additionalย age verification requirements on other sites and platforms.ย
Age verification can be used to ensure parents are actuallyย parentsย and kids areย actually theย age they say they are.ย ย Todayโs standard โenter your birthdayโ age gate is ridiculously easy to circumnavigate – just change your birth year andย youโreย in. Many sites also allow access to inappropriate content without being logged in – like YouTube and social media sites – which we believe is irresponsible and dangerous.ย ย ย
There are ways to provide age verification withoutย disclosingย personalย dataย and we strongly believe there needs to be a low bar for whenย itโsย needed.ย If your social media platform (Instagram, TikTok, Snapchat,ย etc), video game platform (Roblox, Fortnight,ย etc), app or website contains information that isnโt appropriate for young children or high-risk features like chat, DMs, user-generated content, and live streams, then the platform/app/website should be required to age verify users and restrict access to underage users.ย ย
On November 18, 2025, Roblox announced it will use AI to age verification so that users can only chat with other users around their age. They are partnering with technology company Persona, who uses AI to do these age estimations. In early versions of age verification technology, users could cheat the system by using a video of another personโs face to pass the age verification step. However, as the AI age verification technology matures, this loophole is closing. This is a great step in the right directionย albeit we still believe there is more to do – parents still want more control over who theirย kidsย message. A common request among parents is the ability to only let theirย kidsย message with their actual IRL friends on Roblox which is stillย really hardย to do given username handles and clunky parental controls. There are also many games that are not age appropriate and mislabeled with ability to talk and interact with strangers onlineย –ย whichย takes me to the next way to use AI to keep kids safe onlineโฆAI Moderation.ย
AI Moderationย
One of the most visible roles for AI in child safety is content and text moderation. AI is getting increasingly good atย identifyingย nudity, self-harm indicators, hate speech, grooming patterns, and other forms of harmful content. This is true across text, video, image, real-timeย talkย and video. AI tools can analyze language patterns, detect aggression, and alert platform administrators and/or caregivers when concerning behaviorย emerges.ย ย
Even though AI is getting much better, itย isnโtย perfect – so there needs to be a human-in-the-loop and a push by technology companies to continue to improve the gaps. Most systems have flags that are then reviewed by human moderationย teamsย but this feedback loop can be slow. There is an opportunity to design systems that are more proactive with faster feedback loops.ย ย ย
There are a few keys to building child-safe AI:ย
- Testing needs to be robust for use cases that could be harmful to children including bullying, inappropriate content, mental health concerns, grooming/exploitation, and more.ย ย
- Highly activated human-in-the-loop giving feedback on flagged material. For young kids under 13, we believe parents need full transparency so they can supervise messages, chat history, and provide quick, meaningful feedback on alerts.ย ย
- Models that quickly adapt to fast feedback from trusted moderators – like the parents – to adapt to changing slang and close any loopholes quickly.ย
- Parent controls need to be robust and thoughtfully set-upย so theyย canโtย be easily circumvented.ย ย ย
- Blocks and alerts should have context explaining why something is hidden, blocked, orย age-limited.ย
AI-Powered Parental Controlsย
When a kid under 13 is properly age-verified, parentsย should have toย actively approve the platform, app and/or website and set up linked parent/kid accounts. This stillย doesnโtย happen today for most activities online used by kids including ones with high-risk features like chat, DMs, user-generated content, and live streams. We strongly believe we needย additionalย legislative support that requires these companies to age and parent verify!ย ย
AI needs to continue to improve its ability to label contentย thatโsย appropriate for various ages. In some cases, games and sites are putting their age limit too low – and exposing kids to inappropriate content. And in some cases, apps have avoided liability and the need to build COPPA-compliant experiences by just saying their app is only for kids 13 and up. This is problematic because even video streaming platforms like Netflix and Hulu, which have lots of great content for kids under 13, have defaulted toย 13+. This forces parents to lie about their kids’ ages to access that content – and then manage separate parental controls on the Netflix and Hulu side to ensure their kid is only accessingย appropriate contentย for their age.ย ย ย
In addition to AI moderation, AI can be used to provide caregivers with better visibility into digital risks and create real-time alerts to parents to make parental oversight much easier.ย ย ย
AI-Powered Healthier, Less Addictive Interfacesย
With AI, companies can build healthier, non-addictive user experiences for kids with features like educational nudges, prompts, and smart notifications.ย
These nudges provide real-time, dynamic, educational opportunities for kids. Nudges can also helpย establishย healthier experiences andย usageย including time reminders and prompts to both help kids take breaks and be more conscious about how they are feeling and do so in age-appropriate ways.ย ย
One common complaint amongst parents is how popular kid online sites and apps (YouTube, Roblox, messagingย etc) are designed to be addictive.ย ย Weโdย love to see products use new AI features to build less-addictive experiences. One good example is smart notifications. Instead of a messaging app having to choose between โMuteโ and notifications for every emoji and response, imagine smart notifications that only notify kids on time-sensitive plans.ย
Do You Trustย the Techย Company?ย
Perhaps theย most important question for parents to ask themselves is whether they trust the company who made that app, website, or platform. Are financial incentives aligned for those companies to build experiences that are good for your kids? We know of many good companies out there that make money off subscriptions or other business models whichย alignsย incentives towards moderate, healthy use. They build safe-by-design, non-addictive online experiences for kids that really take into the best interest of kids when they design,ย buildย and test new features.ย ย ย
Unfortunately, many of the most popular platformsย donโtย doย nearly wellย enough andย donโtย do enough to protect our kids online. For example, most social media sites have ad-based business models that create incentives to driveย engagementย which leads to addictive experiences.ย ย They have also been accused of not setting up the proper systems to keep kids safe online. In newly unsealed court filings in a big child-safety lawsuit, Metaโs former head of safety and well-being, Vaishnavi Jayakumar, testified that Meta had a โ17-strike policyโ for accounts engaged in sex trafficking.ย I think most parentsย agree this is abhorrent. Meta has denied this characterization, saying it removes accounts suspected of human traffickingย immediatelyย and that the plaintiffsโ brief presents a misleadingย pictureย but Meta has proven itselfย very hardย to trust.ย ย
Iโveย also been outspoken against OpenAI because they are not doing enough to protect our kids.ย They are making all of the same mistakes that social media made.ย Theyย donโtย think ahead and prioritize safety, and they continue to launch products that are harmful to kids. Their parental controls are reactive on the heels of lawsuits (launched after a family alleging OpenAI “actively helped” their son explore suicide methods), are the bare minimum, and are insanely easy to circumvent –ย ย just open up a new tab as an anonymous user and donโt use the account linked to your parents account.ย Thatโsย not good enough!ย ย AI responses should not be given on any questions that are inappropriate for young kids without proper age verification and parental controls in place.ย ย
There are big opportunities for companies to use AI to build safer online experiences for kids and win over the trust of parents. When thoughtfully designed and implemented, AI can serve as a powerful tool – one that scales to the size of todayโs digital ecosystems – supporting kids as they explore, grow, and connect safely online.ย



