[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"site-settings":3,"perspective-series":30,"perspective-related":98,"perspective-why-self-regulation-fails":119},{"siteTitle":4,"siteTagline":5,"siteDescription":6,"organizationName":4,"personName":7,"defaultOgImageUrl":8,"socialHandles":9,"navLinks":10,"footerText":26,"cursorMode":27,"interactionMode":28,"missionPrompt":29},"Mom Player Character","For parents in the digital deep end","Recon, parent guides, and perspective on the digital world your kid already lives in.","Shannon","\u002Fog-default.svg",{},[11,14,17,20,23],{"label":12,"href":13},"Recon","\u002Frecon",{"label":15,"href":16},"Guides","\u002Fguides",{"label":18,"href":19},"Perspective","\u002Fperspective",{"label":21,"href":22},"Live","\u002Flive",{"label":24,"href":25},"About","\u002Fabout","Actively investigating the internet your kid already lives in",true,"tasteful","Pick the parent problem, get the clearest next move.",[31],{"_id":32,"_type":33,"kindLabel":34,"title":35,"slug":36,"path":37,"excerpt":38,"status":39,"featured":27,"publishedAt":40,"featuredImageUrl":40,"featuredImageAlt":41,"pieceCount":42,"pieces":43},"43ee44c6-4b34-4403-94ad-c3b5fdeb8432","series","Series","Regulating AI & Social Media for Real-World Safety","regulating-ai-social-media-real-world-safety","\u002Fperspective\u002Fseries\u002Fregulating-ai-social-media-real-world-safety","A practical look at what it would actually take to regulate AI and social platforms for kids’ real-world safety—drawing on lessons from other industries and naming the tradeoffs.","active",null,"Conceptual image of AI and social media symbols surrounded by caution markers, suggesting regulation and child safety.",7,[44,52,59,66,75,83,91],{"key":45,"type":46,"kindLabel":18,"label":47,"title":48,"path":49,"excerpt":50,"publishedAt":40,"featuredImageUrl":40,"featuredImageAlt":51},"369592fb78a2","perspective","Part 1","Why Self-Regulation Fails (And Always Has)","\u002Fperspective\u002Fwhy-self-regulation-fails","Companies don't change because they want to — they change when the cost of not changing outweighs the cost of changing. The case against hoping tech fixes itself.","Conceptual image representing technology self-regulation, with corporate skyscrapers overshadowing small regulatory road signs.",{"key":53,"type":46,"kindLabel":18,"label":54,"title":55,"path":56,"excerpt":57,"publishedAt":40,"featuredImageUrl":40,"featuredImageAlt":58},"06a047ff76d5","Part 2","What GDPR & Canadian Anti-Spam Laws Actually Teach Us","\u002Fperspective\u002Fwhat-gdpr-and-casl-teach-us","The 'regulation won't work because the internet is global' argument sounds convincing until you look at what GDPR and CASL actually did.","Stylized map highlighting Europe and Canada connected by data lines, representing the global impact of GDPR and Canadian anti-spam laws.",{"key":60,"type":46,"kindLabel":18,"label":61,"title":62,"path":63,"excerpt":64,"publishedAt":40,"featuredImageUrl":40,"featuredImageAlt":65},"e3389b135226","Part 3","The Only Thing That Works: Hitting the Bottom Line","\u002Fperspective\u002Fthe-only-thing-that-works-hitting-the-bottom-line","If you want companies like Meta or OpenAI to change their behavior around kids, the conversation has to reach revenue. Everything else is negotiable.","Conceptual image of a tech company headquarters overlaid with a red downward revenue arrow.",{"key":67,"type":68,"kindLabel":69,"label":70,"title":71,"path":72,"excerpt":73,"publishedAt":40,"featuredImageUrl":40,"featuredImageAlt":74},"538b6afba961","parentGuide","Parent Guide","Part 4","Age-Gating That Actually Works (Not Just a Checkbox)","\u002Fguides\u002Fage-gating-that-actually-works","Typing in a fake birthday isn't age verification — it's just a question. Here’s what real, working age-gating would need to look like, and what parents can do right now.","Illustration of a sign-up screen with an age checkbox, overlaid with icons for ID cards, credit cards, and device settings.",{"key":76,"type":46,"kindLabel":18,"label":77,"title":78,"path":79,"excerpt":80,"publishedAt":81,"featuredImageUrl":40,"featuredImageAlt":82},"483d15c14a8b","Part 5","Should Free Access Exist for Kids?","\u002Fperspective\u002Fshould-free-access-exist-for-kids","Free access removes every barrier — including the ones that are supposed to keep younger users out. That's not accidental. It's part of how these platforms grow.","2026-05-07T17:59:00.000Z","Conceptual image of layered digital gates in front of a social platform interface, representing conditional free access for kids.",{"key":84,"type":46,"kindLabel":18,"label":85,"title":86,"path":87,"excerpt":88,"publishedAt":89,"featuredImageUrl":40,"featuredImageAlt":90},"267591a6959f","Part 6","AI and Privacy vs Safety — Where Should the Line Be?","\u002Fperspective\u002Fprivacy-vs-safety-where-should-the-line-be","Privacy isn't absolute in any other context. The question isn't whether to protect it — it's where the threshold should be when behavior starts signaling real risk.","2026-05-07T22:07:00.000Z","Scales balancing a padlock on one side and a safety shield on the other, representing the tradeoff between privacy and safety in digital platforms.",{"key":92,"type":46,"kindLabel":18,"label":93,"title":94,"path":95,"excerpt":96,"publishedAt":81,"featuredImageUrl":40,"featuredImageAlt":97},"6589bdb8eb91","Part 7","The Mental Impact No One Is Regulating Yet","\u002Fperspective\u002Fthe-mental-impact-no-one-is-regulating-yet","The most measurable harms get regulated first. But the slow accumulation of distorted thinking, constant comparison, and shallow engagement is harder to name — and probably matters more.","Abstract illustration of overlapping social media windows and AI chat bubbles surrounding a human head silhouette, representing mental overload and distorted thinking.",[99,103,106,109,112,115],{"_id":100,"_type":46,"kindLabel":18,"title":55,"slug":101,"path":56,"excerpt":57,"publishedAt":40,"updatedAt":40,"authorName":40,"featuredImageUrl":40,"featuredImageAlt":58,"featured":102},"8af62ad5-be5b-4e69-aeb4-a358474652d5","what-gdpr-and-casl-teach-us",false,{"_id":104,"_type":46,"kindLabel":18,"title":48,"slug":105,"path":49,"excerpt":50,"publishedAt":40,"updatedAt":40,"authorName":40,"featuredImageUrl":40,"featuredImageAlt":51,"featured":102},"b00e8692-3b67-42f7-8dfd-bb373a38323e","why-self-regulation-fails",{"_id":107,"_type":46,"kindLabel":18,"title":62,"slug":108,"path":63,"excerpt":64,"publishedAt":40,"updatedAt":40,"authorName":40,"featuredImageUrl":40,"featuredImageAlt":65,"featured":102},"bef82def-d299-4bef-a518-13376c75f74f","the-only-thing-that-works-hitting-the-bottom-line",{"_id":110,"_type":46,"kindLabel":18,"title":86,"slug":111,"path":87,"excerpt":88,"publishedAt":89,"updatedAt":40,"authorName":40,"featuredImageUrl":40,"featuredImageAlt":90,"featured":102},"68118e39-b171-4b3f-b1ad-e744c5a3b18f","privacy-vs-safety-where-should-the-line-be",{"_id":113,"_type":46,"kindLabel":18,"title":78,"slug":114,"path":79,"excerpt":80,"publishedAt":81,"updatedAt":40,"authorName":40,"featuredImageUrl":40,"featuredImageAlt":82,"featured":102},"193e8fe4-8adb-4ed1-b20f-74cb5c7ec6e6","should-free-access-exist-for-kids",{"_id":116,"_type":46,"kindLabel":18,"title":94,"slug":117,"path":95,"excerpt":96,"publishedAt":81,"updatedAt":40,"authorName":118,"featuredImageUrl":40,"featuredImageAlt":97,"featured":102},"5fcbd76d-d6bc-4def-bae2-713dc9fc528d","the-mental-impact-no-one-is-regulating-yet","Shannon @ MPC",{"_id":104,"_type":46,"kindLabel":18,"title":48,"slug":105,"path":49,"excerpt":50,"publishedAt":40,"updatedAt":40,"authorName":40,"featuredImageUrl":40,"featuredImageAlt":51,"featured":102,"body":120,"seo":385},[121,132,141,149,157,165,173,181,189,197,205,213,221,229,237,245,253,261,269,277,285,293,301,309,317,325,333,341,349,357,365,373,381],{"_key":122,"_type":123,"children":124,"markDefs":130,"style":131},"d256e6161e08","block",[125],{"_key":126,"_type":127,"marks":128,"text":129},"4a0adefe654e","span",[],"Why Self-Regulation Fails",[],"h2",{"_key":133,"_type":123,"children":134,"markDefs":139,"style":140},"997c70811428",[135],{"_key":136,"_type":127,"marks":137,"text":138},"412fcd5d408f",[],"I keep hearing this argument come up whenever people talk about AI and social media, especially when kids are involved.",[],"normal",{"_key":142,"_type":123,"children":143,"markDefs":148,"style":140},"b64e113b1a07",[144],{"_key":145,"_type":127,"marks":146,"text":147},"6e82340418b7",[],"The idea is basically that companies should just make better choices. That if we give them time, or pressure them socially, or trust that the people building these tools care enough, things will sort themselves out.",[],{"_key":150,"_type":123,"children":151,"markDefs":156,"style":140},"9f985795dd45",[152],{"_key":153,"_type":127,"marks":154,"text":155},"510863c7ca90",[],"And I get why people want that to be true. It’s a much easier solution than regulation. It feels less heavy, less invasive, more optimistic.",[],{"_key":158,"_type":123,"children":159,"markDefs":164,"style":140},"84009cc9c6ac",[160],{"_key":161,"_type":127,"marks":162,"text":163},"c9d18e9edb01",[],"But when you actually look at how these companies operate, it just doesn’t hold up.",[],{"_key":166,"_type":123,"children":167,"markDefs":172,"style":140},"ee2ed7573732",[168],{"_key":169,"_type":127,"marks":170,"text":171},"caa0a2445e8a",[],"Companies like OpenAI or Meta aren’t public services. They’re not built to sit there and ask, “what’s the healthiest outcome for society over the next 20 years?” They’re built to grow.",[],{"_key":174,"_type":123,"children":175,"markDefs":180,"style":140},"8fe86aa2c5ce",[176],{"_key":177,"_type":127,"marks":178,"text":179},"e700bed36902",[],"That means more users, more engagement, more data, more revenue. Not because anyone sat in a room and decided to harm people, but because that’s literally what the system rewards.",[],{"_key":182,"_type":123,"children":183,"markDefs":188,"style":140},"167471a15c5c",[184],{"_key":185,"_type":127,"marks":186,"text":187},"7aa6efaca2ff",[],"So when we say they should self-regulate, what we’re really asking is for them to voluntarily slow their own growth. And that’s where the whole thing starts to break down for me.",[],{"_key":190,"_type":123,"children":191,"markDefs":196,"style":140},"2e4243470a67",[192],{"_key":193,"_type":127,"marks":194,"text":195},"0dbd104ff91b",[],"There’s this version of the conversation where people point to ethics. Better design. More responsible AI. And again, I don’t think those things are fake. I’m sure there are people inside these companies who genuinely care about building something good.",[],{"_key":198,"_type":123,"children":199,"markDefs":204,"style":140},"f7b8d10f0717",[200],{"_key":201,"_type":127,"marks":202,"text":203},"f151520563a5",[],"But caring doesn’t change incentives.",[],{"_key":206,"_type":123,"children":207,"markDefs":212,"style":140},"f6dc1b5a2ef8",[208],{"_key":209,"_type":127,"marks":210,"text":211},"d0b71f67242a",[],"If a feature keeps people on the platform longer, it’s going to get prioritized. If something increases engagement, even if it has some negative side effects, it’s very hard for that not to win internally. Not because anyone is evil, but because the numbers will always justify it.",[],{"_key":214,"_type":123,"children":215,"markDefs":220,"style":140},"331fc52cd860",[216],{"_key":217,"_type":127,"marks":218,"text":219},"87382135c681",[],"We’ve already seen this play out over and over again. Social media feeds didn’t become addictive by accident. Recommendation systems didn’t drift into extreme content randomly. Email marketing didn’t magically become respectful of your inbox one day.",[],{"_key":222,"_type":123,"children":223,"markDefs":228,"style":140},"4f77732454df",[224],{"_key":225,"_type":127,"marks":226,"text":227},"4244e94e0779",[],"It changed when it had to.",[],{"_key":230,"_type":123,"children":231,"markDefs":236,"style":140},"a53a62fdf256",[232],{"_key":233,"_type":127,"marks":234,"text":235},"e5915d894a37",[],"A really simple example is email. Before Canada’s Anti-Spam Legislation, companies sent whatever they wanted. No unsubscribe, no consent, nothing. It worked, so they kept doing it. It only stopped when there were actual consequences tied to it. Same with General Data Protection Regulation. Companies didn’t suddenly decide to respect user data out of the goodness of their hearts. They changed because they would lose access to entire markets if they didn’t.",[],{"_key":238,"_type":123,"children":239,"markDefs":244,"style":140},"9326f719e17a",[240],{"_key":241,"_type":127,"marks":242,"text":243},"ee323bafa110",[],"That’s the pattern. Not intention, but pressure.",[],{"_key":246,"_type":123,"children":247,"markDefs":252,"style":140},"6345d18fee56",[248],{"_key":249,"_type":127,"marks":250,"text":251},"b50f9ceefa2d",[],"And I think this becomes way more serious when we’re talking about kids, because the “just make better choices” argument doesn’t apply the same way.",[],{"_key":254,"_type":123,"children":255,"markDefs":260,"style":140},"9ae1a201b1b7",[256],{"_key":257,"_type":127,"marks":258,"text":259},"db325922a613",[],"Kids don’t understand what’s happening under the hood. They don’t understand how their data is being used, or how algorithms are shaping what they see, or why certain content keeps showing up. They’re just interacting with something that feels fun or interesting or validating in the moment.",[],{"_key":262,"_type":123,"children":263,"markDefs":268,"style":140},"8005afb53740",[264],{"_key":265,"_type":127,"marks":266,"text":267},"87aeaa2eb5c0",[],"At the same time, the barrier to entry on most of these platforms is basically nothing. You can sign up for free, say you’re older than you are, and you’re in. And that free model is actually a huge part of the problem, because it’s not really free. You’re paying with your attention, your behavior, your data. That’s the whole engine.",[],{"_key":270,"_type":123,"children":271,"markDefs":276,"style":140},"329d9407c68c",[272],{"_key":273,"_type":127,"marks":274,"text":275},"2b12ded663b6",[],"So now you’ve got a system where it’s incredibly easy for kids to get in, incredibly profitable for companies to keep them there, and almost no real enforcement around whether they should be there in the first place.",[],{"_key":278,"_type":123,"children":279,"markDefs":284,"style":140},"72f961838238",[280],{"_key":281,"_type":127,"marks":282,"text":283},"a21091950747",[],"And we’re still expecting self-regulation to solve that.",[],{"_key":286,"_type":123,"children":287,"markDefs":292,"style":140},"4df8e9dc2b5c",[288],{"_key":289,"_type":127,"marks":290,"text":291},"e7aea4eef78c",[],"I don’t think the issue is that companies don’t care at all. I think it’s that the system they’re operating in doesn’t reward them for caring in the ways that actually matter here.",[],{"_key":294,"_type":123,"children":295,"markDefs":300,"style":140},"988e1d6a1219",[296],{"_key":297,"_type":127,"marks":298,"text":299},"ae4126f5cec1",[],"If growth is the goal, growth will win. If engagement is the metric, engagement will win. If collecting more data improves the business, that’s what the product will move toward. Every time.",[],{"_key":302,"_type":123,"children":303,"markDefs":308,"style":140},"a33903f4b14e",[304],{"_key":305,"_type":127,"marks":306,"text":307},"9349a854a0d6",[],"Until something outside the company forces a different outcome.",[],{"_key":310,"_type":123,"children":311,"markDefs":316,"style":140},"1ab2512e7c2a",[312],{"_key":313,"_type":127,"marks":314,"text":315},"892c64b2a2f5",[],"That’s the part I keep coming back to. The only time we’ve really seen meaningful shifts is when there’s something at stake that the company can’t ignore.",[],{"_key":318,"_type":123,"children":319,"markDefs":324,"style":140},"169260898345",[320],{"_key":321,"_type":127,"marks":322,"text":323},"0ad21419ae1d",[],"Not a suggestion. Not public pressure that fades after a news cycle. Something that directly affects whether they can operate, or how much money they can make.",[],{"_key":326,"_type":123,"children":327,"markDefs":332,"style":140},"c277e8346039",[328],{"_key":329,"_type":127,"marks":330,"text":331},"ad92e1a9a0ce",[],"That could look like losing access to a market entirely. It could be restrictions on advertising. It could be fines that actually scale with revenue instead of being treated like a cost of doing business. It could be ongoing compliance requirements that make it more expensive not to follow the rules than to just build the system properly in the first place.",[],{"_key":334,"_type":123,"children":335,"markDefs":340,"style":140},"78ea9327083a",[336],{"_key":337,"_type":127,"marks":338,"text":339},"04417c130c62",[],"Once you hit that level, the conversation changes internally. It’s not “should we do this,” it’s “we have to.”",[],{"_key":342,"_type":123,"children":343,"markDefs":348,"style":140},"ea1806cfe543",[344],{"_key":345,"_type":127,"marks":346,"text":347},"8be3de69f608",[],"And that’s really the core of it. Self-regulation is optional. Accountability isn’t.",[],{"_key":350,"_type":123,"children":351,"markDefs":356,"style":140},"e94a4ed224f9",[352],{"_key":353,"_type":127,"marks":354,"text":355},"c1e53ed0a9c2",[],"From a parent perspective, this matters because we’re not just handing our kids devices or apps. We’re putting them into systems that are very, very good at holding attention and shaping behavior, and those systems are not neutral.",[],{"_key":358,"_type":123,"children":359,"markDefs":364,"style":140},"ccd674114eee",[360],{"_key":361,"_type":127,"marks":362,"text":363},"49c605f8c0f4",[],"I’m not saying the solution is simple or that regulation fixes everything. It doesn’t. There are tradeoffs, and some of them are uncomfortable.",[],{"_key":366,"_type":123,"children":367,"markDefs":372,"style":140},"c1ac653aa62c",[368],{"_key":369,"_type":127,"marks":370,"text":371},"64a45c87d1bd",[],"But hoping that companies will just choose to limit themselves in ways that hurt their own growth, especially when it comes to kids, doesn’t feel like a real strategy.",[],{"_key":374,"_type":123,"children":375,"markDefs":380,"style":140},"5f6bf9a6bc03",[376],{"_key":377,"_type":127,"marks":378,"text":379},"76b1d99e846f",[],"It feels like something we say because the alternative is harder to deal with.",[],{"_key":382,"_type":383,"alt":384,"children":40,"imageUrl":40,"markDefs":40},"28ca0b0fd73c","image","Abstract illustration of a child holding a smartphone surrounded by overlapping interface windows and attention-grabbing notifications.",{"canonicalUrl":40,"metaDescription":386,"metaTitle":48,"metaTitlePrefix":40,"noIndex":102,"ogDescription":50,"ogImageUrl":40,"ogTitle":48},"Why tech companies rarely change on their own, why self-regulation keeps failing—especially for kids—and why real accountability has to come from outside pressure."]