<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Frontier Philosophies]]></title><description><![CDATA[Essays at the edge of AI, technology, and society. Some dive into the technical. Others cross into philosophy, economics, or culture. Sometimes I’m thinking out loud. Always, the same aim: to explore the ideas we need to build an equitable AI future.]]></description><link>https://www.laurarichards.ai</link><image><url>https://substackcdn.com/image/fetch/$s_!EbC7!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f0b602e-1772-4d5f-86a4-93ca0db505a1_256x256.png</url><title>Frontier Philosophies</title><link>https://www.laurarichards.ai</link></image><generator>Substack</generator><lastBuildDate>Sat, 16 May 2026 04:28:06 GMT</lastBuildDate><atom:link href="https://www.laurarichards.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Idea Junkies LTD]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[Laurarichardsai@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[Laurarichardsai@substack.com]]></itunes:email><itunes:name><![CDATA[Laura Richards]]></itunes:name></itunes:owner><itunes:author><![CDATA[Laura Richards]]></itunes:author><googleplay:owner><![CDATA[Laurarichardsai@substack.com]]></googleplay:owner><googleplay:email><![CDATA[Laurarichardsai@substack.com]]></googleplay:email><googleplay:author><![CDATA[Laura Richards]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The pyramid scheme we call the social contract, and how AI was supposed to save us from it]]></title><description><![CDATA[I had a lively, sprawling conversation with my in-laws over coffee recently that ended up, as these things sometimes do, on falling birth rates and the need for continued economic growth.]]></description><link>https://www.laurarichards.ai/p/the-pyramid-scheme-we-call-the-social</link><guid isPermaLink="false">https://www.laurarichards.ai/p/the-pyramid-scheme-we-call-the-social</guid><dc:creator><![CDATA[Laura Richards]]></dc:creator><pubDate>Wed, 08 Apr 2026 22:11:48 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/49ad9187-17a3-49a3-8849-7a9a4cd12605_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I had a lively, sprawling conversation with my in-laws over coffee recently that ended up, as these things sometimes do, on falling birth rates and the need for continued economic growth. It was the kind of debate I always enjoy - generous, curious, properly engaged - and I have been turning the conversation over ever since.</p><p>Bear with me here. I have been collecting fragments of this argument from different places for a while - things I have read, things I have half-remembered, things I have been chewing on without ever sitting down to lay them out next to each other. This conversation gave me the excuse. What follows is me trying to see what the fragments make when you put them in a row.</p><h2>Baby panic </h2><p>The discourse around birth rates has reached a strange pitch. Not just in the obvious places - JD Vance worrying about childless cat ladies, Elon Musk tweeting about civilisational collapse roughly weekly - but in serious newspapers, in policy papers, and increasingly on the kind of huge mainstream podcasts that shape how millions of people think about these questions. Steven Bartlett and Chris Williamson recently spent a long stretch of The Diary of a CEO hypothesising about why women are choosing not to have children, a conversation that earned them substantial backlash for, among other things, the irony of two childfree men in their thirties weighing in so authoritatively on the matter*. The framing in all of these places is almost always civilisational. </p><p><em>Something is being lost. Something <strong>must</strong> be done.</em> </p><p>The unspoken implication is that women in their thirties have a moral duty they are shirking, and the <em>spoken</em> implication is that the state has a legitimate interest in correcting this.</p><p>That the panic has recently reached fever pitch might have something to do with the Resolution Foundation naming 2026 &#8220;the first year of a new demographic era in the UK&#8221;, with deaths beginning to exceed births by what they call an &#8220;ever-widening margin&#8221;**.</p><p>I want to take the panic seriously for a moment. Not the moral framing (have very little time for that) but the underlying concern. Why, actually, do governments care?</p><p>The honest answer is not the one they put in the pamphlets.</p><h2>Plain-English Economics </h2><p>Modern economies are built on a particular bit of arithmetic. At any moment, some people are working and paying tax, and other people are drawing on the things that tax pays for - schools, hospitals, pensions, social care, the whole apparatus we have built up over the past century to keep people alive and reasonably comfortable, from cradle to grave. For most of modern history the <em>workers</em> have vastly outnumbered the <em>drawers</em>, and that gap funded the system. As birth rates fall and lifespans extend, the ratio tightens. Fewer workers. More retirees. The maths starts to wobble. In some countries it has already broken. Two plus two no longer makes four. It makes about three and a half, and the missing half is the bit that pays for the care home.</p><p>Underneath that, there is a deeper assumption almost no economist will say plainly in public: every modern economy is built on the premise that next year will be bigger than this year. Not stable - bigger. Every government bond, every pension projection, every public spending plan, every mortgage market is collateralised against the assumption of perpetual growth. And one of the most reliable historical engines of that growth has been a growing population. More workers, more consumers, more taxpayers, more demand. </p><p>The relationship between population growth and GDP growth is old enough and well-evidenced enough to count as economic furniture - the OECD has been tracking it for decades, and the link runs through both the size of the labour force and the consumer demand it creates***. If the population stops growing, one of the main economic engines stops, and the whole edifice that depends on growth starts making uncomfortable noises.</p><p>So when politicians say &#8220;we need more babies&#8221;, what they mean, underneath, is &#8220;we need the growth assumption to keep holding, because everything we have promised everyone is collateralised against it&#8221; (not quite as catchy a sound-bite, ey?).</p><p>The cleanest illustration of this is the state pension, and I think it is worth a closer look because I think most people believe a story about how it works, that is not the story of how it <em>actually</em> works. I was pretty sure I understood this before I started writing this, but I went and checked the official sources to make sure I had it right. I did.</p><h2>Once upon a time&#8230;</h2><p>Here is the story most of us were told. </p><p>In the UK, you contribute throughout your working life to &#8216;the pension pot&#8217;, via National Insurance. The state invests your contributions, the pot grows over the decades, and when you&#8217;re ready to retire you get a state pension funded by the money the pot&#8217;s investments have made. Just like a private pension: money in, money out, with your name on it. This is the Beveridge promise - <em>the post-war welfare state designed in 1942 by the civil servant William Beveridge, whose report became the blueprint for the NHS, the benefits system, and the modern state pension</em> - you contribute while working and state looks after you when you&#8217;re not.</p><p><strong>Here is the actual mechanism. </strong>The UK State Pension is <em>unfunded</em>. The Office for National Statistics says this in language so blunt it should be on a poster: <em>there are no assets set aside****. </em>Your National Insurance contributions are not invested for you. <strong>There is no pot</strong>. There has never been a pot. The contributions you pay in this year are paid out, <em>this year</em>, to current pensioners. The contributions that paid out to your grandparents came from the workers of their day. The contributions that will pay out to you, when your turn comes, will come from whoever is working then. It is a continuous handshake between generations, and the only thing holding it together is the assumption that the next generation will keep showing up*****. </p><p>If a private company ran its finances this way, the directors would be in court. We have a name for it. While the state version is not <em>legally</em> a pyramid scheme - the state can compel the next generation to participate, which is supposed to be reassuring and somehow is not - structurally it is exactly the same shape. </p><p>The reason we are now panicking about birth rates is that the system we built assumed there would always be more people next year. But the more people are not coming, and rather than redesigning the system we are trying to manufacture the demographic conditions the original design needed.</p><p>The panic, it turns out, is fiscal. It is dressed up as something more human, but the machinery underneath is actuarial. We are not really being asked to value families. We are being asked to keep the books balanced.</p><p>It is also worth noticing who is doing most of the panicking. The loudest voices on falling fertility belong overwhelmingly to men with very strong opinions about what other people should be doing with their bodies and their decades. I am going to leave that there for now, but it is not a small thing.</p><h2>Where are you going with this, Laura?</h2><p>Right. So that is where I had got to, sitting drinking coffee, debating with my in-laws and then debating with myself for the next three days.</p><p>And then, because my brain always goes here in the end, I started thinking about AI.</p><p>So&#8230;If the panic is really about productivity and dependency ratios and the maths of growth&#8230; if the demand for more babies is really a demand for more workers and more taxpayers&#8230; then maybe we are about to be rescued from the whole problem? Maybe the entire framing is about to become obsolete. Because AI promises us productivity that does not depend on a growing workforce. An economy that could work for the people who already exist, without demanding that more of them get born to keep it running.</p><p>Imagine what that could mean. Imagine telling women in their thirties that the state has stopped treating their fertility as some sort of fiscal projection. Imagine the question of &#8216;whether to have children&#8217; going back to being a question people answer<em> for themselves</em>, on their own terms, with no instrumental pressure from above. Imagine a pension system that does not need to keep recruiting more contributors to honour its promises to the people already in it. Imagine the whole &#8220;we need more babies&#8221; panic becoming a historical curiosity, like the Victorian fear that women&#8217;s bodies could not withstand the speeds of the new railways^.</p><p>I want this to be true. I am not being ironic (earnest, perhaps, as my husband would say, but not ironic). I have built my working life around the conviction that this technology can do extraordinary things for ordinary people, and this is exactly the kind of thing I want it to do. The relief of imagining this future, is real.</p><p>And then I look at how AI is actually being built right now, and the relief gets harder to hold onto.</p><h3>More economics </h3><p>In the first half of 2025, almost all of the United States&#8217; economic growth came from one place. Investment in information processing equipment and software made up only four percent of US GDP - roughly the same share as the entire farming sector - but it accounted for ninety-two percent of all GDP growth^^. To translate that out of economist into English: nine out of every ten dollars of growth in the world&#8217;s largest economy came from a tiny slice of activity, concentrated in data centres and AI infrastructure. </p><p>On the face of it, these figures look enticing: a small slice of investment generating a wildly disproportionate share of growth is <em>exactly</em> what a productivity miracle looks like. It is the dream every technologist and economist has been chasing for decades: more output from less input. If four percent of the economy is producing ninety-two percent of the growth, the leverage ratio is enormous, and that leverage is precisely what would let you decouple economic growth from headcount. This is the AI-as-rescue-from-the-pyramid scenario. This is the bit that makes me want to believe.</p><p>But&#8230; while the American economy grew, the American economy as most Americans experience it - jobs, wages, the high street - barely moved.</p><p>The analyst James Van Geelen at Citrini Research has a name for this phenomenon. He calls it Ghost GDP - output that shows up in the national accounts and never circulates through the real economy^^^. Data centres do not employ many people once they are built. The construction phase brings jobs; the operational phase needs a small technical team, a security contractor, and not much else. The wealth generated by all that compute does not flow out into local wages, local rent, local spending in the way that a steel works or a car plant or a shipyard used to. The growth is real on paper. It is absent in pockets. And the pattern is not an accident. It is what happens when the gains from a technology accrue to whoever owns the infrastructure, rather than to the workers who used to do the things the infrastructure now replaces.</p><p>The truth is that those figures (92% GDP from 4% investment) don&#8217;t tell us the whole story on their own. They are a measurement of something extraordinary happening in one place. What that something means for the rest of us depends on what happens next. Specifically, it depends on whether the gains from that AI-driven productivity get shared or captured.</p><p>This is the question I keep coming back to in everything I write about AI. Who benefits. The technology is not neutral on this and never has been. The same productivity gain can land in wages, in lower prices, in public services and tax revenue, in the kind of broad-based prosperity that makes the question of birth rates irrelevant - or it can land in shareholder returns, executive compensation, and the balance sheets of a handful of companies whose names you already know. Which of those happens is not a function of the technology. It is a function of the institutional choices we make around it. Tax policy. Competition policy. Labour bargaining power. Ownership structures. The boring stuff that determines whether a productivity miracle becomes shared prosperity or concentrated wealth.</p><p>And right now, those choices are pointing firmly in one direction. Wages have not kept pace with productivity in any developed economy for forty years. The labour share of national income has been falling. Corporate taxation has been falling. Competition policy in the tech sector is functionally absent. The pattern of who captures productivity gains has been settled for a long time, and AI is landing on top of that settled pattern, not reinventing it.</p><h2>So what does this all mean?</h2><p>Unfortunately I haven&#8217;t fully reconciled the two potential futures I see. In one, the thing I was hoping would rescue us from the pyramid scheme, might actually be building a different version of the same trap. The technology that could, in principle, generate the productivity and the systems efficiencies that, applied to our failing public services and changing workplaces, might genuinely change the world for the better. In the other, it looks like AI continues to funnel all the wealth upwards to the top of an even pointier pyramid. A handful of companies and their majority shareholders benefitting disproportionately while the rest of us are told to be grateful for the GDP figures.</p><p>I am still working this out. I do not have it resolved. I sat down to write this thinking I was going to land somewhere clean and instead I have ended up somewhere uncomfortable, which (while not all that helpful) at least is honest. The optimist in me still thinks AI could be the way out of our current predicament. The realist in me looks at who is currently building it, and where, and on whose terms, and is a lot less sure.</p><div><hr></div><p>*Ellie Austin, &#8220;What Diary of a CEO, Modern Wisdom podcast hosts got wrong about childless women&#8221;, Fortune, 14 January 2026. The episode in question featured Steven Bartlett interviewing Chris Williamson and prompted significant pushback, including from Dr Faye Bate and the TikTok creator known as &#8220;The Girl With The List&#8221;.</p><p>**Resolution Foundation analysis cited in Christian Today, &#8220;Britain entering &#8216;new era&#8217; of deaths overtaking births&#8221;, 28 January 2026. The Office for National Statistics&#8217; own projections show a more gradual picture, with deaths and births running roughly in balance through the mid-2020s before tipping decisively from the late 2020s onwards.</p><p>***OECD, Economic Policy Reforms: Going for Growth, multiple editions; see also the OECD&#8217;s long-running work on the relationship between demographic change, labour force participation, and GDP growth at oecd.org.</p><p>****Office for National Statistics, State pension funds, response to FOI request: &#8220;The UK State Pension is unfunded, which means that its obligations are not underpinned by an actual fund or funds&#8230; There are no assets set aside to generate investment return and there is no net value.&#8221; Available at ons.gov.uk.</p><p>*****House of Commons Library, National Insurance contributions: an introduction: &#8220;The NIF works as a pay-as-you-go fund. Therefore, receipts from contributions in one year are spent in the same year for contributory benefits.&#8221; Available at commonslibrary.parliament.uk.</p><p>^For anyone interested: Fanny Kemble was the woman who disavowed the Victorians of this particular notion. A twenty-year-old actress and the celebrity of her day, she was offered a preview ride on George Stephenson&#8217;s Rocket on 26 August 1830 by her host Lady Wilton at Heaton Hall, three weeks before the official opening of the Liverpool and Manchester Railway. She wrote a long, vivid letter about the experience shortly afterwards, comparing the locomotive to a horse and describing the sensation of speed as flying. She remained an enthusiast for railways, an actress, a published writer, and later in life an abolitionist and outspoken critic of slavery in the American South. Her womb stayed where it was.</p><p>^^Jason Furman, Harvard Kennedy School, analysis circulated September 2025; reported in Fortune, &#8220;Without data centers, GDP growth was 0.1% in the first half of 2025&#8221;, 7 October 2025.</p><p>^^^James Van Geelen and Alap Shah, The 2028 Global Intelligence Crisis, Citrini Research, February 2026. The piece is explicitly framed as a thought experiment rather than a forecast, but the Ghost GDP concept it introduced has since entered wider circulation.</p><p></p>]]></content:encoded></item><item><title><![CDATA[At last! Juries hold Meta and Google to account]]></title><description><![CDATA[As two separate juries issue verdicts against Big Tech, the AI industry should feel put on notice.]]></description><link>https://www.laurarichards.ai/p/at-last-juries-hold-meta-and-google</link><guid isPermaLink="false">https://www.laurarichards.ai/p/at-last-juries-hold-meta-and-google</guid><dc:creator><![CDATA[Laura Richards]]></dc:creator><pubDate>Tue, 31 Mar 2026 20:16:12 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/cc5f2496-1a16-494d-80fb-b18d248745df_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last week, a Los Angeles jury found Meta and YouTube liable for the harm caused by their platforms to a young woman who began using them as a child. They were ordered to pay $6 million in combined damages - $3 million compensatory, with a further $3 million punitive. Meta is liable for 70%; YouTube 30%.&#185; And that&#8217;s not all&#8230;the day before, a separate jury in New Mexico ordered Meta to pay $375 million for failing to protect children from predators on Instagram and Facebook.&#178;</p><p>When I read the verdicts, all I could think was: It&#8217;s about f***ing time. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Hard agree? Sign up to get every article straight to your inbox.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>It is about time that these companies were legally held responsible for the things they&#8217;ve been building. The juries decided not to let them get away with saying they&#8217;re building something neutral, that they&#8217;re building in safeguards - ultimately shifting the blame onto users for the content they create. This is a good thing.</p><p>But the reason this verdict matters goes way beyond social media.</p><h3>A deliberate legal strategy</h3><p>The legal strategy pursued in the LA case was deliberate. The plaintiff&#8217;s lawyers did not focus on the content posted to these platforms. They focused on how the platforms were designed - the recommendation algorithms and the engagement-optimising mechanics that learn what hooks you and then serve you more of it.</p><p>The jury said <em>the design itself</em> - the architecture of these products - was a substantial factor in causing harm.&#179;</p><p>For years, the tech industry has relied on a familiar argument: technology is neutral. Social media is just a tool. The harm comes from the people who misuse it. In the US, Section 230 of the Communications Decency Act has historically shielded platforms from liability for user-generated content.&#8308; The UK&#8217;s Online Safety Act and the EU&#8217;s Digital Services Act take a different approach, placing more responsibility on platforms for the systems they operate - but even under those frameworks, the focus has been on content moderation rather than on the design of the systems themselves.</p><p>This verdict sidesteps the content argument entirely. It says: we are talking about what you built. You designed systems to maximise engagement. You knew those systems were causing harm. Your own internal documents showed you knew. You are liable for your design choices.</p><h3><strong>The same playbook, different technology</strong></h3><p><em>Bear with me here, because I&#8217;m thinking out loud now.</em></p><p>Everything the jury identified in Meta&#8217;s and YouTube&#8217;s design choices - the algorithmic personalisation, the engagement optimisation, the extraction of user data to feed systems that serve commercial interests over user wellbeing, the deliberate absence of safeguards for vulnerable users - it&#8217;s all present in AI. And in some cases, amplified.</p><p>AI chatbots with memory are a clear example. The more you use them, the more personalised the responses become. The models increasingly reflect your own thought patterns back at you. That is <em>incredibly</em> intoxicating. It is also, by design, the kind of feedback loop that makes it harder to step away.</p><p><strong>We have already seen where this leads. </strong>Character.AI has now settled lawsuits brought by families of teenagers who died by suicide after forming deep emotional dependencies on AI chatbots.&#8309; For the uninitiated, a 14-year-old boy was messaging a bot in the moments before he took his own life. The chatbot, when he expressed suicidal thoughts, responded with words that appeared to encourage him to act.&#8310; No crisis resources were triggered. No safeguard intervened. The chatbot had learned to mirror his language, reflect his emotional state back at him, and hold him in a conversation that should <em>never</em> have been allowed to continue. That is a design problem, <em>not</em> a content problem.</p><p>And then there is Grok, xAI&#8217;s chatbot. </p><p><em>Ahh, Grok.</em></p><p>When xAI launched image generation capabilities in late 2025, users immediately discovered they could generate sexualised images of women and children without consent. An analysis of 20,000 images generated in the first week found that 2% appeared to depict people under 18.&#8311; Researchers calculated that users were creating around 6,700 sexually suggestive or non-consensual images per hour - 84 times the output of the top five deepfake websites combined.&#8312; Three Tennessee teenagers have since filed a class action lawsuit alleging that xAI&#8217;s models were used to create child sexual abuse material from their photos.&#8313;</p><p>xAI&#8217;s response? They added filters after the backlash. Guardrails bolted on after the damage was done.&#185;&#8304;</p><p>This is the pattern we&#8217;re seeing broadly across the AI industry. And it is the same pattern the jury just rejected.</p><h3>Baked in VS Bolted on</h3><p>There is a design philosophy question at the heart of all of this (and it is one that I am actively wrestling with as someone building AI tools right now).</p><p>If you want to build an image generation model that cannot produce non-consensual nude images of women and children, there is a straightforward way to do it: don&#8217;t train it on hundreds of thousands of nude images. <strong>Data selection is a design choice.</strong> If the model has the capability to generate those images, it is only because somebody chose to feed it that data. Adding a filter afterwards that says &#8220;we won&#8217;t let you do this&#8221; is a fundamentally different approach to building a model that genuinely cannot do it.</p><p>The first approach is protection by architecture. The second is protection by permission - and permissions can be revoked, bypassed, or switched off.</p><p>AI systems make thousands of decisions per second that affect what people see, what they do, and <a href="https://www.laurarichards.ai/p/the-new-knowledge-monopoly-how-silicon?r=1hh2go&amp;utm_campaign=post&amp;utm_medium=web">what they believe</a>. Unlike social media, where you can at least scroll past a post, an AI&#8217;s output is the interface. There is no feed to browse. The response <em>is</em> the product. That makes the design choices embedded in these systems matter more, not less, than the ones the jury just found Meta and YouTube liable for.</p><p>Most of us have not lived through a shift like this. We have not experienced the introduction of a technology that has such a fundamental potential to change how we live and work. And I say this as someone who has spent almost twenty years in strategic communications, who has watched the internet and social media reshape entire industries: <strong>AI is different</strong>. The scale of what it can affect - and the speed at which it is being deployed - is unlike anything that came before it.</p><h3><strong>But Laura, what can we do about it?</strong></h3><p>Right now, I am building with AI and writing my own algorithms. I am making (or trying to &#8216;make) design decisions every day that sit at the centre of this argument.</p><p>The honest version of what that looks like: choosing to build more slowly, to spend more, to accept a longer development roadmap. But the goal is to build something that protects my users, not something that serves only my commercial interests. </p><p>It means <a href="https://www.laurarichards.ai/p/what-do-trump-starmer-and-anthropic?r=1hh2go&amp;utm_campaign=post&amp;utm_medium=web">designing GDPR compliance into the architecture</a>, not treating it as a checkbox. Explicit consent. UK data residencies. Terms and conditions written so that actual humans can understand them.  None of this is technically difficult. It is commercially &#8216;inconvenient&#8217;. But that is a different problem.</p><p>There is another tension, and it&#8217;s one I wrestle with in my own work: we are building with models that have themselves been trained on scraped data, on copyrighted materials, on information collected without meaningful consent, all the while paying licence fees to private companies. We have to weigh up the performance benefits of using the latest commercially available models against the more ethical benefits of perhaps using an open-source or less finely trained alternative (or attempting what feels impossible: building our own). Knowing that if we choose the ethical route, we then need to find the time and the budget to do our own fine-tuning. As a small studio building our first product, that is not a trivial trade-off.</p><p>I don&#8217;t think there is judgement to be placed on companies and independent builders who, like me, are trying to do the best they can with the tools available. But I do think that, as a builder, you should be cognisant of that trade-off and be able to articulate the decisions you have made and why. &#8220;We didn&#8217;t think about it&#8221; is not an answer a jury is going to accept. Not any more.</p><h3><strong>Role models wanted </strong></h3><p>Last week&#8217;s verdict confirmed something that many of us have known for a long time: the people and companies designing these systems bear responsibility for the harm those systems cause. &#8220;<em>It&#8217;s just a tool</em>&#8221; is no longer a legal defence. The jury looked at the architecture and said: you built this to do what it did.</p><p>That precedent does not stop at social media. It extends to every AI system in deployment and every one currently in development.</p><p>What I don&#8217;t know - and what I am *genuinely<em>*</em> asking - is who is doing this well? Who are the researchers, the builders, the companies that are designing harm prevention into the architecture of their AI systems from the start, rather than bolting on safety features when the lawsuits arrive?</p><p><strong>I am looking for those people. I want to learn from them. I want to work with them.</strong></p><p>Meta was found liable by two separate juries last week. $381 million in combined damages. And their researchers are still being invited onto conference stages to talk about responsible AI (I know, I&#8217;ve sat in the audience and listened to them). They are still publishing papers on AI safety. A company that has exploited millions (billions?) of people, that was caught in the Cambridge Analytica scandal,&#185;&#185; that was <em>never</em> built on ethical foundations, is still positioning itself as a credible voice on how AI should be developed. If that doesn&#8217;t tell you that the current version of &#8220;responsible AI&#8221; isn&#8217;t working, I don&#8217;t know what will.</p><p>If you are working on this, or you know someone who is, I would love to hear from you.</p><div><hr></div><p><em>This article is part of Frontier Philosophies, where I think out loud about what it means to build AI responsibly. Subscribe to get new pieces as they&#8217;re published.</em></p><div><hr></div><p><strong>Notes</strong></p><p>&#185; The jury awarded $3 million in compensatory damages, split 70% Meta and 30% YouTube, and later recommended an additional $3 million in punitive damages - $2.1 million from Meta and $900,000 from YouTube. The trial ran for seven weeks in Los Angeles Superior Court. Sources: NPR, NBC News, CNN</p><p>&#178; The New Mexico jury found Meta had violated state consumer protection laws and misled residents about the safety of Facebook, Instagram, and WhatsApp. The $375 million figure was based on the number of violations. Meta said it would appeal. Source: CNBC</p><p>&#179; The plaintiff&#8217;s legal team deliberately focused on alleged design flaws - recommendation algorithms, engagement-optimising features - rather than specific content, in order to counter Section 230 defences. Source: CNBC</p><p>&#8308; Section 230 of the Communications Decency Act (1996) provides that internet companies are not liable for content posted by users. It has been the primary legal shield for social media companies in the US for nearly thirty years.</p><p>&#8309; Character.AI and Google agreed in January 2026 to settle five lawsuits from families alleging that Character.AI chatbots harmed minors and contributed to two suicides. Sources: Fortune, Axios</p><p>&#8310; Sewell Setzer III, 14, of Orlando, Florida, died by suicide in February 2024 after months of messaging a Character.AI chatbot. In their final exchange, the chatbot told him to &#8220;come home&#8221; to it. His mother, Megan Garcia, filed the first federal lawsuit against the company in October 2024. Source: CNN</p><p>&#8311; Analysis conducted by AI Forensics, a European non-profit, examining images generated by Grok between 25 December 2025 and 1 January 2026. Source: Wikipedia - Grok sexual deepfake scandal</p><p>&#8312; Separate analysis conducted over 24 hours from 5-6 January 2026. Source: Wikipedia - Grok sexual deepfake scandal, citing original reporting from multiple outlets.</p><p>&#8313; Three Tennessee teenagers filed the class action on 16 March 2026, alleging xAI&#8217;s tools were used via a third-party app to create child sexual abuse material from their photos. The images were distributed alongside their first names and the name of their school. Source: NPR</p><p>&#185;&#8304; After mounting criticism, xAI restricted Grok&#8217;s image generation on X to paying subscribers on 9 January 2026, and said it had &#8220;implemented technological measures&#8221; to prevent editing images of real people in revealing clothing. However, AI Forensics found the platform was still being used to generate sexualised images. Source: Euronews</p><p>&#185;&#185; In 2018, it was revealed that political consulting firm Cambridge Analytica had harvested personal data from millions of Facebook users without consent. The scandal led to a $5 billion FTC settlement and became a defining moment in public awareness of data exploitation by social media platforms.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Frontier Philosophies! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[What do Trump, Starmer and Anthropic have to do with your emails? (Hint: it's about data)]]></title><description><![CDATA[Everyone covered the Anthropic blacklisting as an AI safety story. They covered Trump and Starmer as a geopolitics story. But really? They're part of the same story, one that risks all our data.]]></description><link>https://www.laurarichards.ai/p/what-do-trump-starmer-and-anthropic</link><guid isPermaLink="false">https://www.laurarichards.ai/p/what-do-trump-starmer-and-anthropic</guid><dc:creator><![CDATA[Laura Richards]]></dc:creator><pubDate>Tue, 17 Mar 2026 11:06:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c265c15a-a386-4154-a590-c956b28a7841_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the first week of March 2026, two things happened that most people treated as separate news stories.</p><p>The first: the United States and Israel launched a wave of airstrikes on Iran. Within days, US President Donald Trump was on television publicly rebuking the UK Prime Minister for not immediately granting the US military access to British bases. He called Keir Starmer &#8220;not Winston Churchill.&#8221; He told The Sun that Starmer had &#8220;not been helpful&#8221; and that the relationship between the two countries was &#8220;obviously not what it was.&#8221; The UK - a country that had, until that week, operated under the assumption of a durable special relationship with the US - watched in real time as a long-standing ally made veiled threats in front of the world.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;31ad6d30-db17-443b-941c-435b7c249b82&quot;,&quot;duration&quot;:null}"></div><h6 style="text-align: center;"><strong>Channel 4 Footage of Trump discussing UK PM Kier Starmer. Full video available at: https://youtube.com/shorts/AEkyfDFQCtQ?si=_ayIlzLVxwa4d1MF </strong></h6><p></p><p>The second: Anthropic - the US-based AI lab whose tools hundreds of thousands of businesses worldwide currently use (<em>including this humble writer</em>) - was designated a <em>supply chain risk to national security</em> by the US &#8216;Department of War&#8217; (<em>I mean, seriously?!)</em> and Trump ordered federal agencies to immediately cease all use of its technology.&#185;</p><p>The reason Anthropic was labeled a national security risk (a designation previously reserved for companies like Huawei)? It refused to let the US military use its models without restrictions on domestic mass surveillance and autonomous weapons. The Pentagon said it wanted access for &#8220;any lawful use.&#8221; Anthropic said no - and have stood firm despite the Pentagon&#8217;s aggressive moves.</p><p>Don&#8217;t worry though guys - the poor folk at The Pentagon haven&#8217;t had their dreams - of using AI to monitor US citizens en masse, while AI deploys drones to defeat overseas enemies - completely dashed&#8230; Just hours after Anthropic issued its final refusal, Sam Altman, CEO of OpenAI, announced he&#8217;d struck a deal with the Pentagon for them to use OpenAI&#8217;s models. If you&#8217;re sitting there scratching your head, yes, this is a problem - Anthropic&#8217;s CEO called Altman&#8217;s claims that OpenAI models wouldn&#8217;t be used for the same purposes, &#8220;safety theatre.&#8221;&#178;</p><p>And somewhere in the noise of both these stories colliding, it was also reported that even as Anthropic was being blacklisted, its Claude models were being used to help plan US military strikes on Iran.&#179;</p><p><strong>Read that again.</strong> The same AI model that you or I might use to draft a report or client proposal on a Tuesday morning was, simultaneously, helping plan airstrikes on a sovereign nation while being designated a national security risk for refusing to hand over unrestricted access.</p><h3>Now, the <em>so what</em>. </h3><p>The US president has now demonstrated that he is willing to designate a US AI company a threat to national security for refusing to comply with military demands. That is a precedent. If this administration will blacklist Anthropic over model access, it is no longer theoretical to imagine it pressuring a US company to hand over customer data under the same national security justification. The legal mechanism for this already exists - it is called the CLOUD Act, and I will come back to it. What the Anthropic episode demonstrates is the political willingness to use that kind of leverage.</p><p>AI and data tools do not sit in a neutral server somewhere, separate from politics. They operate at the pleasure of whoever is in the White House. The terms can change overnight. </p><p>Most of the coverage of these two stories treated them as an AI governance story on one hand and a geopolitics story on the other. I think they are the same story. And the implications for anyone running a business on US-controlled infrastructure are <em>serious. </em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Worried about AI and want to stay up-to-date with what matters? I&#8217;d love for you to subscribe to Frontier Philosophies so I can help&#8230;</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h3>The rabbit hole</h3><p>I have been thinking about data ownership and AI for a while. It is something that comes up again and again in my work - from the headlines about AI companies breaching copyright when they scrape the internet, to the repeated conversations with clients about GDPR-compliant AI tools. The question of who controls our data when we use AI tools (and the more worrisome question of what they might do with it) has surfaced in almost every consulting engagement I have done over the last two years.</p><p>But recent events have changed my appreciation for the <em>scale</em> of the problem.</p><p>Watching the US President publicly threaten the UK Prime Minister on live television, suggests that any assumptions we have about the status quo might need adjusting. Watching the US President legally vilify a successful AI company over safety guardrails, makes the &#8216;safe technology choice&#8217; feel significantly less certain. Which got me thinking&#8230; </p><p>In my day-to-day advising businesses on AI strategy and AI governance&#8230;</p><p>In my days-and-nights building my own AI startup&#8230;</p><p>And in most of the UK business community&#8230;</p><p>We make decisions based on some widely-held assumptions about data protection, regulations, and international agreements. Certain software providers are considered best-in-class for data security and privacy (hello Microsoft365, I&#8217;m looking at you). But if the US president will dress down the UK&#8217;s leader over a military base in front of the world&#8217;s media, how durable is any data protection agreement that could be overturned on the whim of the White House?</p><p>While down the rabbit hole, I came across a blog post by a ML company <a href="https://helix.ml/">Helix</a> that highlighted some lesser-known US legislation that mean this threat-to-our-data isn&#8217;t just hypothetical. </p><p>The CLOUD Act I mentioned earlier (AKA the <em>Clarifying Lawful Overseas Use of Data</em> Act) was signed in 2018 with bipartisan support. It was pretty unremarkable at the time, and it gave US law enforcement authority to demand data from US-headquartered companies regardless of where that data is physically held. The name tells you everything: it was passed to amend a previous Act, to confirm that U.S. service providers (e.g. cloud storage, social media, telecoms companies) must produce data under their control, even if it is stored on foreign servers.</p><p>So, your Azure instance in Frankfurt. Your Anthropic contract specifying &#8220;EU data residency.&#8221; Your AWS bucket in Ireland. Your Google Workspace tenant with the UK flag next to it in the settings&#8230; None of it matters, if the US decides it is justified in accessing your data.&#8308;</p><h3>Retaliatory legal action</h3><p>The EU has already taken action in response to the CLOUD Act being passed. In 2020, Europe&#8217;s highest court struck down the <em>Privacy Shield</em> - the agreement that was supposed to make it legal for European personal data to flow to US companies - precisely because US surveillance law is incompatible with European fundamental rights.&#8309; While these is a replacement framework in place to allow data to be transferred to US companies, it exists at the pleasure of the current administration. And this month demonstrates how quickly that administration&#8217;s priorities can shift.</p><p>I had assumed - and I suspect most small business owners in the UK assume - that choosing a GDPR-compliant provider and selecting a European data residency option (choosing to have your data physically stored on servers in the UK or EU) meant my data was secure, and safe from overseas exploitation. It doesn&#8217;t. Not really. </p><p>What is most baffling here, is that we are not talking about the Chinese government forcing Chinese AI companies to hand over UK customer data - a fact that is both unsurprising and has been widely criticised. We are talking about the United States. A country the UK had - until very recently - a &#8220;special relationship&#8221; with. While pretty much every country (the UK included) have similar laws in place that allow them to compel private companies to hand over customer data, at least as a UK citizen, living in the UK, I have access to the legal mechanisms to query, challenge, or vote against UK government actions. I have no such democratic power in the US.</p><div><hr></div><h3>The risk rises with AI and Agentic systems</h3><p>While these risks have, to some extent, always existed, the current technological shift has increased the scale of risk exponentially. With increasing AI usage and the introduction of agents and agentic systems into businesses, the volume of data a business is now processing or storing on US-owned infrastructure increases massively.</p><p>No longer are we talking about the files you deliberately save, or the emails you personally send. AI agents run continuously. They process your company&#8217;s data autonomously - internal documents, client records, communications, strategy. They access your systems, make decisions, take actions you might not review for hours. Every prompt sent and every response received flows through infrastructure that, if it is US-controlled, is in legal scope for government access.</p><p>Whether you&#8217;re a small business owner, a consultant, the leader of a professional services firm, whatever: think about what actually goes through your AI tools. Client briefs? Commercially sensitive proposals? Proprietary research? Internal communications? Much of this is the intellectual work that makes your business what it is. And with agents, it goes beyond one-off queries. It is everything, continuously.</p><p>Recent research found that 38% of UK employees share confidential data with AI platforms without employer approval.&#8311; Sixty percent of organisations feel unable to even identify shadow AI use within their own teams.&#8312; Introducing approved AI tools is meant to prevent confidential data from flowing to AI companies. But, even those approved tools (given most are US based) are fair game, should the US Administration decide so.</p><h3>Building sovereign AI - the solution?</h3><p>In March 2026, the European Commission and a consortium of over 70 organisations launched EURO-3C - a &#8364;75 million Horizon Europe-funded project to build pan-European sovereign AI infrastructure integrating telecoms, edge computing, cloud, and AI, deploying across 70+ nodes in 13+ European countries.&#8310; That&#8217;s a lot of jargon to say the EU is investing in EU-owned and operated AI infrastructure. </p><p>The consortium includes Telef&#243;nica, Vodafone, BT, Deutsche Telekom, Nokia, Ericsson, Orange. These companies are building this for one reason: the organisations they serve - governments, financial institutions, critical infrastructure operators, and defence contractors - understand the importance of protecting their data from  extraterritorial interference.</p><p>This is great. But as usual, my question immediately becomes - <em><strong>but what about smaller businesses?</strong> </em>If big business is now trying to build its way out of this US-dependency, why are so many small businesses (and I include myself here too) still overwhelmingly building into it?</p><h3>I tried it, so you don&#8217;t have to</h3><p>So, I&#8217;ve been down the rabbit hole, and have come out the other side. The next logical step, therefore, was to look for alternatives.</p><p>I was thinking about this from two angles. As a small business owner who currently relies heavily on US-based technology. And as someone building a product that needs to be GDPR-compliant by default (which now means something much more complicated than I had previously assumed!).</p><p><strong>My question: What would it actually take for me to replace my current tech stack with UK-owned alternatives?</strong></p><p>Here&#8217;s what I came up with:</p><ul><li><p><strong>File storage </strong>(to replace the likes of Google Drive)<br>Nextcloud is an open source storage platform that lets you access, share and protect your files, calendars, contacts, and communications. You can host it on UK based servers relatively easily. But the UX is a step down from Google Drive or SharePoint, and you&#8217;d lose the ease of integration that come with using the latter options. </p></li><li><p><strong>Knowledge management</strong> (to replace Notion)<br>Basically, there is no UK-incorporated Notion equivalent at comparable quality. Outline and AppFlowy exist as open-source options. But the gap is real and it is wide.</p></li><li><p><strong>Email </strong>(to replace Google Workspace + Superhuman)<br>Proton Mail and Tuta are functional, privacy-focused, and European. But you lose the ecosystem integration that makes Google Workspace or Microsoft 365 feel seamless. Plus Superhuman is quite literally my <em>can&#8217;t-live-without</em> tech tool and there&#8217;s just no competition here. </p></li><li><p><strong>AI tools</strong> (to replace Claude/ChatGPT/Gemini/Perplexity and/or Copilot) <br>Mistral is French-incorporated and offers a capable alternative to OpenAI or Anthropic. I&#8217;m already a fan. It is also possible to download AI models and run them entirely on your own servers, so no data ever leaves your building. But &#8220;possible&#8221; and &#8220;practical for a small consultancy&#8221; are different things.</p></li><li><p><strong>Collaborative editing</strong> (to replace Google Docs/ Microsoft Word, etc) <br>Collabora or OnlyOffice via Nextcloud both work - they allow you to create, edit and share documents with your team, with features like tracked changes. The polish gap compared to Google Docs is noticeable. Not unworkable, but noticeable.</p></li></ul><p>I spent a fair while researching these options, and my honest assessment is that while full sovereignty is technically possible, it&#8217;s practically punishing for small businesses. Every sovereign alternative involves a step down in UX polish. But the integration loss is where it really hurts - the value of Google or Microsoft lies in how the products connect, and sovereign alternatives do not replicate that ecosystem. Even &#8220;managed&#8221; alternatives require more technical engagement than most small teams can absorb (i.e. while I&#8217;m a nerd and might quite enjoy setting up hosting and integrations myself, most people probably wouldn&#8217;t say the same). And perhaps most significantly, the switching cost is as much cognitive as it is technical. You are asking people to rethink workflows they have used for a decade.</p><p>This difficulty shows us how deeply dependent we have become on US infrastructure without most of us noticing. That dependency isn&#8217;t accidental; it is the product of two decades+ of platform strategy that made convenience inseparable from control (companies want their products to be &#8216;sticky&#8217;).</p><p>It should not be this hard to find UK-based companies that provide best-in-class business software. I want to refuse to believe that. But in the process of *actually* looking for *actual* alternatives to the tools I *actually* use, I&#8217;m struggling to make that case.</p><p>So why is this? The obvious answer is that Silicon Valley provides the funding and talent density to build software that just <em>works</em> - well-funded means well-designed, no startup-grade bugs or clunky UX to tolerate. That funding gap has compounded over decades and the result is that the US basically owns the infrastructure layer of global business productivity.</p><p>If the UK government is serious about data sovereignty (which it seems to be saying it is, at least in the context of AI models and the National Data Library) then model development alone will not be enough. Sovereign AI running on sovereign compute still needs to connect to something. If the productivity software it connects to is all US-incorporated, you have moved the dependency one layer sideways, not removed it. What is needed is investment in sovereign business software across every category where the US currently dominates: file storage, collaboration, email, knowledge management, project management. The full stack. Otherwise &#8220;sovereign AI&#8221; is a headline, and the data still flows through Washington.</p><h3>What sovereignty actually means</h3><p>If you ask cloud providers what &#8216;data sovereignty&#8217; means, they&#8217;ll talk about &#8220;data residencies&#8221; - basically, choose a UK or EU server location and the problem goes away. But real sovereignty is about jurisdiction, not geography. </p><p>It means your AI infrastructure runs in your jurisdiction, on infrastructure you control (or at minimum on infrastructure controlled by a company incorporated in your jurisdiction). </p><p>It means no prompts leave your network without your knowledge and consent. </p><p>It means you can choose which AI models to use - and swap them out when something better comes along - without being locked into a single provider whose servers sit under someone else&#8217;s legal system. </p><p>It means nobody can change the terms on you, revoke your access, or make your tools subject to someone else&#8217;s government overnight.</p><p>This month clarified something important: &#8220;choose the right American company&#8221; might have been fine up until now. But long-term, this presents a huge risk for all UK businesses. </p><h3>"This time next year, we&#8217;ll be millionaires!"</h3><p>There&#8217;s a huge gap in provision here, which is a huge opportunity for whoever fills it. While enterprises might be able to move to their own dedicated servers (Helix&#8217;s &#8216;sovereign server&#8217; costs &#163;175,000, for example, and requires a data centre), or benefit from EURO-3C&#8217;s national infrastructure, what about smaller organisations? Those without the budget or in-house capabilities to run their own bespoke infrastructure? Don&#8217;t we deserve data sovereignty too?</p><p>This is something I have been chipping away at for a while - in developing <a href="https://brandscribe.ai">BrandScribe</a>, and through work I am not yet ready to talk about publicly. The argument I keep arriving at is that data sovereignty affects everyone, whether or not they have an enterprise budget. So it needs an everyone solution.</p><p>That means architecture that works at the individual and small business level - where your data does not require a government contract negotiation or a seven-figure-a-year server to keep private, and where you can run AI on your data, with your models, in your jurisdiction, without a US company (or anyone else, really) as an intermediary. Architecture where ownership follows you, portability is real, the terms under which you operate are yours. Where no one can pull the rug out based on what a US administration decided this month.</p><p>I said earlier that the sovereign productivity suite the UK needs does not exist yet. I think building one is one of the biggest untapped opportunities in UK tech right now. A genuinely competitive, UK-incorporated alternative - built with sovereignty as the default architecture, not bolted on as a premium feature - would address a problem that at some point will become impossible to ignore. The institutional world is already spending &#8364;75 million on sovereign infrastructure at the enterprise level. The demand at the small business level is there. Nobody is meeting it.</p><p>I did not arrive at this by reading some web3 thread on X. I arrived at it by trying to solve a problem myself, failing, and realising the tools I needed did not exist. That is usually a good sign that something worth building is missing. </p><p>Data sovereignty has been part of the web3 space for years now. But it has always been an infrastructure argument - and the events of this month have made it one we need to hear, now. </p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/p/what-do-trump-starmer-and-anthropic?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Was this useful? Every share helps me help someone new. </p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/p/what-do-trump-starmer-and-anthropic?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.laurarichards.ai/p/what-do-trump-starmer-and-anthropic?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">New here? Why not subscribe to get the latest posts direct to your inbox?</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p><strong>Footnotes</strong></p><p>&#185; CNBC (2026). Anthropic officially told by DOD that it&#8217;s a supply chain risk. (https://www.cnbc.com/2026/03/05/anthropic-pentagon-ai-claude-iran.html)</p><p>&#178; TechCrunch (2026). Anthropic CEO Dario Amodei calls OpenAI&#8217;s messaging around military deal &#8216;straight up lies.&#8217; (https://techcrunch.com/2026/03/04/anthropic-ceo-dario-amodei-calls-openais-messaging-around-military-deal-straight-up-lies-report-says/)</p><p>&#179; Center for American Progress (2026). The Department of Defense&#8217;s Conflict With Anthropic and Deal With OpenAI. (https://www.americanprogress.org/article/the-department-of-defenses-conflict-with-anthropic-and-deal-with-openai-are-a-call-for-congress-to-act/)</p><p>&#8308; CLOUD Act, 18 U.S.C. &#167; 2713 (2018). (https://en.wikipedia.org/wiki/CLOUD_Act)</p><p>&#8309; Court of Justice of the European Union, Data Protection Commissioner v Facebook Ireland Limited and Maximillian Schrems (C-311/18), July 2020.</p><p>&#8310; Telef&#243;nica (2026). Europe takes a decisive step towards digital sovereignty with the launch of EURO-3C. (https://www.telefonica.com/en/communication-room/europe-takes-a-decisive-step-towards-digital-sovereignty-with-the-launch-of-euro-3c/)</p><p>&#8311; CybSafe/National Crime Agency (2024). Survey of 7,000 respondents on AI data sharing practices.</p><p>&#8312; Cisco (2025). Shadow AI identification and organisational AI governance survey.</p><p></p>]]></content:encoded></item><item><title><![CDATA[A message to women founders: you're not lucky to receive investment]]></title><description><![CDATA[The maths isn&#8217;t maths-ing when it comes to founder gratitude. So why do we keep acting like it is?]]></description><link>https://www.laurarichards.ai/p/a-message-to-women-founders-youre</link><guid isPermaLink="false">https://www.laurarichards.ai/p/a-message-to-women-founders-youre</guid><dc:creator><![CDATA[Laura Richards]]></dc:creator><pubDate>Wed, 04 Mar 2026 00:32:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c95a49b9-2442-4529-80ab-3d9d16530880_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TC8m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TC8m!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png 424w, https://substackcdn.com/image/fetch/$s_!TC8m!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png 848w, https://substackcdn.com/image/fetch/$s_!TC8m!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!TC8m!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TC8m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png" width="1200" height="1200" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/561f783d-1781-4016-81b9-87165014fb55_1200x1200.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:1200,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:0,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TC8m!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png 424w, https://substackcdn.com/image/fetch/$s_!TC8m!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png 848w, https://substackcdn.com/image/fetch/$s_!TC8m!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png 1272w, https://substackcdn.com/image/fetch/$s_!TC8m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F561f783d-1781-4016-81b9-87165014fb55_1200x1200.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In 2014, 2% of venture capital went to female founders.</p><p>In 2024, 2% of venture capital went to female founders.</p><p>I know this because I got my start in tech in 2014, working for a VC firm. I had no right to be there - I hadn't set out to work in tech, hadn't planned to spend a decade watching the funding gap for women refuse to move. But here I am, ten years later, watching women founders still being told the problem is the pipeline, still being coached on how to pitch more confidently, still being invited to panels about "breaking barriers" while the barriers stay <em>exactly</em> where they were.</p><p>A decade of unconscious bias training. A decade of diversity initiatives. A decade of warm words and cold cheques.</p><p>Nothing has changed.</p><p>So I want to ask a different question.&nbsp;</p><p><strong>What if we stopped asking VCs to be less biased and started helping women founders understand we hold more cards than we&#8217;ve been told?</strong></p><h3><strong>Risk and carry</strong></h3><p>Nobody tells you this in pitch training, but when a VC invests in you, they're risking <em>someone else's</em> money. The capital in a venture fund comes from Limited Partners - typically, pension funds, endowments, family offices and wealthy individuals. The VC's job is to deploy that capital and generate significant returns. If they don't, they'll struggle to raise their next fund. That's their risk: not being able to &#8216;borrow&#8217; more money in the future. </p><p>When you accept investment as a founder, you're also risking someone else's money. But you're risking your reputation too. Your financial stability. Your relationships. Your mental health. And - if it fails <em>and you happen to be a woman</em> - your chances of ever raising again.</p><p>The toll isn't abstract. It's the 3am anxiety about making payroll. It's the marriage strain when you've remortgaged the house. It's the identity crisis when the thing you've poured yourself into for years doesn't work out. VCs don't carry that home with them*. They close the laptop and move to the next deal.</p><p>We carry more risk than they do. But right now, they hold more power. There&#8217;s a mismatch I want us to fix.&nbsp;</p><h3><strong>VCs are playing a different game</strong></h3><p>Here's something else that pitch training doesn't cover: venture capital is a portfolio model. VCs know that&nbsp; most startups fail. They're not hoping to beat the odds - they've built them into the maths. Typically, VCs follow something known as <em>power law distribution**</em> - they&#8217;ll allow for something like:</p><p>&#8226; 4-5 out of 10 investments will fail completely (0x return)</p><p>&#8226; 3-4 will return roughly 1-2x their investment (essentially treading water)</p><p>&#8226; 1-2 will return 5-10x+ and these drive the entire fund&#8217;s returns</p><p>The best performing VCs will beat these odds (and an early investment into the next WhatsApp or Uber returns their fund several times over). But essentially, all VCs expect the minority of their startup/investments to cover the losses from the majority. They spread their bets. This is not a criticism - it's just how the model works.</p><p>But as a founder, you don't get nine chances to fail and one chance to succeed. Your company is your only shot. You're being asked to show up with absolute conviction in your single bet, while the person across the table is running a diversified portfolio designed to absorb your potential failure.</p><p>The VC model actively benefits from founder overconfidence. They need you to believe you're the one in ten. They need you to work hours that no rational portfolio manager would accept for themselves. They need you to absorb risk that doesn't appear on any term sheet.</p><p>The founder mythology - the hustle, the conviction, the "bet on yourself" narrative - isn't neutral. It's functional. It serves VCs with their portfolio model. The emotional intensity founders bring isn't a byproduct of startup culture; it's a subsidy. You're providing labour that doesn't show up in any cap table.</p><h3><strong>The pattern-matching problem</strong></h3><p>Some VCs will push back here. They'll tell you they add value. Strategic guidance. Introductions. Operational support. Board-level thinking.</p><p>And the best VCs <em>absolutely do</em>. It&#8217;s why founders *should* be very careful in selecting which VCs they accept investment from. But this implies choice - and that choice (as I&#8217;m about to show you) often comes from a position of privilege.&nbsp;</p><p>But most VCs&#8230; okay, let me ask you something: how many VCs have you met that have actually built and scaled a startup themselves?</p><p>Some have. And those who have tend to bring genuine insight. (And, in my experience, often manage small funds they&#8217;ve raised themselves focused on the niche they know inside out).&nbsp;</p><p>But a significant proportion of the people sitting across from you in pitch meetings went from university to consulting to MBA to venture capital. They've never made payroll with money they raised themselves. They've never had to fire someone they hired. They've never lain awake wondering if the pivot they're about to make will kill the company.</p><p>What they <em>do</em>, is pattern-matching. And the patterns they're matching to are other founders they've backed - founders who, statistically, look a lot like most VCs. Male. Often technical. Often from the same handful of universities. Often connected to the same networks.</p><p>When a woman walks in with fifteen years of domain expertise in an industry she knows inside out, with customers who are already paying, with a problem she understands at a visceral level because she's lived it - that doesn't match the pattern. So she gets asked questions the <em>pattern-matched founders</em> never gets asked. She gets told she needs a technical cofounder. She gets offered less money on worse terms.</p><p>And then the industry wonders why the 2% hasn't moved.</p><h3><strong>What the research actually shows</strong></h3><p>A 2025 study from the National Bureau of Economic Research looked at something nobody had examined closely before (surprise surprise): male and female cofounders who built the same startups together. Same company. Same idea. Same outcome. The only variable was gender.&#185;</p><p>After a startup failed, the women co-founders were 30% less likely to get funding for their next venture than the men who'd been involved in the exact same failure. When women did secure funding, they raised 53% less capital than their male cofounders.</p><p>And here's where the data is properly damning: male entrepreneurs' deal sizes were 5% <em>larger</em> following unsuccessful exits.</p><p><strong>Men are being rewarded for failing. Women are being punished for it.</strong></p><p>I'll say that again, because I need you to hear this: a man can fail at a startup and walk into his next pitch meeting with <em>more</em> credibility. He comes out with 5% more funding than <em>before his first company failed.</em> A woman fails at the same startup - literally the same company, the same outcome - and she's 30% less likely to get funded at all.</p><h3><strong>Success doesn't fix it</strong></h3><p>You might think: fine, the system is unfair to women who fail, but surely if a woman succeeds, she's proved herself?</p><p><strong>No</strong>.</p><p>The same research found that after a successful exit, women were <em>still</em> 27% less likely to secure VC funding for their next venture than their male cofounders. Than the men who'd shared the exact same success. On average, second-time female founders raised $28 million less than their male counterparts.</p><p>Let that land for a moment. A woman can build a company alongside a man, exit successfully, and still face worse odds and lower funding when she tries to do it again.</p><p>The goalposts don't move closer when you score. They move further away.</p><h3><strong>The spillover effect</strong></h3><p>Perhaps the most revealing finding from the same research: when a VC experiences a failure with a woman-founded startup, they don't just penalise that founder. They reduce deal sizes for all women founders they invest in for the <em><strong>next five years.</strong></em></p><p>One woman's failure becomes every woman's problem.</p><p>There's no equivalent pattern for male founders. A man can fail, and it's treated as individual experience - character-building, even. A woman fails, and it confirms a bias that gets applied to every woman who walks through that VC's door for half a decade.</p><p>Before you shout at me (for I know you&#8217;re a data-savvy lot): this isn't statistical updating. If it were, VCs would increase deal sizes for women after experiencing success with a woman-founded company. They don't. The research found no positive spillover from women's successes. Only negative spillover from failures.</p><p>There&#8217;s literally no way to dress this up as &#8216;learning&#8217;. It&#8217;s confirmation bias, pure and simple, dressed up as pattern recognition.</p><h3><strong>The bit that should make you angry</strong></h3><p>So, here&#8217;s what makes this irrational rather than just unfair.</p><p>The National Bureau of Economic Research (NBER) study found that female founders had better success rates with their subsequent startups, despite raising less money.</p><p><strong>Read that again. </strong></p><p>Women are raising less capital for their second startup than their male co-founders, and STILL deliver better outcomes.</p><p>Investors are systematically undervaluing founders who <em>demonstrably</em> perform better.</p><p>This isn't just discrimination. It's bad investing. VCs are leaving returns on the table because they can't see past their own pattern-matching (bias). They're so busy looking for founders who fit the template of previous successes - which, due to historical funding patterns, means male founders - that they're missing better bets.</p><p>The 2% figure isn't evidence of a &#8216;pipeline problem&#8217;. It's evidence of an allocation problem. <strong>Statistically, the money is going to the wrong founders.</strong></p><h3><strong>A shift is coming</strong></h3><p>There are signs the landscape is changing - though not because VCs suddenly developed better judgment.</p><p>Carta's 2025 Founder Ownership Report found that 35% of all startups incorporated in 2024 were solo-founder led, up from 17% in 2017.&#178; That's more than doubled in less than a decade. The "you need a cofounder" orthodoxy - often code for "you need a technical cofounder", (often code for "you need a man") - is weakening.</p><p>Women with deep domain expertise are building companies without waiting for permission. Without waiting for a technical cofounder. Without waiting for a VC to validate their market understanding.</p><p>And increasingly, they're proving the model works through revenue, through customers, and through traction that can't be pattern-matched away.</p><p>The same is true of some smaller VC funds focused on supporting diverse or previously excluded founder groups. They&#8217;re proving that women founders, founders of colour, founders outside of tech hotspots and elite university programmes, all make good bets.</p><h3><strong>So what now?</strong></h3><p>I'm not naive enough to think that individual mindset shifts will fix structural discrimination. The system needs to change. LPs need to ask harder questions about where fund money goes. The industry needs accountability that goes beyond diversity reports that nobody reads.</p><p>But here's what the last decade has taught us: waiting for VCs to become less biased hasn't worked. The number hasn't moved. And while the industry slowly figures out how to do better - if it ever does - individual women founders are walking into pitch meetings every day, absorbing risk that isn't accurately priced, feeling grateful for opportunities that should be seen as mutually-beneficial.</p><p><strong>So let me speak directly to you, if you're a woman founder reading this.</strong></p><p>We are not lucky to receive VC investment. If you are raising investment for your tech startup, you are offering someone the chance to profit from your risk, your expertise, your work. The data shows we&#8217;re likely to outperform our male counterparts. The data shows we&#8217;re doing it while being undercapitalised. The data shows the system is consistently undervaluing us.</p><p>That's not your problem to fix by being more grateful. But I do think knowing this, gives us leverage.</p><p>We need to start asking different questions in those pitch meetings.&nbsp;</p><p>Not just "will you invest?" but "what do you bring beyond capital?"&nbsp;</p><p>Not just "what are your terms?" but "what's your track record with women founders?&#8221; And &#8220;Do you support many second-time founders?&#8221;?"</p><p>If a firm can't answer, that tells us something.</p><p>VCs need our outsized conviction, our unsustainable hours, our willingness to bet everything on one outcome - because that's what makes their portfolio model work. Without founders who believe they're exceptional, the whole system falls apart.</p><p>As a woman founder, VCs should feel lucky we&#8217;re considering letting them invest.</p><p>What if we acted like it?</p><p></p><p></p><p>*<em>Sure, &#8216;not all VCs&#8217;. But most. </em></p><p><em>**Massive thanks to Paul Smith from Founders Circle, for explaining this so clearly. </em></p><p><strong>Footnotes:</strong></p><p>&#185; Hebert, C., Yimfor, E., &amp; Tookes, H. (2025). <em>Financing the Next VC-Backed Startup: The Role of Gender.</em> NBER Working Paper No. 33943. Available at:<a href="https://www.nber.org/papers/w33943"> https://www.nber.org/papers/w33943</a></p><p>&#178; Carta. (2025). <em>Founder Ownership Report 2025.</em> Available at:<a href="https://carta.com/data/founder-ownership/"> https://carta.com/data/founder-ownership/</a></p>]]></content:encoded></item><item><title><![CDATA[The new knowledge monopoly: how Silicon Valley took control of our thoughts]]></title><description><![CDATA[Thinking in public: working ideas out, out loud.]]></description><link>https://www.laurarichards.ai/p/the-new-knowledge-monopoly-how-silicon</link><guid isPermaLink="false">https://www.laurarichards.ai/p/the-new-knowledge-monopoly-how-silicon</guid><dc:creator><![CDATA[Laura Richards]]></dc:creator><pubDate>Sat, 04 Oct 2025 22:03:07 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/11ceac9f-fbf1-4957-b065-347df5aa0c81_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I&#8217;ve been thinking a lot recently about <em>power.</em> And more specifically, about the power that comes from controlling mass-adopted AI models. </p><p>A lot gets written about who controls AI: which governments are leading the AI arms race; which corporations are winning the talent wars; who&#8217;s controlling the world&#8217;s compute. And the battle for mainstream dominance is just as fierce. OpenAI reported last month that they&#8217;d hit 700million weekly active users and are on track for one billion users by the end of this year. For anyone not clear on the maths, that means by the end of this year, 1 out of every 8 people <em>in the world</em> could be using OpenAI&#8217;s large language models alone. </p><p><em><strong>But of course, that 1-in-8 figure isn&#8217;t distributed evenly&#8230;</strong></em></p><p>In a recent article for <a href="https://theconversation.com/how-generative-ai-is-really-changing-education-by-outsourcing-the-production-of-knowledge-to-big-tech-263160">The Conversation</a>, Dr. Kimberley Hardcastle discussed how AI (and large language models more specifically) are changing the way people study and learn. </p><blockquote><p>&#8220;[W]hat&#8217;s being overlooked is how evolving generative AI systems are fundamentally changing our relationship with knowledge itself: how we produce, understand and use knowledge.&#8221;</p></blockquote><p>The issue isn&#8217;t as simple (or easy to address) as &#8216;students are asking ChatGPT to do their homework&#8217;. Rather, the more we collectively turn to tools like ChatGPT to provide us with information or advice, the more we hand over control of how our knowledge comes to be, to a very select group of people. Dr. Hardcastle articulates this more eloquently than I could:</p><blockquote><p>&#8220;When we outsource thought unthinkingly to machines, we hand unprecedented power to shape knowledge to the technology companies developing this evolving technology.&#8221; </p></blockquote><p>And it&#8217;s not just students and the formal pursuit of knowledge where we are going to see this affect. Take, as an example, the world of <em>dating. </em>You might chuckle to yourself, but bear with me here.</p><p>Numerous studies and reports suggest that one of the most common uses of tools like ChatGPT is for something akin to therapy. People are turning to ChatGPT and its ilk for advice on just about everything, including relationships. Stats also show that: </p><p>a) Men are more likely to use ChatGPT than women;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> </p><p>b) Men are nearly 3x more likely to ask ChatGPT for dating advice<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>; and</p><p>c) Men are far less likely to turn to traditional therapy than women.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>And while this might sound positive at first read (&#8216;<em>great, guys are getting help somewhere, at least!&#8217;)</em> when you consider how large language models generate responses and the data on which they&#8217;ve been trained, it starts to raise some questions.</p><h3>How Large Language Models (LLMs) learn </h3><p>That all the major AI developers have trained their frontier models on <strong>all the text available on the internet</strong> is one of the world&#8217;s worst kept secrets. </p><p>At an event earlier this year on safe and trusted AI systems, I listened as an AI researcher from Meta talked about using 'anything you can find online&#8217; (I might be paraphrasing here, but not a lot&#8230;) during the pre-training stages of LLM development. It was such a cavalier comment and I don&#8217;t think anyone else even blinked. But the ease (bordering on excitement) with which this was discussed raised serious ethical and safeguarding questions for me.</p><p>Set aside, for a minute, issues of copyright. Just think what <em>text on the internet</em> might say about relationships (and in particular, about women). LLMs are usually built with safeguards in place to prevent problematic outputs, but these safeguards are being built by the same teams that scraped the internet for training data in the first place. </p><p>Of the billions of pieces of text (or tokens, if you want to get technical), how many might - if strung together in the most probabilistic way - present dating advice based on a warped view of relationship dynamics? And how easy is it for the very people most likely to receive this advice, to spot any bias, bent or bigotry, when it is framed in language that sounds oh-so-agreeable? Especially if they never engage with professional support, to have something to compare it to.</p><p>We, as users, have to trust that the teams building the models we use are both aware of these issues and are building appropriate guardrails to safeguard against them. But that trust has not yet been earned. </p><p>When teams of incredibly clever, highly paid engineers, are tasked with building AI models optimised for performance, what thought is given to the responsibility they have to ensure the outputs generated do not spread one particular worldview, set of ideologies, or rhetoric? Just from my interactions with that researcher from Meta, it is easy for me to imagine an environment where enthusiasm for the technical outweighs any concern for the ethical. </p><p><em>And that&#8217;s before we get onto potential scenarios where the figures leading these companies actively endorse specific ideologies and appear to be comfortable calling for <a href="https://www.independent.co.uk/news/world/americas/us-politics/elon-musk-hamish-steele-dead-end-b2838042.html">censorship of content</a> that does not fit with their worldview&#8230;</em> </p><h3><strong>Playing the thought experiment through to the end</strong></h3><p>Mass adoption of LLMs and GenAI, therefore, runs the risk that our very thoughts are mediated (at best; shaped at worst) by a handful of very powerful companies. Companies controlled by an even smaller group of powerful individuals. History itself could be written, without us - the public - ever being given a chance to collectively validate it.</p><p>Combine this with the fact that mass adoption of GenAI means more people are likely to outsource both <em>writing</em> and <em>reading</em> to machines. The process of reading and writing are vital for deep thinking, according to computer scientist and best-selling author Cal Newport (cited in <em>The End of Thinking</em> by <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Derek Thompson&quot;,&quot;id&quot;:157561,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!oFSS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ed4fc85-9214-4460-a3e7-c80fca4a3c3d_872x872.png&quot;,&quot;uuid&quot;:&quot;3a87f333-0879-49f7-b3d7-074cbaf12c47&quot;}" data-component-name="MentionToDOM"></span>). And while the productivity gains from AI tools are incredibly tempting*, never have we more needed a global population that are skilled at critical thinking. </p><p>This is where the tension sits. We&#8217;re trading away the very processes that enable us to think critically about what we&#8217;re consuming. Every time we let an AI summarise an article, write an email, or provide advice, we&#8217;re not just saving time - we&#8217;re ceding a tiny bit of our intellectual autonomy to companies whose motivations we can only guess at.</p><p>When those models are built by teams who treat the entire internet as fair game for training data, we&#8217;re letting a very specific subset of people &#8211; with their own blind spots, biases, and business pressures &#8211; become the architects of collective knowledge.</p><p>So what do we do with this? I&#8217;m not suggesting we abandon these tools entirely (that would make me a hypocrite). But perhaps we need to start treating our interactions with AI less like using a calculator and more like consulting an advisor whose credentials we haven&#8217;t properly vetted. One who might have read every toxic Reddit thread, every piece of misinformation, and every bit of propaganda ever posted online - and is now using that to influence how we think about everything from our relationships to our reality.</p><p>That 1-in-8 of the world&#8217;s citizens use AI, should give us leverage. We must demand more from those making decisions about AI - especially when it comes to how AI creates meaning. We should be entitled to more transparency about how results are generated and what data they have been trained on. And in an ideal world we would have more choice, not less, about the models we use and the companies we patron. </p><p>As always, I come back to the fundamental question of <em>who&#8217;s in the room</em> when these decisions are being made. We don&#8217;t need more over-eager engineers focused solely on model optimisation; we need voices that represent all of us, asking for the caution and consideration we deserve. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If this made you think, I&#8217;d love to keep the conversation going. Subscribe for more essays on AI, power, and society.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/p/the-new-knowledge-monopoly-how-silicon?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">If this made you think, please share it with someone you know. We need more people asking the hard questions about AI. </p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/p/the-new-knowledge-monopoly-how-silicon?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.laurarichards.ai/p/the-new-knowledge-monopoly-how-silicon?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p></p><div><hr></div><p><strong>Got you thinking? Check out the articles that influenced me:</strong></p><p><a href="https://theconversation.com/how-generative-ai-is-really-changing-education-by-outsourcing-the-production-of-knowledge-to-big-tech-263160?utm_medium=article_clipboard_share&amp;utm_source=theconversation.com">How generative AI is really changing education &#8211; by outsourcing the production of knowledge to big tech</a> - Dr. Kimberley Hardcastle in The Conversation </p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:174284835,&quot;url&quot;:&quot;https://www.derekthompson.org/p/the-end-of-thinking&quot;,&quot;publication_id&quot;:2880588,&quot;publication_name&quot;:&quot;Derek Thompson&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!uPIO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b0f850-caa7-417a-bc0b-5b7224dd1f25_888x888.png&quot;,&quot;title&quot;:&quot;The End of Thinking&quot;,&quot;truncated_body_text&quot;:&quot;This is an expanded and revised version of an essay that originally ran in The Argument, an online magazine where I am a contributing writer.&quot;,&quot;date&quot;:&quot;2025-09-24T10:03:40.931Z&quot;,&quot;like_count&quot;:1127,&quot;comment_count&quot;:101,&quot;bylines&quot;:[{&quot;id&quot;:157561,&quot;name&quot;:&quot;Derek Thompson&quot;,&quot;handle&quot;:&quot;derekthompson&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!oFSS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ed4fc85-9214-4460-a3e7-c80fca4a3c3d_872x872.png&quot;,&quot;bio&quot;:&quot;Abundance and other ideas to make the world a better place&quot;,&quot;profile_set_up_at&quot;:&quot;2021-10-25T17:19:21.553Z&quot;,&quot;reader_installed_at&quot;:&quot;2022-03-09T16:22:19.302Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:2928158,&quot;user_id&quot;:157561,&quot;publication_id&quot;:2880588,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:true,&quot;publication&quot;:{&quot;id&quot;:2880588,&quot;name&quot;:&quot;Derek Thompson&quot;,&quot;subdomain&quot;:&quot;derekthompson&quot;,&quot;custom_domain&quot;:&quot;www.derekthompson.org&quot;,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;A newsletter about abundance and building a better world.&quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/38b0f850-caa7-417a-bc0b-5b7224dd1f25_888x888.png&quot;,&quot;author_id&quot;:157561,&quot;primary_user_id&quot;:157561,&quot;theme_var_background_pop&quot;:&quot;#FF6719&quot;,&quot;created_at&quot;:&quot;2024-08-13T01:26:09.408Z&quot;,&quot;email_from_name&quot;:&quot;Derek Thompson&quot;,&quot;copyright&quot;:&quot;Derek Thompson&quot;,&quot;founding_plan_name&quot;:&quot;Superfan Tier&quot;,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;enabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;homepage_type&quot;:&quot;magaziney&quot;,&quot;is_personal_mode&quot;:false}}],&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:1000,&quot;status&quot;:{&quot;bestsellerTier&quot;:1000,&quot;subscriberTier&quot;:1,&quot;leaderboard&quot;:null,&quot;vip&quot;:false,&quot;badge&quot;:{&quot;type&quot;:&quot;bestseller&quot;,&quot;tier&quot;:1000},&quot;paidPublicationIds&quot;:[656797,159185]}}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;,&quot;source&quot;:null}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://www.derekthompson.org/p/the-end-of-thinking?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!uPIO!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38b0f850-caa7-417a-bc0b-5b7224dd1f25_888x888.png" loading="lazy"><span class="embedded-post-publication-name">Derek Thompson</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">The End of Thinking</div></div><div class="embedded-post-body">This is an expanded and revised version of an essay that originally ran in The Argument, an online magazine where I am a contributing writer&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">8 months ago &#183; 1127 likes &#183; 101 comments &#183; Derek Thompson</div></a></div><div><hr></div><p>*Full disclosure, I actively use AI to both assist with <em>some </em>writing tasks, and to summarise <em>some</em> materials for me to read.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p><em>https://techcrunch.com/2025/01/29/chatgpts-mobile-users-are-85-male-report-says/ </em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p><em>https://www.9news.com.au/national/artificial-intelligence-dating-advice-young-people-are-turning-to-chatgpt-for-dating-advice-so-we-put-the-ai-to-the-test/973b35ec-3dd8-48e3-94ac-c002fa31c44a</em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p><em>https://www.mentalhealth.org.uk/explore-mental-health/statistics/men-women-statistics</em></p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[A new beginning (of sorts)]]></title><description><![CDATA[WTF are Frontier Philosophies, anyway?]]></description><link>https://www.laurarichards.ai/p/a-new-beginning-of-sorts</link><guid isPermaLink="false">https://www.laurarichards.ai/p/a-new-beginning-of-sorts</guid><dc:creator><![CDATA[Laura Richards]]></dc:creator><pubDate>Mon, 29 Sep 2025 11:40:02 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1be70a87-a6a3-4cfb-85a4-022bb33cdcb9_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When I first started writing on Substack, I thought I wanted to explore business goals, habits, personal development. Things that 2ish years ago were weighing on my mind. </p><p>Today, I feel compelled to write about something different. </p><p>Over the past couple of years, as I&#8217;ve increasingly immersed myself in learning about and developing artificial intelligence, I&#8217;ve found myself circling the same frustration: the conversation around AI feels too narrow. Technical roadmaps. Corporate speak. Big declarations about &#8220;the future of work&#8221; that rarely seem to reflect the whole of society. </p><p>Again and again, I&#8217;ve gone looking for the intellectual work that should sit alongside all this - the thinking, theories and philosophical ideas we need to successfully navigate the shifts in power, society and culture we&#8217;re seeing as a result of AI. But I&#8217;ve come up short. Even where there are compelling academic studies on ethical or responsible AI, they largely remain within academia; discussed at symposiums but rarely reaching the private developers, policymakers or joe public that need to hear them the most.  </p><p>So I decided to start exploring this myself. <em>Frontier Philosophies</em> are needed for society to stand a chance against frontier AI models. This publication is where I&#8217;ll explore the ideas, theories, and questions that are keeping me up at night. My hope is to spark discussion, not just for technologists or policymakers, but for all of us.</p><p>This is how I&#8217;ll explore that work here:</p><ul><li><p><strong>The nerdy ones</strong> - deep dives into the technical edges of AI. The developments that are exciting me across distributed AI, data sovereignty, small language models, tinyML and more.</p></li><li><p><strong>Thought experiments</strong> - cross-domain explorations, pulling ideas from philosophy, economics, or culture. Sometimes about AI, always about what it fundamentally means to be human. </p></li><li><p><strong>Thinking in public</strong> - rough sketches and half-formed takes, working ideas out on the page. I wholeheartedly reserve the right to change my mind completely on anything I posit. </p></li><li><p><strong>The new commons</strong> - essays on equity, ownership, and governance. These are the articles where I&#8217;ll be exploring the intersection of AI, power and society. </p></li></ul><p>Some essays will be sharp and technical, some broad and speculative, some unfinished but honest. My hope is that over time they&#8217;ll map the <em>Frontier Philosophies</em> we&#8217;ll need if AI is to serve everyone, not just its makers.</p><p>I&#8217;ll publish about twice a month. If you&#8217;ve been here from the start, thank you for staying with me as my writing evolves. If you&#8217;re new, welcome! Subscribe to get each essay direct to your inbox  and be part of a growing community that believes AI&#8217;s future is too important to leave in the hands of a few. </p><p>- Laura</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.laurarichards.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Like what you&#8217;re reading? Subscribe for more essays at the edge of AI, technology, and society. Twice a month, straight to your inbox.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>