<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Stelia AI]]></title><description><![CDATA[The distributed intelligence platform that transforms AI concepts into global applications. ]]></description><link>https://stelia.substack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!neo5!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74b9145f-82b9-40c5-b831-af45fe27540a_3342x3342.png</url><title>Stelia AI</title><link>https://stelia.substack.com</link></image><generator>Substack</generator><lastBuildDate>Mon, 06 Apr 2026 14:13:55 GMT</lastBuildDate><atom:link href="https://stelia.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Stelia]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[stelia@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[stelia@substack.com]]></itunes:email><itunes:name><![CDATA[Stelia]]></itunes:name></itunes:owner><itunes:author><![CDATA[Stelia]]></itunes:author><googleplay:owner><![CDATA[stelia@substack.com]]></googleplay:owner><googleplay:email><![CDATA[stelia@substack.com]]></googleplay:email><googleplay:author><![CDATA[Stelia]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[ChatGPT and the trojan horse of agentic commerce]]></title><description><![CDATA[OpenAI&#8217;s new Instant Checkout looks like user convenience, but the deeper play is laying rails for agent-to-agent commerce at planetary scale.]]></description><link>https://stelia.substack.com/p/chatgpt-and-the-trojan-horse-of-agentic</link><guid isPermaLink="false">https://stelia.substack.com/p/chatgpt-and-the-trojan-horse-of-agentic</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Wed, 01 Oct 2025 07:30:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!l4Y4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!l4Y4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!l4Y4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!l4Y4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!l4Y4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!l4Y4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!l4Y4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5002407,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/174942635?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!l4Y4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!l4Y4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!l4Y4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!l4Y4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64a9a698-764b-4c68-a252-9c29bcb425b2_3840x2160.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>OpenAI framed ChatGPT&#8217;s Instant Checkout as obvious: more than 700 million people use ChatGPT each week for everyday tasks, many of them shopping queries. Now, the assistant can help you buy directly. Powered by the Agentic Commerce Protocol (ACP), co-developed with Stripe, the feature debuts with U.S. Etsy sellers and soon over a million Shopify merchants, from Glossier to SKIMS.</p><p>The pitch is user simplicity: describe a product, see the best matches, tap to buy without leaving chat. For merchants, the promise is reach and relevance, with ACP making integration lightweight. For users, the experience is free, secure, and ranked purely on relevance, not fees.</p><p>Many conversational shopping assistants, comparison bots, plugin SDKs, and browser tools have become endangered species within the last 24 hours. The moment OpenAI can go from &#8220;help me find&#8221; to &#8220;help me buy&#8221; inside chat is a major shift in funnel control &#8211; those that don&#8217;t pivot quickly get squeezed.</p><p>And will retailers and brands quickly adapt to the new reality of yet-another-channel in a multichannel world? Or will they forget the lessons of mobile web and app adoption and take years to catch up with consumer preferences?</p><h2>In vitro</h2><p>2024&#8211;25 studies demonstrated that conversational AI is not just feasible but advantageous in e-commerce.</p><ul><li><p>Graph-enhanced retrieval raised factual accuracy by 23% and improved first-contact resolution by 28%.</p></li><li><p>LLM-based assistants like Flippi reached &gt;90% accuracy across 1.3m users, with measurable conversion uplift.</p></li><li><p>User research shows AI-driven shopping doubles repeat purchase intent by reducing decision fatigue.</p></li></ul><h2>In vivo</h2><p>The market followed: conversational AI commerce is projected at $290B for 2025, with generative AI referrals to retail sites up 4,700% year-on-year. Instant Checkout brings those findings to market scale. Key moves include:</p><ul><li><p>Coverage: Etsy live today, Shopify&#8217;s million-plus merchants next.</p></li><li><p>Capability roadmap: single-item purchases now; multi-item carts, more geographies, and more processors coming.</p></li><li><p>Open standard: ACP is open-sourced, designed to work across processors via Stripe&#8217;s Shared Payment Token API or ACP&#8217;s Delegated Payments Spec.</p></li><li><p>Merchant control: sellers remain merchants of record; they handle payments, fulfilment, returns, and support.</p></li><li><p>Trust guarantees: encrypted tokens, minimal data sharing, explicit user confirmation on every step.</p></li></ul><p>The effect is a closed loop: discovery, recommendation, and purchase inside the assistant, with governance claims embedded at protocol level.</p><h2>Trojan horse</h2><p>Behind the convenience is strategic leverage. To compete in this new channel, merchants must structure their data, expose APIs, and otherwise raise their AI game. In effect, OpenAI is inducing the retail ecosystem to modernise.</p><p>This dovetails with NVIDIA&#8217;s ambition: every upgrade to personalisation, inventory prediction, or recommendation is another GPU workload. ChatGPT&#8217;s checkout feature is thus a Trojan Horse: pulling enterprises into deeper AI adoption under the banner of consumer convenience.</p><h2>From human-in-loop to A2A</h2><p>Looking out to the near future we can see a 3 step strategy unfolding:</p><ol><li><p><strong>Human-in-loop commerce</strong> where users confirm each purchase.</p></li><li><p><strong>Policy-guided automation</strong> for example &#8220;auto-reorder under $30,&#8221; &#8220;book flights under $500.&#8221;</p></li><li><p><strong>Full agent-to-agent commerce</strong> when assistants transact directly with merchant agents under policy and audit, not per-click consent.</p></li></ol><p>ACP is architected for all three. Today it enforces human confirmation. Tomorrow it enables autonomous negotiation between agents.</p><p>This isn&#8217;t just speculation. Early frameworks are emerging that attempt to map what a genuine agent-to-agent economy looks like, distinguishing today&#8217;s AI-assisted checkout from fully autonomous A2A transactions. <a href="https://stelia.ai/deep-research/project-kahneman">Work from Stelia</a> points to the need for a holistic approach: defining taxonomies of agents, building economic and legal architectures, and stress-testing failure modes with scenario modeling. The research agenda spans technical feasibility, market mechanisms, governance, and social impacts out to 2035. In that context, OpenAI&#8217;s Instant Checkout looks less like a convenience feature and more like the first policy-constrained implementation on a much longer trajectory.</p><h2>Platform gravity</h2><p>As shopping, payments, and conversation converge, workflow gravity increases. Just as GPT-5&#8217;s tool use will increasingly pull SaaS workflows inside the model, commerce flows will settle in ChatGPT. Control shifts to the platform that captures intent and closes the sale in a single chat window.</p><h2>Safety that travels well</h2><p>OpenAI&#8217;s PR emphasises trust: minimal data sharing, explicit confirmation, secure tokens. But as ACP scales to A2A, safety must expand: protocol-level constraints on agent behavior, auditable logs, and jurisdiction-aware compliance. Without verifiable safety, merchants and regulators will resist deeper automation.</p><h2>Orchestration</h2><p>Global vertical orchestration platforms such as Stelia are becoming the benchmark substrate beneath this shift. They orchestrate &#8220;ground to prompt&#8221;: racks, network kernels, runtimes, and interfaces. In agentic commerce, these platforms guarantee:</p><ul><li><p>Latency low enough for real-time checkout.</p></li><li><p>Compliance with payment and data laws across regions.</p></li><li><p>Resilience across multiple processors.</p></li><li><p>Auditability from GPU kernel to consumer receipt.</p></li></ul><p>They will be the execution fabric for agentic commerce at planetary scale.</p><h2>Strategic, not opportunistic</h2><p>On the surface, Instant Checkout is a new feature and revenue stream. In substance, it is a deliberate long-term bet:</p><ul><li><p>Capture intent at the source.</p></li><li><p>Normalise ACP as the open transaction rail.</p></li><li><p>Pressure merchants to upgrade their AI maturity.</p></li><li><p>Indirectly advance NVIDIA&#8217;s goal of AI saturation across industries.</p></li></ul><p>Convenience is the Trojan horse. Control of the rails is the strategy.</p><p><em>By Paul Morrison, Chief Experience Officer, <a href="http://stelia.ai">Stelia</a></em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[GPT-5 and the coordination problem]]></title><description><![CDATA[As GPT-5 moves into autonomous tool use, the focus turns to governance, brand integrity, and safety in a world of tool-driven AI workflows.]]></description><link>https://stelia.substack.com/p/gpt-5-and-the-coordination-problem</link><guid isPermaLink="false">https://stelia.substack.com/p/gpt-5-and-the-coordination-problem</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Tue, 19 Aug 2025 11:31:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Wh34!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Wh34!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Wh34!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Wh34!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Wh34!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Wh34!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Wh34!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5216182,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/171358350?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Wh34!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Wh34!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Wh34!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Wh34!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27fab18b-4bf4-40cf-842f-a07277abf3f4_3840x2160.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Sam Altman opened the recent launch with a simple arc. Whilst GPT-3 felt like a bright sixth former and GPT-4o like a capable undergraduate. GPT-5, he said, behaves like a PhD-level expert on demand. He also said 700 million people now use ChatGPT each week. That scale changes the product story. It also changes the governance story.</p><p>What follows looks from the Stelia perspective at what gets built, who holds leverage, and which architectural choices age well when intelligence starts to act through tools as much as words.</p><h2><strong>SaaS &#8211; canary in the coal mine</strong></h2><p>For two decades, SaaS companies froze workflows into persistent products. GPT-5 shows a different rhythm. In the launch demos, an agent turned a plain-language request into an interactive aerodynamics simulator in minutes. It produced hundreds of lines of running code with no IDE in sight. The SaaS example illustrates how a fixed product can be replaced by a temporary configuration that vanishes once the job is done. However, the more important shift is the growth of tool use.</p><h2><strong>GPT5 edge: autonomous toolchains</strong></h2><p>GPT-5 is producing even more human-like text, yet it is inow also initiating multi-step processes.</p><ul><li><p>Opening files, parsing and rewriting them.</p></li><li><p>Calling APIs with structured parameters.</p></li><li><p>Searching, filtering and reasoning across results.</p></li><li><p>Handing output to other tools without manual intervention.</p></li></ul><p>The demos showed it working for minutes at a time, chaining tools and functions into coherent workflows. It can run the equivalent of an internal build system, a customer research desk, or a legal discovery process. All from a prompt. This level of autonomy moves well beyond the earlier &#8220;search and summarise&#8221; agents.</p><p>Strategically, the point of control has shifted to the interface between the model and its tools. That is the new coordination substrate.</p><h2><strong>Brand in a tool-driven world</strong></h2><p>When models and connected tools operate other systems, brand becomes a claim about execution integrity.<br><br>Enterprises will ask:</p><ol><li><p>Will this agent only call approved tools, in the way I approved?</p></li><li><p>Will it respect the jurisdictional and contractual constraints I operate under?</p></li><li><p>Can I verify and audit what it did, step by step?</p></li></ol><p>Brand in this context is built on governance primitives. These include policy as code, verifiable attestations for every tool call, and cultural-intelligence models that adapt without human babysitting.</p><h2><strong>Platform gravity</strong></h2><p>GPT-5&#8217;s integration with Gmail, Calendar, Python data analysis, and image generation widens the set of jobs that can be completed without leaving the platform. When tools, memory, and history all live together, workflows settle in and stay. Controlling that space is controlling the habit, hence the near-daily cadence of model upgrades and habit-forming feature enhancements.</p><h2><strong>Safety that travels well</strong></h2><p>Once a model can act through tools, the potential blast radius grows. Guardrails must move from output filtering to protocol-level constraints on tool invocation and data flow. Post-hoc review helps but cannot cover long-running or multi-system processes.</p><p>Safety here means:</p><ul><li><p>Pre-deployment ethics checks on all default tools.</p></li><li><p>Context-aware limits on tool combinations.</p></li><li><p>Transparent logging so actions can be replayed and audited.</p></li></ul><p>The coordination challenge is making these properties travel with the task, across every tool it touches.</p><h2><strong>Costs, compute</strong></h2><p>Like increasingly complex SaaS led to inevitably opaque billing, tool-driven workflows are not free. Each API call, file parse, and image render adds latency and cost. The platforms that endure will be the ones that manage predictable economics without hollowing out control.</p><p>That points to three design moves:</p><ol><li><p>Structure reasoning so outputs are easy to constrain and validate.</p></li><li><p>Shift suitable steps to the edge to cut round trips.</p></li><li><p>Cache where privacy allows.</p></li></ol><h2><strong>Planet-scale coordination</strong></h2><p>As tool use scales, so does the need for responsible coordination. A scheduling agent will email your client. A coding agent will merge to main. A finance</p><p> agent will touch a ledger. These scenarios are already in use.</p><p>The substrate, such as Stelia, keeping these agents at a predictable cost, lawful, safe, and interoperable across borders is not optional. It is the foundation for an intelligence layer that can operate anywhere without eroding human dignity or autonomy.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Who’s running the show? Inside entertainment’s AI power shift]]></title><description><![CDATA[The entertainment industry is undergoing a full-scale systemic transformation, driven by AI-powered orchestration, emotional insight, and real-time data.]]></description><link>https://stelia.substack.com/p/whos-running-the-show-inside-entertainments</link><guid isPermaLink="false">https://stelia.substack.com/p/whos-running-the-show-inside-entertainments</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Wed, 06 Aug 2025 09:00:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!qHc6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qHc6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qHc6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic 424w, https://substackcdn.com/image/fetch/$s_!qHc6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic 848w, https://substackcdn.com/image/fetch/$s_!qHc6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic 1272w, https://substackcdn.com/image/fetch/$s_!qHc6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qHc6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:745529,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/170249001?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qHc6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic 424w, https://substackcdn.com/image/fetch/$s_!qHc6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic 848w, https://substackcdn.com/image/fetch/$s_!qHc6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic 1272w, https://substackcdn.com/image/fetch/$s_!qHc6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e4df9a5-c3b7-4758-bdab-f388b973e9a8_3840x2160.heic 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In my last article, I explored the orchestration challenges facing modern media companies. We looked at how global workflows and fragmented toolchains are forcing publishers and broadcasters to rethink how they produce, localise, and distribute content at scale. Now, as I meet with partners and clients across the entertainment industry in New York, through meetings, studio visits, and events like the Monks happy hour, I am seeing the shift move further. What began as media transformation is rapidly becoming a full-scale restructuring of entertainment. And this time, the change is not coming from inside the studio walls. It is being shaped at the system level.</p><p>The world of entertainment used to be studio driven, rooted in the comfort of predictable, linear consumption. Now, it is being driven by behaviour, emotion and cultural dynamics that change by the hour. The old question: what are we producing is being replaced by a more urgent one: how are we adapting what we produce to real-time attention? That question is reshaping production cycles, marketing strategy, rights management, and even how executives evaluate ROI. The infrastructure of entertainment, not just the content itself, is being rebuilt.</p><h2><strong>Netflix creative scale as competitive edge</strong></h2><p>A recent case from Netflix highlights how different that infrastructure now looks. In their new Argentinian sci-fi series The Eternaut, Netflix used AI-powered VFX through Eyeline Studios to create a building collapse scene that would have been prohibitively expensive to film traditionally. According to Co-CEO Ted Sarandos, the sequence was delivered ten times faster and at a significantly lower cost. The results speak for themselves. In the same quarter, Netflix posted over 11 billion dollars in revenue, up 16% year over year. More than visual effects, this lesson is about how infrastructure enables creative scale. Netflix&#8217;s investment in real-time, AI-enhanced tools is allowing them to tell stories they otherwise could not afford and to do so with local teams producing globally resonant results. This infrastructure-first approach to creative scale is now spreading across the industry, but it requires new ways of measuring what actually works.</p><h2><strong>What MediaScience taught me about emotion</strong></h2><p>Understanding what works means measuring real human response. Last week I visited MediaScience&#8217;s lab in Manhattan, where they test content by measuring real human response eye movement, facial expression, electrodermal activity, and heart rate. Their research shows that attention is shaped in the first seconds and influenced by emotion, context, and platform. I watched how small creative choices &#8211; a change in music, a tighter edit &#8211; triggered measurable shifts in engagement. Studios are already using it to green-light projects, fine-tune trailers, and adjust platform recommendations. MediaScience found that adapting stories based on behavioural signals can boost viewer engagement by over 30%.</p><p>What struck me most is how context and psychophysiological data has become a strategic input. Rather than focusing solely on views or completion, the emphasis has shifted towards understanding and quantifying emotional context. MediaScience is measuring what the viewer feels, right now, and how content responds. This is the new layer of intelligence shaping how entertainment is made and delivered. Phillip Lomax, MediaScience EVP, explained that context is shaped by four pillars, Audiences, Platforms, Content, and Ad Formats, and how MediaScience has gathered significant intelligence across these subjects which have been leveraged by their clients at Disney, Comcast, Google, Amazon, Netflix and many others.</p><h2><strong>Why platforms like Tubi are winning</strong></h2><p>That emotional intelligence MediaScience measures is exactly what&#8217;s driving the fastest-growing platforms. During my meetings here in New York, it has become obvious that many of these platforms are not focused on subscribers alone. They are building ecosystems around engagement. Tubi, for example, recently surpassed 100 million monthly active users, with over a billion hours of content watched in a single month. These are ad-supported businesses with no subscription revenue. Instead of payment, their metric of success is attention, relevance and repeat engagement. And they are winning by making their AI-enabled infrastructure flexible enough to adapt to viewer behaviour in real time. That adaptability is increasingly powered by AI systems that interpret how, when, and why people engage.</p><p>In this new wave of entertainment, the real competition isn&#8217;t just between streamers, it&#8217;s between formats. Premium, paid experiences on one side; free, algorithm-driven discovery on the other. And the future likely sits somewhere in between. The next generation of platforms is betting on frictionless, culturally-tuned content that feels as easy and diverse as scrolling TikTok, while still delivering the narrative pull of long-form entertainment. There&#8217;s also a shift in mindset: from producing more to producing what connects. Experiments are becoming institutional like studios that green-light based on fan input, or libraries that blend creator-led stories with licensed hits. What&#8217;s emerging is a new hybrid model where the old Hollywood playbook is rewritten for a leaner, faster, more emotionally aware viewer economy.</p><h2><strong>Studios focus on convergence, not channels</strong></h2><p>That same mindset is shaping strategy at the studio level. With linear viewership down and theatrical releases under pressure, the big players are thinking convergence. Disney is in the final stages of consolidating its operations across Hulu, ESPN, and Disney+ into a single unified offering making Disney+ its one-stop shop in streaming. The same time Netflix is investing heavily in music with reports of deeper partnerships spanning video, music, and live events. The experience is now the differentiator. And experiences require infrastructure that can handle rights, regionalisation, creative variation, billing, and measurement all in one real-time loop.</p><h2><strong>Fox moves toward one spine</strong></h2><p>FOX is also embracing both ends of the format spectrum. On one side is Tubi, their ad-supported platform built for scale and discovery. On the other is FOX One, a premium, subscription-based service launching August 21. Meanwhile, their internal OneFOX initiative unifies content, advertising, and metadata into a single infrastructure layer. When companies of this size start operating from a single spine, it changes not just internal speed. It changes what becomes possible for audiences and advertisers alike. Together, these moves reflect FOX&#8217;s twin approach to market: accessible at scale, integrated at depth.</p><h2><strong>When AI orchestration works</strong></h2><p>Yet most companies remain stuck in what I call the pilot trap. They are experimenting with artificial intelligence tools here and there, layering plugins or point solutions into isolated workflows. But they are not integrating. They are not feeding data back into their systems. And as a result, they are not scaling. True orchestration does not come from tool trials. It comes from connected systems that allow real-time response to audience behaviour, creative feedback loops, and operational decision-making.</p><p>You can see what this looks like when it works. In June, Netflix released the animated film KPop Demon Hunters. The movie was watched over one hundred thirty million times within weeks, and its soundtrack dominated the charts across the United States and Asia. Fan art began circulating within days. Cosplay, TikTok remixes, and micro-fandoms exploded globally. What made the real difference was the way AI models tracked and adapted in real time to audience engagement, shaping marketing and creative simultaneously. Much of that responsiveness came from AI models tracking sentiment and engagement signals across platforms.</p><h2><strong>From creative execution to strategic infrastructure</strong></h2><p>These successes are forcing teams across the industry to ask serious questions. How do we adapt narratives across cultures without reshooting entire scenes? How do we automate marketing assets in multiple languages within hours? How do we ensure our rights logic and licensing structures do not become bottlenecks to creative agility? These strategic questions are the ones shaping who moves forward and who gets left behind.</p><h2><strong>What the Paramount-Skydance deal reveals</strong></h2><p>The newly sealed merger between Paramount and Skydance confirms precisely this shift. It brings together Hollywood legacy and Silicon Valley agility, creating a powerhouse built for an AI-shaped era where flexibility and infrastructure matter as much as storytelling. Editorial integrity, platform integration, and operational scalability are now board-level priorities. For the industry, it indicates a new phase - one where collaboration is creative and systemic.</p><h2><strong>Programming for cultural moments</strong></h2><p>During one of my conversations this week, an executive said something that stuck with me. He said, we are no longer in the business of programming for time slots. We are programming for cultural moments. That line captures exactly where entertainment is heading. The winners will not be those who simply produce more content. It will be those who can connect that content to emotion, context, and culture &#8211; fast, at scale, and across borders.</p><p>Whether it's a media agency like Havas aligning production and performance, or a contextual tech company like GumGum measuring attention with AI-powered contextual signals, the lines between storytelling, infrastructure, and intelligence are blurring fast.</p><h2><strong>Entertainment&#8217;s new operating system</strong></h2><p>Across all of these examples from Netflix&#8217;s AI-powered production to Tubi&#8217;s adaptive engagement engine and GumGum&#8217;s contextual intelligence, a new layer of AI infrastructure is quietly becoming the foundation of modern entertainment. It&#8217;s not just content being transformed, but the entire operating system behind it: one powered by intelligent agents, audience signals, and dynamic orchestration.</p><p>I am seeing this first hand across every conversation I am having now in New York. Whether I am speaking with creative studios, rights managers, marketing leaders, or platform heads, the same trend keeps emerging. The entertainment industry is undergoing a complete restructure. The systems behind the scenes, the tools, workflows, orchestration engines, and attention models, are now what make or break success.</p><p>It is no longer enough to have good content. You need good systems. You need the ability to operate across markets, adapt in real time, personalise without compromising quality, and measure value not just in views but in cultural impact. The companies that build this AI-native infrastructure now will not only lead in entertainment. They will define how entertainment itself is experienced in the years to come.</p><p><em>by Ula Nairne, VP Media &amp; Entertainment, Stelia</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[How AI satellites prevent flood and wildfire disasters]]></title><description><![CDATA[Stelia shows how AI satellite constellations cut disaster costs with real-time alerts and forecasting for governments, insurers, and enterprises.]]></description><link>https://stelia.substack.com/p/how-ai-satellites-prevent-flood-and-ef9</link><guid isPermaLink="false">https://stelia.substack.com/p/how-ai-satellites-prevent-flood-and-ef9</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Thu, 31 Jul 2025 07:30:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_8be!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_8be!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_8be!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic 424w, https://substackcdn.com/image/fetch/$s_!_8be!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic 848w, https://substackcdn.com/image/fetch/$s_!_8be!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic 1272w, https://substackcdn.com/image/fetch/$s_!_8be!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_8be!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1405511,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/169667149?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_8be!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic 424w, https://substackcdn.com/image/fetch/$s_!_8be!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic 848w, https://substackcdn.com/image/fetch/$s_!_8be!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic 1272w, https://substackcdn.com/image/fetch/$s_!_8be!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff77453d0-3a30-4f2e-aa3b-0b2851357ed4_3840x2160.heic 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>Climate disaster costs</strong></h2><p>In&#8239;2024, weather&#8209;driven catastrophes wiped out <strong>US&#8239;$320&#8239;billion</strong> in global wealth, more than the GDP of Greece, with floods and wildfires leading the tally.<sup>1</sup> Insured losses have risen about <strong>7&#8239;% each year since&#8239;2015</strong>, outpacing world GDP. The planet is paying a premium for information it still receives too late.</p><h2><strong>Defences falling behind</strong></h2><p>Warmer oceans, erratic rainfall and expanding wild&#8209;land-urban interfaces have doubled flood&#8209; and fire&#8209;related losses over the past decade. Policymakers are scrambling: the UN&#8209;backed <strong>Early&#8239;Warnings&#8239;for&#8239;All</strong> initiative pledges universal, life&#8209;saving alerts by&#8239;the end of 2027.<sup>2</sup> Capital is flowing the same way - venture funding for AI&#8209;powered Earth&#8209;observation (EO) start&#8209;ups jumped again in 2024 because early, trusted intelligence is fast becoming critical infrastructure.</p><h2><strong>Satellite breakthroughs</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!oxfk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!oxfk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic 424w, https://substackcdn.com/image/fetch/$s_!oxfk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic 848w, https://substackcdn.com/image/fetch/$s_!oxfk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic 1272w, https://substackcdn.com/image/fetch/$s_!oxfk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!oxfk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic" width="1217" height="562" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:562,&quot;width&quot;:1217,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:57105,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/169667149?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!oxfk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic 424w, https://substackcdn.com/image/fetch/$s_!oxfk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic 848w, https://substackcdn.com/image/fetch/$s_!oxfk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic 1272w, https://substackcdn.com/image/fetch/$s_!oxfk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe21c7c-4d83-4221-a132-bfe05e0d7ba9_1217x562.heic 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>Why only AI can keep up</strong></h2><p>EO satellites now pump out <strong>hundreds of terabytes of imagery every day</strong>, far beyond what any analyst corps or rule&#8209;based software can triage in real time. Deep&#8209;learning pipelines such as ICEYE&#8217;s convolutional flood classifiers and OroraTech&#8217;s FireDetect&#8482; transformer sift noise from signal in milliseconds, flagging a 25&#8239;m&#178; hotspot before smoke is visible and estimating flood depth to actionable precision. Classical post&#8209;processing would take days, rendering the UN&#8217;s early&#8209;warning mandate impossible. <strong>Without AI, revolutionary hardware would decay into an unread archive.</strong></p><h2><strong>Real&#8209;time decisions</strong></h2><p><strong>Governments</strong> can evacuate earlier and allocate crews with surgical precision. Google&#8217;s Flood Hub and related early warning initiatives in India have delivered flood alerts to millions of people, improving local preparedness and response.<sup>5</sup></p><p><strong>Insurance:</strong><br>Insurers are increasingly using rapid-impact satellite maps, such as ICEYE flood data, to develop parametric insurance policies that can speed up claims settlements. MAPFRE RE&#8217;s agreement to license ICEYE flood data is intended to accelerate claims workflows and reduce settlement times. <sup>6</sup></p><p><strong>Enterprise:</strong><br>Timber firms and utilities are using near-real-time satellite analytics to improve environmental risk management and demonstrate ESG performance. These tools support efforts to track burn areas and optimise operations in regions such as Bavaria.</p><p><strong>Society at large</strong> wins most. When Rio&#8239;Grande&#8239;do&#8239;Sul flooded in&#8239;2025, SAR maps let officials funnel aid to vulnerable families within days, proof that fast data can translate directly into faster relief.<sup>7</sup></p><h2><strong>Market and ethics</strong></h2><p>ICEYE fields the largest commercial SAR fleet (48 satellites today; 54 planned by year&#8209;end). OroraTech is sprinting toward a dedicated, always&#8209;on thermal network. Planet and Maxar keep optical dominance; Capella edges up on high&#8209;res SAR. Yet speed alone is not a moat; the real edge lies in <strong>algorithms that convert raw pixels into trusted decisions in minutes</strong>.</p><p>Acceleration raises new questions. Models can under&#8209;predict in data&#8209;poor regions, and cross&#8209;border imagery still triggers sovereignty debates. Transparent accuracy scores, bias audits and privacy controls must travel with every API call if the sector is to keep public trust.</p><h2><strong>The coordination layer</strong></h2><p>Insight becomes impact only when thousands of dispersed actors share the same picture fast enough to act. Emerging coordination platforms, including Stelia, ingest multi&#8209;sensor feeds and surface a single operational view to responders and insurers alike. The result is a shift from reactive recovery to proactive resilience, delivered over secure, enterprise&#8209;grade rails.</p><h2><strong>Digital&#8209;twin forecasting</strong></h2><p>The next leap is already taking shape. Europe&#8217;s <strong>Destination&#8239;Earth</strong> programme is fusing satellite streams, super&#8209;computing and AI to build a live digital twin of the planet, able to &#8220;test&#8209;fly&#8221; disaster scenarios before they happen.<sup>8</sup> On the modelling front, DeepMind&#8217;s <strong>GraphCast</strong> forecasts global weather up to 10&#8239;days ahead with record accuracy, running in minutes on commodity chips.<sup>9</sup> When these AI engines merge with near&#8209;real&#8209;time EO, decision&#8209;makers won&#8217;t just see hazards sooner; they&#8217;ll simulate response plans, budget impacts and supply&#8209;chain knock&#8209;ons before the first raindrop or ember lands.</p><h2><strong>Build the warning network</strong></h2><p>The UN&#8217;s 2027 target for universal early warnings will not be met by satellites alone. It will take shared standards, interoperable APIs and boundary&#8209;less collaboration. <strong>Platforms that harmonise diverse data sources into decisive, trustworthy intelligence are the missing layer.</strong></p><p>Stelia is committed to partnering across that ecosystem, helping turn today&#8217;s patchwork of orbital eyes into tomorrow&#8217;s global nervous system. Because the world already has the eyes in the sky; what it needs now is the brain that turns sight into safety.</p><div><hr></div><h3>References</h3><ol><li><p><em>Natural Disasters in 2024 &#8211; Full&#8209;Year Fact Sheet</em>, Munich Re, January 2025. <a href="https://www.munichre.com/en/company/media-relations/media-information-and-corporate-news/media-information/2025/natural-disaster-figures-2024.html">Climate change is showing its claws: The world is getting hotter, resulting in severe hurricanes, thunderstorms and floods | Munich Re</a></p></li><li><p><em>Early Warnings for All</em>, United Nations. <a href="https://www.un.org/en/climatechange/early-warnings-for-all">https://www.un.org/en/climatechange/early-warnings-for-all</a></p></li><li><p><em>Flood Rapid Impact Launch</em>, ICEYE press release, July 2025. <a href="https://www.iceye.com/newsroom/press-releases/iceye-unveils-ml-powered-flood-rapid-impact-product">ICEYE unveils machine learning-powered Flood Rapid Impact Product to revolutionize response</a></p></li><li><p><em>OroraTech Wildfire Constellation</em>, eoPortal, June 2025 <a href="https://www.eoportal.org/satellite-missions/ororatech-wildfire-constellation#launch">OroraTech Wildfire Constellation - eoPortal</a></p></li><li><p><em>AI for Reliable Flood Forecasting</em>, Google blog, March 2024. <a href="https://blog.google/technology/ai/google-ai-global-flood-forecasting/">How we are using AI for reliable flood forecasting at a global scale</a></p></li><li><p><em>MAPFRE RE Licenses Global Flood Data</em>, ICEYE press release, July 2025. <a href="https://www.iceye.com/newsroom/press-releases/mapfre-re-signs-agreement-to-license-iceyes-global-flood-data">MAPFRE RE signs agreement to license ICEYE&#8217;s global flood data</a></p></li><li><p><em>Supporting Rio Grande do Sul Government</em>, ICEYE press release, July 2025. <a href="https://www.iceye.com/newsroom/press-releases/iceye-supports-rio-grande-do-sul-government-in-mapping-flooded-areas-and-mobilizing-humanitarian-aid">ICEYE supports Rio Grande do Sul government in mapping flooded areas and mobilizing humanitarian aid</a></p></li><li><p><em>Destination Earth: The Digital Twin Helping to Predict &#8211; and Prevent &#8211; Climate Change</em>, IT Pro, July 2025. <a href="https://www.itpro.com/technology/artificial-intelligence/destination-earth-the-digital-twin-helping-to-predict-and-prevent-climate-change">Destination Earth: The digital twin helping to predict &#8211; and prevent &#8211; climate change</a></p></li><li><p><em>GraphCast: AI Model for Faster and More Accurate Global Weather Forecasting</em>, DeepMind blog, November 2023. <a href="https://www.deepmind.com/discover/blog/graphcast-ai-model-for-faster-and-more-accurate-global-weather-forecasting">GraphCast: AI model for faster and more accurate global weather forecasting</a></p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Engineering medical digital twins for faster clinical trials]]></title><description><![CDATA[Unlearn.ai uses generative models to simulate patient counterfactuals, cutting trial sizes 30&#8211;50% while meeting regulatory and data quality challenges.]]></description><link>https://stelia.substack.com/p/engineering-medical-digital-twins</link><guid isPermaLink="false">https://stelia.substack.com/p/engineering-medical-digital-twins</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Wed, 30 Jul 2025 07:30:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!pp6j!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pp6j!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pp6j!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pp6j!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pp6j!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pp6j!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pp6j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:821949,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/169559565?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pp6j!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pp6j!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pp6j!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pp6j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd97c564c-9999-4c1d-8c96-c57e242bea30_3840x2160.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>What if you could generate a synthetic version of every patient in a clinical trial&#8202;&#8212;&#8202;one that shows exactly what would have happened to them without treatment?</p><p>Unlearn.ai built exactly that system. Using conditional restricted Boltzmann machines, they create counterfactual patient trajectories that reduce trial sizes by 30&#8211;50% without losing statistical power.</p><p>The engineering challenge isn&#8217;t just building generative models. It&#8217;s doing so with medical data that has 15&#8211;40% missing values and irregular temporal sampling, while meeting regulatory requirements that would break most ML pipelines. The result is a production system that fundamentally changes how clinical evidence gets generated&#8202;&#8212;&#8202;but building it required solving problems that don&#8217;t exist in typical ML deployments.</p><h3>The architecture: conditional generation at scale</h3><p>Unlike standard ML models that predict P(outcome|features), this system learns P(outcomes | baseline_features, treatment_assignment, time_horizon)&#8202;&#8212;&#8202;essentially building a time-series generator conditioned on medical context. Think of it like dependency injection for probabilistic models: instead of hardcoding assumptions about patient populations, you inject specific patient contexts at runtime.</p><p>The breakthrough methodology, called PROCOVA (Prognostic Covariate Adjustment), transforms the entire statistical inference process. Instead of comparing treatment groups against population averages, each patient gets compared to their digital twin:</p><pre><code># Traditional approach
P(outcome | features)

# Unlearn.ai's conditional generation
P(outcomes | baseline_features, treatment_assignment, time_horizon)

# Per-patient comparison
Treatment_Effect_i = (Observed_Outcome_i - Predicted_Natural_History_i) / Predicted_Natural_History_i</code></pre><p>This shifts from a between-group comparison problem to a within-patient prediction problem. The baseline heterogeneity that creates noise in traditional trials becomes useful signal for personalised modelling&#8202;&#8212;&#8202;turning what was previously statistical noise into valuable training data.</p><p>The conditional restricted Boltzmann machine architecture splits its internal representation across different timescales. Similar to how modern systems separate hot and cold storage paths, separate hidden unit populations capture short-term clinical fluctuations versus long-term disease progression&#8202;&#8212;&#8202;like having different model components specialised for high-frequency and low-frequency patterns in the same probabilistic framework.</p><p>Training requires sliding temporal windows across massive datasets: electronic health records spanning decades, clinical trial databases, real-world evidence from insurance claims. Each training example effectively answers: &#8220;Given this patient profile at this time point, here&#8217;s how their disease progressed over the following months.&#8221;</p><p>The cRBM training pipeline resembles large language model pretraining&#8202;&#8212;&#8202;massive datasets, GPU-accelerated optimisation, generation of synthetic outputs that preserve statistical structure from training data. The critical difference is conditioning and consequence: where language models autocomplete text, this system forecasts disease trajectories with direct clinical and regulatory implications.</p><h3>Data engineering reality check</h3><p>This conditional generation approach sounds elegant in theory. But like most ML systems, the mathematics are only as good as the training data&#8202;&#8212;&#8202;and medical data presents failure modes that would cripple models trained on cleaner domains like computer vision or natural language processing.</p><p>When a patient&#8217;s lab results are missing, it could be a clinical decision (test wasn&#8217;t ordered), operational failure (test was ordered but not performed), documentation gap (test performed but not recorded), or technical failure (test recorded but not transmitted). Each scenario carries different implications for model training, because treating systematic gaps as random missingness introduces bias that propagates through every generated trajectory.</p><p>Like handling eventual consistency in distributed databases, you can&#8217;t assume all medical data arrives on time or at all, so the system must gracefully handle partial information. The platform implements multi-layered data validation that operates continuously rather than as a preprocessing step:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R2sj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R2sj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg 424w, https://substackcdn.com/image/fetch/$s_!R2sj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg 848w, https://substackcdn.com/image/fetch/$s_!R2sj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!R2sj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R2sj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg" width="1179" height="990" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:990,&quot;width&quot;:1179,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:201616,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/169559565?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!R2sj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg 424w, https://substackcdn.com/image/fetch/$s_!R2sj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg 848w, https://substackcdn.com/image/fetch/$s_!R2sj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!R2sj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c99c20d-37e9-4ab8-ae49-2d3968757656_1179x990.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This approach builds quality assessment directly into the generative process so that model uncertainty reflects actual data reliability rather than algorithmic confidence&#8202;&#8212;&#8202;ensuring that every synthetic patient trajectory is grounded in reliable medical evidence.</p><h3>Production-scale engineering challenges</h3><p>With reliable data pipelines established, the next challenge swiftly arises: generating thousands of synthetic patient trajectories in real-time while maintaining the statistical validity that regulators demand. This creates computational and memory constraints that don&#8217;t exist in typical ML deployments.</p><p><strong>Distributed computation:</strong> The architecture distributes inference across GPU clusters, with each processing unit handling distinct patient cohorts. Shared model parameters ensure consistent probabilistic behaviour while enabling high-throughput trajectory generation. The parallelisation strategy accounts for the fact that patient trajectories can be generated independently once the model is trained.</p><p><strong>Memory-efficient state management:</strong> Long-horizon simulations for conditions like cardiovascular disease can span two-year periods with dozens of clinical variables per patient. The system minimises memory overhead through trajectory compression and selective caching strategies that store only clinically significant state transitions.</p><p><strong>Adaptive trajectory updates:</strong> As real trial data streams in, each digital twin incorporates new information to refine its predictions. A patient&#8217;s unexpected lab result or adverse event updates their model state, adjusting the synthetic trajectory moving forward. This creates a feedback loop between trial execution and generative interface&#8202;&#8212;&#8202;unlike typical batch prediction systems that operate on static datasets.</p><p>The regulatory requirements create constraints that most ML systems never face:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hcfE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hcfE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hcfE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hcfE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hcfE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hcfE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg" width="1180" height="469" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:469,&quot;width&quot;:1180,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:90093,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/169559565?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hcfE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hcfE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hcfE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hcfE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d8c6fc1-c129-4046-ac4a-1a6aa7e3b260_1180x469.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Explainability requirements:</strong> While cRBMs excel at capturing complex statistical dependencies, they don&#8217;t naturally produce the interpretable outputs that regulators require. The platform implements gradient-based attribution methods that trace outcome predictions back to specific baseline variables and prior clinical events, building transparency into a fundamentally generative system.</p><h3>Technical validation at scale</h3><p>All this infrastructure investment raises a critical question: how do you prove that synthetic patients are statistically equivalent to real ones? The validation approach operates under a simple principle: model sophistication cannot compensate for data unreliability, so continuous validation monitors multiple dimensions simultaneously.</p><p>This resembles chaos engineering for ML models&#8202;&#8212;&#8202;continuously testing edge cases and failure modes rather than assuming training performance generalises to production scenarios:</p><p><strong>Primary validation:</strong></p><ul><li><p><strong>Data drift detection:</strong> Real-time comparison of incoming patient data against training distribution parameters. When new patients exhibit characteristics outside the model&#8217;s training envelope, the system flags potential extrapolation errors and adjusts confidence intervals accordingly.</p></li><li><p><strong>Cross-population validation:</strong> The platform maintains separate validation datasets for different demographic groups, disease stages, and treatment protocols. Before generating synthetic trajectories for any patient cohort, the system verifies that adequate training data exists for that specific population segment.</p></li></ul><p><strong>Continuous monitoring</strong>: For ongoing trials, the system continuously compares its synthetic predictions against actual patient outcomes as they become available. Significant deviations trigger automated investigation into whether the discrepancy reflects model limitations, data quality issues, or genuine clinical surprises that require model updates.</p><p><strong>Statistical calibration:</strong> High statistical power only provides value if prediction confidence is properly calibrated across diverse patient demographics and disease stages. The platform continuously tests and adjusts its uncertainty estimates, ensuring confidence intervals reflect actual prediction accuracy rather than model confidence.</p><p>Each of these validation capabilities serves the larger goal of reducing trial sizes while maintaining statistical rigour&#8202;&#8212;&#8202;proving that synthetic patients can reliably substitute for real ones in clinical research.</p><h3>Quantified engineering impact</h3><p>This comprehensive validation framework enables Unlearn.ai to make a bold claim: their technical improvements translate into measurable changes in trial efficiency.</p><pre><code>PHASE II ONCOLOGY TRIALS
Typical: 200-400 patients &#8594; With Digital Twins: 120-280 patients
&#8595; 30-40% reduction

PHASE III CARDIOVASCULAR TRIALS  
Typical: 2000-4000 patients &#8594; With Digital Twins: 1400-2800 patients
&#8595; 30-50% reduction

TIMELINE IMPACT: 25-40% faster completion
STATISTICAL POWER: Maintained at equivalent levels</code></pre><h3>Broader engineering implications</h3><p>This system demonstrates how generative modelling built on conditional probability structures enables not just inference, but simulation. It moves beyond prediction into counterfactual reasoning at population scale&#8202;&#8212;&#8202;a pattern that opens applications far broader than medical AI.</p><p>The technical achievement lies not in abstract algorithmic complexity, but in solving concrete real-world constraints through generative computation. The mathematics transforms a fundamental limitation in how medical evidence is generated, while the engineering makes it operationally viable at the scale required for regulatory acceptance.</p><p>For developers working on other domains requiring counterfactual reasoning&#8202;&#8212;&#8202;recommendation systems that need to model alternative user journeys, financial models that simulate different market conditions, or any system where you need to answer &#8220;what would have happened if&#8221;&#8202;&#8212;&#8202;this architecture provides a production-tested pattern for conditioning generative models on complex, multi-dimensional contexts while maintaining statistical rigour at scale.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Why genomics proves AI's true potential for humanity]]></title><description><![CDATA[BioAro&#8217;s genomics breakthrough delivers real-time insights and unlocks precision medicine - redefining global healthcare and advancing humanity.]]></description><link>https://stelia.substack.com/p/why-genomics-proves-ais-true-potential</link><guid isPermaLink="false">https://stelia.substack.com/p/why-genomics-proves-ais-true-potential</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Thu, 24 Jul 2025 07:30:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xrTS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xrTS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xrTS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic 424w, https://substackcdn.com/image/fetch/$s_!xrTS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic 848w, https://substackcdn.com/image/fetch/$s_!xrTS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic 1272w, https://substackcdn.com/image/fetch/$s_!xrTS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xrTS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:749641,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/169051215?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xrTS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic 424w, https://substackcdn.com/image/fetch/$s_!xrTS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic 848w, https://substackcdn.com/image/fetch/$s_!xrTS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic 1272w, https://substackcdn.com/image/fetch/$s_!xrTS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb43e48c8-b637-4051-b99b-910ec19c3e43_3840x2160.heic 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>By 2030, more than one in six people globally will be over 60. Healthcare costs are climbing 5% annually, outpacing GDP growth in developed nations. Meanwhile, non-communicable diseases account for 74% of global deaths, yet early detection and prevention remain limited by a fundamental constraint: our inability to understand disease origins at the molecular level quickly enough to matter.</p><p>The promise of genomics has existed for decades. Sequence a person's DNA, understand their biological blueprint, intervene before disease takes hold. The science works. What hasn't worked is the operationalisation. Traditional genomic analysis takes weeks, costs thousands, and requires specialised facilities accessible to a fraction of the global population.</p><p>This gap between scientific possibility and practical accessibility defines one of the most pressing challenges in healthcare today. It's also becoming a defining test case for AI's role in addressing planetary-scale human needs.</p><h2>The speed revolution</h2><p>A Canadian biotechnology company called BioAro recently achieved something that reframes this entire discussion. Their PanOmiQ platform completed whole genome sequencing analysis in under two hours, generating variant call format files in under five minutes. To put this in perspective: the first human genome took 13 years to sequence. Current industry standard is about four weeks. BioAro compressed this to seven hours with 100% accuracy verified by the College of American Pathologists.</p><p>This isn't just faster computing. Speed at this scale transforms what genomic medicine can be. Emergency room diagnoses. Rural clinic interventions. Real-time treatment decisions in developing nations. When analysis time drops from weeks to hours, genomic medicine stops being a specialised service and starts becoming standard care.</p><p>The technical achievement rests on AI systems that can process multi-omics data in real time, integrating genomics, proteomics, metabolomics, and microbiome analysis into unified insights. More significantly, these systems are designed for global deployment from the start, generating clinical reports in six languages and operating across diverse healthcare infrastructures.</p><p>This represents a pattern emerging across AI applications: the organisations succeeding at scale aren't just building better algorithms. They're building systems architected for planetary-scale distribution from day one.</p><h2>Democratisation through AI</h2><p>The implications extend far beyond genomic analysis speed. BioAro's platform eliminates the high costs and complexity that have historically made whole genome sequencing accessible only to major medical centres. When AI can automate what previously required teams of specialists, advanced genomics becomes deployable in hospitals, research institutions, and underserved communities worldwide.</p><p>This democratisation effect illustrates something crucial about AI's potential impact on humanity. The technology's greatest contribution may not be its raw capabilities, but its ability to make sophisticated human knowledge accessible at scale. Complex expertise that once required years of training and expensive infrastructure can now be embedded in systems that operate reliably across diverse environments and user contexts.</p><p>The multilingual reporting capability provides another glimpse of what global AI deployment requires. Building systems that work across cultural and linguistic boundaries demands more than translation. It requires understanding how medical knowledge is communicated, interpreted, and acted upon in different healthcare contexts. Success at planetary scale means accounting for this diversity from the architectural level, not adding it as an afterthought.</p><p>Organisations attempting to operationalise AI globally are discovering that the technical challenge of making algorithms work is often simpler than the human challenge of making them work everywhere, for everyone.</p><h2>From code to cure</h2><p>BioAro has also introduced something they call an "omics-to-therapeutics" ecosystem. Their AI doesn't just analyse genetic data; it identifies novel therapeutic targets and uses both structure-based and ligand-based modelling to predict how potential drugs might behave in the human body. The system generates chemical candidates from biological insights at what they describe as "machine speed."</p><p>This integration reveals another dimension of AI's potential impact. Rather than replacing human expertise in drug discovery, the platform amplifies it by handling the computational heavy lifting that currently creates bottlenecks. Researchers can focus on the creative and strategic aspects of therapeutic development while AI manages the systematic analysis of massive biological datasets.</p><p>The economic implications are substantial. Traditional drug discovery takes decades and costs billions, with high failure rates. When AI can compress the timeline from biological insight to chemical candidate, it changes the fundamental economics of developing treatments for rare diseases, personalised therapies, and conditions that affect populations in developing nations.</p><p>This points to a broader principle: AI's most transformative applications may emerge when it enables entirely new approaches to existing problems, rather than simply optimising current processes.</p><h2>Redefining human potential</h2><p>What becomes possible when genetic analysis shifts from a specialised service to a standard capability available anywhere healthcare is practised? The answer goes beyond faster diagnoses or more targeted treatments.</p><p>Predictive medicine becomes feasible at population scale. Instead of waiting for symptoms to appear, healthcare systems can identify genetic predispositions and intervene before diseases develop. This shift from reactive to preventive care could fundamentally alter how we think about health and ageing.</p><p>The microbiome analysis capabilities built into BioAro's platform illustrate another frontier. By understanding the interplay between genetics and the microbial ecosystems in our bodies, AI enables interventions tailored not just to our genes, but to our complete biological context. Personalised nutrition, targeted immunotherapy, and chronic disease management all become more precise when guided by comprehensive biological data.</p><p>These advances point towards a future where medicine becomes truly individualised. Rather than treating populations with standardised protocols, healthcare providers can design interventions specific to each person's unique biological profile.</p><p>The challenge is ensuring these capabilities benefit everyone, not just those with access to advanced medical systems.</p><h2>The global deployment challenge</h2><p>Achieving this vision requires solving what may be AI's defining challenge: moving from promising laboratory results to reliable performance across diverse real-world environments. This is where many AI initiatives stumble. Systems that work brilliantly in controlled settings often struggle when deployed globally.</p><p>BioAro's approach offers lessons for organisations grappling with this transition. Their platform is designed for multiple deployment models, supporting both cloud-based and on-premise implementations. This flexibility allows healthcare systems with different technical infrastructures and regulatory requirements to adopt the technology in ways that work for their specific contexts.</p><p>The quantum computing integration they're developing represents another approach to scale-out challenges. By leveraging quantum processing for genomic analysis, they're building systems that can handle exponentially larger datasets while maintaining real-time performance. Beyond commodity computational power, creating the coordination layer needed to serve billions of people simultaneously is the grander challenge.</p><p>Organisations attempting to deploy AI at planetary scale consistently encounter this same fundamental challenge: how to build systems that work reliably across the full spectrum of human contexts, technical environments, and cultural frameworks.</p><h2>The time is now</h2><p>The transformation BioAro is driving in genomics illustrates what becomes possible when complex systems are designed with both scientific rigour and human impact in mind. At Stelia, we see this as part of a broader movement - one that recognises AI&#8217;s future isn&#8217;t just technical, but societal, and that true progress lies in harmonising complex systems with collective wellbeing. The organisations succeeding in this transition share common characteristics: they prioritise deployed systems over promised capabilities, integrate diverse perspectives into their development processes, and focus on measurable outcomes rather than aspirational claims.</p><p>BioAro&#8217;s work reflects a deeper shift already underway - one where the value of AI is measured not just by what it can optimise, but by the real-world outcomes it enables for human lives. This is no longer just a question of technical acceleration, but of ethical integration.</p><p>The future belongs to those who can architect for both.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Eric Schmidt’s case for ‘sovereign AI’ misses the point]]></title><description><![CDATA[Schmidt predicts superintelligence in a decade, but his case for sovereign AI ignores how portable models will outpace national control and regulation.]]></description><link>https://stelia.substack.com/p/eric-schmidts-case-for-sovereign</link><guid isPermaLink="false">https://stelia.substack.com/p/eric-schmidts-case-for-sovereign</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Tue, 22 Jul 2025 09:12:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Owzx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Owzx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Owzx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Owzx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Owzx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Owzx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Owzx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2209234,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/168846861?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Owzx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Owzx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Owzx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Owzx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ecdf242-577f-464b-8dc2-82d7f5805943_3840x2160.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>Schmidt, super&#8209;intelligence, and sovereignty</strong></h2><p>Last week former Google chief executive <strong>Eric&#8239;Schmidt</strong> spent seventy&#8209;five minutes on the <em>Moonshots</em> podcast with Peter&#8239;Diamandis and Dave&#8239;Blundin. The headline moment came when he declared that <strong>digital super&#8209;intelligence is &#8220;within ten years.&#8221;</strong> The other take&#8209;away: those future models, he said, will live in giga&#8209;scale data centres guarded like nuclear stockpiles.</p><p>The remark has landed because policy circles from Washington to Brussels are already debating <strong>&#8220;sovereign&#8239;AI&#8221;.</strong> If national power is set to hinge on who controls the biggest models, Schmidt&#8217;s comments sound like a strategic field manual. They also, as this article shows, rest on a shaky reading of how machine&#8209;learning systems actually spread.</p><h2><strong>Schmidt&#8217;s vision in a nutshell</strong></h2><blockquote><p><em><strong>&#8220;If the structure of the world in five to ten years is ten models &#8211; five in the United States, three in China, two elsewhere &#8211; those data centres will be all nationalised in some way. In China they will be owned by the government. The stakes are too high.&#8221;</strong></em><strong> (31&#8239;min&#8239;57&#8239;s&#8239;&#8211;&#8239;32&#8239;min&#8239;28&#8239;s)</strong></p></blockquote><p>Schmidt&#8217;s argument has four main pillars:</p><ol><li><p><strong>Ten&#8209;year countdown</strong><br><em>&#8220;When do you see what you define as digital super&#8209;intelligence? &#8211; Within ten years.&#8221;</em> (1&#8239;h&#8239;22&#8239;m&#8239;09&#8239;s)</p></li><li><p><strong>Energy as the bottleneck</strong><br><em>&#8220;The natural limit is electricity, not chips&#8230; The United States will need ninety&#8209;two gigawatts more power.&#8221;</em> (2&#8239;m&#8239;13&#8239;s&#8239;&#8211;&#8239;3&#8239;m&#8239;46&#8239;s)</p></li><li><p><strong>Nation&#8209;scale data centres as strategic assets</strong><br>A training run for xAI&#8217;s Grok used a <em>&#8220;ten&#8209;billion&#8209;dollar super&#8209;computer in one building.&#8221;</em> (34&#8239;m&#8239;05&#8239;s&#8239;&#8211;&#8239;34&#8239;m&#8239;34&#8239;s)</p></li><li><p><strong>Deterrence through &#8220;mutual&#8239;AI malfunction&#8221;</strong><br>A state&#8209;level balance of power where each side can crash the other&#8217;s mega&#8209;centre if lines are crossed (25&#8239;m&#8239;56&#8239;s&#8239;&#8211;&#8239;27&#8239;m&#8239;12&#8239;s).</p></li></ol><p>In this frame sovereignty equals location: keep the decisive compute inside your borders and guard it with armed security.</p><h2><strong>Two AI races, one sovereignty problem</strong></h2><p>Eric&#8239;Schmidt actually sketches two separate competitions.</p><p><strong>Race&#8239;1: the fortress sprint</strong><br><em>&#8220;These data centres will be nationalised in some way&#8230; the stakes are too high.&#8221;</em> (31&#8239;min&#8239;57&#8239;s)<br>In other words, only a handful of states or mega&#8209;firms can afford a gigawatt hall and a ten&#8239;billion&#8209;dollar training run.</p><p><strong>Race&#8239;2: the diffusion scramble</strong><br>Minutes later he warns that, once the weights are distilled, <em>&#8220;the final brain can run on four or eight GPUs &#8211; a box about this size,&#8221;</em> making powerful models <em>&#8220;a hundred, a thousand, a million&#8221;</em> in number (33&#8239;min&#8239;19&#8239;s; 41&#8239;min&#8239;50&#8239;s).</p><p>During the episode Dave&#8239;Blundin recalls a visit to OpenAI where researcher <strong>Noam&#8239;Brown</strong> (not on the podcast) predicted that models will soon <strong>write their own step&#8209;by&#8209;step workplans, or &#8220;scaffolding,&#8221;</strong> making multi&#8209;hour tasks autonomous by 2025. If true, diffusion in Race&#8239;2 speeds up because the planning layer travels with the weights, not with the cloud provider.</p><h2><strong>Why this matters for &#8216;sovereign&#8239;AI&#8217;</strong></h2><ul><li><p>Fortress sovereignty may work during Race&#8239;1 while the capability lives in guarded sites.</p></li><li><p>Race&#8239;2 dissolves that physical anchor. Portable checkpoints flow across borders faster than regulators can stamp passports.</p></li><li><p>Schmidt offers a deterrence doctrine for Race&#8239;1 but admits the diffusion phase is <em>&#8220;a set of unknown questions.&#8221;</em></p></li><li><p>Location&#8209;based sovereignty will have little time to prove its worth before portable, self&#8209;planning checkpoints flood every jurisdiction.</p></li></ul><p>Stelia calls this the <strong>sovereignty paradox</strong>: legal control framed around location collapses at the very moment the technology scales to everyday hardware.</p><h2><strong>Three gaps in the Schmidt approach</strong></h2><h3><strong>1. Compute does not equal control</strong></h3><p>Schmidt&#8217;s fortress protects the factory, not the product. Moments after describing armed guards he concedes that the same model, once <strong>distilled</strong>, can run on a desktop server:</p><blockquote><p><em><strong>&#8220;The final brain can be ported and run on four or eight GPUs &#8211; a box about this size.&#8221;</strong></em><strong> (41&#8239;m&#8239;50&#8239;s&#8239;&#8211;&#8239;42&#8239;m&#8239;25&#8239;s)</strong></p></blockquote><p>When weights leave the building they lose all jurisdictional tags. The link between physical residency and real control vanishes.</p><h3><strong>2. Deterrence needs a target &#8211; what if there is none?</strong></h3><p>Nuclear doctrine worked because silos were fixed on maps. Schmidt&#8217;s &#8220;mutual&#8239;AI malfunction&#8221; assumes similar visibility. Yet portable checkpoints allow powerful models to live on laptops, research clusters, open&#8209;source mirrors and even criminal botnets. A strategy that threatens single buildings cannot restrain actors who own none.</p><h3><strong>3. Whose sovereignty counts?</strong></h3><p>The Schmidt narrative is Washington versus Beijing. Missing are Indigenous communities whose language archives feed multilingual models, European citizens protected by GDPR, or African start&#8209;ups seeking local agency. National control over hardware does not answer their claims to participation and benefit.</p><h2><strong>A governance path that could work</strong></h2><p>Stelia&#8217;s engineers operate multi&#8209;jurisdiction deployments every day. Three design principles align with technical reality:</p><p><strong>Residency first:</strong> Keep raw data and first&#8209;training runs inside legally recognised regions using geo&#8209;fenced compute and region&#8209;tied encryption keys.</p><p><strong>Transparency of weights:</strong> Publish weight hashes and lineage metadata so any checkpoint can be traced, wherever it runs. Add inference watermarks for audit.</p><p><strong>Participation hooks:</strong> Build consent, veto rights and benefit&#8209;sharing into dataset pipelines so communities influence use before models go live.</p><p>These tools do not pretend to project domestic law across borders. Instead they create verifiable constraints that travel with the artefact itself &#8211; a sharper fit for how machine&#8209;learning actually works.</p><h2><strong>Why this matters now</strong></h2><p>With Race&#8239;1 already under way and Race&#8239;2 close behind, the window for workable governance is short.</p><blockquote><p><em><strong>&#8220;We are saying it is 1938&#8230; we need to start the conversation now, well before the Chernobyl events.&#8221;</strong></em><strong> (28&#8239;m&#8239;20&#8239;s&#8239;&#8211;&#8239;28&#8239;m&#8239;41&#8239;s)</strong></p></blockquote><p>If policymakers chase the mirage of fortress sovereignty, genuine levers of control will ossify behind optics and compliance theatre. Residency, transparency and participation offer a route that communities can verify and industry can implement.</p><p>Stelia will publish a technical white paper on data sovereignty later this month.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p><em>Source: Moonshots with Eric&#8239;Schmidt (YouTube, June&#8239;2025): </em><a href="https://www.youtube.com/watch?v=qaPHK1fJL5s">Ex-Google CEO: What Artificial Superintelligence Will Actually Look Like w/ Eric Schmidt &amp; Dave B</a></p><p></p>]]></content:encoded></item><item><title><![CDATA[The AI strategy that could rescue AdTech compliance]]></title><description><![CDATA[AdTech faces a compliance crisis. Federated learning enables privacy-safe AI by training models locally &#8211; no data movement, no violations, global performance.]]></description><link>https://stelia.substack.com/p/the-ai-strategy-that-could-rescue</link><guid isPermaLink="false">https://stelia.substack.com/p/the-ai-strategy-that-could-rescue</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Fri, 18 Jul 2025 14:35:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!pVNo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pVNo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pVNo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pVNo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pVNo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pVNo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pVNo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6144558,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/168640113?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pVNo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pVNo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pVNo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pVNo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83f0e360-d46b-4f1d-a916-ef1933c3b9f8_3840x2160.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>AdTech has reached a critical disconnect. As AI models become more sophisticated at predicting user behaviour and optimising campaigns, the data required to train these systems becomes increasingly restricted by privacy regulations. Centralised cloud AI collides with per&#8209;country data rules, creating a fundamental mismatch<sup>1</sup>.</p><h2>The compliance trap</h2><p>This creates what industry veterans are calling the "compliance trap" &#8211; where the most effective AI strategies become legally impossible to implement using legacy cloud platforms.</p><p>Recent policy developments underscore how profoundly policymakers misunderstand this challenge. Europe&#8217;s &#8364;200&#8239;billion &#8216;AI gigafactory&#8217; keeps training inside its borders. Yet a finished model is just numbers; once copied, it operates anywhere. This reflects what researchers term the "sovereignty paradox": the more sophisticated AI systems become, the less meaningful traditional geo&#8209;territorial control mechanisms prove to be<sup>2</sup>.</p><p>For AdTech, this paradox reveals a fundamental flaw in compliance strategies based on geographical data control. A trained targeting model is simply a matrix of numbers. Once training completes, those weights can be perfectly copied and deployed on smartphones, connected TVs, and programmatic platforms worldwide, entirely beyond the jurisdiction where training occurred.</p><h2>How federated learning actually works</h2><p>To understand why this matters for AdTech, consider how federated learning solves the fundamental data&#8209;access problem. Traditional AI training requires collecting all user data in one central location, training a model on the combined dataset, then deploying globally. This approach risks breaching local privacy legislation, creates security risks, and requires expensive data movement across jurisdictions.</p><p>Federated learning flips this model entirely. Instead of moving data to where computation happens, it sends models to where data lives. Here's the process for a global AdTech platform wanting to improve click prediction across EU and US markets:</p><p><strong>Clone</strong> &#8211; A central server creates an initial click&#8209;prediction model and distributes identical copies to EU data centres (Frankfurt) and US data centres (Virginia).</p><p><strong>Train&#8239;locally</strong> &#8211; The EU facility trains the model on German, French, and Italian user data locally, while the US model trains on American user data locally. Raw user data never leaves its regional boundaries.</p><p><strong>Send&#8239;weights</strong> &#8211; After training, each data centre sends back only model parameters &#8211; mathematical weights and biases that represent learned patterns.</p><p><strong>Merge</strong> &#8211; The central coordinator combines EU and US updates using algorithms such as Federated Averaging, creating an improved global model without moving personal data across borders.</p><p><strong>Re&#8209;deploy</strong> &#8211; The updated model returns to both regions and the cycle repeats for continuous improvement.</p><h2>The compliance reality</h2><p>This architecture directly addresses core compliance challenges. The EU centre processes millions of EU users' clicks and sends only weights to the coordinator; the US centre does the same with American data. What flows are mathematical insights, not personal information<sup>3</sup>.</p><p>The European Data Protection Board&#8217;s Opinion&#8239;28/2024 clarifies that parameter exchange aligns with GDPR&#8217;s data&#8209;minimisation and purpose&#8209;limitation principles<sup>4</sup>. In 2023, Ireland&#8217;s Data Protection Commission fined Meta &#8364;405&#8239;million for GDPR violations tied to behavioural advertising<sup>5</sup>. Federated learning offers a potential privacy&#8209;enhancing alternative, keeping raw data local while enabling global optimisation.</p><h2>Legacy platform limitations</h2><p><strong>General&#8209;purpose clouds support federated learning only as an add&#8209;on, not a design principle.</strong> AWS, Azure, and Google Cloud each provide federated toolkits, yet they rely on centralised foundations that re&#8209;introduce compliance vulnerabilities and add DevOps overhead<sup>6</sup>.</p><p>The mathematical reality makes geographical training control meaningless. A personalised ad model trained on European infrastructure is identical to any copy of its numerical weights. Once deployed globally, these models operate beyond territorial governance, turning &#8220;Sovereign&#8239;AI&#8221; into compliance theatre. Standard cloud tools also struggle with multi&#8209;party coordination; authentication and synchronisation often fail when a TV OEM, streaming service, and programmatic platform attempt joint training.</p><h2>Technical performance reality</h2><p><strong>Federated learning delivers up to&#8239;30&#8239;% higher accuracy</strong> because local data diversity stays intact. Traditional centralised training homogenises insights, whereas federated approaches preserve regional signal and enable nuanced targeting, crucial in contexts like connected&#8209;TV viewing, which varies by culture, device, and content genre.</p><h2>The coordination challenge</h2><p>Successful adoption requires orchestrating training across many independent systems. Identity management, version control, and rollback mechanisms must be standardised to avoid conflict and ensure auditability, needs that exceed the scope of most legacy ML tooling.</p><h2>Strategic implementation necessity</h2><p>Regulations are tightening while third&#8209;party data dwindles; the gap between distributed and centralised AI will widen. Valued at $151&#8239;million in&#8239;2024, the federated&#8209;learning market is projected to reach $507&#8239;million by&#8239;2033 (CAGR&#8239;13.6&#8239;%)<sup>7</sup>. By embedding governance directly into system architecture, federated learning addresses real regulatory risk rather than relying on geographic workarounds.</p><p>Partnership strategy is crucial: consortiums of streaming services, publishers, and measurement vendors can pool insights while sharing costs. The real question is which organisations will secure first&#8209;mover advantages before legacy approaches become untenable.</p><p>Organisations that build federated capabilities now gain a sustainable edge that competitors will struggle to replicate as regulations harden and competitive pressures intensify. The compliance trap is real, but so is the solution for those prepared to implement governance through architecture rather than geography.</p><h2>Next steps for decision&#8209;makers</h2><ul><li><p><strong>Evaluate existing data flows</strong> against current and pending regulations.</p></li><li><p><strong>Run a limited federated pilot</strong> (e.g., one EU market, one US market) to measure performance uplift and confirm compliance reporting.</p></li><li><p><strong>Establish a cross&#8209;functional team</strong> combining legal, data science, and platform engineering to formalise governance processes.</p></li></ul><p>For organisations ready to explore this approach in detail, a short technical assessment can clarify requirements, expected ROI, and integration timelines.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p><h2>References</h2><p><sup>1</sup> <em>Opinion 28/2024: Certain Data Protection Aspects of the Use of AI Systems in the EU</em>, European Data Protection Board (EDPB), June 2024.<br><a href="https://www.edpb.europa.eu/our-work-tools/our-documents/opinion-board-art-64/opinion-282024-certain-data-protection-aspects_en">https://www.edpb.europa.eu/our-work-tools/our-documents/opinion-board-art-64/opinion-282024-certain-data-protection-aspects_en</a></p><p><sup>2</sup> <em>EDPB Opinion 28/2024 &#8211; Territorial Limits of AI Model Governance</em>, European Data Protection Board (EDPB), December 2024.<br><a href="https://www.edpb.europa.eu/system/files/2024-12/edpb_opinion_202428_ai-models_en.pdf">https://www.edpb.europa.eu/system/files/2024-12/edpb_opinion_202428_ai-models_en.pdf</a></p><p><sup>3</sup> <em>Opinion 28/2024: GDPR Compliance and Parameter Exchange in AI Training</em>, European Data Protection Board (EDPB), June 2024.<br><a href="https://www.edpb.europa.eu/our-work-tools/our-documents/opinion-board-art-64/opinion-282024-certain-data-protection-aspects_en">https://www.edpb.europa.eu/our-work-tools/our-documents/opinion-board-art-64/opinion-282024-certain-data-protection-aspects_en</a></p><p><sup>4</sup> <em>EDPB Guidance: How AI Parameter Sharing Aligns with GDPR Principles</em>, European Data Protection Board (EDPB), 2024.<br><a href="https://www.edpb.europa.eu/news/news/2024/edpb-opinion-ai-models-gdpr-principles-support-responsible-ai_en">https://www.edpb.europa.eu/news/news/2024/edpb-opinion-ai-models-gdpr-principles-support-responsible-ai_en</a></p><p><sup>5 </sup><em>Irish DPC Fines Meta &#8364;405M for Instagram GDPR Breaches</em>, Burges Salmon, September 2022.<br><a href="https://www.burges-salmon.com/articles/102hxck/irish-data-protection-commissioner-fines-meta-405m-for-violation-of-childrens-p/">https://www.burges-salmon.com/articles/102hxck/irish-data-protection-commissioner-fines-meta-405m-for-violation-of-childrens-p/</a><br><a href="http://www.dataprotection.ie/en/news-media/press-releases/data-protection-commission-announces-decision-instagram-inquiry">http://www.dataprotection.ie/en/news-media/press-releases/data-protection-commission-announces-decision-instagram-inquiry</a></p><p><sup>6</sup> <em>Federated Learning Systems: Challenges with Multi-Party Coordination</em>, arXiv preprint, February 2025.<br><a href="https://arxiv.org/html/2502.05273v1">https://arxiv.org/html/2502.05273v1</a></p><p><sup>7</sup> <em>Federated Learning Market Report 2024&#8211;2033</em>, IMARC Group, 2024.<br><a href="https://www.imarcgroup.com/federated-learning-market">https://www.imarcgroup.com/federated-learning-market</a></p>]]></content:encoded></item><item><title><![CDATA[How smart media companies orchestrate AI for real results]]></title><description><![CDATA[Agentic AI is ending pilot purgatory in media. Orchestration is now essential for scaling creativity, agility, and real-time decisions.]]></description><link>https://stelia.substack.com/p/how-smart-media-companies-orchestrate</link><guid isPermaLink="false">https://stelia.substack.com/p/how-smart-media-companies-orchestrate</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Wed, 16 Jul 2025 12:50:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bNZi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bNZi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bNZi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bNZi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bNZi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bNZi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bNZi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3858857,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://stelia.substack.com/i/168457278?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bNZi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bNZi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bNZi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bNZi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7949860-e119-49c4-9a30-023611ff6b4c_3840x2160.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>The illusion of progress</h3><p>For nearly a decade, artificial intelligence has been touted as the future of media and entertainment. Netflix has invested heavily in AI-driven personalisation and operational systems, continuously testing new models for engagement and optimisation. Disney continues to test automation tools. Warner Bros. runs ongoing proof-of-concept projects. Yet for all the headlines, panels, and prototypes, meaningful integration has remained frustratingly elusive. Endless pilots and proof of concept projects have filled the last few years each full of promise, most ending in silence. Now, as I work alongside companies facing intensified pressure for profitability, audience retention, real-time personalisation, and operational agility, it is clear that isolated AI experiments are no longer enough. The pilot era is over. What comes next is intelligent orchestration, an approach fundamentally different, inherently more strategic, and critical for survival in a rapidly evolving marketplace.</p><h3>The trap of pilot purgatory</h3><p>A recent industry survey conducted in June 2025 painted a sobering picture. Despite rising AI budgets, over 60 percent of media executives are still stuck in pilots, spending more time testing than transforming and only 4 percent operationalised AI at scale&#185;. Even more telling, nearly half of industry leaders cited unclear use cases and ROI uncertainty as the largest barriers to effective adoption. I&#8217;ve seen this first-hand. At Cannes this year, I heard the same refrain across panels and beachside meetings: &#8216;We&#8217;re testing, but not scaling.&#8217; But testing without commitment creates stagnation. This is what I call pilot purgatory, an endless loop of testing, proving, and pausing. It&#8217;s safe. It&#8217;s structured. But it doesn&#8217;t build anything. And in a landscape defined by rising costs, real-time engagement, and global pressure, staying in purgatory is not a neutral choice. It&#8217;s a risk.</p><h3>Why fragmented AI isn&#8217;t enough anymore</h3><p>Strategic pressure is mounting. In my work with content and technology teams, I hear it constantly. Streaming platforms are battling churn and the growing expectation for hyper-personalisation. Game publishers are navigating the complexity of live-service operations, compliance, and localisation. Production teams are buried in rights tracking and versioning. And agencies? They&#8217;re sprinting to deliver market-specific creative faster than ever. Fragmented workflows and isolated automation just can&#8217;t keep up. While platforms like TikTok and Meta are rolling out AI-driven DIY creative tools aimed at small businesses, the companies I work with don&#8217;t need more tools. They need orchestration with intelligence. They&#8217;re not looking to just generate more content. They want systems that adapt, decide, and deploy at scale.</p><h3>The new runway for competitive advantage</h3><p>What makes this moment different is not just that AI is better. It&#8217;s that the infrastructure to support true orchestration is finally emerging. Until recently, there were no systems built to connect content, data, compliance, and logic in real time across the full lifecycle of media creation and distribution. But that&#8217;s changing. AI native platforms are now being designed for orchestration. Built from the ground up to enable intelligence that is secure, governable, and affordable.</p><h3>Beyond automation: the rise of agentic orchestration</h3><p>The media and entertainment industry&#8217;s next evolutionary step is obvious: a strategic transition from isolated automation to systemic orchestration. Unlike traditional automation, which assists, agentic systems act. They decide, execute, adapt and respond across multiple workflows simultaneously without waiting for instruction. They are fit for a world that moves in real time.</p><p>When designed well, agentic orchestration becomes the intelligent backbone of a company&#8217;s operations. Not an add-on, not a tool, but the infrastructure through which decisions flow. Orchestration enables streaming platforms to dynamically tailor promotional content instantly across dozens of international markets, cutting localisation from weeks to hours. In gaming, real-time live-service updates, content translation, and compliance happen automatically, boosting player engagement while reducing friction. Film and audio production companies utilising orchestration gain real-time management of compliance, rights tracking, and asset metadata, significantly accelerating production timelines and reducing operational risk. Creative agencies implementing agentic orchestration rapidly iterate campaign assets, optimising creative output and minimising repetitive manual labour, freeing teams to concentrate on strategic innovation.</p><p>This is how media organisations move from friction to flow. From linear content operations to adaptive ecosystems that scale with demand.</p><h3>The real shift is in leadership</h3><p>But orchestration requires more than infrastructure. It demands a new kind of leadership. I often tell executives this: the real leap is not building smarter systems but becoming a different kind of leader. One who doesn&#8217;t manage tasks but designs systems that manage themselves. The most powerful leaders aren&#8217;t asking, &#8216;What should we automate?&#8217; They&#8217;re asking, &#8216;What can I offload so I can focus on what actually moves us forward?&#8217; What I can observe is that companies with high AI maturity get even 3&#215; higher ROI than those just testing the waters.</p><p>Today&#8217;s CROs are no longer just chasing CPMs. They&#8217;re re-engineering revenue operations through agentic systems that adapt in real-time to audience, regulation, and creative context. The most effective COOs aren&#8217;t managing workflows they&#8217;re architecting feedback loops. And CMOs, if they&#8217;re smart, are investing in infrastructure that doesn&#8217;t just push creative, but shapes it through intelligence.</p><p>Governance, compliance, and performance monitoring are no longer separate layers. They must be embedded directly within orchestration frameworks. In every project I advise, we insist on real-time auditing, automated metadata, and rights tracking&#8202;&#8212;&#8202;because trust can&#8217;t be an afterthought.</p><p>Real-time feedback loops are integral, continuously feeding operational performance data, market signals, and audience insights into orchestration systems. These loops ensure constant optimisation and refinement of workflows, enabling rapid adaptation to changing market dynamics and customer expectations.</p><p>One overlooked shift is the need to shape brand presence not just for consumers, but for the AI agents now parsing the internet. Training data matters. Influencing LLMs may soon require content designed not to be read, but to be remembered by algorithms. Orchestration is how that happens.</p><p>Adopting orchestration strategically involves incremental implementation and rapid scaling. Initial pilot orchestration projects should target high-impact operational areas, clearly demonstrating value to build internal momentum and confidence. Modular integration of orchestration technology into existing infrastructures allows rapid deployment without large-scale disruptions, ensuring immediate and measurable operational improvements.</p><p>Strategic patience and iterative agility are equally essential. Leaders must communicate incremental successes transparently, remain patient with evolving results, and continually refine orchestration frameworks to align with shifting industry landscapes and organisational priorities.</p><p>The rise of agentic AI is also reshaping how we think about talent. As digital labour becomes embedded, leaders must invest in AI literacy, workflow design, and soft skills to make the most of these systems. Technology alone doesn&#8217;t transform companies, people trained to wield it do. This new orchestration should liberate talent, empower teams to shift from routine operations to strategic creativity, innovation, and storytelling, thus amplifying human creativity rather than replacing it.</p><h3>Where to begin</h3><p>The temptation is always to start small. A workflow. A tagging system. A campaign tool. I would say: don&#8217;t start with isolated tools. Instead, pick one high-friction area where multiple systems collide and you can deliver exponential returns. Orchestration only creates real value when it addresses connected complexity. For media and entertainment, the most powerful starting points are content localisation, live-service updates, multi-region rollouts, and rights-bound metadata.</p><p>As generative agents begin to shape not just content creation but content discovery, we&#8217;re entering a phase where orchestration needs to consider not only human workflows, but AI-to-AI ecosystems. The next recommendation engine might not be a human but an LLM choosing one brand over another.</p><p>These orchestration pipelines must be modular, governed, and integrated with real-time data sources. They must evolve as the audience, platforms, and content evolve. They must be designed not as tools, but as systems.</p><h3>Not just faster: Smarter. Freer. Future-Ready.</h3><p>The companies that embrace this shift won&#8217;t just move faster. They&#8217;ll operate differently. With less friction. With more clarity. And with systems that scale intelligence as easily as they scale content. The next decade in media and entertainment won&#8217;t be defined by who tests AI the longest. It will be shaped by who orchestrates it best. The pilot culture is over. What we build next is what will last.</p><p><em>by Ula Nairne, VP Media &amp; Entertainment, Stelia</em></p><h3>References</h3><p>&#185; <em>AI Adoption in the Media Industry: Where It&#8217;s Working and Where It&#8217;s Stuck</em>, Amagi &amp; Dan Rayburn, June 2025. <a href="https://www.amagi.com/resources/insights-ai-impact-media-industry">https://www.amagi.com/resources/insights-ai-impact-media-industry</a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stelia.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://stelia.substack.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[AI governance is still broken, and the clock is ticking]]></title><description><![CDATA[Let&#8217;s not sugarcoat it; the state of AI governance is a mess.]]></description><link>https://stelia.substack.com/p/ai-governance-is-still-broken-and-the-clock-is-ticking-52c5c3a736d7</link><guid isPermaLink="false">https://stelia.substack.com/p/ai-governance-is-still-broken-and-the-clock-is-ticking-52c5c3a736d7</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Fri, 11 Jul 2025 07:16:55 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3ee696b5-e787-470f-a3d8-4b99eaf4a0b4_800x450.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!of0y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!of0y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg 424w, https://substackcdn.com/image/fetch/$s_!of0y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg 848w, https://substackcdn.com/image/fetch/$s_!of0y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!of0y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!of0y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!of0y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg 424w, https://substackcdn.com/image/fetch/$s_!of0y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg 848w, https://substackcdn.com/image/fetch/$s_!of0y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!of0y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6d0f4a2-c67a-4c88-89bf-2ab914ea2ba8_800x450.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Let&#8217;s not sugarcoat it; the state of AI governance is a mess. Worse yet, it&#8217;s showing no signs of getting better anytime soon.</p><p>Just this month (July 2025), Washington saw a dramatic clash over the future of AI regulation. The Trump-aligned proposal to block states from enacting their own AI laws (a move washed in to the &#8220;Big, Beautiful Bill&#8221;) was flatly rejected in the Senate. The message from lawmakers&#8202;&#8212;&#8202;States should be free to protect their residents, even if it means a messy web of conflicting rules.</p><p>Meanwhile, over in Brussels, the European Union is forging ahead with its own AI Act, brushing aside rumours of delays to enactment and signalling its intent to lead on AI governance.</p><p>This transatlantic divergence should worry anyone paying attention. AI is a borderless, general-purpose technology, yet we&#8217;re attempting to regulate it with tools and mindsets built for a pre-digital world, fractured by political agendas and local priorities.</p><p>Lawmakers writing these AI regulations lack the technical understanding needed to grasp the technology they&#8217;re governing, ignoring highly experienced legal professionals and academics with years of experience in the sector. This raises the question: are these laws designed to protect people or simply to create frameworks governments can navigate? Without expert insight, regulation risks becoming symbolic, not practical and ultimately ineffective. This isn&#8217;t protection, this is chaos.</p><h3>A patchwork that is already&nbsp;tearing</h3><p>The current regulatory approach is a tangled mess; a chaotic patchwork of state-level and national laws, each with its own definitions, demands and deadlines. The problem isn&#8217;t just complexity, it&#8217;s incoherence.</p><p>AI doesn&#8217;t neatly respect industry boundaries, let alone state or national ones. A single model can be deployed globally, retooled across sectors and evolve faster than any regulator can track. Yet instead of crafting a unified framework, lawmakers are scrambling to patch together fragmented responses while the ground shifts beneath them.</p><p>In the US, proponents of the failed federal AI moratorium (including some prominent tech figures) warned that a labyrinth of 50 state regulations could cripple innovation and hand China a competitive edge. OpenAI&#8217;s own Sam Altman bluntly told senators that managing compliance across so many jurisdictions would be practically impossible.</p><p>But their warnings weren&#8217;t enough.</p><p>The AI freeze was stripped from the bill leaving America exactly where it was to begin with: fragmented, uncertain and without a national game plan.</p><h3>The big AI&nbsp;bet</h3><p>Across the pond the EU is betting big on regulation. The AI Act, modelled on the bloc&#8217;s approach to the GDPR, aims to impose a sweeping risk-based framework on AI systems used within its borders. On paper it sounds like progress with transparency for low-risk systems, tight controls for high-stakes use cases like hiring or policing and new rules for the looming threat of general-purpose AI (GPAI).</p><p>But the tech is evolving faster than the rules. A single AI model might help you write an email today and guide a cancer diagnosis tomorrow. Trying to classify these systems by risk level is like nailing jelly to an ever moving wall. And the Act&#8217;s GPAI provisions? Vague, reactive and already chasing a moving target with consultations still being undertaken on what GPAI even is (and we&#8217;ve not even got to artificial <em>general</em> intelligence yet&#8230;).</p><p>To their credit, EU lawmakers have acknowledged the problem and tried to adjust late in the game. But even as they push ahead, the uncomfortable truth remains: Europe&#8217;s AI rules may be outdated before the ink is dry.</p><h3>One world, many&nbsp;rules</h3><p>At the heart of the problem is a basic contradiction; AI is global, regulation isn&#8217;t.</p><p>A generative model might be trained in California, fine-tuned in London, hosted in Singapore, and used by someone in Nairobi. So whose laws apply? Whose values shape the outcomes? Who is accountable when things go wrong? And how does this even become &#8216;Sovereign&#8217;.</p><p>Right now there are no clear answers. The US and EU are each pursuing their own philosophies&#8202;&#8212;&#8202;decentralised experimentation versus centralised oversight. Other countries like the UK, are forging different paths altogether. And while there are talks at the G7 and the UN, global coordination remains little more than a diplomatic soundbite.</p><p>What we&#8217;re left with is a high-stakes experiment playing out in real time. The world is testing governance models in parallel, without consensus, without interoperability and without a safety net.</p><h3>Time to wake&nbsp;up</h3><p>Here&#8217;s the hard truth; local rules are no match for global systems.</p><p>AI is rapidly becoming foundational to how we live, work, make decisions and frankly exist. If we don&#8217;t find a way to align on basic principles of safety and accountability, we risk letting this technology outpace our ability to control it.</p><p>Yes, fragmented governance offers room for experimentation. Yes, different regions have different values. But let&#8217;s not pretend that a patchwork of uncoordinated rules is a long-term solution. At best, it&#8217;s a stopgap. At worst, it&#8217;s a recipe for regulatory chaos and ethical disaster.</p><p>The world doesn&#8217;t need one-size-fits-all regulation, but it <em>does</em> need a shared vision. Without it, we&#8217;re not governing AI, we&#8217;re just reacting to it. And sooner or later, the consequences of that passivity will catch up with us.</p><p>The question is no longer whether we need global coordination; it&#8217;s whether we&#8217;ll get there before it&#8217;s too late.</p><p><em>by Imogen Armstrong, Chief Legal Officer, Stelia</em></p>]]></content:encoded></item><item><title><![CDATA[3.5× Faster Inference with Smarter Quantisation: The QServe Playbook]]></title><description><![CDATA[We asked our VP Engineering what&#8217;s the latest thing that&#8217;s made you stop and think: &#8220;wow that&#8217;s clever&#8221;. Here&#8217;s what he had to say:]]></description><link>https://stelia.substack.com/p/3-5-faster-inference-with-smarter-quantisation-the-qserve-playbook-46e35db553a3</link><guid isPermaLink="false">https://stelia.substack.com/p/3-5-faster-inference-with-smarter-quantisation-the-qserve-playbook-46e35db553a3</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Tue, 08 Jul 2025 10:58:08 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/7e92709e-030c-4ba8-b83d-481d80a2e135_800x450.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BeNJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BeNJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BeNJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BeNJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BeNJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BeNJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BeNJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BeNJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BeNJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BeNJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60a6f76d-b861-4e8a-ad8a-53d38020d584_800x450.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>We asked our VP Engineering what&#8217;s the latest thing that&#8217;s made you stop and think: &#8220;wow that&#8217;s clever&#8221;. Here&#8217;s what he had to say:</p><blockquote><p><em>&#8220;I tend to be pretty selective about what catches my attention.&#8221; </em>David Hughes, Stelia VP Engineering, <em>&#8220;Everyone these days is obsessed with &#8216;We don&#8217;t get enough tokens out of this GPU, let&#8217;s just buy a bigger, more powerful GPU.&#8217; That&#8217;s not the answer. There&#8217;s an interesting bit of software from a MIT HAN Lab project at the minute called QServe that flips the whole problem on its head [1].&#8221;</em></p></blockquote><h3>Challenging Scale-Up&nbsp;Thinking</h3><p>That frustration with brute-force scaling is exactly why QServe, the project Dave called out from MIT, stood out to him. QServe is an inferencing library developed by researchers at MIT, specifically focusing on reducing bottlenecks, and increasing throughput in LLM serving&#8202;&#8212;&#8202;a clear rejection of the <em>&#8220;just throw power at it&#8221; </em>mentality.</p><p>In real-world terms, quantisation can meaningfully accelerate LLM inference.</p><h3>Quantisation as a Runtime&nbsp;Strategy</h3><p>Traditionally, deep learning relies on precision formats like FP16, FP32, FP64, and INT8. As research has advanced, there has been a clear push beyond INT8, exploring even lower precision such as INT4. However, INT4 quantisation techniques have mostly been effective in low-batch, edge-style LLM inference, while struggling to deliver gains for large-batch cloud-based serving. That shortfall is primarily due to runtime penalties introduced during dequantisation of either weights or partial sums on GPUs.</p><p>To address this, QServe implements a W4A8KV4 quantisation algorithm (meaning 4-bit weight, 8-bit activation, 4-bit KV cache), following a 4&#8211;8&#8211;4 pattern known as QoQ. This approach achieves measurable speedup by recognising a key insight: the efficiency of LLM serving is often critically bottlenecked by operations on low-throughput CUDA cores. QoQ tackles this with progressive quantisation, lowering dequantisation overhead and pairs it with SmoothAttention to mitigate the accuracy losses typical of 4-bit quantisation. Additional strategies include compute-aware weight reordering, register-level parallelism to reduce quantisation latency, and making fused attention memory-bound to fully harness the performance gains of KV4 quantisation.</p><h3>The Performance Picture</h3><p>Benchmarks from the QServe team [1] show throughput gains of 1.2&#8211;1.4x on LLaMA-3&#8211;8B and up to 3.5x on Qwen-1.5&#8211;72B, with L40S GPUs even outperforming A100-class setups in several scenarios. These results tie back to QServe&#8217;s efforts to minimise dequantisation bottlenecks on GPU cores, which are a well-known source of INT4 inefficiency. Reported token costs were up to three times lower, thanks to reduced kernel overhead and higher hardware utilisation. That said, the paper notes remaining challenges around runtime stability under high concurrency and large context windows. It&#8217;s promising but not guaranteed to hold under every production scenario.</p><h3>The Implementation Trade-Offs</h3><p>Beyond consistency, this kind of adaptive quantisation also isn&#8217;t trivial to roll out. The entire inference stack&#8202;&#8212;&#8202;from tokenizers to kernel-level scheduling&#8202;&#8212;&#8202;needs to tolerate non-uniform precision at runtime. That demands deeper model introspection, more robust calibration workflows and a willingness to invest in kernel-level engineering, especially for workloads beyond the benchmarked standard models. It&#8217;s not a drop-in fix, but for anyone willing to trade a bit of complexity for massive efficiency gains, QServe offers a solid blueprint.</p><h3>What History Teaches&nbsp;Us</h3><p>Dave noted the similarities between QServe&#8217;s solution and earlier industry breakthroughs:</p><blockquote><p>&#8220;In a throwback to ye olden days of the storage world, we were presented with (seemingly) hard limits of, for example, 1TB of physical storage&#8202;&#8212;&#8202;and yet we found ways to work around this in software. Enter: on-the-fly compression which, as algorithms developed further, became more and more efficient and you had different compression algorithms for different use cases, e.g. LZ4 vs GZIP and even continued to innovate in later years with ZSTD becoming even more efficient and delivering the sweet spots of both the aforementioned.</p></blockquote><blockquote><p>We will undoubtedly reach the same point in AI workloads, where we are presented with (again, seemingly) hard limits of the physical infrastructure involved and we will have a need to get clever. QServe is a glimpse of what getting clever might actually look like.&#8221;</p></blockquote><p>Ultimately, QServe is a wake-up call about breaking habits. While the industry has been conditioned to worship bigger GPUs on reference architectures&#8202;&#8212;&#8202;driving up costs and complexity in the name of vendor roadmaps&#8202;&#8212;&#8202;QServe is a reminder that smart software still comes out on top.</p><p>As Dave accurately puts it:<br><em>&#8221;We&#8217;ve accepted that hardware has limitations, but there&#8217;s no reason to just lie down. You can be clever and work around it with software. Ingenuity always wins in the end.&#8221;</em></p><h3>Why This Matters for&nbsp;Stelia</h3><p>This is just one example of what happens when ingenuity reclaims center stage, refusing to be boxed in by hardware limitations. That principle resonates strongly with Stelia&#8217;s own purpose. At Stelia, we are committed to advancing artificial intelligence that is not just more powerful, but more purposeful&#8202;&#8212;&#8202;adaptive, transparent, and human-centered. While we continue to explore advanced quantisation, co-designed inference stacks, and systems-level optimisations like those demonstrated by QServe, our broader goal is to ensure these technologies serve humanity with responsibility, fairness and long-term impact. It is systems-led ideas like QServe that illustrate a path toward this goal&#8202;&#8212;&#8202;by proving that thoughtful engineering can break barriers and expand what AI is capable of.</p><h3>References</h3><p>[1] QServe Project&#8202;&#8212;&#8202;<em>W4A8KV4 Quantization and System Co-design for Efficient LLM Serving</em>: <a href="https://hanlab.mit.edu/projects/qserve">QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving</a><br>GitHub: <a href="https://github.com/mit-han-lab/omniserve">GitHub&#8202;&#8212;&#8202;mit-han-lab/omniserve: [MLSys&#8217;25] QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving; [MLSys&#8217;25] LServe: Efficient Long-sequence LLM Serving with Unified Sparse Attention</a><br>Benchmarks Dataset: <a href="https://huggingface.co/datasets/mit-han-lab/QServe-benchmarks">mit-han-lab/QServe-benchmarks &#183; Datasets at Hugging Face</a><br>Lead Researcher: Song Han</p><p>Lab: MIT HAN Lab</p>]]></content:encoded></item><item><title><![CDATA[The Artificial Tides Are Changing]]></title><description><![CDATA[The first AI era built the engine; the second is about using it. Stelia is architected for this shift.]]></description><link>https://stelia.substack.com/p/the-artificial-tides-are-changing-bc39e5a49e5d</link><guid isPermaLink="false">https://stelia.substack.com/p/the-artificial-tides-are-changing-bc39e5a49e5d</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Fri, 04 Jul 2025 11:04:49 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/cce8d10b-9004-461f-8490-dcb6279bc4da_800x450.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BFfL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BFfL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BFfL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BFfL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BFfL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BFfL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BFfL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg 424w, https://substackcdn.com/image/fetch/$s_!BFfL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg 848w, https://substackcdn.com/image/fetch/$s_!BFfL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!BFfL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be0aaf0-34a7-4b9b-922a-d71de6fa490f_800x450.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p><em>Tobias Hooton, CEO and Founder, Stelia</em></p><h3>A founder&#8217;s view on the next era of&nbsp;AI</h3><p>Most industry observers still believe we&#8217;re in the early days of AI. They&#8217;re wrong.</p><p>The first era of artificial intelligence was a triumph of engineering. It was defined by breakthroughs in scale: GPUs got faster, models got bigger, projects like Stargate were launched. We built the infrastructure. We watched transformers reshape natural language, diffusion models redraw images and a totally new breed of AI-native products enter the mainstream. The energy of that period was electric. But it was also experimental, erratic and incomplete.</p><p>Today, the artificial tide is changing direction. We are entering the second era of AI and its shape is beginning to crystallise: not in academic labs or GPU clusters, but in enterprise conference rooms, legal firms, global supply chains, customer support teams, and the daily workflows of billions of people.</p><p>The first era of AI built the engine. The second era is about using it.</p><p>This moment marks a fundamental transition. And like any transition it comes with confusion and inertia. Many of the players who won in the first phase are now trying to force their way into continued relevance. But the centre of gravity has moved. The epicentre has shifted to application.</p><h3>From Compute Wars to Value&nbsp;Wars</h3><p>In the first era, value accrued to those who could build compute. Access to high-end GPUs became a competitive advantage. Entire business models formed around reselling compute and leasing hardware. (sounds similar to the early 2000s of bare metal compute&#8230;)</p><p>But here&#8217;s the uncomfortable truth: compute is not the moat. At best, it&#8217;s table stakes.</p><p>The AI industry is now saturated with infrastructure providers desperate to monetise their place in the stack. They speak the language of tokens and throughput, but these are engineering metrics, not user value. And while they focus on monetising bits and bandwidth, the real opportunity has evolved.</p><p>We are now witnessing a decoupling: those who control compute will not necessarily control the future. Value has shifted away from raw hardware and toward ease of use, access, and deployment at huge scale. AI is about making intelligence accessible: reliably, securely and in the right context.</p><h3>The Second Phase of AI: Distribution at&nbsp;Scale</h3><p>We are now in the &#8220;adoption era.&#8221; In this phase, the building is effectively done. The foundational breakthroughs required to build mass-market AI products have now occurred (research continues, but the core infrastructure exists). We&#8217;re no longer waiting on hardware; we&#8217;re waiting on distribution.</p><p>In this new landscape, the questions have changed: &#8226; The critical question: how well does it integrate into a 100,000-person enterprise? &#8226; Can it reduce legal review time by 40% and meet compliance requirements? &#8226; Can you scale it across 70 languages with consistent tone, brand, and results?</p><p>This is where the next generation of AI companies will win: in the trench warfare of deployment, serviceability and global distribution of intelligence.</p><p>And just as importantly, in usability.</p><p>Ease of use is not a feature; it&#8217;s the differentiator. Organisations want systems that augment their own, with minimal friction and maximum return. We&#8217;re seeing the rise of truly horizontal AI: core operational enhancements that drive measurable efficiency.</p><p>The era of experimentation is giving way to the era of execution.</p><h3>Why the Old Guard Will&nbsp;Struggle</h3><p>Who would be surprised by this prediction? The giants of Phase One who assume they&#8217;ll simply evolve into Phase Two. But history rarely works that way.</p><p>The incentives are misaligned. The largest compute providers (Nvidia partners, hyperscalers, GPU lessors) have built their business models around consumption primarily for a deeply technical niche audience, not value creation. They profit when you use more resources, not when you deliver better results. This creates a fundamental conflict: their growth is not your growth. Their scale is not your efficiency.</p><p>As enterprises begin to demand value instead of volume, these providers are being caught off-guard. They are structured to win hardware cycles, not workload loyalty.</p><p>We are already seeing cracks in their armour. Once-untouchable players are spinning up foundation model labs and launching APIs on top, hoping to capture enterprise spend. But APIs are not ecosystems, and tokens are not outcomes. Without deep understanding of user needs, without vertical integration, and without true control of the full stack, they risk becoming irrelevant middlemen in a fast-consolidating value chain.</p><p>While GPU providers remain a critical part of the AI ecosystem, powering the infrastructure that enables training and inference at scale, they have shifted from being the epicentre of innovation to a foundational layer in a much larger value stack. As the focus moves from raw compute to real-world utility, the differentiators are in how seamlessly intelligence can be deployed, adopted, and trusted. GPUs are essential but the future is decided elsewhere.</p><h3>The Rise of Full-Stack Intelligence</h3><p>In the second AI phase, the winners will be those who own the entire value chain: from silicon to sentiment. That means building systems, not just models.</p><p>This requires radical vertical integration and deep domain expertise. Control of the hardware. Custom foundation models. Scalable orchestration layers. And intuitive interfaces. All bound by a singular goal: make intelligence useable at scale.</p><p>This is where most AI startups fall short. They pick a piece of the puzzle (an LLM wrapper, a UX shell, an API filter) and try to build a business. Without full-stack control, they are at the mercy of upstream providers and downstream constraints. The result is fragmentation, not acceleration.</p><p>To thrive in this new world, you need to abstract the complexity. Hide the infrastructure. Deliver intelligence as a service: in the human sense, not the cloud computing sense.</p><p>And it must work across domains. Across cultures. Across languages. The next phase of AI is multicontextual. It&#8217;s about understanding what a user says, why they say it, when they say it, and how best to respond.</p><p>This is a systems challenge. And only a few players are positioned to solve it.</p><h3>Why Stelia Already Has the&nbsp;Answer</h3><p>At Stelia, we&#8217;ve been building for this moment for over 5 years, through our own deep research labs across multiple disciplines, we&#8217;ve centred ourselves around creating the technology stack truly distributed AI requires. A full-stack AI company: from GPU-level compute infrastructure, to in-house model training, to orchestration systems that serve hundreds of millions of users across sectors and geographies.</p><p>We didn&#8217;t chase hype cycles. We architected for scale. Our platforms were trained for the complexities of real-world deployment, not just benchmarks.</p><p>Where others scrambled to bolt on value, we built it into the foundation.</p><p>Stelia&#8217;s stack is an intelligence platform designed for mass adoption. From the kernel to the agentic interface, every component is optimised for enterprise use, regulatory clarity, and world-class experience.</p><p>This is Phase Two, purpose-built.</p><p>The artificial tides are changing and with them, the old paradigms are eroding. The future will be owned by those who can deliver the deepest impact.</p><p>At Stelia, we are the shift. And we&#8217;re just getting started.</p>]]></content:encoded></item><item><title><![CDATA[The €200 Billion AI Gigafactory Delusion: Why Europe Misunderstands the Algorithmic Age]]></title><description><![CDATA[In February 2025, European leaders announced InvestAI, a &#8364;200 billion initiative to build AI &#8220;gigafactories&#8221; across the continent&#8230;]]></description><link>https://stelia.substack.com/p/the-200-billion-ai-gigafactory-delusion-why-europe-misunderstands-the-algorithmic-age-cb618c23d4cf</link><guid isPermaLink="false">https://stelia.substack.com/p/the-200-billion-ai-gigafactory-delusion-why-europe-misunderstands-the-algorithmic-age-cb618c23d4cf</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Wed, 02 Jul 2025 06:01:26 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9253de06-c166-479b-be8f-fba29b18e7d1_800x450.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!N7xk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!N7xk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg 424w, https://substackcdn.com/image/fetch/$s_!N7xk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg 848w, https://substackcdn.com/image/fetch/$s_!N7xk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!N7xk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!N7xk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!N7xk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg 424w, https://substackcdn.com/image/fetch/$s_!N7xk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg 848w, https://substackcdn.com/image/fetch/$s_!N7xk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!N7xk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5beacb0-f7ae-4701-8701-62e49fba6899_800x450.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In February 2025, European leaders announced InvestAI, a &#8364;200 billion initiative to build AI &#8220;gigafactories&#8221; across the continent. Commission President Ursula von der Leyen unveiled the programme on February 11th, describing it as creating &#8220;unprecedented capital through InvestAI for European AI gigafactories&#8221; and &#8220;a unique public-private partnership, akin to a CERN for AI.&#8221; France pledged &#8364;109 billion in private investment, whilst &#8364;20 billion was allocated specifically for four gigafactories housing approximately 100,000 last-generation AI chips. The rhetoric was stirring: technological sovereignty through infrastructure ownership, democratic oversight through geographical control, European values embedded in European silicon.</p><p>This represents one of the most expensive category errors in the history of technology policy.</p><p>The fundamental premise underlying Europe&#8217;s gigafactory strategy (that controlling physical infrastructure translates to meaningful governance of AI systems) reflects a profound misunderstanding of where power actually resides in algorithmic societies. This creates what we term the &#8220;sovereignty paradox&#8221;: the more sophisticated AI systems become, the less meaningful traditional territorial control mechanisms prove to be. Distributed AI systems resist the centralised monitoring and control that territorial governance assumes, whilst AI models exist as mathematical objects that can be instantly copied and deployed anywhere without degradation.</p><p>Recent events underscore this disconnect perfectly. When German authorities moved to block the DeepSeek AI app for unlawful data transfers to China, they targeted app store distribution rather than the underlying AI model itself. The mathematical weights powering DeepSeek&#8217;s capabilities remain unchanged and deployable through countless other channels, whilst enforcement focuses on controlling access points rather than algorithmic behaviour. This illustrates precisely how territorial governance mechanisms prove inadequate for governing distributed AI systems.</p><h3>The Mathematical Absurdity of Territorial Control</h3><p>The gigafactory approach rests on a fundamental category error about the nature of AI models themselves. Once training completes, an AI system exists as nothing more than a specific configuration of numerical parameters: billions or trillions of floating-point numbers arranged in precise mathematical relationships. These weights carry no inherent geographical identity, jurisdictional markers, or territorial constraints.</p><p>Consider the mathematical reality: a neural network trained on European data using European computational resources becomes functionally indistinguishable from an identical configuration of weights derived through any other process. The model&#8217;s &#8220;European-ness&#8221; exists only in its provenance metadata, not in its mathematical structure or functional characteristics. Two models with identical weights will produce identical outputs regardless of where they were trained, what data was used, or which regulations governed their creation.</p><p>This mathematical universality renders territorial control mechanisms absurd. A healthcare AI model trained in a European gigafactory using data from EU citizens under EU regulations can be perfectly replicated by simply copying its numerical weights. Once copied, those weights can be embedded in medical devices sold globally, integrated into healthcare systems operating under entirely different regulatory frameworks, or deployed on consumer smartphones without any technical mechanism for maintaining European governance.</p><p>The copying process itself reveals the deeper absurdity. Unlike physical goods that degrade through reproduction, mathematical objects can be perfectly duplicated infinitely without loss of fidelity. A single European-trained model can spawn millions of identical copies operating simultaneously across every jurisdiction on Earth, each mathematically indistinguishable from the &#8220;sovereign&#8221; original yet entirely beyond European control.</p><h3>Distribution Trends and Governance Fragmentation</h3><p>Whilst European policymakers focus on centralised training infrastructure, compelling technical and economic forces are driving AI inference towards distributed edge deployment. Real-time applications cannot tolerate the latency inherent in cloud-based inference: autonomous vehicles cannot wait 200 milliseconds for cloud processing during emergency braking; medical devices monitoring cardiac rhythm cannot pause for network connectivity during critical moments; industrial control systems require local processing capability that operates independently of network infrastructure and regulatory oversight.</p><p>Privacy requirements accelerate this distribution trend. Healthcare AI processing sensitive patient data locally avoids both regulatory complexity and data transmission risks. Financial algorithms running on consumer devices eliminate the legal and technical vulnerabilities associated with transmitting sensitive information to centralised processing facilities. Industrial systems processing proprietary operational data avoid competitive intelligence risks by keeping computation local.</p><p>The economics further reinforce distribution. Edge inference eliminates ongoing cloud computing costs, reduces bandwidth requirements, and enables offline operation. The global edge AI market, valued at $8.2 billion in 2024 and projected to reach $55.6 billion by 2030, reflects genuine economic value creation rather than regulatory compliance theatre.</p><p>These technical and economic forces create governance fragmentation that strikes at the heart of the gigafactory model. A single AI model trained in a European facility might simultaneously operate on German automobiles, French medical devices, Italian manufacturing equipment, and American smartphones, each subject to different regulatory regimes with no technical mechanism for unified governance. The temporal dimension intensifies this challenge: when AI models make split-second decisions on mobile devices, traditional accountability mechanisms prove inadequate for computational events occurring too quickly and too locally for meaningful oversight using territorial governance frameworks.</p><p>Early indications of ephemeral agent networks and autonomous agent-to-agent frameworks suggest even more challenging governance scenarios ahead. As AI systems evolve towards autonomous coordination patterns, these temporal and jurisdictional challenges will intensify exponentially.</p><h3>The Sovereignty Paradox in&nbsp;Practice</h3><p>At the architectural level, distributed systems resist the centralised monitoring and control mechanisms that territorial governance assumes. Edge AI operates through peer-to-peer networks, mesh topologies, and intermittently connected devices that collectively create computational capabilities without central coordination. Emerging agent-to-agent frameworks suggest even more autonomous coordination patterns ahead. These emergent system properties cannot be governed through control over any specific infrastructure component, including the facilities where models were originally trained.</p><p>The sovereignty paradox thus reveals a category error at the heart of European AI policy: attempting to govern inherently global, distributed, and stateless systems through territorial control mechanisms designed for physical, localised, and persistent objects. The &#8364;200 billion gigafactory investment doubles down on this error by strengthening the least relevant aspect of AI governance (control over training location) whilst ignoring the most consequential challenge (governance of distributed deployment).</p><h3>Opportunity Cost and Alternative Approaches</h3><p>Perhaps the most damaging aspect of the gigafactory approach lies not in what it attempts to achieve, but in what it prevents. The &#8364;200 billion investment represents resources that could fund governance innovations addressing the genuine challenges of democratic accountability in algorithmic societies.</p><p>Consider what &#8364;200 billion could accomplish if directed towards governance capability rather than training infrastructure: comprehensive legal frameworks for algorithmic transparency; technical infrastructure for citizen participation in AI development; international institutions for democratic AI governance; research and development for privacy-preserving collaborative AI; community capacity building for algorithmic accountability.</p><p>These investments would address the actual governance challenges posed by AI systems: ensuring algorithmic decisions can be explained and contested; enabling meaningful citizen participation in the development of systems that shape their lives; creating accountability mechanisms that work across jurisdictional boundaries; building technical infrastructure that embeds democratic values into system architecture.</p><p>Technical transparency mechanisms can embed governance properties directly into AI models through cryptographic attestation, making compliance verifiable regardless of deployment context. Rather than relying on territorial control over training facilities, these approaches create technical constraints that ensure algorithmic behaviour aligns with democratic values wherever deployment occurs. Participatory governance mechanisms can enable meaningful citizen input into algorithmic system design without requiring territorial control over computational infrastructure.</p><p>This misallocation occurs precisely when the window for governance innovation remains open. AI systems are still developing; technical standards remain fluid; international norms are still forming. European leadership in governance innovation could establish frameworks that other democracies adopt, creating global infrastructure for algorithmic accountability.</p><h3>Beyond the Gigafactory Delusion</h3><p>The gigafactory delusion represents more than misguided technology policy; it embodies a fundamental misunderstanding of power in algorithmic societies. By investing &#8364;200 billion in training infrastructure that provides no meaningful governance capability over distributed deployment, European leaders reveal dangerous confusion about where control actually resides in AI systems.</p><p>The technical realities of distributed AI systems, and the emerging prospect of ephemeral agent networks operating beyond traditional oversight mechanisms, demand nothing less than a fundamental reimagining of how democratic societies govern algorithmic power. The gigafactory approach represents exactly the wrong answer to exactly the right question.</p><p>Democratic societies deserve better: governance frameworks that enhance rather than undermine democratic accountability, investment priorities that strengthen rather than weaken citizen agency, and technology policies that serve democratic values rather than political theatre. Recognition of this error creates opportunity for approaches that could actually work if democratic societies have the intellectual courage to pursue them.</p><h3>About Stelia</h3><p>Stelia develops AI platforms that embed democratic accountability into technical architecture, proving that transparency and cooperation deliver superior outcomes to territorial control. Our work demonstrates that governance wisdom combined with technical authority can create AI systems that serve democratic values through design rather than oversight: exactly what distributed algorithmic societies require.</p>]]></content:encoded></item><item><title><![CDATA[Beyond Centralised AI: How Lyra Powers Global Distributed Intelligence]]></title><description><![CDATA[2025 isn&#8217;t even halfway through yet and it&#8217;s clear the shift from centralised to distributed inference is well underway. This represents&#8230;]]></description><link>https://stelia.substack.com/p/beyond-centralised-ai-how-lyra-powers-global-distributed-intelligence-4e83f12658a0</link><guid isPermaLink="false">https://stelia.substack.com/p/beyond-centralised-ai-how-lyra-powers-global-distributed-intelligence-4e83f12658a0</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Thu, 29 May 2025 07:00:18 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3dea3ef1-67e0-4d90-bb23-01603712dc7e_800x448.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hBXu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hBXu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hBXu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hBXu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hBXu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hBXu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hBXu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hBXu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hBXu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hBXu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0a61a19-79fb-481a-8edf-cbf3b7fff725_800x448.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>2025 isn&#8217;t even halfway through yet and it&#8217;s clear the shift from centralised to distributed inference is well underway. This represents perhaps the most significant AI execution trend of the year, and it&#8217;s at the heart of what Stelia will showcase as a Gold Sponsor at NVIDIA GTC Paris 2025 this June.</p><p>According to recent industry data, the global AI inference market is projected to reach $1.3 trillion by 2032, with distributed architectures driving remarkable cost reductions. Beyond technical tweaks this is new approach to how AI delivers business value.</p><h3>Pilotitis</h3><p>If you&#8217;ve been following the enterprise AI journey, you&#8217;re likely familiar with what industry analysts call pilotitis or the &#8220;prototype problem&#8221;&#8202;&#8212;&#8202;AI applications that work brilliantly in controlled environments but struggle when exposed to real-world conditions and scaling requirements.</p><p>&#8220;European enterprises have proven AI concepts but now face the challenge of reaching millions of users while maintaining performance,&#8221; explains Dan Scarbrough, Chief of Staff at Stelia. &#8220;The gap between today&#8217;s approaches and tomorrow&#8217;s AI-powered world creates bottlenecks that traditional methods cannot address.&#8221;</p><p>These represent fundamental limitations in latency, cost, and deployment when using centralised inference models:</p><ul><li><p><strong>Latency issues</strong> when transferring data across distributed environments</p></li><li><p><strong>High operational costs</strong> from cross-region traffic and compute resources</p></li><li><p><strong>Data privacy and regulatory compliance challenges</strong></p></li><li><p><strong>Inability to deploy effectively</strong> for real-time applications requiring millisecond responses</p></li></ul><h3>Lyra: The Distributed Intelligence Runtime</h3><p>At NVIDIA GTC Paris 2025, Stelia will showcase how their Lyra distributed intelligence runtime transforms AI concepts into adaptive applications capable of reaching millions of users without boundaries.</p><p>Lyra addresses the fundamental challenges of centralised models through its agentic-led approach. The results are compelling:</p><ul><li><p><strong>Latency reduction of up to 90%</strong>, from 150ms to as low as 10&#8211;15ms in latency-sensitive applications</p></li><li><p><strong>Bandwidth optimisation of 30&#8211;60%</strong> in content-heavy applications</p></li><li><p><strong>Agent orchestration</strong> with existing AI workflows and models</p></li><li><p><strong>Future-ready architecture</strong> through embedded, live-learning systems</p></li></ul><p>What makes Lyra particularly powerful is its intelligence engine, which orchestrates autonomous agents and multi-agent systems-providing real-time intelligence as a core component of distributed AI applications.</p><h3>The European AI&nbsp;Context</h3><p>GTC Paris 2025 represents a unique convergence of AI development in the European context. Co-located with VivaTech (Europe&#8217;s largest startup and technology event), the conference will bring together developers, researchers, and business leaders from across the continent.</p><p>The timing couldn&#8217;t be more significant. According to recent data, 78% of organisations reported using AI in 2024, up from 55% in 2023. In Europe specifically, over 45% of EU enterprises intend to adopt edge-based AI by 2026.</p><p>NVIDIA CEO Jensen Huang&#8217;s live keynote on June 11 at the D&#244;me de Paris is expected to unveil exclusive European AI advancements, setting the stage for three days of intensive knowledge exchange and partnership development.</p><h3>Collaborate to&nbsp;Win</h3><p>Perhaps most compelling about Stelia&#8217;s presence at GTC Paris is their emphasis on design partnerships over pure technology provision. Their collaborative approach acknowledges that building transformative AI applications requires more than just powerful technology, demanding a shared journey from concept to global scale.</p><p>This design partnership model brings organisations from concept to global reach through a structured approach:</p><ol><li><p><strong>Understanding unique challenges</strong> and AI vision</p></li><li><p><strong>Architecting applications</strong> on the Lyra intelligence engine</p></li><li><p><strong>Deploying autonomous agents</strong> to enterprise customers or consumer audiences</p></li><li><p><strong>Achieving global scale</strong> on a platform designed for distributed intelligence</p></li></ol><p>For enterprises facing the reality that legacy approaches aren&#8217;t built for dynamic AI applications and that development velocity slows when trying to scale beyond initial success. This represents a compelling alternative to fragmented approaches.</p><h3>Real-World Impact Across&nbsp;Sectors</h3><p>The distributed intelligence revolution that Lyra enables is already transforming key sectors:</p><p><strong>Media &amp; Entertainment</strong>: Content delivery platforms require end-to-end latency below 50 milliseconds to maintain unified experiences. Distributed intelligence enables processing that reduces server load and bandwidth consumption by up to 40%.</p><p><strong>Healthcare &amp; Telemedicine</strong>: Real-time diagnostics, predictive alerts, and personalised treatments rely on fast, private AI processing. The market for AI-embedded healthcare wearables is projected to reach 350 million units by 2026.</p><p><strong>Retail &amp; E-commerce</strong>: AI-driven personalisation is expected to contribute $800 billion in new global retail revenue by 2025, with e-commerce retailers generating 10&#8211;30% of revenue from AI-driven suggestive selling.</p><h3>Connect with Stelia at GTC Paris&nbsp;2025</h3><p>For organisations looking to scale their AI initiatives, Stelia&#8217;s team will be available throughout GTC Paris 2025, June 10&#8211;12 at the Paris Expo Porte de Versailles. To learn more about Stelia or to book a meeting during the event: visit Stelia, <a href="https://stelia.typeform.com/GTCParis2025?utm_source=newsroom&amp;utm_medium=organic&amp;utm_campaign=content">book a meeting</a> or contact <a href="mailto:lyra@stelia.ai">lyra@stelia.ai</a></p><p><em>Originally published at <a href="https://newsroom.stelia.ai/beyond-centralised-ai-how-lyra-powers-global-distributed-intelligence/">https://newsroom.stelia.ai</a> on May 29, 2025.</em></p>]]></content:encoded></item><item><title><![CDATA[Stelia Announced as Gold Sponsor for NVIDIA GTC Europe 2025]]></title><description><![CDATA[Stelia, the distributed runtime for AI, is proud to announce their participation as a Gold Sponsor of NVIDIA GTC Paris 2025, colocated with&#8230;]]></description><link>https://stelia.substack.com/p/stelia-announced-as-gold-sponsor-for-nvidia-gtc-europe-2025-0d21029268f0</link><guid isPermaLink="false">https://stelia.substack.com/p/stelia-announced-as-gold-sponsor-for-nvidia-gtc-europe-2025-0d21029268f0</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Tue, 27 May 2025 13:52:16 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a9687924-2801-431d-b0f1-23a6cd98bb90_800x448.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!X3Lu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!X3Lu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!X3Lu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!X3Lu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!X3Lu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!X3Lu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!X3Lu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!X3Lu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!X3Lu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!X3Lu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0458b867-8e04-4fdd-8780-1d32bace7f48_800x448.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Stelia, the distributed runtime for AI, is proud to announce their participation as a Gold Sponsor of NVIDIA GTC Paris 2025, colocated with VivaTech from June 10&#8211;12 at the Paris Expo Porte de Versailles.</p><p>As organisations move beyond AI experimentation to production deployment, they face fundamental challenges that traditional approaches cannot solve. At NVIDIA GTC Paris 2025, Stelia will showcase Lyra, the distributed AI runtime that transforms concepts into adaptive applications capable of reaching millions of users without technical barriers or complexity.</p><p>Lyra&#8217;s intelligence engine orchestrates all components of AI, providing borderless deployment and simple integration as a core component of scale-out AI applications. This architecture enables organisations to deliver consistent experiences regardless of location or device, eliminating the cost unpredictability and fragmentation that typically occurs when scaling beyond initial success.</p><p>Stelia experts will be available throughout the three-day event to demonstrate how Lyra&#8217;s distributed intelligence approach enables European enterprises and developers to build sophisticated AI applications that adapt, evolve, and create unprecedented value without rebuilding their underlying infrastructure.</p><h3>Key Capabilities of&nbsp;Lyra</h3><p>Lyra is purpose-built for intelligence at global scale, delivering:</p><ul><li><p>Borderless AI deployment&#8202;&#8212;&#8202;Simply distribute AI capabilities across regions and environments</p></li><li><p>Adaptive runtime orchestration&#8202;&#8212;&#8202;Real-time personalisation and intelligence across global deployments</p></li><li><p>Simplified integration&#8202;&#8212;&#8202;Connect with existing AI workflows and models without technical barriers</p></li><li><p>Future-proof architecture&#8202;&#8212;&#8202;As AI evolves, your applications evolve with it</p></li></ul><p>&#8220;European enterprises have proven the AI concept but now face the challenge of scaling to millions of users while maintaining performance,&#8221; said Dan Scarbrough, Chief of Staff at Stelia. &#8220;By partnering with forward-thinking organisations at GTC Paris, we&#8217;re demonstrating how our collaborative approach delivers tomorrow&#8217;s AI-powered experiences and measurable business outcomes, without boundaries.&#8221;</p><p>Meet the Stelia executive team, including Tobias Hooton CEO and Dan Scarbrough Chief of Staff, at NVIDIA GTC Paris 2025, June 10&#8211;12 at the Paris Expo Porte de Versailles.</p><p>For more information about Stelia and how we can benefit your organisation, please visit <a href="http://www.stelia.io/">Stelia</a> or contact <a href="mailto:connect@stelia.io">lyra@stelia.ai</a></p><h3>About Stelia:</h3><p>Stelia delivers intelligence at global scale through Lyra, the distributed AI runtime. Lyra transforms AI concepts into adaptive applications, reaching millions of users without barriers or complexity. Lyra orchestrates AI components for borderless deployment and simple integration, driving experiences that adapt, evolve, and unlock unprecedented value for innovators ready to harness AI&#8217;s full potential. Stelia delivers measurable business outcomes, without boundaries.</p><h3>About NVIDIA GTC Paris&nbsp;2025</h3><p><a href="https://www.nvidia.com/en-eu/gtc/session-catalog/?tab.allsessions=1700692987788001F1cG#/">NVIDIA GTC Paris 2025</a> brings together Europe&#8217;s most innovative AI developers, engineers, and researchers to explore advancements in AI, accelerated computing, and related fields. Held from June 10&#8211;12 at the Paris Expo Porte de Versailles, the conference features a keynote by NVIDIA CEO Jensen Huang on June 11 at the D&#244;me de Paris. The event emphasises real-world AI applications, agentic AI, generative AI, and cloud infrastructure, fostering collaboration between NVIDIA experts and the European tech community.</p><h3>About VivaTech&nbsp;2025</h3><p><a href="https://vivatechnology.com/">VivaTech</a> is Europe&#8217;s largest startup and technology event, held annually in Paris, France. The 2025 edition, running from June 11&#8211;14, will welcome over 165,000 attendees from more than 160 countries, featuring 2,800+ exhibitors and 13,500+ startups. The event focuses on technological innovation, with special emphasis on AI advancement, sustainability, and cross-border collaboration.</p><p><em>Originally published at <a href="https://newsroom.stelia.ai/stelia-announced-as-gold-sponsor-for-nvidia-gtc-europe-2025/">https://newsroom.stelia.ai</a> on May 27, 2025.</em></p>]]></content:encoded></item><item><title><![CDATA[US Forges Historic AI Partnerships With UAE and Saudi Arabia]]></title><description><![CDATA[Middle East Emerges as Key Hub for Distributed Intelligence]]></description><link>https://stelia.substack.com/p/us-forges-historic-ai-partnerships-with-uae-and-saudi-arabia-f8ce7c866cb1</link><guid isPermaLink="false">https://stelia.substack.com/p/us-forges-historic-ai-partnerships-with-uae-and-saudi-arabia-f8ce7c866cb1</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Mon, 19 May 2025 10:09:47 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5e3dfed0-0034-43cf-9753-90c2cafe96dc_800x448.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nv-l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nv-l!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Nv-l!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Nv-l!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Nv-l!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nv-l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Nv-l!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Nv-l!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Nv-l!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Nv-l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ae8ea78-6c9d-401d-8ef9-e5b5c3fbe091_800x448.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><h3>Middle East Emerges as Key Hub for Distributed Intelligence</h3><p>The United States has signed landmark agreements with the United Arab Emirates and Saudi Arabia that will fundamentally reshape the global AI landscape. These deals unlock a trillion-dollar capital pool while positioning the Middle East as a crucial node in the emerging distributed intelligence ecosystem.</p><h3>Reshaping the Global AI Landscape</h3><p>According to SemiAnalysis, &#8220;The new U.S. pacts with the UAE and Saudi Arabia reshape the AI landscape on three fronts: Macro, Geopolitical, and Infrastructure.&#8221; By sidestepping previous export-control frameworks, Washington has opened significant capital flows for AI infrastructure development.</p><p>The UAE agreement centers around Abu Dhabi&#8217;s G42, securing an import quota of 500,000 Nvidia top-tier chips annually. The deal includes a planned 5GW AI datacenter campus, with the first 1GW phase already under construction, utilizing solar, gas, and nuclear power sources.</p><h3>Trillion-Dollar Capital Mobilization</h3><p>Saudi Arabia&#8217;s parallel $600 billion economic package includes DataVolt&#8217;s $20 billion investment in U.S. datacenters, alongside major commitments from Oracle, Google, and other tech giants. The newly formed Saudi AI firm HUMAIN will deploy approximately $10 billion in AMD systems and an equal investment in Nvidia hardware.</p><p>These agreements address critical infrastructure challenges. As SemiAnalysis notes, &#8220;Energy-rich Gulf nations join the roster of trusted partners just as U.S. data-center grids hit their physical limits. Europe could have relieved the bottleneck but stumbled on power shortages and slow permitting.&#8221;</p><h3>From Centralized to Distributed Intelligence</h3><p>&#8220;We&#8217;re moving from centralized AI models to distributed intelligence networks that can scale globally,&#8221; observes Tobias Hooton, CEO of Stelia. &#8220;These agreements go way beyond building more datacenters towards a new understanding of how intelligence is distributed and activated across geography.&#8221;</p><p>The deals also reshape the competitive landscape with China. As the Department of Commerce issued guidance declaring the use of Huawei Ascend chips an export control violation, these partnerships effectively extend U.S. technological influence in the region.</p><h3>Security and Governance Frameworks</h3><p>Security considerations remain important. The White House AI Czar David Sacks has expressed confidence in management strategies, noting that physical verification is straightforward: &#8220;all one would have to do is send someone to a datacenter and count the server racks to make sure the chips are still there.&#8221;</p><h3>Industry Leaders Convene in&nbsp;Dubai</h3><p>As the industry processes these developments, key stakeholders will gather at the Digital Infrastructure Investment Forum in Dubai. Stelia CEO Tobias Hooton will participate in a panel discussion on Monday, May 19th at 2:15pm at Madinat Jumeirah, alongside representatives from TAWAL, Maerifa Solutions, and Khazna Data Centers. The panel will explore capital flows, connectivity, scaling demands, and investment models.</p><h3>Activating the Platform Intelligence Layer</h3><p>&#8220;While much focus has been on building physical infrastructure, the real transformation will come from how we activate these resources,&#8221; says Hooton. &#8220;The most innovative companies are focusing on delivering business outcomes from platform intelligence that can activate, distribute, and scale in-market.&#8221;</p><p>These historic agreements create powerful new vectors for innovation while reshaping geopolitical relationships. The trillion-dollar question becomes how effectively this distributed intelligence architecture can address our most pressing challenges at global scale.</p><p><em>Originally published at <a href="https://newsroom.stelia.ai/us-forges-historic-ai-partnerships-with-uae-and-saudi-arabia/">https://newsroom.stelia.ai</a> on May 19, 2025.</em></p>]]></content:encoded></item><item><title><![CDATA[Part 3: From Pilot to Production — Overcoming Distributed AI Challenges]]></title><description><![CDATA[Distributed intelligence provides the foundational architecture for scaling enterprise AI, delivering the latency, cost, and performance&#8230;]]></description><link>https://stelia.substack.com/p/part-3-from-pilot-to-production-overcoming-distributed-ai-challenges-6f0e84d3e966</link><guid isPermaLink="false">https://stelia.substack.com/p/part-3-from-pilot-to-production-overcoming-distributed-ai-challenges-6f0e84d3e966</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Fri, 09 May 2025 09:51:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/78172fee-21e6-4757-9f77-30bec6c5b843_800x448.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g61k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g61k!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!g61k!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!g61k!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!g61k!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g61k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g61k!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!g61k!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!g61k!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!g61k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee3324cb-7b18-4be0-8222-221097dcc7b0_800x448.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Distributed intelligence provides the foundational architecture for scaling enterprise AI, delivering the latency, cost, and performance advantages discussed in our previous analysis. However, successful implementation requires navigating significant organisational and technical challenges.</p><p>Our exploration of enterprise AI scaling has revealed both the implementation challenges facing organizations and how distributed intelligence provides the architectural foundation for success. In this final installment, we examine how Stelia&#8217;s agent-led platform converts these concepts into operational reality.</p><h3>The Build vs. Buy&nbsp;Decision</h3><p>The demands for scaled AI are substantial, with system complexity challenges cited by 66% of executives in the KPMG survey.</p><p>Organisations face two pathways:</p><ol><li><p><strong>Internal development</strong>&#8202;&#8212;&#8202;Requiring substantial capital investment and specialised expertise</p></li><li><p><strong>Platform partnership</strong>&#8202;&#8212;&#8202;Leveraging purpose-built solutions from specialised providers</p></li></ol><p>While both approaches can work, the second option significantly accelerates time-to-value. Stelia&#8217;s distributed intelligence platform removes the need for complex management, enabling enterprises to scale AI capabilities efficiently without extensive capital outlays or specialised technical teams.</p><h3>Orchestration Through Abstraction</h3><p>The key innovation driving Stelia&#8217;s approach is resource abstraction&#8202;&#8212;&#8202;a software layer that shields business users from underlying complexity while dynamically optimising system performance. This orchestration layer, powered by Stelia, delivers four critical advantages:</p><ol><li><p><strong>Simplified Management</strong>: Abstracts complex configurations, reducing the expertise needed for system orchestration and addressing the skills gaps (51%) identified in the KPMG survey.</p></li><li><p><strong>Dynamic Scalability</strong>: Automatically allocates resources to meet fluctuating AI workload demands.</p></li><li><p><strong>Cost Optimisation</strong>: Continuously evaluates and optimises resource usage for maximum efficiency.</p></li><li><p><strong>Deployment Flexibility</strong>: Supports hybrid deployment models that address data sovereignty requirements (a priority for 54% of organisations) while minimising latency (56% prioritise edge computing).</p></li></ol><h3>From Technical Complexity to Business&nbsp;Value</h3><p>Managing distributed AI systems traditionally requires specialised expertise in orchestration and scalability&#8202;&#8212;&#8202;a significant barrier given the technical skills gaps (51%) identified in the KPMG survey.</p><p>The Stelia advantage lies in transforming this complexity into intuitive workflows. Rather than requiring teams to manage intricate system configurations, Stelia&#8217;s agent-led approach provides natural language interfaces that enable non-specialists to deploy and manage sophisticated AI capabilities.</p><p>This approach extends beyond system management to several optimisation techniques that industry experts identify as critical for AI efficiency:</p><ol><li><p><strong>Model Optimisation</strong>: Stelia&#8217;s intelligent orchestration applies approaches that can reduce model complexity while preserving accuracy.</p></li><li><p><strong>Continuous Monitoring</strong>: The platform tracks utilisation in real-time, identifying bottlenecks before they impact performance.</p></li><li><p><strong>Automated Resource Management</strong>: Dynamically adjusts resources based on demand patterns, ensuring optimal performance and cost efficiency.</p></li></ol><p>These capabilities directly address the data consistency and quality concerns (64%) identified as major barriers to AI adoption. By automating synchronisation and validation across distributed nodes, Stelia ensures reliable model performance even in real-time applications.</p><h3>From Possibility to Production</h3><p>Enterprise AI stands at an inflection point. Organisations have validated AI&#8217;s business value through pilots but struggle to scale these initiatives securely and cost-effectively across operations.</p><p>Distributed intelligence, powered by Stelia&#8217;s adaptive orchestration layer, provides the foundation for this transition&#8202;&#8212;&#8202;transforming AI from experimental projects to operational capabilities. By simplifying management, enhancing scalability, and optimising costs, Stelia enables organisations to overcome the financial, technical, and regulatory barriers to enterprise-wide AI deployment.</p><p>The KPMG data makes clear that competitive advantage in AI will not come from incremental experimentation but from operational excellence at scale. Organisations that can bridge the gap between pilot success and enterprise deployment will capture disproportionate value in the AI economy.</p><p>The future of AI-driven enterprises will be defined not by those who experiment most broadly, but by those who execute most precisely.</p><p><em>Originally published at <a href="https://newsroom.stelia.ai/part-3-from-pilot-to-production-overcoming-distributed-ai-challenges/">https://newsroom.stelia.ai</a> on May 9, 2025.</em></p>]]></content:encoded></item><item><title><![CDATA[Part 2: Distributed Intelligence — The Foundation for Scale AI]]></title><description><![CDATA[To overcome the scaling challenges outlined in Part 1, leading enterprises are embracing distributed intelligence &#8212; a fundamentally&#8230;]]></description><link>https://stelia.substack.com/p/part-2-distributed-intelligence-the-foundation-for-scale-ai-d49c3a4bdf42</link><guid isPermaLink="false">https://stelia.substack.com/p/part-2-distributed-intelligence-the-foundation-for-scale-ai-d49c3a4bdf42</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Tue, 06 May 2025 12:14:28 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/83abaea7-d528-4efb-bbc7-42224cb16899_800x448.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dlIn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dlIn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dlIn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dlIn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dlIn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dlIn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dlIn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dlIn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dlIn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dlIn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27e8d436-4821-4563-9c1f-4784d004b79b_800x448.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>To overcome the scaling challenges outlined in Part 1, leading enterprises are embracing distributed intelligence&#8202;&#8212;&#8202;a fundamentally different approach to AI deployment that transcends traditional limitations.</p><p>In our previous analysis, we examined the KPMG data revealing a critical execution gap in enterprise AI, with 65% of organizations stuck in pilot phases. Now we explore how Stelia&#8217;s distributed intelligence architecture fundamentally transforms this AI scaling equation.</p><h3>The Hidden Cost of AI Inference</h3><p>While much attention focuses on model training, recent research reveals that over 80% of computational demand in AI actually comes from inference tasks&#8202;&#8212;&#8202;the everyday predictions and decisions that deliver business value. Inefficient inference operations lead to wasted resources that diminish ROI and can ultimately discourage broader AI adoption.</p><p>Organisations must recognise that scaling AI isn&#8217;t merely about performance but about sustainable economics. Stelia&#8217;s distributed intelligence platform addresses this precise challenge.</p><h3>Beyond Single-Node Computing</h3><p>Conventional AI deployments concentrate computing resources in centralised locations&#8202;&#8212;&#8202;creating bottlenecks, latency issues, and prohibitive scaling costs. Distributed intelligence takes a different approach by orchestrating AI workloads across optimal computational resources based on performance, cost, and compliance requirements.</p><p>This orchestration layer enables continuous, high-throughput processing and real-time decision-making by intelligently directing AI models across multiple nodes. The approach addresses fundamental deployment barriers by:</p><p><strong>Ensuring data sovereignty compliance</strong>&#8202;&#8212;&#8202;Processing data where regulations require while maintaining unified workflows</p><p><strong>Reducing latency through proximity</strong>&#8202;&#8212;&#8202;Critical for real-time applications in customer experience and operational analytics</p><p><strong>Optimising resource utilisation</strong>&#8202;&#8212;&#8202;Directing workloads to the most efficient processing environment for each specific task</p><h3>The Adaptive Advantage</h3><p>Stelia&#8217;s distributed intelligence approach provides significant performance advantages across critical metrics. Independent benchmarks have demonstrated that similar distributed approaches can improve:</p><ul><li><p><strong>Latency</strong>&#8202;&#8212;&#8202;Delivering superior end-user experiences by minimising the time from prompt submission to result delivery</p></li><li><p><strong>Throughput</strong>&#8202;&#8212;&#8202;Processing substantially more AI inferences within the same timeframe</p></li><li><p><strong>Cost efficiency</strong>&#8202;&#8212;&#8202;Translating directly to improved ROI and freed resources for innovation</p></li></ul><p>These aren&#8217;t theoretical projections but reflect the real-world advantages of intelligent workload orchestration. The approach directly addresses the execution challenges revealed in the KPMG survey, enabling enterprises to accelerate AI deployment beyond pilot phases.</p><h3>Direct Business&nbsp;Impact</h3><p>The business case for distributed intelligence becomes increasingly compelling as enterprises seek to scale their AI initiatives:</p><ol><li><p><strong>Cost Efficiency</strong>: Stelia&#8217;s distributed orchestration delivers up to 86.4% savings compared to traditional cloud approaches, addressing the financial barriers to scaled AI deployment identified in the KPMG survey.</p></li><li><p><strong>Data Security &amp; Compliance</strong>: By processing sensitive information locally while maintaining unified workflows, Stelia mitigates the risk management concerns cited by 82% of executives.</p></li><li><p><strong>Operational Resilience</strong>: Stelia&#8217;s intelligent orchestration layer ensures AI systems can operate efficiently across diverse workflows, critical for supporting the expanding use cases in recruitment (26%), call centers (61%), and data analysis (78%).</p></li><li><p><strong>Simplified Deployment</strong>: Stelia&#8217;s adaptive intelligence layer abstracts away infrastructure complexity, addressing the technical skills gaps (51%) that hinder adoption.</p></li></ol><p>The market validates this approach. The AI inference segment is projected to grow from $106.15 billion in 2025 to $254.98 billion by 2030 (19.2% CAGR), driven by enterprises seeking real-time insights to maintain competitive advantage in data-intensive sectors.</p><p>Having examined the technical foundation of distributed intelligence, in our final installment we address the practical implementation challenges organizations face and how Stelia&#8217;s agentic approach transforms complex orchestration into intuitive business workflows.</p><p><em>Originally published at <a href="https://newsroom.stelia.ai/part-2-distributed-intelligence-the-foundation-for-scale-ai/">https://newsroom.stelia.ai</a> on May 6, 2025.</em></p>]]></content:encoded></item><item><title><![CDATA[Stelia at POSSIBLE: Paving the Way from Pilots to Platforms]]></title><description><![CDATA[Stelia was on the ground at POSSIBLE 2025, meeting face-to-face with brand and media leaders grappling with how to make AI work at scale&#8230;]]></description><link>https://stelia.substack.com/p/stelia-at-possible-paving-the-way-from-pilots-to-platforms-24feb83f90e2</link><guid isPermaLink="false">https://stelia.substack.com/p/stelia-at-possible-paving-the-way-from-pilots-to-platforms-24feb83f90e2</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Fri, 02 May 2025 18:59:28 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/82fc5b4c-0116-4ebf-808f-c8325769a454_800x448.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kbDU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kbDU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kbDU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kbDU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kbDU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kbDU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kbDU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kbDU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kbDU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kbDU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2a1ec54-ecfe-4f48-9216-6edaed16e763_800x448.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Stelia was on the ground at POSSIBLE 2025, meeting face-to-face with brand and media leaders grappling with how to make AI work at scale. Energy was high&#8202;&#8212;&#8202;but urgency was higher. With 33% of attendees representing brand leadership (a sharp rise from 21% in 2023) and over 2,700 meetings booked through the app, this wasn&#8217;t a trade show. It was a high-stakes summit on the future of media, content, and technology. And AI was at the center of nearly every conversation.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!P_PO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!P_PO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg 424w, https://substackcdn.com/image/fetch/$s_!P_PO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg 848w, https://substackcdn.com/image/fetch/$s_!P_PO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!P_PO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!P_PO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!P_PO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg 424w, https://substackcdn.com/image/fetch/$s_!P_PO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg 848w, https://substackcdn.com/image/fetch/$s_!P_PO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!P_PO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18b9383b-0981-4faf-afe6-2eb314a18514_800x600.jpeg 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><p>But despite all the excitement around agent-led orchestration and next-gen creative ops, one message cut through: <strong>Media &amp; Entertainment isn&#8217;t struggling with ambition&#8202;&#8212;&#8202;it&#8217;s stuck in execution.</strong></p><p>We heard it repeatedly in meetings and on the floor:</p><blockquote><p><em>&#8220;We&#8217;re stuck in pilots.&#8221;</em></p></blockquote><p>This frustration wasn&#8217;t theoretical. Companies have AI initiatives in motion&#8202;&#8212;&#8202;but they&#8217;re moving slowly. Many are siloed, reliant on internal engineering, and bound by the complexity of cloud costs and rollout delays. One executive summed it up plainly:</p><blockquote><p><em>&#8220;We need to personalize globally without killing the team.&#8221;</em></p></blockquote><p>That tension&#8202;&#8212;&#8202;between scale and sustainability&#8202;&#8212;&#8202;is now defining the industry&#8217;s relationship with AI. Creative leads want autonomy. Ops teams want velocity. CMOs want content out the door faster than ever. But the stack beneath them -piecemeal tooling, experimentation-heavy workflows, and operational overhead -isn&#8217;t keeping up.</p><p>And it&#8217;s leading to hesitation:</p><blockquote><p><em>&#8220;Cloud costs and rollout delays are making us hesitate.&#8221;</em></p></blockquote><p>These weren&#8217;t isolated complaints. They surfaced across panels like <em>&#8220;The Proximity Shift: Agentic AI and the Future of All Brands&#8221;</em> and <em>&#8220;The Agentic Revolution: Reimagining Visual Content Marketing with AI.&#8221;</em></p><p>The window for &#8220;trying AI&#8221; has closed. If you&#8217;re not scaling, you&#8217;re stalling.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LirV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LirV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg 424w, https://substackcdn.com/image/fetch/$s_!LirV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg 848w, https://substackcdn.com/image/fetch/$s_!LirV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!LirV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LirV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LirV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg 424w, https://substackcdn.com/image/fetch/$s_!LirV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg 848w, https://substackcdn.com/image/fetch/$s_!LirV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!LirV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89464dc0-6a72-4d5d-b3ef-56f42cdc6a14_800x600.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>That&#8217;s where our presence at POSSIBLE was most felt&#8202;&#8212;&#8202;not in promoting a product, but in sharing a model that directly addresses the scale gap. The goal is simple: get platforms out of pilot mode and into production, faster.</p><p>At Stelia, we&#8217;ve designed for exactly these M&amp;E constraints:</p><ul><li><p><strong>Flat-fee platform models</strong> that eliminate budget ambiguity and align with how content teams actually operate</p></li><li><p><strong>Fast-deployment architecture</strong> that avoids month-long engineering timelines and cloud integration purgatory.</p></li><li><p><strong>Agent-led orchestration</strong> that liberates creative teams from manual dependencies and accelerates global personalization without team burnout.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tG4S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tG4S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg 424w, https://substackcdn.com/image/fetch/$s_!tG4S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg 848w, https://substackcdn.com/image/fetch/$s_!tG4S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!tG4S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tG4S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tG4S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg 424w, https://substackcdn.com/image/fetch/$s_!tG4S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg 848w, https://substackcdn.com/image/fetch/$s_!tG4S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!tG4S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F703bde3b-fbb6-4314-b543-1212c14cea78_800x600.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>At POSSIBLE, what connected wasn&#8217;t concept&#8202;&#8212;&#8202;it was clarity. A model that maps directly to the pain points leaders are living. When we described a platform approach that replaces engineering-heavy workflows with intelligent orchestration&#8202;&#8212;&#8202;and does it at predictable, non-extractive pricing&#8202;&#8212;&#8202;the response was immediate: <em>&#8220;This is what we&#8217;ve been waiting for.&#8221;</em></p><p>For Media &amp; Entertainment teams pushing toward AI maturity, it&#8217;s time to move from ambition to orchestration. From pilots to platforms. From bottlenecks to flow.</p><p><em>Originally published at <a href="https://newsroom.stelia.ai/stelia-at-possible-paving-the-way-from-pilots-to-platforms/">https://newsroom.stelia.ai</a> on May 2, 2025.</em></p>]]></content:encoded></item><item><title><![CDATA[Mastering Enterprise AI: Beyond Pilots to Operational Excellence]]></title><description><![CDATA[Artificial intelligence has moved beyond hype to become a critical competitive differentiator. The latest KPMG AI Quarterly Pulse Survey&#8230;]]></description><link>https://stelia.substack.com/p/mastering-enterprise-ai-beyond-pilots-to-operational-excellence-4b2dd2a53843</link><guid isPermaLink="false">https://stelia.substack.com/p/mastering-enterprise-ai-beyond-pilots-to-operational-excellence-4b2dd2a53843</guid><dc:creator><![CDATA[Stelia]]></dc:creator><pubDate>Wed, 30 Apr 2025 14:27:02 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/738e75d8-66d6-43b1-bbe2-c1ca87347543_800x448.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XC-u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XC-u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!XC-u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!XC-u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!XC-u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XC-u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XC-u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg 424w, https://substackcdn.com/image/fetch/$s_!XC-u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg 848w, https://substackcdn.com/image/fetch/$s_!XC-u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!XC-u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff18f5f5e-e35c-4605-830f-636ca7518a94_800x448.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Artificial intelligence has moved beyond hype to become a critical competitive differentiator. The latest KPMG AI Quarterly Pulse Survey (Q1 2025) confirms what forward-thinking organisations already know: 93% of enterprise leaders report that generative AI investments have strengthened their competitive position. Yet a striking paradox exists. Despite widespread experimentation and proven business value, only 11% of organisations have fully deployed AI agents, while 65% remain stuck in pilot phases. This execution gap represents the central challenge facing enterprise AI initiatives today</p><h3>The Shift from Experiment to Execution</h3><p>The data shows clear momentum in AI adoption across critical business functions:</p><ul><li><p>78% are leveraging AI for data analysis (up from 70%)</p></li><li><p>66% for administrative duties (up from 27%)</p></li><li><p>61% for call centers (up from 16%)</p></li><li><p>26% for recruitment (up from 15%)</p></li></ul><p>These impressive growth rates reflect a maturing understanding of AI&#8217;s potential. Organisations increasingly view AI as an augmentation tool rather than a replacement technology&#8202;&#8212;&#8202;76% believe it will automate tasks without replacing roles, while 69% see it empowering high performers.</p><h3>The Technical Inflection Point</h3><p>As AI initiatives move from conceptual to operational, leadership dynamics are shifting dramatically. CIO involvement in AI strategies has surged from 31% to 86%, while CEO involvement has decreased from 34% to 8%. This technical handoff reflects the evolution from strategic vision to execution reality&#8202;&#8212;&#8202;precisely where many organisations encounter significant barriers</p><h3>Investment vs. Implementation</h3><p>Organisations are backing their AI ambitions with substantial capital, planning to invest $114 million in generative AI over the coming year (up from $89 million). With 97% reporting measurable profitability, the business case is clear. Yet executives consistently identify five critical barriers preventing scaled deployment:</p><ol><li><p><strong>Risk Management (82%)</strong>: Cybersecurity vulnerabilities, compliance issues, and operational risks</p></li><li><p><strong>Data Quality (64%)</strong>: Inconsistent, siloed data undermining model accuracy</p></li><li><p><strong>Trust (35%)</strong>: Building workforce confidence amidst concerns about job displacement</p></li><li><p><strong>Technical Complexity</strong>: System architecture challenges (66%), rapid technology evolution (56%), and skills gaps (51%)</p></li><li><p><strong>Cost Efficiency</strong>: Substantial complexity and financial exposure in scaling AI operations</p></li></ol><p>The paradox is clear: enterprises have validated AI&#8217;s business value through pilots but lack the orchestration layer needed to scale these initiatives securely, reliably, and cost-effectively across global operations.</p><p>With the enterprise AI scaling challenges clearly established, our next article explores how Stelia&#8217;s distributed intelligence approach creates the architectural foundation necessary to bridge the gap between AI experimentation and enterprise-wide deployment.</p><p><em>Originally published at <a href="https://newsroom.stelia.ai/winning-with-enterprise-ai-a-3-part-report/">https://newsroom.stelia.ai</a> on April 30, 2025.</em></p>]]></content:encoded></item></channel></rss>