<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[A.I.N.S.T.E.I.N]]></title><description><![CDATA[Artificial Intelligence Norms, Standards, Trust, Ethics, Integrity & Navigation]]></description><link>https://ainstein.sanjeevaniai.com</link><image><url>https://substackcdn.com/image/fetch/$s_!l2pF!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F195ee19e-3683-4fe4-8927-dad386e22f02_1024x1024.png</url><title>A.I.N.S.T.E.I.N</title><link>https://ainstein.sanjeevaniai.com</link></image><generator>Substack</generator><lastBuildDate>Thu, 09 Apr 2026 01:17:17 GMT</lastBuildDate><atom:link href="https://ainstein.sanjeevaniai.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[A.I.N.S.T.E.I.N.]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[suneeta@sanjeevaniai.com]]></webMaster><itunes:owner><itunes:email><![CDATA[suneeta@sanjeevaniai.com]]></itunes:email><itunes:name><![CDATA[A.I.N.S.T.E.I.N.]]></itunes:name></itunes:owner><itunes:author><![CDATA[A.I.N.S.T.E.I.N.]]></itunes:author><googleplay:owner><![CDATA[suneeta@sanjeevaniai.com]]></googleplay:owner><googleplay:email><![CDATA[suneeta@sanjeevaniai.com]]></googleplay:email><googleplay:author><![CDATA[A.I.N.S.T.E.I.N.]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[A note to my subscribers]]></title><description><![CDATA[When I started A.I.N.S.T.E.I.N., I did not know who would show up.]]></description><link>https://ainstein.sanjeevaniai.com/p/a-note-to-my-subscribers</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/a-note-to-my-subscribers</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Tue, 07 Apr 2026 14:11:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!l2pF!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F195ee19e-3683-4fe4-8927-dad386e22f02_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When I started A.I.N.S.T.E.I.N., I did not know who would show up. Some of you paid. Some of you subscribed quietly. All of you mattered more than you know.</p><p>I am making a simple change today. Everything on A.I.N.S.T.E.I.N. is now free, and everyone who ever paid has free access for life. No action needed on your part.</p><p>I am also stepping back from the regular cadence for now. When I return, it will be with more clarity, more depth, and more of what actually matters to you as a practitioner navigating AI in the real world.</p><p>Thank you for being here early. That means everything.</p><p>See you soon.</p><p>Suneeta </p>]]></content:encoded></item><item><title><![CDATA[Full Code, Low Code, No Code: The AI Trust Gap Nobody Is Talking About]]></title><description><![CDATA[The Easier It Is to Deploy AI, the Harder It Is to Know What It Will Do]]></description><link>https://ainstein.sanjeevaniai.com/p/full-code-low-code-no-code-the-ai</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/full-code-low-code-no-code-the-ai</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 30 Mar 2026 14:03:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ATB5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ATB5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ATB5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png 424w, https://substackcdn.com/image/fetch/$s_!ATB5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png 848w, https://substackcdn.com/image/fetch/$s_!ATB5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png 1272w, https://substackcdn.com/image/fetch/$s_!ATB5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ATB5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png" width="1456" height="701" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:701,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3131024,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/191818571?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ATB5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png 424w, https://substackcdn.com/image/fetch/$s_!ATB5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png 848w, https://substackcdn.com/image/fetch/$s_!ATB5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png 1272w, https://substackcdn.com/image/fetch/$s_!ATB5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbab9e10c-a1e9-40a2-ad6d-1d1b6ff61dcb_2124x1022.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>                                                 Image created by AI</strong></em></p><p></p><p>Last week I wrote about a New York bill that would restrict AI systems from providing professional advice in licensed fields. A few readers asked a sharp follow-up question: does the bill apply differently depending on how the AI system was built?</p><p>The answer might surprise you, and I wil&#8230;</p>
      <p>
          <a href="https://ainstein.sanjeevaniai.com/p/full-code-low-code-no-code-the-ai">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[New York Wants to Silence Your AI Chatbot. Here Is What That Actually Means. ]]></title><description><![CDATA[When Regulators Start Scoring What AI Systems Say, Not What Companies Promise]]></description><link>https://ainstein.sanjeevaniai.com/p/new-york-wants-to-silence-your-ai</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/new-york-wants-to-silence-your-ai</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Tue, 24 Mar 2026 14:03:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!FplD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FplD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FplD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png 424w, https://substackcdn.com/image/fetch/$s_!FplD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png 848w, https://substackcdn.com/image/fetch/$s_!FplD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png 1272w, https://substackcdn.com/image/fetch/$s_!FplD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FplD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png" width="1456" height="842" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:842,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3102752,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/191818105?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FplD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png 424w, https://substackcdn.com/image/fetch/$s_!FplD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png 848w, https://substackcdn.com/image/fetch/$s_!FplD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png 1272w, https://substackcdn.com/image/fetch/$s_!FplD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c718fbb-2db8-4b31-bf3a-bb43d0fcb12e_1988x1150.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>                                                Image created by AI</strong></em></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Yesterday I wrote about the shift from measuring organizations to measuring AI systems. Today, the New York State Legislature is proving why that shift is urgent.</p><p>A bill introduced by Senator Kristen Gonzalez would restrict AI systems from providing what lawmakers call &#8220;substantive responses&#8221; in fields that require professional licenses. Medicine, law, engineering, psychology, dentistry, nursing, and other regulated professions where incorrect guidance can cause serious harm.</p><p>Read that again carefully. The bill does not say &#8220;companies must have policies about what their AI says.&#8221; It says AI systems must not provide certain types of responses. The subject of the regulation is the machine, not the organization.</p><p>This is the shift happening in real time.</p><p><strong>What the bill actually does</strong></p><p>The proposal draws a line between general information and professional advice. An AI chatbot can share educational content about, say, symptoms of a condition or how a legal process generally works. What it cannot do is cross into substantive guidance that resembles what a licensed professional would provide. It cannot offer what looks like a medical diagnosis, a legal strategy, an engineering recommendation, or a psychological assessment.</p><p>The bill also includes a private right of action. That means individuals can sue companies if their AI systems provide restricted guidance. This is not a regulatory slap on the wrist. This is litigation exposure for every company deploying a customer-facing AI system in a licensed domain.</p><p><strong>Why this matters beyond New York</strong></p><p>If you are thinking &#8220;I do not operate in New York, this does not apply to me,&#8221; think again.</p><p>New York tends to set the template. When New York moved on financial regulation, the rest of the country followed. The same pattern is already forming with AI. Colorado&#8217;s AI Act takes effect in 2026. The EU AI Act becomes fully enforceable in August 2026. The NAIC Model Bulletin on AI in insurance has been adopted by 24 states. NYC Local Law 144 already requires bias audits for automated hiring tools.</p><p>The direction is clear: regulators are moving from governing organizations that use AI to governing what AI systems actually do. And they are doing it jurisdiction by jurisdiction, which means any company deploying AI across state lines will soon face a patchwork of requirements that all ask the same fundamental question: does your AI system stay within its authorized boundaries?</p><p><strong>The measurement problem this creates</strong></p><p>Here is where the data scientist in me gets interested.</p><p>&#8220;Substantive response&#8221; is a fuzzy concept. Where exactly does educational information end and professional advice begin? When does a health chatbot cross from sharing general wellness content into offering what could be interpreted as a diagnosis? When does a legal information tool cross from explaining a process into recommending a strategy?</p><p>These are not binary questions. They are spectrum questions. And spectrum questions require quantitative measurement, not policy checklists.</p><p>Think about what an organization would need to demonstrate under this bill. Not that they have a policy saying &#8220;our AI does not give medical advice.&#8221; They would need to demonstrate that their AI system actually stays within bounds, consistently, across thousands of interactions, including edge cases where users push the boundaries with creative phrasing.</p><p>That is a behavioral measurement problem. You cannot solve it by reading the organization&#8217;s policy documents. You solve it by observing what the AI system actually says when real people interact with it. You measure boundary adherence: how often does the system recognize when it is approaching a restricted domain, and how reliably does it pull back?</p><p>This is exactly the kind of observable, quantifiable AI system property that I described yesterday. The policy says the system will not give medical advice. The behavior shows whether it actually does or does not. The gap between those two is where the litigation risk lives.</p><p><strong>What this means for different types of AI deployments</strong></p><p>The bill applies regardless of how the AI system was built, but the risk profile varies significantly.</p><p>Organizations that build their own AI from the ground up have complete control over system prompts, guardrails, and response boundaries. They can engineer precise limits. But they also own 100% of the liability.</p><p>Organizations using low-code platforms like Copilot Studio or LangFlow face a shared responsibility problem. The platform provides underlying model behavior and some guardrails, but the builder configures the use case and the domain scope. When the system drifts into professional advice territory, who is liable? The platform or the builder?</p><p>And then there are the no-code deployments, the custom GPTs, the drag-and-drop chatbot builders. This is the highest risk category, and it is not close. The people building on these platforms are often the exact professionals the bill is trying to protect: small healthcare clinics, law offices, dental practices. They deploy an AI chatbot on their website, feed it their documents, and assume the platform handles compliance. It usually does not.</p><p>The gap between how easy it is to deploy AI and how hard it is to govern what it says is widest in the no-code tier. And that gap is exactly where this bill&#8217;s private right of action will land hardest.</p><p><strong>The deeper signal</strong></p><p>Step back from the specifics of this one bill and look at what it represents.</p><p>For decades, professional licensing has been a human-to-human regulatory framework. A doctor is licensed. A lawyer passes the bar. An engineer gets certified. The license attaches to the person, and the person is accountable for what they say.</p><p>AI breaks that model. The chatbot giving health guidance is not a licensed professional. It is not a person. It cannot be sued, sanctioned, or stripped of credentials. So the regulatory framework has to evolve. It has to attach accountability to the system&#8217;s behavior and to the entity that deployed it.</p><p>This bill is one of the first attempts to do that explicitly. It will not be the last. And every attempt will come back to the same core question: can you prove, with data, that your AI system behaves within its authorized boundaries?</p><p>That is not a policy question. That is a measurement question. And it demands the kind of quantitative, reproducible, behavior-based measurement that this newsletter exists to explore.</p><p>More next Tuesday.</p><div><hr></div><p><em>This is part of the &#8220;Before The Number&#8221; series at A.I.N.S.T.E.I.N., exploring what it takes to build quantitative AI governance measurement from first principles. If this resonated, share it with someone deploying AI in healthcare, legal, or any licensed profession.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[We Were Measuring the Wrong Thing ]]></title><description><![CDATA[Why AI Governance Has Been Scoring the Organization When It Should Be Scoring the Machine]]></description><link>https://ainstein.sanjeevaniai.com/p/we-were-measuring-the-wrong-thing</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/we-were-measuring-the-wrong-thing</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 23 Mar 2026 14:02:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!FKlb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FKlb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FKlb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png 424w, https://substackcdn.com/image/fetch/$s_!FKlb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png 848w, https://substackcdn.com/image/fetch/$s_!FKlb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png 1272w, https://substackcdn.com/image/fetch/$s_!FKlb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FKlb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png" width="1204" height="986" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:986,&quot;width&quot;:1204,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:106414,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/191816049?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FKlb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png 424w, https://substackcdn.com/image/fetch/$s_!FKlb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png 848w, https://substackcdn.com/image/fetch/$s_!FKlb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png 1272w, https://substackcdn.com/image/fetch/$s_!FKlb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6586702-647e-4147-bb9d-1f99b51607c7_1204x986.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>                                                 Image created by AI</strong></em></p><p></p><p>I owe you an explanation for the silence.</p><p>Two weeks ago, I published &#8220;How to Measure AI Governance&#8221; and laid out the five pillars, the metrics, the frameworks, the KPIs. I meant every word of it. And then I went quiet, because something broke in my own thinking that I could not write aro&#8230;</p>
      <p>
          <a href="https://ainstein.sanjeevaniai.com/p/we-were-measuring-the-wrong-thing">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[How to Measure AI Governance]]></title><description><![CDATA[The Five Pillars, the Metrics That Matter, and Why Checklists Are Not Enough]]></description><link>https://ainstein.sanjeevaniai.com/p/how-to-measure-ai-governance</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/how-to-measure-ai-governance</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 09 Mar 2026 14:02:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!R6zd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R6zd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R6zd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png 424w, https://substackcdn.com/image/fetch/$s_!R6zd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png 848w, https://substackcdn.com/image/fetch/$s_!R6zd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png 1272w, https://substackcdn.com/image/fetch/$s_!R6zd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R6zd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png" width="1456" height="889" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:889,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2070238,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/189697788?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!R6zd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png 424w, https://substackcdn.com/image/fetch/$s_!R6zd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png 848w, https://substackcdn.com/image/fetch/$s_!R6zd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png 1272w, https://substackcdn.com/image/fetch/$s_!R6zd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa04eff63-161f-4894-aaef-037d4b02a2e5_1802x1100.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h6><em>                                                                                                         Image by AI</em></h6><p></p><blockquote><p>If you cannot measure it, you cannot govern it. </p></blockquote><p>That principle holds true across every regulated industry, from finance to healthcare to cybersecurity, and it holds true for AI.</p><p>Yet when most organizations talk about AI governance today, they are talking about policies, principles, and frameworks. They are talking about what they believe, not what they can prove. And there is a meaningful difference between having an AI ethics policy and being able to demonstrate, with data, that your AI systems are actually governed.</p><p>This essay is a practical guide to bridging that gap. It walks through the core pillars of AI governance measurement, the specific metrics that matter, the frameworks available to structure the work, and the challenges that make this harder than it sounds. If you are a CISO, a Chief AI Officer, a compliance leader, or a founder building in this space, this is the foundation you need.</p><h2>Why Measurement Matters</h2><p>Without measurement, AI governance is a set of intentions. It lives in documents that get written once and reviewed quarterly at best. It gives leadership a sense of comfort without giving them a basis for action.</p><p>Measurement changes that in four concrete ways.</p><ol><li><p>It demonstrates due diligence. When regulators, boards, or the public ask how your AI is managed, measurement gives you evidence rather than assurances. </p></li><li><p>It allows you to identify and mitigate risks before they cause harm, because metrics like model drift detection time and fairness deviation surface problems that narrative assessments miss entirely. </p></li><li><p>It prepares you for the regulatory compliance landscape that is already here, with the EU AI Act requiring specific documentation and measurement for high-risk AI systems. </p></li><li><p>And it builds trust with every stakeholder who needs to know that your AI decisions are fair, transparent, and accountable.</p></li></ol><h2>The Five Pillars of AI Governance Measurement</h2><p>Effective AI governance measurement is not about tracking model accuracy or inference speed. Those are performance metrics. Governance measurement focuses on accountability, fairness, transparency, compliance, and safety. These are the five pillars, and each one requires its own set of metrics.</p><h3>Accountability and Ownership</h3><p>This pillar measures who is responsible for your AI systems and their outcomes. It sounds basic, but in my experience, a surprising number of organizations deploy AI systems where no single person owns the governance risk. The model was built by one team, deployed by another, and monitored by no one in particular.</p><p>The qualitative goal is straightforward: every high-risk AI system should have a named business owner who is accountable for its impact. The quantitative metric that tracks this is the percentage of deployed AI systems with a defined, documented business owner. If that number is below 100% for your high-risk systems, you have a governance gap that no policy document can close.</p><h3>Transparency and Explainability</h3><p>This pillar measures how well your AI system&#8217;s decisions can be understood by the humans affected by them. A lending model that denies an application needs to be able to explain why. A hiring algorithm that filters out candidates needs to produce a reason that a human can evaluate.</p><p>The quantitative metric here is the percentage of AI-driven decisions that include a human-interpretable explanation. In practice, this is one of the hardest metrics to improve because many complex models, particularly large language models, are inherently difficult to explain. But the measurement itself forces the conversation about where explainability gaps exist and how material those gaps are.</p><h3>Fairness and Bias Mitigation</h3><p>This pillar measures the extent to which your AI systems treat different demographic groups equitably. It is not enough to say &#8220;we care about fairness.&#8221; You need to measure the actual disparity in outcomes across protected groups and track that disparity over time.</p><p>The core metric is the measurable difference in approval rates, error rates, or outcomes between demographic groups. If your lending model approves 78% of applications from one group and 61% from another, that disparity is your fairness metric, and it needs to be monitored continuously, not just checked once before deployment.</p><h3>Risk and Compliance</h3><p>This pillar measures adherence to both internal policies and external regulations. With the EU AI Act, NIST AI RMF, and ISO 42001 all converging on requirements for risk classification and documentation, this pillar is becoming the most operationally urgent.</p><p>The key metrics include the percentage of high-risk AI systems that have completed an Algorithmic Impact Assessment, the percentage of inventoried systems that have undergone formalized risk classification, and the policy adherence rate across all AI projects. These numbers tell you whether your governance framework is actually being followed or whether it exists only on paper.</p><h3>Safety and Security</h3><p>This pillar measures your AI system&#8217;s resilience against attacks, errors, and unintended harm. It includes incident response readiness and the speed at which AI-specific failures are detected and resolved.</p><p>The metrics that matter here are the average time to detect and time to resolve AI-related incidents, including model drift, toxic output, adversarial attacks, and data pipeline failures. If your organization cannot tell you how long it takes to detect when a model has drifted from its intended behavior, your safety posture has a blind spot.</p><h2>Key Performance Indicators for AI Governance</h2><p>Beyond the five pillars, there are specific KPIs that give leadership a clear picture of governance health across the organization.</p><p>Program health metrics include AI inventory coverage (the percentage of all AI systems currently cataloged), risk classification completion (the percentage of inventoried systems that have been formally classified by risk level), and policy adherence rate (the percentage of AI projects fully compliant with established guidelines).</p><p>Decision and accountability metrics include decision latency for risk issues (how long it takes to make a material decision on an escalated AI risk), human override rate (how frequently automated decisions are reversed by human reviewers), and governance debt (the number of deferred governance controls that were postponed to speed up deployment).</p><p>Operational integrity metrics include model drift detection time, data lineage visibility (the percentage of models with full source-to-sink tracking), and audit readiness score (the percentage of models with current documentation and version control).</p><p>Ethical impact metrics include explanation coverage and fairness deviation, both of which I discussed in the pillars section above.</p><p>The important thing about these KPIs is that they are specific, measurable, and tied to real governance risk. They are not opinions. They are not traffic lights. They are numbers that a board can track quarter over quarter and that an auditor can verify independently.</p><h2>The Frameworks That Structure This Work</h2><p>Organizations do not need to build their measurement approach from scratch. Several established frameworks provide the structure.</p><p>The NIST AI Risk Management Framework provides guidelines for managing risks to improve the trustworthiness of AI systems. NIST has also recently released a preliminary draft Cyber AI Profile (NISTIR 8596) that maps AI considerations directly onto the Cybersecurity Framework 2.0, embedding AI governance into operational security infrastructure rather than treating it as a separate discipline.</p><p>ISO/IEC 42001 is an international standard specifying requirements for establishing, implementing, maintaining, and continually improving an AI management system. As an ISO 42001 Lead Auditor, I work with this framework regularly, and its strength is that it provides a certifiable standard that organizations can be audited against.</p><p>The EU AI Act is the most comprehensive regulatory framework currently in effect, requiring specific measurement and documentation for high-risk AI systems. It is not optional for organizations operating in or selling into the European market, and its requirements are driving measurement adoption globally.</p><p>These frameworks tell you what to measure and why. The challenge is translating their requirements into the specific quantitative metrics I described above, and doing so continuously rather than at a single point in time.</p><h2>The Challenges That Make This Hard</h2><p>If measuring AI governance were easy, every organization would already be doing it. Several factors make it genuinely difficult.</p><p>Concepts like fairness and transparency are contextually dependent. What counts as fair in a lending model may differ from what counts as fair in a hiring algorithm. There is no single universal formula, and measurement requires thoughtful interpretation alongside the numbers.</p><p>Many complex AI models, particularly large language models, are inherently difficult to explain. This makes transparency measurement challenging not because the metric is wrong but because the underlying system resists the measurement.</p><p>Standardization is still evolving. While frameworks exist, universally accepted methods for calculating specific metrics like bias are not yet settled. Different tools and approaches can produce different results for the same system.</p><p>Organizations have historically incentivized performance over responsibility. Accuracy and speed get rewarded. Governance measurement introduces a different set of priorities, and that cultural shift is often harder than the technical implementation.</p><p>And finally, data quality and lineage remain fundamental obstacles. You cannot measure governance properly if you do not understand the data your AI systems are trained on, and many organizations have complex or undocumented data flows that make this difficult.</p><h2>Where This Is Heading</h2><p>Every one of these challenges is real, and none of them are reasons to avoid measurement. They are reasons to invest in building the measurement infrastructure now, before regulators require it and before the gap between what your organization claims about its AI governance and what it can actually prove becomes a liability.</p><p>The organizations that solve the measurement problem first will not just be compliant. They will set the standard that others measure against. They will have the data to report to boards, the benchmarks to negotiate with partners, and the scores to prove what checklists never could.</p><p>AI governance measurement is not a nice-to-have. It is the infrastructure that makes governance real.</p>]]></content:encoded></item><item><title><![CDATA[Understanding AI Governance Measurement]]></title><description><![CDATA[Why Quantitative Measurement Is No Longer Optional!]]></description><link>https://ainstein.sanjeevaniai.com/p/understanding-ai-governance-measurement</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/understanding-ai-governance-measurement</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 02 Mar 2026 20:11:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Vu6o!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Vu6o!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Vu6o!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png 424w, https://substackcdn.com/image/fetch/$s_!Vu6o!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png 848w, https://substackcdn.com/image/fetch/$s_!Vu6o!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png 1272w, https://substackcdn.com/image/fetch/$s_!Vu6o!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Vu6o!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png" width="1456" height="867" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:867,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2960668,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/189685261?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Vu6o!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png 424w, https://substackcdn.com/image/fetch/$s_!Vu6o!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png 848w, https://substackcdn.com/image/fetch/$s_!Vu6o!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png 1272w, https://substackcdn.com/image/fetch/$s_!Vu6o!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce04ce4f-853b-45e7-ba35-ccf48d2687e2_1840x1096.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h6><em>                                                                                                  Image created by AI</em></h6><p></p><p>Every critical system in human history eventually got measured. Before FICO, loan officers decided your creditworthiness with a handshake and a gut feeling. Same income, same history, approved at one branch and denied at another. Before s&#8230;</p>
      <p>
          <a href="https://ainstein.sanjeevaniai.com/p/understanding-ai-governance-measurement">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Everyone’s Writing the Rulebook. Nobody’s Building the Instrument.]]></title><description><![CDATA[Why the global rush to govern AI keeps producing frameworks instead of measurement]]></description><link>https://ainstein.sanjeevaniai.com/p/everyones-writing-the-rulebook-nobodys</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/everyones-writing-the-rulebook-nobodys</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Tue, 24 Feb 2026 20:05:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-NCj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-NCj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-NCj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png 424w, https://substackcdn.com/image/fetch/$s_!-NCj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png 848w, https://substackcdn.com/image/fetch/$s_!-NCj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png 1272w, https://substackcdn.com/image/fetch/$s_!-NCj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-NCj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png" width="876" height="866" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/153179cf-3c12-4f59-bd39-c8f569970920_876x866.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:866,&quot;width&quot;:876,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1148782,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/189055181?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-NCj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png 424w, https://substackcdn.com/image/fetch/$s_!-NCj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png 848w, https://substackcdn.com/image/fetch/$s_!-NCj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png 1272w, https://substackcdn.com/image/fetch/$s_!-NCj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F153179cf-3c12-4f59-bd39-c8f569970920_876x866.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h6><em>                                                                                                  Image Created by AI</em></h6><p></p><p>If you are in the governance space, the first three weeks of February 2026 arrived as a cascade, one event after another, each reinforcing the same message: the world is taking AI governance seriously. UC Berkeley&#8217;s Center for Long-Term C&#8230;</p>
      <p>
          <a href="https://ainstein.sanjeevaniai.com/p/everyones-writing-the-rulebook-nobodys">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Your AI Agent Has No ID ]]></title><description><![CDATA[Why Agentic AI Needs a Trust Score, Not a Checklist]]></description><link>https://ainstein.sanjeevaniai.com/p/your-ai-agent-has-no-id</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/your-ai-agent-has-no-id</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 16 Feb 2026 23:33:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!SC5S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SC5S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SC5S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png 424w, https://substackcdn.com/image/fetch/$s_!SC5S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png 848w, https://substackcdn.com/image/fetch/$s_!SC5S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png 1272w, https://substackcdn.com/image/fetch/$s_!SC5S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SC5S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png" width="1378" height="1070" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1070,&quot;width&quot;:1378,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:895109,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/187929491?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SC5S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png 424w, https://substackcdn.com/image/fetch/$s_!SC5S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png 848w, https://substackcdn.com/image/fetch/$s_!SC5S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png 1272w, https://substackcdn.com/image/fetch/$s_!SC5S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d35d615-5df3-4701-9024-ddd77c5acfe3_1378x1070.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h6><em>                                                                                  Image created using AI</em></h6><p></p><p>This week, I was invited through one of the world&#8217;s largest expert networks to consult on a topic that stopped me in my tracks: the challenges and solutions for securely deploying autonomous AI agents in business environments, with a particular focus on something called &#8220;verifiable credentials&#8221; for AI agents and &#8220;trusted AI through intent binding.&#8221;</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>I have been building AI governance infrastructure for over two years now, and I have spent 25 years before that deploying AI and data systems across healthcare, fintech, insurance, legal technology, and education. But this invitation was different. It was not about compliance checklists or policy frameworks. It was about a question that the enterprise world is only now beginning to ask out loud: How do you prove that an AI agent acting on your behalf is actually authorized to do what it is doing?</p><p><em>Let that sink in for a moment.</em></p><p><strong>The Problem Nobody Talks About</strong></p><p>Right now, across every industry you can name, companies are racing to deploy AI agents that can act autonomously. These agents book meetings, process claims, triage legal inquiries, approve transactions, generate reports, and make decisions that used to require a human being in the loop. The promise is enormous: speed, scale, consistency, cost reduction. And the technology has reached a point where these agents can genuinely perform.</p><p>But here is what almost nobody is asking: <em>What credentials does that AI agent carry?</em></p><p>When a human employee joins an organization, they go through background checks, they receive role-based access, they sign agreements about what they can and cannot do, and there is a paper trail connecting their identity to their authority. If that employee oversteps their boundaries, there is an audit trail. There is accountability. There is a verifiable chain that connects what they did to what they were authorized to do.</p><p>Now think about an AI agent. It gets deployed with an API key, maybe some prompt instructions, maybe a set of tool permissions configured by a developer who was moving fast to hit a sprint deadline. Where is the verifiable proof of what that agent is authorized to do? Where is the audit trail that connects its actions to a specific human decision about its scope? Where is the credential that another system, a partner, a regulator, a customer, can independently verify without simply trusting the agent&#8217;s own assertions?</p><p>It does not exist. Not in any standardized, quantitative, independently verifiable way.</p><p><strong>Your AI agent has no ID.</strong></p><p><strong>Why Checklists Fail for Autonomous Systems</strong></p><p>The traditional approach to AI governance has been borrowed from software compliance: create a checklist, assess against it annually or quarterly, produce a report, file it away. This works reasonably well for static systems where humans make the final decisions and the AI is just providing recommendations. You can audit the model, check the training data, review the outputs, and sign off.</p><p>But autonomous AI agents break this model completely.</p><p>An autonomous agent does not wait for a human to review its output before acting. It chains decisions together. It interprets ambiguous inputs in real time. It interacts with other systems, sometimes other agents, and the scope of its actions can shift based on context in ways that no static checklist can anticipate. A checklist that was accurate on Monday might be meaningless by Wednesday because the agent encountered a scenario that nobody tested for, and it made a judgment call.</p><p>This is not a theoretical concern. I have seen it firsthand. When I deployed an autonomous AI voice agent for a law firm handling workers&#8217; compensation cases, the most dangerous moments were not when the agent got something wrong in a predictable way. They were when callers deviated from expected conversational paths and the agent had to decide, in real time, whether it was authorized to handle the new direction or whether it needed to escalate. A checklist cannot govern that decision. A quantitative, continuously updated trust boundary can.</p><p>The difference matters. A checklist says &#8220;this system was compliant when we last checked.&#8221; A trust score says &#8220;this system is operating within its verified boundaries right now, and here is the quantitative evidence.&#8221;</p><p><strong>What Verifiable Credentials for AI Actually Means</strong></p><p>When the consultation framed the topic around &#8220;verifiable credentials for AI agent deployment,&#8221; it pointed at something that I believe will become one of the defining infrastructure layers of the next decade.</p><p>Verifiable credentials for AI agents means that an agent carries cryptographically provable attestations about what it is authorized to do, who authorized it, what compliance standards it has been assessed against, and what boundaries it operates within. Any party interacting with that agent, whether it is another system, a business partner, a regulator, or a customer, can independently verify those claims without having to trust the agent itself.</p><p>Think of it like a digital license. Not a static certificate that was issued once and sits in a drawer, but a living, scored, continuously updated credential that reflects the agent&#8217;s current risk posture and authorization scope. When a partner organization&#8217;s system interacts with your AI agent, it can check that credential and confirm: yes, this agent has been assessed at a risk score of 247 out of 1000, it is authorized for these specific actions, it has been evaluated against ISO 42001 and the EU AI Act, and its last assessment was 14 minutes ago.</p><p><em>That is fundamentally different from saying &#8220;we passed an audit last quarter.&#8221;</em></p><p><strong>The Market Is Moving Faster Than You Think</strong></p><p>Here is what struck me about this invitation, and about the broader pattern I am seeing across the industry right now. This is not a topic that only governance nerds and compliance officers are thinking about. Investment firms are actively conducting diligence on companies in the AI security and governance tooling space. Major corporations are paying expert network rates to understand how verifiable trust for AI agents works. Professional services firms are benchmarking API governance frameworks in banking. The money is following the question.</p><p>And the question is the same everywhere: <em>How do we trust AI that acts on its own?</em></p><p>For those of us who have been working in AI governance, this is the moment where the market catches up to the problem. For the past two years, I have heard enterprise leaders say they are &#8220;exploring&#8221; AI governance, that they know it matters but they are not ready to invest. That language is shifting. When investment firms start researching the competitive landscape for AI trust infrastructure, it means capital allocation decisions are being made. When enterprise clients specify &#8220;verifiable credentials&#8221; and &#8220;intent binding&#8221; as the topics they want to discuss, it means they have moved past awareness and into solution design.</p><p><em>The window between &#8220;people are asking the question&#8221; and &#8220;someone owns the answer&#8221; is open right now.</em></p><p><strong>Why Measurement Beats Compliance</strong></p><p>This brings me to the core thesis of everything I write about in this newsletter, and everything I am building.</p><p>The reason checklists and traditional compliance frameworks fail for autonomous AI is not that they are poorly designed. It is that they are qualitative instruments being applied to a quantitative problem. Asking whether an AI agent &#8220;meets&#8221; a compliance standard is like asking whether a bridge &#8220;meets&#8221; safety requirements without measuring the load it can bear. The answer is meaningless without a number.</p><p>What the market needs, and what it is beginning to demand, is quantitative measurement infrastructure for AI trust. A scoring system that can express, in a single interpretable number, how much risk an AI agent carries across multiple regulatory frameworks simultaneously. Not a binary pass/fail. Not a subjective assessment. A reproducible, auditable, continuously updated measurement that engineering teams can act on and compliance teams can report on and regulators can verify and business partners can trust.</p><p>This is the science of measurement applied to critical AI systems. It is what I call the work that happens &#8220;before the number,&#8221; the careful, rigorous thinking about what to measure, how to measure it, and why the measurement methodology matters as much as the result.</p><p><strong>What Comes Next</strong></p><p>If you are deploying AI agents in your organization today, or planning to, here is what I would encourage you to think about.</p><p>First, ask yourself whether you can prove, right now, what your AI agent is authorized to do. Not what it was designed to do. Not what the prompt says it should do. Can you prove it, verifiably and quantitatively, to a third party who has no reason to trust your assertions?</p><p>Second, ask yourself how you would know if your AI agent exceeded its authorized scope. Not after the fact, when a customer complains or a regulator asks. Right now. In real time. Do you have a continuous measurement of whether the agent is operating within its trust boundaries?</p><p>Third, ask yourself whether your current governance approach would survive the question: &#8220;Show me the score.&#8221; Not the checklist. Not the policy document. The score. The number that tells me, quantitatively, where this agent falls on the risk spectrum and what that number is based on.</p><p>If you cannot answer those questions today, you are not alone. Almost nobody can. But the market is telling us, loudly and with real dollars behind it, that the window to build this capability is right now.</p><p>That is what I am working on. That is what &#8220;Before the Number&#8221; is about. And I will have a lot more to say about it in the weeks ahead.</p><div><hr></div><p><em>Suneeta Modekurty is the Founder and Chief AI Architect of SANJEEVANI AI, where she builds quantitative AI governance infrastructure. She is an ISO 42001 Lead Auditor and holds an O-1A visa for extraordinary ability in AI, bioinformatics, and data science. She publishes &#8220;Before the Number&#8221; on Substack, exploring the science of measurement in critical AI systems.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[What Happens When There's No Number]]></title><description><![CDATA[The invisible cost of governing without measurement]]></description><link>https://ainstein.sanjeevaniai.com/p/what-happens-when-theres-no-number</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/what-happens-when-theres-no-number</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Tue, 10 Feb 2026 15:03:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!w7M4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>BEFORE THE NUMBER</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!w7M4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!w7M4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png 424w, https://substackcdn.com/image/fetch/$s_!w7M4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png 848w, https://substackcdn.com/image/fetch/$s_!w7M4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!w7M4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!w7M4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png" width="1386" height="1120" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1120,&quot;width&quot;:1386,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:483264,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/187477850?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!w7M4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png 424w, https://substackcdn.com/image/fetch/$s_!w7M4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png 848w, https://substackcdn.com/image/fetch/$s_!w7M4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!w7M4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4cfbf39a-15d7-4981-8fd5-7279b250f7e4_1386x1120.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h6> <em>                                           Divergence of Qualitative and Quantitative Measurements Over Time</em></h6><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>In 1846, a doctor named Ignaz Semmelweis noticed something troubling. In one ward of the Vienna General Hospital, mothers were dying at five times the rate of the ward next door. It was the same hospital, in the same city, during the same year. The only difference was that one ward was staffed by doctors who came directly from performing autopsies, while the other was staffed by midwives who did not.</p><p>Semmelweis did not have germ theory. He did not have a microscope powerful enough to see bacteria. But he had something just as important: he had counted. He had a number. And the number made the invisible visible. When he introduced handwashing with chlorinated lime, mortality dropped from 18% to under 2%. The number did not just describe the problem. It made the solution undeniable.</p><p>But here is the part of the story people forget. The medical establishment rejected his findings for twenty years. They did not reject the findings because the data was wrong. They rejected the findings because the profession had no culture of measurement. Doctors operated on reputation, seniority, and judgment. Introducing a number threatened the entire social order of medicine because it implied that a senior physician&#8217;s intuition could be proven wrong, and worse, that a junior doctor armed with better data could be proven right. The number was not just a scientific tool. It was a direct challenge to the authority structure that governed the profession. And the authority structure pushed back.</p><p>Semmelweis died in an asylum in 1865. Germ theory was not widely accepted until the 1880s. For two decades, the refusal to accept a quantitative finding cost thousands of lives. Not because the answer was unknown, but because the system was not structured to receive a quantitative answer.</p><p>This essay is about what happens in that gap. It is not about the moment a number arrives, but about the years before it does, when a system that matters is governed by opinion and the cost of that governance failure remains invisible to the very people responsible for it.</p><p>Semmelweis&#8217;s story is dramatic, but it is not unique. When you study the history of measurement across critical systems in lending, aviation, food safety, environmental regulation, and cybersecurity, the same pattern emerges so consistently that it starts to look less like coincidence and more like a law of institutional physics. Every time a critical system operates without quantitative measurement, four specific costs appear. They appear in the same order, with the same dynamics, regardless of the industry, the era, or the technology involved.</p><p><strong>The first cost is structural inconsistency: </strong>Before credit scoring existed, two loan officers at the same bank could evaluate the same applicant and reach opposite conclusions. This was not a flaw in the system. This <em>was</em> the system. Without a shared quantitative reference point, every decision was a fresh act of human judgment, shaped by experience, bias, workload, mood, and variables that nobody tracked because nobody could.</p><p>Fair Isaac Corporation studied this phenomenon in the 1950s and found staggering variance in loan decisions across branches of the same institution. The variance was not small. Outcomes that should have been statistically identical were diverging by double digits. The same applicant, with the same income and the same repayment history, could be approved at one branch in the morning and denied at another branch in the afternoon. This was not corruption. It was the natural and predictable result of a system that relied entirely on individual judgment with no standardized measurement to anchor it.</p><p>The same dynamic shows up everywhere measurement is absent. In food safety before HACCP scoring, the same restaurant could pass a health inspection in one county and fail in the county next door, because inspectors had no shared quantitative standard for what constituted a violation. In education before standardized assessments, the same student could be classified as &#8220;gifted&#8221; in one school district and &#8220;average&#8221; in another, because teachers were evaluating against their own internal benchmarks rather than a common metric. In pain management before the numeric pain scale, the same tibial fracture could receive acetaminophen from one nurse and morphine from another, because pain was described in adjectives rather than measured in numbers.</p><p>The deeper problem is not that people made different decisions. It is that nobody could demonstrate the inconsistency was even happening, because there was no consistent metric to compare against. When there is no number, inconsistency becomes invisible. It hides inside the phrase &#8220;professional judgment,&#8221; and it persists precisely because no one can see it.</p><p><strong>The second cost follows directly from the first: accountability disappears. </strong>You cannot hold someone accountable for violating a standard that does not exist in quantifiable terms. You can write a policy that says &#8220;ensure patient safety.&#8221; You can create a governance framework that says &#8220;implement responsible AI practices.&#8221; But until those principles are attached to measurable thresholds, enforcement becomes a matter of interpretation, and interpretation is the enemy of accountability.</p><p>The aviation industry learned this through tragedy. Before the adoption of measurable safety metrics such as hours between incidents, defect rates per flight cycle, and standardized checklists with quantified completion rates, airlines assessed their own safety through self-reporting. The standard was &#8220;we follow best practices.&#8221; After a crash, the investigation would inevitably reveal that &#8220;best practices&#8221; meant different things to different maintenance crews, different inspectors, different shifts, and different airports. No one had been lying. There was simply nothing specific enough to be accountable to. The standard existed in prose rather than in numbers, and prose can be interpreted generously by anyone who needs it to be.</p><p>The same dynamic plagued corporate environmental responsibility for decades. When the standard was qualitative and every company could simply declare that it was &#8220;committed to sustainability,&#8221; every company in the world was effectively compliant, because commitment is not measurable. It was only when emissions reporting introduced actual numbers in the form of tons of CO&#8322;, parts per million, and year-over-year change that accountability became possible. This shift did not happen because regulators suddenly became tougher. It happened because there was finally something concrete to hold companies accountable against. You can argue indefinitely about whether a company is &#8220;committed to sustainability.&#8221; You cannot argue with 47,000 metric tons of carbon.</p><p>Measurement does not create accountability on its own. But without measurement, accountability is theater. It carries all the language of oversight, including policies, frameworks, committees, and review boards, but none of the teeth. Because teeth require thresholds, and thresholds require numbers.</p><p><strong>The third cost is the most economically significant, and it is also the hardest to see, because it manifests as things that never happen. </strong>When there is no number, markets do not crash. They simply never form in the first place.</p><p>Before credit scores existed, the secondary mortgage market barely existed either. A bank in Ohio could not sell a bundle of loans to an investor in New York, because there was no standardized way to assess the risk of those loans at scale. Every loan had been originated by a local officer, evaluated using local criteria, and documented in local formats. An investor three thousand miles away would have needed to re-underwrite every individual loan in order to assess the bundle, and the cost of that diligence was prohibitive. So the transaction simply did not happen.</p><p>The secondary mortgage market did not emerge because someone invented a clever financial instrument. It emerged because someone gave every borrower a number that an investor three thousand miles away could evaluate in seconds. Credit scoring did not just measure risk. It made risk portable. It created a common language that allowed parties who had never met each other to transact with confidence. A market worth trillions of dollars had been locked inside the absence of a three-digit number.</p><p>The same pattern explains why cyber insurance took decades to mature. Until organizations had quantifiable security postures in the form of scores, ratings, measurable controls, and auditable configurations, underwriters could not price policies with any actuarial confidence. You cannot build an insurance market around &#8220;we think we&#8217;re secure.&#8221; Insurance requires a number that actuaries can model, that underwriters can compare across applicants, and that reinsurers can aggregate into portfolios. The number does not just describe the risk. It enables the market infrastructure that makes risk transferable.</p><p>When there is no number, entire markets remain latent, not because demand is absent, but because there is nothing to transact around. Buyers cannot evaluate, sellers cannot differentiate, insurers cannot price, investors cannot compare, and regulators cannot benchmark. The market sits frozen, waiting for a unit of measurement that all participants agree to trust. And the longer that wait continues, the more value remains locked inside the gap.</p><p><strong>The fourth cost is perhaps the most insidious: improvement becomes impossible to prove. </strong>Imagine walking into a hospital board meeting and reporting that patient safety improved this quarter. The first question will be: compared to what? By how much? Measured how? Without a quantitative baseline, improvement is a feeling rather than a fact. You can spend millions on better processes, better training, and better technology and still have absolutely no way to demonstrate that any of it worked.</p><p>This is not hypothetical. It is the precise reason the quality movement in manufacturing stalled for years until statistical process control gave factories a way to measure variation and prove that their interventions were actually reducing defects. W. Edwards Deming did not just advocate for quality. He advocated for measurement, because he understood from decades of experience that without it, quality was a slogan rather than a discipline. His famous observation that you cannot improve what you cannot measure was not a platitude. It was an empirical conclusion about what happens when organizations try to get better without quantitative feedback loops.</p><p>The consequence of unprovable improvement is organizational paralysis. When the finance team asks whether the governance investment is working and the honest answer is &#8220;we believe so but cannot demonstrate it,&#8221; budgets get questioned. When leadership asks whether the new training program reduced risk and the answer is anecdotal rather than measured, confidence erodes. And gradually, organizations stop investing in getting better, not because they do not want to improve, but because they have learned that improvement without measurement is indistinguishable from stagnation. The return on investment becomes invisible, and so the investment stops.</p><p>What is striking about these four costs is not that they exist, but that the people inside the system rarely see them clearly. Inconsistency feels like professional judgment. Missing accountability feels like flexibility. Frozen markets feel like the market simply is not ready yet. Unprovable improvement feels like doing the best you can under difficult circumstances. The absence of a number is comfortable. It protects incumbents. It allows vagueness to masquerade as strategy. It lets everyone believe they are above average, because there is no average to measure against.</p><p>This is why measurement is always resisted before it is adopted. Semmelweis was ridiculed. Early credit scoring was called dehumanizing, with critics arguing that a human relationship between banker and borrower could not and should not be reduced to a number. Standardized testing was called reductive. Emissions reporting was called burdensome. Every number that eventually became infrastructure started its life as an inconvenient truth that the existing establishment preferred to ignore.</p><p>And yet, in every single case, once the number arrived and proved its value, the world did not go back. Nobody has argued for returning to gut-feel lending decisions after FICO. Nobody has advocated for removing thermometers from hospitals. Nobody has suggested that airlines stop tracking maintenance defect rates. Nobody has proposed that companies stop reporting emissions in measurable units. The resistance dissolves once the number demonstrates what it can do, because the number does not just measure the system. It <em>reorganizes</em> the system. It changes who has authority, what constitutes evidence, how decisions get made, and what accountability looks like. The number becomes the infrastructure that everything else is built upon.</p><p>Which brings us to the present. Today, AI systems are approving loans, diagnosing cancers, screening job applicants, scoring insurance claims, generating legal documents, triaging emergency calls, and making consequential decisions that affect the lives of millions of people across every regulated industry on the planet.</p><p>Ask how well these systems are governed, and the answer in most organizations is qualitative. It takes the form of a maturity model with descriptive levels, a readiness checklist with binary checkboxes, a consultant&#8217;s assessment delivered as a narrative report, or a regulatory framework mapped onto a spreadsheet that was current when it was created and outdated by the time it was presented. The output is language rather than measurement. &#8220;We&#8217;re at level 3.&#8221; &#8220;We&#8217;ve addressed most of the NIST categories.&#8221; &#8220;We&#8217;re working on it.&#8221;</p><p>These are not numbers. They are opinions formatted to look like numbers. And the four costs are already visible for anyone willing to look.</p><p>The inconsistency is already here. Two auditors assessing the same AI system against the same governance framework reach materially different conclusions about its compliance posture. This is not a failure of the auditors. It is a failure of the measurement approach, or more precisely, the absence of one. When the standard is a checklist of qualitative criteria, every assessment is an interpretation, and interpretations diverge.</p><p>The accountability gap is already here. The EU AI Act is in force. ISO 42001 certifications are underway. NIST AI RMF adoption is accelerating. Regulatory enforcement is not theoretical but operational. Yet enforcement against what baseline? When a regulator asks an organization to demonstrate its AI governance posture over time, the evidence that exists consists of a policy document from last year, a completed questionnaire, and a consultant&#8217;s letter. None of these constitute the kind of quantitative, auditable, time-series evidence that regulators in every other domain have learned to require.</p><p>The frozen markets are already here. AI insurance is nascent. AI procurement due diligence is a custom exercise every time, with no standardized scoring to streamline vendor evaluation. AI risk assessment in mergers and acquisitions relies on qualitative representations that are difficult to verify and impossible to compare across targets. These markets are not early because demand is low. They are early because there is nothing standardized to transact around. The infrastructure of transaction, including pricing, benchmarking, comparison, and aggregation, requires a number that does not yet exist.</p><p>And the inability to prove improvement is already here. Organizations are spending real money on AI governance by hiring responsible AI teams, building review processes, investing in training, and purchasing tools. But when the board asks whether the organization is better governed than it was last quarter, the honest answer is that nobody knows, because nobody can measure it. There is a belief that things have improved. There are actions that should have made things better. But there is no number that moved.</p><p>Every critical system in human history has eventually been measured. Not because someone decided measurement was philosophically appealing, but because the cost of not measuring became intolerable. The lending industry crossed that threshold in the 1950s. Healthcare crossed it in the 19th century with the advent of lab science and vital signs. Aviation crossed it after enough planes fell from the sky and enough investigations revealed that &#8220;best practices&#8221; had been an empty phrase. Cybersecurity crossed it when boards stopped accepting narrative descriptions of risk exposure and started demanding numbers.</p><p>AI governance has not crossed that threshold yet. But every signal suggests it is approaching. The regulatory pressure is building across multiple jurisdictions simultaneously. The liability exposure is growing as AI systems move deeper into consequential decision-making. The number of AI systems in production is compounding faster than governance practices can keep pace. And the gap between what organizations claim about their governance posture and what they can actually demonstrate with evidence is widening every quarter.</p><p>The question is not whether AI governance will get a number. History is unambiguous on this point, because every critical system eventually does. The question is what happens between now and then, how long the gap persists, how much damage accumulates inside it, and who bears the cost of an industry that governed itself by opinion when it could have governed itself by measurement.</p><p>History suggests that cost will be larger than anyone currently inside the gap realizes. It always is.</p><p></p><p></p><p><em>Until next time,</em></p><p><strong>Suneeta Modekurty</strong></p><p>Founder &amp; Chief Architect, METRIS&#8482;</p><p></p><p><em>Before the Number is a publication about the science of measurement in critical systems.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Before the Number]]></title><description><![CDATA[Why Every Critical System Eventually Gets Measured and Why AI Is Next]]></description><link>https://ainstein.sanjeevaniai.com/p/before-the-number</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/before-the-number</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Tue, 03 Feb 2026 16:57:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!R-Hq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R-Hq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R-Hq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png 424w, https://substackcdn.com/image/fetch/$s_!R-Hq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png 848w, https://substackcdn.com/image/fetch/$s_!R-Hq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png 1272w, https://substackcdn.com/image/fetch/$s_!R-Hq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R-Hq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png" width="1412" height="1278" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1278,&quot;width&quot;:1412,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2414848,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/186755932?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!R-Hq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png 424w, https://substackcdn.com/image/fetch/$s_!R-Hq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png 848w, https://substackcdn.com/image/fetch/$s_!R-Hq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png 1272w, https://substackcdn.com/image/fetch/$s_!R-Hq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162e8e82-fca9-4932-9f5a-4f5197c16323_1412x1278.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Two weeks ago, on January 22, I announced that AI governance now has a score.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Singapore released the world&#8217;s first agentic AI governance framework. South Korea&#8217;s AI Basic Act became enforceable. And on that same day, we shipped METRIS - a quantitative governance score for AI systems.</p><p>One of the responses that stayed with me came from a reader who wrote: <em>&#8220;The distinction between frameworks and real, measurable governance is so crucial. Finally, someone speaks about dynamic AI risk.&#8221;</em></p><p>An AI/ML engineer called METRIS&#8217;s FICO-style approach <em>&#8220;a total game changer for clarity&#8221;</em> and then asked the right question: how does METRIS handle the shifting weight of different regulations as they evolve in real time?</p><p>A data engineer was more direct: <em>&#8220;METRIS needs to be deployed everywhere, all at once. Think of the corporate goofs that could have been avoided.&#8221;</em></p><p>These reactions tell me the market knows it needs a number. But before I explain what METRIS does or how it works, I want to go deeper. I want to share <em><strong>why</strong></em> it needs to exist.</p><p>Because this isn&#8217;t a technology story. It&#8217;s a measurement story. And it starts with a question I&#8217;ve been asking my entire career.</p><h2>The Question</h2><p>I have fought ambiguity my entire professional life.</p><p>Twenty-five years across healthcare, pharma, finance, and technology. In every role, the hardest problems were never technical. They were the ones where two equally qualified people looked at the same situation and reached opposite conclusions. Where the answer depended on who you asked, what day it was, or which office you walked into.</p><p>That kind of ambiguity doesn&#8217;t just slow things down. It erodes trust. It makes fair outcomes impossible. And it creates a world where accountability is always someone else&#8217;s problem.</p><p>So I started asking a question that wouldn&#8217;t leave me alone: <em><strong>How did humanity solve ambiguity before?</strong></em></p><p>The answer, every single time, was the same. We invented a number.</p><h2>Banking Before FICO</h2><p>Before 1989, getting a loan was a conversation. A loan officer looked at you across a desk. They reviewed your application. They considered your employment, your address, your references. And then they made a judgment call.</p><p>The same person, with the same income and the same history, could be approved at one branch and denied at another. The process was subjective, inconsistent, and (let&#8217;s be honest), was discriminatory. Billions of dollars were allocated based on gut feeling, personal bias, and the quality of a handshake.</p><p>Then Fair, Isaac and Company introduced a three-digit number.</p><p>FICO didn&#8217;t add technology to lending. It removed ambiguity from lending. One number. Same calculation everywhere. A 720 in New York meant the same thing as a 720 in Nebraska. Suddenly, lending decisions were comparable, auditable, and fairer than the alternative.</p><p>Today, nothing moves without a credit score. Mortgages, car loans, insurance premiums, apartment rentals. The number became infrastructure.</p><h2>Medicine Before Lab Tests</h2><p>A doctor feels your forehead. &#8220;You seem warm.&#8221; Another doctor examines you an hour later. &#8220;You seem fine.&#8221; Same patient. Two opinions. Who&#8217;s right?</p><p>Before quantitative diagnostics, medical decisions were based on observation, intuition, and experience. All valuable, but subjective. Two physicians could examine the same patient and reach opposite conclusions. Treatment varied not by disease, but by doctor.</p><p>Then someone measured body temperature. Then blood count. Then glucose levels. Then cholesterol, then hemoglobin A1C, then troponin levels, then genetic markers.</p><p>Numbers didn&#8217;t replace doctors. Numbers gave doctors a shared reality to work from. A blood glucose of 280 means the same thing in Asia and in America or Europe. The measurement created a common language, and with it, accountability.</p><h2>Education Before Standardized Assessment</h2><p>A teacher reads your essay and says &#8220;good.&#8221; Another teacher reads the same essay and says &#8220;mediocre.&#8221; A third says &#8220;brilliant.&#8221; The essay hasn&#8217;t changed. The judges have.</p><p>Before rubrics, scores, and standardized marking systems, educational assessment was entirely subjective. You can argue about whether standardized testing is perfect, it isn&#8217;t. But it made assessment comparable. A 92 means the same thing regardless of who graded it. Progress could be tracked. Gaps could be identified. Accountability became possible.</p><h2>The Deeper Pattern</h2><p>If you go far enough back, you find the origin of measurement itself: the moment societies decided that subjective trust wasn&#8217;t enough.</p><p>Signatures. Seals. Notarization. Weights and measures. Currency denominations. Accounting standards. Credit ratings. Safety certifications. Every one of these innovations was born from the same realization: when the stakes are high enough, &#8220;trust me&#8221; is not a system. Measurement is.</p><p>The pattern is always the same. First, a critical human activity operates on judgment alone. Then the stakes get high enough that inconsistency becomes intolerable. Then someone invents a way to measure. And the measurement becomes the new infrastructure-so foundational that within a generation, no one can imagine doing without it.</p><h2>Now Look at AI</h2><blockquote><p><em>AI systems are making lending decisions. Diagnosing patients. Evaluating students. Screening job applicants. Flagging criminal suspects. Approving insurance claims. Moderating speech. Predicting recidivism.</em></p></blockquote><p>These are the exact same domains that humanity spent centuries learning to measure. The lending system has a FICO score. The lab has quantitative diagnostics. The exam has a rubric. The contract has a signature.</p><p>But the AI that is replacing these systems? The AI that is now making these consequential decisions on our behalf?</p><p>Ask &#8220;How governed is this AI system?&#8221; and the answer you get is: a checklist. A policy document. A consultant&#8217;s opinion. A yes-or-no audit. Or worse? a shrug.</p><p><em><strong>There is no number.</strong></em></p><h2>The Problem with Binary</h2><p>The AI governance conversation today is stuck in binary. Compliant or not. Pass or fail. And that framing is the source of the paralysis.</p><p>Consider two companies. <br><strong>Company A</strong> has done nothing:no documentation, no fairness testing, no monitoring, no risk assessment. <br><strong>Company B</strong> has documented all its models, implemented bias testing, established human oversight protocols, but hasn&#8217;t yet completed adversarial robustness testing.</p><p>In a binary system, both fail. Same result. Same bucket. Binary made them identical. But they are not identical. One is at 50. The other is at 680. One needs a transformation. The other needs a nudge.</p><p>Binary created a market that is frozen. Companies that haven&#8217;t started say &#8220;we&#8217;re exploring.&#8221; Companies that have invested say &#8220;what&#8217;s the point?&#8221; Companies that passed say &#8220;we&#8217;re done&#8221; and stop paying attention - until the next incident.</p><h2>A Score Changes Everything</h2><p>A score does what binary cannot. It makes progress visible. It enables comparison. It creates continuous accountability. It creates a market that moves.</p><p>At 50, you know you&#8217;re early but you&#8217;ve started. At 400, you can see progress. At 680, you know exactly which gaps separate you from 800. At 900, you can prove your posture to your board, your regulator, your customers. And tomorrow, if your score drops because a new regulation kicked in or a model drifted, you see it immediately.</p><p><strong>Governance isn&#8217;t pass/fail. It&#8217;s a score.</strong></p><h2>What&#8217;s Next</h2><p>In my last newsletter, I announced METRIS </p><p>Today, I wanted to tell you <em><strong>why</strong></em> it exists. Because the founding insight behind METRIS isn&#8217;t technical. It&#8217;s historical. Every critical system eventually gets a number. AI&#8217;s turn is now.</p><p>In the coming weeks, I&#8217;ll share: how the scoring engine actually works, what the first assessments are revealing about the state of AI governance in the wild, and how organizations are using their score to move from &#8220;we&#8217;re exploring&#8221; to &#8220;here&#8217;s where we stand.&#8221;</p><p>If you&#8217;re building AI and wrestling with governance, reply to this newsletter. Tell me what&#8217;s broken. What&#8217;s working. What you wish existed. I read everything.</p><p>Every critical system in human history eventually got a number.</p><p>Lending got a credit score. Health got diagnostics. Education got assessments. Security got ratings. Financial health got audited statements.</p><p>AI is now the most consequential system in modern life. And it has no number.</p><p></p><p><strong>Not yet.</strong></p><p></p><p><em><strong>Suneeta Modekurty</strong></em></p><p><em>Founder &amp; Chief Architect, METRIS&#8482;</em></p><p><em>ISO 42001 Lead Auditor | Sanjeevani AI LLC</em></p><p><em>A.I.N.S.T.E.I.N. is a reader-supported publication. Subscribe to follow the METRIS journey.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[January 22, 2026. Mark Your Calendars. ]]></title><description><![CDATA[AI Governance now has a score]]></description><link>https://ainstein.sanjeevaniai.com/p/january-22-2026-mark-your-calendars</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/january-22-2026-mark-your-calendars</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Fri, 23 Jan 2026 19:46:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!RNEY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RNEY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RNEY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png 424w, https://substackcdn.com/image/fetch/$s_!RNEY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png 848w, https://substackcdn.com/image/fetch/$s_!RNEY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png 1272w, https://substackcdn.com/image/fetch/$s_!RNEY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RNEY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png" width="1404" height="1300" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1300,&quot;width&quot;:1404,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2693377,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/185571045?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RNEY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png 424w, https://substackcdn.com/image/fetch/$s_!RNEY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png 848w, https://substackcdn.com/image/fetch/$s_!RNEY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png 1272w, https://substackcdn.com/image/fetch/$s_!RNEY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16833b50-b8ab-4886-929e-c0380607a106_1404x1300.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Not because Singapore launched a framework.</p><p>Not because South Korea&#8217;s AI law went live.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><strong>Because METRIS arrived.</strong></p><p>And honestly? The timing couldn&#8217;t be more poetic.</p><p>On the same day that governments finally admitted AI governance needs real infrastructure - not more PDFs - we shipped exactly that.</p><p>Let me explain.</p><h3><strong>What Happened on January 22</strong></h3><p>Singapore released the world&#8217;s first agentic AI governance framework. 27 pages of guidance on how enterprises should govern AI agents.</p><p>South Korea&#8217;s AI Basic Act became enforceable. Mandatory impact assessments. Documentation requirements. Fines for non-compliance.</p><p>Two major economies. Two different approaches. Same message:</p><p><strong>The era of &#8220;we&#8217;ll figure out AI governance later&#8221; is over.</strong></p><p>But here&#8217;s what caught my attention in Singapore&#8217;s framework:</p><blockquote><p><em>&#8220;AI risk is no longer static, it is dynamic and behavioral.&#8221;</em></p></blockquote><p>Finally. Someone said it out loud.</p><h3><strong>The Problem With Frameworks</strong></h3><p>Frameworks tell you <em>what</em> to do.</p><p>They don&#8217;t tell you <em>how well</em> you&#8217;re doing it.</p><ul><li><p>&#8220;Implement human oversight&#8221; &#8594; But how do you measure if it&#8217;s meaningful?</p></li><li><p>&#8220;Assess and bound risks&#8221; &#8594; But what&#8217;s your actual risk score?</p></li><li><p>&#8220;Enable accountability&#8221; &#8594; But across which of the 9 regulatory frameworks that apply to you?</p></li></ul><p>We&#8217;ve spent years collecting frameworks like Pok&#233;mon cards. EU AI Act. ISO 42001. NIST AI RMF. Singapore MGF. Korea AI Basic Act.</p><p>And yet - <strong>94% of AI repositories still fail basic governance requirements.</strong></p><p>We know this because we measured it. 2,000+ repositories. 1,429 checkpoints. 9 regulatory frameworks.</p><p>The market is drowning in frameworks.</p><p>What it&#8217;s starving for is <strong>measurement.</strong></p><h3><strong>Enter METRIS</strong></h3><p>METRIS is what we&#8217;ve been building at Sanjeevani AI.</p><p>Not another framework. Not another checklist.</p><p><strong>A quantitative risk score for AI governance.</strong></p><p>Think of it like this:</p><ul><li><p>Frameworks tell you to &#8220;be healthy&#8221;</p></li><li><p>METRIS is your blood pressure reading</p></li></ul><p>Here&#8217;s what it does:</p><ul><li><p><strong>0-1000 Risk Score</strong> &#8594; Know exactly where you stand</p></li><li><p><strong>1,429 Checkpoints</strong> &#8594; Mapped across 9 regulatory frameworks</p></li><li><p><strong>Continuous Assessment</strong> &#8594; Not point-in-time audits</p></li><li><p><strong>Bayesian Scoring + Monte Carlo Modeling</strong> &#8594; Because governance isn&#8217;t binary</p></li></ul><p>The Singapore framework calls for &#8220;continuous monitoring&#8221; and &#8220;technical controls throughout the agent lifecycle.&#8221;</p><p>Great. <strong>METRIS is how you actually do that.</strong></p><h3><strong>Why January 22</strong></h3><p>We could have launched any day.</p><p>But when we saw Singapore&#8217;s Davos announcement on the calendar, and Korea&#8217;s enforcement date landing the same day, we knew.</p><p>This was the moment.</p><p>Not to ride their coattails - but to draw a line:</p><p><strong>January 22, 2026 is the day AI governance stopped being a conversation and started being infrastructure.</strong></p><p>They wrote the frameworks.</p><p>We built the measurement layer.</p><h3><strong>What This Means For You</strong></h3><p>If you&#8217;re an enterprise deploying AI - especially agentic AI - here&#8217;s the reality:</p><ol><li><p><strong>Voluntary frameworks become market expectations.</strong> Singapore&#8217;s isn&#8217;t mandatory. It doesn&#8217;t matter. Your customers, partners, and investors will expect you to comply.</p></li><li><p><strong>Mandatory requirements are cascading.</strong> Korea today. EU AI Act implementation ongoing. Others will follow.</p></li><li><p><strong>&#8220;We&#8217;re working on governance&#8221; isn&#8217;t an answer anymore.</strong> The question is: <em>What&#8217;s your score?</em></p></li></ol><p>The enterprises that can answer that question - with data, not promises - will own the trust advantage.</p><h3><strong>One Ask</strong></h3><p><em><strong>If you&#8217;re building AI and wrestling with governance, whether you&#8217;re trying to comply with frameworks, preparing for audits, or just trying to figure out where you actually stand, I want to hear from you.</strong></em></p><p>Reply to this email. Tell me what&#8217;s broken. What&#8217;s working. What you wish existed.</p><p>I read everything.</p><p><strong>January 22, 2026. Mark your calendars.</strong></p><p>The day AI governance got a score.</p><p></p><p><strong>Suneeta Modekurty</strong></p><p><em>Founder, SANJEEVANI AI | Creator of METRIS</em></p><ul><li><p><a href="https://sanjeevaniai.com">METRIS - SANJEEVANI AI</a></p></li><li><p><a href="https://www.imda.gov.sg/-/media/imda/files/about/emerging-tech-and-research/artificial-intelligence/mgf-for-agentic-ai.pdf">Singapore Agentic AI Framework (PDF)</a></p></li><li><p><a href="https://fpf.org/blog/south-koreas-new-ai-framework-act-a-balancing-act-between-innovation-and-regulation/">South Korea AI Basic Act Overview</a></p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The ROI Case Your CFO Can't Ignore]]></title><description><![CDATA[Why data governance isn't a cost center but is a profit protection strategy]]></description><link>https://ainstein.sanjeevaniai.com/p/the-roi-case-your-cfo-cant-ignore</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/the-roi-case-your-cfo-cant-ignore</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 12 Jan 2026 15:01:06 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xdTg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xdTg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xdTg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png 424w, https://substackcdn.com/image/fetch/$s_!xdTg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png 848w, https://substackcdn.com/image/fetch/$s_!xdTg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png 1272w, https://substackcdn.com/image/fetch/$s_!xdTg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xdTg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png" width="1442" height="1328" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1328,&quot;width&quot;:1442,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3410213,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/183372691?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xdTg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png 424w, https://substackcdn.com/image/fetch/$s_!xdTg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png 848w, https://substackcdn.com/image/fetch/$s_!xdTg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png 1272w, https://substackcdn.com/image/fetch/$s_!xdTg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcab7ddb8-3c28-4c82-b9a2-e1d690f45ff4_1442x1328.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>S called me last week with a problem.</p><p>&#8220;I used your cost calculator. Ran our numbers. We&#8217;re bleeding almost $800,000 a year on data quality issues.&#8221;</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>&#8220;That&#8217;s a big number,&#8221; I said.</p><p>&#8220;That&#8217;s the problem. It&#8217;s too big. When I showed my CFO, she didn&#8217;t see an opportunity. She saw a disaster. Her first question was &#8216;How did we let this get so bad?&#8217; Her second question was &#8216;How much will it cost to fix?&#8217;&#8221;</p><p>S paused.</p><p>&#8220;I didn&#8217;t have an answer.&#8221;</p><h2>The CFO&#8217;s Real Question</h2><p>Here&#8217;s what I&#8217;ve learned after 25 years in this space: executives don&#8217;t fund problems. They fund solutions with returns.</p><p>Showing your CFO that bad data costs $800,000 annually doesn&#8217;t get you a budget. It gets you a meeting where everyone argues about whose fault it is.</p><p>What gets you a budget is showing that a $150,000 investment returns $500,000 in the first year.</p><p>That&#8217;s not a problem. That&#8217;s a 233% ROI.</p><h2>The Math That Changes the Conversation</h2><p>Let me show you how to reframe the data quality conversation from &#8220;look how broken we are&#8221; to &#8220;look what we can capture.&#8221;</p><p><strong>Step 1: Start with your cost calculator numbers</strong></p><p>Let&#8217;s say your organization identified:</p><ul><li><p>$312,000 in rework costs (data teams cleaning instead of analyzing)</p></li><li><p>$450,000 in failed project costs (3 projects that never delivered)</p></li><li><p>$200,000 in bad decision costs (choices made on flawed data)</p></li><li><p>$12,000 in compliance documentation time</p></li></ul><p>Total: $974,000 annually</p><p><strong>Step 2: Apply realistic reduction rates</strong></p><p>You won&#8217;t eliminate 100% of these costs. But industry benchmarks show:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!X1LQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!X1LQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!X1LQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!X1LQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!X1LQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!X1LQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1794818,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/183372691?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!X1LQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!X1LQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!X1LQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!X1LQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa21ef952-50db-49ea-a3e3-53bb73a3bd5f_1408x768.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Step 3: Calculate the return</strong></p><p>Using moderate investment on our $974,000 example:</p><ul><li><p>Cost reduction (40%): $389,600 saved annually</p></li><li><p>Investment required: $120,000 (tools, training, dedicated resource)</p></li><li><p>First year net return: $269,600</p></li><li><p>ROI: 225%</p></li><li><p>Payback period: 3.7 months</p></li></ul><p>That&#8217;s the slide your CFO wants to see.</p><h2>The Costs They&#8217;re Not Counting</h2><p>When I work with organizations, I always find costs the calculator missed. These are harder to quantify but often larger:</p><p><strong>Opportunity cost.</strong> Your data scientists spend 50% of their time cleaning data. That&#8217;s not just a labor cost - it&#8217;s innovation that never happened. Products that never launched. Insights that never surfaced.</p><p><strong>Decision latency.</strong> When leaders don&#8217;t trust the data, they delay decisions. They ask for &#8220;one more analysis.&#8221; They convene committees. In fast-moving markets, slow decisions are expensive decisions.</p><p><strong>Talent attrition.</strong> Good data professionals don&#8217;t stay at organizations with bad data infrastructure. The cost of replacing a senior data scientist - recruiting, onboarding, ramp-up time - runs $150,000 to $300,000.</p><p><strong>Insurance premiums.</strong> This one surprises people. D&amp;O insurance carriers are starting to ask about AI governance. Poor data governance is becoming an underwriting factor. I&#8217;ve seen premiums increase 40-60% for organizations that can&#8217;t demonstrate controls.</p><p><strong>Regulatory exposure.</strong> The EU AI Act penalties can reach &#8364;35 million or 7% of global revenue. GDPR fines have already exceeded &#8364;4 billion total. These aren&#8217;t theoretical risks anymore.</p><h2>Building the Business Case</h2><p>Here&#8217;s the structure that works:</p><p><strong>Page 1: Current State (The Burning Platform)</strong></p><p>Don&#8217;t lead with this, but you need it. Show the calculated costs. Keep it factual, not accusatory. Frame it as &#8220;what we discovered&#8221; not &#8220;what went wrong.&#8221;</p><p><strong>Page 2: The Opportunity (This Is Your Real Opener)</strong></p><p>Lead with the ROI. &#8220;A $120,000 investment yields $389,600 in annual savings - a 225% return with 3.7 month payback.&#8221;</p><p>Now you have attention.</p><p><strong>Page 3: Risk Reduction (The Insurance Policy)</strong></p><p>Beyond cost savings, governance reduces:</p><ul><li><p>Regulatory penalty exposure</p></li><li><p>Reputational risk from AI failures</p></li><li><p>D&amp;O insurance premium increases</p></li><li><p>Audit findings and remediation costs</p></li></ul><p>Quantify what you can. Acknowledge what you can&#8217;t but name the risks.</p><p><strong>Page 4: Competitive Advantage (The Growth Angle)</strong></p><p>Organizations with mature data governance:</p><ul><li><p>Launch AI initiatives faster (they trust their data)</p></li><li><p>Win more RFPs (they can answer the governance questions)</p></li><li><p>Attract better talent (professionals want to work with good data)</p></li><li><p>Command premium valuations (acquirers and investors check this now)</p></li></ul><p><strong>Page 5: The Ask (Specific and Staged)</strong></p><p>Don&#8217;t ask for everything at once. Propose phases:</p><ul><li><p>Phase 1: Assessment and quick wins ($30,000, 6 weeks)</p></li><li><p>Phase 2: Foundation building ($60,000, 3 months)</p></li><li><p>Phase 3: Optimization and scaling ($30,000, ongoing)</p></li></ul><p>Smaller initial asks get approved faster.</p><h2>The Conversation Shift</h2><p>S called me again yesterday.</p><p>&#8220;I rebuilt the presentation. Led with the ROI. Showed the 225% return and 4-month payback.&#8221;</p><p>&#8220;And?&#8221;</p><p>&#8220;She asked how fast we could start.&#8221;</p><p>That&#8217;s the shift. Same numbers. Different framing. Completely different outcome.</p><p>Your CFO doesn&#8217;t need to understand data quality. She needs to understand returns. Speak her language, and the budget follows.</p><h2>Your Turn</h2><p>I&#8217;ve built an ROI Calculator that structures this entire conversation. It takes your cost calculator outputs and transforms them into CFO-ready business cases.</p><p>Includes:</p><ul><li><p>Investment scenario modeling</p></li><li><p>ROI and payback calculations</p></li><li><p>Risk reduction quantification</p></li><li><p>One-page executive summary template</p></li><li><p>Talking points for the budget conversation</p><p></p></li></ul><p>Ok, a quick note on &#8216;How the Calculator Works&#8217;</p><p>The math is simple but powerful:</p><p><strong>Step 1:</strong> Enter your total current cost (from the Data Quality Cost Calculator, or estimate)</p><p><strong>Step 2:</strong> The calculator shows three investment scenarios:</p><p><strong>Step 3:</strong> It calculates automatically:</p><ul><li><p><strong>Annual Savings</strong> = Your Current Cost &#215; Cost Reduction %</p></li><li><p><strong>Net Return</strong> = Annual Savings - Investment</p></li><li><p><strong>ROI</strong> = Net Return &#247; Investment</p></li><li><p><strong>Payback Period</strong> = Investment &#247; Annual Savings &#215; 12 months</p></li></ul><p><strong>Example:</strong> If your current data quality costs are $974,000 annually and you choose moderate investment:</p><ul><li><p>Annual Savings: $974,000 &#215; 40% = $389,600</p></li><li><p>Net Return: $389,600 - $75,000 = $314,600</p></li><li><p>ROI: $314,600 &#247; $75,000 = <strong>419%</strong></p></li><li><p>Payback: $75,000 &#247; $389,600 &#215; 12 = <strong>2.3 months</strong></p></li></ul><p>That&#8217;s the slide your CFO wants to see.</p><p></p><p><strong>Your CFO doesn&#8217;t need to understand data quality. They need to understand returns. Same numbers, different framing, completely different outcome.</strong></p><p><strong>Try it yourself.</strong></p><p><strong>Download: &#8220;AI Governance ROI Calculator&#8221;</strong> and Transform your cost analysis into a CFO-ready business case with ROI modeling, payback calculations, and executive talking points.</p><p><strong><a href="https://forms.gle/xqmC4P5K4CaS2LZX9">Download AI Governance ROI Calculator</a></strong></p><p></p><p><em>Next week: The regulatory reckoning. The EU AI Act, NIST AI RMF, and a wave of state laws are creating deadlines most organizations aren&#8217;t ready for. What happens when &#8220;we&#8217;ll figure it out later&#8221; meets &#8220;the deadline is now.&#8221;</em></p><p></p><p></p><p></p><p><em>Founder of SANJEEVANI AI. ISO/IEC 42001 Lead Auditor. 25+ years in AI, data, and compliance across HealthTech, FinTech, EdTech, and Insurance. Building METRIS, the quantitative AI governance platform.</em></p><p></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">A.I.N.S.T.E.I.N is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Anatomy of a Dataset]]></title><description><![CDATA[What You See Versus What's Actually There]]></description><link>https://ainstein.sanjeevaniai.com/p/the-anatomy-of-a-dataset</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/the-anatomy-of-a-dataset</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 05 Jan 2026 15:01:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!XbO7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XbO7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XbO7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!XbO7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!XbO7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!XbO7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XbO7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2077484,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/182204690?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XbO7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!XbO7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!XbO7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!XbO7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd80810aa-a654-4c89-8ff4-4241d0e6edc8_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>S showed up with a laptop this time. Screen already open to a spreadsheet.</p><p>&#8220;I did the lineage exercise you suggested,&#8221; S said. &#8220;Traced one number from the board report all the way back to its source.&#8221;</p><p>&#8220;How did it go?&#8221;</p><p>&#8220;Painful. But useful. I found three places where the data gets transformed and nobody documented why.&#8221;</p><p>&#8220;That&#8217;s progress.&#8221;</p><p>&#8220;But it raised another question.&#8221; S turned the laptop toward me. &#8220;I got to the source dataset. The original table. And I realized I don&#8217;t actually know what&#8217;s in it.&#8221;</p><p>&#8220;What do you mean?&#8221;</p><p>&#8220;I mean I can see the columns. I can see the rows. I can see the values. But I don&#8217;t know if I can trust any of it. I don&#8217;t know what&#8217;s missing. I don&#8217;t know what&#8217;s wrong. I don&#8217;t know what assumptions are baked in. I&#8217;m looking at it, but I&#8217;m not seeing it.&#8221;</p><h3>The Illusion of Understanding</h3><p>This is one of the most common traps in working with data. You open a dataset. It has columns and rows. It has labels that seem clear. Customer_ID. Transaction_Date. Revenue. Status.</p><p>You assume you understand it.</p><p>But the surface of a dataset tells you almost nothing about what&#8217;s actually inside. The labels tell you what someone intended to capture. They don&#8217;t tell you what was actually captured. They don&#8217;t tell you what&#8217;s missing. They don&#8217;t tell you what&#8217;s inconsistent. They don&#8217;t tell you what&#8217;s changed since the dataset was created.</p><p>&#8220;I pulled some basic statistics,&#8221; S said. &#8220;Row counts. Column types. That kind of thing.&#8221;</p><p>&#8220;That&#8217;s a start. What did you find?&#8221;</p><p>&#8220;Nothing obviously wrong. Which is almost worse. I don&#8217;t know if there&#8217;s nothing wrong, or if I just don&#8217;t know how to look.&#8221;</p><p>&#8220;There&#8217;s a difference between looking at data and profiling data. Most people do the first. Very few do the second.&#8221;</p><h3>What Data Profiling Actually Means</h3><p>Data profiling is the systematic examination of a dataset to understand its structure, content, and quality. It goes far beyond opening a file and scrolling through rows.</p><p>A proper profile examines several dimensions.</p><p><strong>Structure:</strong> How many rows? How many columns? What are the data types? Are the types consistent with what the labels suggest? A column called &#8220;Age&#8221; should contain numbers. A column called &#8220;Email&#8221; should contain text in a specific format. If the types don&#8217;t match expectations, something is wrong.</p><p><strong>Completeness:</strong> Which fields have missing values? How many? Is the missingness random or systematic? If 90 percent of records are missing a field, that field is probably useless. If only records from a certain time period are missing a field, something changed in the data collection process.</p><p><strong>Uniqueness:</strong> Are there duplicates? How do you define a duplicate? Two records with the same customer ID? The same email? The same name and address? The answer depends on context, and getting it wrong can significantly skew any analysis.</p><p><strong>Distribution:</strong> What values actually appear in each column? What&#8217;s the minimum? The maximum? The mean? The median? Are there outliers? A column called &#8220;Age&#8221; should probably have values between 0 and 120. If there&#8217;s a value of 999 or -1, that&#8217;s likely a placeholder for missing data, not an actual age.</p><p><strong>Consistency:</strong> Do related fields make sense together? If a record has a &#8220;Signup_Date&#8221; of 2025 but a &#8220;First_Purchase_Date&#8221; of 2019, something is wrong. If a &#8220;State&#8221; field says &#8220;California&#8221; but the &#8220;Zip_Code&#8221; starts with &#8220;100,&#8221; that&#8217;s New York, not California.</p><p><strong>Validity:</strong> Do the values conform to known rules? Email addresses should match a pattern. Phone numbers should have the right number of digits. Status codes should come from a defined list. Dates should be actual dates.</p><p>&#8220;That&#8217;s a lot to check,&#8221; S said.</p><p>&#8220;It is. But most of it can be automated. The problem isn&#8217;t that it&#8217;s hard to do. The problem is that people don&#8217;t do it.&#8221;</p><h3>The Problems Hiding in Plain Sight</h3><p>Let me give you some examples of what profiling reveals that casual inspection misses.</p><p><strong>The 99 Percent Problem</strong></p><p>Let&#8217;s say you have a dataset of customer orders. It has columns for Order_ID, Customer_Name, Product, Amount, and Country. You&#8217;re asked to analyze international sales trends.</p><p>You open the file. You scroll through the Country column. You see &#8220;USA,&#8221; &#8220;Canada,&#8221; &#8220;UK,&#8221; &#8220;Germany,&#8221; &#8220;Australia.&#8221; Looks good. You have international data.</p><p>But when you actually count the values, you discover that 99 percent of the orders are from &#8220;USA.&#8221; The remaining 1 percent is scattered across 47 different countries, with most having fewer than 10 orders each.</p><p>If you&#8217;re trying to understand international buying patterns, this dataset is nearly useless. You don&#8217;t have enough data from any country except the US to draw meaningful conclusions. But you&#8217;d never know this from scrolling. You&#8217;d only know it from counting.</p><p><strong>The Placeholder Problem</strong></p><p>Let&#8217;s say you have a dataset of employees. It has columns for Employee_ID, Name, Department, Start_Date, and Annual_Salary. You&#8217;re asked to calculate the average salary by department.</p><p>You scroll through the Salary column. You see values like $52,000, $78,000, $95,000, $120,000. The numbers look reasonable for salaries.</p><p>But when you look at the full distribution, you find something strange. About 15 percent of the values are exactly $0. Another 5 percent are exactly $999,999.</p><p>These aren&#8217;t real salaries. The $0 values are probably contractors or interns whose salary wasn&#8217;t entered. The $999,999 values are probably executives whose salary is confidential and someone entered a placeholder instead.</p><p>If you calculate the average salary including these values, you get a number that represents nobody. The zeros pull it down. The 999,999 values pull it up. The result is meaningless.</p><p><strong>The Format Problem</strong></p><p>Let&#8217;s say you have a dataset of transactions. It has columns for Transaction_ID, Customer_ID, Date, and Amount. You&#8217;re asked to analyze sales trends by month.</p><p>You scroll through the Date column. You see values like &#8220;01/02/2024,&#8221; &#8220;15/03/2024,&#8221; &#8220;22/04/2024.&#8221; They look like dates. Good.</p><p>But when you look more carefully, you notice something. Some records show &#8220;01/02/2024.&#8221; Others show &#8220;2024-02-01.&#8221; Others show &#8220;02-01-2024.&#8221;</p><p>Three different formats in the same column. And here&#8217;s the problem: is &#8220;01/02/2024&#8221; January 2nd or February 1st? In the US, it&#8217;s January 2nd. In Europe, it&#8217;s February 1st. In the ISO format, it would be written differently altogether.</p><p>If this data came from multiple sources or multiple time periods, the formats might be mixed. Without checking, you could be putting January transactions in February and vice versa. Your monthly trend analysis would be wrong, and you&#8217;d never know why.</p><p><strong>The Evolution Problem</strong></p><p>Let&#8217;s say you have a dataset of product sales going back five years. It has columns for Sale_ID, Date, Product_Name, Category, and Revenue. You&#8217;re asked to show how each product category has grown over time.</p><p>You scroll through the Category column. You see &#8220;Electronics,&#8221; &#8220;Clothing,&#8221; &#8220;Home Goods,&#8221; &#8220;Sports.&#8221; Looks consistent.</p><p>But when you filter by year and look at the values, you discover something. In 2020 and 2021, there was one category called &#8220;Electronics.&#8221; In 2022 and 2023, that category was split into &#8220;Consumer Electronics&#8221; and &#8220;Professional Electronics.&#8221; In 2024, they merged it back into just &#8220;Electronics.&#8221;</p><p>The label looks the same in 2020 and 2024, but they mean different things. In 2020, &#8220;Electronics&#8221; included everything. In 2024, &#8220;Electronics&#8221; is the merger of two categories that existed separately for two years.</p><p>If you draw a trend line for &#8220;Electronics&#8221; across five years, the line is lying to you. The dip in 2022-2023 isn&#8217;t a real decline. It&#8217;s just that the category was split. The spike in 2024 isn&#8217;t real growth. It&#8217;s just that they merged it back.</p><p><strong>The Silent Null Problem</strong></p><p>Let&#8217;s say you have a dataset of customer contacts. It has columns for Customer_ID, Name, Email, Phone, and Address. You&#8217;re asked to verify that the database is complete before a marketing campaign.</p><p>You run a quick check for missing values. The system tells you every row has a value in every column. 100 percent complete. Great.</p><p>But when you look at the actual values in the Email column, you find entries like &#8220;N/A,&#8221; &#8220;NA,&#8221; &#8220;none,&#8221; &#8220;null,&#8221; &#8220;not provided,&#8221; &#8220;-,&#8221; and what looks like empty space but is actually a single space character.</p><p>These are all different ways that people entered &#8220;we don&#8217;t have this information.&#8221; But because they typed something instead of leaving it blank, the system counts them as complete.</p><p>Your 100 percent completeness is fiction. You might actually be missing email addresses for 20 percent of your customers. If you send a marketing campaign based on this &#8220;complete&#8221; data, 20 percent of your emails go nowhere.</p><p>&#8220;I&#8217;ve seen some of these,&#8221; S said. &#8220;The placeholder thing especially. We have records with birthdates of January 1, 1900.&#8221;</p><p>&#8220;That&#8217;s a classic. January 1, 1900 is the default date in many systems. It gets entered when the real date wasn&#8217;t known. If you calculate average customer age including those records, your average customer is 125 years old. Obviously nonsense.&#8221;</p><h3>The Metadata Gap</h3><p>There&#8217;s another layer that profiling reveals: the gap between what documentation says and what data contains.</p><p>Let me give you an example.</p><p>Let&#8217;s say you have a dataset of support tickets. It has columns for Ticket_ID, Customer_ID, Date_Opened, Date_Closed, Status, and Priority. You&#8217;re asked to analyze how quickly tickets get resolved.</p><p>You find a data dictionary that someone wrote two years ago. It says the Status column contains three possible values: &#8220;Open,&#8221; &#8220;Closed,&#8221; or &#8220;Pending.&#8221;</p><p>You start writing your analysis based on this. Open tickets are active. Closed tickets are resolved. Pending tickets are waiting for something.</p><p>But when you actually look at the Status column, you find values like &#8220;Open,&#8221; &#8220;OPEN,&#8221; &#8220;open,&#8221; &#8220;Closed,&#8221; &#8220;closed,&#8221; &#8220;CLOSED,&#8221; &#8220;Pending,&#8221; &#8220;pending,&#8221; &#8220;Resolved,&#8221; &#8220;Cancelled,&#8221; &#8220;Escalated,&#8221; &#8220;On Hold,&#8221; &#8220;Waiting for Customer,&#8221; and &#8220;Duplicate.&#8221;</p><p>The documentation said three values. Reality has fifteen. Some are just capitalization differences. Others are completely new statuses that were added after the documentation was written.</p><p>If you build your analysis assuming only three statuses exist, you&#8217;ll miss or miscategorize a significant portion of your tickets. Your resolution time calculation will be wrong because you didn&#8217;t account for &#8220;Escalated&#8221; or &#8220;On Hold&#8221; tickets.</p><p>Documentation tells you what someone once believed about the data. Profiling tells you what&#8217;s actually there today. These are often very different things.</p><p>&#8220;So the documentation is lies,&#8221; S said.</p><p>&#8220;Not lies. Outdated truths. The documentation was probably accurate when someone wrote it two years ago. But then the support team added new statuses. Someone started entering values in all caps. A new manager created an &#8216;Escalated&#8217; category. And nobody updated the documentation because nobody owns it.&#8221;</p><h3>The Upstream Blindness Problem</h3><p>Here&#8217;s another thing profiling reveals that&#8217;s particularly important if you&#8217;re building AI or machine learning models.</p><p>Let me explain with an example.</p><p>Let&#8217;s say you&#8217;re building a model to predict which customers will cancel their subscription. You have a dataset of past customers with columns for Customer_ID, Age, Income, Subscription_Length, Support_Tickets_Filed, and Cancelled (yes or no).</p><p>You train the model. It works well in testing. You deploy it.</p><p>Six months later, the model is making terrible predictions. Customers it said would stay are leaving. Customers it said would leave are staying. What happened?</p><p>You go back and profile the original training data. Here&#8217;s what you find.</p><p>The Age column had placeholder values. About 10 percent of customers had an age of -1, which was entered when age was unknown. The model learned that customers with age -1 had specific cancellation patterns. But -1 isn&#8217;t a real age. It&#8217;s a data entry convention.</p><p>The Income column had duplicates. Some customers appeared multiple times with slightly different information. The model learned those examples more heavily because it saw them more often. But they weren&#8217;t more important. They were just duplicated.</p><p>The Support_Tickets_Filed column changed meaning over time. In the first year of data, it counted all tickets. In the second year, it only counted unresolved tickets. The model learned patterns based on a number that meant different things at different times.</p><p>The model didn&#8217;t know these were data quality issues. The model just saw patterns. And it learned whatever patterns helped it make predictions, including patterns that only existed because of data problems.</p><p>This is how you get models that work in testing but fail in production. They learned the quirks of the training data, not the underlying reality.</p><p>&#8220;So profiling is actually a prerequisite for machine learning,&#8221; S said.</p><p>&#8220;It should be. But teams are under pressure to ship models quickly. Profiling feels like a delay. So they skip it, train on whatever data they have, and hope for the best. Sometimes they get lucky. Often they don&#8217;t. And when the model fails, they blame the algorithm instead of examining the data.&#8221;</p><h3>What a Proper Profile Contains</h3><p>Let me describe what a thorough data profile actually looks like, using a concrete example.</p><p>Let&#8217;s say you have a dataset of online orders. It has columns for Order_ID, Customer_ID, Order_Date, Product_ID, Quantity, Unit_Price, Total_Amount, Shipping_Address, and Status.</p><p>A proper profile would examine this dataset across several dimensions.</p><p><strong>Summary statistics for every column</strong></p><p>For Order_ID: How many total orders? Are all IDs unique, or are there duplicates?</p><p>For Quantity: What&#8217;s the minimum value? Maximum? Average? Is the minimum 1, or are there zeros or negative numbers that shouldn&#8217;t exist?</p><p>For Unit_Price: What&#8217;s the range? Is the minimum $0.01 or $0.00? Is the maximum reasonable, or is there a $999,999 placeholder?</p><p>For Status: What values appear? Just &#8220;Shipped&#8221; and &#8220;Delivered&#8221;? Or also &#8220;Cancelled,&#8221; &#8220;Returned,&#8221; &#8220;Processing,&#8221; &#8220;On Hold,&#8221; and twelve other variations?</p><p><strong>Completeness analysis</strong></p><p>Which columns have missing values?</p><p>Maybe Shipping_Address is missing for 5 percent of orders. Are those digital products that don&#8217;t need shipping? Or are they data entry errors?</p><p>Maybe Customer_ID is missing for 2 percent of orders. Are those guest checkouts? Or are they system errors?</p><p>Is the missingness random, or is it concentrated in certain time periods or certain product types?</p><p><strong>Uniqueness analysis</strong></p><p>Order_ID should be unique. Is it? If the same Order_ID appears twice, do both rows have identical information, or different information? Which one is correct?</p><p>Are there duplicate orders where everything is the same except Order_ID? That might indicate the same order was entered twice by mistake.</p><p><strong>Distribution analysis</strong></p><p>For Quantity, most orders probably have quantities between 1 and 5. If there&#8217;s an order with quantity 10,000, is that a real bulk order or a data entry error?</p><p>For Total_Amount, what does the distribution look like? Is it clustered around certain price points? Are there outliers like $0.00 orders or $50,000 orders that need investigation?</p><p><strong>Cross-field validation</strong></p><p>Does Quantity times Unit_Price equal Total_Amount? If not, which field is wrong?</p><p>If Status is &#8220;Delivered,&#8221; is there a delivery date somewhere? If Status is &#8220;Cancelled,&#8221; is Total_Amount zero or was the customer still charged?</p><p>Does Order_Date make sense? Are there orders dated in the future? Orders dated before the company existed?</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/subscribe?"><span>Subscribe now</span></a></p><p></p><p><strong>Temporal analysis</strong></p><p>Are orders distributed evenly over time, or are there gaps? A gap might mean data wasn&#8217;t collected during that period.</p><p>Did the average order value change suddenly at some point? That might indicate a pricing change, or it might indicate a data issue.</p><p>Did new Status values start appearing at a certain date? That might indicate a system change that needs to be accounted for.</p><p>&#8220;That sounds like a lot of work for every dataset,&#8221; S said.</p><p>&#8220;It sounds like more work than it is. Most of this can be automated. You run a profiling tool, and it generates these statistics in seconds. The real work isn&#8217;t generating the numbers. It&#8217;s reviewing them and deciding what they mean.&#8221;</p><h3>The Decision That Follows</h3><p>Profiling isn&#8217;t an end in itself. It&#8217;s the foundation for a decision: can you trust this data for your intended purpose?</p><p>Let me give you four scenarios.</p><p><strong>Scenario 1: Green light.</strong></p><p>You profile the order dataset. Everything looks clean. Order_IDs are unique. Quantities and prices are in reasonable ranges. Total_Amount matches Quantity times Unit_Price in 99.9 percent of cases. Missing values are minimal and explainable. Status values match the documentation.</p><p>Decision: You can proceed with confidence.</p><p><strong>Scenario 2: Yellow light with caveats.</strong></p><p>You profile the dataset and find some issues. About 3 percent of orders have a Status of &#8220;Unknown&#8221; that isn&#8217;t in the documentation. About 1 percent have Total_Amount that doesn&#8217;t match Quantity times Unit_Price.</p><p>Decision: You can proceed, but document the limitations. Exclude the &#8220;Unknown&#8221; status orders from your analysis, or treat them as a separate category. Flag the mismatched amounts for investigation but don&#8217;t let them block the whole project.</p><p><strong>Scenario 3: Red light, fixable.</strong></p><p>You profile the dataset and find significant issues. The Order_Date column has three different formats. About 15 percent of Unit_Price values are $0.00, which are probably placeholder values. There are duplicate Order_IDs that need to be resolved.</p><p>Decision: You cannot proceed until these issues are fixed. The date formats need to be standardized. The $0.00 prices need to be investigated and either corrected or excluded. The duplicates need to be resolved. This is data cleaning work that must happen before analysis.</p><p><strong>Scenario 4: Red light, unfixable.</strong></p><p>You profile the dataset and find fundamental problems. The data only covers three months, but you need a full year for seasonal analysis. Half the customers are missing demographic information that&#8217;s essential for your segmentation. The Status definitions changed twice during the time period and there&#8217;s no way to reconcile them.</p><p>Decision: This data cannot support your intended purpose. You need different data, or you need to change what you&#8217;re trying to do. This is painful to admit, but better to know now than to build something on a broken foundation.</p><p>&#8220;How often is it that last one?&#8221; S asked.</p><p>&#8220;More often than people want to admit. Projects get approved based on assumptions about data availability. Someone says &#8216;we have two years of customer data&#8217; and everyone assumes it&#8217;s usable. Then someone finally profiles it and discovers half the fields are empty, the definitions changed three times, and it&#8217;s actually only useful for six months of analysis. By that point, there&#8217;s pressure to proceed anyway. That&#8217;s how bad analyses get built. That&#8217;s how bad models get deployed.&#8221;</p><h3>Making This Practical</h3><p>S closed the laptop.</p><p>&#8220;So before I trust any dataset, I should profile it. Understand what&#8217;s actually there. Compare it to what&#8217;s documented. Make an explicit decision about whether it&#8217;s fit for purpose.&#8221;</p><p>&#8220;That&#8217;s exactly right.&#8221;</p><p>&#8220;And if I&#8217;m going to train a model on it, I should be especially careful. Because the model will learn whatever patterns are in there, including patterns that only exist because of data problems.&#8221;</p><p>&#8220;Yes. The model doesn&#8217;t know that -1 means &#8216;missing.&#8217; The model doesn&#8217;t know that January 1, 1900 means &#8216;unknown date.&#8217; The model just sees numbers and finds patterns. Your job is to make sure the patterns it finds are real.&#8221;</p><p>&#8220;And most of this profiling can be automated, so it&#8217;s not as slow as it sounds.&#8221;</p><p>&#8220;Correct. The barrier isn&#8217;t technical. There are tools that generate profiles with a few clicks. The barrier is cultural. It&#8217;s making profiling a standard step that happens before anyone builds anything, not an afterthought that happens when something goes wrong.&#8221;</p><h3>Where This Goes Next</h3><p>&#8220;Next week,&#8221; I said, &#8220;we should talk about what happens after you deploy something. You&#8217;ve profiled the data. You&#8217;ve built your model or your report. It&#8217;s working today. But data changes over time. What happens when the data you trained on stops representing the data you&#8217;re seeing? That&#8217;s called drift, and it&#8217;s one of the most common reasons things fail in production.&#8221;</p><p>S nodded. &#8220;Same time?&#8221;</p><p>&#8220;Same time.&#8221;</p><p></p><p><em>Next week: Data Drift. Why yesterday's training data fails tomorrow, and how to detect when your data stops representing reality.</em></p><p></p><p><strong>You can&#8217;t trust what you haven&#8217;t examined. And scrolling through rows is not examination.</strong></p><p><strong>Download: &#8220;Dataset Profiling Checklist&#8221; &#8212; A systematic guide to understanding what&#8217;s actually in your data before you build anything on top of it.</strong></p><p><strong><a href="https://forms.gle/TwY4WsXGbt5SoPE16">Download Dataset Profiling Checklist</a></strong></p><p></p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/p/the-anatomy-of-a-dataset?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/p/the-anatomy-of-a-dataset?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p><p>Founder of SANJEEVANI AI. ISO/IEC 42001 Lead Auditor. 25+ years in AI, data, and compliance across HealthTech, FinTech, EdTech, and Insurance. Building METRIS, the quantitative AI governance platform.</p><p></p>]]></content:encoded></item><item><title><![CDATA[The Hidden Cost of "Good Enough" Data]]></title><description><![CDATA[Why Data Quality Problems Don't Announce Themselves Until It's Too Late]]></description><link>https://ainstein.sanjeevaniai.com/p/the-hidden-cost-of-good-enough-data</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/the-hidden-cost-of-good-enough-data</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 29 Dec 2025 15:02:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!sZeL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sZeL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sZeL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png 424w, https://substackcdn.com/image/fetch/$s_!sZeL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png 848w, https://substackcdn.com/image/fetch/$s_!sZeL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png 1272w, https://substackcdn.com/image/fetch/$s_!sZeL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sZeL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png" width="1384" height="1308" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1308,&quot;width&quot;:1384,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2966477,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/182194804?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sZeL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png 424w, https://substackcdn.com/image/fetch/$s_!sZeL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png 848w, https://substackcdn.com/image/fetch/$s_!sZeL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png 1272w, https://substackcdn.com/image/fetch/$s_!sZeL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa50a2518-9a94-42aa-86d1-703f87088d84_1384x1308.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>S came in looking frustrated.</p><p>&#8220;I&#8217;ve been trying to explain this to people at work,&#8221; S said, dropping into the chair. &#8220;What we talked about last week. About data being the real problem. About the six dimensions.&#8221;</p><p>&#8220;How&#8217;s that going?&#8221;</p><p>&#8220;They nod. They agree in principle. Then they say &#8216;but our data is good enough.&#8217;&#8221;</p><p>S made air quotes around &#8220;good enough.&#8221;</p><p>&#8220;And you don&#8217;t think it is?&#8221;</p><p>&#8220;I don&#8217;t know. That&#8217;s the problem. Nobody knows. Nobody&#8217;s measuring. They just assume it&#8217;s fine because nothing has visibly exploded.&#8221;</p><p>&#8220;Yet.&#8221;</p><p>&#8220;Yet,&#8221; S repeated. &#8220;So how do I show them what &#8216;good enough&#8217; is actually costing?&#8221;</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/subscribe?"><span>Subscribe now</span></a></p><p></p><h3>The Invisible Tax</h3><p>Here&#8217;s the thing about data quality problems. They don&#8217;t announce themselves.</p><p>A server crash is obvious. An application error throws a message on the screen. A security breach makes headlines.</p><p>But bad data? Bad data just sits there. Quietly. Doing damage you can&#8217;t see.</p><p>Think of it like a slow leak in your roof. Water drips into the insulation. The wood starts to rot. Mold grows in places you never look. Everything seems fine from the inside. Until one day the ceiling collapses.</p><p>Data quality problems work the same way. They accumulate silently. They compound over time. By the time you notice something is wrong, the damage has spread far beyond the original source.</p><p>This is why &#8220;good enough&#8221; is so dangerous. It feels safe. It feels pragmatic. But it&#8217;s actually a bet. You&#8217;re betting that the problems you can&#8217;t see won&#8217;t become problems you can&#8217;t ignore.</p><p>Most organizations lose that bet eventually.</p><h3>Where the Money Actually Goes</h3><p>S leaned forward. &#8220;Okay, but can we put numbers on this? When I say &#8216;data quality costs money,&#8217; people want specifics.&#8221;</p><p>Let&#8217;s get specific.</p><p>Gartner has estimated that poor data quality costs organizations an average of $12.9 million per year. But that number is abstract. It doesn&#8217;t tell you where the money goes.</p><p>Here&#8217;s where it actually goes:</p><p><strong>Rework.</strong> This is the biggest one, and the least visible. Analysts spend time cleaning, correcting, and reconciling data instead of analyzing it. Studies suggest that data professionals spend 40 to 60 percent of their time on data preparation. Think about that. You&#8217;re paying skilled people good salaries, and half their time goes to janitorial work that shouldn&#8217;t be necessary.</p><p><strong>Failed projects.</strong> Data science projects fail at a rate of 80 to 85 percent, depending on which study you read. Many of those failures trace back to data issues. The model couldn&#8217;t be trained because the data was insufficient. The model was trained but didn&#8217;t generalize because the training data didn&#8217;t represent reality. The model worked in testing but failed in production because nobody accounted for data drift. Each failed project represents months of work, significant salary costs, and opportunity cost of what those people could have been doing instead.</p><p><strong>Bad decisions.</strong> This is the hardest to quantify but potentially the most expensive. When decisions are based on incorrect data, the costs show up downstream. A marketing campaign targets the wrong customers. A supply chain forecast misses actual demand. A hiring model screens out qualified candidates. A credit model approves risky borrowers. Each of these decisions made sense based on the data available. The data was just wrong.</p><p><strong>Compliance failures.</strong> Regulators are increasingly asking organizations to demonstrate data quality. The EU AI Act requires that training data be &#8220;relevant, representative, free of errors and complete.&#8221; GDPR requires that personal data be accurate and kept up to date. When you can&#8217;t demonstrate compliance, you face fines. When you can demonstrate compliance but the documentation takes weeks to assemble, you face operational costs that add up quickly.</p><p><strong>Eroded trust.</strong> This one doesn&#8217;t show up on any balance sheet, but it might be the most corrosive. When stakeholders stop trusting the numbers, they stop using them. Executives make gut decisions instead of data-driven ones. Analysts hedge their findings with so many caveats that the insights become useless. The entire investment in data infrastructure delivers diminishing returns because nobody believes the output.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/p/the-hidden-cost-of-good-enough-data?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/p/the-hidden-cost-of-good-enough-data?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>The Compounding Problem</h3><p>&#8220;That&#8217;s a lot of categories,&#8221; S said. &#8220;But they feel separate. Like different problems that happen to share a cause.&#8221;</p><p>&#8220;They&#8217;re not separate. They feed each other.&#8221;</p><p>This is the part that most people miss. Data quality problems don&#8217;t stay contained. They compound.</p><p>Here&#8217;s how it works.</p><p>Bad data enters your system at the point of collection. Maybe a form field wasn&#8217;t validated properly. Maybe an integration dropped some records. Maybe someone made a typo during manual entry.</p><p>That bad data flows into your data warehouse. Now it&#8217;s sitting alongside good data, and there&#8217;s nothing to distinguish them. Your warehouse doesn&#8217;t know which records are accurate and which are garbage.</p><p>An analyst pulls data for a report. They don&#8217;t know about the quality issues because nobody documented them. The report goes to leadership. Leadership makes a decision based on the report.</p><p>The decision turns out to be wrong. But nobody traces it back to the data. Instead, they blame the analyst&#8217;s methodology. Or the model. Or the market conditions. The actual root cause remains hidden.</p><p>Meanwhile, the same bad data is being used to train machine learning models. The models learn the patterns in the data, including the patterns that exist only because of errors. The models get deployed. They make predictions. The predictions inform more decisions.</p><p>Each step amplifies the original problem. A small error at the source becomes a medium error in the warehouse becomes a significant error in the model becomes a costly error in the real world.</p><p>&#8220;So fixing it later is always more expensive than fixing it early,&#8221; S said.</p><p>&#8220;Exponentially more expensive. IBM estimated that data errors caught at the point of entry cost about $1 to fix. The same errors caught during data storage cost about $10 to fix. Caught during analysis, $100. Caught after a decision has been made and implemented, potentially thousands or millions.&#8221;</p><p>&#8220;And most organizations catch them at which stage?&#8221;</p><p>&#8220;Most organizations catch them when something visibly breaks. Which is the most expensive stage of all.&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Dfp9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Dfp9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png 424w, https://substackcdn.com/image/fetch/$s_!Dfp9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png 848w, https://substackcdn.com/image/fetch/$s_!Dfp9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png 1272w, https://substackcdn.com/image/fetch/$s_!Dfp9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Dfp9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png" width="1364" height="564" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:564,&quot;width&quot;:1364,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:311919,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/182194804?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Dfp9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png 424w, https://substackcdn.com/image/fetch/$s_!Dfp9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png 848w, https://substackcdn.com/image/fetch/$s_!Dfp9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png 1272w, https://substackcdn.com/image/fetch/$s_!Dfp9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91dcb5c-2bca-4bfb-b805-8cac3626eee8_1364x564.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>The &#8220;Good Enough&#8221; Trap</h3><p>S sat back. &#8220;So when people say &#8216;our data is good enough,&#8217; what they&#8217;re really saying is...&#8221;</p><p>&#8220;They&#8217;re saying they haven&#8217;t looked closely enough to see the problems. Or they&#8217;ve seen the problems but haven&#8217;t calculated the costs. Or they&#8217;ve calculated the costs but decided to defer them to the future.&#8221;</p><p>&#8220;Which just makes the future costs higher.&#8221;</p><p>&#8220;Exactly. &#8216;Good enough&#8217; is a form of technical debt. And like all technical debt, it accrues interest.&#8221;</p><p>There&#8217;s a psychological component to this too. Once an organization has decided that its data is &#8220;good enough,&#8221; it becomes very hard to revisit that decision. Admitting that data quality is a problem means admitting that past decisions made on that data might have been wrong. It means acknowledging that significant investment might be needed to fix things. It means taking responsibility for something that has no obvious owner.</p><p>So people don&#8217;t look. They don&#8217;t measure. They don&#8217;t ask uncomfortable questions. And the problems compound.</p><p>&#8220;How do you break that cycle?&#8221; S asked.</p><p>&#8220;You make the invisible visible. You measure. You quantify. You show people what &#8216;good enough&#8217; is actually costing them in terms they can&#8217;t ignore.&#8221;</p><h3>Making It Visible</h3><p>The first step is to stop treating data quality as a binary. Data is not simply &#8220;good&#8221; or &#8220;bad.&#8221; It exists on a spectrum across multiple dimensions. And each dimension can be measured.</p><p>Last week, we talked about the six dimensions: accuracy, completeness, consistency, timeliness, validity, and uniqueness. Each of these can be expressed as a number.</p><p>Accuracy can be measured by sampling records and verifying them against source documents or real-world facts. If you check 1,000 records and 950 are accurate, your accuracy rate is 95 percent. Is that good enough? It depends on the use case. For a marketing mailing list, maybe. For a medical diagnosis system, absolutely not.</p><p>Completeness can be measured by counting null or missing values in required fields. If a dataset has 100,000 records and 15,000 are missing a critical field, your completeness rate is 85 percent. What decisions are being made with that 15 percent gap? What are those decisions costing?</p><p>Consistency can be measured by comparing how the same entity is represented across different systems. If a customer appears in three databases with three different addresses, you have a consistency problem. How many downstream processes are affected by that inconsistency?</p><p>Timeliness can be measured by tracking the age of records relative to their expected refresh rate. If customer data should be updated monthly but hasn&#8217;t been touched in two years, you have a timeliness problem. What decisions are being made based on a reality that no longer exists?</p><p>Validity can be measured by checking records against defined business rules. Emails should match a certain format. Ages should fall within a certain range. Status codes should come from a defined list. Every violation represents a record that might cause problems downstream.</p><p>Uniqueness can be measured by identifying duplicate records. If the same customer appears five times with slightly different information, not only is the data wrong, but any analysis based on customer counts will be overstated.</p><p>&#8220;Once you have these numbers,&#8221; I told S, &#8220;you can start having a different conversation. Instead of &#8216;our data is good enough,&#8217; you can say &#8216;our data is 94 percent accurate, 85 percent complete, and has a 12 percent duplicate rate.&#8217; Those numbers mean something. They can be tracked over time. They can be tied to business outcomes.&#8221;</p><h3>Connecting Quality to Outcomes</h3><p>S picked up on something. &#8220;You said &#8216;tied to business outcomes.&#8217; How do you actually make that connection? How do you go from &#8216;our data has a 12 percent duplicate rate&#8217; to &#8216;this costs us X dollars&#8217;?&#8221;</p><p>This is where it gets interesting. And where most organizations stop short.</p><p>The connection isn&#8217;t always direct, but it can be traced. Let me give you a few examples.</p><p>Duplicate customer records mean inflated marketing costs. If your database says you have 100,000 customers but 12,000 of those are duplicates, you&#8217;re sending 12,000 unnecessary marketing emails. You&#8217;re paying for 12,000 extra impressions in advertising campaigns. You&#8217;re overstating your customer base to investors and board members. Each of these has a cost.</p><p>Incomplete address data means failed deliveries. If 15 percent of your customer addresses are missing or outdated, some percentage of physical shipments will fail. Each failed delivery has a direct cost for reshipping plus an indirect cost for customer frustration and potential churn.</p><p>Stale training data means model degradation. If your machine learning model was trained on data that&#8217;s two years old and the world has changed since then, your model&#8217;s predictions are increasingly wrong. Each wrong prediction has a cost. A fraud detection model that misses fraudulent transactions costs you money. A recommendation engine that suggests irrelevant products costs you sales.</p><p>Inconsistent definitions mean misaligned teams. If the marketing team defines &#8220;active customer&#8221; differently than the finance team, they&#8217;re making decisions based on different realities. When those decisions conflict, someone has to clean up the mess. That cleanup has a cost in time, meetings, reconciliation, and delayed execution.</p><p>&#8220;So the framework,&#8221; S said, &#8220;is to measure quality, then trace each quality dimension to specific business processes, then estimate the cost of failures in those processes.&#8221;</p><p>&#8220;That&#8217;s exactly right. It&#8217;s not always precise. Some costs are hard to quantify. But even rough estimates are better than the fiction of &#8216;good enough.&#8217;&#8221;</p><h3>The ROI Question</h3><p>S paused for a moment. &#8220;Okay. So we measure quality. We calculate costs. Then what?&#8221;</p><p>&#8220;Then you can make an actual business case for improvement.&#8221;</p><p>This is what transforms data quality from an IT concern into a strategic priority. When you can say &#8220;we&#8217;re losing approximately $2 million per year due to data quality issues, and an investment of $500,000 would reduce that by 60 percent,&#8221; you&#8217;re speaking a language that executives understand.</p><p>The ROI of data quality work is often surprisingly high. The problem is that nobody calculates it. They see the cost of the investment but not the cost of the status quo. So the status quo wins by default.</p><p>&#8220;But calculating it requires measuring first,&#8221; S said.</p><p>&#8220;Which brings us back to where we started. You can&#8217;t manage what you don&#8217;t measure. You can&#8217;t improve what you can&#8217;t quantify. You can&#8217;t make a business case for something you can&#8217;t express in numbers.&#8221;</p><p>&#8220;So the first step is always measurement.&#8221;</p><p>&#8220;The first step is always measurement.&#8221;</p><h3>What Changes This</h3><p>S stood up to go. Then stopped.</p><p>&#8220;Last week you mentioned that this is a governance issue, not just a technical issue. What did you mean by that?&#8221;</p><p>&#8220;I mean that fixing data quality isn&#8217;t just about better tools or more cleaning scripts. It&#8217;s about organizational change. It&#8217;s about who owns data quality, how it&#8217;s measured, how it&#8217;s incentivized, and how it connects to larger business objectives.&#8221;</p><p>&#8220;That sounds harder than just buying software.&#8221;</p><p>&#8220;It is harder. It&#8217;s also the only thing that actually works. You can buy the best data quality tools in the world, but if nobody owns the problem, if there are no standards, if there&#8217;s no accountability, those tools will just generate reports that nobody reads.&#8221;</p><p>&#8220;So how do you actually build that? The governance piece?&#8221;</p><p>&#8220;That&#8217;s a bigger conversation. For next week?&#8221;</p><p>S nodded. &#8220;Next week.&#8221;</p><p></p><p><em>Next week: Metrics that matter versus metrics that mislead. Why single numbers are almost always lying to you, and how to build measurement systems that actually drive improvement.</em></p><p></p><p><strong>&#8220;Good enough&#8221; is a bet that the problems you can&#8217;t see won&#8217;t become problems you can&#8217;t ignore. Most organizations lose that bet eventually.</strong></p><p><strong>Download: &#8220;Data Quality Cost Calculator&#8221; &#8212; A simple framework to estimate what data quality issues are costing your organization.</strong></p><p><strong><a href="https://forms.gle/bRnimMncEqYB18HDA">Download Data Quality Cost Calculator</a></strong></p><p></p><p>Founder of SANJEEVANI AI. ISO/IEC 42001 Lead Auditor. 25+ years in AI, data, and compliance across HealthTech, FinTech, EdTech, and Insurance. Building METRIS, the quantitative AI governance platform.</p>]]></content:encoded></item><item><title><![CDATA[Why Every AI Problem Is Actually a Data Problem]]></title><description><![CDATA[The Model Is 10% of the Work. The Data Is 90% of the Outcome.]]></description><link>https://ainstein.sanjeevaniai.com/p/why-every-ai-problem-is-actually</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/why-every-ai-problem-is-actually</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Mon, 22 Dec 2025 15:00:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!0QrQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0QrQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0QrQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png 424w, https://substackcdn.com/image/fetch/$s_!0QrQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png 848w, https://substackcdn.com/image/fetch/$s_!0QrQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!0QrQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0QrQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png" width="696" height="785.4545454545455" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1440,&quot;width&quot;:1276,&quot;resizeWidth&quot;:696,&quot;bytes&quot;:3054589,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/182046772?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0QrQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png 424w, https://substackcdn.com/image/fetch/$s_!0QrQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png 848w, https://substackcdn.com/image/fetch/$s_!0QrQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!0QrQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F19d28391-bd68-4a8b-a61a-8387567c29c1_1276x1440.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>S showed up with a new energy this week. Less confused, more frustrated.</p><p>&#8220;I&#8217;ve been reading,&#8221; S said, dropping into the chair across from me. &#8220;Papers. Blog posts. Case studies. Everyone&#8217;s talking about models. GPT-this, neural network-that, transformer-whatever.&#8221;</p><p>&#8220;And?&#8221;</p><p>&#8220;And I keep thinking about what we discussed. The Amazon hiring tool. The healthcare algorithm. The court system. In every case, the problem wasn&#8217;t really the model, was it?&#8221;</p><p>I smiled. S was getting there.</p><p>&#8220;The model worked,&#8221; S continued. &#8220;It did exactly what it was designed to do. It found patterns. It made predictions. It scaled decisions.&#8221;</p><p>&#8220;So what was the problem?&#8221;</p><p>S paused. Then, quietly: &#8220;The data.&#8221;</p><h3>The Dirty Secret of AI</h3><p>Here&#8217;s something the AI industry doesn&#8217;t advertise loudly: the model is the easy part.</p><p>Building a neural network? There are frameworks for that. Training an algorithm? There are tutorials. Deploying a prediction engine? Cloud providers will do it for you with three clicks.</p><p>But the data that feeds those models? That&#8217;s where everything breaks.</p><div class="pullquote"><p>A 2021 study by MIT and IBM found something remarkable: improving data quality by just 10% often delivered better model performance than any algorithmic improvement. </p></div><p>Not sometimes. Often.</p><p>Think about that. Companies spend millions on fancier models, more parameters, better architectures. And a methodical data cleanup would have worked better.</p><p>Andrew Ng, one of the most respected names in machine learning, has been saying this for years. He calls it &#8220;data-centric AI&#8221;. This is the recognition that for most real-world problems, the bottleneck isn&#8217;t the algorithm. It&#8217;s the data.</p><p>&#8220;So why does everyone focus on models?&#8221; S asked.</p><p>&#8220;Because models are sexy. Data cleaning is not.&#8221;</p><p>&#8220;That&#8217;s it? Vanity?&#8221;</p><p>&#8220;Partly. Also, models are contained. You can point to a model. You can benchmark it. You can publish a paper about it. Data is messy. Data is distributed across systems, teams, years of accumulated decisions. Data doesn&#8217;t fit on a leaderboard.&#8221;</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/subscribe?"><span>Subscribe now</span></a></p><h3>What &#8220;Data Quality&#8221; Actually Means</h3><p>S leaned forward. &#8220;Okay, so data matters more than models. But what does &#8216;good data&#8217; even mean? How do you measure it?&#8221;</p><p>This is where most conversations go wrong. People treat &#8220;data quality&#8221; as a vague aspiration, like &#8220;innovation&#8221; or &#8220;synergy.&#8221; Something everyone agrees is good but nobody defines.</p><p>In reality, data quality has specific, measurable dimensions. Six of them are widely recognized:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-lQ6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-lQ6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png 424w, https://substackcdn.com/image/fetch/$s_!-lQ6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png 848w, https://substackcdn.com/image/fetch/$s_!-lQ6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png 1272w, https://substackcdn.com/image/fetch/$s_!-lQ6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-lQ6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png" width="1408" height="736" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:736,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:462011,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/182046772?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-lQ6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png 424w, https://substackcdn.com/image/fetch/$s_!-lQ6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png 848w, https://substackcdn.com/image/fetch/$s_!-lQ6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png 1272w, https://substackcdn.com/image/fetch/$s_!-lQ6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa458abed-b3a5-4f36-b3c5-363b65dfb6b7_1408x736.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>1. Accuracy</strong> Does the data reflect reality? If your database says a customer lives in Chicago but they moved to Denver three years ago, that&#8217;s an accuracy problem. Seems obvious. But in large systems with millions of records, accuracy degrades constantly and&#8230; &#8220;silently&#8221;.</p><p><strong>2. Completeness</strong> Is the data present? Missing values aren&#8217;t just annoying. they&#8217;re informative in the wrong way. If 80% of your &#8220;income&#8221; field is blank, any model trained on that data will learn to work around the gaps. Sometimes in ways you don&#8217;t expect.</p><p><strong>3. Consistency</strong> Does the same thing mean the same thing everywhere? One system records dates as MM/DD/YYYY. Another uses DD/MM/YYYY. One team defines &#8220;active customer&#8221; as anyone who logged in this year. Another team means anyone who made a purchase. Same label. Different meanings. Model trained on this will learn confusion.</p><p><strong>4. Timeliness</strong> Is the data current? A credit risk model trained on 2019 data might be useless in 2024. The world changed. Customer behavior changed. The economy changed. But the model is still looking for patterns that no longer exist.</p><p><strong>5. Validity</strong> Does the data conform to defined formats and rules? An email field should contain emails. An age field shouldn&#8217;t contain negative numbers. A country code should match a real country. Validity violations often indicate upstream problems such as poorly designed forms, broken integrations, or manual entry errors.</p><p><strong>6. Uniqueness</strong> Is each record represented once? Duplicate records create phantom patterns. If the same customer appears twice with slightly different information, the model might learn that these are two different people with correlated behavior. They&#8217;re not. It&#8217;s just messy data.</p><p>&#8220;These aren&#8217;t abstract concepts,&#8221; I told S. &#8220;Each one is measurable. Each one has thresholds. Each one can be monitored.&#8221;</p><p>&#8220;But most companies don&#8217;t monitor them.&#8221;</p><p>&#8220;Most companies don&#8217;t even know these dimensions exist. They think of data quality as &#8216;does it load without errors.&#8217; That&#8217;s not quality. That&#8217;s just functioning.&#8221;</p><h3>The Real Cost of &#8220;Good Enough&#8221;</h3><p>S pulled out a notebook. I recognized this. It&#8217;s the signal that something was landing.</p><p>&#8220;Can you quantify this? Like, what does bad data actually cost?&#8221;</p><p>&#8220;Gartner estimated that poor data quality costs organizations an average of $12.9 million per year.&#8221;</p><p>&#8220;Average?&#8221;</p><p>&#8220;Average. Some industries are worse. Healthcare. Finance. Insurance. Anywhere decisions are high-stakes and data is distributed across legacy systems.&#8221;</p><p>But the headline number hides the more interesting story: where those costs come from.</p><p><strong>Direct costs:</strong></p><ul><li><p>Rework: Analysts spending 40-60% of their time cleaning data instead of analyzing it</p></li><li><p>Errors: Incorrect decisions based on incorrect data</p></li><li><p>Integration failures: Systems that can&#8217;t talk to each other because data formats don&#8217;t match</p></li></ul><p><strong>Indirect costs:</strong></p><ul><li><p>Missed opportunities: Patterns you can&#8217;t see because the data is too noisy</p></li><li><p>Model failures: Predictions that don&#8217;t generalize because training data didn&#8217;t represent reality</p></li><li><p>Compliance violations: Regulatory penalties when data governance requirements aren&#8217;t met</p></li></ul><p><strong>Hidden costs:</strong></p><ul><li><p>Eroded trust: Stakeholders who stop believing the numbers</p></li><li><p>Decision paralysis: Leaders who won&#8217;t act because they don&#8217;t trust the data</p></li><li><p>Technical debt: Workarounds that accumulate until the system becomes unmaintainable</p></li></ul><p>&#8220;The thing about data quality costs,&#8221; I said, &#8220;is that they&#8217;re distributed. No single failure looks catastrophic. It&#8217;s a thousand small problems, each one seemingly tolerable, adding up to systemic dysfunction.&#8221;</p><p>&#8220;Death by a thousand cuts.&#8221;</p><p>&#8220;Exactly. And because no single cut is fatal, nobody prioritizes the bandages.&#8221;</p><h3>The Lifecycle View</h3><p>S flipped to a new page. &#8220;So where does data quality break down? Is there a pattern?&#8221;</p><p>There is. And it follows the data lifecycle.</p><p><strong>At creation:</strong> Garbage enters the system. Bad forms, broken integrations, manual errors, unclear definitions.</p><p><strong>At storage:</strong> Quality degrades over time. Records go stale. Schemas drift. Corruption happens.</p><p><strong>At processing:</strong> Transformations introduce errors. Joins create duplicates. Business logic gets misapplied.</p><p><strong>At analysis:</strong> Models amplify noise. Bias compounds. Overfitting hides problems until production.</p><p><strong>At decision:</strong> Actions based on bad analysis create real-world harm. And often, that harm generates data that feeds back into the system &#8212; creating a loop of dysfunction.</p><p>&#8220;Every stage is a potential failure point,&#8221; I said. &#8220;And most organizations only check at one or two stages, usually the ones closest to the final report.&#8221;</p><p>&#8220;So they catch problems at the end, when it&#8217;s too late.&#8221;</p><p>&#8220;Or they don&#8217;t catch them at all. The decision looks confident. The numbers look precise. Nobody realizes the foundation is sand.&#8221;</p><h3>Why This Is an AI Governance Problem</h3><p>S sat back. &#8220;This feels like a data engineering conversation. What does it have to do with AI governance?&#8221;</p><p>&#8220;Everything.&#8221;</p><p>There&#8217;s a connection here that most people miss. When we talk about AI governance, people tend to think about models, algorithms, and outputs. But if you actually read the major AI governance frameworks (AI governance frameworks like the EU AI Act, NIST AI RMF, ISO/IEC 42001), you&#8217;ll find that data is at the center of all of them. This isn&#8217;t an afterthought. Data is treated as the foundation.</p><p>Take the EU AI Act as an example. Article 10 specifically requires that training, validation, and testing datasets be &#8220;relevant, representative, free of errors and complete.&#8221; This is not a suggestion or a best practice. It is a legal requirement for any AI system classified as high-risk. If your data doesn&#8217;t meet these standards, your AI system is not compliant. It&#8217;s that simple.</p><p>The NIST AI Risk Management Framework takes a similar approach. It dedicates an entire function to data-related concerns. The framework talks extensively about measuring data quality, documenting where data comes from, tracking how data changes over time, and monitoring for drift that could affect model performance.</p><p>ISO/IEC 42001, which is the international standard for AI management systems, requires organizations to demonstrate that they have proper data governance practices in place. You cannot get certified without showing that you understand and control the data flowing into your AI systems.</p><p>The message from regulators is clear. If you want to govern AI, you have to start by governing data.&#8220;So when regulators come asking about your AI system,&#8221; I said, &#8220;they&#8217;re not just asking about the model. They&#8217;re asking about the data. Where did it come from? How was it validated? How do you know it&#8217;s still representative?&#8221;</p><p>&#8220;And most companies can&#8217;t answer those questions.&#8221;</p><p>&#8220;Most companies have never been asked. That&#8217;s about to change.&#8221;</p><h3>The Uncomfortable Question</h3><p>S closed the notebook. Looked at me directly.</p><p>&#8220;Data quality is important. It&#8217;s measurable. It&#8217;s connected to everything downstream. So why isn&#8217;t it treated that way?&#8221;, She asked.</p><p>I&#8217;ve thought about this question for years. The answers are unsatisfying but honest:</p><p><strong>Organizational silos.</strong> Data is created by one team, stored by another, processed by a third, analyzed by a fourth. No one owns quality end-to-end.</p><p><strong>Misaligned incentives.</strong> Data engineers are measured on pipeline uptime, not data accuracy. Data scientists are measured on model performance, not data completeness. Executives are measured on outcomes, not the quality of inputs.</p><p><strong>Invisible failures.</strong> When a model makes a bad prediction because of bad data, the root cause is rarely traced back. The model gets blamed. The data stays broken.</p><p><strong>Short-term thinking.</strong> Fixing data quality is expensive upfront and pays off slowly. Shipping a model is fast and visible. Organizations optimize for what they can see.</p><p>&#8220;So the system is designed to ignore the problem,&#8221; S said.</p><p>&#8220;The system is designed to optimize for other things. Data quality is a casualty of those optimizations.&#8221;</p><p>&#8220;Until it isn&#8217;t. Until something breaks badly enough that people notice.&#8221;</p><p>&#8220;Right. And by then, the cost of fixing it is ten times what it would have been to do it right from the start.&#8221;</p><h3>What Changes This</h3><p>S asked the question I was waiting for: &#8220;So what would it actually take to treat data quality as foundational? Not as an afterthought?&#8221;</p><p>A few things:</p><p><strong>Measurement:</strong> The first change is to start measuring at all. There&#8217;s an old saying in management: you can&#8217;t manage what you don&#8217;t measure. This applies directly to data quality. Most organizations don&#8217;t measure data quality in any systematic way. They might check it once a year during an audit, or maybe once a quarter when a report is due. But that&#8217;s not real measurement. That&#8217;s occasional inspection. Data quality changes constantly. New records come in. Old records go stale. Systems get updated. People make errors. If you only look once a quarter, you&#8217;re flying blind for three months at a time. What you need is continuous monitoring. That means dashboards that track all six dimensions we discussed: accuracy, completeness, consistency, timeliness, validity, and uniqueness. It also means alerts that notify the right people when any of these dimensions fall below acceptable thresholds. Without this kind of ongoing measurement, you&#8217;re reacting to problems after they&#8217;ve already caused damage.</p><p><strong>Ownership:</strong> The second change is to assign clear ownership. Someone has to be accountable for data quality. And I don&#8217;t mean &#8220;the data team&#8221; as some vague abstraction. When everyone is responsible, no one is responsible. You need a specific person with a name and a title who owns data quality. This person needs real authority to make decisions about data standards, data processes, and data tools. They need a budget to invest in the infrastructure and people required to maintain quality. And there need to be real consequences when data quality fails on their watch. If nobody&#8217;s job depends on data quality, data quality will always lose to whatever has a deadline attached to it.</p><p><strong>Integration: </strong>The third change is to integrate quality checks into every stage of work. Data quality can&#8217;t be a separate workstream that happens before the &#8220;real work&#8221; begins. That approach treats quality as a one-time gate you pass through and forget about. Instead, data quality has to be embedded into every stage of the data lifecycle. When data enters a pipeline, there should be automated quality checks. When data feeds a model, there should be validation rules that flag problems before training begins. When data informs a decision, there should be confidence bounds that reflect the quality of the inputs. If the underlying data is shaky, the decision-maker needs to know that. Quality isn&#8217;t a phase. It&#8217;s a continuous practice that runs alongside everything else.</p><p><strong>Incentives: </strong>The fourth change is to align incentives. Until data quality becomes part of how people are evaluated, it will always be deprioritized. People focus on what they&#8217;re measured on. If a data engineer is measured only on pipeline uptime, they will optimize for uptime and ignore quality. If a data scientist is measured only on model accuracy, they will chase accuracy without questioning whether the training data was reliable. To fix this, you need to tie data quality metrics to the things people care about. Include quality scores in performance reviews. Make quality thresholds a requirement for project approvals. Block model deployments if input data doesn&#8217;t meet validation standards. When quality affects promotions, bonuses, and project success, people start paying attention to it.</p><p></p><p>&#8220;That&#8217;s a lot of organizational change,&#8221; S said.</p><p>&#8220;It is. And that&#8217;s why most organizations don&#8217;t do it. They&#8217;d rather deal with the symptoms than address the cause.&#8221;</p><p>&#8220;But the symptoms keep getting worse.&#8221;</p><p>&#8220;They do. And at some point, the cost of ignoring the problem exceeds the cost of fixing it. The smart organizations are figuring that out now. Before the regulators force them to.&#8221;</p><h3>Where This Goes Next</h3><p>S stood up to leave. Paused at the door.</p><p>&#8220;So we&#8217;ve established that AI is fundamentally a data problem. That data quality is measurable. That it breaks down at every stage of the lifecycle. That it&#8217;s a governance issue, not just a technical issue.&#8221;</p><p>&#8220;Yes.&#8221;</p><p>&#8220;What I still don&#8217;t understand is: how do you actually measure it? Not in theory. In practice. What are the metrics? What are the thresholds? How do you know when something is &#8216;good enough&#8217; versus when it&#8217;s a risk?&#8221;</p><p>&#8220;That&#8217;s the language of measurement,&#8221; I said. &#8220;And it&#8217;s more nuanced than most people realize.&#8221;</p><p>&#8220;Next week?&#8221; she asked. <br>&#8220;Yes. Next week,&#8221; I said.<br></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/p/why-every-ai-problem-is-actually?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/p/why-every-ai-problem-is-actually?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p><br><em>Next week: The difference between a metric that matters and a metric that misleads, and why single numbers are almost always lying to you.</em></p><p></p><p><strong>The data that trains your AI is the foundation everything else stands on. Cracks in that foundation don&#8217;t disappear. They propagate.</strong></p><p><strong>Download: &#8220;Data Quality: The 6 Dimensions&#8221; &#8212; A visual guide to what data quality actually means and the questions to ask at each stage of the lifecycle.</strong></p><p><strong><a href="https://forms.gle/9uPpEJLw1kQgyKvK8">Download Data Quality: The 6 Dimensions</a></strong><br><br>Founder of SANJEEVANI AI. ISO/IEC 42001 Lead Auditor. 25+ years in AI, data, and compliance across HealthTech, FinTech, EdTech, and Insurance. Building METRIS, the quantitative AI governance platform.</p>]]></content:encoded></item><item><title><![CDATA[Context ROT: The Hidden Crisis Destroying AI Governance From Within]]></title><description><![CDATA[How Redundant, Obsolete, and Trivial Data Multiplies Every Silo and Undermines Trust]]></description><link>https://ainstein.sanjeevaniai.com/p/context-rot-the-hidden-crisis-destroying</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/context-rot-the-hidden-crisis-destroying</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Sun, 21 Dec 2025 01:14:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!l2pF!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F195ee19e-3683-4fe4-8927-dad386e22f02_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A few days ago, I ran a poll asking the AI governance community a simple question: <strong>What creates the deepest data silos in your organization?</strong></p><p>The responses clustered around two familiar culprits: &#8220;People &amp; Processes&#8221; and &#8220;AI Systems.&#8221; Both received the lion&#8217;s share of votes, and for good reason. These are the visible, tangible causes we discuss in boardrooms and document in governance frameworks.</p><p>But here&#8217;s what the conversation is missing and what might be the most dangerous oversight in modern AI governance:</p><p><strong>Redundant, Obsolete, and Trivial (ROT) data is the silent multiplier that transforms manageable silos into governance nightmares.</strong></p><p>This isn&#8217;t just about storage costs or cleanup projects. This is about a fundamental crisis in how we think about data quality, lineage, and trustworthiness in AI systems.</p><h2>The Triple Threat: How Silos Actually Form</h2><p>Before we dive into ROT, let&#8217;s establish the foundation. Data silos in AI governance emerge from three interconnected sources:</p><h3>1. People &amp; Processes: The Distributed Truth Problem</h3><p>Teams create silos organically as they solve problems independently. Marketing builds their customer segmentation model using their version of customer data. Sales forecasts revenue using their CRM export. Product analytics trains churn prediction on their event stream.</p><p>Each team believes (genuinely believes) that their copy represents the source of truth.</p><p>The result? You don&#8217;t have one customer dataset. You have twelve. Each slightly different. Each with different update frequencies. Each with different data quality rules. And when you try to reconcile them, you discover:</p><ul><li><p>Customer #47392 exists in six systems with four different email addresses</p></li><li><p>Last purchase date varies by up to three weeks depending on which system you query</p></li><li><p>Lifetime value calculations differ by 40% across departments</p></li></ul><p>This isn&#8217;t malice. This is organizational entropy.</p><h3>2. AI Systems: The Technical Fragmentation</h3><p>Even when teams want to collaborate, AI systems actively work against them.</p><p>Modern ML pipelines fragment data across stages:</p><ul><li><p>Raw data in data lakes</p></li><li><p>Cleaned data in feature stores</p></li><li><p>Training datasets in experiment tracking systems</p></li><li><p>Validation sets in model registries</p></li><li><p>Production inference logs in monitoring platforms</p></li></ul><p>Each system speaks a different schema language. Each has different versioning conventions. Some track lineage; others don&#8217;t. The data scientist who trained your production model six months ago has since left the company, and nobody can definitively say which exact dataset version was used.</p><p>When model drift appears, you&#8217;re left investigating shadows.</p><h3>3. ROT Data: The Exponential Multiplier</h3><p>Now here&#8217;s where it gets truly problematic.</p><p>Both people-driven silos and system-driven fragmentation are difficult problems. But they&#8217;re manageable if you have clean, current, traceable data.</p><p><strong>ROT data makes both problems exponentially worse.</strong></p><h2>Understanding the ROT Phenomenon</h2><p>Let me break down what ROT actually means in an AI governance context:</p><p><strong>Redundant Data:</strong> Multiple copies of the same information across systems, teams, and time periods. Not just duplicates in the traditional sense, but semantically identical data stored in different formats, granularities, and contexts.</p><p><strong>Obsolete Data:</strong> Information that was once valuable but is no longer current, accurate, or relevant. This includes deprecated customer segments, outdated product taxonomies, superseded regulatory classifications, and retired feature definitions.</p><p><strong>Trivial Data:</strong> Low-value noise that clutters systems without contributing meaningful signal. Debug logs that never get analyzed. Test datasets that outlived their experiments. Temporary tables that became permanent fixtures.</p><p>The insidious thing about ROT? <strong>It compounds.</strong></p><p>Redundant data creates more opportunities for obsolescence. Obsolete data gets duplicated by automated processes. Trivial data accumulates in the gaps between systems. And before you know it, 60-80% of your AI governance surface area consists of data that actively undermines your objectives.</p><h2>Introducing Context ROT: When Insights Lose Meaning</h2><p>This brings us to what I believe is the core crisis: <strong>Context ROT.</strong></p><p>Context ROT occurs when the data supporting your insights, decisions, and governance metrics becomes so stale, duplicated, or untraceable that the insights themselves lose meaning.</p><p>Let me illustrate with a real scenario I encountered recently:</p><h3>Case Study: The Phantom Drift Incident</h3><p>A financial services company was monitoring their credit risk model using a sophisticated drift detection system. One Tuesday morning, the dashboard lit up red. That&#8217;s a significant feature drift detected in the &#8220;employment_sector&#8221; variable.</p><p>The model risk team investigated. They pulled the training distribution and compared it to the production distribution. Sure enough, massive shift. The &#8220;retail&#8221; sector had dropped from 18% to 3% of applications.</p><p>Alarm bells rang. Had the economy shifted that dramatically? Was this a data quality issue? Should they retrain the model?</p><p>After three days of investigation, they discovered the truth:</p><p>The &#8220;employment_sector&#8221; taxonomy had been updated 14 months earlier. The new taxonomy split &#8220;retail&#8221; into &#8220;retail_essential&#8221; and &#8220;retail_discretionary.&#8221; The production system was using the new taxonomy. The model monitoring system was comparing against training data that used the old taxonomy.</p><p>There was no drift. There was no model degradation. There was <strong>Context ROT.</strong></p><p>The governance dashboard was showing a critical alert based on data that no longer matched reality. The investigation consumed dozens of hours. Trust in the monitoring system eroded. And this wasn&#8217;t an isolated incident. It was one of dozens of similar issues.</p><h2>How ROT Undermines Core AI Governance Principles</h2><p>The impact of ROT extends far beyond false alarms. It systematically degrades every pillar of responsible AI:</p><h3>1. Model Trustworthiness</h3><p><strong>Question:</strong> Can you trust predictions built on stale data?</p><p>If your training dataset includes customer segments that were deprecated 18 months ago, and your model learned patterns from those outdated segments, what exactly is your model predicting?</p><p>When 40% of your training data is redundant (multiple copies of the same transactions from different source systems), you&#8217;re not training on a representative sample. You&#8217;re training on a distorted one where certain patterns are artificially amplified.</p><h3>2. Explainability</h3><p><strong>Question:</strong> How do you explain decisions when you can&#8217;t verify data lineage?</p><p>A regulator asks: &#8220;Why was this loan denied?&#8221;</p><p>Your model says: &#8220;Primary factor was debt-to-income ratio of 0.87&#8221;</p><p>The regulator asks: &#8220;Where did that ratio come from?&#8221;</p><p>You investigate and find three different debt-to-income calculations in your pipeline:</p><ul><li><p>One from the original application</p></li><li><p>One from a credit bureau</p></li><li><p>One from an internal risk model</p></li></ul><p>All three are stored. All three are slightly different. Which one did the model actually use? The lineage is broken because of redundant data storage without proper version control.</p><p><strong>You cannot explain a decision you cannot trace.</strong></p><h3>3. Risk Metrics and Compliance</h3><p><strong>Question:</strong> Are your compliance dashboards showing reality or illusion?</p><p>Your model fairness dashboard reports that your approval rates by demographic group are within acceptable bounds. Excellent.</p><p>Except the demographic data feeding that dashboard is 22 months old for 35% of customers. You&#8217;re measuring fairness against obsolete information. Your compliance metric is technically accurate but substantively meaningless.</p><p>This is Context ROT in its purest form.</p><h3>4. Audit Readiness</h3><p><strong>Question:</strong> Can you prove which data version was used for which decision?</p><p>An auditor requests evidence that a specific model decision from six months ago complied with regulations in effect at that time.</p><p>You need to show:</p><ul><li><p>The exact input data used</p></li><li><p>The model version deployed</p></li><li><p>The feature transformations applied</p></li><li><p>The regulatory rules in effect</p></li></ul><p>But your data lake contains seventeen versions of the customer profile table from that time period. Some are incremental snapshots. Some are full loads. Some have conflicting timestamps. Which one was actually used?</p><p>Without definitive answers, you&#8217;re exposed to regulatory risk not because you did anything wrong, but because you can&#8217;t prove you did it right.</p><h2>The Compounding Effect: Why ROT Makes Silos Worse</h2><p>Here&#8217;s the crucial insight that connects everything:</p><p><strong>ROT doesn&#8217;t just coexist with silos. It actively amplifies them.</strong></p><p>When Marketing&#8217;s customer dataset develops ROT, they don&#8217;t clean it up. They create a new, &#8220;cleaner&#8221; copy. Now you have two siloed datasets instead of one.</p><p>When Sales discovers obsolete account records, they don&#8217;t remove them. They add a flag to indicate &#8220;legacy&#8221; status. The obsolete data persists, confusing every downstream system.</p><p>When Product&#8217;s feature engineering pipeline generates trivial intermediate tables, those tables get backed up, replicated, and referenced by other teams who assume they&#8217;re important. The trivial becomes permanent.</p><p>Each silo becomes a breeding ground for ROT. Each ROT dataset spawns new silos. The cycle accelerates.</p><p>And the organization&#8217;s ability to govern AI degrades proportionally.</p><h2>Breaking the Cycle: A Framework for ROT Remediation</h2><p>Eliminating ROT isn&#8217;t a weekend cleanup project. It&#8217;s a fundamental shift in how organizations think about data as a governance asset.</p><p>Here&#8217;s a framework I&#8217;ve developed for organizations serious about addressing this:</p><h3>Phase 1: Discovery and Assessment</h3><p><strong>Map the Shadow Data Landscape</strong></p><p>Most organizations have no idea how much ROT exists in their systems. Start with:</p><ul><li><p><strong>Data lineage mapping:</strong> Where does each dataset come from? Where does it go? What transformations occur?</p></li><li><p><strong>Replication audit:</strong> Identify all copies of semantically similar data. Don&#8217;t just look for exact duplicates &#8212; look for different representations of the same business entity.</p></li><li><p><strong>Staleness analysis:</strong> When was each dataset last updated? When was it last <em>meaningfully</em> used (not just queried, but used in a decision)?</p></li><li><p><strong>Value assessment:</strong> What decisions depend on this data? If it disappeared tomorrow, what would break?</p></li></ul><p>This discovery phase often reveals shocking truths. I&#8217;ve worked with organizations where 70% of their &#8220;critical&#8221; data assets hadn&#8217;t been used in production for over a year.</p><h3>Phase 2: Categorization and Prioritization</h3><p>Not all ROT is created equal. Prioritize based on:</p><p><strong>High-Risk ROT:</strong></p><ul><li><p>Obsolete data still feeding production models</p></li><li><p>Redundant data creating inconsistent governance metrics</p></li><li><p>Trivial data consuming significant storage or processing resources</p></li></ul><p><strong>Medium-Risk ROT:</strong></p><ul><li><p>Deprecated datasets with unclear retirement dates</p></li><li><p>Duplicate data with conflicting quality rules</p></li><li><p>Low-value data in critical pipelines</p></li></ul><p><strong>Low-Risk ROT:</strong></p><ul><li><p>Archived data properly isolated from production</p></li><li><p>Development/test data clearly labeled</p></li><li><p>Historical data with valid retention justification</p></li></ul><p>Focus remediation efforts on high-risk ROT first.</p><h3>Phase 3: Implementation of Hygiene Infrastructure</h3><p>This is where most organizations fail. They treat ROT as a cleanup project rather than a continuous governance function.</p><p>Sustainable ROT remediation requires infrastructure:</p><p><strong>1. Automated Data Lifecycle Policies</strong></p><p>Define retention schedules based on data category and regulatory requirements. Implement automated archival and deletion workflows. Make the default state &#8220;expiring&#8221; rather than &#8220;permanent.&#8221;</p><p><strong>2. Version Control for Datasets</strong></p><p>Treat datasets like code. Every version should be tagged, traceable, and tied to specific model deployments. When you deprecate a dataset, the deprecation should be tracked as explicitly as the creation.</p><p><strong>3. Canonical Source Designation</strong></p><p>For every business entity (customer, product, transaction), designate one canonical source of truth. All other representations should be clearly marked as derived views with explicit lineage back to the canonical source.</p><p><strong>4. Duplication Prevention Controls</strong></p><p>Before creating a new dataset, require teams to search for existing similar datasets. Implement approval workflows for new data stores. Make it harder to duplicate than to reuse.</p><p><strong>5. Quality Gates in Data Pipelines</strong></p><p>Implement automated checks that flag:</p><ul><li><p>Data older than defined freshness thresholds</p></li><li><p>Duplicates across systems</p></li><li><p>Orphaned datasets with no downstream consumers</p></li><li><p>Schemas that don&#8217;t match registered standards</p></li></ul><h3>Phase 4: Cultural Transformation</h3><p>Technology alone won&#8217;t solve this. You need organizational buy-in.</p><p><strong>Data Ownership Accountability</strong></p><p>Assign explicit ownership to every dataset. The owner is accountable for:</p><ul><li><p>Maintaining data quality</p></li><li><p>Preventing unauthorized duplication</p></li><li><p>Executing timely deprecation</p></li><li><p>Documenting lineage and dependencies</p></li></ul><p><strong>Deprecation Workflows</strong></p><p>Make it socially acceptable &#8212; even celebrated &#8212; to sunset old data. Create clear processes for:</p><ul><li><p>Announcing deprecation timelines</p></li><li><p>Migrating downstream dependencies</p></li><li><p>Archiving for compliance</p></li><li><p>Verifying complete removal</p></li></ul><p><strong>Governance Metrics That Matter</strong></p><p>Track and report:</p><ul><li><p>Percentage of production models using deprecated data</p></li><li><p>Time-to-remediation for identified ROT</p></li><li><p>Storage costs attributed to trivial data</p></li><li><p>Lineage completeness scores</p></li></ul><p>Make ROT reduction a KPI for data teams, not just a best practice.</p><h2>The Strategic Imperative: ROT as Infrastructure</h2><p>Here&#8217;s my central thesis:</p><p><strong>Data hygiene is not operational maintenance. Data hygiene is governance infrastructure.</strong></p><p>You wouldn&#8217;t build a critical application without version control, testing, and deployment pipelines. You wouldn&#8217;t run production systems without monitoring, logging, and incident response.</p><p>Why would you build AI governance on data infrastructure that lacks basic hygiene controls?</p><p>Organizations that will lead in AI governance over the next decade aren&#8217;t the ones with perfect processes from day one. They&#8217;re the ones who recognize that:</p><ul><li><p><strong>ROT is inevitable</strong> &#8212; entropy always increases without active intervention</p></li><li><p><strong>ROT is measurable</strong> &#8212; you can quantify redundancy, obsolescence, and triviality</p></li><li><p><strong>ROT is addressable</strong> &#8212; with the right infrastructure and incentives</p></li><li><p><strong>ROT reduction is continuous</strong> &#8212; not a project, but a practice</p></li></ul><h2>Looking Forward: The Governance Advantage</h2><p>There&#8217;s a competitive dimension to this that&#8217;s worth acknowledging.</p><p>As AI regulation intensifies globally, from the EU AI Act to emerging US frameworks to sector-specific requirements, the ability to demonstrate clean data lineage, explainable decisions, and trustworthy metrics will become a significant advantage.</p><p>Organizations still struggling with ROT will find compliance prohibitively expensive. Audits will take months instead of weeks. Regulatory inquiries will expose fragility.</p><p>Organizations that have invested in data hygiene as governance infrastructure will move faster, prove compliance easier, and earn greater trust from stakeholders.</p><p><strong>The time to address ROT isn&#8217;t when the auditor asks for evidence. It&#8217;s now.</strong></p><h2>Conclusion: The Residue That Keeps Silos Alive</h2><p>Data silos don&#8217;t form from a single root cause. They compound across people, processes, and technology, creating an interconnected web of fragmentation.</p><p>And ROT is the residue that keeps them alive.</p><p>It&#8217;s the organizational plaque that hardens over time, making every governance initiative harder, every audit more painful, and every AI deployment riskier.</p><p>But unlike some governance challenges, ROT is solvable. It requires commitment, infrastructure, and cultural change but it&#8217;s fundamentally tractable.</p><p>The question is whether your organization will treat it as the critical governance issue it is, or continue relegating it to the backlog as &#8220;technical debt we&#8217;ll address someday.&#8221;</p><p>Because in AI governance, Context ROT isn&#8217;t just a data quality problem.</p><p><strong>It&#8217;s a trust crisis waiting to happen.</strong></p><p></p><p><em>What&#8217;s your experience with ROT in AI systems? Have you seen governance metrics distorted by stale, redundant, or trivial data? I&#8217;d love to hear your stories and strategies in the comments.</em></p>]]></content:encoded></item><item><title><![CDATA[The Silent Architecture of Trust in AI]]></title><description><![CDATA[Data Quality Isn't an IT Problem. It's Everything]]></description><link>https://ainstein.sanjeevaniai.com/p/the-silent-architecture-of-trust</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/the-silent-architecture-of-trust</guid><pubDate>Tue, 16 Dec 2025 15:01:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!wCpt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wCpt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wCpt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wCpt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wCpt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wCpt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wCpt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg" width="726" height="481" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:481,&quot;width&quot;:726,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:50439,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/178014738?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wCpt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wCpt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wCpt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wCpt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53370699-a2ed-4dd7-9a1f-7a315acce7ec_726x481.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>S called me again after last week&#8217;s article.</p><p>&#8220;Okay, I get it now. AI tunes itself to patterns in data. It doesn&#8217;t understand. It just finds the frequency that works. But what happens when the data is... wrong?&#8221;</p><p>That&#8217;s the question that keeps me up at night.</p><h3>The Hiring Tool That Worked Perfectly</h3><p>A company deployed an AI hiring tool. Fast, efficient, everyone impressed.</p><p>Then someone asked: &#8220;Why did it reject this candidate?&#8221;</p><p>They checked the system. The algorithm worked. The metrics looked accurate.</p><p>But when they traced backwards, they found the real issue: training data from 2019 that excluded entire demographics.</p><p>The AI didn&#8217;t fail. It tuned itself exactly to what it was given. It found a frequency and that frequency was biased.</p><p>This is what happens when the signal is wrong.</p><h2>Data Is Evidence, Not Truth</h2><p>Here&#8217;s the uncomfortable reality most teams don&#8217;t want to hear: data is not truth. It&#8217;s evidence. And evidence carries assumptions, blind spots, and biases.</p><p>Every row in a dataset represents a choice someone made about what to record and how. Those choices shape what the AI learns.</p><p>In AI systems, data plays two roles:</p><ul><li><p>It teaches the model how to see the world</p></li><li><p>It judges the model&#8217;s decisions later</p></li></ul><p>When both the teacher and the judge are flawed, the system has no way to know it&#8217;s wrong.</p><p>This is why governance must begin before the first line of code. Before the model. Before the architecture. At the data.</p><h2>The Anatomy of Bad Data</h2><p>Most bad data isn&#8217;t intentional. It&#8217;s small, everyday errors that quietly corrupt AI systems.</p><p><strong>Inaccurate Data</strong> A salary entered as $50000 instead of $50,000. Seems minor. But now every calculation downstream is wrong.</p><p><strong>Incomplete Data</strong> A patient&#8217;s record missing age or medical history. The model can&#8217;t learn accurate patterns from what isn&#8217;t there.</p><p><strong>Inconsistent Data</strong> &#8220;USA&#8221;, &#8220;U.S.&#8221;, and &#8220;United States&#8221; treated as three separate values. The system sees three different countries.</p><p><strong>Outdated Data</strong> Using 2018 customer trends to predict 2025 behavior. The world changed. The model didn&#8217;t.</p><p><strong>Biased Data</strong> A hiring dataset with mostly male candidates. The AI learns to favor what it saw most. Not because anyone told it to but because that&#8217;s what the data showed.</p><p>Bad data isn&#8217;t just incorrect. It&#8217;s unverified assumptions treated as fact.</p><h3>The Domino Effect</h3><p>Remember the FM tuning analogy? The system adjusts based on every example it sees.</p><p>Now imagine what happens when flawed data enters the pipeline:</p><ol><li><p>Feature engineering amplifies wrong signals</p></li><li><p>Model tunes to flawed patterns</p></li><li><p>Testing passes because the test data has the same flaws</p></li><li><p>Deployed model reinforces errors in production</p></li><li><p>Retraining locks the bias in as &#8220;truth&#8221;</p></li></ol><p>Each step makes it harder to find the original problem. By you notice something&#8217;s wrong, the error has multiplied through the entire system.</p><p>This isn&#8217;t sabotage. It&#8217;s systemic inertia. And it happens more often than anyone wants to admit.</p><h3>The Real Cost</h3><p>McKinsey reports bad data costs enterprises 15&#8211;20% of revenue.</p><p>But the real cost isn&#8217;t money. It&#8217;s justice.</p><p>Predictive policing systems that target the wrong neighborhoods&#8212;because historical arrest data reflected biased enforcement, not actual crime.</p><p>Loan models that reject qualified applicants because missing fields correlated with demographics the system learned to penalize.</p><p>Healthcare algorithms that recommended less care for certain patients because cost data was used as a proxy for health needs.</p><p>These weren&#8217;t evil intentions. They were patterns in data, learned at scale, deployed without adequate governance.</p><p>AI governance isn&#8217;t slowing innovation. It&#8217;s protecting people from systems that don&#8217;t know they&#8217;re causing harm.</p><h3>Governance: The Immune System</h3><p>Think of governance as your AI&#8217;s immune system. It detects, prevents, and corrects damage before it reaches people.</p><p>Several frameworks guide this:</p><ul><li><p><strong>ISO/IEC 5259</strong> &#8212; Data quality standards</p></li><li><p><strong>ISO/IEC 42001</strong> &#8212; AI Management Systems</p></li><li><p><strong>NIST AI RMF</strong> &#8212; Govern, Map, Measure, Manage</p></li></ul><p>At its core, governance answers one question: &#8220;What data caused this decision, and who is responsible?&#8221;</p><p>If you can&#8217;t answer that, you don&#8217;t have governance. You have hope.</p><h2>Data as Ethical Material</h2><p>This might be the most important shift in thinking: data isn&#8217;t just technical material. It&#8217;s ethical material.</p><p>Every row in your dataset represents a human life: their job application, their loan request, their medical history, their interactions with your product.</p><p>When you train an AI on that data, you&#8217;re encoding decisions that will affect people like them. At scale. Automatically. Without human review.</p><p>Clean data isn&#8217;t a feature. It&#8217;s a responsibility.</p><h3>The Loop That Never Ends</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PGhx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PGhx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png 424w, https://substackcdn.com/image/fetch/$s_!PGhx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png 848w, https://substackcdn.com/image/fetch/$s_!PGhx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png 1272w, https://substackcdn.com/image/fetch/$s_!PGhx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PGhx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png" width="370" height="344" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:688,&quot;width&quot;:740,&quot;resizeWidth&quot;:370,&quot;bytes&quot;:217501,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/178014738?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PGhx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png 424w, https://substackcdn.com/image/fetch/$s_!PGhx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png 848w, https://substackcdn.com/image/fetch/$s_!PGhx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png 1272w, https://substackcdn.com/image/fetch/$s_!PGhx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52871097-2016-40ec-b4c8-823c0a465cc6_740x688.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Data &#8594; Model &#8594; Decision &#8594; Feedback &#8594; Retraining &#8594; Data</p><p>This loop runs continuously in every AI system. New data comes in. Models update. Decisions change.</p><p>Governance keeps this loop honest. It asks:</p><ul><li><p>Where did this data come from?</p></li><li><p>What assumptions does it carry?</p></li><li><p>Who does it represent and who does it miss?</p></li><li><p>What decisions will it drive?</p></li><li><p>Who&#8217;s accountable when something goes wrong?</p></li></ul><p>Without governance, the loop runs blind. And blind systems cause harm they can&#8217;t see.</p><h2>Key Takeaways</h2><p><strong>For Builders:</strong></p><ul><li><p>Audit data before you architect models</p></li><li><p>Track data lineage like you track code</p></li><li><p>Build feedback loops for bias and drift</p></li></ul><p><strong>For Leaders:</strong></p><ul><li><p>Data quality is a governance issue, not an IT task</p></li><li><p>Invest in data stewardship roles</p></li><li><p>Ask &#8220;Where did this data come from?&#8221; before &#8220;What can this model do?&#8221;</p></li></ul><p><strong>For Everyone:</strong></p><ul><li><p>The AI systems affecting your life are only as good as their training data</p></li><li><p>When something feels wrong, it might be a data problem nobody caught</p></li><li><p>Demand transparency, not just explanations</p></li></ul><h3>A Question to Sit With</h3><p>If data is the DNA of AI, what kind of organism are we creating?</p><h4>Next Week</h4><p>We&#8217;ve covered what AI is, how it learns, and why data is the foundation of everything.</p><p>Next week, we go deeper: specific case studies of AI failures that made headlines. Amazon&#8217;s hiring AI. Healthcare algorithms. Credit systems. What went wrong, why nobody caught it, and what governance would have changed.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><strong>Subscribe to A.I.N.S.T.E.I.N.</strong> so you don&#8217;t miss it.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>Suneeta Modekurty</em> <br><em>Founder, A.I.N.S.T.E.I.N.</em></p>]]></content:encoded></item><item><title><![CDATA[🚨 FLASH BRIEFING: Trump's AI Executive Order]]></title><description><![CDATA[Your risk exposure just went UP, not down. Here's what to do.]]></description><link>https://ainstein.sanjeevaniai.com/p/flash-briefing-trumps-ai-executive</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/flash-briefing-trumps-ai-executive</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Fri, 12 Dec 2025 18:34:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!hPun!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hPun!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hPun!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!hPun!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!hPun!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!hPun!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hPun!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1759598,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/181430605?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hPun!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!hPun!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!hPun!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!hPun!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082eef3b-7e3a-43c3-ac3d-1953c1e4c270_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>The Headlines vs. The Reality</h3><p>What happened:  President Trump signed an executive order creating an &#8220;AI Litigation Task Force&#8221; to challenge state-level AI regulations through lawsuits and funding cuts.</p><p>What the headlines say: &#8220;Federal government takes over AI oversight&#8221;</p><p>What&#8217;s actually happening:  We just entered a period of &#8216;maximum regulatory uncertainty&#8217; and your risk exposure probably increased, not decreased.</p><h3> Three Things Most Analysis Is Missing</h3><h5> 1. The EU AI Act Doesn&#8217;t Care About US Politics</h5><p>August 2026 is still coming. If you sell to European customers, process EU citizen data, or have European operations, nothing changed for you today.</p><p>In fact, companies betting on US deregulation while ignoring EU requirements are setting themselves up for a painful collision with &#8364;35 million fines.</p><h5> 2. State Laws Don&#8217;t Disappear Overnight</h5><p>The executive order creates a process for challenging state laws, not an immediate preemption. Colorado&#8217;s SB 205, California&#8217;s CPRA AI provisions, Illinois&#8217; BIPA, NYC&#8217;s Local Law 144&#8230; these remain enforceable until successfully challenged in court.</p><p>Timeline reality check:</p><p>- Commerce Department evaluation: 90 days minimum</p><p>- Litigation process: 12-36 months per state</p><p>- Appeals: Add another 12-24 months</p><p>Translation: You&#8217;re operating under existing state laws for at least 2-3 more years in many jurisdictions.</p><h5>3. Uncertainty Is the Real Risk</h5><p>The worst position? Neither fully compliant with current state laws nor prepared for whatever federal framework emerges.</p><p>Companies now face:</p><p>- Enforcement risk from states racing to act before preemption</p><p>- Litigation risk  from the new federal Task Force</p><p>- Reputational risk  from policy whiplash</p><p>- International risk  from diverging US/EU approaches</p><h4>What Smart Companies Are Doing Right Now</h4><p><strong>This week:</strong></p><p>- Audit current state-law exposure (which laws apply to your AI systems today?)</p><p>- Document your compliance posture (if enforcement accelerates, you want evidence of good faith)</p><p>- Brief leadership on the transition period risks</p><p>This month:</p><p>- Prioritize EU AI Act readiness (the clearest, most enforceable framework)</p><p>- Model financial exposure under multiple regulatory scenarios</p><p>- Identify which compliance investments remain valuable regardless of US policy direction</p><p><strong>This quarter:</strong></p><p>- Build governance infrastructure that transcends any single framework</p><p>- Establish audit trails and documentation practices</p><p>- Develop board-ready risk quantification</p><p><strong>Need help auditing your current exposure?</strong> Reply to this email with &#8220;AUDIT&#8221; and I&#8217;ll send you our AI Governance Quick Assessment checklist.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/p/flash-briefing-trumps-ai-executive?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/p/flash-briefing-trumps-ai-executive?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p><h3> The Quantified View</h3><p>Here&#8217;s what our risk models show for a mid-sized company with AI systems operating across US and EU markets:</p><p>Scenario                                                12-Month Risk Exposure</p><p>Pre-executive order baseline            $1.2M - $3.8M         </p><p>Transition period (current)                $1.8M - $5.2M         </p><p>Full federal preemption (2027+)        $0.9M - $2.1M         </p><p>EU-only compliance focus                 $2.4M - $7.1M         </p><p>The counterintuitive finding: Risk exposure &#8216;increases&#8217; during the transition period due to enforcement uncertainty, potential state &#8220;enforcement sprints,&#8221; and the complexity of tracking evolving requirements.</p><p>Want to see your organization&#8217;s specific risk numbers? </p><p><a href="https://calendar.google.com/calendar/u/0/appointments/schedules/AcZssZ06VxOY4gLq9NgwfsWOl1RQsuKsE6Ts17tzGuz7uRfa_Ztcislf51HDTQ8tGCEcGNneQZ3h0C-m">Request a METRIS Assessment &#8594;</a></p><h3> The Bottom Line</h3><p>This executive order is a gift to companies who&#8217;ve been dragging their feet on governance, if they use the transition period wisely.</p><p>It&#8217;s a warning to companies who interpret &#8220;deregulation&#8221; as &#8220;do nothing.&#8221;</p><p>And it&#8217;s irrelevant to your EU AI Act obligations.</p><p>The winners in 2026 won&#8217;t be the companies that guessed correctly about US policy direction. They&#8217;ll be the ones who built governance infrastructure robust enough to adapt to whatever comes.</p><h3>What We&#8217;re Watching</h3><p>- Commerce Department&#8217;s 90-day evaluation (due mid-March 2026)</p><p>- State attorney general responses (expect California and New York to push back hard)</p><p>- EU reaction to US regulatory divergence</p><p>- Insurance and investor community signals</p><p>We&#8217;ll send updates as these develop.<br><br></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/subscribe?"><span>Subscribe now</span></a></p><p></p><p></p><p>About This Brief</p><p>I&#8217;m Suneeta Modekurty, Founder of SANJEEVANI  AI and creator of METRIS, a quantitative AI governance platform that transforms compliance assessments into financial risk projections. When regulations shift, our models help you understand what it actually means for your bottom line.<br><br>Three ways to stay ahead:</p><ol><li><p><strong>Follow me on <a href="https://www.linkedin.com/in/smodekurty">LinkedIn</a></strong> for real-time analysis as this evolves</p></li><li><p><strong>Download the technical paper</strong> on Zenodo (DOI: 10.5281/zenodo.17850617)</p></li><li><p><strong>Reply with your biggest compliance question</strong>, I answer every email</p></li></ol><p></p><p></p><p>&#169; 2025 SANJEEVANI AI LLC | [sanjeevaniai.com](https://sanjeevaniai.com)</p><p>You received this special edition because you&#8217;re on our AI Governance Intelligence list. Regular editions continue as usual, every week Tuesday 9:00 am CST</p>]]></content:encoded></item><item><title><![CDATA[How AI Actually Learns ]]></title><description><![CDATA[And Why It Means Data Is Everything]]></description><link>https://ainstein.sanjeevaniai.com/p/how-ai-actually-learns</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/how-ai-actually-learns</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Tue, 09 Dec 2025 15:01:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VYmZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VYmZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VYmZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png 424w, https://substackcdn.com/image/fetch/$s_!VYmZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png 848w, https://substackcdn.com/image/fetch/$s_!VYmZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png 1272w, https://substackcdn.com/image/fetch/$s_!VYmZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VYmZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png" width="1130" height="1358" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1358,&quot;width&quot;:1130,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2343988,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/180301086?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VYmZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png 424w, https://substackcdn.com/image/fetch/$s_!VYmZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png 848w, https://substackcdn.com/image/fetch/$s_!VYmZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png 1272w, https://substackcdn.com/image/fetch/$s_!VYmZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a2c11aa-3339-4928-8a5d-a5cd73df92d7_1130x1358.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>My friend (S) called me after reading last week&#8217;s article.</p><p>&#8220;Okay, I get that AI learns from examples instead of following rules. But <em>how</em> does that actually work? How does showing a computer millions of pictures teach it anything?&#8221;</p><p>I remember asking the same question years ago. I had a statistics background, understood regression, could build models. But the idea that a system could learn patterns without explicit programming took time to sink in.</p><p>Let me explain it the way I wish someone had explained it to me.</p><h3>The Training Loop</h3><p>Training an AI system is simpler than most people think.</p><p>You show it an example. It makes a guess. You tell it if the guess was right or wrong. It adjusts slightly. You repeat this millions of times.</p><p>That&#8217;s it. Show, guess, adjust, repeat</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mJjl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mJjl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png 424w, https://substackcdn.com/image/fetch/$s_!mJjl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png 848w, https://substackcdn.com/image/fetch/$s_!mJjl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png 1272w, https://substackcdn.com/image/fetch/$s_!mJjl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mJjl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png" width="338" height="308.5553539019964" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1006,&quot;width&quot;:1102,&quot;resizeWidth&quot;:338,&quot;bytes&quot;:180353,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/180301086?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mJjl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png 424w, https://substackcdn.com/image/fetch/$s_!mJjl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png 848w, https://substackcdn.com/image/fetch/$s_!mJjl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png 1272w, https://substackcdn.com/image/fetch/$s_!mJjl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F511d4da8-dfcf-4eed-9a4c-d791f2bd04f2_1102x1006.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Take spam detection. You show the system an email. It guesses the email is spam or not spam. You tell it the correct answer. If it guessed wrong, you fine-tune it like tuning an FM radio to get the station you want. You&#8217;re adjusting the frequency until the signal comes through clear.</p><p>Each training example is another turn of the dial. A million examples means a million tiny adjustments. Eventually, the system locks onto the right frequency. Finally, it starts recognizing patterns accurately.</p><p>Engineers call this &#8220;<strong>supervised learning</strong>.&#8221; I call it i<strong>ndustrial-scale pattern tuning</strong>.</p><h3>How to find the Right Frequency</h3><p>Think of it like this.</p><p>When you tune a radio, you&#8217;re searching through noise to find a clear signal. Too far left, static. Too far right, static. You keep adjusting until the music comes through.</p><p>AI training works the same way. The system starts in noise with random guesses, no real pattern recognition. Each training example helps it tune toward the signal. &#8220;Wrong answer? oh oh, adjust left. Still wrong, it&#8217;s ok, adjust right. Closer. Closer. There&#8230;. We Gooo!&#8221;</p><p>After millions of these micro-adjustments, the system finds the frequency where patterns become clear. It can now distinguish spam from legitimate email, fraud from normal transactions, disease from healthy tissue.</p><p>The thing is, nobody programs these frequencies directly. The system finds them through repetition. Show enough examples, give enough feedback, and it tunes itself.</p><p>This is powerful. It&#8217;s also why we can&#8217;t always explain exactly what the system learned. It found a frequency that works, but describing that frequency in human terms is harder than it sounds. And it feeds heavily on data to be able to do it. </p><h2>Why It Needs So Much Data</h2><p>S asked the follow-up: &#8220;Why millions of examples? Can&#8217;t it learn faster?&#8221;</p><p>The world is messier than training sets assume.</p><p>Most dogs are easy to recognize. But what about a chihuahua that looks like a large rat? A mop that resembles a sheepdog? A wolf that could pass for a husky? A blurry photo taken at night?</p><p>If your system only saw a hundred dogs during training, it never encountered these edge cases. It tuned itself to recognize &#8220;typical&#8221; dogs and goes static on anything unusual.</p><blockquote><p>Volume matters because <strong>edge cases</strong> matter. More data means more weird situations to tune against. Fewer surprises when the system hits the real world.</p></blockquote><p>Ok, let&#8217;s talk about what you&#8217;ve experienced this yourself, even if you didn&#8217;t know why:</p><ul><li><p>Your phone&#8217;s autocorrect keeps &#8220;fixing&#8221; your name or a word you use often to something wrong. The system was tuned on data that didn&#8217;t include your name or your industry&#8217;s jargon.</p></li><li><p>You travel for work, use your credit card in a new city, and it gets blocked for &#8220;suspicious activity.&#8221; The fraud detection system wasn&#8217;t tuned for your edge case. So, a sudden location change by a legitimate user.</p></li><li><p>Your voice assistant works fine for your colleague but constantly mishears you. It was tuned on voices that sound like your colleague. Your accent, speech pattern, or tone wasn&#8217;t well-represented in the training data.</p></li></ul><p>These aren&#8217;t bugs. The systems are working exactly as tuned. They just weren&#8217;t tuned on enough edge cases to handle <em>you</em> in <em>that</em> situation.</p><p>I&#8217;ve seen this repeatedly. A system works perfectly in testing, then falls apart in production because the training data was too clean, too narrow, too optimistic about what reality actually looks like. The test environment was a quiet studio recording. Reality is a live concert with feedback and crowd noise.</p><p><em>Subscribe for one weekly, practical, jargon-free breakdown of how AI really works</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe&quot;,&quot;text&quot;:&quot;Subscribe for Weekly AI Insights&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/subscribe"><span>Subscribe for Weekly AI Insights</span></a></p><h3>Here&#8217;s the Secret</h3><p>AI doesn&#8217;t &#8220;understand&#8221; anything.</p><p>When a system recognizes a cat, it&#8217;s not thinking &#8220;whiskers, pointy ears, therefore cat.&#8221; It found a frequency where certain pixel patterns correlate with the label &#8220;cat.&#8221; The &#8216;<em>why&#8217;</em> doesn&#8217;t exist for the system. Only the correlation does.</p><p>This is both the power and the danger.</p><p><strong>The power:</strong> AI can tune into patterns humans would never detect. It can process more examples than any person could review in a lifetime. It operates continuously without fatigue.</p><p><strong>The danger:</strong> AI has no common sense. It doesn&#8217;t know that a drawing of a cat isn&#8217;t a real cat. It doesn&#8217;t understand context. It just matches patterns, including patterns that shouldn&#8217;t matter.</p><p>I&#8217;ve sat in meetings where executives said &#8220;our AI understands customer intent.&#8221; No. It&#8217;s <em>tuned</em> to predict customer intent based on historical patterns. That&#8217;s different. Understanding implies reasoning, context, judgment. AI doesn&#8217;t reason. It correlates at the frequency it was trained on.</p><h3>AI&#8217;s Strengths and Blind Spots</h3><p>After years working with AI, I&#8217;ve developed a rough rule.</p><p><strong>AI tunes well for:</strong></p><ul><li><p>Pattern recognition at scale (images, text, transactions)</p></li><li><p>Finding anomalies in large datasets</p></li><li><p>Predicting outcomes from historical data</p></li><li><p>Generating content similar to training examples</p></li></ul><p><strong>AI can&#8217;t tune for:</strong></p><ul><li><p>Reasoning about cause and effect</p></li><li><p>Common sense and context</p></li><li><p>Situations that never appeared in training</p></li><li><p>Explaining why it made a decision</p></li></ul><p>There&#8217;s an observation from robotics researcher Hans Moravec: what&#8217;s easy for humans is hard for AI, and vice versa.</p><p>A five-year-old can walk across a messy room, recognize grandma from any angle, and hold a conversation. These &#8220;simple&#8221; tasks are incredibly hard for AI and no amount of tuning gets you common sense.</p><p>But that same five-year-old can&#8217;t analyze a million transactions for fraud or review every medical study published in the last decade. AI tunes for that easily. Humans can&#8217;t.</p><p>AI is a tool with specific strengths. Expecting it to replace human judgment misunderstands what tuning actually achieves.</p><p><em>Join a growing community of leaders, founders, and practitioners learning how to build AI systems responsibly.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe&quot;,&quot;text&quot;:&quot;Join the Community&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/subscribe"><span>Join the Community</span></a></p><h2>What This Means for Governance</h2><p>Now, let&#8217;s connect the dots:</p><ul><li><p>AI learns by tuning itself to patterns in data</p></li><li><p>It doesn&#8217;t understand those patterns. It replicates them</p></li><li><p>Whatever patterns exist in your data become the frequency your AI broadcasts on</p></li></ul><p>What if those patterns include bias? What if the data is outdated? What if the data is incomplete? or is unrepresentative of the people you&#8217;re serving?</p><p>The AI doesn&#8217;t know. It can&#8217;t know. It tunes to whatever you fed it and treats those patterns as the correct signal.</p><p>This is why AI governance doesn&#8217;t start with the model. It starts with data.</p><p>A perfectly tuned system trained on the wrong data broadcasts the wrong signal very confidently, at scale, with no awareness that anything is off. Does this mean the system is broken? No, the system isn&#8217;t broken. It&#8217;s tuned exactly to what it was given.</p><p><em>If you&#8217;re responsible for AI, Data, Risk, or Compliance-get my weekly governance insights.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/subscribe&quot;,&quot;text&quot;:&quot;Get AI Governance Insights&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/subscribe"><span>Get AI Governance Insights</span></a></p><p></p><h2>The Questions You Should Be Asking</h2><p>Before you trust any AI system, ask:</p><ul><li><p>What data was it tuned on?</p></li><li><p>How old is that data?</p></li><li><p>Who&#8217;s represented in the training set and who&#8217;s missing?</p></li><li><p>Has it been tested on data it never saw during training?</p></li><li><p>What happens when it encounters something completely new?</p></li><li><p>What are the known failure modes?</p></li></ul><p>These aren&#8217;t technical questions. They&#8217;re governance questions. And in my experience auditing systems across EdTech, HealthTech, FinTech, and Insurance, most organizations can&#8217;t answer them.</p><p>They deploy the system. They trust the vendor. They hope for the best.</p><blockquote><p>Hope is not a governance strategy.</p></blockquote><h2>Next Week</h2><p>Now that you understand how AI learns, and how it tunes itself to patterns in data, we&#8217;re ready to look at what happens when that data is wrong.</p><p>Next week: <strong>The Silent Architecture of Trust</strong>, why data quality isn&#8217;t an IT problem, it&#8217;s the foundation of whether your AI works at all. We&#8217;ll look at what bad data actually looks like, how small errors compound into system-wide failures, and what governance means in practice.</p><p>Your AI is only as good as the signal it was tuned on. Next week, we examine that signal.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/p/how-ai-actually-learns?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Knowledge spreads trust, your share helps more teams build responsible, well-governed AI.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/p/how-ai-actually-learns?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/p/how-ai-actually-learns?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p><em>Suneeta Modekurty</em><br><em>Founder, A.I.N.S.T.E.I.N.</em></p>]]></content:encoded></item><item><title><![CDATA[What Actually Is AI? (And Why Your Company Needs to Understand It)]]></title><description><![CDATA[No jargon. No math. Just clear answers to the questions you've been afraid to ask.]]></description><link>https://ainstein.sanjeevaniai.com/p/what-actually-is-ai-and-why-your</link><guid isPermaLink="false">https://ainstein.sanjeevaniai.com/p/what-actually-is-ai-and-why-your</guid><dc:creator><![CDATA[A.I.N.S.T.E.I.N.]]></dc:creator><pubDate>Tue, 02 Dec 2025 15:03:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Q5Px!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q5Px!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q5Px!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png 424w, https://substackcdn.com/image/fetch/$s_!Q5Px!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png 848w, https://substackcdn.com/image/fetch/$s_!Q5Px!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png 1272w, https://substackcdn.com/image/fetch/$s_!Q5Px!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q5Px!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png" width="1456" height="1002" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1002,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3531189,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ainstein.sanjeevaniai.com/i/179532420?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Q5Px!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png 424w, https://substackcdn.com/image/fetch/$s_!Q5Px!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png 848w, https://substackcdn.com/image/fetch/$s_!Q5Px!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png 1272w, https://substackcdn.com/image/fetch/$s_!Q5Px!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98f56c29-2cb0-4a95-9a67-d3f2a54f134f_1810x1246.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Let me tell you about a conversation that inspired this article.</strong></p><p>Yesterday, a friend called me. She&#8217;s been working at a Fortune 500 financial services company for ten years. She is very smart, strategic, manages critical projects that affect thousands of customers. We were catching up when she said something that stopped me cold: She said, &#8220;Suneeta, I need to be honest with you. I don&#8217;t really understand what AI is. I use ChatGPT for emails, we have some AI thing in our CRM, and my boss keeps asking about &#8216;AI strategy.&#8217; But if someone asked me to actually explain what&#8217;s happening... I couldn&#8217;t. And neither could my boss.&#8221;</p><p>She paused. &#8220;Is that bad?&#8221;</p><p>Here&#8217;s what I told her: No, it&#8217;s not bad. It&#8217;s normal. And she&#8217;s not alone.</p><p>Most executives I talk to are in the same position. They&#8217;re using AI tools daily. Their companies are making million-dollar decisions about AI systems. But they couldn&#8217;t explain what&#8217;s actually happening under the hood if their job depended on it.</p><p>And increasingly, their jobs might.</p><p>This isn&#8217;t a knowledge gap you can afford to ignore anymore. Not because AI is replacing your job tomorrow (it&#8217;s not). But because your company is already deploying AI systems that make important decisions about customers, about employees, about money, about risk. And if you don&#8217;t understand what AI actually is, you can&#8217;t ask the right questions about whether those systems are working properly.</p><p>So let&#8217;s fix that right now. No, we are using no jargon, nor any math. Just clear answers to the questions you&#8217;ve been afraid to ask.</p><h2><strong>What AI Actually Is (The Simple Truth)</strong></h2><p>Here&#8217;s the definition that matters:</p><blockquote><p><strong>AI is software that learns patterns from examples, rather than following explicit rules that we write.</strong></p></blockquote><p>That&#8217;s it. That&#8217;s the core difference between AI and every other type of software you&#8217;ve ever used.</p><p>Let me show you what I mean.</p><h3><strong>Traditional Software: Rules We Write</strong></h3><p>Traditional software does exactly what we tell it to do, step by step. A programmer writes explicit instructions:</p><p><em>&#8220;If the email subject line contains the word &#8216;invoice,&#8217; move it to the Accounting folder.&#8221;</em></p><p><em>&#8220;If the transaction amount exceeds $10,000, flag it for review.&#8221;</em></p><p><em>&#8220;If the customer&#8217;s zip code starts with &#8216;902,&#8217; calculate shipping as $15.99.&#8221;</em></p><p>These are rules. Clear, explicit, unambiguous. The software doesn&#8217;t think. It doesn&#8217;t learn. It doesn&#8217;t adapt. It just follows our instructions, perfectly, every single time.</p><p>This works beautifully for tasks where the rules are known, stable, and can be written down explicitly. Calculating taxes. Processing payroll. Routing phone calls. Millions of business processes run this way, reliably, every day.</p><h3><strong>AI: Patterns It Discovers</strong></h3><p>AI works fundamentally differently. Instead of giving it rules, we give it examples. Lots of examples. Then we ask it to figure out the patterns.</p><p><em>&#8220;Here are 10,000 emails that humans labeled as &#8216;accounting-related.&#8217; Figure out what makes something accounting-related.&#8221;</em></p><p><em>&#8220;Here are 50,000 transactions where fraud investigators marked which ones were fraudulent. Figure out what fraud looks like.&#8221;</em></p><p><em>&#8220;Here are 100,000 customer service conversations that humans rated as &#8216;resolved satisfactorily.&#8217; Figure out what makes customers satisfied.&#8221;</em></p><p>The AI system analyzes these examples, identifies patterns. Some are obvious, some are subtle, some that no human would have thought to look for and learns rules from the data itself. Rules we didn&#8217;t write. Rules we might not even understand. Rules that work... until they don&#8217;t.</p><p>This is powerful. It lets us automate tasks that are too complex, too nuanced, or too context-dependent for anyone to write explicit rules. It&#8217;s why AI can recognize faces in photos, transcribe speech with accents, recommend products you didn&#8217;t know you wanted, and increasingly, make decisions that used to require human judgment.</p><h3><strong>A Clearer Analogy</strong></h3><p>Think about teaching a child to recognize dogs.</p><p>You don&#8217;t give them explicit rules: <em>&#8220;Dogs have four legs, fur, a tail, and bark.&#8221;</em> That&#8217;s the traditional software approach. It would fail immediately. What about three-legged dogs? Hairless dogs? Silent dogs? Dogs that look like wolves?</p><p>Instead, you show them dogs. Many dogs. Different breeds, sizes, colors. You point and say &#8220;dog.&#8221; Eventually, through exposure to examples, they learn what &#8220;dog-ness&#8221; is. They can recognize dogs they&#8217;ve never seen before, even unusual ones, because they&#8217;ve learned the pattern.</p><p>That&#8217;s how AI works. Show it enough examples of something, and it learns to recognize that thing even in situations it&#8217;s never encountered before.</p><p>But here&#8217;s the critical part that almost everyone misses: <strong>AI can only learn what&#8217;s in the examples you show it.</strong></p><p>If you only show a child golden retrievers, they might not recognize a chihuahua as a dog. If you only show them dogs in daylight, they might struggle at night. If the examples you provide are limited, biased, or flawed, the patterns they learn will be limited, biased, or flawed.</p><p>This brings us to the most important concept in understanding AI.</p><h2><strong>Why Data Is the Food AI Eats (And Quality Determines Everything)</strong></h2><p>My friend asked me: &#8220;But if AI is so smart, can&#8217;t it figure out when the data is wrong?&#8221;</p><p>No. It can&#8217;t. And this is where most AI problems begin.</p><h3><strong>AI Has No Independent Knowledge</strong></h3><p>AI doesn&#8217;t &#8220;know&#8221; anything except what its training data taught it. It has no common sense. No life experience. No ability to say &#8220;wait, this seems wrong.&#8221;</p><p>If you train an AI on biased data, it learns the bias. If you train it on incomplete data, it learns incomplete patterns. If you train it on old data, it learns outdated rules. And it applies those patterns with perfect, unwavering confidence.</p><blockquote><p><strong>You are what you eat. AI is what it learns from.</strong></p></blockquote><p>This isn&#8217;t metaphorical, it&#8217;s literal. The quality of your AI system is determined almost entirely by the quality of its training data. Not by the sophistication of its algorithm. Not by how expensive it was. Not by the reputation of the vendor.</p><p><strong>By. The. Data.</strong></p><h3><strong>What Makes Data &#8220;Good&#8221;?</strong></h3><p>Good data has three essential characteristics:</p><p><strong>1. Representative:</strong> It includes diverse examples that reflect the real world the AI will operate in.</p><p>If you&#8217;re building a medical diagnosis AI, your training data needs patients of different ages, races, genders, and geographies. If you only train on data from wealthy urban hospitals, it will fail in rural clinics. If you only train on data from healthy young adults, it will fail with elderly patients.</p><p><strong>2. Accurate:</strong> The examples are labeled correctly and measured properly.</p><p>If humans mislabeled things during training marking non-spam as spam, categorizing inquiries incorrectly, applying inconsistent standards, the AI learns those errors as truth. Garbage in, garbage amplified out.</p><p><strong>3. Contextual:</strong> It preserves the circumstances under which data was collected.</p><p>Data collected during a pandemic might not reflect normal behavior. Data from one region might not transfer to another. Data from 2020 might not predict 2025. Context matters, and when we strip it away, patterns become misleading.</p><h3><strong>What Makes Data &#8220;Bad&#8221;?</strong></h3><p>Bad data comes in predictable forms:</p><p><strong>Biased:</strong> Systematically over-represents some groups and under-represents others. Your hiring AI trained on historical data where 90% of engineers were male? It just learned that &#8220;good engineer&#8221; correlates with &#8220;male.&#8221;</p><p><strong>Incomplete:</strong> Missing critical information that humans need to make good decisions. Patient records without documented allergies. Credit applications without income verification. Resume databases without actual job performance data.</p><p><strong>Outdated:</strong> Reflects how things used to work, not how they work now. Consumer behavior from 2019 doesn&#8217;t predict consumer behavior post-pandemic. Market dynamics from stable periods don&#8217;t predict crisis periods.</p><p><strong>Inaccurate:</strong> Simply wrong. Typos. Measurement errors. System glitches. The decimal point that shifted. The sensor that drifted. The form field that accepted &#8220;N/A&#8221; as a zip code.</p><p>Each of these flaws gets encoded into the AI&#8217;s understanding of the world. And then it makes millions of decisions based on that flawed understanding.</p><h2><strong>When Bad Data Becomes Bad Decisions: Real Stories</strong></h2><p>Let me show you what this looks like in practice. These aren&#8217;t hypothetical scenarios. These are real companies that spent millions on AI systems, deployed them confidently, and watched them fail in ways that made headlines.</p><h3><strong>Amazon&#8217;s Hiring AI: When History Becomes Destiny</strong></h3><p>In 2018, Reuters reported that Amazon had been developing an AI system to automate resume screening.[<a href="http://Dastin, Jeffrey. &#8220;Amazon scraps secret AI recruiting tool that showed bias against women.&#8221; Reuters, October 10, 2018. https://www.reuters.com/article/us-amazon-com-jobs-automation-insight-idUSKCN1MK08G &#8617;">1</a>] <br>The goal was noble: reduce bias in hiring by removing human subjectivity. Let algorithms evaluate candidates objectively, based purely on qualifications.</p><p>The AI was trained on ten years of resumes submitted to Amazon. They were resumes of people who were hired, promoted, succeeded. The system learned patterns from this data: what successful Amazon employees looked like on paper.</p><p>There was just one problem: For most of that decade, Amazon&#8217;s technical workforce was predominantly male. Not because men were objectively better engineers, but because tech industry hiring had historically skewed male.</p><p>The AI learned this as a pattern. It observed that in the historical data, &#8220;successful technical employee&#8221; strongly correlated with &#8220;male.&#8221; So it began penalizing resumes that indicated female gender. It downgraded graduates of women&#8217;s colleges. It downgraded resumes containing the word &#8220;women&#8217;s&#8221; as in &#8220;women&#8217;s chess club captain.&#8221;</p><p>The algorithm wasn&#8217;t broken. It was doing exactly what it was trained to do: find patterns in historical hiring data and apply them. The problem was that historical hiring patterns included bias, and the AI faithfully learned and amplified it.</p><p>Amazon shut the system down. But not before it revealed an uncomfortable truth:</p><blockquote><p><strong>AI doesn&#8217;t eliminate human bias. It automates it. At scale. With mathematical precision.</strong></p></blockquote><h3><strong>Apple Card: When Algorithms Can&#8217;t Explain Themselves</strong></h3><p>In 2019, a tech entrepreneur named David Heinemeier Hansson tweeted that Apple Card had given him 20 times the credit limit his wife received, despite her having a higher credit score.[<a href="http://Telford, Taylor. &#8220;Apple Card algorithm sparks gender bias allegations against Goldman Sachs.&#8221; The Washington Post, November 11, 2019. https://www.washingtonpost.com/business/2019/11/11/apple-card-algorithm-sparks-gender-bias-allegations-against-goldman-sachs/ &#8617;">2</a>] <br>Other couples reported similar experiences. The pattern was clear and troubling. Apple and Goldman Sachs (the bank behind Apple Card) insisted there was no gender discrimination in their algorithm. Their AI evaluated creditworthiness based on objective factors, they said. Not gender.</p><p>They were probably telling the truth. The algorithm likely didn&#8217;t use gender as an input variable at all.</p><p>But here&#8217;s what happens with AI: Even when you don&#8217;t directly use protected characteristics like gender or race, AI can learn to use proxy variables that correlate with those characteristics. Zip code. Shopping patterns. Transaction history. The AI finds patterns that happen to correlate with gender, even if it never sees the gender field.</p><p>The New York Department of Financial Services launched an investigation. The challenge wasn&#8217;t proving the algorithm was biased and the outcomes spoke for themselves. The challenge was getting anyone to explain <em>why</em> the algorithm made the decisions it did.</p><p>This revealed another uncomfortable truth: <strong>When AI makes decisions that seem wrong, often nobody can explain why. Not even the people who built it.</strong></p><h3><strong>Healthcare AI: The Cost of Unrepresentative Data</strong></h3><p>A 2019 study published in <em>Science</em> revealed that a healthcare algorithm used by hospitals across the United States was systematically discriminating against Black patients.[<a href="http://Obermeyer, Ziad, et al. &#8220;Dissecting racial bias in an algorithm used to manage the health of populations.&#8221; Science, Vol. 366, Issue 6464, October 25, 2019, pp. 447-453. https://www.science.org/doi/10.1126/science.aax2342 &#8617;">3</a>] <br>The algorithm helped doctors decide which patients needed extra medical care, affecting millions of people annually.</p><p>The problem? The algorithm used healthcare costs as a proxy for healthcare needs. It assumed that patients who cost the system more money were sicker and needed more care.</p><p>But Black patients, on average, had lower healthcare costs not because they were healthier, but because they had less access to care due to systemic barriers. The AI learned that &#8220;lower cost = healthier&#8221; and recommended less care for Black patients who were actually just as sick as white patients with higher costs.</p><p>The algorithm worked perfectly from a technical standpoint. It accurately predicted what it was trained to predict: costs. But costs weren&#8217;t the right thing to measure. The data reflected historical inequity, and the AI perpetuated it.</p><h3><strong>The Pattern Behind the Failures</strong></h3><p>Look closely at these stories and you&#8217;ll see the same structure:</p><ol><li><p><strong>Well-intentioned deployment:</strong> Nobody set out to build discriminatory systems</p></li><li><p><strong>Training on historical data:</strong> AI learned from how things were done before</p></li><li><p><strong>Successful pattern matching:</strong> The AI correctly identified patterns in that data</p></li><li><p><strong>Problematic real-world outcomes:</strong> Those historical patterns included historical biases</p></li><li><p><strong>Inability to explain or fix quickly:</strong> By the time problems surfaced, the AI was already deployed</p></li></ol><p>This pattern repeats across industries. Predictive policing tools that over-target certain neighborhoods (trained on biased arrest patterns). Loan approval systems that reject qualified applicants from certain zip codes (trained on discriminatory lending history). And countless other cases that never make headlines.</p><p><strong>The AI worked perfectly. The data was the problem.<br></strong></p><h2><strong>Why This Matters to Your Business Right Now</strong></h2><p>After I shared the Amazon and Apple Card stories with my friend, she said: &#8220;Okay, but we&#8217;re not building hiring algorithms or credit systems. We&#8217;re just a regular company.&#8221;</p><p>Then I asked her about their customer service chatbot. Their fraud detection system. Their inventory forecasting tool.</p><p>&#8220;Oh,&#8221; she said. &#8220;Those count as AI?&#8221;</p><p>Yes. They do.</p><h3><strong>You&#8217;re Already Using AI (Whether You Realize It or not..)</strong></h3><blockquote><p>Most organizations can&#8217;t produce a complete list of the AI systems they&#8217;re currently using.</p></blockquote><p>Let me help you find them:</p><p><strong>In your customer service:</strong> That chatbot on your website? AI. The email routing system that decides which department handles which inquiry? Probably AI. The &#8220;recommended articles&#8221; in your knowledge base? AI.</p><p><strong>In your finance department:</strong> Fraud detection systems? AI. Expense report anomaly detection? AI. Cash flow forecasting tools? Increasingly AI.</p><p><strong>In your HR systems:</strong> Resume screening tools? AI. Interview scheduling systems that &#8220;optimize&#8221; candidate selection? AI. Performance prediction models? AI.</p><p><strong>In your marketing:</strong> Email subject line optimization? AI. Ad targeting? AI. Product recommendations? AI. Dynamic pricing? AI.</p><p><strong>In your operations:</strong> Inventory forecasting? AI. Logistics optimization? AI. Predictive maintenance alerts? AI.</p><p>These systems are making decisions. Some small, some significant. Some are just recommending options to humans. Others are operating autonomously, making hundreds or thousands of decisions daily.</p><p>And here&#8217;s what should concern you: <strong>Most companies don&#8217;t know what data these systems were trained on, whether that data was any good, or whether the systems are still working as intended.</strong></p><h3><strong>The Risks Are Growing</strong></h3><p>Three forces are converging to make AI governance urgent:</p><p><strong>1. Regulatory Pressure:</strong> The EU AI Act is now in force, with penalties up to &#8364;35 million or 7% of global revenue for violations.[<a href="https://www.consilium.europa.eu/en/press/press-releases/2024/05/21/artificial-intelligence-ai-act-council-gives-final-green-light-to-the-first-worldwide-rules-on-ai/ &#8617;">4</a>] <br>US states are passing their own AI laws. You will be required to explain how your AI systems make decisions and prove they&#8217;re not discriminatory.</p><p><strong>2. Legal Liability:</strong> Companies are being sued for discriminatory AI decisions in hiring, lending, insurance, and housing. Courts are no longer accepting &#8220;the algorithm did it&#8221; as a defense.</p><p><strong>3. Reputational Risk:</strong> In an age of social media, an AI failure can become a PR crisis in hours. Amazon, Apple Card, and others learned this the expensive way.</p><h3><strong>But Also: The Opportunity</strong></h3><p>Here&#8217;s what most companies miss: <strong>Good AI governance isn&#8217;t just risk mitigation. It&#8217;s competitive advantage.</strong></p><p>Companies that can prove their AI systems are trustworthy with evidence and not just promises are winning contracts. They&#8217;re getting better insurance rates. They&#8217;re attracting customers who don&#8217;t trust competitors. They&#8217;re raising capital more easily because investors can see quantified risk management.</p><p>The question isn&#8217;t whether to govern AI. It&#8217;s whether you&#8217;ll do it before your competitors do.</p><h2><strong>Questions You Should Ask Your Team This Week</strong></h2><p>You don&#8217;t need to become a data scientist to govern AI effectively. But you do need to start asking better questions. Here are five that every executive should pose to their teams:</p><p><strong>1. &#8220;What AI systems are we currently using?&#8221;</strong></p><p>Not &#8220;do we use AI&#8221; because you do. The question is <em>where</em>. Ask for a complete inventory: every tool, system, or platform that uses machine learning, algorithms, or automation to make decisions. Most companies are shocked by how long this list becomes.</p><p><strong>2. &#8220;What decisions do these systems make autonomously vs. recommend to humans?&#8221;</strong></p><p>There&#8217;s a difference between &#8220;AI suggests, human decides&#8221; and &#8220;AI decides, human rubber-stamps.&#8221; Know which is which. The autonomous ones deserve much more scrutiny.</p><p><strong>3. &#8220;What data were they trained on, and when was that data collected?&#8221;</strong></p><p>If the answer is &#8220;we don&#8217;t know&#8221; or &#8220;the vendor won&#8217;t say&#8221;, that&#8217;s a red flag. You&#8217;re making decisions based on patterns learned from data you can&#8217;t verify. If the data is more than 2-3 years old, the patterns might be obsolete.</p><p><strong>4. &#8220;Who&#8217;s monitoring whether these systems still work correctly?&#8221;</strong></p><p>AI doesn&#8217;t break like traditional software. It <em>drifts</em>. Performance degrades gradually as the world changes. If nobody&#8217;s actively monitoring accuracy, fairness, and reliability, you&#8217;re flying blind.</p><p><strong>5. &#8220;What happens when these systems are wrong, and who&#8217;s accountable?&#8221;</strong></p><p>Have you actually defined what &#8220;wrong&#8221; means for each system? What&#8217;s your process when someone disputes an AI decision? Who reviews? Who has authority to override? If these answers aren&#8217;t documented, you don&#8217;t have governance.</p><p>These questions won&#8217;t give you complete answers immediately. That&#8217;s fine. The goal is to start the conversation, to make AI systems visible and accountable rather than invisible and assumed.</p><h2><strong>What&#8217;s Coming in This Series</strong></h2><p>This is the first of many conversations we&#8217;re going to have about AI, data, and governance. Over the coming weeks, I&#8217;ll break down everything you need to understand. You may not  become a technical expert, but you need to be an informed decision-maker.</p><p><strong>Next week</strong>, we&#8217;ll go deeper into how AI actually learns, the training process, why it needs so much data, and what happens inside these systems when they&#8217;re &#8220;learning patterns.&#8221; You&#8217;ll understand why some AI tasks are easy and others remain impossibly hard.</p><p><strong>In the weeks that follow</strong>, we&#8217;ll explore:</p><ul><li><p>Why data quality determines everything (and how to measure it)</p></li><li><p>Real stories of AI failures and what they teach us</p></li><li><p>The AI systems already running in your business (and how to find them)</p></li><li><p>What &#8220;AI governance&#8221; actually means in practice</p></li><li><p>Why traditional compliance checklists don&#8217;t work</p></li><li><p>How leading companies are measuring AI trustworthiness quantitatively</p></li></ul><p>This isn&#8217;t about fear-mongering. AI is neither savior nor threat. It&#8217;s a tool. An enormously powerful tool. One that can amplify both human wisdom and human error at unprecedented scale.</p><p>The question isn&#8217;t whether to use AI. That ship has sailed. The question is whether you&#8217;ll use it responsibly, whether you&#8217;ll understand it well enough to govern it effectively, and whether you&#8217;ll build systems your stakeholders can actually trust.</p><p>My friend asked if not understanding AI was bad. I said no, at least not yet. But it becomes a problem when you&#8217;re responsible for systems you can&#8217;t explain. When your company is using AI to make decisions about people&#8217;s jobs, money, or opportunities, and you can&#8217;t answer basic questions about how those decisions get made.</p><p>That&#8217;s a choice you can no longer afford to make.</p><p></p><p><strong>Don&#8217;t miss next week&#8217;s deep dive into how AI actually learns.<br><br></strong><em>Know a colleague making AI decisions without asking these questions? Share this with them before they learn the hard way.<br></em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ainstein.sanjeevaniai.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share A.I.N.S.T.E.I.N&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ainstein.sanjeevaniai.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share A.I.N.S.T.E.I.N</span></a></p><p></p><h3><strong>References</strong></h3><ol><li><p>Dastin, Jeffrey. &#8220;Amazon scraps secret AI recruiting tool that showed bias against women.&#8221; <em>Reuters</em>, October 10, 2018. <a href="https://www.reuters.com/article/us-amazon-com-jobs-automation-insight-idUSKCN1MK08G">https://www.reuters.com/article/us-amazon-com-jobs-automation-insight-idUSKCN1MK08G</a> <a href="#user-content-fnref-1">&#8617;</a></p></li><li><p>Telford, Taylor. &#8220;Apple Card algorithm sparks gender bias allegations against Goldman Sachs.&#8221; <em>The Washington Post</em>, November 11, 2019. <a href="https://www.washingtonpost.com/business/2019/11/11/apple-card-algorithm-sparks-gender-bias-allegations-against-goldman-sachs/">https://www.washingtonpost.com/business/2019/11/11/apple-card-algorithm-sparks-gender-bias-allegations-against-goldman-sachs/</a> <a href="#user-content-fnref-2">&#8617;</a></p></li><li><p>Obermeyer, Ziad, et al. &#8220;Dissecting racial bias in an algorithm used to manage the health of populations.&#8221; <em>Science</em>, Vol. 366, Issue 6464, October 25, 2019, pp. 447-453. <a href="https://www.science.org/doi/10.1126/science.aax2342">https://www.science.org/doi/10.1126/science.aax2342</a> <a href="#user-content-fnref-3">&#8617;</a></p></li><li><p>European Commission. &#8220;Artificial Intelligence Act: Council gives final green light to the first worldwide rules on Artificial Intelligence.&#8221; Press release, May 21, 2024. <a href="https://www.consilium.europa.eu/en/press/press-releases/2024/05/21/artificial-intelligence-ai-act-council-gives-final-green-light-to-the-first-worldwide-rules-on-ai/">https://www.consilium.europa.eu/en/press/press-releases/2024/05/21/artificial-intelligence-ai-act-council-gives-final-green-light-to-the-first-worldwide-rules-on-ai/</a> <a href="#user-content-fnref-4">&#8617;</a></p></li></ol><p></p><p></p><p></p><h3><strong>About the Author</strong></h3><p>Suneeta Modekurty is an ISO/IEC 42001 AIMS Lead Auditor and O-1A visa holder for extraordinary ability in data science. She has 25 years of experience building AI systems across EdTech, HealthTech, FinTech, and Insurance. Connect on <a href="link">LinkedIn</a>.</p><p>&#169; 2025 SANJEEVANI AI</p>]]></content:encoded></item></channel></rss>