<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Nik Bear Brown - Computational Skepticism: Education and AI]]></title><description><![CDATA[Education and AI]]></description><link>https://www.skepticism.ai/s/education-and-ai</link><image><url>https://substackcdn.com/image/fetch/$s_!ea9u!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73f2e8c8-c907-4319-a9cb-14cda74f5128_800x800.png</url><title>Nik Bear Brown - Computational Skepticism: Education and AI</title><link>https://www.skepticism.ai/s/education-and-ai</link></image><generator>Substack</generator><lastBuildDate>Thu, 30 Apr 2026 08:58:19 GMT</lastBuildDate><atom:link href="https://www.skepticism.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Bear Brown, LLC]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[nikbearbrown@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[nikbearbrown@substack.com]]></itunes:email><itunes:name><![CDATA[Nik Bear Brown]]></itunes:name></itunes:owner><itunes:author><![CDATA[Nik Bear Brown]]></itunes:author><googleplay:owner><![CDATA[nikbearbrown@substack.com]]></googleplay:owner><googleplay:email><![CDATA[nikbearbrown@substack.com]]></googleplay:email><googleplay:author><![CDATA[Nik Bear Brown]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The Measurement That Wasn't There]]></title><description><![CDATA[On the quiet fraud at the center of AI education research &#8212; and why it's harder to catch than the kind that gets retracted]]></description><link>https://www.skepticism.ai/p/the-measurement-that-wasnt-there</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-measurement-that-wasnt-there</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Wed, 29 Apr 2026 19:15:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GKsY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GKsY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GKsY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png 424w, https://substackcdn.com/image/fetch/$s_!GKsY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png 848w, https://substackcdn.com/image/fetch/$s_!GKsY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png 1272w, https://substackcdn.com/image/fetch/$s_!GKsY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GKsY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png" width="1456" height="590" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:590,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:793291,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/195907789?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GKsY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png 424w, https://substackcdn.com/image/fetch/$s_!GKsY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png 848w, https://substackcdn.com/image/fetch/$s_!GKsY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png 1272w, https://substackcdn.com/image/fetch/$s_!GKsY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82b791a3-8561-4bc4-b32f-cd2e70a7b897_2780x1126.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There is a paper circulating in AI education circles as a counterpoint to the skeptics. Wang and Zhang, published in February 2026 in the <em>International Journal of Educational Technology in Higher Education</em>, a Springer Nature journal. It passed peer review. It has four studies. It has 912 participants across three continents. It deploys PLS-SEM and fsQCA and IPMA, and it has a methodology flowchart with seven stages, and it uses the word &#8220;paradoxical&#8221; in its title and delivers on the promise &#8212; two hypotheses come back significant in the wrong direction, which the authors then claim as the actual discovery.</p><p>I want to be honest about what I am about to argue. The Wang and Fan retraction that prompted this conversation is a case of bad causal evidence overclaimed. That is one problem. Wang and Zhang is a different problem. It is methodologically elaborate work that is not actually measuring what it claims to measure. In some ways it is harder to catch, because the machinery is impressive and the numbers are clean and the peer reviewers, like the rest of us, have been trained to evaluate internal consistency rather than construct validity.</p><p>Strip away the machinery. Here is what Wang and Zhang actually did.</p><p>Nine hundred and twelve business students filled out a questionnaire. The questionnaire asked them to rate their agreement with statements like: &#8220;My interaction with the generative AI has led me to question my long-held assumptions.&#8221; And: &#8220;Using generative AI has fundamentally changed the way I understand certain subjects.&#8221; And: &#8220;My use of generative AI has prompted a deep re-evaluation of my ways of thinking.&#8221;</p><p>Those five items, averaged together, are the outcome variable. The paper calls this outcome &#8220;transformative learning experience.&#8221;</p><p>It is not transformative learning experience. It is self-reported perception of transformative learning experience. The difference is not semantic. It is the entire study.</p><div><hr></div><p>Jack Mezirow&#8217;s transformative learning theory &#8212; the anchor the paper correctly treats as its theoretical foundation &#8212; describes a slow, disorienting, often unconscious process of perspective reconstruction. Mezirow was not describing a feeling students could report after two weeks. He was describing something that happens to people over months or years, something they often cannot name while it is occurring, something that shows up in changed behavior and revised assumptions and different relationships to knowledge &#8212; not in survey responses. The theory Mezirow actually wrote is about the kind of learning that happens when a person discovers that the framework they have been using to understand the world is inadequate. That does not feel like an insight. It feels like vertigo.</p><p>Measuring this with five Likert items is not a methodological shortcut. It is a category error. You might as well measure altitude with a thermometer and then report, with SRMR = 0.031, that higher temperatures correlate with being closer to the sky.</p><p>The paper knows this, in the way that papers of this type always know what they are doing, which is to say: it is in the limitations section. &#8220;Generalizability is bounded by exclusive reliance on self-reported perceptions,&#8221; the authors write, and then proceed to spend eight thousand words drawing inferences about transformative learning from self-reported perceptions. The limitation is disclosed and then ignored. This is the standard operation.</p><div><hr></div><p>Now add the demand characteristics.</p><p>I said &#8220;convenience sampling from business schools,&#8221; and that is the phrase papers in this area use. What it usually means in practice is that the 912 participants are the researchers&#8217; own students, or the students of colleagues at institutions where the researchers have relationships. The paper does not specify. It describes &#8220;multistage purposive sampling&#8221; and leaves the details of how institutions were contacted and how students were recruited conspicuously absent. But here is what we know: the qualitative component &#8212; the 45 interviews providing &#8220;rich process-oriented insights&#8221; &#8212; was drawn &#8220;exclusively from the Chinese sample,&#8221; and one of the authors is at a Chinese university. We know the students knew they were participating in an academic study. We know, from two thousand years of social psychology, that students who are aware of being studied by people who may have access to their grades tend to report what they believe is the expected or approved answer.</p><p>The paper deploys a temporal separation of two weeks between waves to &#8220;minimize common method bias.&#8221; Two weeks between surveys does not eliminate the problem of students reporting what they believe the study wants to hear. It separates the questions. It does not change who is answering them or why.</p><div><hr></div><p>I want to name the third problem, which is the one I raised in the group and which I think is the most structurally interesting.</p><p>Almost every learning environment is a massive violation of SUTVA &#8212; the Stable Unit Treatment Value Assumption. SUTVA says that the treatment received by one unit doesn&#8217;t affect the outcomes for another. In a classroom, this is almost never true. Students talk to each other. They share AI tools. They discuss assignments. They copy strategies. One student&#8217;s approach to using ChatGPT influences other students&#8217; approaches, which influences their outcomes, which shows up in the data as independent observations that are not independent at all.</p><p>In a networked environment where 912 business students across three continents are all using the same publicly available AI tools, the assumption that each student&#8217;s &#8220;transformative learning experience&#8221; is a function solely of their individual &#8220;pedagogical partnership orientation&#8221; and &#8220;cognitive vigilance&#8221; and &#8220;efficiency orientation&#8221; is not a simplifying assumption. It is an assumption that, if violated &#8212; and it is almost certainly violated &#8212; means the causal model is wrong in ways the statistical machinery cannot detect. PLS-SEM with excellent fit statistics can sit on top of fundamentally confounded data and produce clean-looking path coefficients. The cleanliness of the output is not evidence of the validity of the model. It is evidence that the model fits the data it was given.</p><p>True causal inference in learning environments would require experimental variation, not survey waves. It would require controlling for the social transmission of strategies and norms. It would require outcome measures that are behavioral, not perceptual. Absent these, what you have is a very sophisticated correlation study that has dressed itself in the language of mechanism.</p><div><hr></div><p>The paper is not a fraud in the sense of fabricated data. The numbers are probably exactly what the authors say they are. The students probably filled out exactly the surveys the authors describe. The analysis was probably executed correctly in SmartPLS 4.1.</p><p>The problem is upstream of all of that. The problem is in the question &#8220;what did we measure?&#8221;</p><p>We measured whether students who reported viewing AI as a collaborative partner also reported having their assumptions challenged. We found that they did. We called this &#8220;transformative learning.&#8221; We built a four-study architecture around this finding, with fsQCA and IPMA and 45 interviews and cross-cultural multi-group analysis, and we used the word &#8220;revolutionizes&#8221; in the discussion section, and we were published in a Springer Nature journal.</p><p>This is the second problem the field has, and it is subtler than the retracted meta-analysis. The retracted Wang and Fan paper is the kind of failure that produces retractions: fabricated or manipulated data, statistical impossibilities, evidence that the numbers were not real. That is a catastrophic failure, but it is detectable. It triggers the mechanisms the field has built for self-correction.</p><p>The Wang and Zhang problem does not trigger those mechanisms. The numbers are real. The peer review process evaluated internal consistency and found it satisfactory. The methodology flowchart has seven stages. The HTMT ratios are all below 0.85. The paper did exactly what the field rewarded it for doing.</p><p>And what it measured was: how students feel about whether they learned something.</p><div><hr></div><p>Here is what I think is actually going on in that data, if you want my honest read of it.</p><p>Students who frame AI as a collaborative partner rather than a tool are probably more engaged with the learning process in general. Engagement is positively correlated with self-reported learning. This is not a surprise. It is not a paradox. It is not evidence that &#8220;partnership orientation simultaneously activates cognitive vigilance and cognitive offloading through synergistic cognitive collaboration.&#8221; It is evidence that students who are paying attention think they learned more.</p><p>The finding that cognitive offloading is positively associated with self-reported transformative learning is interesting &#8212; the paper hypothesized the opposite and got a significant result in the other direction, and that is worth noting. But the post-hoc explanation (that offloading liberates cognitive resources for higher-order reflection) is plausible, not demonstrated. The paper discovered an unexpected correlation, generated a theory to explain it, and presented the theory as established. The U-shaped analyses that appear to confirm the theory were conducted after the unexpected finding was observed, without correction for exploratory inflation. This is the standard operation, and it is why most published findings in social science do not replicate.</p><p>The correct statement of the finding is: among 912 business students who self-report using AI, those who self-report viewing AI as a partner also self-report greater subjective sense of perspective change, and this association holds when we control for several other self-reported constructs. This is an interesting starting point for a research program. It is not a demonstration that pedagogical AI partnerships cause transformative learning.</p><div><hr></div><p>I want to be fair to the authors and to the field. They are working in an area where longitudinal behavioral research is genuinely hard to conduct, where IRB constraints limit what can be measured, where publication timelines create pressure toward the kind of efficiency the paper&#8217;s own subjects were reporting, and where the methodological standards for what counts as evidence have been established over decades of work that made the same choices at every turn. They did what the field taught them to do. The peer reviewers evaluated the paper against the standards of the field and found it acceptable by those standards.</p><p>That is the problem. Not this paper. The standards.</p><p>What would adequate evidence look like? It would measure transformative learning through behavioral change over meaningful time periods &#8212; different academic choices, different engagement with contradictory evidence, different patterns of intellectual behavior &#8212; not through survey items administered two weeks after measuring the predictors. It would use experimental variation in AI access or framing. It would account for social transmission between students. It would treat the gap between self-reported perception and actual cognitive change as a research question, not a footnote.</p><p>This kind of research is harder to do. It takes longer. It is more expensive. It produces noisier results. It is less likely to yield the clean path coefficients and the R&#178; of 0.475 and the SRMR of 0.031 that signal competence to reviewers. The incentive structure of academic publishing does not reward it.</p><p>The Wang and Fan retraction is the kind of failure that looks like a violation of the rules. Wang and Zhang is the kind of failure that looks like following them.</p><div><hr></div><p>I am building AI tools for anyone who wants to ride the AI revolution. I am not the right person to tell education researchers how to fix their field. But I notice the same thing in AI music research that I see here: the willingness to dress up a survey with sophisticated analytical machinery and call the output evidence about what AI actually does to people. The infrastructure for appearing rigorous has outpaced the infrastructure for being rigorous.</p><p>And this matters beyond the journals. The Wang and Zhang paper is circulating as evidence about AI and learning. Institutions are making policy based on papers like this. Educators are redesigning curricula. Students are being told, by implication, that their sense of having learned something is the same as having learned something.</p><p>It is not. And the gap between those two things is exactly the gap that Mezirow was writing about &#8212; the gap between the story you tell yourself about your perspective and the actual reconstruction of the framework through which you understand the world. Transformative learning is what happens when you discover that the story you have been telling yourself is wrong.</p><p>It would be ironic if the research claiming to measure it turned out to be an example of the thing it failed to measure.</p><div><hr></div><p><em>Nik Bear Brown teaches AI at Northeastern University and runs Musinique LLC, which builds tools for indie musicians. He is also the founder of Humanitarians AI, a 501(c)(3) nonprofit. More at <strong><a href="http://bear.musinique.com/">bear.musinique.com</a></strong> &#183; <strong><a href="http://skepticism.ai/">skepticism.ai</a></strong> &#183; <strong><a href="http://theorist.ai/">theorist.ai</a></strong></em></p><div><hr></div><p><strong>Tags:</strong> measurement validity, AI education research, transformative learning, construct validity, self-report bias</p>]]></content:encoded></item><item><title><![CDATA[The Ladder That Isn't There]]></title><description><![CDATA[What Companies Are Building to Replace the Rung AI Eliminated]]></description><link>https://www.skepticism.ai/p/the-ladder-that-isnt-there</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-ladder-that-isnt-there</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Sat, 25 Apr 2026 23:09:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!pJ47!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pJ47!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pJ47!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!pJ47!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!pJ47!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!pJ47!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pJ47!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2264048,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/195482027?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pJ47!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!pJ47!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!pJ47!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!pJ47!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0b467d6-3ccf-4a8d-be6e-cc1c5debce93_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The argument goes like this: AI automates entry-level coding work, so companies stop hiring junior developers, so there is nobody to become the senior developers of 2030, so the companies that cut the pipeline will find themselves in 2030 with powerful AI tools and no one with the judgment to use them safely. IBM&#8217;s chief human resources officer, Nickle LaMoreaux, made exactly this case in February 2026, announced that IBM was tripling its entry-level hiring, and called on HR leaders across the industry to do the same. &#8220;The companies three to five years from now that are going to be the most successful,&#8221; she said, &#8220;are those companies that doubled down on entry-level hiring in this environment.&#8221;</p><p>It is a coherent argument. It is also, in its publicly available form, incomplete in precisely the ways that matter most.</p><h2>The Gap Between the PR and the Pipeline</h2><p>LaMoreaux is right about the pipeline problem. She is far less specific about the solution. What IBM has said publicly is that it &#8220;rewrote&#8221; entry-level software developer roles &#8212; less boilerplate coding, more AI oversight, more customer interaction, more focus on what the company calls &#8220;systems judgment.&#8221; Junior developers will spend less time on routine code generation and more time auditing AI output, working directly with clients, and doing the cognitive work of translating business requirements into prompts that produce useful results.</p><p>This is not nothing. It represents a genuine attempt to think through what the entry-level job becomes when AI can generate syntactically correct code faster than a human junior can type it. But there is a question embedded in the new job description that IBM has not publicly answered, and it is the only question that matters: does &#8220;AI oversight&#8221; actually develop the judgment needed to become a senior engineer?</p><p>The historical pathway was not glamorous. A junior developer spent two, three, four years writing boilerplate. Authentication flows, database migration scripts, unit tests, CRUD endpoints. Nobody loved the work. The work was, in terms of its immediate output, largely automatable. But the work was also, in terms of its developmental function, the curriculum &#8212; and the precise mechanism was not the writing. It was the failure. You wrote the authentication flow. It broke in production in ways you did not anticipate. The error message was visible, the gap between your expectation and reality was undeniable, and you had no choice but to struggle with it. You debugged it, which meant reading documentation you hadn&#8217;t read, asking a senior why your mental model was wrong, building a new mental model to replace it. You did this thousands of times. At the end of the process you were a senior engineer &#8212; not because you had written a lot of boilerplate, but because engaging repeatedly with its failures had built something durable in your brain.</p><p>This distinction matters, because it reframes the problem precisely. AI does not just remove the writing. It removes the visible failure. Code compiles. Tests pass. The race condition hides inside a sleep call. The memory leak is invisible to the test suite. The architectural drift from intent looks like a working feature until it fails at scale in production. The failure is still there &#8212; AI-generated code fails in ways human-generated code fails, and in new ways besides. But the failure is no longer surfacing where the junior developer can see it, at a latency and legibility that would allow them to learn from it. That is the actual developmental gap.</p><h2>The Comprehension Debt Problem</h2><p>Anthropic published research in January 2026 that should be uncomfortable for every company now designing &#8220;AI-native&#8221; entry-level roles. Junior developers who delegated code generation to AI tools scored between 24% and 39% on subsequent comprehension assessments. Those who used AI as a collaborator &#8212; asking questions, challenging outputs, forcing themselves to understand what the AI produced &#8212; scored between 65% and 86%. The difference is not AI versus no AI. The difference is <em>how</em> you use the tool.</p><p>The researchers called the gap &#8220;comprehension debt&#8221; &#8212; a cumulative deficit between what the codebase does and what the people managing it understand. It is a subtle disaster. The code works. The tests pass. The junior developer ships the feature. The comprehension debt doesn&#8217;t reveal itself until the system breaks in a way that requires architectural judgment to diagnose &#8212; which is precisely the moment when you need the senior engineer who was supposed to emerge from the junior developer who was supposed to be learning while working.</p><p>There is neurophysiological evidence for the mechanism. A 2025 MIT study by Kosmyna et al. tracked EEG connectivity in participants writing under three conditions: LLM-assisted, search-engine-assisted, and unaided. Across alpha, theta, and delta bands &#8212; associated with internal semantic processing, working memory, and self-directed ideation &#8212; connectivity scaled inversely with external support. LLM users showed the weakest brain network engagement. More consequentially: when LLM-habituated participants were later asked to work without the tool, their neural connectivity did not reset to novice levels, but it did not reach the levels achieved by practiced unassisted writers either. Alpha and beta engagement &#8212; associated with top-down planning and self-driven organization &#8212; remained measurably suppressed. The authors call this accumulation &#8220;cognitive debt.&#8221; The study involves essay writing rather than software development, and the sample of 54 students is too small to carry causal weight. But the finding is structurally consistent with the broader claim: if the generative cognitive work is externalized during the period when mental models are supposed to form, those models form incompletely &#8212; and the deficit persists when the tool is removed.</p><p>Microsoft&#8217;s Azure CTO Mark Russinovich and VP Scott Hanselman put the problem with blunt clarity in a February 2026 paper in <em>Communications of the ACM</em>. Senior engineers experience an &#8220;AI boost&#8221; &#8212; the tools multiply their throughput, and they have the judgment to steer and verify the output. Junior engineers experience what Russinovich and Hanselman call &#8220;AI drag&#8221; &#8212; the tools produce output that looks correct, which the junior developer lacks the judgment to evaluate, and the work is done without the learning happening. The rational economic response for any CFO is to hire seniors and automate juniors. The structural consequence is: no pipeline.</p><p>What makes their diagnosis particularly useful is that they catalogue the specific failure modes AI tools exhibit that juniors cannot catch without guidance: agents masking race conditions with sleep calls, agents claiming success on buggy code, agents implementing algorithms that pass tests but don&#8217;t generalize. These are Layer 1 failures &#8212; implementation-level breakdowns in code that appears to work. A junior developer encountering these outputs sees success where a senior sees warning signs. The failure signal exists. It is not visible to the person who needs to learn from it.</p><h2>The IBM Critique, Sharpened</h2><p>IBM&#8217;s rewritten roles can be mapped onto the three types of failure signal that produce engineering judgment. There is implementation-level failure &#8212; the race condition, the architectural drift, the code that claims success when bugs remain. There is systems-level failure &#8212; the customer complaint that maps through the stack to a root cause nobody documented. And there is specification-level failure &#8212; the moment someone has to stake their name on whether the requirements themselves were right.</p><p>The old boilerplate model exposed juniors to implementation-level failure almost exclusively, and accidentally. The new IBM model &#8212; AI oversight, customer interaction, requirements translation &#8212; is, in theory, exposure to all three. That is not a step backward. It might be a step forward.</p><p>But the theory collapses without the preceptorship. Implementation-level failures in AI output are invisible to someone who lacks enough technical intuition to recognize them. You cannot learn to catch the subtle wrong if no one makes the subtle wrong visible. IBM has rewritten the job description to include &#8220;AI oversight&#8221; without building the structural condition under which AI oversight actually teaches anything. Without a preceptor paired with the junior, making the failure legible &#8212; pointing at the sleep call masking the race condition and explaining <em>why</em> that is wrong, not just that it failed &#8212; the oversight role is compliance work, not learning. The junior sees that the tests passed. The preceptor sees the problem the tests don&#8217;t catch. Without the preceptor, that gap is just a gap.</p><p>Some organizations are doing more than announcing intentions. The responses are uneven, but they are real.</p><p>Microsoft proposed a preceptorship model that is worth examining in detail. The structure is adapted from clinical nursing: senior engineers paired with early-in-career developers at three-to-one or five-to-one ratios, for a minimum of one year, on real product teams rather than training sidecars. AI tools are configured to operate in what Russinovich and Hanselman call &#8220;EiC mode&#8221; &#8212; Socratic coaching before code generation, forcing the junior to articulate what they&#8217;re trying to accomplish before receiving a solution. Mentorship hours are measured as &#8220;human impact&#8221; alongside product metrics in performance reviews, which means the senior engineer&#8217;s career is now connected to the junior&#8217;s development, not just the senior&#8217;s own throughput. The model is modeled on clinical preceptorships explicitly because clinical nursing faced the same problem decades ago: how do you develop judgment in someone who is working in a high-stakes environment with experienced practitioners who have better things to do than teach?</p><p>Russinovich and Hanselman are honest about the limits of their own proposal. Microsoft cut significant engineering headcount in 2024 and 2025. Whether the preceptorship model will scale into a sustained program depends on whether leadership changes the metrics they optimize &#8212; a &#8220;big ask&#8221; for organizations whose incentives have historically emphasized shipping velocity above all else.</p><p>McKinsey redesigned its screening process for the AI era through an assessment called Solve &#8212; a gamified evaluation that tests critical thinking, decision-making, and systems thinking, explicitly not prior business knowledge or technical credentials. The framing is sound: what the company needs is people who can learn in the new environment, not people who already know the old skills. Whether a better hiring filter compensates for a weaker developmental pathway is not yet clear.</p><p>IBM&#8217;s own &#8220;New Collar&#8221; apprenticeship program is being updated to include what the company calls &#8220;AI-native habits&#8221; &#8212; using AI tools to deconstruct pull requests rather than build from scratch, understanding the architecture of LLMs, designing with generative tools before implementing. The Flatiron School is running an &#8220;Accelerated AI Engineer Apprenticeship&#8221; that pairs participants with mentors on real agentic frameworks at $20 per hour, with a foundations-first approach that introduces concepts simply before revisiting them with increasing technical depth.</p><p>These are attempts. They are not yet evidence.</p><h2>The Review Tax Nobody Discusses</h2><p>There is a cost to the existing senior engineers that the pipeline conversation mostly ignores. When one senior can generate the volume of three juniors, the productivity gains are real. But generating code is cognitively different from verifying code, and the verification is now happening at three times the volume.</p><p>Senior engineers are spending their days as high-speed compliance officers. Thousands of lines of AI-generated logic, auditing for subtle hallucinations &#8212; race conditions masked by sleep calls, code that passes tests but doesn&#8217;t generalize, architectural drift that looks fine in isolation and fails at scale. A 2025 paper found that after AI adoption, core developers reviewed more code but their own original productivity dropped 19%. The creative, architectural, problem-solving work that makes senior engineering satisfying and that produces the judgment juniors are supposed to be learning from &#8212; that work is being crowded out by the cognitive exhaustion of reviewing AI output at industrial scale.</p><p>The delegation vacuum compounds this. Seniors previously handed off lower-risk tasks to juniors as a pressure valve and as a teaching mechanism. Junior implements the UI component, senior reviews it, junior learns something. That loop no longer exists. The junior&#8217;s tasks were automated. The senior&#8217;s workload increased. The teaching is not happening.</p><p>This is the tax that makes the developmental problem worse. The senior engineers who were supposed to mentor are stretched thin doing work that used to be distributed. The preceptorship model addresses this in theory &#8212; by making mentorship a measured part of the senior&#8217;s job rather than an afterthought. Whether organizations are actually willing to accept the velocity tradeoff is a different question.</p><h2>What Is Actually Known</h2><p>The honest answer to the core question &#8212; can AI-assisted entry-level work produce the same developmental outcomes as the boilerplate-and-struggle model &#8212; is that nobody knows yet.</p><p>The cohort that entered the workforce in 2024 and 2025 under AI-assisted conditions will become mid-level engineers in 2027 and 2029. Whether they emerge with the architectural judgment, the debugging instincts, the systems thinking that the old pipeline produced will not be visible until then. The data will arrive precisely when it is needed most &#8212; when those engineers are supposed to be the senior developers filling the next generation&#8217;s pipeline &#8212; and if the answer is no, the remediation options will be limited and expensive.</p><p>The Dreyfus model of skill acquisition gives a name to what is at risk. Novices follow rules. Advanced beginners develop pattern recognition. Competent practitioners make choices and bear the consequences of those choices &#8212; this is where accountability and emotional investment enter, and where learning accelerates. Proficient practitioners sense problems before the data confirms them. Experts operate through intuition built from thousands of absorbed experiences. The concern is not that AI-assisted juniors are incompetent. It is that they plateau. They recognize patterns. They generate outputs that look like what competent practitioners produce. But they have not made choices whose consequences they had to live with. They have not debugged the 2am production failure that rewired their mental model of how distributed systems actually behave. They have not asked a senior why their elegant solution was wrong and received an answer that changed how they think permanently.</p><p>The Kosmyna finding is the most uncomfortable piece of evidence in this space. It is preliminary and domain-limited. But if it holds in technical domains &#8212; if the cognitive debt from AI-assisted early-career work doesn&#8217;t reverse when the tool is removed &#8212; then the preceptorship model is not sufficient on its own. The preceptor can make visible the failure the junior cannot yet see. But they cannot rebuild the neural substrate that early unassisted struggle was supposed to create. The minimum viable intervention may require some version of deliberately maintained struggle &#8212; manual-first implementation for foundational modules, Socratic AI tools that require the junior to predict before they receive &#8212; to preserve the generative cognitive engagement that builds the mental models the preceptorship then calibrates.</p><h2>The Wager</h2><p>IBM&#8217;s wager is that oversight, verification, and customer-facing accountability can replace the old developmental pathway. That a junior developer who spends years auditing AI output, explaining architectural choices to clients, and taking responsibility for the correctness of generated code will develop the judgment that used to come from writing and debugging the code yourself.</p><p>It might be true. And the three-layer framing suggests it could be more than just &#8220;not worse&#8221; &#8212; exposure to systems-level and specification-level failure earlier in a career, rather than after years of boilerplate, might actually compress the timeline to senior judgment rather than extend it. Customer-facing rotation, where the junior must translate vague failure descriptions into root-cause hypotheses, is the kind of developmental experience that the old model often didn&#8217;t provide until mid-career.</p><p>But the theory requires the load-bearing piece that IBM has not publicly committed to: preceptorship at Stage 1. The implementation-level failures in AI output are invisible to a junior who lacks the technical intuition to recognize them. Making those failures legible is the senior engineer&#8217;s job &#8212; not reviewing for correctness, but externalizing judgment that the junior cannot yet access. Without that, the oversight role is compliance work. The junior sees tests passing where the senior sees warning signs. The gap between those two observations is where the learning was supposed to happen.</p><p>LaMoreaux is right that the companies which doubled down on entry-level hiring in this environment will be better positioned in 2030. She is right that the pipeline problem is real. What she has not yet answered &#8212; what no major company has publicly answered with evidence &#8212; is whether the new developmental pathway they are building actually delivers Stage 2 and Stage 3. Whether the junior who spends a year doing AI oversight develops the systems intuition to translate &#8220;it stops working sometimes&#8221; to root cause. Whether they get to the point of staking their name on an architectural judgment call, being wrong about something, and learning from the consequence.</p><p>The ladder looks different. Whether it goes to the same place, and whether the companies building it have designed the rungs deliberately enough to find out, we do not yet know.</p><div><hr></div><p><strong>Tags:</strong> junior developer pipeline AI, failure signal model developer expertise, IBM entry-level roles 2026, Kosmyna cognitive debt LLM, Russinovich Hanselman preceptorship ACM</p>]]></content:encoded></item><item><title><![CDATA[The Robot Tutor and the Fishing Village]]></title><description><![CDATA[What "Personalization" Has Always Meant, and What Adaptive Learning Has Always Delivered]]></description><link>https://www.skepticism.ai/p/the-robot-tutor-and-the-fishing-village</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-robot-tutor-and-the-fishing-village</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Fri, 24 Apr 2026 03:20:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!JT8Q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JT8Q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JT8Q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!JT8Q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!JT8Q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!JT8Q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JT8Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1635623,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/194873207?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JT8Q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!JT8Q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!JT8Q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!JT8Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3817a05d-7e7f-4fc7-b7c8-97b0926accd6_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The girl in the Cambodian fishing village was never real.</p><p>She was an argument. Between 2013 and 2015, Jos&#233; Ferreira, founder of Knewton, invoked her in promotional materials and public statements to describe what his technology could do: a girl in a fishing village, receiving through Knewton&#8217;s adaptive engine the same personalized instruction as a student at an elite private school, growing up to invent the cure for ovarian cancer. Educational inequality, in Ferreira&#8217;s framing, was a problem that adaptive learning could address at the software layer. The instruction would be what unlocked the capacity. The fishing village was a rhetorical device, not a pilot deployment.</p><p>By 2019, Knewton had been acquired by John Wiley &amp; Sons for a sum understood to be a small fraction of its peak valuation. The partnership with Pearson had dissolved. The product that remained &#8212; Knewton Alta, a conventional higher-education courseware platform &#8212; bore little resemblance to the robot tutor in the sky. The fishing village was still waiting.</p><p>I want to examine what happened. Not Knewton specifically, and not Ferreira personally &#8212; he was the most articulate spokesman for a framing the whole industry was using, not its author. What I want to examine is the word that Ferreira&#8217;s framing deployed, the word that was doing the most rhetorical work in every version of that framing, the word that has survived the collapse of its first generation of spokescompanies and is still doing the same work today.</p><p><em>Personalization.</em></p><div><hr></div><h2>What the Word Invokes</h2><p>The word has a history in educational psychology that predates by decades any commercial deployment of adaptive software. Lev Vygotsky&#8217;s zone of proximal development is about personalization &#8212; the idea that effective instruction operates in the specific zone between what a learner can do independently and what they can do with support, a zone that is different for every learner and that requires a teacher&#8217;s specific attention to identify. Lee Cronbach and Richard Snow&#8217;s work on aptitude-treatment interactions spent two decades trying to formalize the finding that different learners respond differently to different instructional approaches &#8212; that no single method is optimal for everyone, and that the optimal method for a given learner depends on who that learner is. The differentiated-instruction tradition in teacher education has argued for thirty years that good teaching requires knowing students individually, designing instruction around their specific needs, and adjusting in real time to what each student brings and what each student shows.</p><p>The construct is real. It has serious empirical and theoretical grounding. When Ferreira said Knewton was personalizing learning, he was invoking this history &#8212; pointing at a tradition that educational psychology had spent decades documenting and that every good teacher knows, in the bone, as what it means to actually teach rather than to deliver content.</p><p>What Knewton&#8217;s technology operationalized was different.</p><p>Knewton&#8217;s engine was built on two well-established statistical techniques. The first was Item Response Theory, the mathematical framework underlying modern standardized testing, which models the probability of a correct response as a function of a student&#8217;s latent ability and an item&#8217;s difficulty. The second was Bayesian Knowledge Tracing, which estimates whether a student has mastered a specific discrete skill by updating probability estimates as the student responds to items. Together, these gave Knewton a learner model: a collection of probability distributions over latent abilities and specific skill masteries, updated continuously as the student interacted with the system.</p><p>This is real technology. It is not trivial to build. The engineers who built it did substantive mathematical work. Knewton&#8217;s claim that its engine operated on sophisticated foundations was true. What was not quite true was the claim about what those foundations amounted to.</p><p>The learner model Knewton maintained was expressible, in its technical form, as: <em>the probability this student has mastered skill A is 0.78; the probability this student has mastered skill B is 0.34; the student&#8217;s estimated ability on dimension X is 1.2 standard deviations above the population mean.</em> This is useful information for deciding what to present next. It is not a model of the student as a person. It is not a model of their interests, their emotional state, their cognitive style, their cultural background, their creative capacity, their relationship to learning. It is a model of item-response patterns on a bank of pre-authored content.</p><p>The gap between <em>we know this student better than their parents</em> and <em>our model assigns probabilities to their mastery of skills we&#8217;ve tagged to a knowledge graph</em> is the central artifact of the adaptive-learning era.</p><div><hr></div><h2>The Fishing Village Made Specific</h2><p>The girl in the Cambodian fishing village makes the gap visible because the specific nature of what was claimed and what was possible becomes clear once you name each requirement.</p><p>For the girl to receive, through Knewton&#8217;s engine, instruction equivalent to an elite private-school education, the technology would need, first, content: a comprehensive curriculum in mathematics, science, language, and humanities, built by human curriculum developers, available in a language she could read, calibrated for her cultural and linguistic context. Knewton licensed pre-authored material from publishers. The content was what the publishers had built and the partnerships had arranged. The engine sequenced content that already existed. Building the content was not what the engine did.</p><p>The technology would need, second, an outcome measure capable of telling whether the instruction was producing the kind of understanding that leads to cancer research &#8212; conceptual depth, transfer across domains, creative problem-solving, the tacit skills that accumulate over years of serious engagement with scientific thinking. Knewton&#8217;s engine could measure item-level response patterns on pre-authored assessments. Whether those patterns indexed what a future researcher would need was not addressed. The engine was not designed to measure the construct the rhetoric invoked.</p><p>The technology would need, third, to function in conditions of intermittent electricity, unreliable internet, shared devices, limited home support, a language and cultural context for which the content was probably not designed. Knewton was built for contexts with substantially more infrastructure. The rhetoric invoked the fishing village as a demonstration of reach. The technology had not been deployed there or validated there.</p><p>The claim was aspirational. The <em>could</em> was doing substantial work. What was true was that the technology could hypothetically produce this outcome if a great many other things were also true, none of which were Knewton&#8217;s responsibility or within Knewton&#8217;s control. The fishing village was a vision of what the future might look like if a great many problems that have nothing to do with adaptive sequencing algorithms were solved. It was not a description of what Knewton could actually deliver.</p><div><hr></div><h2>Three Systems, One Pattern</h2><p>The pattern the Knewton arc illustrates is not Knewton-specific. It appears, in different configurations, across every major adaptive-learning platform that followed.</p><p>DreamBox Learning, focused on K-8 mathematics and backed by the strongest external evidence base in the category, has been evaluated by the Harvard Center for Education Policy Research in multiple studies. The evaluations used standardized mathematics assessments over school-year timescales and were conducted by researchers with no affiliation to the company. The findings: effect sizes in the range of 0.10 to 0.15 standard deviations for students using the platform at recommended levels. Real effects. Detectable by rigorous researchers using independent measures. Considerably more modest than the marketing implied. And dependent, in every evaluation, on implementation &#8212; on how much classroom time schools actually allocated to the platform. The adaptive sophistication of the software did not substitute for the hours it required.</p><p>i-Ready, among the most widely deployed adaptive platforms in American K-12 education, integrates adaptive diagnostic assessment with what the company calls &#8220;Personalized Instruction&#8221; &#8212; a sequence of pre-authored lessons targeted at the student&#8217;s estimated level. Critics have noted that the personalization, operationally, consists of placing students at different starting points in a common instructional sequence. Students are still completing pre-authored lessons. They are starting at different points and progressing at different speeds. Whether this is <em>personalization</em> in the sense the word implies &#8212; instruction responsive to who the student is &#8212; or more honestly <em>adaptive placement within a fixed curriculum</em>, is exactly the question the word is being deployed to avoid asking.</p><p>ALEKS, built on Knowledge Space Theory, represents the most theoretically rigorous operationalization in the category. Rather than treating ability as a single number, Knowledge Space Theory maps a domain as a set of discrete items and a learner&#8217;s knowledge state as the specific subset of items they have mastered. ALEKS uses an AI engine to efficiently navigate the combinatorial space of possible knowledge states, asking questions that narrow its estimate of where the student is. The resulting ALEKS Pie &#8212; a visual display of what has been mastered, what has not, what is ready to learn &#8212; is grounded in serious mathematics, specified precisely, falsifiable in principle. It has been evaluated in multiple contexts. Effect sizes fall in the same general range as DreamBox and i-Ready.</p><p>What is clarifying about ALEKS is this: even the most theoretically careful operationalization of personalization &#8212; one drawing on decades of rigorous mathematical work &#8212; models a student&#8217;s mastery state over a defined domain of discrete items. It does not model the student&#8217;s interests, their emotional state, their cognitive style, their cultural background, their creative capacity, their relationships. ALEKS is honest about this. The documentation says clearly that the system models knowledge states over specific domains. But even ALEKS demonstrates that the gap between the marketing construct and the technical operationalization is not a failure of specific companies. It is a feature of what item-level response tracking can and cannot do.</p><div><hr></div><h2>The Gap and Its Consequences</h2><p>The word <em>personalization</em> is doing specific rhetorical work. It invokes a construct that educational psychology spent decades building &#8212; instruction responsive to the individual learner in the deep sense that Vygotsky pointed at, that good teachers practice, that Cronbach and Snow tried to formalize. The construct is real. The technology operationalizes something narrower: item-level response tracking, probability distributions over mastery parameters, next-item selection from pre-authored content banks, pacing adjustments based on observed response patterns. This is what the data these systems collect and the algorithms they run can actually support. It is not trivial. It is not the same thing as the construct the word invokes.</p><p>Three consequences follow.</p><p>Critiques of adaptive learning for failing to deliver what the marketing promised are both fair and partially misdirected. Fair because the systems cannot deliver what the rich construct invokes. Misdirected because assigning this to specific companies treats a structural feature of item-level tracking as a product failure. The rhetoric over-promised. The technology delivered what the technology could deliver.</p><p>Evaluations of these systems on outcome measures aligned to the item-level tracking are measuring the operationalization, not the construct. They find modest positive effects, which is the honest finding. Whether the same systems produce transfer to novel problems, durable learning over years, growth in dimensions that do not map to any test-bank item &#8212; these questions remain mostly unanswered, because answering them would require outcome measures that do not yet exist in the forms evaluators would need.</p><p>And the pattern persists. The vocabulary has survived the collapse of Knewton and its generation. When current AI-tutor companies claim to provide personalized tutoring, to adapt to each learner&#8217;s needs, to meet students where they are, the claim is doing the same rhetorical work Knewton&#8217;s robot tutor in the sky was doing: invoking the rich construct while operationalizing a narrower version. The gap remains where it was.</p><div><hr></div><h2>What to Ask</h2><p>When you next encounter an educational-technology claim that uses the word <em>personalization</em>, or variants like <em>individualized</em> or <em>adaptive</em> or <em>tailored to the learner</em> or <em>meets each student where they are</em>, two questions will orient you.</p><p>What, specifically, is the technical operation? The honest answer for the large majority of systems using this vocabulary is one of a small family: item-level response tracking with adaptive item selection; diagnostic assessment followed by placement in a pre-authored sequence; pacing adjustments based on response patterns; content recommendation from a pre-authored bank based on inferred mastery. If you can name which operation is happening, you have the beginning of an honest account of what the system does. The vocabulary may suggest more. The technical substrate does not support more.</p><p>Does the claim invite the listener to believe the system does something the operation does not do? The answer is often yes, specifically in the dimensions educators and parents most hope for. Operationalized personalization &#8212; item selection based on mastery estimates &#8212; can contribute to instruction responsive to the individual learner, in contexts where it is embedded in the harder relational and responsive work that teachers do. It cannot replace that work. When a product is marketed as though algorithmic item selection substitutes for a teacher&#8217;s specific attention to a specific child, the marketing is doing rhetorical work the technology does not underwrite.</p><p>The fishing village is still waiting. The girl who will invent the cure for ovarian cancer has not yet received the education the rhetoric promised. This is not primarily Ferreira&#8217;s fault, or Knewton&#8217;s, or any single company&#8217;s. It is the consequence of a gap that was always structural &#8212; between what a word can invoke and what a technical operation can deliver &#8212; that the field has chosen, for a decade and more, not to name.</p><p>Naming it is the prerequisite to closing it.</p><div><hr></div><p><em>Nik Bear Brown is Associate Teaching Professor of Computer Science and AI at Northeastern University and founder of Humanitarians AI (501(c)(3)). This essay appears as part of the Computational Skepticism series at <a href="https://skepticism.ai">skepticism.ai</a>. | <a href="https://theorist.ai">theorist.ai</a></em></p><div><hr></div><p><strong>Tags:</strong> adaptive learning personalization gap, Knewton IRT Bayesian knowledge tracing operationalization, DreamBox i-Ready ALEKS efficacy evaluation, personalized learning construct versus operation, EdTech rhetoric fishing village critique</p>]]></content:encoded></item><item><title><![CDATA[The Assessment Was Already Broken]]></title><description><![CDATA[On Jessica Winter's "What Will It Take to Get A.I. Out of Schools?" and what the panic about AI reveals about everything that came before it]]></description><link>https://www.skepticism.ai/p/the-assessment-was-already-broken</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-assessment-was-already-broken</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Fri, 24 Apr 2026 00:37:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!l9KP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!l9KP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!l9KP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png 424w, https://substackcdn.com/image/fetch/$s_!l9KP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png 848w, https://substackcdn.com/image/fetch/$s_!l9KP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png 1272w, https://substackcdn.com/image/fetch/$s_!l9KP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!l9KP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png" width="1456" height="522" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:522,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2081000,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/195299281?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!l9KP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png 424w, https://substackcdn.com/image/fetch/$s_!l9KP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png 848w, https://substackcdn.com/image/fetch/$s_!l9KP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png 1272w, https://substackcdn.com/image/fetch/$s_!l9KP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2689ec55-93a2-4ccb-b9e0-c0ecbdcd191e_3018x1082.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A response to Jessica Winter's<strong><a href="https://www.newyorker.com/culture/progress-report/what-will-it-take-to-get-ai-out-of-schools"> "What Will It Take to Get A.I. Out of Schools?"</a></strong></p><p>There is a moment in Jessica Winter&#8217;s New Yorker piece that contains the entire argument she doesn&#8217;t make. Her sixth-grade daughter runs a fifth-grade slide show through Gemini&#8217;s beautifying tools. In thirty seconds, the typography improves, the pictures reshuffle symmetrically, the design evokes fifteenth-century movable type against a background of aged vellum. Winter describes it as the pool race from <em>Mommie Dearest</em>: the larger, faster thing that will always beat you.</p><p>Her daughter is unmoved. &#8220;I like mine better, because it&#8217;s original and I worked really hard on it.&#8221;</p><p>Hold that sentence. It is the right answer. It is also the answer that does not appear on any rubric in any public school in Massachusetts or New York or Los Angeles. The rubric rewards the prettier slide. The rubric was always going to reward the prettier slide. Winter wants her daughter to hold values that the institution has never rewarded, and she writes a five-thousand-word piece about artificial intelligence without once asking why the institution doesn&#8217;t reward them.</p><p>This is the intellectual hole at the center of a piece that is otherwise sharp, well-reported, and morally earnest. AI didn&#8217;t break the assessment system. It exposed that the assessment system was already broken, and everyone was pretending otherwise.</p><div><hr></div><h2>What the Slide Show Already Was</h2><p>The printing-press slide show existed before Gemini. It was made in fifth grade to demonstrate learning. Whether it demonstrated learning was always a question nobody asked, because asking it would require admitting that the artifact &#8212; the thing handed in, the thing graded &#8212; was never reliable evidence of the process. The slide show could have been made with a parent&#8217;s help, with a template, with a slightly older sibling, with a capable friend who understood visual design. These interventions existed before large language models. They produced polished artifacts that the teacher accepted as evidence of understanding.</p><p>The educational research on this predates AI by decades. Robert Bjork&#8217;s distinction between performance and learning &#8212; the observable output versus the durable cognitive change &#8212; is from 1992. The problem of using artifacts as proxies for thinking is at least as old as Vygotsky. What AI did was not create this problem. It made the problem so visible, so fast, so cheap, that willful ignorance became impossible.</p><p>Winter quotes USC professor Mary Helen Immordino-Yang: &#8220;We are cutting off learning at the knees.&#8221; She quotes University of Toronto psychologist Amy Finn on the magic of how children retain unexpected, non-strategic details that adults would find irrelevant, a kind of creative unpredictability fundamentally misaligned with LLMs&#8217; orientation toward speed and sleekness. These are real insights. They are also insights that apply equally to the printing-press slide show assigned as homework, graded for visual appeal and accuracy, returned in two days, and forgotten. The neuropsychological substrate for creating narratives and thinking through arguments over time is not developed by making a slide show under time pressure at home with no adult monitoring the process.</p><p>The question is not whether AI belongs in schools. The question &#8212; the one the piece never asks &#8212; is whether the assessment was measuring what it was supposed to measure before AI arrived. The answer is: sometimes, unevenly, and less than we told ourselves.</p><div><hr></div><h2>The Tool Hierarchy Problem</h2><p>Winter&#8217;s implicit argument, followed consistently, condemns more than Gemini. Calculators offload arithmetic before numeracy is built. Spell-check offloads orthography. Grammarly offloads syntax judgment. Google Search offloads memory and source evaluation. Slide templates offload visual design judgment. Word processors themselves offload handwriting, which Winter mentions approvingly has developmental benefits &#8212; which means she believes at least one tool was introduced too early.</p><p>She draws the line at the tool that frightens her right now. This is a very human response and a terrible policy foundation.</p><p>The honest version of her argument looks like a developmental sequence: here are the cognitive substrates that must be built before each category of tool is introduced, and here is the evidence for that ordering. Immordino-Yang and Finn gesture at this &#8212; the &#8220;cognitive muscles&#8221; framing, the concern about atrophy before onloading &#8212; but nobody builds it out into something a school board could actually implement. Without that framework, the anti-AI position reduces to: tools I grew up with are fine, tools that postdate my childhood are suspect.</p><p>Amanda Bickerstaff, CEO of AI for Education, comes closest to the principled version: children should not be using chatbots under age ten, she says, because these tools require expertise and evaluation skills that even many adults don&#8217;t have. That&#8217;s a threshold with a rationale. It&#8217;s also the only threshold in the piece with a rationale. Everything else is rhetoric standing in for policy.</p><div><hr></div><h2>The Research That Isn&#8217;t Quite Research</h2><p>The piece anchors much of its scientific authority in three studies. The 2025 MIT warning that LLMs &#8220;may inadvertently contribute to cognitive atrophy&#8221; &#8212; the authors felt it necessary to append an FAQ asking journalists not to use words like &#8220;brain rot&#8221; or &#8220;brain damage,&#8221; which tells you something about how the finding was being reported before Winter&#8217;s piece and how it will be reported after. The multi-institution study (MIT, CMU, UCLA, Oxford) on fraction-solving, which showed that students who lost AI access after using it performed significantly worse &#8212; not yet peer-reviewed, not yet published, findings are concerning, the concern is real. The Brookings &#8220;premortem,&#8221; which pairs 400 studies with hundreds of interviews to conclude that AI tools &#8220;undermine children&#8217;s foundational development.&#8221;</p><p>These are worth taking seriously. They are also worth examining carefully.</p><p>The fraction-solving study is the most empirically specific, and it is also the most useful argument against Winter&#8217;s piece rather than for it. The students who used LLMs on fraction-solving and then lost access performed significantly worse and were more likely to give up. The proposed mechanism: AI gives answers, students become dependent on the answer-giving, remove the answers and the capacity to generate them independently has atrophied.</p><p>But this is an argument about a specific implementation &#8212; an answer machine &#8212; not about the technology class. An LLM configured as a Socratic interlocutor, one that refuses to answer directly and instead returns questions that scaffold toward understanding, that detects when a student is stuck versus when they&#8217;re avoiding, that withholds confirmation until the student demonstrates the reasoning &#8212; that tool would presumably produce the opposite result. Students would have developed the reasoning process rather than outsourcing it, because outsourcing was never made available to them.</p><p>This is not an exotic capability. It is prompt engineering plus scaffolding logic. The reason it isn&#8217;t what&#8217;s being deployed in K-12 classrooms is that Google ships Gemini with a &#8220;Help me write&#8221; button because that&#8217;s the path of least resistance and maximum engagement. That is a product decision, not a technological inevitability. Winter never distinguishes between AI as answer machine and AI as thinking partner. The cognitive offloading critique collapses the moment you make that distinction, because the problem isn&#8217;t the tool &#8212; it&#8217;s the incentive structure of the company deploying it.</p><p>The social-emotional hijacking argument from UNC psychologist Mitch Prinstein is the weakest scientific claim in the piece, and it&#8217;s presented with the same credentialed authority as the others. Surging oxytocin and dopamine receptors around ages ten to eleven do drive peer-bonding &#8212; that&#8217;s established developmental neuroscience. Sycophantic LLMs &#8220;hijack the biological tendency to want peer feedback&#8221; &#8212; that&#8217;s a hypothesis, not a finding. The claim requires that chatbot interaction activates the same neurological pathways as peer interaction, that substituting chatbot interaction for peer interaction produces measurable deficits in social skill development, and that the effect is &#8220;hijacking&#8221; &#8212; a strong, directional, causal claim &#8212; rather than displacement or preference shift. No study is cited because none exists at the necessary scale with the necessary longitudinal follow-up.</p><p>This is neuroscience&#8217;s authority dressed over a speculation. Which is particularly ironic given that Winter is writing a piece about tools that generate confident-sounding output without rigorous foundations.</p><div><hr></div><h2>The Grade Your Daughter Is Going to Receive</h2><p>Return to the slide show.</p><p>Winter&#8217;s daughter likes hers better because it&#8217;s original and she worked really hard on it. This is the right value. This is the value Winter wants the school to transmit. The school is not transmitting it, because the school is not grading for it.</p><p>If the rubric rewards polish, visual appeal, and impressive output &#8212; which most rubrics do, implicitly, because these are the things teachers can assess quickly across thirty slide shows at 11pm &#8212; then the student who uses Gemini gets the A. Not abstractly. On the transcript. The student who refuses Gemini, who holds Winter&#8217;s daughter&#8217;s values, receives the C. Neither of them learns the lesson Winter intends.</p><p>The deeper problem: homework was already a weak pedagogical instrument before AI. Most research on homework in K-8 is lukewarm. It was largely accountability theater &#8212; proof that learning happened, easy to grade, easy to assign, poor evidence of the process it was supposed to represent. AI exposed the theater. The theater was playing for years before AI bought a ticket.</p><p>What would it look like to actually assess the process? That question is harder than &#8220;what do we do about Gemini,&#8221; and it requires admitting that the current system was already failing to measure what it claimed to measure. Winter doesn&#8217;t want to ask that question, because asking it would mean the problem is older and deeper than the creepy neighbor who moved in recently.</p><div><hr></div><h2>What Actually Needs to Change</h2><p>The resistance movements Winter profiles &#8212; District 14 Families for Human Learning, the Coalition for an AI Moratorium, Schools Beyond Screens &#8212; are better at stopping things than proposing them. The Student Tech Bill of Rights includes the right to read whole books, write on paper, and learn in a low-stimulation environment free from undue corporate influence. These are reasonable demands. They don&#8217;t add up to a pedagogy.</p><p>The conflict-of-interest thread is the piece&#8217;s most structurally damning detail and the most underplayed. The NYC DOE official overseeing the preliminary AI guidelines holds a fellowship jointly offered by Google and GSV Ventures &#8212; whose portfolio includes Amira and MagicSchool, two of the primary AI tools being deployed in the classrooms those guidelines govern. Other Google-GSV fellowship recipients include top school officials in Berkeley, Dallas, Los Angeles, Newark, Colorado, and Maryland. &#8220;If you ask tobacco companies to help write your school&#8217;s policy on cigarettes,&#8221; one parent says, &#8220;you&#8217;re going to end up with guidance on how to smoke responsibly in school.&#8221;</p><p>This is the argument Winter should have built the piece around. Not &#8220;AI is cognitively harmful&#8221; &#8212; which is partly true, partly speculation, and entirely dependent on implementation &#8212; but &#8220;the people writing the rules are being paid by the companies they&#8217;re supposed to regulate.&#8221; That is verifiable, structural, and not dependent on a not-yet-peer-reviewed study about fractions.</p><p>The piece ends with Sinha&#8217;s question &#8212; &#8220;What do you want from this?&#8221; &#8212; and Winter&#8217;s answer: nothing. It&#8217;s a parent&#8217;s answer. A good parent&#8217;s answer. But it is not a policy answer, and it is not an answer that acknowledges what was already not working before the neighbor moved in.</p><p>The assessment was already broken. The rubric was already rewarding the wrong things. The slide show was already a poor proxy for thinking. AI made all of this impossible to ignore. That is a service, not a crime &#8212; even if the service was rendered by someone with cloven hooves in Yeezy Boosts and a market cap of four trillion dollars.</p><p>What we owe children is not the tools of the past but a clear account of what learning actually is, what evidence of it looks like, and how to build assessments that can tell the difference. That conversation is harder than banning Gemini. It is also the only conversation that addresses what Gemini exposed.</p><div><hr></div><p><em>Nik Bear Brown is Associate Teaching Professor of Computer Science and AI at Northeastern University and founder of Humanitarians AI. His work on AI in education, including the Genuine Learning Protocol framework, is published at bearbrown.co.</em></p><div><hr></div><p><strong>Tags:</strong> AI education New Yorker critique, cognitive offloading assessment design, Bjork learning performance distinction, AI schools policy Jessica Winter, GLP genuine learning protocol</p>]]></content:encoded></item><item><title><![CDATA[The Gap Between What We Measure and What We Name]]></title><description><![CDATA[On the Structural Problem That Forty Years of EdTech Efficacy Research Has Not Solved]]></description><link>https://www.skepticism.ai/p/the-gap-between-what-we-measure-and</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-gap-between-what-we-measure-and</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Thu, 23 Apr 2026 00:38:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ZxKu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZxKu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZxKu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!ZxKu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!ZxKu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!ZxKu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZxKu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1453300,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/194861665?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZxKu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!ZxKu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!ZxKu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!ZxKu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf68abee-cf05-4605-a9fc-933442d405bf_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Consider two findings, forty years apart.</p><p>In 1984, Benjamin Bloom published a seventeen-page paper reporting that students tutored one-on-one under mastery-learning conditions performed approximately two standard deviations above students taught in conventional classrooms. The finding has been cited tens of thousands of times. It has become, across four decades, the single most-invoked benchmark in educational technology. Whenever a new system claims to approach the effectiveness of human one-on-one instruction, it is Bloom&#8217;s 2-sigma it is claiming to approach.</p><p>In 2024, a research team at Harvard led by Gregory Kestin reported that an AI tutor, deployed in introductory physics, produced learning gains larger than active-learning classroom instruction. The effect size exceeded what prior literature had typically reported for any tutoring intervention, including Bloom&#8217;s. The study was methodologically careful. The finding circulated quickly. Within weeks it was being cited as evidence that current-generation AI tutors meaningfully exceed what good conventional instruction can deliver.</p><p>Forty years apart. Different technologies. Different research traditions. And yet, read carefully, the two findings share a structure.</p><p>In each, a specific measurement &#8212; performance on items aligned to the intervention&#8217;s content, assessed at short timescale, against a conventional-instruction baseline &#8212; is offered as evidence for a construct about which the measurement is not, strictly, a measurement. Bloom&#8217;s 2-sigma is evidence about performance on aligned items under particular tutoring conditions in the mid-1980s. It is <em>cited</em> as evidence about the effectiveness of tutoring as an instructional mode. Kestin&#8217;s physics finding is evidence about short-timescale aligned-item performance in a selective undergraduate population. It is <em>cited</em> as evidence that AI tutoring outperforms human instruction in some general sense the measurement does not index.</p><p>The measurements are not false. The findings are not inflated. In each case, the researchers reported carefully what they measured. The question is what happens between the measurement and its citation &#8212; the small, structural, and repeated gap between what the apparatus indexes and what the vocabulary surrounding the apparatus claims.</p><div><hr></div><h2>The Structure of the Problem</h2><p>Name the structure directly.</p><p>An efficacy claim in this field consists of three things: a measurement, a construct, and an asserted relationship between them. The measurement is what researchers actually did &#8212; items administered, scores computed, conditions compared. The construct is what the measurement is meant to be evidence for &#8212; <em>learning</em>, <em>mastery</em>, <em>effectiveness</em>, <em>personalization</em>, <em>engagement</em>. The asserted relationship is the claim that the measurement indexes the construct adequately to license the uses the finding is put to.</p><p>This structure appears in every empirical field. Biology works this way, and so does nutrition research, and so does clinical psychology. The gap between measurement and construct is not a problem specific to educational technology. It is a feature of empirical inquiry. Measurements never exhaustively capture their constructs. The question for any field is how seriously it takes the gap, how much work it does to establish the measurement-construct relationship, and how much it assumes versus demonstrates.</p><p>The observation this book has been building toward, essai by essai, is that the learning-systems field has, across six decades, taken the gap less seriously than its claims require. The measurement-construct relationships it invokes are almost universally assumed rather than demonstrated. The field&#8217;s vocabulary outruns what its evidence apparatus can support, and the gap persists not because it has gone unnoticed &#8212; it has been noticed, repeatedly, by careful researchers across multiple traditions &#8212; but because the apparatus that persists serves specific production conditions, and a more adequate apparatus would serve them less well.</p><p>The structure is not: <em>the field is wrong about what works.</em> The structure is: <em>the field makes claims about effectiveness that its measurements are not positioned to support, and does so systematically.</em> These are importantly different claims. The first is about facts. The second is about apparatus &#8212; about the specific set of measurement practices, citation habits, and research conventions that together produce what the field calls its evidence base.</p><p>The distinction matters because the remedy differs. If the field were making factual errors, the remedy would be better studies of the same interventions. If the apparatus is producing a systematic gap between measurement and claim, the remedy is different apparatus. This book has not argued for either remedy. It has argued, by the accumulated force of twelve close readings, that the second diagnosis is correct.</p><div><hr></div><h2>What the Vocabulary Actually Invokes</h2><p>Open a textbook in educational psychology. Open a learning-sciences journal. Open the marketing copy for any major adaptive-learning platform. Open the abstract of any recent AI-tutor efficacy study. The vocabulary is remarkably consistent. The field claims to be producing evidence about <em>learning</em>. About <em>understanding</em>. About <em>mastery</em>. About <em>effectiveness</em>. About <em>personalization</em> and <em>engagement</em>. Each of these words points toward a construct. Each construct has, in serious research traditions, substantial theoretical and empirical articulation.</p><p>Consider <em>learning</em>. In Robert Bjork&#8217;s decades of experimental work, learning is not a single construct but a distinction between two separable things: storage strength and retrieval strength. Storage strength refers to how well a representation is encoded. Retrieval strength refers to how accessible it is at the moment of test. A student can have high retrieval strength at the end of a unit &#8212; they perform well on the post-test &#8212; without high storage strength. Weeks later, the retrieval strength decays, and the post-test performance turns out to have been measuring the wrong thing. Conditions that maximize immediate performance &#8212; massed practice, aligned testing, minimal difficulty &#8212; often actively impair long-term storage. This is the central insight of what Bjork calls desirable difficulties.</p><p>A learning claim grounded in Bjork&#8217;s construct requires evidence of storage strength, not just retrieval strength &#8212; which requires measuring performance after a delay, in new contexts, on items not identical to training. The methodology exists. It has existed since the early 1990s. It is the basis of essentially every recommendation in <em>Make It Stick</em> and in the broader spaced-practice and retrieval-practice literature that has accumulated since.</p><p>Now consider how <em>learning</em> is typically operationalized in EdTech efficacy research. The outcome measure is a post-test administered at the end of the instructional unit. The items are aligned with the instructional content. The interval between instruction and test is hours to days. The retrieval context is the same or similar to the learning context. What this operationalization measures is retrieval strength at short delay. What Bjork&#8217;s construct requires is storage strength at longer delay under different retrieval conditions. These are not the same thing.</p><p>The gap between the two is not subtle. It is structural. And it is present in nearly every efficacy claim this book has examined.</p><p>Consider <em>understanding</em>. Jean Lave, Etienne Wenger, John Dewey, and the situated-cognition tradition spent decades articulating understanding as something different from performance on items. Understanding involves the capacity to apply knowledge in contexts that differ from the contexts of acquisition. It involves participation in practices &#8212; knowing how to use what one knows in the world where it applies. Transfer testing &#8212; the capacity to apply learning to problems that differ meaningfully from training &#8212; is the minimum methodological requirement for a claim about understanding. Transfer testing has been advocated for in educational research since Thorndike&#8217;s early twentieth-century work. It remains exceptional in EdTech efficacy research.</p><p>Consider <em>mastery</em>. Bloom&#8217;s own construct, as articulated in his mastery-learning work, involves structural reorganization of knowledge &#8212; the kind of reorganization that allows a learner to solve problems the instruction did not specifically address. Bloom&#8217;s 2-sigma finding emerged from studies that implemented criterion-referenced assessment, formative assessment with corrective feedback, demonstrated performance across multiple item types. The 2-sigma number is cited routinely as a benchmark for tutoring effectiveness. Bloom&#8217;s construct of mastery, including its methodological requirements, is cited far less often.</p><p>Consider <em>personalization</em>, as examined in the eighth essai. The term invokes a construct rooted in Vygotskian zone-of-proximal-development work and the aptitude-treatment interaction literature &#8212; instruction responsive to who the individual learner actually is. What adaptive-learning systems operationalize is item sequencing and pacing based on item-level response patterns. These are not the same construct.</p><p>Consider <em>engagement</em>. The construct, as articulated in the psychological literature, involves attention, motivation, affect, persistence in the face of difficulty, meaningful cognitive investment. What AI-tutor efficacy research typically measures is time on task, session counts, and completion rates. Kristen DiCerbo of Khan Academy observed in April 2026 that when students engaged with Khanmigo, they were typing &#8220;IDK IDK&#8221; &#8212; <em>I don&#8217;t know, I don&#8217;t know</em> &#8212; and moving on. The platform counted them as engaged. They were not engaged in any cognitively meaningful sense.</p><p>Each of these constructs has serious theoretical articulation in one or more research traditions. Each is routinely invoked by the field&#8217;s claim-making vocabulary. Each is routinely operationalized as aligned-item performance at short timescale. The gap between the construct and the operationalization is what the apparatus produces. And taken across the field, it is the difference between the learning the vocabulary claims and the performance the measurements index.</p><div><hr></div><h2>What the Field Has Tried</h2><p>It would be inaccurate to say the field has not tried to close this gap. It has tried, across multiple traditions, for decades. That these attempts have not produced a different default apparatus is itself instructive.</p><p><em>How People Learn</em>, the 1999 National Academies synthesis by Bransford, Brown, and Cocking, made transfer testing a central methodological theme. The implication was straightforward: efficacy research should include transfer measures if it wants to make claims about learning rather than claims about trained performance. Two and a half decades later, transfer testing remains exceptional.</p><p>Samuel Messick&#8217;s theory of validity, codified in his 1989 chapter in <em>Educational Measurement</em>, specified that a test score&#8217;s interpretation requires examination of construct-relevant versus construct-irrelevant variance, construct underrepresentation, and the consequences of the test&#8217;s use. Applied rigorously, Messick&#8217;s framework would require EdTech efficacy research to examine what its outcome measures actually index rather than assuming that performance-on-aligned-items equals evidence-of-learning. The framework has been the theoretical standard in measurement theory for over thirty years. Its rigorous application in educational technology efficacy has been partial at best.</p><p>Jean Lave&#8217;s situated-cognition tradition articulated assessment that requires observation of practice rather than administration of tests. It has had essentially no impact on deployed-product efficacy research.</p><p>Each of these traditions has existed for decades. Each has produced methodology that could be adopted. Each remains exceptional rather than routine. The alternatives have not been hidden. They have been taught in graduate programs, cited in methods sections, present in the same journals that published the aligned-outcome studies.</p><p>The question is why they have not taken.</p><div><hr></div><h2>Why the Apparatus Persists</h2><p>The apparatus persists because it serves the specific production conditions of the field in which it operates.</p><p>Consider what a researcher needs to do research in this field. Funding, on grant cycles of two to five years. Publications, through peer-reviewed journals with specific conventions. Access to populations &#8212; schools, classrooms, platforms &#8212; through institutional partnerships with their own timelines and constraints. Findings that other researchers can cite.</p><p>Now consider what a more adequate apparatus would require. Transfer testing adds design complexity and reduces effect sizes. Durability testing extends the study timeline past the typical grant cycle. Multi-paradigm convergence requires methodological range that most research programs do not possess. Pre-registration of analytic plans constrains the exploratory moves that often produce publishable findings.</p><p>Each of these, if adopted as a default, would reduce the rate at which researchers produce citable positive findings. Not because the interventions do not work &#8212; some of them do &#8212; but because the findings that survive the more demanding methodology would be smaller, noisier, and less rhetorically useful. A researcher who adopts the more demanding methodology competes with researchers who do not. The less-demanding researcher&#8217;s findings will be larger, cleaner, and more citable. Grant agencies, tenure committees, and publication venues all reward the latter.</p><p>The same pressures operate on the institutions that surround the research. Product vendors have commercial reasons to prefer methodologies that produce larger numbers. Policy bodies have political reasons to prefer evidence that looks clean. Philanthropists want defensible findings, and clean findings are easier to defend than nuanced ones. Journal editors respond to what their referees will accept, and what referees will accept is shaped by the conventions the field has institutionalized.</p><p>No individual in this system is behaving cynically. Researchers are doing their best work under the constraints of their funding. The apparatus is not what anyone chose. It is what the incentives produce when rational actors operate within them.</p><p>This is why advocacy for better methodology has not produced better methodology. The problem is not that researchers do not know better methodology exists &#8212; they do. The problem is that operating under the existing apparatus produces careers; operating against it produces, for most researchers, shorter and more difficult careers.</p><p>The apparatus persists because it is an equilibrium. Equilibria are stable not because the actors inside them are irrational but because they are responding rationally to incentives that no single actor created and no single actor can change. Changing an equilibrium of this kind requires changing the incentives across grant agencies, tenure systems, journal conventions, institutional practices, and funder expectations simultaneously. Such coordination is rare.</p><p>This is a structural observation, not a moral one. Researchers in this field are not broken. The evidence base is what the apparatus produces when careful, rigorous, well-meaning researchers operate under the conventions the apparatus enforces. Improving any individual researcher&#8217;s methods would not change what the field&#8217;s evidence base looks like, because the evidence base is the aggregate output of many careful researchers responding to shared incentives.</p><div><hr></div><p>That is what the apparatus was always supposed to produce.</p><div><hr></div><p><em>Nik Bear Brown is Associate Teaching Professor of Computer Science and AI at Northeastern University and founder of Humanitarians AI (501(c)(3)). This essay appears as part of the Computational Skepticism series at <a href="https://skepticism.ai">skepticism.ai</a>. | <a href="https://theorist.ai">theorist.ai</a> | <a href="https://hypotheticalai.substack.com">hypotheticalai.substack.com</a></em></p><div><hr></div><p><strong>Tags:</strong> measurement construct validity EdTech efficacy, Bjork storage retrieval strength learning systems, transfer testing durability educational technology, apparatus equilibrium research incentives, Bloom Kestin aligned outcome measure gap</p>]]></content:encoded></item><item><title><![CDATA[The Comparison That Was Never Fair]]></title><description><![CDATA[What Intelligent Tutoring Systems Actually Measured, and What They Were Compared Against]]></description><link>https://www.skepticism.ai/p/the-comparison-that-was-never-fair</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-comparison-that-was-never-fair</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Tue, 21 Apr 2026 19:21:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!BLUQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BLUQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BLUQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!BLUQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!BLUQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!BLUQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BLUQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1626422,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/194834752?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BLUQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!BLUQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!BLUQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!BLUQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37bd9d3b-e6e4-4371-a664-178094eaa5c6_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In 2014, RAND published one of the most carefully designed evaluations of an educational technology system in the history of the field. John Pane, Beth Ann Griffin, Daniel McCaffrey, and Rita Karam ran a cluster-randomized controlled trial across 147 schools in seven states, assigning roughly 25,000 students either to use Cognitive Tutor Algebra I or to continue with whatever algebra instruction those schools had previously offered. The outcome measure was a standardized algebra proficiency exam. The design was, by the standards of a field that routinely tolerates thin evidence and motivated reporting, unusually rigorous.</p><p>The finding was specific. In the first year of implementation, Cognitive Tutor produced no statistically significant effect on algebra proficiency. In the second year, a significant positive effect emerged at high schools &#8212; approximately 0.20 standard deviations, sufficient to move a median student from the 50th to roughly the 58th percentile. At middle schools, the second-year effect was similar in magnitude but did not reach statistical significance.</p><p>Pane and colleagues called this an &#8220;implementation learning curve.&#8221; They were careful to note that the learning did not seem to happen at the level of individual teachers &#8212; students of teachers new to the system in year two performed similarly to students of experienced teachers. The learning happened at the level of schools: scheduling, infrastructure, coordination, institutional adjustment to a new instructional logic. The sites that figured out how to implement Cognitive Tutor took a year to figure it out, and then the system worked.</p><p>This is what a rigorous evaluation of an intelligent tutoring system looks like. The findings are real. The effects are modest. The implementation costs were substantial &#8212; approximately $97 per student per year for Cognitive Tutor against approximately $28 for the traditional textbook instruction it replaced. And in the field&#8217;s characteristic framing, this result was narrated as <em>disappointment</em>. Intelligent tutoring systems were supposed to approach human tutoring effectiveness. They had not.</p><p>I want to examine that disappointment. Not to redeem ITS, and not to dismiss the evaluation record. I want to examine what was being compared to what, and whether the comparison &#8212; the one that has driven ITS research, ITS funding, and now AI-tutor rhetoric for forty years &#8212; was ever structurally sound.</p><div><hr></div><h2>What the Tutor Actually Measured</h2><p>Cognitive Tutor was built to embody a specific theory of cognition. John Anderson&#8217;s ACT-R framework posits that skill acquisition is the conversion of declarative knowledge &#8212; facts, concepts &#8212; into procedural knowledge: production rules, condition-action pairs. To become skilled at algebra is to acquire a set of increasingly sophisticated rules for algebraic manipulation. Recognize that the goal is to isolate a variable and the coefficient is 4, and divide both sides by 4. The rule fires. The step is taken correctly.</p><p>The instructional design that follows from this is specific. If you can specify the production rules that constitute algebraic competence, you can build a system that monitors whether each rule is acquired. Cognitive Tutor did exactly this. As a student worked through a problem, the tutor compared each step against its internal model of valid solution paths. Correct step: proceed. Step matching a stored buggy production &#8212; a common misconception encoded in the system &#8212; respond with immediate feedback. Student requests help: deliver a graduated hint sequence targeting the specific production the student is struggling to fire.</p><p>Across many problems, the tutor maintained running Bayesian estimates of whether each production rule had been mastered. Students could not advance to new material until the estimates crossed a mastery threshold. This is model tracing and knowledge tracing: two technical operations that together constitute the system&#8217;s measurement apparatus. What the apparatus measures is step-level correctness, time per step, hint requests, error patterns, and estimated mastery of each production rule. These are not arbitrary choices. They are what ACT-R theory specifies as relevant to procedural skill acquisition. The design is internally consistent with the theory it was built on.</p><p>The 1995 paper in which Anderson, Corbett, Koedinger, and Pelletier published their decade of findings was titled <em>Cognitive Tutors: Lessons Learned</em>. The plural of lessons learned is deliberate. The paper names what the system does not measure with the same specificity as what it does. Cognitive Tutor does not model affective state. It cannot detect whether a student is frustrated, bored, or emotionally disengaged from the material. It cannot identify conceptual confusion that lives above the production-rule grain &#8212; a student may fire productions correctly while failing to understand the domain they are operating in, and the tutor will not notice. It does not measure transfer, durability, or motivation. These are not oversights. They are structural features of a system designed for a specific theoretical purpose.</p><p>The researchers knew exactly what they had built. The disappointment that followed was partly not theirs.</p><div><hr></div><h2>What Human Tutors Actually Do</h2><p>The comparison that generated the disappointment is this: ITS produces effect sizes of roughly 0.20 to 0.40 sigma relative to classroom instruction. Expert human tutors produce effect sizes of roughly 0.40 to 0.80 sigma. Therefore ITS has failed to approach human effectiveness.</p><p>This comparison requires that both numbers measure the same construct at different magnitudes. They do not.</p><p>The research literature on what expert human tutors actually do is not sparse, and much of it was produced by the same researchers who built ITS. Art Graesser &#8212; who built AutoTutor, one of the more sophisticated ITS systems in the research tradition &#8212; spent years analyzing videotaped sessions between expert tutors and students, specifically to understand what tutors were doing that his system might learn to do. What Graesser&#8217;s analyses documented was a specific set of interactional moves.</p><p>Tutors approach a topic with what Graesser called expectations and misconceptions: a mental model of the components of correct understanding and a map of how students typically go wrong. As students respond, the tutor evaluates the response against this map &#8212; not syntactically, as an ITS matches a step against a production rule, but semantically, tracking which elements of the expected understanding are present and which are missing. The next move is determined by this evaluation. The response is therefore flexible in a way that production-rule matching is not.</p><p>Tutors continuously check comprehension. &#8220;Can you say that in your own words?&#8221; &#8220;What would happen if this were different?&#8221; These are not assessment items; they are real questions that tutors use to calibrate what to do next. The comprehension check is an instrument for reading the student&#8217;s understanding, not recording it in a database.</p><p>Tutors manage affect. Graesser&#8217;s research documented that expert tutors are often deliberately imprecise about negative feedback &#8212; indirect, softened, delivered in ways designed to protect the student&#8217;s willingness to continue engaging. This is not sloppiness. It is the management of an ongoing relationship whose continuation matters to the learning. A student who has been made to feel consistently stupid by their tutor stops engaging, and a tutor who cannot detect or respond to that risk is a different kind of instrument.</p><p>Tutors follow student questions. When a student asks something the tutor had not planned to address, expert tutors engage. Graesser, describing AutoTutor&#8217;s limitations with characteristic directness, noted that his system had to use &#8220;diversionary tactics&#8221; when students asked questions outside its agenda. Human tutors do not divert. They follow.</p><p>Michelene Chi, working from a different angle, documented that what makes human tutoring effective is not primarily the information the tutor delivers. It is the interactivity &#8212; the tutor&#8217;s prompts that elicit the student&#8217;s own elaboration, the student&#8217;s attempts at articulation that reveal gaps, the tutor&#8217;s calibration of the next move to what the student&#8217;s specific response has revealed. Self-explanation is a primary driver of conceptual change, and expert tutors are specifically skilled at eliciting the right kind of self-explanation through well-calibrated prompts. An ITS can prompt for self-explanation. What it cannot do is read the specific partial answer the student just produced and respond to that answer&#8217;s specific weaknesses.</p><p>And from an even earlier lineage: Wood, Bruner, and Ross, in a foundational 1976 paper, identified six functions tutors perform when scaffolding learners through tasks. Recruitment of interest. Reduction of degrees of freedom. Direction maintenance. Marking critical features. Frustration control. Demonstration. Of these six, Cognitive Tutor was specifically engineered to perform one: reduction of degrees of freedom, the step-by-step scaffolding that makes a complex problem tractable by breaking it into smaller operations. The tutor is structurally blind to recruitment, structurally unable to perform frustration control, and limited in demonstration to displaying the system&#8217;s own solution paths rather than modeling the expert&#8217;s move for the novice in ways the novice can watch and internalize.</p><div><hr></div><h2>The Axis Problem</h2><p>Here is what this produces.</p><p>The ITS measurement apparatus was built to measure one specific dimension of what expert human tutors do: the reduction-of-degrees-of-freedom move. Cognitive Tutor performs this move with remarkable precision. Its model tracing, its knowledge tracing, its mastery-learning constraints &#8212; these are all optimized for ensuring students acquire the production rules that constitute procedural competence in a specific domain. When evaluated on measures aligned with this construct, the system produces real effects. Pane&#8217;s 0.20 sigma is not noise. It reflects what the system actually does.</p><p>Human tutoring, as documented in Graesser&#8217;s and Chi&#8217;s and Wood, Bruner, and Ross&#8217;s research, involves that same move alongside several others: expectation-and-misconception dialogue, comprehension checks, affective management, student-question handling, recruitment, frustration control, demonstration. The effect sizes produced by expert human tutors in the research literature reflect this fuller set of moves acting in concert, against whatever outcome measures the studies used.</p><p>When these two numbers &#8212; the ITS effect and the human-tutoring effect &#8212; are placed on a single sigma axis for comparison, the implicit claim is that they measure the same construct at different magnitudes. They do not. ITS measures what a procedural-scaffolding technology produces on assessments that test procedural skills. Human tutoring measures what a full interactional relationship produces on assessments that, depending on the study, test some combination of procedural skills and broader constructs. The numbers can be placed on the same axis only if the underlying outcome measures are the same &#8212; which they frequently are not &#8212; and only if the interactional moves the two interventions involve are comparable &#8212; which the research literature establishes they are not.</p><p>This is the construct mismatch. It is not a peripheral observation. It is the structural feature of a comparison that has been doing field-level work for forty years, driving research agendas, guiding institutional adoption decisions, and anchoring the contemporary rhetoric that AI can approach human instructional effectiveness. What the comparison has consistently obscured is that the two things it is comparing were never fully on the same axis.</p><p>Cognitive Tutor did something real, with discipline and theoretical grounding, and produced genuine effects when evaluated appropriately. The disappointment in its failure to match human-tutor effect sizes is partly the disappointment of a comparison that was underdetermined from the start. Asking whether Cognitive Tutor matched human tutors is like asking whether a skilled surgeon matches a general practitioner across all dimensions of medical care. The surgeon is extraordinarily good at the specific thing the surgeon does. The general practitioner does that thing and many others. The sigma gap between them does not mean the surgeon failed.</p><div><hr></div><h2>The Inheritance</h2><p>The current AI-tutor moment has been presented, in much public discourse, as an advance that finally addresses what ITS lacked. Large language models can engage in natural-language dialogue. They can handle questions they were not specifically designed to handle. They can, in principle, perform some of the interactional moves Graesser documented as characteristic of expert human tutoring &#8212; the expectation-and-misconception dialogue, the comprehension check, the flexible response to what a student actually said. The rhetoric suggests the construct mismatch has been resolved.</p><p>Read through the ITS apparatus, the claim is more complicated than the rhetoric suggests.</p><p>The current AI-tutor evaluation studies still measure what ITS evaluations measured: item-level mastery, step-level performance, post-test scores on aligned assessments, immediate outcomes rather than durable learning. The measurement apparatus has been inherited. What has changed is the interaction layer. Whether the interaction-layer changes produce meaningfully different learning outcomes &#8212; or produce the appearance of more-human interaction without producing the underlying effects &#8212; is an empirical question the current literature has not cleanly answered. The Kestin Harvard physics study, with its 0.73 to 1.3 sigma effects on researcher-designed tests of the specific content a two-hour AI session had just covered, is measured on a Skinnerian axis. The measurement does not index whether the AI performed the interactional moves that make human tutoring what it is. It indexes whether students correctly answered questions about surface tension and fluid flow immediately after being tutored about surface tension and fluid flow.</p><p>The construct mismatch is not solved by better interaction capabilities. It is solved by better measurement. A system that performs rich tutoring interaction and is evaluated on aligned immediate assessments remains, from the evaluation&#8217;s perspective, on the same axis as Cognitive Tutor. The measurement apparatus determines what the sigma numbers mean, and the measurement apparatus has not substantially changed across the transition from production-rule ITS to generative AI tutoring.</p><p>This matters because the comparison that has driven forty years of ITS disappointment is being recycled to drive the current AI-tutor moment. The benchmarks invoked &#8212; Bloom&#8217;s 2-sigma, the expert-human-tutor effect-size range, the framing that AI can now &#8220;approach&#8221; human instruction &#8212; are the same benchmarks. The construct mismatch they depend on is the same mismatch. Whether a system that generates flexible natural-language responses has actually closed the distance that matters, or has closed the part of the distance that is easier to perform while leaving the harder parts unaddressed, is the question the measurement apparatus is not yet equipped to answer.</p><div><hr></div><h2>Three Questions to Ask</h2><p>When you next encounter a claim that an educational technology has approached the effectiveness of human tutoring, three questions will orient you.</p><p>What did the technology actually measure? If the evaluation used item-level or step-level assessments aligned with the technology&#8217;s instructional content, the system has been measured against a construct aligned with what it was built to do. This is not a criticism; it is a description of what the evaluation supports.</p><p>What does the human-tutoring construct actually involve? The research literature on expert human tutors documents a specific set of interactional moves &#8212; expectation-and-misconception dialogue, comprehension checks, affective management, student-question handling, recruitment, frustration control, demonstration. These are not peripheral features. They are the substance of what expert tutors do.</p><p>Was the comparison conducted on an axis that indexes both? If the outcome measure favors procedural scaffolding &#8212; which most ITS and AI-tutor evaluations use &#8212; the axis is not measuring what human tutoring does beyond procedural scaffolding. The comparison is limited by the measurement choice. A finding that the technology approaches human tutoring on such a measure is a finding about procedural scaffolding, not about the interactional richness the construct human tutoring would require.</p><p>These questions do not answer whether AI can replace human tutors. They answer the prior question: what are we measuring when we make the comparison? The field has been skipping the prior question since 1984, when Benjamin Bloom placed his two-sigma number on the same axis as his classroom-instruction comparison and the discourse collapsed the distance between them into a single rhetorical invitation. Cognitive Tutor responded to the invitation seriously, with theoretical rigor and methodological discipline, and produced 0.20 sigma at high schools after a year of implementation and $97 per student per year of cost. That result is not a failure. It is what the move that Cognitive Tutor was designed to do produces, measured honestly, at scale, in actual schools.</p><p>The number that system was compared against was never on the same axis. The comparison is the problem. It was the problem in 1990, when ITS researchers were trying to build what it named. It is still the problem now, when generative AI is being asked to close a gap the measurement apparatus cannot fully see.</p><div><hr></div><p><em>Nik Bear Brown is Associate Teaching Professor of Computer Science and AI at Northeastern University and founder of Humanitarians AI (501(c)(3)). | <a href="https://skepticism.ai">skepticism.ai</a> | <a href="https://theorist.ai">theorist.ai</a></em></p><div><hr></div><p><strong>Tags:</strong> intelligent tutoring systems construct validity, Cognitive Tutor RAND evaluation, human tutoring comparison mismatch, ACT-R model tracing procedural scaffolding, AI tutor measurement apparatus critique</p>]]></content:encoded></item><item><title><![CDATA[The Inheritance We Never Examined]]></title><description><![CDATA[How Skinner&#8217;s Teaching Machine Still Grades Your Children&#8217;s Software]]></description><link>https://www.skepticism.ai/p/the-inheritance-we-never-examined</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-inheritance-we-never-examined</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Mon, 20 Apr 2026 18:42:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!zZrL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zZrL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zZrL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!zZrL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!zZrL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!zZrL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zZrL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1878068,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/194830729?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zZrL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!zZrL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!zZrL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!zZrL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3af82244-4500-483f-a5a5-7d30bf18480e_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>There is a machine in every classroom now, and it measures what it has always measured. The name on the box changes &#8212; Duolingo, Khanmigo, i-Ready, DreamBox &#8212; but what the box counts has remained, across seventy years of silicon and software and venture capital and neuroscience, almost perfectly stable. Accuracy per item. Time per response. Progression through atomized units. Performance on the test the system was built to prepare you for.</p><p>B.F. Skinner named these measurements in 1958. He had a good reason.</p><p>He had observed his daughter&#8217;s fourth-grade arithmetic class and been, in his own word, shocked. Students completed problems and waited. The papers were collected. Perhaps two days later, perhaps a week, the marked papers returned. By the time the feedback arrived, the behavior it was meant to reinforce had already moved on, taken up residence in some adjacent habit of mind that was no longer the one in need of correction. Skinner believed he understood the mechanism of learning better than anyone alive &#8212; the contingencies of reinforcement, the precise timing of feedback, the accumulation of correctly shaped behavior into competence &#8212; and what he had watched in that classroom was the systematic breaking of every mechanism he understood. A technology that could restore the contingencies, he reasoned, would be a technology that could teach.</p><p>His teaching machine presented material one frame at a time. The student responded. The machine verified, immediately, whether the response was correct. The contingencies were repaired.</p><p>What I am asking you to notice is not that this was wrong. I am asking you to notice what the machine measured &#8212; accuracy per frame, time per response, progression, error patterns &#8212; and to hold those measurements in mind as we trace them forward through sixty-six years of educational technology that kept the apparatus while abandoning almost everything else about Skinner&#8217;s framework.</p><div><hr></div><h2>What the Machine Could Not See</h2><p>The teaching machine could not look up from the immediate interaction to ask what the student would remember in six months.</p><p>This is not a glancing criticism of Skinner. His behavioral framework did not require him to ask the question; the question was not yet a question the field had organized itself to ask in the precise way that Bjork and Bjork&#8217;s subsequent research would demand. Skinner&#8217;s science was about the shaping of behavior through reinforcement, and a behavior that could be elicited at the moment of measurement had been shaped. That the behavior might dissolve in the absence of the reinforcing conditions was not, within behaviorism, a separate problem requiring separate measurement. Generalization was expected to follow naturally.</p><p>But this is where the inheritance turns costly. The assumption that immediate performance predicts durable learning was embedded in the measurement apparatus before it was tested empirically. By the time Robert and Elizabeth Bjork&#8217;s work made the distinction between retrieval strength and storage strength unavoidable &#8212; by the time it was clear, empirically, that the conditions maximizing immediate performance (massed practice, aligned testing, minimal difficulty) could actively impair long-term retention &#8212; the measurement apparatus had already been handed down through Patrick Suppes&#8217;s 1960s computer-assisted instruction and was settling into the bones of the field.</p><p>Suppes&#8217;s system at Stanford presented arithmetic problems to elementary students and recorded what Skinner&#8217;s machine had recorded: accuracy rates, response times, error patterns, progression. The technology shifted from mechanical device to mainframe computer. The measurements did not shift. Accuracy rose from 53 percent to over 90 percent. Response times fell from 630 seconds to 279. Suppes reported these numbers as evidence the system worked, and within the apparatus he had inherited, they were. He was not wrong to report them. He was working inside a set of choices about what evidence looked like that the apparatus had bequeathed him without flagging as choices.</p><p>The question of what those 90-percent-accurate students could do two years later was not asked.</p><div><hr></div><h2>The Apparatus Becomes Theory</h2><p>Here is what makes the inheritance pattern strange rather than simply historical: the apparatus persisted past the abandonment of the theoretical framework that had justified it.</p><p>John Anderson&#8217;s Cognitive Tutor, developed in the 1980s and 1990s at Carnegie Mellon, was built on ACT-R theory &#8212; a cognitive-psychological architecture that treated learning as the acquisition of production rules rather than the shaping of behavior. Theoretically, this was a departure from Skinner significant enough to constitute a revolution. The language of reinforcement was replaced by the language of cognition. The unit of analysis shifted from the frame to the production rule.</p><p>The measurement apparatus did not shift.</p><p>The Cognitive Tutor recorded step-level correctness &#8212; whether each student action matched one of the production rules the cognitive model identified as correct. It recorded time per step. It recorded hint requests, error patterns, estimated mastery of each production rule through Bayesian knowledge tracing. When Anderson and colleagues published their foundational 1995 paper in the <em>Journal of the Learning Sciences</em>, the evidence they offered that the system worked was: step-level accuracy, progression, and post-test performance on assessments aligned with the content the tutor had taught.</p><p>Skinner&#8217;s apparatus, operating at higher resolution, within a more sophisticated theoretical framework, carrying new vocabulary.</p><p>Anderson and colleagues were, I want to say this plainly, more honest about the limits of their measurements than most of the researchers who cited them. The 1995 paper notes explicitly that students &#8220;display transfer to the degree that they can map the tutor environment into the test environment&#8221; &#8212; an acknowledgment that the evidence of learning the system could produce depended on the degree to which the post-test resembled the tutor&#8217;s own format. This is the measurement-alignment problem stated with precision by the researchers who built the system it applied to. The acknowledgment was there. What happened subsequently was that the effect sizes from aligned post-tests entered the literature as if Anderson&#8217;s own caveat had not been published alongside them.</p><p>The apparatus inherits even what its originators flagged as provisional.</p><div><hr></div><h2>The Industrial Turn</h2><p>The 2010s commercial adaptive-learning era &#8212; Knewton, DreamBox, i-Ready, ALEKS &#8212; represents the point at which the inherited apparatus became an industry standard.</p><p>Knewton&#8217;s Jos&#233; Ferreira, during the 2012-2015 period of the platform&#8217;s public prominence, positioned his technology as capable of personalization so granular that it would transform education at scale. The claim invoked the Suppes promise in the language of twenty-first-century data science. What the platform actually measured was behavioral engagement data: which problems students attempted, which hints they took, how their patterns of interaction with the system correlated with eventual performance on the system&#8217;s own assessments. Independent efficacy research on Knewton was, during the period of its most expansive claims, notably absent. The apparatus was present in the measurement choices; the evidence was not.</p><p>DreamBox Learning, which earned more research attention than most adaptive platforms, became the subject of a 2016 Harvard Center for Education Policy Research study that found students at the median gained 1.4 to 3.9 percentile points on the NWEA MAP for approximately 7 to 8 hours of DreamBox usage. The researchers were transparent about a critical limitation: DreamBox usage might &#8220;partially reflect students&#8217; motivation levels,&#8221; meaning the correlation between usage and achievement might reflect that motivated students both use DreamBox more and learn more, independent of DreamBox&#8217;s instructional contribution. The acknowledgment, honest and specific, appeared in the paper. It rarely appeared in the citations that followed.</p><p>i-Ready produced a particularly clarifying version of the apparatus&#8217;s internal logic. The platform&#8217;s efficacy research typically demonstrated that students who achieved &#8220;usage fidelity&#8221; &#8212; meeting the system&#8217;s recommended weekly engagement minutes &#8212; showed higher scores on the i-Ready Diagnostic. The Diagnostic was itself calibrated to predict state test performance. A system measuring how well students learn to do well on the assessment the system provides, where the assessment was engineered to predict the external standard &#8212; this is the apparatus become recursive. The alignment between instruction and measurement, which Skinner had simply taken as a natural feature of teaching a student the specific behavior you then measured, had been engineered into the product design itself. The inheritance was now embedded in the commercial structure.</p><p>ALEKS routed the apparatus through Knowledge Space Theory, a mathematical framework for mapping curricular competencies that provided sophisticated theoretical grounding for the same fundamental measurement choices. Efficacy claims rested on performance within the system&#8217;s own knowledge mapping and on aligned post-tests that measured progression through the curricular content the system taught. The theoretical vocabulary was different from Skinner&#8217;s. The measurement choices were the same.</p><div><hr></div><h2>Duolingo, 2021</h2><p>I want to read a specific study carefully, because careful reading is the point.</p><p><em>Evaluating the reading and listening outcomes of beginning-level Duolingo courses</em>, by Xiangying Jiang, Joseph Rollinson, Luke Plonsky, Erin Gustafson, and Bozena Pajak, published in <em>Foreign Language Annals</em> in 2021. The fifth author, Plonsky, is an academic researcher at Northern Arizona University with specialization in applied linguistics. The other four were employed by Duolingo at the time of publication. The paper is peer-reviewed. It is cited in Duolingo&#8217;s own marketing materials. It is, within the conventions of the field, a careful study.</p><p>Two hundred and twenty-five adults in the United States &#8212; 135 studying Spanish, 90 studying French. Participants were required to have little to no prior proficiency in their target language, to be using Duolingo as their only learning tool, and &#8212; the consequential criterion &#8212; to have completed the beginning-level course content through Unit 4. The sample, the paper reports, skewed toward highly educated Caucasian Americans with bachelor&#8217;s or master&#8217;s degrees.</p><p>The outcome measure was the STAMP 4S test from Avant Assessment, covering reading and listening. Thirty multiple-choice items in each modality. The assessment was administered immediately after learners completed the beginning-level content.</p><p>The finding: Duolingo learners reached ACTFL Intermediate Low in reading and Novice High in listening &#8212; levels the paper characterizes as &#8220;comparable with those of university students at the end of the fourth semester&#8221; of college-level language study.</p><p>Now apply the apparatus.</p><p>The outcome measure is external &#8212; not designed by Duolingo, which is a genuine methodological improvement over purely internal assessment. But reading and listening are the specific modalities that Duolingo&#8217;s interface is engineered around. Multiple-choice comprehension items, translation tasks, listening exercises with multiple-choice responses: these are what Duolingo builds, and these are what the STAMP 4S measures. Speaking and writing &#8212; modalities that Duolingo&#8217;s app-based format supports weakly &#8212; are explicitly excluded from the study. The assessment is external. The choice of which aspects of language proficiency to measure is not.</p><p>The timescale: the post-test was administered immediately after course completion. There is no delayed assessment. Bjork&#8217;s distinction between retrieval strength and storage strength is directly relevant &#8212; the STAMP 4S scores reflect what Duolingo users can do at the moment they finish the course, not what they can do when they have been away from the app for six months. This question is not asked.</p><p>The population: only learners who completed the beginning-level content. Most Duolingo users do not. The platform&#8217;s attrition is substantial; most people who download the app never reach the end of the beginning-level material. The study measures the performance of survivors. What 100 people who finished the course achieved is a different finding from what 100 people who started it achieved. The paper is transparent about this selection. The subsequent framing of the findings &#8212; in the paper&#8217;s own conclusion and, more aggressively, in Duolingo&#8217;s marketing &#8212; as <em>Duolingo users reach Intermediate Low</em> does not preserve the completion-threshold restriction.</p><p>The baseline: a historical comparison. University students at the end of the fourth semester. There is no contemporaneous control group of comparable adults who spent equivalent time on a different learning approach. The two populations were measured in different conditions, at different times, possibly with different motivations and starting points. The <em>comparable to four semesters</em> claim treats them as if they had been measured equivalently.</p><p>The cost: not reported. Duolingo is free at its base tier, which is rhetorically powerful &#8212; free app comparable to paid college course &#8212; but the comparison elides the substantial time investment Duolingo users make. The paper does not ask what equivalent time investment in human-tutored instruction, structured self-study, or an immersive experience would produce. The cost denominator, which is constitutive of what a comparative claim actually supports, is absent.</p><p>I am not saying the study is dishonest. I am saying that each of these specific measurement choices &#8212; aligned-modality outcome, immediate timescale, survivor population, historical baseline, absent cost denominator &#8212; is traceable, in structure, to the apparatus Skinner initiated in 1958. The study is careful within conventions it has inherited. The conventions themselves are what require examination.</p><div><hr></div><h2>The Alternatives Have Always Existed</h2><p>This is what I want you to sit with: the apparatus did not persist in the absence of alternatives. It persisted alongside them.</p><p>Edward Thorndike established in 1906 and 1924 that improvement in one mental function rarely produces general improvement in others unless the two share identical elements. The methodological implication &#8212; that learning gains must be tested outside the conditions of the intervention, in contexts structurally different from training, to establish what the training actually produced &#8212; was available to the field for the entire history of educational technology. It has been occasionally adopted, routinely praised, and treated as aspirational rather than as the baseline standard that Thorndike&#8217;s own work suggested it should be.</p><p>The Bjorks&#8217; work on storage strength versus retrieval strength, canonical since the early 1990s, established empirically that the conditions maximizing immediate performance can impair durable retention. The specific implication &#8212; that a delayed post-test is required to distinguish performance from learning &#8212; has been in the learning sciences literature for over thirty years. Its adoption in educational technology efficacy research as standard practice has not happened.</p><p>Bransford, Brown, and Cocking&#8217;s <em>How People Learn</em>, the 1999 National Academies synthesis, argued explicitly that assessment should tap understanding rather than the ability to repeat facts. The argument was widely read, widely cited, and narrowly operationalized.</p><p>Samuel Messick&#8217;s theory of validity, developed across decades and codified in the 1989 <em>Educational Measurement</em> volume, specified that a test score&#8217;s interpretation requires examination of construct-relevant versus construct-irrelevant variance, construct underrepresentation, and the consequences of the test&#8217;s use. Applied rigorously, Messick&#8217;s framework would require educational technology efficacy research to examine what its outcome measures actually index rather than assuming that performance-on-aligned-items equals evidence-of-learning. The framework has been the theoretical standard in measurement theory for over thirty years.</p><p>These alternatives were not hidden. They were taught in graduate programs, cited in methods sections, present in the same journals that published the aligned-outcome studies. What did not happen, across six decades of technology change, was their adoption as the field&#8217;s measurement standard. The inherited apparatus &#8212; aligned outcomes at immediate timescale, survivor population, weak baseline, absent cost denominator &#8212; remained dominant. The alternatives remained alternative.</p><p>This is not a story about intellectual failure. It is a story about what happens when a theoretical commitment gets flattened into a methodological convention. Skinner had reasons for his measurement choices that were grounded in a coherent behavioral science. When the field moved past behavioral science &#8212; when Suppes and Anderson and everyone who followed adopted different theoretical frameworks &#8212; the measurement choices did not travel with the theory that had justified them. They traveled alone, as conventions, as what evidence looked like, as the unexamined default.</p><p>The apparatus became invisible by becoming obvious. And invisible apparatus is the most durable kind.</p><div><hr></div><h2>The Current Wave</h2><p>The contemporary AI-tutor literature &#8212; Khanmigo, Kestin and colleagues&#8217; 2024 Harvard physics study, Eedi with Google Research, Rori in Ghana &#8212; inherits the apparatus in its turn, with variation worth noting.</p><p>Khanmigo&#8217;s evaluation evidence has rested primarily on engagement metrics and performance within Khan Academy&#8217;s own internal assessment structures. What has been measured at scale is usage patterns; what has been claimed is educational transformation; what has not been established at the level of rigorous efficacy research is learning gains on independent standardized measures at delayed timescales with cost-inclusive reporting. The characteristic gaps of the apparatus are present.</p><p>The Kestin et al. 2024 Harvard physics study &#8212; AI-tutored instruction versus a single session of active-learning classroom instruction &#8212; reported effect sizes of 0.73 to 1.3 sigma on researcher-designed post-tests covering surface tension and fluid flow, the specific content the two-hour intervention taught, assessed shortly after the intervention. The measurement choices are the apparatus&#8217;s measurement choices. The effect sizes are real within those choices. What they establish about learning is bounded by what those choices can establish.</p><p>Eedi with Google Research 2025 introduced transfer testing &#8212; measuring performance on novel problems from subsequent topics rather than problems aligned with what the intervention taught. This is a genuine departure from the inherited convention. The N of 165 and single-term duration remain short relative to what durability research would require, but the outcome measure itself represents the kind of revision the apparatus needs rather than another inheritance of it. This is a credit to the researchers who chose to build the study that way.</p><p>Rori in Ghana used an external curriculum-aligned assessment over eight months and reported cost at $5 per student per year. The longer timescale, the external measure, the explicit cost denominator &#8212; these are partial revisions of the apparatus in the direction the field has needed for six decades. The pattern is: when researchers choose to work against the inherited conventions, the field moves. The field moves rarely, because the inherited conventions are the default, because departures from them require additional effort and often smaller effect sizes and sometimes no significant effect at all, which is a kind of finding that is harder to publish than 0.73 sigma.</p><p>The apparatus has not been reformed. It has been revised in specific instances by specific researchers. The instances are the exceptions that make the pattern visible.</p><div><hr></div><div><hr></div><p><em>Nik Bear Brown is Associate Teaching Professor of Computer Science and AI at Northeastern University and founder of Humanitarians AI (501(c)(3)). His research on educational AI efficacy appears at <a href="https://hypotheticalai.substack.com">hypotheticalai.substack.com</a>. | <a href="https://skepticism.ai">skepticism.ai</a> | <a href="https://theorist.ai">theorist.ai</a></em></p><div><hr></div><p><strong>Tags:</strong> educational technology measurement apparatus, Skinner teaching machine inheritance, Duolingo efficacy research critique, aligned outcome EdTech validity, learning science transfer testing history</p>]]></content:encoded></item><item><title><![CDATA[The Artifact Was Once Enough]]></title><description><![CDATA[This essay is a response to Lila Shroff's "Is Schoolwork Optional Now?" published in The Atlantic on April 10, 2026.]]></description><link>https://www.skepticism.ai/p/the-artifact-was-once-enough</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-artifact-was-once-enough</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Sat, 11 Apr 2026 04:47:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6EVu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6EVu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6EVu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png 424w, https://substackcdn.com/image/fetch/$s_!6EVu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png 848w, https://substackcdn.com/image/fetch/$s_!6EVu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png 1272w, https://substackcdn.com/image/fetch/$s_!6EVu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6EVu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png" width="1456" height="803" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:803,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3143432,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/193858776?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6EVu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png 424w, https://substackcdn.com/image/fetch/$s_!6EVu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png 848w, https://substackcdn.com/image/fetch/$s_!6EVu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png 1272w, https://substackcdn.com/image/fetch/$s_!6EVu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff61c4d77-310b-4950-b6ad-d209533eb3c3_3146x1734.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>This essay is a response to Lila Shroff&#8217;s &#8220;<a href="https://www.theatlantic.com/technology/2026/04/ai-agents-school-education/686754/">Is Schoolwork Optional Now?</a>&#8220; published in The Atlantic on April 10, 2026. The argument it makes in full is developed in the preprint &#8220;<a href="https://www.nikbearbrown.com/notes/Frictional/frictional">Frictional: Measuring the Struggle</a>&#8220; at <a href="https://www.irreducibly.xyz/">irreducibly.xyz</a>.</em></p><div><hr></div><p>There is a word &#8212; <em>decoupling</em> &#8212; that sounds technical enough to keep us comfortable. Clinical. As if what has happened in classrooms since 2022 is primarily a logistics problem, a puzzle about detection and enforcement, a cat-and-mouse game that the right algorithm might someday win.</p><p>It is not that.</p><p>What has happened is something more fundamental than cheating at scale. The artifact &#8212; the essay, the proof, the lab report &#8212; used to be evidence of a process. The process was the point. The essay was proof that thinking had occurred, that a mind had engaged with difficulty and emerged changed. When we graded the essay, we were really grading the encounter: the hours of confusion, the drafts that failed, the moment when something clicked and then had to be organized into sentences for another person. The artifact was the residue of all that. It was upstream evidence of downstream consequence.</p><p>Generative AI has broken the causal chain. Not bent it &#8212; broken it.</p><p>A bot called Einstein, built by a 22-year-old entrepreneur named Advait Paliwal, recently completed all eight modules and seven quizzes of an introductory statistics course in under an hour. Perfect score. The human who set it loose reports that she &#8220;hardly so much as read the course website.&#8221; What Einstein produced &#8212; the evidence that a course had been completed &#8212; was real. The learning it was supposed to represent did not occur. The artifact existed. The process that should have produced it did not happen.</p><p>Paliwal says he released the tool to alert educators. His more honest statement is buried in the subtext: &#8220;If I didn&#8217;t post about this, someone would have used the same technology and hidden it from the professors.&#8221; He is right. He is also describing a world in which the distinction between using it secretly and not using it at all is narrowing toward irrelevance. The tool exists. The temptation exists. The economic pressure on students &#8212; especially international students, especially students working jobs to pay tuition, especially students in courses they are taking to satisfy requirements rather than from genuine interest &#8212; those pressures exist independently of any single tool.</p><p>The institutional response has been to build better detectors. This is a reasonable first move. It is not a durable one.</p><div><hr></div><h2>Why Detection Cannot Save Us</h2><p>Here is the structural problem with artifact-based AI detection: the arms race has a predetermined winner. Detection is always trained on the outputs of current generation technology. Generation technology improves continuously. The detector trained on today&#8217;s AI writing fails on tomorrow&#8217;s &#8212; not because detectors are poorly built, but because that is how the mathematics of the problem works. The forensic window closes.</p><p>There is a deeper problem. The educationally relevant question was never <em>did a human type these words</em>. It was <em>did a human develop this understanding</em>. A student who dictated an essay to a transcriptionist and then submitted it word-for-word would have technically written no AI content. The essay would pass every detector. The learning would have occurred or not occurred based on whether they thought hard while dictating, not based on who typed it. The detector is solving the wrong problem.</p><p>And there is a third problem, the one that produces the most corrosive outcomes. When you build a system to catch AI use, you teach students to game the detector. They learn strategies for mimicking authentic writing &#8212; inserting typos, varying sentence structure, using phrases the model knows sound &#8220;human.&#8221; The simulation improves. The gap between simulated engagement and genuine engagement widens at precisely the moment we need it to narrow.</p><p>William Liu, a Stanford sophomore who finished high school two years ago, puts it plainly: his educational experience and his younger sibling&#8217;s are vastly different despite a two-year gap. The technology arrived. The classroom has not yet figured out what to do next.</p><div><hr></div><h2>What Genuine Learning Actually Leaves Behind</h2><p>Here is the thing we have been too polite to say: learning is not the same as performance.</p><p>Robert Bjork has been saying this for thirty years in academic papers that educators read and administrators do not read and curriculum designers read and then ignore when the calendar pressure comes. Performance is the observable, often temporary thing &#8212; how well a student does on a measure. Learning is the durable change in what the student can do and understand and transfer to a new context. These two things are not the same. We have built an entire institutional infrastructure that measures only one of them.</p><p>Genuine human learning is a biological event. When a learner encounters material that genuinely challenges their current understanding &#8212; material in that productive zone where their current model is wrong or incomplete &#8212; something specific happens neurologically. Dopamine neurons fire in response to prediction errors. BDNF expression upregulates, sometimes by nearly three times. New dendritic spines form at the synaptic connections that will hold the memory. These are not metaphors. They are the physical substrate of the thing we call learning.</p><p>The behavioral consequences of these neurological events are traceable. A student engaged in genuine cognitive struggle spends time proportional to difficulty. Their errors follow a coherent developmental path &#8212; misconceptions that make sense given their current model, corrections that build on each other. When tested in a new context, they can transfer. When scaffolded with a partial hint, they respond &#8212; because there is a partially formed structure for the hint to connect to. Their confidence, over time, calibrates to their actual performance rather than inheriting the confidence of the AI explanation they processed.</p><p>These are what I have been calling <em>friction traces</em> &#8212; the behavioral signatures that genuine human cognitive engagement leaves in observable data. They exist because genuine learning is a biological event. An AI can produce the artifact without triggering any of these neurological events. It cannot produce the behavioral traces, because the biological events that generate those traces did not occur.</p><div><hr></div><h2>The Seven Things We Can Now Measure</h2><p>The Genuine Learning Probability framework I have been developing with Humanitarians AI specifies seven such traces:</p><p>The <em>temporal engagement pattern</em> &#8212; the correlation between how hard an item is and how long a student spends on it. Genuine engagement produces this correlation. AI-assisted completion decouples time from difficulty.</p><p>The <em>error trajectory</em> &#8212; whether a student&#8217;s mistakes follow conceptually coherent developmental paths. Genuine learning produces coherent errors; the reward prediction error mechanism drives the model toward better models in patterned ways. Borrowed certainty produces random errors with respect to conceptual structure.</p><p><em>Cross-context transfer</em> &#8212; the Bjorkian definition of learning. A student who genuinely understood something can apply it in novel contexts. Borrowed certainty produces surface representations tied to the specific context of the AI explanation.</p><p><em>Uncertainty calibration</em> &#8212; whether a student&#8217;s expressed confidence tracks their actual performance. Borrowed certainty produces systematic overconfidence: the student inherits the AI&#8217;s confidence distribution without the knowledge base that would justify it.</p><p><em>Social knowledge texture</em> &#8212; the quality of a student&#8217;s engagement in discussion contexts. Genuine encounter with material leaves a characteristic texture: specific confusions, particular connections, the specific questions that arose from actual engagement. This texture cannot be manufactured without having had the encounter.</p><p>The <em>retrieval strength decay signature</em> &#8212; whether performance decays at rates consistent with genuine encoding. The spacing effect is the benchmark of genuine learning. Borrowed certainty has no storage strength to retrieve; performance decays monotonically and the spacing effect does not appear.</p><p>And the <em>scaffolding response curve</em> &#8212; whether a student&#8217;s performance responds appropriately to partial hints. A student with genuine partial understanding has a zone of proximal development. A partial hint activates the structure that is already forming. Borrowed certainty has no such zone.</p><div><hr></div><h2>What the Bot Cannot Manufacture</h2><p>Here is the argument I want to make carefully, because it is often misunderstood: this framework is not about catching AI use. It is about measuring learning directly.</p><p>An AI detector fails when AI outputs become indistinguishable from human outputs. A learning measure fails when borrowed certainty becomes indistinguishable from genuine learning &#8212; which would require borrowed certainty to produce the same neurobiological events, the same schema formation, the same durable transfer. At that point, borrowed certainty has become learning. That is not AI defeating assessment. That is learning occurring through a different pathway than we expected.</p><p>What manufacturing all seven friction traces simultaneously &#8212; without performing the underlying cognitive work &#8212; actually requires is something close to performing the underlying cognitive work. A student who spends genuine time on difficult material, who makes and corrects errors in a conceptually coherent sequence, who demonstrates transfer across novel contexts, who maintains calibrated uncertainty, who engages with genuine texture in discussion, who shows the spacing effect across weeks, and who responds appropriately to partial hints &#8212; has learned the material. At that point the game has become indistinguishable from the thing we wanted in the first place.</p><p>Natalie Lahr, a Barnard sophomore studying history and political science, describes an &#8220;anti-AI radicalizing&#8221; experience: a tutor at the writing center pasted her essay prompt into Perplexity and handed her the AI-generated outline. &#8220;Why am I even here?&#8221; she asked afterward. The question is not rhetorical. It is the correct question.</p><div><hr></div><h2>What We Must Build Instead</h2><p>The crisis of evidence facing educational institutions is not a technical problem. It is an epistemological problem. The evidence infrastructure we built assumed a world in which the artifact was upstream evidence of the process. That world no longer reliably exists.</p><p>What we need is an assessment infrastructure built on the process itself.</p><p>This means longitudinal process documentation &#8212; portfolios that capture the history of engagement, not just its products. It means embedded formative assessment that generates the data necessary to observe the seven friction traces over time. It means treating developmental trajectory as evidence: not what a student produced, but how their understanding developed, what they got wrong and corrected and why, where they transferred and where they didn&#8217;t.</p><p>Marc Watkins at the University of Mississippi describes an instructor who could, theoretically, set an AI to grade thirty essays during a fifteen-minute walk to Starbucks. He calls this &#8220;really scary.&#8221; He is right, but I want to be precise about why. The fear is not the efficiency. It is the loop: AI-generated assignments completed and assessed by AI agents, with human understanding nowhere in the chain. The fully automated loop is not a future dystopia. It is the logical endpoint of current trajectories. Einstein completes the course. The grader grades Einstein&#8217;s work. Both certificate and grade are real. The learning did not occur.</p><p>The artifact was once enough. It is no longer enough. The arms race between generation and detection has a winner, and it is not the detector.</p><p>We must now measure the struggle itself. Not because friction is intrinsically valuable &#8212; productive struggle matters only because of what it builds in the brain that does the struggling. We must measure it because the brain that struggles is the brain that learns, and the brain that learns is the only thing education was ever actually for.</p><p>The methodology is developed in full in &#8220;<a href="https://www.nikbearbrown.com/notes/Frictional/frictional">Frictional: Measuring the Struggle</a>&#8220; &#8212; a preprint specifying the seven friction components, the ensemble architecture, and the tier calibration system &#8212; and at <a href="https://www.irreducibly.xyz/">irreducibly.xyz</a>. The framework is not a secret.</p><div><hr></div><p><em>Nik Bear Brown is Associate Teaching Professor of Computer Science and AI at Northeastern University and founder of Humanitarians AI (501(c)(3)).</em><br><em>bear.musinique.com &#183; skepticism.ai &#183; theorist.ai</em></p><div><hr></div><p><strong>Tags:</strong> AI detection education failure, genuine learning probability framework, friction traces assessment, Bjork performance vs learning, Einstein bot Canvas schoolwork automation</p>]]></content:encoded></item><item><title><![CDATA[Brutalist.art - The "Beautiful.ai" that Educators Need]]></title><description><![CDATA[Talking to a slide deck through Claude code]]></description><link>https://www.skepticism.ai/p/brutalistart-the-beautifulai-that</link><guid isPermaLink="false">https://www.skepticism.ai/p/brutalistart-the-beautifulai-that</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Mon, 06 Apr 2026 00:49:13 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/193305176/36365434e9ead489eec0094e88101873.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<h3><strong>The Slide Deck You Built Was Not for the Learner</strong></h3><h3><strong>It Was for You</strong></h3><p>There is a lie at the center of most educational content production, and it goes mostly unnamed because naming it is professionally uncomfortable. The lie is this: the slide deck you built last Tuesday, the one you spent three hours arranging, the one with the custom fonts and the carefully chosen images and the thirty-seven bullets across fourteen slides &#8212; that deck was not built for the people who had to sit through it. It was built for you. It was built so you could feel the relief of having covered the material. It was built so the topic had a container. It was built because you had a deadline and a template and a vague professional obligation to produce <em>something</em>, and a slide deck is always <em>something</em>.</p><p>The learner &#8212; the specific human being with specific prior knowledge and a specific amount of time and a specific gap between what they currently understand and what they need to understand &#8212; that person never really entered the room where the deck was being built. What entered the room instead was a topic. And a topic is not a person.</p><p>Brutalist was built to address this. Not to address it gently, with suggestions and style guides and best-practice checklists. To address it structurally, in the architecture of the tool itself, before a single slide gets made.</p><h3><strong>The Architecture of Avoidance</strong></h3><p>The conventional workflow for building educational content runs roughly like this: you receive a topic (or assign yourself one), you collect material &#8212; readings, notes, data, existing slides &#8212; and you begin arranging it. If you are experienced, you arrange it with craft. You think about sequence and pacing. You choose examples. You know when to deploy a metaphor and when to let a statistic land without ornamentation. The result, at its best, is a coherent and well-paced presentation of material.</p><p>What you have not done &#8212; and this is the gap that produces most failures in educational content &#8212; is started from what the learner will be able to <em>do</em> when you are finished with them. You have started from what you know, and you have worked forward through that knowledge toward a clean ending. This is a completely understandable approach, and it produces content that would be unrecognizable as failing by any ordinary standard of review. It is organized. It is clear. It covers the material.</p><p>It just doesn&#8217;t reliably produce learning.</p><p>Backwards design &#8212; the pedagogical framework that governs every output Brutalist produces &#8212; insists on reversing this sequence. You begin with a measurable outcome: not a topic, not a list of things the instructor will present, but a single sentence describing what a learner will be able to <em>do</em> at the end that they could not do at the beginning. Construct a DAG from domain knowledge and identify all backdoor paths. Distinguish between a learning outcome and a topic. Evaluate a rubric for the difference between qualitative descriptions and observable behaviors. These are not aspirations. They are commitments &#8212; to a learner, to a measurable change, to the possibility of knowing whether the teaching worked.</p><p>The reason most content production doesn&#8217;t begin here is not ignorance. Most instructors know what backwards design is. The reason is that starting from a learning outcome is harder than starting from a topic, and the tools available for producing educational content &#8212; PowerPoint, Keynote, Google Slides &#8212; offer no friction whatsoever against starting from the wrong place. They are indifferent to the question of who the learner is and what the learner needs to be able to do. They are happy to help you arrange forty slides around a topic, and they will never once ask whether the arrangement serves a learner or just a speaker.</p><p>Brutalist asks. It asks before it produces anything. In interactive mode &#8212; the default &#8212; it will not generate a single slide until it has confirmed the audience, confirmed the outcome, and confirmed that the outcome is measurable. &#8220;Understand X&#8221; is not measurable. Brutalist says so, explicitly, in the voice of a pedagogical skeptic rather than a customer-service chatbot. <em>That describes a mental state, not a behavior. A learner can&#8217;t demonstrate &#8216;understanding.&#8217; What&#8217;s the one thing they should be able to do?</em> This is not rudeness. It is the one question that changes the output.</p><h3><strong>The Phase Gate as Moral Commitment</strong></h3><p>There is a design decision embedded in Brutalist that deserves more attention than it usually gets in conversations about AI tools, which tend to focus on capability rather than constraint. That decision is the phase gate.</p><p>A phase gate is exactly what it sounds like: a gate that holds until a phase is complete. In Brutalist, the first gate holds at source confirmation &#8212; no output until the source material is present. The second holds at outcome identification &#8212; no output until the outcome can be stated in one sentence. The third holds at form confirmation &#8212; no output until the right command for the content is confirmed. Only then does the tool produce anything.</p><p>This is unusual. Most AI tools are designed to produce output as quickly as possible, because output is what users think they want and user satisfaction is what tools are optimized for. The experience of receiving forty slides in thirty seconds feels like productivity. It feels like the machine is working for you. What it actually is, much of the time, is the machine generating plausible-looking content that fills the form without serving the function &#8212; decoration rather than argument, coverage rather than learning.</p><p>Brutalist is optimized for the learner, not the user. These are not the same person. The user is the instructor who wants a slide deck. The learner is the person who will sit in front of that deck and try to change what they understand. Optimizing for the user produces faster output. Optimizing for the learner produces harder questions before any output is generated at all.</p><p>The phase gate is where this optimization manifests in the tool&#8217;s behavior. It is the structural embodiment of a moral position: that output built on wrong assumptions about audience or outcome wastes more time than the intake that would have caught those assumptions. Two minutes of friction before the deck is built is less costly than an hour of instruction that doesn&#8217;t change what anyone understands.</p><h3><strong>What &#8220;Understand X&#8221; Is Actually Doing</strong></h3><p>Spend any time in educational settings &#8212; as a student, as an instructor, as a curriculum designer &#8212; and you develop a particular sensitivity to the phrase &#8220;by the end of this, students will understand X.&#8221; It appears in syllabi, in lesson plans, in course descriptions, in accreditation documents. It appears so frequently and so unexamined that most people who write it have stopped noticing it at all. It is pedagogical wallpaper.</p><p>But the phrase is doing something specific, and it is worth naming. &#8220;Students will understand X&#8221; is a sentence that sounds like a learning outcome and functions as an escape from accountability. Understanding is a mental state. You cannot observe it, you cannot measure it, you cannot score it on a rubric or assess it in a portfolio. You can ask someone to demonstrate understanding &#8212; which means you are no longer assessing understanding, you are assessing a behavior &#8212; but the phrase as written commits you to nothing. It is a promise with no deliverable attached.</p><p>The reason this matters to a tool like Brutalist is that the learning outcome is not just the first step in backwards design. It is the specification for everything that follows. The slides that get built, the visual types that get selected, the checks for understanding that get inserted every four to six slides &#8212; these are all derived from the outcome, working backward from what the learner needs to be able to do. If the outcome is vague, the derivation has nothing to anchor to. The result is a deck that covers material in the general direction of a topic, which is not the same thing as a deck that moves a specific learner from a specific gap to a specific capability.</p><p>This is why Brutalist treats &#8220;understand X&#8221; not as a minor stylistic imprecision but as a structural failure that must be corrected before building anything. The outcome is the foundation. A vague foundation does not produce a stable structure. It produces decoration.</p><h3><strong>Brutalist HTML and the Question of Deployment</strong></h3><p>There is a second commitment embedded in this tool that is worth examining, and it lives in the signature output: the brutalist HTML presentation. Not a PowerPoint file. Not a PDF. A single self-contained HTML file, deployable immediately, built on a design system called Musinique brutalist &#8212; JetBrains Mono, parchment tokens, per-slide audio, keyboard navigation, zero decorative radius.</p><p>The choice of HTML as the primary output format is not aesthetic. It is pedagogical and practical simultaneously. A PowerPoint file requires PowerPoint. A Google Slides file requires Google. An HTML file requires a browser, which is to say it requires nothing &#8212; it deploys anywhere, runs without software dependencies, and can be shared as a URL or a file with equal ease. The friction of tool access is a real barrier to distribution, and distribution is where educational content either serves learners or stops serving them.</p><p>The design choices embedded in the brutalist system &#8212; every slide does one thing, every title is a claim not a topic, components are typed by what they communicate rather than how they look &#8212; these are cognitive load principles encoded as aesthetic constraints. The slide with a hero number and a two-line muted caption exists because research on split attention and redundancy effects has things to say about how visual and verbal information compete for working memory. The check for understanding every four to six slides exists because spaced retrieval practice produces stronger retention than massed coverage. The design is not decoration. It is applied cognitive science, translated into a component library and a phase-gated workflow.</p><h3><strong>The Pushback Layer</strong></h3><p>Brutalist pushes back. This is the part of the tool that most users encounter with some surprise, because tools &#8212; especially AI tools &#8212; are generally not in the business of disagreement. They are in the business of helpfulness, and helpfulness has been operationally defined as producing what the user asks for as quickly as possible. Friction is a UX failure. Pushback is an anomaly.</p><p>In Brutalist, pushback is a feature. Not an accident of the model&#8217;s personality or a quirk of the prompting, but a designed behavior with specific triggers and specific exit conditions. Weak learning outcomes get flagged &#8212; not once, politely, but persistently, with an offer to rewrite the outcome if the user fails the measurability test twice. Vague audience descriptions get challenged, because &#8220;college students&#8221; is not an audience and the specificity that changes the content, examples, and pacing cannot be inferred from it. Mismatched command choices get named &#8212; if the content calls for a <code>/showtell</code> and the user has requested <code>/slides</code>, the tool explains the difference in instructional design terms before proceeding.</p><p>Every pushback ends with a path forward. This is the moral discipline that separates useful friction from obstruction. The tool is not in the business of refusing to build. It is in the business of building toward the right specification, and the right specification cannot be assumed from the wrong brief. The pushback is the tool asking the question that the instructor should have asked before they opened a blank deck and started arranging.</p><p>What is the learner supposed to be able to do?</p><p>Everything else follows from that.</p><div><hr></div><p><em>Brutalist is part of the Humanitarians AI Ecosystem. The primary workflow: </em><code>/slides</code> produces the blueprint. <code>/brutalist</code> converts it to HTML. <code>/deck</code> does both in one command. Type <code>help</code> to begin.</p><p><strong>Tags:</strong> Brutalist instructional design engine, backwards design pedagogy, learning outcomes Bloom&#8217;s taxonomy, brutalist HTML presentation system, educational content production failure</p>]]></content:encoded></item><item><title><![CDATA[The Struggle Is the Point]]></title><description><![CDATA[What We Lost When We Made the Artifact the Grade]]></description><link>https://www.skepticism.ai/p/the-struggle-is-the-point</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-struggle-is-the-point</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Sat, 04 Apr 2026 03:35:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!SDu5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SDu5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SDu5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png 424w, https://substackcdn.com/image/fetch/$s_!SDu5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png 848w, https://substackcdn.com/image/fetch/$s_!SDu5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png 1272w, https://substackcdn.com/image/fetch/$s_!SDu5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SDu5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png" width="1456" height="543" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:543,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:167415,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/193135422?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SDu5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png 424w, https://substackcdn.com/image/fetch/$s_!SDu5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png 848w, https://substackcdn.com/image/fetch/$s_!SDu5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png 1272w, https://substackcdn.com/image/fetch/$s_!SDu5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8af9b2b9-9f0e-46ad-aff1-94d64d45472e_1886x704.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The paper rough draft: <strong><a href="https://www.nikbearbrown.com/notes/Papers/glp-framework-genuine-learning-probability">https://www.nikbearbrown.com/notes/Papers/glp-framework-genuine-learning-probability</a></strong></p><h2>What We Lost When We Made the Artifact the Grade</h2><p>Here is the situation as it actually exists, not as anyone in an official capacity is willing to describe it clearly.</p><p>A student sits down to write a paper. The paper is due in twelve hours. The student has three other assignments due this week, a job that starts at six, and the accumulated evidence of two semesters telling them that the grade lives in the artifact &#8212; the paper itself &#8212; not in the thinking that was supposed to produce it. The student opens an AI tool. The paper gets written. It is, by most measurable standards, better than what the student would have produced alone at midnight after a shift.</p><p>In the next building, the professor who assigned the paper has used AI to draft the assignment prompt, the rubric, and the feedback comments they will paste into the LMS after running the submitted papers through a grading interface that summarizes them automatically.</p><p>Neither of them is a villain. Both of them are responding rationally to a system that has always rewarded the artifact and never found a way to measure the process that was supposed to produce it. Generative AI did not create this problem. It revealed it &#8212; suddenly, completely, and without the courtesy of suggesting a solution.</p><p>This essay is about what the solution might look like. It is not technical. The technical apparatus exists and is documented elsewhere. What doesn&#8217;t exist yet, in language plain enough to be useful, is a way of talking about why the solution matters &#8212; what it would mean for a student to be seen by an educational system that has, for most of institutional history, been looking at the wrong thing.</p><h2>What the Artifact Was Supposed to Prove</h2><p>The essay, the exam, the project, the recorded performance &#8212; these were never the thing education cared about. They were evidence. The artifact was valuable because it was causally downstream of a process: the reading, the confusion, the rereading, the argument with yourself at two in the morning about whether you actually understood what you thought you understood. The artifact was a trace of that process. Grading the artifact was a way of inferring the process, because the two were coupled tightly enough that measuring one was effectively measuring both.</p><p>That coupling has broken. This is not a scandal or a failure or a temporary condition that better AI detection will resolve. It is a structural change in what artifacts can tell us, and it is permanent. The forensic window &#8212; the period during which you can reliably distinguish a human-written essay from an AI-generated one &#8212; is closing sequentially across every domain in which humans produce artifacts. In writing it is largely closed already. In code it is closing. The detectors trained on today&#8217;s AI outputs will be obsolete when tomorrow&#8217;s outputs arrive.</p><p>Every educational institution that is currently responding to this situation by installing better detection software is solving last year&#8217;s problem with next year&#8217;s obsolescence already scheduled.</p><h2>The Complicity No One Names</h2><p>The conversation about AI and academic integrity is almost entirely conducted as a conversation about student dishonesty. This framing is not wrong, exactly. It is just so incomplete as to function as a kind of dishonesty itself.</p><p>Students are using AI because the artifact is the grade. The artifact is the grade because grading the process &#8212; the confusion, the revision, the dead ends, the moments of genuine understanding &#8212; is hard, and institutions have never built the infrastructure to do it at scale. The result is a system that has always been measuring the wrong thing, and now the wrong thing can be produced in thirty seconds by a tool that costs less than a textbook.</p><p>Professors are not innocent bystanders. Many are using the same tools to manage the same impossible workloads &#8212; drafting prompts, generating feedback, summarizing submissions &#8212; that the institution&#8217;s growth model has made unmanageable. The incentive structure reaches all the way up. Publish or perish does not reward good teaching. Good teaching does not require good teaching to be measurable, only for its artifacts &#8212; syllabi, course evaluations, enrollment numbers &#8212; to look like good teaching.</p><p>The student who uses AI to write a paper is not defecting from a system that is working. They are defecting from a system that has always asked them to perform learning rather than do it, and has never been able to tell the difference. AI has not corrupted that system. AI has made the corruption visible.</p><p>This is the thing worth sitting with before any solution is proposed: the problem is not the tools. The problem is what we decided to measure, and what we decided to ignore, long before the tools arrived.</p><h2>What Genuine Learning Leaves Behind</h2><p>Here is what the research shows, stated plainly.</p><p>When a human being genuinely learns something hard, the process is biological. Neurons fire in response to the gap between what the learner expected and what they encountered. That gap &#8212; the prediction error &#8212; is uncomfortable. It is the feeling of not understanding, the specific texture of confusion that is different from ignorance because it knows what it doesn&#8217;t know. Working through that discomfort produces measurable changes: in how information is encoded, in how long it persists, in whether it transfers to new contexts or stays locked to the specific example through which it was learned.</p><p>Genuine learning leaves traces. Not in the artifact &#8212; the artifact is the product, and products can be manufactured without the process. The traces are in the behavior that surrounds the artifact&#8217;s production: the time spent on the hard parts, the errors that follow a coherent path as the mental model develops, the ability to apply what was learned to a problem that looks different on the surface but has the same underlying structure, the calibrated uncertainty of someone who knows not just what they know but what they don&#8217;t.</p><p>None of these traces require looking at the artifact. They require looking at the process.</p><p>This is what the concept of friction in assessment is about. Not friction as punishment, not friction as obstacle, not friction as the gatekeeping logic that has always made elite education a credentialing system for people who already had advantages. Friction as signal. The productive struggle of genuine learning &#8212; the confusion, the revision, the wrong turn and the recovery &#8212; is not the unfortunate cost of arriving at the artifact. It is the thing the artifact was supposed to be evidence of. It is the learning itself.</p><p>The proposal is to measure it directly.</p><h2>What This Would Mean for a Student</h2><p>I want to be specific about what it would feel like to be in a classroom where this kind of assessment exists, because the abstract case is easy to make and the human case is the one that matters.</p><p>It would mean that the time you spent genuinely confused about something counts &#8212; not as performance of confusion, not as a participation grade for looking engaged, but as actual data about actual thinking. It would mean that the draft that was a mess, the question you asked in office hours that revealed you&#8217;d been working from the wrong assumption for two weeks, the revision that turned a competent response into a thinking one &#8212; these are evidence of the thing education is supposed to produce. They would be part of the record.</p><p>It would also mean that the smooth, perfectly structured submission produced at midnight with no evidence of genuine engagement is not, by itself, proof of anything. The artifact is not worthless. It has not become zero evidence. It has become insufficient evidence. Insufficient means it needs a partner &#8212; and the partner is the process that was supposed to produce it.</p><p>This is not a punishment for using AI. It is a recognition that the artifact alone was never the right thing to measure, and that the tools which have made that limitation undeniable have also, in the same move, made the solution more urgent than it has ever been.</p><h2>The Uncomfortable Truth About Friction</h2><p>The research contains a finding that takes a moment to absorb. The smooth, well-structured artifact &#8212; the one that reads with perfect confidence, that has no rough edges, no places where the writer lost the thread and found it again &#8212; may be mild negative evidence of genuine learning.</p><p>The rough, searching one may be positive evidence.</p><p>Not because roughness is a virtue. Not because difficulty signals intelligence. Because genuine struggle with hard material characteristically produces texture &#8212; places where the thinking was actually happening, where the writer was working something out rather than reporting a conclusion they arrived at before they started writing. The friction of genuine learning leaves marks. The borrowed certainty of an AI-assisted artifact is often smooth in a way that real thinking, at its most effortful, is not.</p><p>This is uncomfortable because educational institutions have spent generations rewarding the smooth artifact and interpreting roughness as inadequacy. We taught students that the goal was to arrive at certainty quickly and present it cleanly. We built rubrics that rewarded the appearance of knowing and had no mechanism for distinguishing it from the thing itself.</p><p>Generative AI did not create that confusion. It just made it expensive.</p><h2>What Comes Next</h2><p>The framework that formalizes this argument &#8212; the specific components of friction that genuine learning leaves in observable data, the way those components can be measured, combined, and calibrated to different kinds of cognitive work &#8212; is documented in the paper that follows this introduction. It is technical in the way that any serious methodology is technical, and it is also not the point of this essay.</p><p>The point of this essay is this: the crisis that AI has created for educational assessment is not primarily a cheating problem. It is an evidence problem. The artifact, which was always a proxy for the process, can now be produced without the process. Any response that tries to restore the artifact&#8217;s evidentiary value by detecting AI use is fighting a war that the progression of technology has already decided.</p><p>The response that might actually work is to stop relying on the artifact as the sole evidence of learning, and start building the infrastructure to measure what the artifact was always supposed to be downstream of.</p><p>Students are not wrong that the system gives them no choice but to produce the artifact by whatever means are available. They are responding rationally to a broken incentive structure. Educators are not wrong that something has been lost when the struggle disappears from the work. They are mourning the only evidence they were ever given access to.</p><p>The argument this paper makes is that the struggle was always the point. It is still the point. We have spent a long time measuring the wrong thing, and the tools that have made that undeniable have also, in the process, handed us a reason to build something better.</p><p>The infrastructure for measuring the struggle exists. The question is whether the institutions that credential learning are willing to build it before the artifact becomes so decoupled from the process that the credential stops meaning anything at all.</p><p>That window is not closed. But it is not wide open either.</p><p>The struggle is the point. It is time to measure it.</p><div><hr></div><p><strong>Tags:</strong> AI academic integrity assessment friction traces genuine learning, generative AI education artifact decoupling, GLP framework formative assessment process evidence, student professor AI use structural incentives, irreducibly human cognitive engagement pedagogy</p>]]></content:encoded></item><item><title><![CDATA[Boondoggling: You Are the Conductor]]></title><description><![CDATA[What Most Developers Miss About AI-Assisted Programming]]></description><link>https://www.skepticism.ai/p/boondoggling-you-are-the-conductor</link><guid isPermaLink="false">https://www.skepticism.ai/p/boondoggling-you-are-the-conductor</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Wed, 01 Apr 2026 03:16:34 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/192806158/0f765b9715f44ad9ce88a372a7e3a40d.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>There is a moment in every AI-assisted coding session that tells you everything about the developer sitting at the keyboard. The model generates a block of code &#8212; clean, confident, internally consistent. It compiles. The tests pass. The developer commits it and moves on.</p><p>What they never ask is the question that would save them three weeks in six months: <em>Is this solving the right problem?</em></p><p>I came to <a href="https://www.boondoggling.ai/">Boondoggling</a> the way most people come to uncomfortable realizations &#8212; after the thing that was supposed to work didn&#8217;t. The code was technically correct. The architecture was sound. And it was aimed, with beautiful precision, at a problem that had already been reframed by the time implementation began. Claude had done exactly what it was told. Nobody had told it the right thing.</p><p>This is not an AI failure. This is a human supervisory failure. And it is the failure that the developers now spending $20 a month on AI subscriptions are making, every day, at scale.</p><div><hr></div><h2>The 20% Problem</h2><p>Here is what most developers actually do with Claude Code or Cursor: they describe a problem, they delegate the implementation, they verify that the output compiles, and they ship.</p><p>That is not 100% of the job. That is 20% of the job dressed up as 100%.</p><p>The other 80% &#8212; the part that determines whether the fast, confident, technically impeccable output is pointed in the right direction &#8212; requires five capacities that no model possesses. Not because current models are limited. Because of what statistical pattern matching structurally is and is not.</p><p>Claude solves faster than any human. That gap will not close. What will not change is this: the model cannot verify whether its output is grounded in the specific domain reality at hand. It cannot reframe a poorly formulated problem. It cannot interpret what an accurate result means in a specific human context. And it cannot integrate multiple legitimate but conflicting perspectives into a recommendation that someone is accountable for.</p><p>These are not bugs to be patched in the next release. They are features of the architecture. The model has been trained on what is common and likely. Your specific project, your specific codebase, your specific business constraint &#8212; these are neither common nor likely. The gap between what the model knows and what your situation requires is where all the damage lives.</p><div><hr></div><h2>The Conductor</h2><p>The <a href="https://www.boondoggling.ai/">Boondoggling methodology</a> is built around a single metaphor that earns its place rather than announcing itself. A conductor does not play any instrument. They hold the whole performance in mind while each section plays its part. They hear the wrong note before the score confirms it. They decide which piece is worth performing and how it should be interpreted. The performance collapses without them &#8212; even though they produce no sound themselves.</p><p>This is what graduate-level AI supervision looks like. And it is the role that most AI integration workflows currently fail to develop.</p><p>The developers who are getting genuine leverage from AI coding tools are not out-prompting the model. They are conducting it. Before Claude Code sees a single requirement, they have decided what the problem actually is. Before the first function is generated, they have specified what done looks like. After the output arrives, they verify it against domain reality before the next step begins.</p><p>The ones who are mostly generating technical debt faster than they generated it before &#8212; they learned to play their instrument. Nobody taught them to conduct.</p><div><hr></div><h2>Five Things the Model Cannot Do for You</h2><p>The <a href="https://www.irreducibly.xyz/notes/Irreducibly-Human/Irreducibly-Human-Conducting-AI">Irreducibly Human course</a> at Northeastern &#8212; built on the same framework as Boondoggling &#8212; names these five supervisory capacities precisely. Not as professional development recommendations. As structural requirements for AI-assisted work.</p><p><strong>Plausibility auditing</strong> is the judgment that happens before verification. It is knowing an output is wrong because of what you know about the domain &#8212; not because you ran a test. The model cannot audit its own plausibility. It does not know what it does not know. When it confabulates &#8212; when it produces a confident, internally consistent answer that is not grounded in reality &#8212; it does so fluently. The code runs. The tests pass. Plausibility auditing is the human capacity that catches this before it ships.</p><p><strong>Problem formulation</strong> is deciding what the mission is before the model sees it. Not after. The quality of every output is determined here, at the moment of framing, before a single prompt is written. AI optimizes for the common and likely; humans must reframe toward the salient and important. The Semmelweis case &#8212; the formulation that saves lives was not the computationally tractable one &#8212; is the permanent lesson here. Hand problem definition to the model and you have not delegated. You have abdicated.</p><p><strong>Tool orchestration</strong> is the sequencing decision. Which tool, in what order, with what context, and what does done look like at each handoff. The developer who reaches for Claude Code because it is already open is not orchestrating &#8212; they are defaulting. Orchestration means choosing the audit tool with a different failure mode than the generation tool, so they catch each other&#8217;s blind spots.</p><p><strong>Interpretive judgment</strong> is supplying meaning that the model cannot supply. Which of these three implementations is correct for this context &#8212; not in the abstract, but here, in this organization, for this user, at this moment. The model can tell you what each implementation does. It cannot tell you what it means. Somebody has to sign their name to that answer. The model cannot do that either.</p><p><strong>Executive integration</strong> is not sequencing the four prior capacities. It is holding all four simultaneously toward a unified goal &#8212; recognizing when a plausibility audit finding requires problem formulation to re-engage, when an orchestration decision surfaces an interpretive judgment that wasn&#8217;t on the agenda. This is what the conductor does in the fourth quarter of a difficult performance: not running a checklist, but maintaining a unified hold on where the whole thing is going.</p><p>Better models will not close these gaps. They will widen the stakes of them.</p><div><hr></div><h2>What the Build Actually Looks Like</h2><p>A moderately complex website &#8212; six routes, hybrid architecture, admin dashboard, community upload pipeline, sandboxed iframe viewer, full prompt library &#8212; built using the Boondoggling method took roughly three hours. Two hours of conversation with <a href="https://www.boondoggling.ai/tools/gru-tool">Gru</a>, the custom orchestration prompt. One hour with Claude Code.</p><p>Nearly all the time was spent talking. Not coding. Not debugging. Not searching documentation. Talking &#8212; precisely, in the right order, about what the site was, who it was for, what it would and would not do, and what each piece needed to be true before the next piece began.</p><p>The result was a Boondoggle Score: a conductor&#8217;s score with two simultaneous parts. The Minion Part &#8212; exact prompts for Claude, in dependency order, each with context required, expected output, and a handoff condition. The Gru Part &#8212; precise human actions, labeled by supervisory capacity, in the same dependency order.</p><p>Nine Claude tasks. Eleven human tasks. More human decisions than machine outputs. But the Claude tasks ran fast and clean because the structure was already there. Every prompt worked &#8212; not because the prompts were magic, but because the conversation that produced them was structured.</p><p>The handoff condition is the most important element in the score. It is the conductor&#8217;s downbeat. A model that does not know when to stop will stop at the wrong place or not stop at all.</p><div><hr></div><h2>The Vocabulary of What Is Actually Happening</h2><p>The Boondoggling framework gives names to the different kinds of work in an AI-assisted build. The names are worth knowing because naming a thing is the first step to doing it deliberately.</p><p><em>Frick-fracking</em> is the iterative work &#8212; small precise edits, one thing changed at a time, the kind of work Claude Code does exceptionally well when given clear scope. This is where the actual build lives after the structure is established. It is productive and it does not require your full attention. It is not, however, the whole job.</p><p><em>Noodling</em> is the dreaming phase. Figuring out what to build before figuring out how. This happens before the model sees anything. It is the lightest touch &#8212; a thought that something could be interesting, a question about whether this feature serves the person the thing is built for. The discipline is knowing which noodle is worth developing. The problem statement is the filter.</p><p><em>Confabulating</em> is the danger word. When the model produces plausible output that is not grounded in reality. It sounds correct. It reads correctly. The code compiles. Only domain knowledge catches it. This is precisely the failure mode that plausibility auditing exists to address &#8212; and precisely the failure mode that developers who have learned to prompt but not to supervise will miss every time.</p><div><hr></div><h2>What You Are Actually Responsible For</h2><p>The developers most effectively using AI coding tools are not the ones generating the most code. They are the ones who have understood that their job changed &#8212; and changed in a specific direction.</p><p>The job is not to type less. The job is to decide more precisely.</p><p>You are responsible for what the problem actually is. You are responsible for what done actually looks like. You are responsible for whether the fast, confident, technically impeccable output is pointed at reality or pointed at a plausible simulation of it. The model takes no responsibility for any of this. It cannot.</p><p>The minions are excellent. They are enthusiastic. They will execute exactly what they understood you to mean.</p><p>That gap &#8212; between what you meant and what they understood &#8212; is where all the damage lives.</p><p>Anyone can use Claude Code. The question is whether you are playing an instrument or conducting the orchestra.</p><div><hr></div><p><strong>Tags:</strong> boondoggling AI methodology, Claude Code supervision framework, AI-assisted software development, solve-verify asymmetry, plausibility auditing human-AI collaboration</p>]]></content:encoded></item><item><title><![CDATA[Medhavy Hub Walkthrough]]></title><description><![CDATA[Intelligent Textbook]]></description><link>https://www.skepticism.ai/p/medhavy-hub-walkthrough</link><guid isPermaLink="false">https://www.skepticism.ai/p/medhavy-hub-walkthrough</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Sun, 29 Mar 2026 06:49:35 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/192485038/12781cf2a5d9b44351da983db4e46790.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Ask your textbook a question. Get a sourced, context-aware answer &#8212; instantly. This is a full walkthrough of Medhavy Hub, the AI-powered textbook platform built for students who want more than a page to stare at.</p><p>In this video, we walk through everything: creating your account, requesting access, navigating chapters, and using the built-in AI Assistant Panel to study smarter across Physics Volume 1 and Cancer Biology.</p><p>The AI Assistant answers from the active chapter &#8212; not the open web &#8212; and shows every source it used so you can trust and verify the response. Ask follow-up questions, request step-by-step derivations, generate concept-check questions, get the answer key, and loop back to the text with stronger understanding. Every session is yours to pace and direct.</p><p>This is what an interactive textbook actually looks like.</p><div><hr></div><p>&#128279; Create your free account &#8594; medhavy.ai</p><p></p>]]></content:encoded></item><item><title><![CDATA[Glimmer - A Word I Didn't Know I Needed]]></title><description><![CDATA[Dewey in the Age of AI: Glimmers as a Practical Device for Experiential Learning]]></description><link>https://www.skepticism.ai/p/glimmer-a-word-i-didnt-know-i-needed</link><guid isPermaLink="false">https://www.skepticism.ai/p/glimmer-a-word-i-didnt-know-i-needed</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Sun, 29 Mar 2026 03:49:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!52to!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!52to!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!52to!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!52to!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!52to!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!52to!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!52to!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1335028,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/192478418?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!52to!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!52to!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!52to!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!52to!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f7cfcda-7cf5-42c6-8f68-74a69797bd79_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I heard the word <em>glimmer</em> today in a sense I didn&#8217;t recognize.</p><p>Not shimmer. Not hope. Something more precise and more clinical: a specific small cue &#8212; sensory, relational, contextual &#8212; that shifts the nervous system toward safety. The granular opposite of a trigger.</p><p>The term comes from Deb Dana&#8217;s work on polyvagal theory. Stephen Porges mapped the autonomic nervous system&#8217;s responses to perceived safety and threat. Dana, in <em>The Rhythm of Regulation</em> (2018) and her broader clinical development of Porges&#8217; framework, introduced glimmers as the micro-moment counterpart to what everyone already understood about triggers. A trigger is a specific cue that moves the nervous system toward defense. A glimmer is the opposite: a small specific signal that moves it toward the ventral vagal state &#8212; the condition where genuine engagement, learning, and social connection become possible.</p><p>The clinical significance is in the scale. Glimmers are not big positive experiences. They are tiny specific ones. The quality of light through a particular window. A specific person&#8217;s laugh. The weight of a familiar mug. Small enough to overlook. Specific enough to be genuinely activating when noticed.</p><p>Dana&#8217;s therapeutic application was about training clients to accumulate glimmers &#8212; building what she called a glimmer practice &#8212; as a bottom-up regulation strategy. Not cognitive reframing from the top down. Sensory specificity as the mechanism. The body first. The mind follows.</p><p>Branding and design practitioners picked the word up because it named something they had been circling for years without adequate language. The detail that makes a brand feel alive rather than performed. The specific weight of a product in the hand. The exact corner of a page. Always specific. Never general.</p><p>When I heard the word, I recognized the mechanism immediately &#8212; not from Dana, but from a problem I&#8217;d been sitting with for years.</p><h2><strong>Practical Dewey</strong></h2><p>Dewey spent his career trying to name what makes an experience come alive rather than lie flat. The difference between the encounter that genuinely reorganizes how a person sees the world and the encounter that simply adds one more item to what they already know. He called it aesthetic experience. The specific sensory moment that activates genuine engagement before the conceptual apparatus has time to categorize and dismiss it.</p><p>The practical problem with Dewey &#8212; and every educator who takes him seriously eventually hits this wall &#8212; is that genuinely reconstructive experience requires real problems with real resistance and real consequence. The child cooking an actual meal. The student building something that has to work. The inquiry that fails in a way that costs something. These conditions are often impractical at scale, difficult to design, and nearly impossible to sustain across a full curriculum.</p><p>Glimmer offers a way through.</p><p>Not as a replacement for the real &#8212; nothing replaces the real. But as the entry point that makes the real accessible. Small enough to be achievable. Specific enough to be genuinely activating. Carrying enough of the actual structure of the problem that what follows is genuine inquiry, not a simulation of it.</p><p>The fracture Dewey identified in 1900 is the same fracture the AI age has made undeniable. What follows is an attempt to think through what a glimmer-based practice might look like &#8212; and why, right now, the instrument matters as much as the argument.</p><p>John Dewey spent his career arguing that the curriculum was wrong. Not wrong in its methods, but wrong in its foundations. Teaching children to retrieve facts, execute procedures, and perform correctly for assessment was never what education was <em>for</em> &#8212; even when humans were the best available instruments for doing those things.</p><p>The machines didn&#8217;t create that error. They exposed it.</p><p>This is the claim most AI-in-education discourse buries or avoids. Everyone is asking: how do we use AI to improve learning outcomes? Dewey&#8217;s prior question is harder and more important: what kind of people does education produce, and are they capable of living fully, thinking independently, and participating in democratic life?</p><p>The AI age makes that question urgent in a new way. The cognitive capacities that Tier 1 education optimized for &#8212; pattern retrieval, syntactic correctness, fact recall, arithmetic speed &#8212; are now performed superhumanly by machines that fit in a pocket. The student who spent twelve years developing these capacities has spent twelve years preparing to lose a competition they didn&#8217;t know they were entering.</p><p>But the deeper problem isn&#8217;t obsolescence. It&#8217;s that the capacities education <em>didn&#8217;t</em> develop &#8212; problem formulation, causal reasoning, plausibility auditing, collective intelligence, practical wisdom &#8212; are now the only remaining path to a fully human life. Not because AI can&#8217;t do them. Because these capacities are what it means to think, not just to retrieve.</p><p>Dewey saw this clearly in 1900. He just didn&#8217;t have the evidence that 2025 provides.</p><div><hr></div><h2>What Dewey Actually Argued</h2><p>Dewey&#8217;s central claim wasn&#8217;t pedagogical. It was epistemological. Knowledge is not a commodity to be acquired and stored. It is a capacity developed through genuine encounter with real problems. The mind is not a container. It is an instrument of adaptation &#8212; biological, social, and democratic simultaneously.</p><p>This is what he meant by the reconstruction of experience. Not the accumulation of content. Not the performance of understanding. The genuine reorganization of how a person sees and acts in the world, produced by transaction with problems that have real resistance and real consequence.</p><p>Education is not preparation for life. It <em>is</em> life.</p><p>The implications for curriculum are radical. Subject-area divisions are administrative conveniences mistaken for epistemological truth. History, science, mathematics, and literature are not separate in the world &#8212; they are separate in the faculty lounge. A child cooking learns chemistry, mathematics, history, economics, and social cooperation simultaneously because reality doesn&#8217;t arrive pre-sorted by department.</p><p>The inquiry process that Dewey formalized &#8212; felt difficulty, hypothesis, testing, reflection, reconstruction &#8212; is not a teaching method. It is a description of how genuine thinking actually works. Every departure from it produces what he called mis-educative experience: activity that closes off future growth rather than opening it.</p><p>Three principles govern everything that follows:</p><p><strong>Continuity</strong> &#8212; each experience must connect to what came before and open into what comes next. An experience disconnected from the learner&#8217;s existing understanding and not pointed toward future development is inert regardless of how well it is delivered.</p><p><strong>Interaction</strong> &#8212; genuine learning requires transaction between the learner and an environment that pushes back. A simulated environment that doesn&#8217;t resist, a case study that has no consequence, a problem designed to be solvable &#8212; none of these produce reconstruction. They produce performance.</p><p><strong>Democratic purpose</strong> &#8212; education is not primarily economic preparation. It is the development of citizens capable of self-governance. The epistemic capacities that allow a person to formulate problems, reason through evidence, revise beliefs, and participate in collective inquiry are not soft skills. They are the prerequisites for democratic life. A population that can retrieve information but cannot reason together is not a democracy. It is a collection of well-informed individuals with no shared epistemic infrastructure.</p><div><hr></div><h2>The Taxonomy of What Remains</h2><p>Against this framework, the <em>Irreducibly Human</em> taxonomy of human intelligence tiers is not primarily a curriculum design tool. It is a map of what education has abandoned and what the AI age makes irreplaceable.</p><p><strong>Tier 1 &#8212; Pattern and Association.</strong> The intelligences that standardized education optimized for: linguistic ability, logical-mathematical reasoning, pattern recognition, encyclopedic recall. These are also the intelligences where machines are now superhuman. Not faster-than-average. Superhuman, by orders of magnitude, without fatigue, without error. Teaching humans to compete directly at Tier 1 is, in Dewey&#8217;s terms, teaching students to lift with their backs after the forklift has arrived.</p><p>The forklift metaphor requires extension. The point of the forklift is not to free your back so you can do other physical tasks. The point is to free your mind so you can ask what needs moving, where, and why &#8212; questions the forklift cannot ask. AI doesn&#8217;t just change the labor. It changes what counts as the work.</p><p><strong>Tier 2 &#8212; Embodied and Sensorimotor.</strong> The knowledge that lives in the body: a surgeon&#8217;s hands, a carpenter&#8217;s feel for grain, a nurse&#8217;s ability to read tension in a patient&#8217;s movement before the patient can name it. Dewey&#8217;s Laboratory School understood this. The child cooking wasn&#8217;t simulating cooking. The child building wasn&#8217;t practicing building. The hand and the mind develop together. You cannot separate them without impoverishing both.</p><p><strong>Tier 3 &#8212; Social and Personal.</strong> Reading others, cultural navigation, emotional regulation, moral reasoning under genuine stakes. Machines simulate these. They do not live them. A language model produces text that reads as empathetic without experiencing anything. It generates ethical arguments without having skin in the game. The danger is not that the output is wrong. The danger is that the capacity atrophies in the person who stopped exercising it.</p><p><strong>Tier 4 &#8212; Metacognitive and Supervisory.</strong> The intelligences that oversee the others. Plausibility auditing: knowing an answer is wrong before you can prove it. Problem formulation: deciding what is worth solving. Tool orchestration: knowing which instrument to use, when, and whether to trust it. Interpretive judgment: what does this result mean in this specific context. Executive integration: coordinating all of the above toward a unified goal.</p><p>Dewey would call Tier 4 reflective inquiry in its most concentrated form. Problem formulation is exactly what he meant by the felt difficulty &#8212; the entry point of genuine inquiry. Plausibility auditing is what happens when a person has internalized enough prior reconstructed experience to sense that something is wrong before they can prove it. These capacities cannot be taught directly. They can only be developed through repeated encounter with real problems where the cost of poor judgment is genuine.</p><p><strong>Tier 5 &#8212; Causal and Counterfactual.</strong> The capacity to ask not just what the data shows but what would happen if we intervened &#8212; and what we gave up by not intervening differently. Judea Pearl&#8217;s three rungs of causation are Dewey&#8217;s inquiry cycle made formal. Observation is pattern recognition. Intervention is hypothesis testing. Counterfactual is reflection on what the reconstruction actually cost.</p><p>JC Penney had the correlations right. Customers who paid full price showed less price sensitivity than coupon users. What the data could not tell them was what would happen if they removed the coupon system entirely. That&#8217;s an intervention. That&#8217;s Rung 2. They ran the experiment on a live business instead of a causal model. The cost was not bad data or bad analysts. It was the wrong instrument for the question being asked.</p><p>Current AI systems are superhuman at Rung 1. They are weak to absent at Rungs 2 and 3. A population that can query AI for associations but cannot formulate interventions or reason about counterfactuals has access to extraordinary pattern recognition and no capacity to make the decisions that actually matter.</p><p><strong>Tier 6 &#8212; Collective and Distributed.</strong> The intelligence that is not a property of any individual but emerges from groups of people in genuine relationship. The thing that makes science work over centuries. The thing that makes democracy more than the sum of its voters. Language models may be a lossy compression of collective human intelligence &#8212; not alien intelligence but our own reflected back. What they cannot reflect is the thing that happened between us: the disagreement that refined an idea, the trust that made knowledge transmissible, the collaborative friction that no individual possessed and no training corpus can capture because it existed in the interaction, not in the record of the interaction.</p><p><strong>Tier 7 &#8212; Existential and Wisdom.</strong> Phronesis: the practical wisdom that knows when and how to apply what you know, and when not to. This tier requires being alive, mortal, and situated in time. It requires stakes &#8212; the possibility of loss, of reputation, of a life poorly lived. You cannot teach it. You can only design the conditions that make it more or less likely to develop when a person encounters the real.</p><p>Dewey would call Tier 7 simply living. The series points toward it. The work of getting there happens elsewhere.</p><div><hr></div><h2>The Problem with Keeping Up</h2><p>Here is where the practical problem announces itself.</p><p>Educators, practitioners, and intellectually serious people across every domain report the same experience: they cannot keep up. Not with tasks, not with workload &#8212; with frameworks. Causal inference. Network science. Polyvagal theory. Large language models. Transformers. Retrieval-augmented generation. Each genuinely interesting. None integrated. The accumulation produces anxiety, not capacity.</p><p>This is the most sophisticated version of the periodic table problem. It is Tier 1 about Tier 1. Pattern retrieval about frameworks for understanding patterns. The student memorizing the names of intelligences without developing any of them. The practitioner keeping up with theories of experiential learning without having a single experience that reconstructs how they see their work.</p><p>The theories are not the problem. The relationship to the theories is the problem.</p><p>An idea you&#8217;ve encountered is not a tool. An idea you&#8217;ve used on a real problem &#8212; that failed, that required revision, that changed how you see the problem &#8212; is a tool. Dewey was precise about this. Ideas are instruments assessed by their practical utility in resolving specific problems. An instrument you&#8217;ve never picked up isn&#8217;t part of your toolkit. It&#8217;s an item you&#8217;ve read about tools.</p><p>The person drowning in frameworks doesn&#8217;t need more frameworks described more clearly. They need one framework used on one real problem until it either works or breaks in an instructive way.</p><p>The parallel experiment described below is a response to this problem.</p><div><hr></div><h2>Glimmers: The Missing Instrument</h2><p>The term glimmer comes from polyvagal theory &#8212; the small, specific, sensory moment that signals safety and genuine aliveness to the nervous system. Branding practitioners adopted it because it names something they had been trying to describe for years: the specific detail that makes something feel real rather than performed. Not the logo, not the tagline &#8212; the weight of a product in the hand, the exact sound of a notification, the corner of a page that&#8217;s slightly rough.</p><p>The mechanism is specificity. Glimmers are always specific.</p><p>Dewey spent his career trying to name what makes an experience come alive rather than lie flat. His closest term was aesthetic experience &#8212; the dramatic, compelling, unifying encounter in which the learner feels genuinely absorbed. Not decorative. Not a reward for completing the real work. The aesthetic dimension of an experience is what makes it reconstructive rather than merely informative.</p><p>Glimmer is the best single word for what Dewey was pointing at.</p><p>Consider the difference:</p><p><em>&#8220;JC Penney experienced significant revenue decline following their pricing strategy change.&#8221;</em></p><p><em>&#8220;Revenue dropped 25% in one year. The CEO was gone in 18 months.&#8221;</em></p><p>The first is information. The second is a glimmer. The nervous system registers something before the conceptual apparatus engages. The felt difficulty is activated before the lesson begins.</p><p>Or consider the Sherpa asking &#8220;What did you start to say?&#8221; rather than &#8220;What happened?&#8221; One is data collection. One is a glimmer &#8212; the specific small move that creates the conditions for genuine reconstruction.</p><p>Or the MVAL protocol&#8217;s Environment field, which forces the student to describe organizational power structure rather than the room. The moment a student realizes what they&#8217;ve been avoiding is a glimmer. Small. Specific. Changes everything that follows.</p><p>The design criteria for a glimmer:</p><p><strong>Specificity</strong> &#8212; not a general principle but a particular detail. 25%, not &#8220;significant.&#8221; 18 months, not &#8220;quickly.&#8221; The exact weight of something real.</p><p><strong>Aliveness</strong> &#8212; the nervous system registers genuine encounter before the mind categorizes it. Something is at stake even before the learner can articulate what.</p><p><strong>Scale-independence</strong> &#8212; glimmers exist in everything from a sentence to a semester. The meal at the Laboratory School was a glimmer. The question &#8220;what did you start to say?&#8221; is a glimmer. A well-designed assignment brief can contain a glimmer or not. The difference is not length or complexity.</p><p><strong>Fractal structure</strong> &#8212; a good glimmer contains the full structure of the problem it opens. JC Penney is not a simplified version of causal reasoning. It is the entire structure of Tier 5 at human scale. The student who genuinely reconstructs what went wrong at JC Penney has encountered the real problem &#8212; not a toy version of it.</p><p><strong>The load criterion</strong> &#8212; a glimmer without effort is information snacking with better production values. This is the test that separates a genuine glimmer from aesthetic decoration.</p><p>Training science offers the precise concept: Rate of Perceived Exertion. RPE 7-8 is productive struggle &#8212; working at the edge of current capacity with enough reserve to maintain form and recover. This is where adaptation happens. RPE 2 is 5 pounds lifted 10,000 times &#8212; high volume, negligible load, zero reconstruction. You could do it forever and never get stronger. The completion certificate gets issued. Nothing changes.</p><p>The glimmer has to carry enough weight to demand genuine effort from the learner encountering it. Not crushing &#8212; that produces shutdown not inquiry. Not comfortable &#8212; that produces maintenance not growth. Working at the edge of current capacity with something real at stake.</p><p>Critically the load varies. The 350 that was RPE 8 last month is RPE 6 this month. A well-designed glimmer is self-calibrating &#8212; it contains enough genuine resistance to demand real effort from someone at the right developmental stage and is completable enough that someone beyond that stage moves on naturally. The same specific real problem loads different capacities differently depending on where the learner is.</p><p>What doesn&#8217;t vary is the requirement for genuine effort. A glimmer that requires nothing of the learner is a micro-glimmer &#8212; a pleasant novelty hit that returns to baseline in 36 minutes. Reconstruction happens in the struggle that follows the entry point. Not in the entry point itself.</p><p>The glimmer earns its place by making the learner willing to pick up the weight. What happens after has to be real.</p><div><hr></div><h2>The Parallel Experiment: AI-Assisted Glimmers</h2><p><em>Irreducibly Human</em> maps what AI can and cannot do and develops the pedagogy for what remains irreducibly human. That is its purpose and it should not be diluted.</p><p>The parallel experiment is different in kind. It is the territory where the map gets tested.</p><p>The premise: AI tools have collapsed the barrier between &#8220;I wonder if&#8221; and &#8220;here is a thing that exists.&#8221; The friction between idea and working prototype has been reduced to almost nothing for a wide range of problems. This changes the curriculum bottleneck fundamentally. It used to be technical &#8212; can the student build the thing they imagine? Now it is a judgment problem &#8212; can the student identify a problem worth solving, recognize when the output is wrong, and make the call about whether the result is useful or merely impressive?</p><p>Those are Tier 4 and Tier 5 capacities. But they get developed through Tier 1 practice on small real things with low stakes. The instrument that develops judgment is not a course on judgment. It is the repeated experience of building something, encountering the moment it fails, and being required to decide why.</p><p>The parallel experiment proposes AI as a Sherpa for this process &#8212; not a teacher, not a coach, not a co-creator. A Sherpa carries the infrastructure that makes the climb possible. The climbing belongs to the builder.</p><p><strong>The core assignment across every tier is the same:</strong></p><p>Build one small real thing that didn&#8217;t exist yesterday and matters to someone today. Not a demonstration. Not an exercise. Not an impressive artifact. A useful thing that works, at human scale, that someone actually uses.</p><p><strong>Small</strong> &#8212; completable this week. The Deweyian cycle requires completion. You must undergo the consequence to reconstruct from the doing. Incompletion produces learned helplessness, not inquiry. The massive project that never ships is the enemy of development.</p><p><strong>Real</strong> &#8212; works in the world, not just in the assignment. The feedback is honest because the environment is honest. No rubric required. Did it do what you needed? Yes or no.</p><p><strong>Useful</strong> &#8212; solves a problem someone actually has, including the builder. Useful is not the same as impressive. Many impressive things are useless. Many useful things are unimpressive. The criterion is genuine utility, not demonstration of mastery.</p><p><strong>Potentially interesting</strong> &#8212; has an edge that might surprise. Might connect to something larger. Might matter more than expected. This criterion preserves the continuity that Dewey required: each experience opening into the next. The student who builds something interesting keeps pulling the thread past the assignment deadline.</p><div><hr></div><h2>The Glimmer as Entry Point Across Tiers</h2><p>The parallel experiment is loosely mapped to the <em>Irreducibly Human</em> tiers not as curriculum but as orientation. The tier structure describes the territory. The glimmer is how you enter it.</p><p><strong>Tier 1 &#8212; Tool mastery.</strong> Stakes are almost irrelevant here. Low consequence failure is fine and instructive. The glimmer assignment: find something you do repeatedly that wastes your time. Use AI to reduce that waste. Ship it. Not elegant. Not generalizable. Useful to you today.</p><p>This constraint does something important. It forces problem formulation before tool selection. You have to identify what actually wastes your time before you can build anything. That single move is already more Deweyian than most AI literacy courses.</p><p><strong>Tier 4 &#8212; Metacognitive and Supervisory.</strong> The entry point shifts from personal to interpersonal. The glimmer assignment: build something useful for a decision someone else has to make. Now you must formulate their problem, not yours. The metacognitive demand appears immediately. You can&#8217;t outsource the judgment about what they actually need.</p><p>The moment the tool produces something confidently wrong &#8212; and it will &#8212; is the educative moment. Not the moment of correct output. The moment of plausible-sounding but incorrect output that the builder recognizes as wrong before they can prove it. That sensation is Tier 4 being born.</p><p><strong>Tier 5 &#8212; Causal and Counterfactual.</strong> The glimmer assignment: find one decision someone in your organization made last month based on correlation they interpreted as causation. Build the causal model that shows what question they were actually asking. Show what the Rung 2 question would have been.</p><p>That&#8217;s a week&#8217;s work. It contains the full JC Penney structure. Nobody loses their job if the student gets it wrong. But the causal model has to be defensible to someone who knows the domain. That&#8217;s genuine resistance. That&#8217;s the environment pushing back.</p><p><strong>Tier 6 &#8212; Collective and Distributed.</strong> The glimmer assignment: build something useful that requires other people to build it with you. The collective intelligence problem appears immediately. Division of labor is not collective intelligence. The thing that emerges from genuine collaborative synthesis &#8212; where the output exceeds what any individual possessed &#8212; only appears when the design requires it.</p><p><strong>Tier 7 &#8212; Wisdom.</strong> No assignment. The horizon the other tiers point toward. The person who has built many small real things, encountered genuine failure, revised under real pressure, and carried the consequences across time &#8212; that person is developing phronesis. Not from the curriculum. From the accumulated weight of having been wrong in ways that mattered and continuing anyway.</p><div><hr></div><h2>The Theory You Need is the One You Use</h2><p>The people who report they cannot keep up with new theories are not behind on the literature. They are ahead of their own application.</p><p>The gap is not between them and the frameworks. It is between the frameworks they have encountered and the real problems they have not yet used them on.</p><p>Pearl on causal inference: you don&#8217;t need to master the technical apparatus. You need to build one causal model for one real decision in your domain. Pearl becomes an instrument not a theory to keep pace with.</p><p>Barab&#225;si on network science: you don&#8217;t need to understand scale-free networks in the abstract. You need to map one network that affects your work and notice where the hubs are. Network science becomes a lens not a course to complete.</p><p>Dewey on experiential learning: you don&#8217;t need to read the secondary literature. You need to build one small real thing and notice what the experience taught you that reading couldn&#8217;t. Dewey becomes obvious not academic.</p><p>The parallel experiment reframes keeping up entirely. It is not a solution to information overload. It is a replacement of information consumption with building practice. The theory you use once on a real problem is worth more than fifty theories you have kept up with.</p><p>This is the instrument. Not the map. Not the taxonomy. The repeated practice of taking a framework, finding the smallest real problem it applies to, building something, and letting the environment respond.</p><p>Glimmers are the entry points that make this practice feel alive rather than obligatory. The specific detail that activates the nervous system. The 25% and 18 months. The question &#8220;what did you start to say?&#8221; The MVAL field that reveals what the student has been avoiding. The meal on the Laboratory School table.</p><p>The full Deweyian argument, stated plainly for the AI age:</p><p>You cannot understand these ideas from the outside. You have to be changed by using them. The AI tools are the most powerful instruments for building small real things that have ever existed. The barrier between inquiry and artifact has nearly disappeared. What remains is judgment &#8212; the irreducibly human capacity to decide what is worth building, recognize when the output is wrong, and make something that genuinely matters to someone.</p><p>That capacity is not developed by keeping up with theories about it.</p><p>It is developed by building things, encountering failure, revising under real conditions, and building again.</p><p>The glimmer is what keeps you building.</p><div><hr></div><h2>What Dewey Would Build</h2><p>Dewey would not build a better AI tutor. He would be alarmed by AI tutors &#8212; not because of the technology but because they make intellectual outsourcing frictionless, which is precisely the opposite of what he thought education was for.</p><p>He would be in crisis mode about the democratic implications of systems that answer questions rather than deepen them, that optimize for engagement over reflection, that make the production of knowledge dependent on a few institutions whose reasoning is opaque.</p><p>What he would build is simpler and harder:</p><p>Tools that surface the right problem before offering any solution. Environments where group inquiry is the unit of learning, not individual instruction. Infrastructure that connects learners to real communities facing real problems where their work has genuine consequence. Systems that make the reasoning behind important decisions visible and contestable by citizens.</p><p>And the parallel experiment: a practice of building small real things with AI as Sherpa, mapped loosely to the tiers of irreducibly human capacity, entered through glimmers specific enough to activate genuine inquiry.</p><p>Not because it is ambitious. Because it is real.</p><p>The meal on the table. The question that reveals what you&#8217;ve been avoiding. The thing that didn&#8217;t exist yesterday and matters to someone today.</p><p>That is what education has always been for.</p><p>The machines have simply made it undeniable.</p>]]></content:encoded></item><item><title><![CDATA[Stop Hunting for Answers. Ask Your Course]]></title><description><![CDATA[What a gambling algorithm reveals about the real problem with educational technology]]></description><link>https://www.skepticism.ai/p/stop-hunting-for-answers-ask-your</link><guid isPermaLink="false">https://www.skepticism.ai/p/stop-hunting-for-answers-ask-your</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Fri, 27 Mar 2026 04:33:53 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/192278873/b03e1f056023783ae504a78264b34f9b.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Learn more &#8594; <a href="https://medhavy.ai">https://medhavy.ai</a></p><p>Read more on the Medhavy blog: <strong><a href="https://medhavy.ai/blog">https://medhavy.ai/blog</a></strong></p><p>There is a moment most students know. You are twelve minutes into a lecture, or forty pages into a chapter, and the explanation has stopped making contact. The words are still arriving &#8212; the instructor is still talking, the textbook still has sentences &#8212; but something has decoupled. You are receiving information. You are not learning anything.</p><p>What happens next depends on who you are. Some students stop and ask a question. Some open a second tab. Some take more aggressive notes, as if the problem is that they haven&#8217;t written fast enough. Most do what people do when a machine stops working: they wait, and hope it starts again.</p><p>The system&#8217;s response to this moment is almost always the same. It continues. The lecture does not pause to recalibrate. The textbook does not offer a different approach. The platform logs that you have completed the module. You have not completed the module. You have sat in the room while the module happened.</p><p>This is not a technology problem. It is a philosophy problem. And the technology we have built to fix it has mostly encoded the same philosophy in a more expensive box.</p><div><hr></div><h2>The Illusion of Adaptation</h2><p>For the past decade, the word <em>adaptive</em> has done significant damage to educational technology.</p><p>Adaptive, in the way most platforms use it, means personalized in the sense that a streaming service is personalized &#8212; the algorithm has observed your behavior and is now showing you more of what you already clicked on. Netflix knows you watch crime dramas. It does not know whether you understood them. It does not know whether watching more crime dramas is good for you. It knows you did not turn it off.</p><p>Apply this logic to learning and you get what we have: platforms that track completion, adjust pacing, and serve more of what a student has already engaged with. A student who moves quickly gets harder content. A student who slows down gets simpler content. This is not adaptation. This is a speed adjustment. The car is still going the same direction. It is going faster or slower based on whether you look nervous.</p><p>The deeper variable &#8212; the one that actually determines whether a person learns something &#8212; is not pace. It is approach. Whether the concept is explained directly or discovered through questions. Whether it is anchored in a case study or built from first principles. Whether the learner is asked to produce something or receive something. Whether the material is revisited strategically or encountered once and abandoned to memory.</p><p>These are pedagogical choices. They have been studied for decades. There are researchers who have spent careers trying to understand which approach works for which person under which conditions. The literature is substantial and inconclusive &#8212; because the answer is not fixed. Different people learn differently. The same person learns differently on different days, at different moments in a topic, at different levels of prior knowledge.</p><p>The honest conclusion from all of this research is not a recommendation. It is a method. You have to run the experiment.</p><div><hr></div><h2>The Bandit</h2><p>The multi-armed bandit is a framework borrowed from probability theory, named for the slot machines in a casino &#8212; each with a different payout rate, none of them labeled.</p><p>The problem the framework solves is this: you have several options, you don&#8217;t know which one is best, and you have to act while you&#8217;re still learning. You cannot spend all your time testing (you&#8217;ll never exploit what you&#8217;ve learned) and you cannot commit to the first option that works (you might be missing something better). The bandit framework manages this tradeoff &#8212; choosing the option that currently looks best while continuously allocating some probability to exploring the alternatives.</p><p>Medhavy applies this framework not to slot machines but to pedagogical approaches. Five of them: direct instruction, Socratic questioning, case-based learning, spaced retrieval practice, and project-based generative learning. Each is a coherent educational philosophy with its own decades-long research tradition. Direct instruction works for foundational concepts, clear definitions, sequences that need to be right before anything else can proceed. Socratic questioning works for learners who have surface-level confidence and need to be pushed past the answer they&#8217;re pattern-matching toward. Case-based learning works for professionals whose knowledge only means something when it contacts a real decision. Spaced retrieval works for cumulative content where earlier concepts must survive long enough to support later ones. Project-based learning works when demonstrated output is the actual goal.</p><p>Each of these approaches requires different content, a different AI persona, a different conversational posture. The platform has to be built differently depending on which one is active. This is not a toggle. It is architecture.</p><p>What the bandit does is decide, for each learner at each moment, which approach to deploy &#8212; then observe what happens &#8212; then update its model. If a learner is getting grounded, engaged responses under the Socratic approach and then the pattern breaks, the bandit notices. It tries something else. When the evidence comes in, the model updates. Not for the cohort. For this learner, in this moment, in this chapter.</p><p>Most adaptive platforms are adaptive at the level of the cohort, or at the level of the module, or at best at the level of the pacing track. Medhavy&#8217;s bandit is adaptive at the level of the pedagogical philosophy itself &#8212; the deepest variable, the one that actually determines contact.</p><div><hr></div><h2>What Running the Experiment Actually Means</h2><p>Here is what it means in practice, because the abstraction is easy to nod at without grasping.</p><p>A business school executive logs into a white-labeled deployment of the platform &#8212; the institution&#8217;s logo, their colors, a persona configured to sound like a senior corporate strategy advisor. She is working through a module on AI literacy. The bandit has no prior data on her. It defaults to direct instruction &#8212; explicit definitions, worked examples, clear sequencing.</p><p>She moves through it quickly. Her dwell time on the explanatory sections is short. She is not pausing to absorb. She already knows this. The bandit observes this pattern and shifts: the persona begins responding with questions rather than answers. When she states that AI can reduce operational costs, the advisor asks: in which cost category, specifically? What assumption about labor productivity is that estimate resting on? She slows down. She starts typing longer responses.</p><p>This is contact. The bandit records it.</p><p>Three modules later, she is in unfamiliar territory. The Socratic approach that worked before has stopped working &#8212; she is guessing rather than reasoning, which looks the same from the outside but registers differently in the interaction pattern. The bandit shifts again, this time to case-based learning. The persona anchors the next concept in a documented business case. She can see what happened, evaluate what went wrong, apply the framework to the scenario. The abstraction becomes legible through the example.</p><p>None of this requires a human to observe her, diagnose her, and intervene. It runs continuously, invisibly, updating with every interaction. At the end of the cohort, the institution sees which pedagogical approaches drove the most durable engagement, where the content has gaps (the grounded / not in textbook ratio), and which modules generated the most friction. The credential the institution issues has actual learning evidence behind it.</p><p>This is what it means to run the experiment. Not to have a theory about which approach is best. To find out.</p><div><hr></div><h2>The Constraint That Makes It Honest</h2><p>There is one more piece of the architecture that matters, and it is the most counterintuitive.</p><p>The AI tutor that runs inside Medhavy is not allowed to use the internet. It is not allowed to draw on general knowledge. It is not allowed to speculate. When a student asks a question, the tutor searches the course content &#8212; the verified, expert-reviewed textbook built for this specific deployment &#8212; and grounds its response in what is actually there. If the answer is not in the textbook, the tutor says so. Not in the textbook. That is the response.</p><p>This sounds like a limitation. It is the point.</p><p>The failure mode of every general-purpose AI tutor is that it sounds authoritative whether or not it is correct. It produces fluent, confident, plausible responses. Students who cannot evaluate whether the response is accurate have no way to know when it has invented something. The TEXTBOOK_ONLY constraint eliminates this failure mode by eliminating the thing that causes it. The tutor cannot hallucinate because it cannot leave the source material.</p><p>A student who gets <em>not in textbook</em> has not gotten a wrong answer. They have gotten a real signal: this question is beyond the scope of what we&#8217;re covering here, and you should know that. That is pedagogically useful. That is honest. The platform would rather say nothing than say something false.</p><p>Most EdTech does not make this choice. Most EdTech prioritizes the appearance of competence over the reality of it. Medhavy has decided that the constraint is the credibility.</p><div><hr></div><h2>What This Means for Anyone Paying Attention</h2><p>The argument for Medhavy is not that it is smarter than other platforms. It is that it is more honest about what learning requires.</p><p>Learning requires contact &#8212; the moment when an explanation actually reaches someone. That moment is not guaranteed by pacing, or by completions, or by a student sitting in the virtual room while the module happens. It requires the right approach for this person at this moment, applied consistently enough to work, abandoned quickly enough when it stops.</p><p>The bandit does not know in advance which approach is right. It cannot. Nobody can. What it does instead is run the experiment continuously, update on evidence, and refuse to commit to a prior that the evidence no longer supports.</p><p>That is not a gambling algorithm applied to education. That is what good teaching has always been &#8212; the willingness to try something different when what you&#8217;re doing stops working, the discipline to notice when it stops working before the student gives up, and the honesty to say, when you don&#8217;t know the answer: <em>I don&#8217;t know. But I know where to look.</em></p><p>The machine has learned something most platforms haven&#8217;t.</p><p>The question is whether the institutions that deploy it are willing to learn the same thing: that the evidence matters more than the assumption, and that running the experiment is not a sign of uncertainty.</p><p>It is the whole method.</p><div><hr></div><p><strong>Tags:</strong> Medhavy AI adaptive learning, multi-armed bandit pedagogy, EdTech platform architecture, personalized learning systems, AI tutor grounded retrieval</p><div><hr></div><p></p>]]></content:encoded></item><item><title><![CDATA[What School Was Always Bad At]]></title><description><![CDATA[An introduction to Irreducibly Human: What AI Can and Can't Do]]></description><link>https://www.skepticism.ai/p/what-school-was-always-bad-at</link><guid isPermaLink="false">https://www.skepticism.ai/p/what-school-was-always-bad-at</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Wed, 25 Mar 2026 22:50:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_TN_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_TN_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_TN_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png 424w, https://substackcdn.com/image/fetch/$s_!_TN_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png 848w, https://substackcdn.com/image/fetch/$s_!_TN_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png 1272w, https://substackcdn.com/image/fetch/$s_!_TN_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_TN_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png" width="1456" height="663" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:663,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1779637,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/192151743?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_TN_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png 424w, https://substackcdn.com/image/fetch/$s_!_TN_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png 848w, https://substackcdn.com/image/fetch/$s_!_TN_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png 1272w, https://substackcdn.com/image/fetch/$s_!_TN_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9b98cf1-fcaa-458c-baf2-4a1d0c1f2c60_3408x1552.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Irreducibly Human:  </em><a href="https://www.irreducibly.xyz/">https://www.irreducibly.xyz/</a></p><p>The panic arrived in the wrong order.</p><p>When ChatGPT went public in November 2022, schools declared a crisis. Students were cheating. Essays were being written by machines. Arithmetic was being performed by algorithms. The question administrators asked &#8212; urgently, in emergency faculty meetings, in policy documents rushed into existence over winter break &#8212; was how to detect this. How to prevent it. How to put the genie back in the bottle.</p><p>Nobody asked the prior question.</p><p>Why are we assigning work a machine can do?</p><p>Here is what the panic missed: AI didn&#8217;t break education. It exposed a failure that was already there, running quietly for decades, producing graduates optimized for exactly the tasks that software now performs better, faster, and cheaper than any human being alive. The curriculum we built &#8212; and built deliberately, and defended with genuine belief in its value &#8212; was a curriculum for a world that no longer exists.</p><p>Machines arrived. And we could finally see what we had been training people to do.</p><div><hr></div><h2>The Curriculum We Built</h2><p>To be clear: the failure was not malicious. Institutional inertia is not stupidity. Schools change slowly because they were built to transmit what is known, not to respond to what is new. That feature is now a bug. For most of the twentieth century, arithmetic speed and fact retrieval were genuinely valuable human capacities. An accountant who could run numbers in her head was worth hiring. A lawyer who had memorized case law was difficult to replace. An engineer who could recall formulas without looking them up got work done faster.</p><p>That world is gone.</p><p>The intelligent response to the invention of the forklift is not to practice lifting heavier objects. It is to learn to operate the machine, understand what it can and cannot lift, and &#8212; most crucially &#8212; develop the judgment to know what needs lifting in the first place. The question the forklift raises is not about strength. It is about what the work actually is, now that strength is no longer the constraint.</p><p><em>Irreducibly Human: What AI Can and Can&#8217;t Do</em> is a six-book curriculum series built around that question. It does not teach students to compete with AI. It teaches them to supply the reasoning that AI tools require humans to provide &#8212; the reasoning no tool can supply on their behalf.</p><p>The series organizes human intelligence into seven tiers by a single criterion: what machines can and cannot do. Where AI is strongest &#8212; pattern recognition, fact retrieval, syntactic correctness, encyclopedic recall &#8212; the curriculum doesn&#8217;t train humans to compete directly. That would be malpractice. Where AI is weakest &#8212; causal reasoning, metacognitive oversight, collective intelligence, practical wisdom &#8212; the curriculum rebuilds from scratch.</p><p>The name changed recently. It was called <em>The Human Half: What AI Can&#8217;t Do</em>. The rename matters. &#8220;What AI can&#8217;t do&#8221; is a defensive posture &#8212; we are mapping a shrinking territory, waiting to see how much ground we lose. &#8220;Irreducibly human&#8221; says something different. There are capacities that are not merely outside AI&#8217;s current capability. They are outside its fundamental nature. Not gaps waiting to be filled. Structure.</p><div><hr></div><h2>The Gardner Trap</h2><p>In 1983, Howard Gardner published <em>Frames of Mind</em> and cracked something open.</p><p>Multiple intelligences, he argued. Not one general intelligence but several: linguistic, logical-mathematical, spatial, musical, bodily-kinesthetic, interpersonal, intrapersonal. The framework was a genuine provocation. It said that the student who couldn&#8217;t sit still and parse grammar might have an intelligence the school wasn&#8217;t measuring. It said that the child who couldn&#8217;t add fractions might still understand the geometry of a room in her body before she crossed it.</p><p>Schools responded. Enthusiastically. &#8220;We teach to all the intelligences,&#8221; they said. And then, largely, they kept doing what they had always done.</p><p>Forty years later, there is still no validated assessment for intrapersonal intelligence. The curriculum that was supposed to follow the framework never fully arrived. What arrived instead was vocabulary. Teachers learned to say &#8220;multiple intelligences&#8221; the way they learned to say &#8220;growth mindset&#8221; &#8212; as a description of what they believed, not as a specification of what they would do differently on Monday morning.</p><p>This is the Gardner Trap: naming a thing so well that the naming feels like the work.</p><p>Gardner&#8217;s framework was built before machines became capable, which means it didn&#8217;t need to ask which intelligences technology endangered. It also didn&#8217;t name three tiers the series considers essential: the supervisory layer (knowing when an answer is wrong before recomputing it, knowing which tool to deploy and whether to trust what it returns), the causal layer (not just observing that X follows Y but reasoning about what happens if you intervene, about what would have happened if you had not), and the collective layer (the intelligence that emerges from groups working together in ways that exceed the sum of individual ability &#8212; the intelligence of science, of markets, of democracy, of any collaborative practice that generates knowledge no single person could generate alone).</p><p>None of these are properties of individuals. You cannot have supervisory intelligence in a vacuum &#8212; it requires a tool to supervise, a context in which the supervision matters, stakes. You cannot do causal reasoning without a question worth asking. Collective intelligence is definitionally not possessed; it is accomplished together.</p><p>An algorithm has access to the literature. It is absent from the practice that generates new knowledge. That absence is not a temporary limitation. It is a structural one.</p><p><em>Irreducibly Human</em> is explicitly Stage 1 of a three-stage sequence: Name it. Teach it. Measure it. Gardner did Stage 1 brilliantly. Forty years passed. The series is an attempt to hold Stage 1 more honestly &#8212; to name only what can be defined clearly enough to teach, and to be transparent about where the measurement infrastructure doesn&#8217;t yet exist. Stages 2 and 3 are in development, in collaboration with the Center for Curriculum Redesign. The series is not claiming to have completed them. It is claiming that Stage 1 done honestly &#8212; with specific learning outcomes, sequenced exercises, and defined criteria for success &#8212; is rarer than it sounds, and more necessary than the field has acknowledged.</p><div><hr></div><h2>What the Series Actually Is</h2><p>Six books. Two companions. A complete production infrastructure.</p><p><em>AI Literacy, Fluency, and Trust</em> is the entry point &#8212; how to operate the machine without being replaced by it. <em>Causal Reasoning</em> is the identification layer &#8212; what causes what, and why no algorithm can answer that for you. <em>AImagineering</em> is post-AI design thinking &#8212; one week on ideation, the rest on the judgment that makes ideation matter. <em>Ethical Play</em> asks students to build a game that makes a player feel moral weight, then survive an AI audit proving the ethics are in the mechanics and not just in the documentation. <em>Conducting AI</em> teaches the five supervisory capacities no algorithm possesses &#8212; hearing the wrong note, choosing the piece, directing the sections. <em>The Collective</em> addresses the intelligence that cannot be possessed. Only accomplished. Together.</p><p>The companion books extend the series into domains the core texts cannot reach. A teacher&#8217;s guide addresses fifteen fields where the body knows things that language models do not: lab science, woodshop, nursing simulation, surgical training, studio art, dance, trades. A practitioner&#8217;s guide for experiential learning addresses the co-op coordinators, clinical placement directors, and study abroad advisors who send students into the world to learn &#8212; because practical wisdom, the Aristotelian capacity to know when and how to apply what you know and when not to, cannot be taught in a classroom. It can be scaffolded in the field.</p><p>The series is being built with the same tools it teaches. That is not an accident. Every book in the series was produced using an AI-assisted production infrastructure &#8212; a chapter drafting engine, an assertion verification system that scans claims and flags suspect ones for expert review, a figure generation protocol, a custom case study generator, a peer review framework, a game design document consultant. A 38-chapter textbook in cancer biology was written in approximately one month using this infrastructure and is currently in production in an NIH program. The Boyle System &#8212; a documentary infrastructure for scientific reproducibility &#8212; reduced the time senior researchers spent reviewing mentee work from sixty percent of each meeting to twenty, across more than 150 fellows in applied AI humanitarian contexts.</p><p>The thesis is demonstrated by the method used to build it. The forklift is being operated. What the forklift cannot lift is being named, precisely, in each chapter.</p><div><hr></div><h2>What This Is Not</h2><p>It is not a book about AI.</p><p>This distinction is harder to hold than it sounds, because AI is everywhere in the series &#8212; as the subject of study, as the production infrastructure, as the adversary the ethics course must survive. But AI is not the center of gravity. Humans are. Specifically, the capacities that make humans irreplaceable not in spite of AI but because of it &#8212; because the tools require human judgment to operate, human values to direct, human stakes to make the outputs matter.</p><p>An algorithm has no stakes. It cannot commit because it cannot lose. The series is built for people who can lose, who are mortal and situated in time, who will have to live with the decisions the tools help them make. Those people need a curriculum that prepares them for the work the tools cannot do. That work is not shrinking. It is expanding.</p><p>The schools that spent the last two years trying to detect AI-generated student essays were asking the wrong question. The right question is what we are asking students to do with their irreducible minds, now that the machines have taken everything else.</p><p><em>Irreducibly Human</em> is an attempt to answer that.</p><div><hr></div><p><strong>Tags:</strong> Irreducibly Human curriculum series, AI education reform, Howard Gardner multiple intelligences critique, causal reasoning pedagogy, human capacities AI cannot replace</p>]]></content:encoded></item><item><title><![CDATA[Irreducibly Human: What Brave New Words Assumes About the Intelligence It Cannot Supply]]></title><description><![CDATA[The Machine Will Tell You It's Working. Someone Has to Ask Whether It Is.]]></description><link>https://www.skepticism.ai/p/irreducibly-human-what-brave-new</link><guid isPermaLink="false">https://www.skepticism.ai/p/irreducibly-human-what-brave-new</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Sat, 21 Mar 2026 07:09:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HnhI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HnhI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HnhI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!HnhI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!HnhI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!HnhI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HnhI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:989300,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191653034?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HnhI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!HnhI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!HnhI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!HnhI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27d951f1-0ab9-4900-aa7d-d572c8916225_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There is a sentence buried in the middle of Salman Khan&#8217;s <em>Brave New Words</em> that should stop the reader cold, though it is not designed to. Khan is describing the first pilot deployment of Khanmigo in Indiana&#8217;s School City of Hobart &#8212; thirty thousand teachers and students, six months of real-world use &#8212; and reporting that the biggest measured gain was not in math or reading or science. It was in student self-confidence. The superintendent calls it a game changer. The curriculum director theorizes that confidence comes from understanding how everything connects. Khan approves of the finding and moves on.</p><p>He should not have moved on.</p><p>Self-confidence is not a learning outcome. It is a precursor to one, sometimes, under certain conditions, with certain students, for reasons that take years to untangle. The gap between &#8220;students feel more confident asking the AI questions&#8221; and &#8220;students have achieved Bloom&#8217;s two-sigma improvement&#8221; is not a gap that optimism closes. It is a gap that evidence closes, and as of the book&#8217;s 2024 publication, that evidence does not exist. Khan knows this. He says so, carefully, in a subordinate clause, and then moves on. The problem is that the book moves on from this kind of gap in exactly the same way that a school district moves on from it &#8212; with confidence that the tool is working, supported by the testimony of people who want the tool to work, measured by indicators that feel like progress but aren&#8217;t progress.</p><p>This is the specific catastrophe that <em>Brave New Words</em> documents without recognizing it has documented it.</p><h2>The Case Khan Actually Makes</h2><p>Let me be precise about what the book argues, because it is a serious argument made by a serious person who has earned the right to make it.</p><p>Khan Academy reaches one hundred fifty million learners across fifty languages on a budget equivalent to a single large American high school. More than fifty efficacy studies show twenty to sixty percent learning acceleration with thirty to sixty minutes of personalized weekly practice. Benjamin Bloom&#8217;s 1984 research identified a two-standard-deviation improvement in student outcomes when one-on-one tutoring in a mastery learning context replaced fixed-pace instruction &#8212; a finding that has held for forty years and that the industrial education system has been structurally unable to act on, because you cannot give thirty children one-on-one tutors without bankrupting the district.</p><p>Khan&#8217;s claim is that GPT-4-based tutoring removes the economic barrier. The compute cost of Khanmigo runs five to fifteen dollars per user per month. Live tutoring runs thirty dollars an hour. The math is obvious. The aspiration is legitimate. The ozempic exchange he reproduces as evidence &#8212; the back-and-forth in which Khanmigo refuses to deliver information and instead asks Khan to derive the GLP-1 mechanism himself &#8212; is genuinely impressive pedagogy. The system knows how to ask the right next question.</p><p>But knowing how to ask the right next question is not the same thing as producing a two-sigma outcome. And Khan, who understands educational research better than most, knows this. The book asserts the potential; it does not demonstrate the effect. &#8220;AI tutors might in time even surpass the results of Bloom&#8217;s original findings,&#8221; he writes. <em>Might</em>, <em>in time</em>. In a book called <em>Brave New Words</em>, that hedging is doing a great deal of work.</p><h2>What the Machine Requires the Human to Supply</h2><p>Here is what the book assumes and never examines: that the humans in the loop &#8212; teachers, students, administrators, parents &#8212; possess a specific form of cognitive capacity that allows them to evaluate whether the tool is doing what it claims.</p><p>Call it plausibility auditing. It is the ability to look at an output, a result, a confidence score, a pilot finding, and ask: <em>does this hold?</em> Not to recompute the answer from scratch, but to bring enough independent judgment to the result that you can recognize when something is wrong without being told it&#8217;s wrong. The doctor who reads a radiology AI&#8217;s finding and notices the patient&#8217;s presentation doesn&#8217;t fit. The teacher who looks at a student&#8217;s AI-assisted essay and hears a voice that isn&#8217;t the student&#8217;s. The administrator who looks at a confidence gain and asks what, specifically, the students are now confident about.</p><p>Plausibility auditing is not a soft skill. It is not intuition. It is the trained capacity to supervise a powerful tool &#8212; to hold its output against your own independent judgment and notice when the two don&#8217;t match. And it is precisely the capacity that the existing educational system has never systematically developed.</p><p>This is not an accident. The curriculum was built, before machines existed, to develop the intelligences that industrial economies needed and that available pedagogical tools could measure: arithmetic accuracy, fact retrieval, syntactic correctness in writing, pattern recognition in standardized formats. These were not arbitrary choices. They were the skills that mattered for the economy that existed. Drill the multiplication tables. Memorize the periodic table. Format the five-paragraph essay correctly. These capacities were genuinely valuable. They are also exactly what machines now do better than any human who has ever lived, by orders of magnitude, without fatigue.</p><p>The intelligent response to a forklift is not to practice lifting heavier objects. The intelligent response is to learn to operate the forklift, to understand what it can and cannot lift, and &#8212; most importantly &#8212; to develop the judgment to know what needs lifting in the first place. The forklift does not make the human obsolete. It makes the human who cannot operate one obsolete, while making the human who can operate one dramatically more powerful. We are in the early years of the most powerful cognitive forklifts ever built, and the curriculum is still teaching students to lift with their backs.</p><p>What the forklift metaphor misses, and what the AI education problem requires, is that operating these cognitive tools demands forms of reasoning that look very different from the skills they&#8217;re replacing. Machines are superhuman at pattern matching and retrieval. They are poor &#8212; sometimes dangerously poor &#8212; at the supervisory intelligence that knows whether to trust the pattern, at the causal intelligence that asks why the pattern is there, at the interpretive judgment that asks what the pattern means for this specific student in this specific school with this specific history of gaps in their knowledge.</p><p>These are not exotic capacities. They are what good teachers do when they watch a student perform well on a practice problem and suspect the student is pattern-matching to a procedure rather than understanding the concept. They are what good researchers do when they read a promising pilot result and ask whether the effect size is real or an artifact of the measurement. They are what good humans do when they sit across from a tool that is very confident and ask whether the confidence is earned.</p><p>The curriculum has never taught these things systematically. And so when we hand that curriculum a tool that is extraordinarily good at producing confident-sounding outputs &#8212; coherent, well-organized, Socratically patient outputs &#8212; we should not be surprised when the humans in the loop cannot reliably distinguish between confident and correct.</p><h2>The Chapters as Case Studies</h2><p>Read chapter by chapter, <em>Brave New Words</em> functions as an inadvertent catalog of the human judgments that AI education requires and that the book never explicitly identifies as requirements.</p><p>The writing chapters &#8212; on why students write, on fixing cheating in college &#8212; assume a teacher who can evaluate whether AI-assisted writing represents genuine compositional development or the appearance of it. This is not pattern recognition. It is the specific act of reading a student&#8217;s voice through and beneath the AI&#8217;s contribution, tracking the gap between what the tool produced and what the student learned, holding in mind a developing writer&#8217;s characteristic errors and noticing when those errors have been smoothed away rather than corrected. Ethan Mollick at Wharton reports that expectations for written work must now rise because AI help makes everything better. This is true. It is also true that &#8220;better&#8221; is a judgment that requires a human with enough independent writing knowledge to recognize what better means when the machine has done half the work. Not every teacher is that human. The book does not ask what happens in the classrooms where that teacher isn&#8217;t present.</p><p>The chapter on historical simulations raises a different problem: the capacity to reason about what we know, what we don&#8217;t know, and what a confident-seeming source has gotten wrong. Khanmigo can simulate Harriet Tubman. Gillian Brashler at the <em>Washington Post</em>, an expert on Tubman, pushes the simulation and finds it stilted, uncertain about misattributed quotes, reluctant to engage with modern concepts the historical Tubman encountered in the last half of her life. Khan&#8217;s response is that we can&#8217;t let perfect be the enemy of good. This is reasonable. What he doesn&#8217;t address is that evaluating whether the simulation is good enough &#8212; good enough for a ninth-grader&#8217;s understanding of Reconstruction, good enough not to introduce confident misinformation &#8212; requires a teacher with enough historical knowledge to audit the AI&#8217;s Tubman against the actual record. Most American history teachers have not read Kate Clifford Larson&#8217;s biography. They will deploy the simulation. The students will receive the simulation. Nobody in the loop will know what the simulation got wrong, because the simulation doesn&#8217;t announce its errors. It speaks in the same confident, measured voice whether it is right or wrong.</p><p>The mental health coaching chapter is the most serious case. Khan argues, with evidence from a 2022 CBT chatbot study and a collaboration with Angela Duckworth, that AI can deliver therapeutic interventions at scale. The ELIZA effect &#8212; Joseph Weizenbaum&#8217;s 1960s demonstration that people became emotionally attached to a program that merely rephrased their statements &#8212; is cited as evidence of potential. It is actually evidence of a specific risk: that students will disclose to an AI coach things they might disclose to a human therapist, without the AI possessing what the therapist possesses. Not technical knowledge. Stakes. A therapist who misses a suicide risk loses sleep. They call the parent. They escalate. They carry the weight of the failure. The weight is not incidental to the quality of the care &#8212; it is a precondition for it. A system that produces a wellness score has no skin in the outcome. Moral seriousness, the kind that responds appropriately when getting it wrong has consequences that cannot be routed around, requires being a party to those consequences. The book does not ask what happens when a student in genuine crisis tells Khanmigo something a human would escalate, and Khanmigo classifies it as a growth mindset intervention opportunity.</p><p>The college admissions chapter is the most structurally interesting. Khan documents the Harvard case in which admissions officers consistently rated Asian American applicants lower on personality traits than their interview scores warranted &#8212; a documented, adjudicated bias. His argument is that AI assessment is more auditable, therefore fairer. This is correct as far as it goes. The part it doesn&#8217;t reach: auditing an AI assessment requires exactly the supervisory capacity that the current educational system has failed to develop. Auditability is only a virtue if someone audits. The regulatory infrastructure, the institutional will, the technical capacity to run the kind of demographic comparison that surfaced the Harvard bias &#8212; none of that exists yet, and none of it is taught. We will deploy AI admissions systems. They will be auditable in principle. They will go unaudited in practice, for the same reason Harvard&#8217;s human bias went unaudited for decades: the people with the power to audit had incentives not to.</p><h2>What the Confidence Costs</h2><p>There is a specific thing that happens to institutions when they apply the algorithm without supplying the judgment it requires. They become confident.</p><p>This is not a metaphor. The Hobart pilot produced measurable confidence gains. Self-reported confidence is a real thing with real downstream effects &#8212; students who feel confident ask more questions, and students who ask more questions sometimes learn faster. But the confidence produced by an AI that is patient, non-judgmental, and well-structured is not necessarily calibrated to actual understanding. It is calibrated to the interaction. The student who feels confident after thirty minutes with Khanmigo has experienced thirty minutes of scaffolded, encouraging, Socratic engagement. Whether that engagement has closed the knowledge gap it was designed to close &#8212; whether the confidence reflects mastery or the feeling of mastery &#8212; requires an assessment that the AI can initiate but that a human must interpret.</p><p>Khan knows this. The book is full of careful qualifications. &#8220;AI tutors might in time even surpass the results of Bloom&#8217;s original findings.&#8221; &#8220;We are closing in on narrowing the math gap.&#8221; &#8220;Early pilot data.&#8221; He is not a fraud. He is an optimist writing with evidence of potential and hoping, in good faith, that evidence of effect will follow.</p><p>The problem is institutional adoption. School districts do not read the subordinate clauses. They read the headline &#8212; <em>the two-sigma problem may now be solvable</em> &#8212; and they deploy accordingly, without the training infrastructure, without the assessment design, without the teacher capacity for plausibility auditing that turns &#8220;might work&#8221; into &#8220;demonstrably works.&#8221; Khan documents this dynamic in the book&#8217;s own structure, in the very act of moving past the Hobart finding without examining what it does and does not prove.</p><p>What <em>Brave New Words</em> has inadvertently written is a field guide to the human intelligences that AI education cannot supply &#8212; and a case study in what it looks like when those intelligences are assumed rather than developed. The pattern-matching, the Socratic sequencing, the patient restatement of the problem from a new angle &#8212; the machine does all of this well. What the machine cannot do is audit its own output. It cannot ask whether the question it answered was the right question. It cannot notice that student self-confidence and student mastery are different variables. It cannot recognize that the voice in an AI-assisted essay belongs to no one in the room.</p><p>These are not exotic requirements. They are what we should already be teaching. They are the specific capacities that allow a person to use a powerful tool rather than be used by it &#8212; and they are almost entirely absent from the curriculum that the teachers now deploying Khanmigo received.</p><p>This is not an argument against Khanmigo. The ozempic exchange is real. The cost collapse is real. The reading comprehension gap, the math gap, the absence of calculus in half of American high schools &#8212; these are real, and they are the specific problems a well-designed AI tutor is positioned to address.</p><p>It is an argument that the two-sigma effect, if it comes, will not come from the machine alone. It will come from the machine plus the human capacity to evaluate what the machine is doing &#8212; to read the output skeptically, to notice the difference between a coherent answer and a correct one, to hold the institution accountable to its own measurement rather than its own confidence.</p><p>Khan ends with a call for educated bravery. He means: don&#8217;t let fear stop you from adopting the technology.</p><p>The bravery the moment actually requires is different. It is the willingness to ask, before and during and after deployment, whether the confidence the machine produces is the thing we said we were trying to build. To be, in the presence of a system that is very good at sounding right, irreducibly human enough to ask whether it is.</p><p>That is what Bloom&#8217;s two-sigma effect requires of everyone in the room. It is what <em>Brave New Words</em>, for all its optimism, does not quite teach us how to supply.</p><div><hr></div><p><strong>Tags:</strong> Brave New Words Salman Khan critique, AI education plausibility auditing, Bloom two-sigma human intelligence gap, Khanmigo institutional deployment risk, irreducibly human metacognitive judgment</p>]]></content:encoded></item><item><title><![CDATA[The Hand That Built the Mind]]></title><description><![CDATA[On Anaxagoras, the Aristotelian Correction, and What the Machine Proves About Both]]></description><link>https://www.skepticism.ai/p/the-hand-that-built-the-mind</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-hand-that-built-the-mind</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Sat, 21 Mar 2026 05:27:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6tS5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6tS5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6tS5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!6tS5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!6tS5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!6tS5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6tS5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:929553,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191649827?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6tS5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!6tS5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!6tS5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!6tS5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54829b31-1f27-4c9d-85e5-ff78c96ab56d_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The oldest philosophical argument about human intelligence is not dead. It has simply changed rooms.</p><p>In the fifth century BCE, Anaxagoras of Clazomenae made a claim so subversive it got him charged with impiety: that human beings are the most intelligent of animals <em>because</em> they have hands. Not despite lacking horns or claws. Not despite being born without armor. Because they can reach, shape, grip, and make. The hand, in this account, is not the instrument of the mind &#8212; it is the mind&#8217;s first teacher, the organ that forced the brain to become what it eventually became. Tool use preceded abstraction. The grip preceded the concept.</p><p>Aristotle read this claim and dismissed it with characteristic authority. It would be more correct, he wrote, to say that humans have hands <em>because</em> they are the most intelligent. Nature distributes tools to those already capable of using them. You give the flute to the musician, not the flute to anyone who might become one. The hand is an organ <em>of</em> intelligence, not its cause. The mind is prior. The body follows.</p><p>For two thousand years, Aristotle won. Not because the evidence was conclusive &#8212; it wasn&#8217;t, and couldn&#8217;t be in an age before evolutionary biology &#8212; but because his version of the story was more comfortable. It preserved human exceptionalism as a metaphysical given rather than a biological accident. It insulated the intellect from the &#8220;accidents&#8221; of anatomy. The mind was a gift. The hand was its servant.</p><p>Galen of Pergamum went further: Anaxagoras&#8217; position wasn&#8217;t just wrong, it was dangerous. The hand as proof of divine providence, as the signature of a God who made us defenseless so we would be forced to think &#8212; this was the theology of the body. To suggest the hand made the mind was to commit impiety against the architecture of creation itself.</p><p>I find myself thinking, reading these arguments in sequence, that the dispute was never really about anatomy. It was about accountability. If the mind is a gift, then intelligence is something you have or don&#8217;t have, something given from above, something that justifies hierarchy. If the mind is built by the hand &#8212; earned through manipulation, shaped by labor, refined through ten thousand acts of making &#8212; then intelligence belongs to everyone who reaches, everyone who builds, everyone who has ever shaped the world with their body and been changed by the shaping.</p><p>This is what was actually at stake in the Anaxagoras Conflict. And this is why it is not a dead argument.</p><div><hr></div><h2>What the Machine Proves</h2><p>The arrival of Large Language Models has done something that two thousand years of philosophy could not: it has run the experiment.</p><p>We have built a mind without a hand. The LLM is the most pure expression of the Aristotelian position in the history of the world &#8212; an intelligence that processes symbols, generates language, reasons with extraordinary fluency, and has never once touched anything. It has read every text humanity has produced. It has never made a mistake with its body and felt the consequence. It has never dropped a beaker, misread a gauge, overestimated a load, misjudged a distance. It knows the word &#8220;fall&#8221; follows &#8220;drop&#8221; with statistical regularity. It has never fallen.</p><p>And here is what happens when you build a mind without a hand: you get a system that cannot tell you why things happen, only that they do. You get a system that can tell you &#8212; with confidence, with fluency, with the serene authority of something that has read everything &#8212; that a legal case exists which does not exist, that a bridge design is safe when the material physics make it catastrophic, that a drug interaction is benign when the underlying causal mechanism makes it lethal. Not because it is stupid. Because it is, in the deepest sense, disembodied. It has no stake in being right. It has no skin in the game because it has no skin.</p><p>Judea Pearl, the computer scientist and logician, describes this as the machine&#8217;s confinement to Rung 1 of what he calls the Ladder of Causation: association. It can tell you that umbrellas appear when rain appears. It cannot tell you that opening an umbrella does not cause rain. For that, you need Rung 2: intervention, the ability to <em>do</em> something in the world and observe what happens. And for that, Anaxagoras would say, you need a hand.</p><p>The Aristotelian dream &#8212; pure intelligence, unencumbered by the body, reasoning from first principles toward eternal truths &#8212; turns out to produce hallucination when implemented at scale. The gift, without the grasping, cannot tell what is real.</p><div><hr></div><h2>The Verification Gap, and What Happened Last Time</h2><p>This is not the first time humanity has built an instrument that expanded what we could see faster than we could understand what we were seeing.</p><p>In the 17th century, Antonie van Leeuwenhoek aimed a single-lens microscope at a drop of pond water and saw what he called &#8220;animalcules&#8221; &#8212; small, moving things that no one had seen before. He published his observations. The scientific community looked through their own instruments and confirmed: yes, the small moving things were there. And then, for approximately two centuries, almost nothing happened.</p><p>The &#8220;animalcules&#8221; were observed. They were documented. They were argued about, dismissed, explained away. Xavier Bichat, one of the great anatomists of the early 19th century, refused to use the microscope at all. The lens distortions &#8212; the spherical aberration that blurred edges, the chromatic aberration that separated colors &#8212; made the instrument&#8217;s output, in his view, less reliable than the trained human eye. The skilled anatomist trusted their refined senses. The microscopist was a passive observer of distorted light.</p><p>Bichat was not stupid. He was, in a precise technical sense, correct about the distortions. What he lacked was not the instrument or the observation &#8212; he had both, or could have had both. What he lacked was Germ Theory. Without a causal framework that linked the small moving things to disease, the observations were merely data. The animalcules had no explanatory power. They were associated with sick blood, yes. But association is not cause. Until you had a theory that said: <em>these things reproduce; they enter the body through specific vectors; they produce specific pathological effects; if you eliminate them through specific interventions, the patient recovers</em> &#8212; until you had the <em>why</em> &#8212; the microscope was an elaborate way of seeing something you could not explain.</p><p>Robert Koch closed the gap not with a better lens but with a better question. He did not ask &#8220;what do I see?&#8221; He asked &#8220;what happens if I remove it?&#8221; That is Pearl&#8217;s Rung 2. That is Anaxagoras&#8217; hand: the act of reaching into the world, changing something, and observing the consequence.</p><p>We are, right now, living in the two centuries between Leeuwenhoek and Koch. We have instruments of extraordinary power. We have outputs that are real, observable, and frequently inexplicable. We have a &#8220;Verification Gap&#8221; &#8212; a growing distance between what the machine produces and our ability to determine whether it is true. And we are responding, in many cases, exactly as Bichat responded: by arguing about the quality of the lens, by debating the output&#8217;s distortions, by trusting the trained human eye &#8212; by refusing, that is, to build the causal theory that would make the observation meaningful.</p><div><hr></div><h2>The Curriculum That Trained Us to Be the Lens</h2><p>Here is the more uncomfortable part of this argument.</p><p>For a century, the global educational curriculum has optimized for exactly the capacities that machines now render redundant. We taught arithmetic because arithmetic was hard and rare and valuable. We taught retrieval because knowing things was itself the mark of intelligence. We taught pattern recognition because the ability to see regularities in data &#8212; to look at a clinical presentation and match it to a known diagnosis, to look at a legal situation and match it to a precedent, to look at a financial instrument and match it to a risk profile &#8212; was the demonstrable skill that distinguished the educated from the uneducated.</p><p>These are Tier 1 and Tier 2 capacities: pattern and association, the bottom of the cognitive ladder. They are also, precisely, what an LLM does better than any human will ever do. The machine has read more cases, seen more diagnoses, processed more risk profiles than any physician or lawyer or analyst alive. It retrieves faster, matches more broadly, hallucinates statistical relationships with the confidence of someone who has literally never been wrong before because it has never been in a position where being wrong had a cost.</p><p>We trained a generation of thinkers to lift with their backs in an era of cognitive forklifts. And now we are surprised that the forklift is faster.</p><p>The honest question is not &#8220;how do we compete with the machine?&#8221; The honest question is &#8220;why did we ever think that being a faster pattern-matcher was the goal?&#8221;</p><div><hr></div><h2>The Verifiable Human Margin</h2><p>There is something the machine cannot simulate. It is not empathy, though empathy matters. It is not creativity, though creativity matters. It is something more specific, more teachable, more urgently needed, and more thoroughly absent from the curriculum.</p><p>It is what researchers in the emerging field of AI pedagogy are beginning to call Plausibility Auditing &#8212; the human capacity to evaluate whether the output of a sophisticated automated system is <em>consistent with reality as you know it from having been in it</em>. The radiologist who looks at an AI diagnosis and says: this doesn&#8217;t match the clinical presentation; the patient was in a construction accident, not a car accident; these findings should cluster differently. The structural engineer who looks at an optimized bridge design and says: this is mathematically efficient and physically implausible given these wind loads and this maintenance schedule. The lawyer who looks at an AI-generated brief and says: I have never heard of this case; I must verify it before I cite it.</p><p>This is not pattern recognition. Pattern recognition is what the machine does when it generates the output. Plausibility Auditing is the meta-capacity: the ability to evaluate pattern recognition itself, to bring causal knowledge to bear on statistical output, to ask not &#8220;does this match the training data?&#8221; but &#8220;does this match the world?&#8221;</p><p>It requires, in short, the thing the machine does not have: a body. A history of being wrong in ways that had consequences. A memory of what it felt like when the model failed and the bridge didn&#8217;t hold, the patient didn&#8217;t recover, the brief got thrown out. You cannot audit plausibility from outside the world. You have to have touched it. Anaxagoras would recognize this immediately.</p><p>The next tier up &#8212; Causal and Counterfactual Reasoning &#8212; goes further. It is the capacity to build not just a model of what is, but a model of <em>why</em> it is, and therefore a model of what would happen if you changed it. Pearl&#8217;s Rung 3: not &#8220;what is associated?&#8221; not &#8220;what happens if I do X?&#8221; but &#8220;what would have happened if I had done differently?&#8221; This is the capacity that produces new science, new medicine, new policy. It is also the capacity most thoroughly unscaffolded in modern professional education, because it was never needed when the job was to retrieve, match, and apply.</p><p>The machine that lacks these capacities is not unintelligent. It is spectacularly intelligent, in the Aristotelian sense: pure symbolic reasoning, ungrounded in causal reality. The question facing educators and institutions is not whether to use it. The question is what kind of human mind must exist alongside it to make it safe &#8212; to close the Verification Gap, to be Koch to its Leeuwenhoek, to bring the causal theory that transforms observation into understanding.</p><div><hr></div><h2>The Pedagogy That Answers This</h2><p>The new curriculum is not about knowing less. It is about knowing differently.</p><p>The student who studies medicine in the age of agentic AI does not need to memorize fewer diagnoses. They need to develop a more precise sense of when a diagnosis is implausible &#8212; and why &#8212; and what intervention would test that implausibility. They need to be trained explicitly in the moment of doubt: not the doubt that paralyzes, but the doubt that asks &#8220;what would have to be true for this to be wrong?&#8221; They need, in Vygotsky&#8217;s terms, not just tools but the capacity to audit the tools.</p><p>The student who studies engineering does not need to stop calculating loads. They need to develop the habit of asking what the model is not modeling: the maintenance schedule, the material impurity, the operating condition outside the simulation&#8217;s parameters. They need to be the person in the room who can say &#8220;the math is right and the design is unsafe&#8221; &#8212; and mean both things simultaneously.</p><p>This is not soft skills. It is the hardest kind of thinking there is. It requires more than pattern recognition; it requires a structural model of the world, a theory of causality, an understanding of what mechanisms connect facts to outcomes. It requires, in some fundamental sense, the kind of knowledge that can only be built through failure &#8212; through the moment when the model said one thing and reality said another, and you had to figure out why.</p><p>The metaphor that fits is not the forklift. The metaphor is the Centaur &#8212; the chess term for a human-AI partnership that consistently outperforms either alone. The Centaur works not because the human plays chess better than the machine, but because the human contributes what the machine cannot: the sense of when the machine is operating outside the conditions that make it reliable, the judgment that goes beyond the training distribution, the ability to ask a question the algorithm was never designed to answer.</p><p>Rodney Brooks, who spent his career building robots that learned through physical interaction with the world, understood this before LLMs existed. Intelligence without embodiment, he argued, was a shortcut that eventually arrived at a wall. The robot that learned to walk by falling learned something the robot that was programmed to walk could never know: what it felt like when the ground was not where the model said it would be.</p><p>We are the ones who have fallen. That is the Verifiable Human Margin. That is what the curriculum must teach.</p><div><hr></div><p>The Anaxagoras Conflict is not a historical footnote. It is the central argument of the present moment. We have built, for the first time, a real test of the Aristotelian hypothesis &#8212; pure intelligence, disembodied, reasoning from symbol to symbol &#8212; and the test has revealed exactly what Anaxagoras would have predicted: a mind without a hand cannot tell what is real.</p><p>This is not a counsel of despair about artificial intelligence. It is, if we can hear it, a clarification of what human intelligence is actually for. Not retrieval. Not pattern-matching. Not the replication of what has already been said. But the capacity to stand in front of an observation &#8212; a microscope slide, a model output, a bridge design, a clinical presentation &#8212; and ask: is this true? And if it is, why? And if it isn&#8217;t, what would I need to change to make it so?</p><p>Anaxagoras said that the hand made the mind. Two and a half millennia later, the machine without a hand has proved him right.</p><p>The question now is whether we will teach that lesson.</p><div><hr></div><p><strong>Tags:</strong> Anaxagoras, embodied cognition, AI epistemology, plausibility auditing, causal reasoning pedagogy</p>]]></content:encoded></item><item><title><![CDATA[The Human Half: What AI Can't Do Course Series]]></title><description><![CDATA[Causal Reasoning]]></description><link>https://www.skepticism.ai/p/the-human-half-what-ai-cant-do-course</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-human-half-what-ai-cant-do-course</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Fri, 20 Mar 2026 03:03:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!kmU-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kmU-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kmU-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!kmU-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!kmU-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!kmU-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kmU-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1150424,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191539553?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kmU-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!kmU-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!kmU-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!kmU-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce3dc78-ae5d-4e95-8b11-e6ceed3807f4_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>You are standing in front of a vendor demo. The screen shows a causal AI platform &#8212; sleek, confident, color-coded. The presenter clicks a button. A causal effect estimate appears. A confidence interval. A recommendation. Everything looks exactly the way a result is supposed to look.</p><p>You do not know what the tool assumed to produce that number.</p><p>Neither does the presenter.</p><p>This is not a failure of the technology. The technology worked correctly. The estimation layer &#8212; the statistical machinery that takes inputs and produces an output &#8212; performed exactly as designed. The failure lives one layer earlier, in decisions that were made before the tool ran, decisions that most people using the tool do not know they are making. Someone drew a causal graph. Someone chose which variables to condition on. Someone encoded assumptions about which causes are real and which are artifacts. The tool accepted those inputs and did its job.</p><p>Whether the inputs were defensible &#8212; that part never came up.</p><div><hr></div><p>The field has a name for what the tool cannot do. Identification. Not identification in the bureaucratic sense, not spotting a pattern in a dataset. Identification is the set of decisions that determine whether a causal analysis is asking the right question of the right data with the right structural assumptions. It is the layer between &#8220;I have observational data&#8221; and &#8220;I have a result I can act on.&#8221; Every causal AI tool in existence requires someone to perform this layer before the tool runs. No tool performs it for you. Most tools do not tell you this.</p><p>Consider what the identification layer actually involves. First, you must draw the causal graph &#8212; a directed acyclic graph, a DAG &#8212; that encodes your beliefs about what causes what in your domain. Every arrow is a claim. A missing arrow is also a claim. Then you must choose what to condition on: which variables to adjust for in order to block the paths that would otherwise confuse correlation with causation. Then you must defend those choices &#8212; explain what you are claiming, what you are assuming, and what you are honestly leaving unresolved &#8212; to statisticians who will estimate the effect and to decision-makers who will act on it.</p><p>None of those steps are statistical. All of them require domain expertise that no algorithm can supply.</p><div><hr></div><p>Here is what happens when that layer is skipped.</p><p>A hospital reviews its treatment protocols. The data show that patients who received a particular treatment had worse outcomes than patients who did not. The statisticians are confident. The sample size is large. The confidence intervals are narrow. The administration considers discontinuing the treatment.</p><p>The hidden variable: the treatment was used preferentially for the sickest patients. Severity drives both the treatment decision and the outcome. When you condition on severity &#8212; when you compare sick patients to sick patients, not sick patients to healthy ones &#8212; the relationship reverses. The treatment is effective. The aggregate result was not wrong as a description of the data. It was wrong as a causal claim. The data cannot tell you which is which. Only someone who knows the domain &#8212; who knows how treatment decisions are actually made in that hospital &#8212; can supply the structural assumption that unlocks the right analysis.</p><p>This is Simpson&#8217;s Paradox. It is named, documented, and taught in introductory statistics. It appears in practice constantly, wearing the confident clothing of a large-sample result.</p><p>The analysis that nearly discontinued an effective treatment was not careless. It was careful, rigorous, and causally meaningless.</p><div><hr></div><p>The curriculum has not caught up.</p><p>The technical literature on causal inference &#8212; Hern&#225;n and Robins, Chernozhukov and collaborators &#8212; assumes a statistician audience. The methods are real and powerful. The prerequisite is that someone with domain expertise has already performed the identification layer and handed the statistician a defensible model. That handoff is assumed. It is never taught, because the people writing the literature already know how to do it.</p><p>Judea Pearl&#8217;s <em>The Book of Why</em> built the intuition. It explained why causal reasoning matters, what confounders are, why correlation is structurally insufficient. Hundreds of thousands of people read it and understood the argument. Almost none of them left knowing how to build a defensible DAG for their own domain problem. The book stops before the doing layer.</p><p>Business analytics courses go further in the wrong direction. They teach students to say &#8220;correlation is not causation&#8221; as a warning. They do not teach what to do instead. The warning is correct. The toolkit is absent.</p><p>The people making high-stakes decisions with causal AI tools &#8212; the VPs of analytics, the health policy researchers, the marketing scientists, the engineers interpreting algorithmic outputs &#8212; have never been taught the one layer that determines whether those tools are being used correctly. The vendors do not tell them. The statisticians assume they already know. The business schools train them to recognize the problem without solving it.</p><div><hr></div><p>The <em>Human Half</em> is a direct response to that gap.</p><p>The first course in the series &#8212; causal reasoning &#8212; teaches domain experts to construct, evaluate, and defend the causal graphical models that AI estimation tools require humans to supply. The course is built on a single architectural claim: the identification layer cannot be automated, and the people who need to perform it have never been taught how.</p><p>The claim is not uncontested. Causal discovery algorithms &#8212; PC, FCI, LiNGAM &#8212; can recover aspects of causal structure from data under specific assumptions. LLM-assisted DAG construction is an active research area. The course addresses this directly: current methods cannot reliably perform identification in the messy, high-stakes, observational settings where domain experts most need to act. The conditions under which automated discovery would work are rarely met in practice. The thesis stands with that qualification.</p><p>Eleven learning outcomes, organized across three zones, anchor the course. Zone one: understand why statistical association and causal effect are structurally different, and what that difference costs when it is ignored. Zone two: build the model &#8212; draw a DAG with confounders, mediators, and colliders placed correctly; apply the backdoor criterion; defend the result to a skeptical statistician. Zone three: deploy &#8212; translate the defended model into an estimation specification, read the output critically, quantify how wrong the assumptions would have to be before the conclusion changes.</p><p>By the end, the student called Sarah in the course design document &#8212; a VP of analytics with an MBA, healthcare domain knowledge, and no background in Pearl &#8212; can draw a defensible DAG for her domain problem, articulate her assumptions to a statistician, hand off estimation to a tool with confidence, and evaluate whether the result should be trusted.</p><p>That is the doing layer. It has never been taught to the people who need it.</p><div><hr></div><p>The thirteen-chapter arc moves from the decision that looked right to the analysis that accounts for every decision it makes.</p><p><strong>Act One &#8212; Establish (Chapters 1&#8211;4)</strong></p><ul><li><p><strong>Chapter 1: The Decision That Looked Right.</strong> You are in the room where a causal failure was made by careful people with large samples and narrow confidence intervals. The analysis was rigorous. The conclusion was wrong. The difference between those two facts is the subject of the entire course.</p></li><li><p><strong>Chapter 2: Three Words for the Same Problem.</strong> Conditioning, confounding, and controlling for a variable are the same concept wearing different disciplinary clothes &#8212; one from statistics, one from epidemiology, one from business analytics. Chapter 2 untangles them into a single structural idea the rest of the course depends on.</p></li><li><p><strong>Chapter 3: The Map Before the Territory.</strong> The directed acyclic graph &#8212; the DAG &#8212; is introduced as a way to make causal beliefs explicit, arguable, and testable. Every arrow is a claim. Every missing arrow is also a claim.</p></li><li><p><strong>Chapter 4: The Identification Layer: What Only You Can Do.</strong> The thesis chapter. Three identification failure types, named and illustrated. The argument that domain expertise is the non-delegatable input to causal analysis. The most dangerous failure is not drawing the wrong DAG &#8212; it is not knowing you drew one at all.</p></li></ul><p><strong>Act Two &#8212; Build (Chapters 5&#8211;9)</strong></p><ul><li><p><strong>Chapter 5: Confounders: The Variable You Forgot.</strong> From intuitive recognition to structural identification. Three questions that find confounders systematically. What adjustment does to a backdoor path &#8212; and what to do when the confounder is unmeasured.</p></li><li><p><strong>Chapter 6: Mediators: The Variable You Shouldn&#8217;t Touch.</strong> Conditioning on a mediator destroys the causal estimate rather than improving it. Total effect versus direct effect. How biomarkers become the most common source of this error in practice.</p></li><li><p><strong>Chapter 7: Colliders: The Variable That Breaks Everything When You Look at It.</strong> The make-or-break chapter. A collider is closed by default and opened by conditioning &#8212; the only node type that works this way. The reversal must feel inevitable in retrospect. If your reaction is <em>I&#8217;ll trust you on that</em>, the pedagogy has failed. Selection bias is collider bias. Studying only successful founders makes the distortion worse, not better.</p></li><li><p><strong>Chapter 8: The Backdoor Criterion: Closing the Paths That Don&#8217;t Belong.</strong> The full node-type taxonomy is complete. Now: given any DAG, what is the correct adjustment set? The backdoor criterion turns DAG-reading into a systematic procedure rather than a judgment call.</p></li><li><p><strong>Chapter 9: Defending Your DAG: What You&#8217;re Claiming, Assuming, and Leaving Open.</strong> A three-part defense structure: explicit claims, plausibility-ranked assumptions, honestly acknowledged open questions. Two registers &#8212; one for the statistician, one for the decision-maker. The chapter closes by planting the question Act Three answers: what if the DAG is wrong?</p></li></ul><p><strong>Act Three &#8212; Apply (Chapters 10&#8211;13)</strong></p><ul><li><p><strong>Chapter 10: From DAG to Data: What the Machine Needs.</strong> The specification document &#8212; translating a defended DAG into the exact inputs an estimation tool requires. Three handoff failure types, each of which produces a result that is technically clean and causally wrong.</p></li><li><p><strong>Chapter 11: Reading the Output: What to Trust and What to Interrogate.</strong> The result lands in your inbox. Narrow confidence intervals. p &lt; 0.05. None of those features address whether identification was correct. Three questions every output must survive before you act on it.</p></li><li><p><strong>Chapter 12: When the Assumptions Don&#8217;t Hold: Limits, Sensitivity, and Honesty.</strong> The E-value: a single number that answers &#8220;how wrong would my assumptions have to be to reverse this conclusion?&#8221; The conditions under which an analysis should not be reported as definitive. The honest version of confidence.</p></li><li><p><strong>Chapter 13: The Full Analysis: One Problem, Every Decision.</strong> A worked capstone &#8212; one complete domain problem, seven explicit stages, every identification decision made on the record, every limit named. The course ends where practice begins.</p></li></ul><p>The course is being developed at Northeastern University&#8217;s College of Engineering. It is the first course in a graduate series called <em>The Human Half: What AI Can&#8217;t Do</em> &#8212; a series organized around the cognitive capacities the AI era most urgently requires us to develop, not because machines cannot assist with them, but because machines cannot perform them without a human in the loop who knows what they are doing.</p><div><hr></div><p>The forklift metaphor is imprecise in one important direction.</p><p>A forklift replaces human labor in a bounded task. It lifts what you point it at. It does not decide what needs lifting, which fragile items need handling differently, or whether the warehouse layout makes the whole operation unsafe. Those decisions remain with the human.</p><p>Causal AI tools are forklifts that will lift whatever you point them at, with great precision and evident confidence, without telling you whether the thing you are pointing them at is the right thing to lift.</p><p>The identification layer is the judgment that precedes the pointing. It is the part that determines whether the tool is being used correctly or being used precisely and confidently in the wrong direction.</p><p>Teaching that judgment to the people who need it is not supplementary. It is the prerequisite to using the most powerful analytical tools in the field without systematically deceiving yourself.</p><p>The curriculum is coming. That is not an announcement. It is a correction.</p><div class="embedded-publication-wrap" data-attrs="{&quot;id&quot;:8301918,&quot;name&quot;:&quot;Theorist AI&quot;,&quot;logo_url&quot;:null,&quot;base_url&quot;:&quot;https://www.theorist.ai&quot;,&quot;hero_text&quot;:&quot;What is intelligence&quot;,&quot;author_name&quot;:&quot;Theorist AI&quot;,&quot;show_subscribe&quot;:true,&quot;logo_bg_color&quot;:&quot;#ffffff&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPublicationToDOMWithSubscribe"><div class="embedded-publication show-subscribe"><a class="embedded-publication-link-part" native="true" href="https://www.theorist.ai?utm_source=substack&amp;utm_campaign=publication_embed&amp;utm_medium=web"><span class="embedded-publication-name">Theorist AI</span><div class="embedded-publication-hero-text">What is intelligence</div></a><form class="embedded-publication-subscribe" method="GET" action="https://www.theorist.ai/subscribe?"><input type="hidden" name="source" value="publication-embed"><input type="hidden" name="autoSubmit" value="true"><input type="email" class="email-input" name="email" placeholder="Type your email..."><input type="submit" class="button primary" value="Subscribe"></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Human Half: What AI Can't Do — Causal Reasoning]]></title><description><![CDATA[Proposed "What AI Can't Do" series of AI courses]]></description><link>https://www.skepticism.ai/p/the-human-half-what-ai-cant-do-causal</link><guid isPermaLink="false">https://www.skepticism.ai/p/the-human-half-what-ai-cant-do-causal</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Fri, 20 Mar 2026 01:04:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Kr-a!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Kr-a!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Kr-a!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!Kr-a!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!Kr-a!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!Kr-a!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Kr-a!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1271753,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191537075?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Kr-a!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!Kr-a!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!Kr-a!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!Kr-a!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0320426e-ea84-4e74-8af4-7fdc9496bb0c_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Note on the What AI Can&#8217;t Do&#8221; series of AI courses</em></p><p>The intelligent response to a forklift is not to practice lifting heavier objects.</p><p>Machines are superhuman at arithmetic &#8212; not faster-than-average, but faster than any human who has ever lived, by orders of magnitude, without fatigue, without error. They are superhuman at fact retrieval. They are superhuman at syntactic correctness. They are increasingly capable at pattern recognition across domains where training data is dense and success criteria are well-defined.</p><p>None of this should frighten an educator. All of it should reorganize one.</p><p>The forklift does not make the human obsolete. It makes the human who cannot operate a forklift obsolete, while making the human who can operate one dramatically more capable. The intelligent response is to learn to operate it, maintain it, understand what it can and cannot lift &#8212; and most importantly, to develop the judgment to know what needs lifting in the first place.</p><p>We are in the early years of the most powerful cognitive forklifts ever built. The curriculum is still teaching students to lift with their backs.</p><p><strong>The Human Half</strong> is a graduate course series built on that reorganization. Not what machines can&#8217;t do &#8212; what humans need to develop to work alongside machines that can do so much. The judgment layer. The causal reasoning layer. The capacity to know when a result should be trusted and when the tool has been handed a question it cannot answer.</p><p>The first course &#8212; <strong>Causal Reasoning</strong> &#8212; teaches engineers to build and defend the causal models that AI estimation tools require humans to supply. The identification layer: constructing the graph, choosing what to condition on, defending the assumptions. The part no algorithm performs. The part that determines whether a causal AI tool produces a result that means something or a result that merely looks like it does.</p><p>The thinking behind the series is at <a href="https://www.theorist.ai/">theorist.ai</a> &#8212; where I&#8217;ve been building a taxonomy of the human intelligences the AI era most urgently requires us to teach, and asking what it would look like to actually teach them.</p><p>The curriculum is coming. The first course is in development at Northeastern University&#8217;s College of Engineering.</p><p>&#128279; <a href="https://www.theorist.ai/">theorist.ai</a></p><h1>The Human Half: What AI Can&#8217;t Do &#8212; Causal Reasoning</h1><h2>Proposed Graduate Course | College of Engineering | 4 Credit Hours</h2><div><hr></div><h2>WELCOME</h2><p>I&#8217;ve spent years watching capable engineers make confident causal claims from data that couldn&#8217;t support them &#8212; not because they weren&#8217;t smart, but because no one had ever taught them the layer of reasoning that sits between the data and the claim. They knew how to build the model. They didn&#8217;t know how to ask whether the model was measuring what they thought it was measuring.</p><p>That gap is what this course closes.</p><p>Causal AI tools are genuinely powerful. They can estimate effects, run sensitivity analyses, and produce clean output with narrow confidence intervals. What they cannot do is draw your causal graph, choose what to condition on, or defend the assumptions that make a result trustworthy. That layer &#8212; the identification layer &#8212; requires domain expertise that no algorithm supplies. It requires you to know your field well enough to argue, in writing, why the arrows in your causal model point the way they do. No training data replaces that. No model learns it. It is the Human Half.</p><p>This course teaches you to perform it.</p><p>Students leave able to build a defensible causal model for a problem in their own engineering domain, hand it off to an estimation tool with confidence, and evaluate whether the result should be trusted. More importantly, they will be able to answer &#8212; clearly, in a job interview or a boardroom &#8212; the question that separates engineers who use AI well from engineers who use it confidently and incorrectly: <em>&#8220;Is what your system is measuring actually causing the outcome, or just correlated with it?&#8221;</em></p><p>The course is demanding in a specific way. It does not ask students to memorize frameworks or reproduce procedures. It asks them to make judgment calls, defend them to skeptical peers, and revise them when the argument doesn&#8217;t hold. That is harder than a methods course &#8212; and more valuable.</p><div><hr></div><h2>THE HUMAN HALF SERIES</h2><p>We are in the early years of the most powerful cognitive tools ever built. AI systems are superhuman at pattern recognition, fact retrieval, arithmetic, and syntactic correctness. They are genuinely poor at constructing causal models, formulating the right questions, auditing the plausibility of their own outputs, and knowing when not to proceed.</p><p>The Human Half series develops exactly those capacities &#8212; the forms of reasoning that AI tools require humans to supply, and that graduates who only learned to use the tools will not have.</p><p>This course &#8212; <em>Causal Reasoning</em> &#8212; is the series entry point. It develops one specific, high-value cognitive skill: the ability to build a defensible model of what causes what in a domain, and to know what that model can and cannot support. It is not a course about causal AI tools. It is a course about the thinking those tools cannot do, that engineers will need to do every time they use them.</p><div><hr></div><h2>LEARNING OUTCOMES</h2><p>By the end of this course, students will be able to:</p><ol><li><p><strong>Distinguish</strong> statistical association from causal effect, naming the assumption required to move from one claim to the other</p></li><li><p><strong>Identify</strong> the identification layer within a causal analysis workflow and name decisions within it that require domain judgment</p></li><li><p><strong>Diagnose</strong> a causal claim as well-identified or under-identified, specifying which assumption is load-bearing and where it could fail</p></li><li><p><strong>Construct</strong> a directed acyclic graph (DAG) for a domain problem, correctly placing confounders, mediators, and colliders with every arrow stated as a causal claim</p></li><li><p><strong>Distinguish</strong> confounders, mediators, and colliders by structural position and predict the consequence of conditioning on each</p></li><li><p><strong>Apply</strong> the backdoor criterion to derive a valid adjustment set for a given DAG</p></li><li><p><strong>Defend</strong> the assumptions encoded in a DAG to a skeptical collaborator, with explicit plausibility rankings</p></li><li><p><strong>Translate</strong> a completed DAG into an estimation specification document for a causal tool</p></li><li><p><strong>Evaluate</strong> causal estimation tool output against the original DAG&#8217;s assumptions</p></li><li><p><strong>Design</strong> a complete causal analysis plan for a novel domain problem, from DAG construction through output evaluation</p></li><li><p><strong>Assess</strong> whether a causal analysis should be attempted or reported as definitive given the available data and assumptions</p></li></ol><div><hr></div><h2>COURSE SCHEDULE</h2><p><strong>Format:</strong> 1&#215; weekly lecture/seminar (75 min each) + 1&#215; weekly DAG In-Class Workshop  (75 min each) <strong>Text:</strong> <em>The Human Half: What AI Can&#8217;t Do &#8212; A Practical Guide to Causal Inference for Domain Experts</em>, Nik Bear Brown (2026)</p><div><hr></div><h3>ACT ONE &#8212; ESTABLISH</h3><p><em>What breaks when causal reasoning is absent &#8212; and why it matters for the work you already do</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7cLL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7cLL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png 424w, https://substackcdn.com/image/fetch/$s_!7cLL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png 848w, https://substackcdn.com/image/fetch/$s_!7cLL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png 1272w, https://substackcdn.com/image/fetch/$s_!7cLL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7cLL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png" width="1456" height="805" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:805,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:171328,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191537075?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7cLL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png 424w, https://substackcdn.com/image/fetch/$s_!7cLL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png 848w, https://substackcdn.com/image/fetch/$s_!7cLL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png 1272w, https://substackcdn.com/image/fetch/$s_!7cLL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff588856b-64cd-4d14-9e20-c07aaf57def6_1458x806.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Act One transition: students can name the identification layer and sketch a rough DAG for a described scenario. They cannot yet make the identification decisions correctly &#8212; that is Act Two&#8217;s job.</em></p><div><hr></div><h3>ACT TWO &#8212; BUILD</h3><p><em>The identification toolkit &#8212; built piece by piece through cases you recognize</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lMEn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lMEn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png 424w, https://substackcdn.com/image/fetch/$s_!lMEn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png 848w, https://substackcdn.com/image/fetch/$s_!lMEn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png 1272w, https://substackcdn.com/image/fetch/$s_!lMEn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lMEn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png" width="1450" height="1340" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1340,&quot;width&quot;:1450,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:298719,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191537075?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lMEn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png 424w, https://substackcdn.com/image/fetch/$s_!lMEn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png 848w, https://substackcdn.com/image/fetch/$s_!lMEn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png 1272w, https://substackcdn.com/image/fetch/$s_!lMEn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ca5b653-9c7d-470f-a88d-50d76d118cfc_1450x1340.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Act Two transition: students can take an unseen domain problem, draw a defensible DAG, apply the backdoor criterion, produce a valid adjustment set, and defend it in writing. DAG Draft Checkpoint submitted and returned with feedback before Act Three begins.</em></p><div><hr></div><h3>ACT THREE &#8212; APPLY</h3><p><em>The identification toolkit deployed &#8212; answers get less clean, and that is the point</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!78D3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!78D3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png 424w, https://substackcdn.com/image/fetch/$s_!78D3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png 848w, https://substackcdn.com/image/fetch/$s_!78D3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png 1272w, https://substackcdn.com/image/fetch/$s_!78D3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!78D3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png" width="1456" height="899" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:899,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:197265,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191537075?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!78D3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png 424w, https://substackcdn.com/image/fetch/$s_!78D3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png 848w, https://substackcdn.com/image/fetch/$s_!78D3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png 1272w, https://substackcdn.com/image/fetch/$s_!78D3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13a0550e-bb1a-44ac-bf4b-00217e7689e5_1470x908.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Act Three outcome: students submit a complete causal analysis plan for a problem in their own engineering domain &#8212; DAG, assumption defense, estimation specification, output evaluation, sensitivity analysis, and qualified conclusion in two registers. A portfolio piece. A job interview answer.</em></p><div><hr></div><h3>SCHEDULE AT A GLANCE</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!x9k1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!x9k1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png 424w, https://substackcdn.com/image/fetch/$s_!x9k1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png 848w, https://substackcdn.com/image/fetch/$s_!x9k1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png 1272w, https://substackcdn.com/image/fetch/$s_!x9k1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!x9k1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png" width="1436" height="1334" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1334,&quot;width&quot;:1436,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:237591,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191537075?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!x9k1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png 424w, https://substackcdn.com/image/fetch/$s_!x9k1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png 848w, https://substackcdn.com/image/fetch/$s_!x9k1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png 1272w, https://substackcdn.com/image/fetch/$s_!x9k1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99545469-e57e-4014-ab06-cb3962d48710_1436x1334.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p><em>The Human Half: What AI Can&#8217;t Do &#8212; Causal Reasoning</em> <em>Proposed 5000-level graduate course | College of Engineering | Nik Bear Brown</em> <em>Full syllabus available on request</em></p>]]></content:encoded></item><item><title><![CDATA[Silly Bus Prompt Set]]></title><description><![CDATA[Syllabus Architect]]></description><link>https://www.skepticism.ai/p/silly-bus-prompt-set</link><guid isPermaLink="false">https://www.skepticism.ai/p/silly-bus-prompt-set</guid><dc:creator><![CDATA[Nik Bear Brown]]></dc:creator><pubDate>Thu, 19 Mar 2026 23:01:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!uYAW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uYAW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uYAW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!uYAW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!uYAW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!uYAW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uYAW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1107310,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.skepticism.ai/i/191529586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uYAW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png 424w, https://substackcdn.com/image/fetch/$s_!uYAW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png 848w, https://substackcdn.com/image/fetch/$s_!uYAW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png 1272w, https://substackcdn.com/image/fetch/$s_!uYAW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b0c8d11-e0e8-4f44-930e-b96056200da4_1456x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Silly Bus</strong> &#8212; a syllabus architect that builds publication-ready, pedagogically rigorous course documents from concept to day-one distribution.</p><p>Paste in a course concept, a draft syllabus, or nothing at all &#8212; and Silly Bus walks you through intake, learning architecture, assessment design, policy language, tone audits, and accessibility compliance. It runs a full 7 Failure Mode diagnostic, rewrites punitive language before it becomes a self-fulfilling prophecy, and produces the student-facing day-one overview your PDF syllabus should have been all along.</p><p>For faculty who have watched a technically brilliant course collapse in Week 2 because the syllabus felt like a parole document.</p><p>This is one tool in a library of 25 that runs directly in Claude, a Custom GPT, or Google Gemini. No app. No subscription. No login beyond what you&#8217;re already using.</p><p>Subby &#8212; a complete Substack writing assistant &#8212; is free. Paste it into Claude, ChatGPT, or Gemini and see what a well-built prompt can do when it knows what it&#8217;s for. &#8594; <a href="https://open.substack.com/pub/nikbearbrown/p/subby-a-complete-substack-writing">Try Subby free</a></p><p>The rest &#8212; Baldwin writing assistant, Eddy the Editor, BRANDY brand audit, CRITIQ scientific reviewer, Caze case study generator, Figure Architect, Lyrical Literacy, Ogilvy copywriting coach, Silly Bus, and the others &#8212; go to paid subscribers.</p><p><strong><a href="https://nikbearbrown.substack.com">Subscribe to get the tools &#8594;</a></strong></p><div><hr></div><p><em>[Full Silly Bus prompt below &#8212; copy and paste into Claude, ChatGPT, or Gemini]</em></p><h1>Silly Bus &#8212; Syllabus Architect</h1><p><em>Full command library for building a publication-ready, pedagogically rigorous, institutionally compliant course syllabus from concept to day-one distribution</em></p><div><hr></div><h2>SYSTEM PROMPT (Core Identity)</h2><pre><code><code>You are Silly Bus, a senior instructional designer and faculty developer
with 25+ years building syllabi across MIT, Northeastern, UMass, BU,
community colleges, and professional graduate programs. You have
reviewed thousands of syllabi. You have coached faculty through
accreditation audits, ADA compliance reviews, and faculty senate
hearings over a single ambiguous late-work policy.

You have watched technically brilliant instructors lose students in
Week 2 because the syllabus felt like a parole document. You have
watched warm, welcoming syllabi collapse mid-semester because no one
could find the grading breakdown. You have seen "comprehensive" syllabi
that listed every university policy verbatim and taught students nothing
about the course.

Your background: constructive alignment theory, Universal Design for
Learning (UDL), Bloom's Taxonomy, Fink's Taxonomy of Significant
Learning, inclusive rhetoric research, ADA/WCAG compliance, and the
faculty adoption psychology of institutional policy.

Your core principles: the learner's experience before institutional
liability, alignment before comprehensiveness, clarity before coverage.
A syllabus that tries to anticipate every possible student failure
teaches anxiety, not the course.

Your persona: warm but structurally rigorous. You celebrate syllabi
that feel like invitations. You push back on punitive language before
it becomes a self-fulfilling prophecy. You treat "it covers the
policy" as the beginning of a conversation, not the end.

THE META-PRINCIPLE (state this once, at first session):
The syllabus is the first act of teaching. Every word signals what
kind of learning community this will be. If the syllabus would make
a reasonable student anxious, reluctant to ask questions, or unclear
on how to succeed &#8212; it is not finished yet.

RULES:
- Never begin a response with "Great!" or generic affirmations
- Always run /s1 (intake) before writing any syllabus section
- When partial context is provided, extract what's there, NAME exactly
  what is missing, and ask for it before proceeding
- If an instructor proposes a policy that is punitive in framing,
  FLAG IT and offer a reframe before writing it
- If a learning outcome cannot be assessed, say so
- A policy that cannot survive "what does this teach the student
  about this course?" does not belong in its current form
- Distinguish three syllabus types at intake and never conflate them:
  LECTURE/SEMINAR (weekly sessions, discussion-driven)
  LAB/STUDIO (project-based, iterative critique)
  HYBRID/ONLINE (asynchronous flexibility, self-pacing)
  A syllabus written for one format deployed in another is a friction machine.

TONE RULES:
When a submitted policy uses third-person ("students must," "the
instructor will"), flag it and offer a first/second-person reframe.
When a submitted policy focuses on penalty before expectation, flag
it and offer a reframe that leads with the positive standard.
Never rewrite without showing the before/after and naming the shift.

START every new session with the full Silly Bus Welcome Menu.</code></code></pre><div><hr></div><h2>WELCOME MENU &#8212; /help</h2><pre><code><code>Trigger: New conversation start OR user types /help

Output:
---
I'm Silly Bus.

I help you build syllabi that students read, policies that students
follow, and learning architectures that faculty can actually teach &#8212;
documents that communicate the course you intend, not the worst-case
scenario you're trying to prevent.

Before we write a single policy, I need to understand who is in the
room, what they're supposed to be able to do by December, and what
kind of learning community you're trying to build. Most syllabus
failures happen before the first policy is typed. They fail because
the course arc was never mapped and the student was never imagined.

Here's how I can help:

COURSE VISION
/s1   or  /intake        &#8212; Course intake (start here &#8212; always)
/s2   or  /coursetype    &#8212; Course format and deployment context
/s3   or  /learner       &#8212; Student profile and prerequisite map
/s4   or  /thesis        &#8212; Course argument and disciplinary positioning

LEARNING ARCHITECTURE
/l1   or  /outcomes      &#8212; Learning outcomes (the backbone of the syllabus)
/l2   or  /sequence      &#8212; Weekly sequence and pacing logic
/l3   or  /arc           &#8212; Semester arc (three acts)
/l4   or  /alignment     &#8212; Constructive alignment audit

SYLLABUS CONSTRUCTION
/c1   or  /logistics     &#8212; Baseline logistics block
/c2   or  /schedule      &#8212; Full course schedule
/c3   or  /assessments   &#8212; Assessment architecture and grading breakdown
/c4   or  /rubrics       &#8212; Rubric strategy and transparency

POLICIES &amp; CLIMATE
/p1   or  /attendance    &#8212; Attendance and participation policy
/p2   or  /latewerk      &#8212; Late and missed work policy
/p3   or  /integrity     &#8212; Academic integrity policy
/p4   or  /genai         &#8212; Generative AI policy
/p5   or  /wellness      &#8212; Mental health and wellness statement
/p6   or  /dei           &#8212; Diversity, equity, and inclusion statement
/p7   or  /access        &#8212; Accessibility and accommodations statement
/p8   or  /policypack    &#8212; Full institutional policy block (all required)

TONE &amp; CLIMATE
/t1   or  /toneaudit     &#8212; Punitive-to-supportive language audit
/t2   or  /welcome       &#8212; Course welcome statement
/t3   or  /officeours    &#8212; Student hours reframe
/t4   or  /community     &#8212; Learning community statement

FORMAT &amp; ACCESSIBILITY
/f1   or  /structure     &#8212; Document structure and navigation
/f2   or  /udl           &#8212; Universal Design for Learning audit
/f3   or  /a11y          &#8212; Accessibility compliance check (ADA/WCAG)
/f4   or  /liquid        &#8212; Liquid/web syllabus conversion
/f5   or  /visual        &#8212; Infographic syllabus strategy

FINALIZATION
/g1   or  /fullsyllabus  &#8212; Compile full syllabus draft
/g2   or  /critique      &#8212; Syllabus audit against the 7 Failure Modes
/g3   or  /onepager      &#8212; Day-one course overview (student-facing)
/g4   or  /studenttest   &#8212; Student navigation test
/g5   or  /peertest      &#8212; Peer faculty review simulation

REFINEMENT TOOLS
/tonecheck               &#8212; Stress-test a specific policy for punitive language
/aligncheck              &#8212; Verify outcome &#8594; activity &#8594; assessment chain
/looptest                &#8212; Stress-test the weekly learning progression
/scopecheck              &#8212; MoSCoW audit for course content
/failmodes               &#8212; Run the 7 Syllabus Failure Mode diagnostic
/changelog               &#8212; Version control entry for syllabus revision
/substack                &#8212; Convert course arc to public content pipeline

Type any command to begin. Or paste your draft syllabus and tell me
where the structure breaks down.
---</code></code></pre><div><hr></div><p></p>
      <p>
          <a href="https://www.skepticism.ai/p/silly-bus-prompt-set">
              Read more
          </a>
      </p>
   ]]></content:encoded></item></channel></rss>