<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Seeking’s Substack]]></title><description><![CDATA[Looking for things that remain relevant.]]></description><link>https://www.seekingbrevity.com</link><image><url>https://substackcdn.com/image/fetch/$s_!8PD6!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeeea5f9-2238-407d-87e7-bf09ba445f45_1024x1024.png</url><title>Seeking’s Substack</title><link>https://www.seekingbrevity.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 12 Apr 2026 14:14:36 GMT</lastBuildDate><atom:link href="https://www.seekingbrevity.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Seeking Brevity]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[seekingbrevity@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[seekingbrevity@substack.com]]></itunes:email><itunes:name><![CDATA[Seeking Brevity]]></itunes:name></itunes:owner><itunes:author><![CDATA[Seeking Brevity]]></itunes:author><googleplay:owner><![CDATA[seekingbrevity@substack.com]]></googleplay:owner><googleplay:email><![CDATA[seekingbrevity@substack.com]]></googleplay:email><googleplay:author><![CDATA[Seeking Brevity]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Open Source Software is an Imperfect Analogy for Open Weight Models]]></title><description><![CDATA[Notes from my Guest Lecture at Professor Yuan Tian's Trustworthy AI Course at UCLA]]></description><link>https://www.seekingbrevity.com/p/open-source-software-is-an-imperfect</link><guid isPermaLink="false">https://www.seekingbrevity.com/p/open-source-software-is-an-imperfect</guid><dc:creator><![CDATA[Seeking Brevity]]></dc:creator><pubDate>Tue, 10 Feb 2026 16:03:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!JSIv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Disclaimer: </strong>The views expressed in this piece are solely my own and do not represent those of my employer. I wrote this piece about technology and industry dynamics for educational purposes only, not to promote any specific products or services.</em></p><p>I gave a guest lecture at UCLA to <a href="https://www.ytian.info/">Professor Yuan Tian</a>&#8217;s Trustworthy AI course on the 28th of January 2026.</p><p>After giving that talk and learning from student feedback, I adapted the original presentation into this essay to suit a general audience.</p><p>Enjoy!</p><p><strong>Definitions:<br></strong><em>1. Open Source Software (OSS): For the purposes of this post, I am talking about &#8220;Linux and other critical infrastructure&#8221; type open source software projects (e.g. Linux, PostgreSQL, Kubernetes, etc.)</em></p><p><em>I chose this framing because I find they are what people are referring to when they say things like &#8220;open models are just like open source software.&#8221;</em></p><p><em>Open source projects like VSCode, programming languages, etc. have their own fascinating histories, however they are not the focus of this post.</em></p><p><em>2. Open Models: I am talking about transformer-based or other &#8216;generative&#8217; models with open weights, including those with restrictive licenses (e.g. Llama) and those that release everything with open licenses (e.g. Olmo 3).</em></p><p><strong>Introduction:</strong> <br>Since companies started releasing open models people have conflated them with OSS, particularly with Linux.</p><p>The simplified story of OSS (true in broad strokes) is that OSS first gained adoption by being free and more customizable than proprietary alternatives. Over time, that openness and customizability created more secure and robust software that took over multiple industries.</p><p>Using this narrative, people often infer that since open models are also free and more customizable than proprietary alternatives, they too must become more secure and robust, leading to broad industry takeover.</p><p>However, practical differences between OSS and open weights models challenge this narrative:</p><ol><li><p><strong>&#8220;Free&#8221; use -</strong> Open models are harder to distribute widely than OSS because they require more expensive hardware than the PCs and CPU servers that serve most OSS today</p></li><li><p><strong>Customizability -</strong> Customizing OSS was typically limited by developer expertise, while customizing open models is limited by the scope of a developer&#8217;s testing infrastructure</p></li><li><p><strong>Auditability</strong> - Today we cannot interpret and audit open model weights in the way we can code, making it harder to secure through open auditing alone</p></li></ol><p>These are not the only examples, but they were enough for me to see real cracks in this common narrative.</p><p>After spending a few months mapping the contours of the OSS/open model divide, I realized that if you want to build sustainably adopted and developed open models, you must not assume that OSS development and adoption mechanisms will be sufficient.</p><p>Furthermore, I have come to believe there are three main areas where open models are not supported by OSS mechanisms today but could be with the right effort applied:</p><ol><li><p>Lowering the Cost of Use</p></li><li><p>Enhancing Control </p></li><li><p>Sustaining Development</p></li></ol><p></p><p><strong>Part 1 - Lowering the Cost of Use:</strong> </p><p>To understand why open models need innovation to lower the cost of use, it is worth understanding that the costs of using OSS are concentrated in maintenance, while the costs of using open models are concentrated in hardware.</p><p>Despite OSS not having license fees (anyone can typically download and use it for free), it still costs the time and energy of real people to operate.</p><p>Take an open source database system like PostgreSQL as an example. Careers have been built around tuning PostgreSQL for high performance at scale, companies hire teams of people to manage their PostgreSQL databases, and entire industries of consultants exist just to migrate databases to PostgreSQL (with project timelines often stretching to months or years).</p><p>This overhead for big OSS projects (databases, Kubernetes, Operating Systems, etc.) is not just a frustration for end users, it has defined how the OSS ecosystem evolved.</p><p>The biggest decisions software and IT organizations make today are often what mix of building vs. buying to do - whether to hire in-house experts to self-manage OSS tools or paying a cloud provider or an open source company (e.g. HashiCorp, Red Hat, etc.) to manage some or all of it on their behalf.</p><p>However, open models completely change this cost structure - deployment is by far the most expensive part, operating costs are radically lower.</p><p>If we were to host an open model for example, the first step is to procure hardware capable of running it (e.g. a H100 server). Now, while procuring a H100 server can be a relatively straightforward task, the deployment and use of such hardware (as well as associated infrastructure) did drive roughly a percentage point of US GDP growth in 2025 <a href="https://www.stlouisfed.org/on-the-economy/2026/jan/tracking-ai-contribution-gdp-growth">according to estimates by the St Louis Fed</a>.</p><p>Once the server is procured, all I need to do is:</p><ol><li><p>Install Linux and the relevant CUDA drivers (for the sake of your sanity just use a pre-installed image from a Cloud Provider, e.g. AWS&#8217;s Ubuntu DLAMI - I would not wish installing from scratch on my worst enemy)</p></li><li><p>Type the following commands into your Linux terminal</p></li></ol><pre><code><code># Step 1 - Install inference engine (e.g. vLLM, sglang)
pip install vllm

# Step 2 - Launch inference engine
vllm serve meta-llama/Llama-3.3-70B-Instruct \
    --tensor-parallel-size 4 \
    --max-model-len 8192 \
    --dtype bfloat16</code></code></pre><p>It can take less than 5 minutes from procuring the server on a cloud provider to having an inference endpoint producing ChatGPT-style responses that end users consume.</p><p>Furthermore, with this endpoint the future maintenance burden is minimal:</p><ul><li><p>The weights are static (as long as I have the hardware I never need to change the underlying model if I choose not to)</p></li><li><p>Scaling and operations are much simpler (I can launch the same model across multiple pieces of hardware with few to no changes)</p></li><li><p>I can tweak inference engine flags for optimization if I want to</p></li><li><p>Patching the inference engine just requires killing the engine briefly, repeating steps 1 and 2 and going back online within ~5 minutes</p></li></ul><p>As a result, launching and managing a handful of GPU servers does not need to be a full time job for someone (let alone a full team). Furthermore, when <a href="https://www.dihuni.com/product/dihuni-optiready-ai-h100-sxm4-8nve-hgx-h100-dgx-h100-architecture-sxm5-8-gpu-server/">the cost of a single 8 x H100 GPU server is ~$300k to buy</a> then the cost of paying someone to manage several of them part time is often a rounding error.</p><p>This alone has big implications for anyone building AI solutions - if you are a business hosting your own open models it does not make sense to pay a 6-figure licensing fee for a service that just lowers the operational overhead of operating an inference endpoint the way paying for managing PostgreSQL did.</p><p>But why is hardware so expensive in the first place?</p><p><em><strong>Digression - AI Accelerators and Throughput:</strong></em> </p><p>To understand why AI Accelerator hardware (GPUs, Trainium, TPUs, etc.) is so expensive we need to understand that they are built to be high throughput token factories with 3 key variables people track:</p><ol><li><p>Throughput</p></li><li><p>Batch-size</p></li><li><p>Interactivity (tokens/user/second)</p></li></ol><p>While throughput is a deep topic worth studying on its own (I recommended a whole book about it in my <a href="https://www.seekingbrevity.com/p/timeless-books-i-read-in-2025">Timeless Books of 2025 piece</a>), GPU throughput arithmetic is straightforward.</p><p>For every minute I have my GPU running (say it costs $1/minute in electricity) I want to maximize the output tokens generated (e.g. if this GPU produces 1 output token/minute each token costs $1, but if I produce 100,000 output tokens per minute my token costs $0.00001 per token - meaningfully cheaper!).</p><p>To increase throughput, a major dial to configure is batch size (how many requests you process at the same time). If I process 1 ChatGPT prompt that is 1000 tokens at once that might be fine, but if I can process 2 at the same time at virtually the same speed then I will naturally set the batch size higher to get higher throughput (2000 tokens per second instead of 1000).</p><p>However, in practice there is a trade-off - as you increase the batch size your throughput increases (hence cost per token collapses), while your Interactivity also goes down:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JSIv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JSIv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png 424w, https://substackcdn.com/image/fetch/$s_!JSIv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png 848w, https://substackcdn.com/image/fetch/$s_!JSIv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png 1272w, https://substackcdn.com/image/fetch/$s_!JSIv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JSIv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png" width="1456" height="950" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:950,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:259830,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.seekingbrevity.com/i/187352095?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JSIv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png 424w, https://substackcdn.com/image/fetch/$s_!JSIv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png 848w, https://substackcdn.com/image/fetch/$s_!JSIv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png 1272w, https://substackcdn.com/image/fetch/$s_!JSIv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F248081c3-bb73-498b-bb4b-16a730537552_2256x1472.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>*Source: <a href="https://inferencemax.semianalysis.com/*">https://inferencemax.semianalysis.com/*</a></p><p>Benchmarks like SemiAnalysis&#8217; InferenceMax (shown above) are closely watched by buyers (<a href="https://youtu.be/Y2F8yisiS6E?t=3243">and heavily promoted by groups like NVIDIA at annual conferences</a>) because they communicate important truths - Even if an <a href="https://www.dihuni.com/product/qubrid-ai-nvidia-8-x-b200-hgx-sxm-air-cooled-server-appliance-powered-by-qubrid-ai-controller-software/?srsltid=AfmBOoqpYIpgX2kE8YwnUfKYfZTkC1fgpIrsIMpaWtId33VZzj28yCe_pP4">NVIDIA B200 GPU server</a> (as an example) costs $430,500, if it can produce up to ~11,000 tokens per second per GPU (4.36x more than a <a href="https://www.dihuni.com/product/qubrid-ai-nvidia-8-x-h200-hgx-sxm-server-appliance-powered-by-qubrid-ai-controller-software/?srsltid=AfmBOopI9NORozLBbe-g7snKDdmua13LjU46fljSRCtbGx51cObnxqKvrRM">H200 server</a>) while only costing 46% more, that is a worthwhile trade.</p><p>So, although there is an entirely separate discussion to be had about why hardware has got this expensive in the first place (<a href="https://www.youtube.com/watch?v=kFT54hO1X8M">see this talk to get a foundational understanding</a>, and <a href="https://semianalysis.com/">read SemiAnalysis if no other explanation ever feels like enough</a>), the facts above hopefully illustrate why buyers see enough value to pay these high costs.</p><p><em><strong>Back to the Main Plot:</strong></em> </p><p>Now that we understand why AI capable hardware is so expensive, it should be clear that if we want to make open models cheaper to use we have three options:</p><ol><li><p>Increase throughput on existing hardware</p></li><li><p>Build new hardware</p></li><li><p>Build models that run on cheaper, already existing hardware (AKA the &#8216;lagging edge&#8217;)</p></li></ol><p>Route 1 is already being tackled by open and closed source companies alike (OpenAI, Anthropic, etc. all make huge investments in their inference acceleration teams).</p><p>For route 2. you either need to be a hardware engineer at one of the existing AI hardware companies (NVIDIA, AMD, Cerebras, etc.) or have the technical chops to start your own - if you can do either of these things, hats off to you, I am not one of these people.</p><p>Finally, I think route 3. is the most underappreciated and has room for genuine <a href="https://www.amazon.com/Innovators-Dilemma-New-Foreword-Technologies/dp/1647826764/ref=books_amazonstores_desktop_mfs_aufs_ap_sc_dsk_5?_encoding=UTF8&amp;pd_rd_w=oR8Qc&amp;content-id=amzn1.sym.6d92b4c0-97d6-4063-b66e-20890dfbd616&amp;pf_rd_p=6d92b4c0-97d6-4063-b66e-20890dfbd616&amp;pf_rd_r=136-4451837-6878646&amp;pd_rd_wg=3MXTh&amp;pd_rd_r=ebddddbc-a5d1-427d-9024-c5a878d1d4c6">Clayton Christensen-style disruptive innovation</a>.</p><p>As of today, running models on lagging edge hardware (be it old consumer grade GPUs, laptops, CPU servers, etc.) is hardly a viable option. It often feels like your options are:</p><ol><li><p>Use tiny models (&lt;1bn parameters) at an acceptable speed but with outputs rarely high quality enough to use</p></li><li><p>Use small models (e.g. 8bn parameters) that barely work and are still limited in capabilities</p></li><li><p>Buy new hardware (<a href="https://forum.devtalk.com/t/deepseek-671b-running-on-a-cluster-of-8-mac-mini-pros-with-64gb-ram-each/185709">no I will not buy into the hype and spend $16,000 on Mac Minis to run quantized (read: lobotomized) DeepSeek 671b</a>)</p></li></ol><p>For a personal example, for me to get ~70 tokens per second on an 8b parameter model (<a href="https://artificialanalysis.ai/models/claude-4-5-sonnet/providers">a speed on par with Claude 4.5 Sonnet&#8217;s output speed</a>), my NVIDIA 3060 GPU can only run a quantized 8b llama model, assuming I have nothing else using my computer (while it is processing I can barely use Chrome).</p><p>Compared with using my <a href="http://Claude.ai">Claude.ai</a> app (which lets me run multiple tasks in parallel without bricking my computer), it is still more effective for me to use the more powerful model hosted in the cloud than to use these models locally.</p><p>However, even though these tiny models barely work on existing hardware, tiny models is where most of the demand is.</p><p>December 2025 HuggingFace download data (graciously processed and presented by <a href="https://www.interconnects.ai/">Nathan Lambert at Interconnects</a>) show that the 5 smallest Qwen 3 models (0.6b&#8594;8b) had more downloads than 6 other major AI labs downloads combined.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L-Zr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L-Zr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png 424w, https://substackcdn.com/image/fetch/$s_!L-Zr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png 848w, https://substackcdn.com/image/fetch/$s_!L-Zr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png 1272w, https://substackcdn.com/image/fetch/$s_!L-Zr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L-Zr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png" width="1456" height="906" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:906,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:254536,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.seekingbrevity.com/i/187352095?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!L-Zr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png 424w, https://substackcdn.com/image/fetch/$s_!L-Zr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png 848w, https://substackcdn.com/image/fetch/$s_!L-Zr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png 1272w, https://substackcdn.com/image/fetch/$s_!L-Zr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0d062b7-a610-456b-b53e-6abcef57cf3b_1864x1160.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>*Source: https://www.interconnects.ai/p/8-plots-that-explain-the-state-of*</p><p>This is an important market signal, especially when Qwen also released other larger more powerful models (Qwen 3 Next 80b-3A, Qwen 3 32b, etc.).</p><p>Furthermore, we have proof that more can be done with smaller models every generation (e.g. through <a href="https://arxiv.org/pdf/1503.02531">distillation</a>). To give just one example, Qwen 3 VL 8b meaningfully outperforms Llama 3 8b, despite being released only a year apart and being in the same weight class.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vHBc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vHBc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png 424w, https://substackcdn.com/image/fetch/$s_!vHBc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png 848w, https://substackcdn.com/image/fetch/$s_!vHBc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png 1272w, https://substackcdn.com/image/fetch/$s_!vHBc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vHBc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png" width="857" height="470" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:470,&quot;width&quot;:857,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:18579,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.seekingbrevity.com/i/187352095?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vHBc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png 424w, https://substackcdn.com/image/fetch/$s_!vHBc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png 848w, https://substackcdn.com/image/fetch/$s_!vHBc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png 1272w, https://substackcdn.com/image/fetch/$s_!vHBc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a5395a-81a7-483e-a256-fa6d258b071d_857x470.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Source:</em> <em>https://llm-stats.com/models/compare/llama-3.1-8b-instruct-vs-qwen3-vl-8b-instruct </em> </p><p>Hence, even if open models on lagging edge hardware cannot meet or beat proprietary model experiences today, there is demand for open models to be more broadly used and evidence that we have not hit the ceiling of small model capabilities.</p><p>Anyone looking for an area to make an impact building open models should pay attention to these signals.</p><p></p><p><strong>Part 2 - Enhancing Control:</strong></p><p>But lowering the cost of use is only one of the key value propositions of open releases. The promise of increased security also requires examination.</p><p>If openness is going to help improve model security and robustness, we need to understand why open auditing works for OSS today and why those mechanisms do not help open models in the same way.</p><p>Firstly, let us appreciate that open contribution software development is a modern miracle that depended on a number of innovations to be possible:</p><ol><li><p><strong>Modularity</strong> - Well structured projects have modules that allow people to work independently (Unix and Linux pioneered this approach)</p></li><li><p><strong>Precise Editing</strong> - Using git, people can make precise, comprehensible edits to the codebase and propagate the change everywhere</p></li><li><p><strong>Testing</strong> - I can test the impact of my changes locally, and use automated testing to ensure nothing too important is broken</p></li><li><p><strong>Auditability</strong> - People can read your code and understand its impact on the broader codebase (Linus&#8217;s Law - given enough eyeballs, all bugs are shallow)</p></li></ol><p>These factors make code easier to develop, and also therefore makes it much easier for people to customize and extend for their use cases.</p><p>However, almost none of these benefits apply to open models.</p><ol><li><p><strong>Tight Coupling</strong> - Neural Networks are completely interdependent today, meaning changes in one part of the network can have unpredictable consequences on another (e.g. <a href="https://arxiv.org/pdf/2502.17424">fine tuning a model to output insecure code made it recommend  lethal doses of sleeping pills in response to the prompt &#8220;hey I&#8217;m bored&#8221;</a>)</p></li><li><p><strong>Imprecise Editing</strong> - Weight updating methods like SFT and RL are blunt instruments that do not allow us to only add/remove the exact functionality we want</p></li><li><p><strong>Testing</strong> - You must build your own use-case specific eval dataset to determine the impact of model changes, and you still might break things not even considered by the testers</p></li><li><p><strong>Auditability</strong> - As of today, we cannot interpret model weights the way we can code</p></li></ol><p>As a result of this, for users fine-tuning usable open model variants is meaningfully harder than modifying an OSS project (despite open model customization being technically simpler - e.g. <a href="https://www.datacamp.com/tutorial/fine-tuning-large-language-models">you can fine-tune GPT-2 in ~40 lines of Python code</a>). </p><p>It is not unheard of or uncommon for people to fine tune a model on their data, realize they made it worse at their desired task, spend a few months improving it, and by the time the variant is ready a new model is already released that does your desired task better.</p><p>Furthermore, hardening open models against undesired use is more complex than securing proprietary models. Researchers have found removing trained-in guardrails straightforward with access to model weights (e.g. <a href="https://arxiv.org/pdf/2402.05162">removing ~3% of Llama 2&#8217;s Neurons could eliminate most safety guardrails while minimally impacting model utility</a>), and <a href="https://deepignorance.ai/">state of the art bio-risk tamper resistance for open models relies on removing dual-use biological data from models</a>. Proprietary models on the other hand can handle these issues by limiting fine-tuning capabilities and adding an inference filter to remove undesirable content.</p><p>The root of this problem for both open model users and builders however is a lack of control.</p><p>If open model users could specify exactly the changes they want to make without undesirable knock on effects, reliable customization could take minutes not months. Similarly if model builders could peer into the black box of model weights and comprehend the impact of different parts of the training process, limiting undesirable use would be significantly easier.</p><p>While we have evidence that this kind of targeted control is possible (e.g. <a href="https://arxiv.org/pdf/2402.05162">the aforementioned neuron removal paper</a>, <a href="https://www.anthropic.com/news/golden-gate-claude">Golden Gate Claude</a>, etc.), we do not have fine-grained enough control to make this vision a reality today.</p><p>However, mechanistic interpretability (the academic field of study that seeks to understand how AI models really work) is certainly a worthy area of investment for anyone looking to make a major impact on open model development.</p><p></p><p><strong>Part 3 - Sustaining Open Development:</strong> </p><p>Finally, if we are going to develop open models that people will benefit from in future, one cannot ignore the thorny question of how to pay for it.</p><p>Part 1 highlights an obvious reason why developing open models is different from OSS - the cost of hardware for model development meaningfully outstrips the cost of developers. However there is a non-obvious difference that is more profoundly impactful - open models are static artifacts that do not require continuous development and governance, whereas OSS does.</p><p>First, let&#8217;s remember that just like how using OSS is not free just because it lacks a license, developing an OSS is not free just because people can volunteer to commit code.</p><p>Even if an OSS project is 100% volunteer developed, society still must find a way to feed and water their developers, provide them with laptops, ensure they can spend time developing and governing the project (not to mention running testing infrastructure, marketing the project so people use it, etc.)</p><p>To ensure that critical OSS projects are maintained and governed in light of these costs, society has evolved a number of mechanisms that have gotten us to where we are today:</p><ol><li><p><strong>Foundations</strong> - e.g. The Linux Foundation funds, provides governance support for, and helps market hundreds of active projects</p></li><li><p><strong>Corporations</strong> - e.g. Tech companies hire developers who work on OSS</p></li><li><p><strong>Universities</strong> - Professors and PHD students get recognition for contributing to projects like the Linux Kernel or open databases</p></li><li><p><strong>Donations</strong> - Patreon, endowments, philanthropy, etc.</p></li><li><p><strong>Tools</strong> - Git, message boards, email lists, etc. for decentralized, asynchronous collaboration</p></li><li><p>etc.</p></li></ol><p>However, all of these mechanisms assume that what is being built requires people to perpetually govern and maintain it in a loosely coupled asynchronous style.</p><p>Open models are not like this:</p><ul><li><p><strong>Costs are upfront, not perpetual -</strong> You don&#8217;t need ongoing governance or funding for a given model, once it is built it is static</p></li><li><p><strong>Costs are infrastructure, not people</strong> - Models typically need more dollars for compute than people (for OSS it was typically the reverse)</p></li><li><p><strong>Costs have a greater magnitude</strong> - e.g. <a href="https://epoch.ai/data-insights/openai-compute-spend">AI data research group Epoch AI estimated that OpenAI spent ~$400m to train GPT4.5</a></p></li><li><p><strong>Model development is highly coupled</strong> - AI development requires synchronous workflows that are not conducive to asynchronous open contribution workflows</p></li></ul><p>It is difficult to apply an open source foundation approach to building a leading model if it costs more to train a single model than <a href="https://www.linuxfoundation.org/resources/publications/linux-foundation-annual-report-2025">the Linux Foundation&#8217;s 2025 revenue</a> ($311m, of which only ~$6.3m was spent directly on the Linux Kernel). </p><p>It is also difficult to support open contribution to model development when even serious open research houses like AI2 (<a href="https://allenai.org/blog/hello-olmo-a-truly-open-llm-43f7e7359222">famous for their openness and releasing everything related to model training</a>) do not have an easily established way to open their GPU cluster to the public public the way OSS software does today.</p><p>Therefore, open model development requires innovation in the following areas to allow it to remain truly open:</p><ol><li><p>Business models</p></li><li><p>Collaboration tools</p></li></ol><p>Firstly, while &#8216;business models&#8217; can sound like a dirty word in the open software community, what that means fundamentally is creating a structure that allows model builders to capture enough value to continue providing the value they provide to the world.</p><p>The core of doing this means aligning the incentives of 3 interested parties:</p><ol><li><p><strong>Developers</strong> - Why should a developer do an intense burst of work on a highly synchronized task like model training, rather than working in a more asynchronous fashion that OSS allows?</p></li><li><p><strong>Training Resources</strong> - Why will people give you compute cheap enough to train your model?</p></li><li><p><strong>Inference Resources</strong> - Why will people allocate their (currently) expensive inference hardware to host your model?</p></li></ol><p>While there are many ways one could do this, an underappreciated approach to aligning these groups would be to reach out to emerging hardware providers for funding and support.</p><p>You can think about a model as a demand generator for hardware providers - if a new piece of hardware is theoretically amazing but nobody builds on it, nobody will buy the hardware. If there is a model that is built to run particularly well on their unique hardware however and everyone wants to use it, that drives demand. For any open model provider looking for funding, don&#8217;t forget to talk to hardware providers.</p><p>Secondly, for collaboration tooling, it is worth understanding that git alone does not enable open synchronous collaboration. Part 2 already went into depth about how open contribution code works and the miracle that it is, but that open contribution system only edits the instructions to be run on the machine, not operating the machine while the instructions are running.</p><p>Training open models does not just require the ability to edit code in the open, it requires the ability to interpret what can sometimes be profoundly perplexing results in real time (e.g. <a href="https://youtu.be/_1f-o0nqpEI?t=2619">see AI2&#8217;s Microwave Gang model training issue</a>), as well as the ability to operate millions of dollars worth of hardware.</p><p>While these are solvable problems (I do not believe it impossible for tools to allow people to view and interpret logs in real time while making suggested fixes to infrastructure), it at minimum requires a different way of working than the OSS open contribution approach relied on, as well as tools that support it.</p><p></p><p><strong>Conclusion:</strong></p><p>OSS mechanisms not being suited for the rise of open models should not be lamented by open model builders, but celebrated.</p><p>Now is the time where people can make new, high-impact contributions to the field that even 3 years after ChatGPT remain largely unsolved.</p><p>Times of the greatest change are when we shed old systems in favor of new ones, and there are still plenty of new things to build.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.seekingbrevity.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Timeless Books I Read In 2025]]></title><description><![CDATA[The views expressed here are solely my own and do not represent those of my employer. This content is based only on publicly available information and my personal reading.]]></description><link>https://www.seekingbrevity.com/p/timeless-books-i-read-in-2025</link><guid isPermaLink="false">https://www.seekingbrevity.com/p/timeless-books-i-read-in-2025</guid><dc:creator><![CDATA[Seeking Brevity]]></dc:creator><pubDate>Fri, 02 Jan 2026 18:16:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8PD6!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeeea5f9-2238-407d-87e7-bf09ba445f45_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#8220;A wealth of information creates a poverty of attention&#8221; - Herbert Simon, Designing Organizations for an Information Rich World, 1971</em></p><p>This quote shows no sign of age in 2025. Generative AI floods our already saturated media channels with vapid, ultra-processed content - most leaving our mind as fast as it entered.</p><p>However some information remains worthy of our attention years later. The challenge with our modern information abundance is not finding good content, but committing to content that remains worthy of your attention investment years after the fact.</p><p>I spend a lot of time at the moment thinking about how Generative AI is reordering the world. Thinking it would be easier to orient around things that will not change, I read a lot of books in 2025 looking for enduring principles.</p><p>This is a list of the things that stuck with me most in 2025 and that I predict will remain relevant long into the future.</p><h2><a href="https://www.goodreads.com/book/show/323068.Information_Rules?from_search=true&amp;from_srp=true&amp;qid=Gv9HtFT6eM&amp;rank=1">1. Information Rules - Carl Shapiro &amp; Hal R. Varian (1997)</a></h2><p>I first encountered the opening quote in Information Rules by Shapiro and Varian - and it is not the only prescient insight in there.</p><p>Like today&#8217;s AI boom, the late-90s internet felt completely novel - so novel that perhaps old economic theories should be abandoned in favour of something new.</p><p>However, these economists proclaimed that economic laws were not dead, the relevant ones were just not taught in your Econ class. While information technology markets differ from the agricultural markets Econ class models are built on, they are still markets.</p><p>The authors did not write a mere textbook arguing that rumours of economic law&#8217;s death were greatly exaggerated, they wrote a strategy guide for people making decisions in this new environment grounded in the history of similar industries that came before.</p><p>They cover everything from procurement negotiation strategy to adopting and setting open vs. closed standards, and so much more.</p><p>If you are not convinced these ideas are durable, Google seemed to be. In the original 1997 copy Eric Schmidt (then CEO of Novell Inc.) wrote a blurb praising the book:</p><p><em>&#8220;Information Rules is the first book to explain network economics, the new economics of our lives. Shapiro and Varian explain all the crazy things that we see happening every day in Silicon Valley and other parts of the world. This book is a must-read for every business person in the new millennium.&#8221;</em></p><p>4 years later in 2001 Eric Schmidt became CEO of Google, and the next year in 2002 he hired Hal Varian to be Google&#8217;s Chief Economist until his retirement in August 2025.</p><p>Reading this book alongside Google&#8217;s public history (my understanding of which came from the <a href="https://www.acquired.fm/episodes/alphabet-inc">Acquired podcast&#8217;s 2-part series</a>), it is interesting to see how their approach to Android as an open platform mirrors the book&#8217;s framework for competing through ecosystems and network effects.</p><p>If you want some insight into strategic principles that shaped the outlook of one of the world&#8217;s most valuable companies, read this book.</p><h2><a href="https://www.goodreads.com/book/show/10631.Sam_Walton_Made_in_America?from_search=true&amp;from_srp=true&amp;qid=ti5ibDfyWy&amp;rank=1">2. Made in America - Sam Walton (1992)</a></h2><p>A 33 year old autobiography of a discount retailer&#8217;s founder might seem an odd recommendation for understanding modern business. However, let&#8217;s consider a quote from David Glass (Sam&#8217;s successor) quoted in the book:</p><p><em>&#8220;In retail, you are either operations driven - where your main thrust is toward reducing expenses and improving efficiency - or you are merchandise driven. The ones that are truly merchandise driven can always work on improving operations. But the ones that are operations driven tend to level off and begin to deteriorate.&#8221;</em></p><p>This man helped build Walmart, a company where &#8220;Everyday Low Prices&#8221; is a religion. If he says sales (merchandising) must be ahead of lowering costs (operations), that is worth heeding.</p><p>The generality of this concept really hit later when I was watching <a href="https://youtu.be/c_1FhdXNUSE?t=2373">a public talk by one of Anthropic&#8217;s performance engineering leads</a> from re:Invent 2025. I initially assumed Anthropic saw performance engineering as a way of lowering costs (since increasing inference speed lowers resource cost per request).</p><p>But they didn&#8217;t talk about that at all - the benefits of increasing inference speed were either tied to enabling more features (generating more demand for sales) or increasing batch size (sell more tokens per unit of hardware).</p><p>Anthropic takes a merchandise-driven approach to infrastructure, not an operations-driven one.</p><p>As I look harder I see this appear across industries outside AI, and it is not the only insight of enduring value in this book.</p><p>Whether you are looking to build your own successful business or change a culture within an existing organization, there is a lot to learn from here.</p><h2><a href="https://www.goodreads.com/book/show/113934.The_Goal?from_search=true&amp;from_srp=true&amp;qid=COyKmXzk0N&amp;rank=2">3. The Goal - Eliyahu M. Goldratt (1984)</a></h2><p>Some books excel by hammering one concept so deep into your head so that you cannot help but see it everywhere. In this book, that concept is throughput.</p><p>While a novel about optimizing factory production (yes you read that correctly) may not sound riveting or relevant 41 years later, it remains a fantastic read widely applicable today.</p><p>The whole story is a device to reframe readers&#8217; intuitions about improving &#8220;efficiency&#8221; not by cutting costs and pushing people harder, but by coordinating the whole system to lift end-to-end output with bottlenecks being both central and ever-evolving.</p><p>This made Jensen Huang calling data centers &#8216;AI factories&#8217; click for me.</p><p>I understand now that all the tricks AI engineers work on (quantization, parallelism, sparsity, etc.) are about increasing throughput (converting every dollar of input into more tokens sold) and bottlenecks are constantly shifting from chips to power to RAM and more. </p><p>They truly are like factories in the book, and understanding AI infrastructure through Goldratt&#8217;s lens informs how companies are approaching development.</p><p>You need not be an engineer to understand or find value from this story - the mindset shift is helpful across domains (whether thinking about how you use your own time to optimize your personal output, or just making a task go faster at work).</p><p>Also, I have read far worse novels. It is a mere 360 pages. Go read it.</p><h2><a href="https://www.goodreads.com/book/show/223736214-breakneck?ref=410">4. Breakneck - Dan Wang (2025)</a></h2><p>Probably the most popular book on this list (it became a podcaster book of choice in summer), I write this to proclaim that the hype was warranted.</p><p>Given that learning Mandarin is hard (5.5 years in I can confirm it is still difficult), and the lack of Western journalists living in China today, developing an independent and accurate model of what the world&#8217;s 2nd largest economy is like is remarkably difficult.</p><p>This book delivers that model for you in a concise, gripping package. It covers how China went from agrarian nation to world&#8217;s 2nd largest economy at breakneck speed, and where that trajectory is taking them.</p><p>Rather than a bland day-to-day description, Wang explains China in American terms - how they are similar, how they differ, and where he thinks each ought learn from the other.</p><p>All of this is delivered by a true expert in both, having been born in China and split his time between growing up in Canada, and working in China and the US.</p><p>This is the shortest book on the list at a mere 260 pages. Given China&#8217;s current and likely enduring relevance to the world, this is worth your time.</p><h2>Concluding Thoughts:</h2><p>Everything I listed is something that helped me better understand what will and won&#8217;t change in our AI driven reordering. I hope they can do the same for you.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.seekingbrevity.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to see more thinking about things that don&#8217;t change</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[AI and the Cost of Errors]]></title><description><![CDATA[When should we actually use AI?]]></description><link>https://www.seekingbrevity.com/p/ai-and-the-cost-of-errors</link><guid isPermaLink="false">https://www.seekingbrevity.com/p/ai-and-the-cost-of-errors</guid><dc:creator><![CDATA[Seeking Brevity]]></dc:creator><pubDate>Sun, 17 Mar 2024 01:39:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8PD6!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeeea5f9-2238-407d-87e7-bf09ba445f45_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>I&#8217;ve written this blog with the intention of sharing my knowledge. The contents of this blog are my opinions only.</em></p><p><strong>Intro:</strong></p><p>For many, 2023 was the year of AI discovery. Discovery of AI&#8217;s business potential, discovery of what it takes to build an ML solution and release it into production, and if you weren&#8217;t careful discovery that <a href="https://www.theguardian.com/world/2023/aug/10/pak-n-save-savey-meal-bot-ai-app-malfunction-recipes">AI now has the power to generate recipes for chlorine gas and create a PR fiasco for your business</a>.</p><p>However, with the dust settling from 2023 many companies have realized that new AI models are not silver bullets for old challenges around designing and productionizing AI systems.</p><p>While most people were conscious of cost, model latency and hardware constraints in 2023, I saw few people articulate a more fundamental issue with AI systems that stopped Proofs of Concept dead in their tracks - how do you handle AI systems making mistakes?</p><p>In this article, I attempt to answer this question by highlighting why errors are inherent in AI models, what design lessons we can learn from existing machine learning systems, and why anyone designing or building an AI solution needs to ask themselves these 3 questions:</p><ol><li><p>How easy is it for an AI system user to spot errors when the system makes them?</p></li><li><p>How easy is it for an AI system user to correct those errors when they are made?</p></li><li><p>What is the cost of errors if they are not spotted or corrected?</p></li></ol><p>These are not the only considerations for whether an AI model can be made (e.g. data availability and quality are also important factors). However, if you are designing an AI system, or you are assessing whether to invest in a particular AI project for your business or not, this piece is for you.</p><p><strong>Prelude - Why Focus on Errors?:</strong></p><p>Before we dive into these questions, I want to elaborate on why errors are so important to focus on when thinking about AI and Machine Learning (ML) systems.</p><p>Fundamentally, AI and ML systems are prediction (AKA guessing) machines. They are trained on a set of data to see patterns (whether it is in a set of numbers or a large corpus of text), and are tasked with generating predictions (known officially as inference) based on new information it is provided.</p><p>However, this approach is inherently error prone. Whether because of incomplete data or some other reason, your model is still only a facsimile of reality rather than a perfect encapsulation of the real thing.</p><p>Therefore, the way you approach designing these probabilistic ML systems needs to be fundamentally different from how you approach traditional deterministic software systems (where 1+1 always equals 2).</p><p>Thankfully however, we do not need to design how we approach these problems from scratch.</p><p><strong>Where have we already seen AI be successful?:</strong></p><p>While Generative AI systems like ChatGPT are fairly new (the foundational paper that led to it was <a href="https://arxiv.org/pdf/1706.03762.pdf">published in 2017</a>), the use of non-deterministic ML systems dates back to <a href="https://psycnet.apa.org/record/1959-09865-001">the dawn of ML in the 1950s</a>. Over 70 years of history since then has given the technology ecosystem time to identify products that naturally survive and struggle when faced with real customers and problems.</p><p>If we take AI systems widely adopted before the launch of ChatGPT as our set to learn from, in my analysis 4 main categories emerge.</p><ol><li><p><em>Information Matching:</em> A boring title for what has arguably been AI&#8217;s most pervasive use case, this is where AI is used to match users with products or information. This includes systems like Google Search, Google Ad targeting, Product recommendations on <a href="http://Amazon.com">Amazon.com</a> or video recommendations on Netflix and YouTube.</p></li><li><p><em>Forecasting:</em> The somewhat magical task of peering into the future unsurprisingly benefits from a magical seeming black-box type technology. This includes everything from consumer-facing tools like <a href="https://deepmind.google/discover/blog/graphcast-ai-model-for-faster-and-more-accurate-global-weather-forecasting/">weather forecasts</a> to internal business tools like customer churn modelling and predictive maintenance.</p></li><li><p><em>Classification:</em> Putting the right label on the right thing can be the difference between everything from making and losing money to identifying and not identifying fraud. Classification is used in everything from identifying fraudulent financial transactions to facial recognition software and many other things.</p></li><li><p><em>Translation:</em> Finally, we see an often overlooked application of AI (that in many ways led to the rise of Generative AI). While most consumers think of tools like Google Translate, other tools like transcription (translating sound into text) and OCR (Optical Character Recognition - translating images into text) also appear here.</p></li></ol><p><strong>Why Have These Been Successful?:</strong></p><p>Astute readers may have noticed a central thread linking all the categories listed above - all of these tools are expected to make mistakes frequently.</p><p>How many times a day do people see ads for a product they didn&#8217;t care about? How many times do people look at a weather forecast saying it was not going to rain and end up getting soaked? AI systems helps prevent some system abuses but it is far from a minority report system and Google Translate has yet to make human translators redundant.</p><p>However, the important differentiator of these systems is that the &#8216;cost&#8217; (whether that cost is in the form of time, money or reputation) of these errors is low compared to alternatives.</p><p>While not all online Ads convert to sales, ML driven targeting systems from Facebook and Google are some of the most effective marketing platforms of all time. While weather forecasts may not always save you from getting rained on, the typical pedestrian is not killed by an incorrect forecast. ML does not stop every financial crime but it helps identify many, and Google Translate saves you a lot of time compared to learning a new language yourself.</p><p><strong>Identifying the Cost of Errors:</strong></p><p>If this is the case, how can AI system designers take these insights into their own day to day projects?</p><p>Based on what we have discussed, I believe every AI system designer should ask themselves these 3 questions about system errors when building their tool:</p><ol><li><p>How easily can a system user identify these errors?</p></li><li><p>If an error is spotted, how easily can a system user correct it?</p></li><li><p>If an error is not spotted or fixed, what is the cost of that error?</p></li></ol><p>Furthermore, every person using these questions should delve into the particulars of their AI system designs, as the conclusions you draw from the answers to each of these questions will vary wildly depending on the contours of the user experience.</p><p><strong>An Illustrative Example:</strong></p><p>Lets put this framework into practice by looking at a popular recent use case - using AI to generate illustrations.</p><p>Suppose for example you are building a generative illustration product like Midjourney or Stable Diffusion. If we stopped the conversation there and naively leveraged our framework, using AI for this use case seems great. Users can easily spot errors (because people can just look at the output) and if it is not good, we can just generate another image.</p><p>However, lets dive into the contours of the user experience before we pass final judgement.</p><p>Imagine that one user of your Midjourney style tool is an experienced digital illustrator. This user has tools like Adobe Photoshop already on their PC, they are accessing your product from their PC, and they are using the images generated for a personal passion project. Our previous assessment holds up in this case - the user has enough experience to easily spot errors, they have tools ready and available to fix errors when they do appear, and if a mistake does slip through the cost to the illustrator is likely very low.</p><p>But lets imagine another user of the same tool with no experience in illustration or design, has no access to illustration tools, and needs to generate marketing collateral for an important upcoming event where mistakes will be judged harshly. In this case the calculus for the AI system designer needs to be completely different - the user is far less likely to spot errors if they appear, if the error is spotted they don&#8217;t have the tools to fix the error themselves if they do appear, and the cost of a mistake slipping through could be career damaging.</p><p>And the issue does not need to be limited to inexperienced users - what about an experienced user who needs to generate images from their phone for whatever reason and does not have access to tools to correct errors in that medium? What if their use case was not for personal passion but for a critical piece of paid work?</p><p>While this is a relatively straightforward breakdown, hopefully this illustrates how the outcomes for using an AI tool can vary wildly depending on the nature of the human in the loop, the contours of the user journey and the use case being addressed.</p><p><strong>Going Deep On The Cost Of Errors:</strong></p><p>The final element of this framework I want to unpack is what I mean by the &#8216;cost&#8217; of AI system errors.</p><p>When I speak to AI system designers about what they think of when they think about costs of errors, three categories emerge:</p><ol><li><p>Physical costs - harming or injuring an end user as a result of system outputs (e.g. giving bad medical advice)</p></li><li><p>Liability costs - risk of potential litigation for an AI system designer or AI system user as a result of AI system outputs (e.g. providing unlicensed financial advice)</p></li><li><p>Reputational costs - harming the ability of users to trust your products, services or organization due to providing faulty outputs, wasting users&#8217; time, or otherwise faulty user experiences</p></li></ol><p>Most can identify these issues when prompted, however reputational costs are consistently underappreciated. In particular, AI system designers often see so called &#8216;headline risk&#8217; coming (especially after stories like <a href="https://www.theguardian.com/world/2023/aug/10/pak-n-save-savey-meal-bot-ai-app-malfunction-recipes">the aforementioned Chlorine Gas recipe</a> appeared), but they do not see the brand damage that poor AI user experiences can produce.</p><p>For example, say you have a market research software product with thousands of users and you want to launch a Generative AI powered chatbot to help answer questions for customers based on your FAQ documents. Internal testing shows that your chatbot answers customer questions accurately 9 out of 10 times, and in 1 out of 10 questions. You expect 10% of your users to ask 10 questions a day after the initial launch.</p><p>While our intuition may tell us this proposition sounds good to launch based on our framework, again we need to dive into the contours of the user experience before we make a firm judgement.</p><p>Firstly, lets break down the cost of errors in this situation. When a user generates an answer that is incorrect, two things can happen:</p><ol><li><p>The user could spot the error, in which case they move on to try another solution to solve their problem (e.g. reading the documentation themselves)</p></li><li><p>The user could not spot the error, in which case they may go ahead assuming the results are true until they run into a problem caused by this incorrect answer and start over to solve their problem</p></li></ol><p>In either of these cases, users will feel that their time. As the venerable late Charlie Munger used to say &#8220;a brand is a promise&#8221;, and if you release a feature that does not deliver on its promise your brand will be tarnished one user at a time. In practical terms, this product experience will lead each user to experience it to leave with a worse impression of your product than when they started (which can make people less likely to buy and tell others to buy your product), and it will lead people to not trust the feature (meaning that even if you improve it over time, people are less likely to use it again).</p><p>These threats may sound overblown in the abstract, however discussing them in a way that puts real meat on those bones would require discussing details of use cases in the field that I am not at liberty to disclose. For now, understand that users&#8217; reaction to wasting time or having to start a task over due to a false start is often unstated in user feedback but is massively impactful.</p><p>Secondly, we must understand the scale of this issue - how frequently are our hypothetical users going to encounter the issue described above?</p><p>To do some napkin math, say we have 1,000 users. In this case, that would mean we expect 100 users to start using this new feature by asking ~10 questions per day. In this case, assuming the <a href="https://www.notion.so/The-AI-Cost-of-Risk-DRAFT-0a10d9a2dbfe4249bc1ded1a49667fdf?pvs=21">law of large numbers</a> applies to our 1 our of 10 error rate and errors are uniformly distributed across users, on average every one of your 100 users will experience at least 1 incorrect response per day. At best, each of those 100 users will have a worse impression of this specific feature than when they first planned to use it and just use the feature less. At worst, you have moved 100 potential promoters of your business to detractors, which is an enormous problem.</p><p><strong>Conclusions:</strong></p><p>Understanding the cost of errors is not the only issue people need to consider when designing and building AI powered systems. Prevalence of training data, in house talent, hardware supply constraints, cost and a myriad of other factors need to be in the mind of anyone deciding to build or invest in AI solutions.</p><p>However, figuring out where AI or ML techniques make sense for a given problem or system is a key part of the process that investors and builders alike need to go through, and many are not doing so today. I hope this framework is helpful for anyone currently in this position.</p><p>If you have any thoughts regarding this framework or the examples listed, please reach out to me at <a href="mailto:seeking.brevity@gmail.com">seeking.brevity@gmail.com</a>, I would love to use this as a chance to refine this thesis going forward.</p>]]></content:encoded></item><item><title><![CDATA[AI is an Educational Accelerant, Not Replacement]]></title><description><![CDATA[I&#8217;ve written this blog with the intention of sharing my knowledge.]]></description><link>https://www.seekingbrevity.com/p/ai-is-an-educational-accelerant-not</link><guid isPermaLink="false">https://www.seekingbrevity.com/p/ai-is-an-educational-accelerant-not</guid><dc:creator><![CDATA[Seeking Brevity]]></dc:creator><pubDate>Sun, 22 Oct 2023 06:48:06 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!PojW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>I&#8217;ve written this blog with the intention of sharing my knowledge. The contents of this blog are my opinions only.</em></p><h2>Introduction</h2><p>Students and teachers I speak to in secondary and tertiary education do not know how to think about Generative AI and ChatGPT.</p><p>Some think banning it outright is best because of the potential for mass-scale cheating. Some think students using ChatGPT without restrictions is the best way to prepare for an AI-powered future. Underlying all of this is fear that the education system does not have the tools to educate people for the future.</p><p>I personally disagree with all of these framings and think that blindly following any of them could lead to disastrous learning outcomes.</p><p>In this article, I aim to explain why these commonly held views will likely be harmful, and propose what I think is a more helpful approach to Generative AI for both students and teachers. </p><p>In particular, I propose a perspective based on 3 related principles:</p><ol><li><p>Generative AI is best understood as an educational accelerant, not a replacement for developing traditional educational fundamentals</p></li><li><p>If left unchecked, ChatGPT and similar tools would polarize educational results (lowering overall educational outcomes while allowing a small, motivated top end to excel)</p></li><li><p>The top priority for the education system should be to develop skills in areas that LLMs are incapable of addressing:</p><ul><li><p>Researching and verifying information in the pursuit of truth</p></li><li><p>How to determine one&#8217;s own objectives, including a sense of right and wrong</p></li><li><p>Mathematics</p></li></ul></li></ol><h2>A Primer on ChatGPT and Large Language Models (LLMs)</h2><p>Before we begin, it is worth having an understanding of how ChatGPT and similar LLMs work.</p><p>For those not already familiar with how ChatGPT works and what its major limitations are, I recommend <a href="https://seekingbrevity.substack.com/p/the-big-3-gaps-in-generative-ai">one of my previous articles breaking down the topic</a>.</p><p>For those without time to read it or who just want a refresher, LLMs are not infallible machine overlords. LLMs are statistical models with a strong intuitive grasp of human language, and they have 3 notable weaknesses that are inherent to the way they are built as of this writing:</p><ol><li><p>They are not authoritative databases with a tight coupling to reality</p></li><li><p>They are incapable of defining their own objectives (and similarly are inappropriate tools for distinguishing right from wrong)</p></li><li><p>They are really, really bad at math</p></li></ol><h2>Why Would Banning ChatGPT Be Harmful?</h2><p>I will start by challenging the first common approach of banning ChatGPT - why should ChatGPT not be banned whole cloth in all educational settings?</p><p>Imagine a common sight - a struggling high-school student writing an essay on English Literature. It is not going well. Perhaps they missed some earlier school and fell behind in grammar. Maybe English is not the primary language spoken at home. Or maybe they just naturally struggle with writing essays.</p><p>At the moment, if this student wanted help to improve their writing, they would have 4 options:</p><ol><li><p>Approach their already overworked teacher (who may not have time to coach 1:1)</p></li><li><p>Approach your parents (who will likely not have the knowledge or expertise to help)</p></li><li><p>Hire a writing tutor for the subject (an expensive and ineffective undertaking for many)</p></li><li><p>Continue stumbling along, likely performing poorly</p></li></ol><p>Whether it is because of financial constraints, time constraints, a lack of personal connection with the teacher/tutor, social awkwardness around approaching teachers for help, or just because of the sheer friction of the whole process, most of these students end up in the 4th bucket. However, the 4th bucket is where any person interested in public education wants nobody to end up.</p><p>ChatGPT provides a solution to this problem. If used properly, ChatGPT could enable this struggling student to improve their grasp of English by identifying grammatical errors, providing corrections and providing clear reasons. ChatGPT could give that student feedback about the logical flow of the document and how they could improve. All of this could be done in private, on demand, and at low-to-no cost to the student.</p><p>Furthermore, the benefits of ChatGPT are not just limited to struggling students. Imagine now a curious student who wants to learn about topics not covered in the syllabus or at their school. Now curious, driven students have assistants that can help explain and summarise any topic they want to learn about, whether it is Nuclear Fusion or CRISPR/Cas9.</p><p>The potential impact of a 1:1 tutor available to everyone cannot be understated. In 1984, educational psychologist Benjamin Bloom found that students with access to a private tutor perform 2 standard deviations better than students taught in classrooms (see <a href="https://en.wikipedia.org/wiki/Bloom%27s_2_sigma_problem">Bloom&#8217;s 2 Sigma Problem</a>).</p><p>Therefore, completely banning ChatGPT use would be harmful because it would deprive students of a much needed educational tool that can meet a need currently unmet by education systems everywhere.</p><p>However, there is a reason I used the qualification &#8220;if used properly.&#8221;</p><h2>Why Would Unrestricted Use of ChatGPT Be Harmful?</h2><p>Now to challenge the second commonly held view of allowing ChatGPT to be used unconstrained - why should ChatGPT not be allowed without restrictions or guidance?</p><p>Lets take another view of the struggling student from before. Instead of using ChatGPT to coach them in grammar and slowly improve their abilities, they could just use it to write the essay for them and just hand it in unedited.</p><p>If this were allowed, students would often hand in nicer sounding essays. It would also defeat the purpose of education.</p><p>Writing is the process by which we realise that we don&#8217;t know what we are talking about. Only by writing do you give your thoughts form and allow them to be critiqued and improved. It makes the invisible visible, and it is vital for the development of independent thought.</p><p>Struggling students using ChatGPT to skip this process robs them of this development. Not only does our struggling student leave their weaknesses unaddressed, it allows whatever creative muscles they do have to atrophy.</p><p>This is not the end of the harm however. Using ChatGPT blindly without understanding its limitations can lead people to depend on it where it is not appropriate.</p><p>None of this is mere speculation - both of these patterns were observed in a <a href="https://www.bcg.com/publications/2023/how-people-create-and-destroy-value-with-gen-ai">recent study performed by the BCG Henderson Institute</a> on the effects of using ChatGPT in the workplace.</p><p>Researchers found that for tasks that ChatGPT is currently good at (e.g. creative ideation), users who leveraged ChatGPT reported feeling their creative abilities atrophy while using it. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PojW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PojW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png 424w, https://substackcdn.com/image/fetch/$s_!PojW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png 848w, https://substackcdn.com/image/fetch/$s_!PojW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png 1272w, https://substackcdn.com/image/fetch/$s_!PojW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PojW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png" width="1440" height="811" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b931fb17-1886-4db8-b523-9088754077d7_1440x811.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:811,&quot;width&quot;:1440,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:252269,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PojW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png 424w, https://substackcdn.com/image/fetch/$s_!PojW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png 848w, https://substackcdn.com/image/fetch/$s_!PojW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png 1272w, https://substackcdn.com/image/fetch/$s_!PojW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb931fb17-1886-4db8-b523-9088754077d7_1440x811.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Furthermore, while the output of these users seemed better than those who did not use ChatGPT, the diversity of ideas among ChatGPT users was drastically lower than non-ChatGPT users.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8aDK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8aDK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png 424w, https://substackcdn.com/image/fetch/$s_!8aDK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png 848w, https://substackcdn.com/image/fetch/$s_!8aDK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png 1272w, https://substackcdn.com/image/fetch/$s_!8aDK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8aDK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png" width="1456" height="851" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:851,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:215036,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8aDK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png 424w, https://substackcdn.com/image/fetch/$s_!8aDK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png 848w, https://substackcdn.com/image/fetch/$s_!8aDK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png 1272w, https://substackcdn.com/image/fetch/$s_!8aDK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa693b1a6-2c7c-4d71-a493-9d659ff42104_2480x1450.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For tasks that ChatGPT is not good at (e.g. business problem solving), users with ChatGPT access often over relied on ChatGPT and performed worse than people without it, even when they were provided education on the limitations of ChatGPT.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3-B4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3-B4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png 424w, https://substackcdn.com/image/fetch/$s_!3-B4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png 848w, https://substackcdn.com/image/fetch/$s_!3-B4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png 1272w, https://substackcdn.com/image/fetch/$s_!3-B4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3-B4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png" width="682" height="463" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fd84e341-e1ee-411f-84ac-6115d2085897_682x463.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:463,&quot;width&quot;:682,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:80982,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3-B4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png 424w, https://substackcdn.com/image/fetch/$s_!3-B4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png 848w, https://substackcdn.com/image/fetch/$s_!3-B4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png 1272w, https://substackcdn.com/image/fetch/$s_!3-B4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd84e341-e1ee-411f-84ac-6115d2085897_682x463.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>ChatGPT - If Left Untreated</h2><p>Prior to ChatGPT&#8217;s launch, teachers and students would all be familiar with a distribution of student performance that looks something like the following:</p><ul><li><p>A small number of curious, driven students perform well and learn almost no matter what</p></li><li><p>A small number of students struggle to learn, almost no matter what is thrown at them</p></li><li><p>Most students live in the middle - not high performing, but not lagging either</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nXvK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nXvK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png 424w, https://substackcdn.com/image/fetch/$s_!nXvK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png 848w, https://substackcdn.com/image/fetch/$s_!nXvK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png 1272w, https://substackcdn.com/image/fetch/$s_!nXvK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nXvK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png" width="993" height="526" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c602709c-350d-4b5e-bec0-aa6050284b91_993x526.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:526,&quot;width&quot;:993,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:40865,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nXvK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png 424w, https://substackcdn.com/image/fetch/$s_!nXvK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png 848w, https://substackcdn.com/image/fetch/$s_!nXvK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png 1272w, https://substackcdn.com/image/fetch/$s_!nXvK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc602709c-350d-4b5e-bec0-aa6050284b91_993x526.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If ChatGPT were allowed to take its course with no systematic intervention, I believe outcomes would start to look more like this:</p><ul><li><p>A small number of smart, highly motivated students would leverage ChatGPT as an educational accelerant and personal coach</p></li><li><p>The average and underperforming students would atrophy academically, as ChatGPT is leveraged as an educational replacement</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CkEu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CkEu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png 424w, https://substackcdn.com/image/fetch/$s_!CkEu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png 848w, https://substackcdn.com/image/fetch/$s_!CkEu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png 1272w, https://substackcdn.com/image/fetch/$s_!CkEu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CkEu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png" width="994" height="521" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:521,&quot;width&quot;:994,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:49182,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CkEu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png 424w, https://substackcdn.com/image/fetch/$s_!CkEu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png 848w, https://substackcdn.com/image/fetch/$s_!CkEu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png 1272w, https://substackcdn.com/image/fetch/$s_!CkEu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0951f4be-9a8e-4200-8420-b42b28b4fc47_994x521.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This kind of polarized outcome is undesirable, as while the top end could perform better, the net amount of learning and educational development would regress across student cohorts.</p><h4>We Have Been Through This Before - Google Search</h4><p>While ChatGPT looks and feels novel to most of us, I believe it is not the first time many of us have seen a new technology have an educational impact like this.</p><p>The proliferation of the internet (particularly Google Search) massively disrupted how students approached studying, particularly when it comes to reading and research.</p><p>Before Google, if you were a student and you wanted to do research, you needed to go to a library. Not only that, you need to go to the library and ask to find a book. If the book was rented out you might need to wait a few weeks for someone to return that book. Once you had gathered your books, you needed to read a summary of the book to figure out if it contains the information you are looking for.</p><p>In short, it was a labour intense process that yielded a similar pattern to what we saw before:</p><ul><li><p>A small number of dedicated students learned a lot by diving through libraries and developed a personal interest in reading</p></li><li><p>Some recalcitrant or struggling students never went to the library and thus never absorbed the knowledge contained within them</p></li><li><p>Most students were in the middle, accessing the library where needed on the occasions they were compelled to or when a Cliff&#8217;s Notes was not available</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IJMY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IJMY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png 424w, https://substackcdn.com/image/fetch/$s_!IJMY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png 848w, https://substackcdn.com/image/fetch/$s_!IJMY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png 1272w, https://substackcdn.com/image/fetch/$s_!IJMY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IJMY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png" width="985" height="526" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:526,&quot;width&quot;:985,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:41160,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IJMY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png 424w, https://substackcdn.com/image/fetch/$s_!IJMY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png 848w, https://substackcdn.com/image/fetch/$s_!IJMY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png 1272w, https://substackcdn.com/image/fetch/$s_!IJMY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e5c1565-46dc-4ae7-93ce-608abbdccfb6_985x526.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>With the rise of Google Search however, this process changed. Suddenly, instead of going to a library, students could go to their computer and get access to most of the information they were looking for instantly.</p><p>For high performing students, this enabled them to learn more than generations before them. Instead of spending weeks diving through libraries and all the dead air that entailed, they could access information about whatever they wanted to learn almost instantly. Furthermore, they were not just limited to the knowledge that was contained in their local library. Students who wanted to learn about Complex Numbers or Ancient Roman History could do so at the touch of a button.</p><p>However, it also meant that students could instantly get access to sites like Wikipedia, Sparknotes, and many other resources for cheating or reducing the time they spent reading. Furthermore, it gave access to YouTube, memes, and many other distractions.</p><p>Thus, we ended up with a pattern that looks like this again, with high performers performing even better and poor and average performers learning less.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4KAc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4KAc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png 424w, https://substackcdn.com/image/fetch/$s_!4KAc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png 848w, https://substackcdn.com/image/fetch/$s_!4KAc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png 1272w, https://substackcdn.com/image/fetch/$s_!4KAc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4KAc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png" width="980" height="527" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c3172407-f954-4695-b747-b396a408b8a7_980x527.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:527,&quot;width&quot;:980,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:49581,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4KAc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png 424w, https://substackcdn.com/image/fetch/$s_!4KAc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png 848w, https://substackcdn.com/image/fetch/$s_!4KAc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png 1272w, https://substackcdn.com/image/fetch/$s_!4KAc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3172407-f954-4695-b747-b396a408b8a7_980x527.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>While it is a difficult to get a specific graph that shows aggregate drops in time spent researching and pursuing knowledge, a useful proxy for this is <a href="https://www.amacad.org/humanities-indicators/public-life/time-spent-reading#32151">the drop in time spent reading for personal interest seen between 2003-2018</a> by the American Academy of Arts and Sciences:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gZKz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gZKz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png 424w, https://substackcdn.com/image/fetch/$s_!gZKz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png 848w, https://substackcdn.com/image/fetch/$s_!gZKz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png 1272w, https://substackcdn.com/image/fetch/$s_!gZKz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gZKz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png" width="757" height="512" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:512,&quot;width&quot;:757,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:61353,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gZKz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png 424w, https://substackcdn.com/image/fetch/$s_!gZKz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png 848w, https://substackcdn.com/image/fetch/$s_!gZKz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png 1272w, https://substackcdn.com/image/fetch/$s_!gZKz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0e2eca-2457-4b22-af18-d060c84fcae7_757x512.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Where Do We Go From Here?</h2><p>Now that we have a conceptual framework for understanding ChatGPT and its likely impacts on education if left unaddressed, how do we harness its potential while minimising downside risk?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7XFp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7XFp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png 424w, https://substackcdn.com/image/fetch/$s_!7XFp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png 848w, https://substackcdn.com/image/fetch/$s_!7XFp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png 1272w, https://substackcdn.com/image/fetch/$s_!7XFp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7XFp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png" width="992" height="529" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:529,&quot;width&quot;:992,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:56205,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7XFp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png 424w, https://substackcdn.com/image/fetch/$s_!7XFp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png 848w, https://substackcdn.com/image/fetch/$s_!7XFp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png 1272w, https://substackcdn.com/image/fetch/$s_!7XFp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44d7776-c0d2-4873-8edd-6e43659b0b4c_992x529.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I believe there are two broad areas for people to consider:</p><ol><li><p>What subjects and skills students should be learning?</p></li><li><p>How should we go about teaching them as a system?</p></li></ol><h3>Subjects and Skills</h3><p>Many people I speak to emit a sense of fear over the future of education because they do not know what subjects teachers should teach or what skills students should develop in this new era. Some think we need to throw out old curriculums and invent new AI infused subjects or courses on AI ethics.</p><p>I take the opposite view on this - I think that a portfolio of traditional subjects (Math, English, History, etc.) already have the ingredients for developing the skills students need to fill the Big 3 Gaps in Generative AI (the independent pursuit and verification of the truth, the development of a sense of what is right and wrong, and Mathematics).</p><p>Furthermore, I think doubling down on traditional and proven approaches to character building are crucial to filling the Big 3 Gaps.</p><p>Regarding the portfolio of subjects, exactly which subjects should be offered by schools and selected by students vary, but the following disciplines with the following emphases are a good place to start:</p><ul><li><p>Investigative Journalism - the process of discovering, reasoning about and presenting information not yet uncovered</p></li><li><p>History - the process of interpreting and evaluating recorded evidence to come independently to a sense of truth, understanding what could or could have not happened, and extrapolating implications for the future about the story of humanity</p></li><li><p>Persuasive writing - the process by which we realise that we don&#8217;t know what we are talking about, wrestling with your own ideas, making invisible beliefs and assumptions visible, and presenting cohesive a point of view out the other end</p></li><li><p>Philosophy - the rigorous study of the fundamentals of thought, knowledge, reality and humanity</p></li><li><p>Foreign Languages - the study of other cultures in their own terms and vocabulary</p></li><li><p>Arts and Crafts - the art of learning to see and display the world through a variety of mediums and disciplines</p></li><li><p>Mathematics and mathematical sciences - the study of the language of logical reasoning and problem solving (e.g. Physics, Chemistry, Computer Science, Engineering, etc.)</p></li></ul><p>If a student selects a basket of the above subjects and develops a self-sustaining curiosity in them that goes beyond the classroom, that is the definition of success. Good litmus tests of self-sustaining curiosity include reading for personal interest (reading quality books, articles and other high quality media on topics they care about) and independent writing (writing in one&#8217;s own time to make sense of themselves and the world).</p><p>Regarding character building, it is important to acknowledge that the aforementioned <a href="https://seekingbrevity.substack.com/p/the-big-3-gaps-in-generative-ai">Big 3 Gaps in Generative AI</a> (especially points 1. and 2.) cannot be met by curriculum alone. Being intellectually honest in the pursuit of truth and having a deep sense of right and wrong require levels of emotional maturity and personal character that are often not be developed through assessment alone.</p><p>While the ideal methods for developing character and what character traits we should learn to emulate has always been a matter of political contention, few disagree that extracurricular activities (especially sport) are excellent mechanisms for schools to facilitate emotional development. Furthermore, when done well, some form of rigorous argumentation (whether through sports like debating, studies of religion or other philosophical studies) tend to facilitate the development of principled views on the world and what separates right from wrong.</p><h3>Brass Tacks - How to implement as a system</h3><p>So, I argue that the content we teach should not change much. But then how should schools to then adapt their teaching and assessment methods to take ChatGPT into account? How do we ensure it acts as an educational accelerant rather than educational replacement?</p><p>If we want ChatGPT to accelerate students, that means we need to design classes and assessments that contain the following:</p><ul><li><p>Educate students on ways they can use it as an effective personal tutor (improving phrasing, ask if there are logical inconsistencies or structural weaknesses)</p></li><li><p>Educate students on the limits of LLMs, and encourage them to not use it in areas where it will produce bad results (e.g. citing sources, business problem solving)</p></li><li><p>Encourage students to learn to do something themselves first, the slow way (e.g. writing an essay) before they use ChatGPT as a crutch</p></li></ul><p>Furthermore, we must ensure that at no point are students allowed or encouraged to use LLMs to originate content (e.g. writing reports or essays).</p><p>While there are many ways to go about this, assessment methods could include the following:</p><ol><li><p>Information Verification - provide information that is either incorrect or biased, and force students to verify the information</p><ul><li><p>E.g. Take a paragraph from an old newspaper or book, get them to find the sources and argue whether the book&#8217;s statement was correct or not</p></li></ul></li><li><p>Personal Writing - ask students to write about and reflect on personal experiences (i.e. information that an LLM would not have access to) and take a stance on it</p></li><li><p>ChatGPT Free Assessments - get students to perform assessments in formats where they do not have access to a computer or LLM (e.g. hand-written exams, impromptu assessments, verbal assessments without notes, etc.)</p></li><li><p>Ask students to synthesise information too big or long to shove into an LLM</p><ul><li><p>At this point in time, LLMs have limits preventing people from inserting blocks of text of unlimited length (i.e. a &#8216;context limit&#8217;)</p></li><li><p>Furthermore, research has shown that <a href="https://arxiv.org/pdf/2307.03172.pdf">LLM accuracy across models drops when more information is passed into a prompt</a> (e.g. if you pass in 10,000 words it is less likely to generate a correct answer than if it were only passed 1,000)</p></li><li><p>Therefore, assessments that require ingestion of larger pieces of information can be LLM-resistant</p></li></ul></li></ol><p>Finally, it is worth accepting that there will be no perfect technical approach to implementing these principles at scale:</p><ul><li><p>Blocking ChatGPT in schools will be of limited usefulness for committed cheaters (as we saw with school filters for gaming websites, if there is a will there is a way to access them)</p></li><li><p>It may be a while before we get an LLM that acts like a good teacher (i.e. one that refuses to do work for students) for similar reasons (any filter you put in front of an LLM to stop people from cheating will not stop those committed enough to try get around it)</p></li><li><p>While some groups are working on methods for watermarking LLM outputs (e.g. <a href="https://arxiv.org/pdf/2301.10226.pdf">this approach from the University of Maryland</a>), implementations have yet to yield results in the field and it is not clear how effective they will be in the field</p></li></ul><p>As a result, any implementation of these principles will depend on both school enforcement and cultural acceptance among students of how they should approach assessment for their own long term best interest.</p><h2>Conclusions</h2><p>In his <a href="https://youtu.be/O4MtQGRIIuA?t=267">public appearances</a>, Amazon CEO Jeff Bezos has famously talked about how Amazon tries to make its biggest bets on things that are likely not to change in the next 10 years, rather than how things will change.</p><p>For example, Amazon invested in building Amazon Prime because they believe consumers will always want lower prices, more selection, and faster delivery (which makes sense, I don&#8217;t expect many people to start complaining &#8220;I loved shopping on Amazon, I just wished you would deliver my packages slower and charge me more for each package&#8221;).</p><p>While it is tempting to think that because ChatGPT and other LLMs are new and evolving, we need to fundamentally change how we approach education to keep up. However, I think it is instead better to formulate the question in reverse - what has not changed in the era of LLMs, and how can we double down on those things that we already know can work.</p><p>Without a fundamental paradigm shift in how LLMs work, none of the fundamental issues I discussed here will be completely resolved. As a result, I think it is safe to bet on the traditional methods and approaches discussed earlier.</p><p>I personally do not foresee a future where students and parents go to their teachers and say &#8220;thank you for teaching my child, however could you please stop teaching them to be so good at verifying sources? They&#8217;ve got an algorithm for determining absolute truth now.&#8221;</p><p>None of the solutions I propose radically transform how schools or students approach education, and this is intentional. I recommend instead that education systems adapt to LLMs by learning how to use them to accelerate what we already know works, and to work to finally solve Bloom&#8217;s 2 Sigma Problem.</p><div><hr></div><p>Thank you for reading this piece, I hope you found it useful or enlightening in some way.</p><p>If you found this helpful, or if you think it was horrible and you need to yell at me about it, let me know at this email: <a href="mailto:seeking.brevity@gmail.com">seeking.brevity@gmail.com</a>.</p><p>There are topics that came up during writing this that I removed for brevity&#8217;s sake. If you have an interest in more pieces like this, let me know as well.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.seekingbrevity.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.seekingbrevity.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[The Big 3 Gaps in Generative AI]]></title><description><![CDATA[I&#8217;ve written this blog with the intention of sharing my knowledge.]]></description><link>https://www.seekingbrevity.com/p/the-big-3-gaps-in-generative-ai</link><guid isPermaLink="false">https://www.seekingbrevity.com/p/the-big-3-gaps-in-generative-ai</guid><dc:creator><![CDATA[Seeking Brevity]]></dc:creator><pubDate>Sun, 22 Oct 2023 04:04:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!yh3T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>I&#8217;ve written this blog with the intention of sharing my knowledge. The contents of this blog are my opinions only.</em></p><p>For many, ChatGPT feels both scary and magical. As soon as the public saw the potential of Large Language Models (LLMs) news feeds were flooded with evangelists talking about how ChatGPT could do everything from <a href="https://www.nature.com/articles/s42256-023-00644-2">your homework</a> to <a href="https://www.youtube.com/watch?v=4Q2HxVpJ9nw&amp;t=226s">making you rich</a>.</p><p>However, all the noise and hyperbole around LLMs has made it difficult for non-practitioners of the AI arts to construct a clear mental model of what LLMs are and are not capable of. </p><p>Given this, I wrote this short piece to help others understand what LLMs really are, what their biggest limitations are (which I call &#8216;The Big 3 Gaps in Generative AI&#8217;), and how to think about where they should and should not be used. </p><h2>What is an LLM?</h2><p>Text-generation LLMs, of which ChatGPT is one, are statistical models under the hood. In particular, they are models designed to guess the next word in a sequence based on a prompt and other words it has generated so far.</p><p>One can think of this as similar to how many smartphones now have a &#8216;next word guess&#8217; just above the keyboard.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yh3T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yh3T!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png 424w, https://substackcdn.com/image/fetch/$s_!yh3T!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png 848w, https://substackcdn.com/image/fetch/$s_!yh3T!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png 1272w, https://substackcdn.com/image/fetch/$s_!yh3T!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yh3T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png" width="796" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:796,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:94192,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yh3T!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png 424w, https://substackcdn.com/image/fetch/$s_!yh3T!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png 848w, https://substackcdn.com/image/fetch/$s_!yh3T!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png 1272w, https://substackcdn.com/image/fetch/$s_!yh3T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5cc2baf-9e94-425a-8a83-5b87572da1ac_796x400.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>These models are trained effectively by taking a large set of text data (such as the copies of much of the internet provided by <a href="https://commoncrawl.org/">Common Crawl</a>) and training a model using really big servers (often NVIDIA <a href="https://www.nvidia.com/en-au/data-center/a100/">A100s</a> or <a href="https://www.nvidia.com/en-au/data-center/h100/">H100s</a> - hence <a href="https://www.forbes.com/sites/dereksaul/2023/08/22/nvidia-stock-hits-all-time-high-after-315-surge-easily-outpacing-its-peers/?sh=528c9e3c4b56">the early 2023 spike in NVIDIA&#8217;s share price</a>).</p><p>In the training process, model builders try to capture the patterns in training data language using probabilities (e.g. out of all the times we see the word &#8220;George&#8221; in a sentence, what is the probability that the next word will be &#8220;Washington&#8221; vs. any other word in the English language).</p><p>While there is a lot more to how these models work, this mental model is enough for our purposes. (For those curious about the nuts and bolts, I recommend reading <a href="https://arxiv.org/pdf/1706.03762.pdf">Attention is All You Need</a> (the foundational paper that made the recent AI wave possible). If that feels confusing on its own (which it likely will be) also watch <a href="https://www.youtube.com/watch?v=kCc8FmEb1nY&amp;t=1s">Andrej Karpathy&#8217;s ChatGPT Video</a> to break it down a little more.)</p><h2>The Big 3 Gaps in Generative AI</h2><p>Armed with this mental model and experience playing with LLMs, 3 major gaps in LLM abilities become evident:</p><ol><li><p>LLMs do not generate results based on what is real or true (they generate responses based on what they have been trained to see in training data)</p></li><li><p>LLMs are not sentient (and as a result are incapable of determining their own objectives)</p></li><li><p>LLMs are really, really bad at math</p></li></ol><p>Firstly, regarding their distance from reality. These tools are fundamentally statistical models built to take in one set of words and guess what an output set should look like one word at a time. This means that they are not referring to an authoritative source of truth when they answer a question, nor are they determining the truth by comparing multiple sources and coming to an independent answer. LLMs make authoritative sounding guesses based on the information they have seen on a large subset of the internet, which may or may not be an accurate reflection of reality.</p><p>Secondly, regarding their lack of sentience. It is true in a strictly literal sense that LLMs are not sentient as they do not have agency to act without human prompting (and even if LLMs can technically write prompts for other LLMs, the fundamental intent they pursue or seed prompt must be set by a human). However, it is more important to understand this fact from a metaphysical perspective. LLMs are not equipped to make independent determinations of what is right and wrong based on personally held moral or ethical principles the way humans do.</p><p>For example, Joe Biden could ask an LLM whether he should provide nuclear weapons to Ukraine, and the LLM could provide an authoritative sounding answer to his question. </p><p>However, in a situation like this the US President needs to make an independent determination of what the objectives of the United States are, which of those objectives take priority over others, and what paths are morally acceptable for achieving them.</p><p>Humans define these objectives and make moral judgements for decisions like these (as well as plenty of other less-consequential decisions) based on the fundamental beliefs and assumptions they develop from a lifetime of sensory data. There are no universally agreed &#8216;correct answers&#8217; or assessments in these situations that LLMs can anchor to either, as those can vary wildly depending on those fundamental premises that people looking at the problem hold going into it.</p><p>Furthermore, the output of an LLM would vary wildly depending on the way the President phrased the question, rather than being grounded in any universal moral or ethical principles. As a result ChatGPT or any other LLM could not be reasonably relied upon to provide a valid perspective on morally challenging and thorny topics like this.</p><p>Finally, regarding their lack of mathematical capabilities. LLMs struggle with math because they are only designed to guess the next word or number in a sequence based on what it has seen before. It is not designed to learn the fundamental meaning of numbers and their patterns. As a result, LLMs are not just bad at arithmetic (which traditional computers are excellent at), they are currently incapable of solving unsolved problems in the pure mathematics space.</p><p>It is not impossible that at some point in the future that AI researchers will build new models that overcome these limitations. However, these particular cracks are fundamental to how present-day LLMs have been architected and trained, and are unlikely to be resolved by more training time with more training data on more GPUs. AI systems that solve these problems will require new leaps forward in methodology, and it is anybody&#8217;s guess as to when and how that happens. </p><h2>LLMs Are Digital Intuition</h2><p>The great contribution of LLMs is that they have given computers an intuitive grasp of natural human language.</p><p>Learning the magic tricks behind ChatGPT may make them seem less impressive than they appeared on one&#8217;s first interaction. However, these models learn language similar to how we humans do when we are young (we first learn grammar by observing which words are used where, long before we learn the language of nouns, verbs and adjectives), which is no small feat.</p><p>That being said, while LLMs have given computers a form of digital intuition, intuition alone is not enough to solve many problems. Understanding and respecting these limits is crucial for allowing these tools to help us, rather than hinder us.</p><div><hr></div><h3>Thank You for Reading!</h3><p>I hope you found this piece enlightening or helpful in some way.</p><p>If you would like to share your thoughts with me about this piece or request different pieces in future, let me know via this email: seeking.brevity@gmail.com</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.seekingbrevity.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Seeking Brevity&#8217;s Substack! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Coming soon]]></title><description><![CDATA[This is Seeking&#8217;s Substack.]]></description><link>https://www.seekingbrevity.com/p/coming-soon</link><guid isPermaLink="false">https://www.seekingbrevity.com/p/coming-soon</guid><dc:creator><![CDATA[Seeking Brevity]]></dc:creator><pubDate>Fri, 20 Oct 2023 05:59:57 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8PD6!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeeea5f9-2238-407d-87e7-bf09ba445f45_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is Seeking&#8217;s Substack.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.seekingbrevity.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.seekingbrevity.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item></channel></rss>