<?xml version="1.0" encoding="UTF-8"?><rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>computation &amp;mdash; Language &amp; Literacy</title>
    <link>https://languageandliteracy.blog/tag:computation</link>
    <description>Musings about language and literacy and learning</description>
    <pubDate>Fri, 17 Apr 2026 02:25:33 +0000</pubDate>
    <image>
      <url>https://i.snap.as/LIFR67Bi.png</url>
      <title>computation &amp;mdash; Language &amp; Literacy</title>
      <link>https://languageandliteracy.blog/tag:computation</link>
    </image>
    <item>
      <title>Reviewing Claims I’ve Made on LLMs</title>
      <link>https://languageandliteracy.blog/reviewing-claims-ive-made-on-llms?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[Novice bunny and expert bunny on bikes&#xA;When I typically begin a series of blogs to conduct nerdy inquiry into an abstract topic, I don&#39;t generally know where I&#39;m going to end up. This series on LLMs was unusual in that in our first post, I outlined pretty much the exact topics I would go on to cover.&#xA;&#xA;Here&#39;s where I had spitballed we might go:&#xA;&#xA;The surprisingly inseparable interconnection between form and meaning&#xA;Blundering our way to computational precision through human communication; Or, the generative tension between regularity and randomness&#xA;The human (and now, machine) capacity for learning and using language may simply be a matter of scale&#xA;Is language as separable from thought (and, for that matter, from the world) as Cormac McCarthy said?&#xA;Implicit vs. explicit learning of language and literacy&#xA;&#xA;Indeed, we then went on to explore each of these areas, in that order. Cool!&#xA;!--more--&#xA;&#xA;Some Hypotheses from This Series&#xA;&#xA;What theories have we raised through this exploration?&#xA;&#xA;1) LLMs gain their uncanny powers from the statistical nature of language itself; &#xA;2) the meaning and experiences of our world are more deeply entwined with the form and structure of our language than we previously imagined; &#xA;3) LLMs may offer us an opportunity to further the convergence between human and machine language; &#xA;4) AI can potentially extend our cognitive abilities, enabling us to process and understand far more information;&#xA;5) Both human and machine learning progresses from fuzzy, imprecise representations to higher precision, and the greater the precision, the greater the effort and practice (or “compute”) that is required; and &#xA;6) LLMs challenge Chomsykan notions of innateness and suggest that implicit, statistical learning alone can lead to gaining the grammatical structure and meaning of a language.&#xA;&#xA;While I’ve been mostly positive and excited about the potential of AI (aside from pointing out how it is accelerating the looming ecological catastrophe that seems to be our trajectory) I should probably pause here to acknowledge that there may be important counterpoints to many of these (perhaps somewhat starry-eyed) hypotheses. &#xA;&#xA;Onto the Counterclaims&#xA;&#xA;Let&#39;s take a more critical look at some of my claims:&#xA;&#xA;1) I claim that language is fundamental to the generative powers of LLMs. &#xA;&#xA;Yet Andrej Karpathy, who is no stranger to LLM development, tweeted: &#xA;&#xA;  It&#39;s a bit sad and confusing that LLMs (&#34;Large Language Models&#34;) have little to do with language; It&#39;s just historical. They are highly general purpose technology for statistical modeling of token streams. A better name would be Autoregressive Transformers or something.&#xA;&#xA;  They don&#39;t care if the tokens happen to represent little text chunks. It could just as well be little image patches, audio chunks, action choices, molecules, or whatever. If you can reduce your problem to that of modeling token streams (for any arbitrary vocabulary of some set of discrete tokens), you can &#34;throw an LLM at it.&#34;&#xA;&#xA;I agree that LLMs are performing “statistical modeling of token streams,” and that “for any arbitrary vocabulary of some set of discrete tokens, you can ‘throw an LLM at it.’” &#xA;&#xA;We now have multimodal LLMs that are modeling out of token streams of audio, visual, and text, and will no doubt have ones feeding from additional streams of sensory data as they are increasingly paired with cameras on humans, objects, and robots.&#xA;&#xA;Yet I also think Karpathy undersells that when LLMs suddenly exploded into general public awareness and fascination, it was merely a “historical” fact that they were trained upon vast amounts of human generated text and were able to reproduce and generate human language. As we’ve explored in this and a previous series, there is something about human language itself uniquely adapted for our brain circuitry and the propagation of our culture within social interaction in our world. And being able to communicate with a powerful computational model through the medium of conversational human language has been a revolutionary advent. We are just in the beginning stages of grokking it.&#xA;&#xA;As I tweeted in response to Karpathy, token streams may be applied to anything, but human language seems to be uniquely suited to the advancement of combined human and machine learning. Not only because we rely on it for communication – but furthermore due to the algebraic and statistical nature of our language.&#xA;&#xA;Recent case in point: the viral attention currently on NotebookLM’s Audio Overview. Listening to a conversation, however artificial, resonates with us, because that&#39;s what&#39;s in our social nature. And, surprisingly, it does a fairly good job of surfacing information from across multiple multimodal sources (and soon, across languages) that we find interesting, relevant, and meaningful.&#xA;&#xA;Speaking of NotebookLM Audio Overview. . . here’s one derived from all the blog posts (except this one) from this series, as well as the sources–outlined in post 1–that inspired them all: https://notebooklm.google.com/notebook/a4f35399-e288-4293-b2d2-0489e6b1f037/audio&#xA;&#xA;4) I claim there is great potential for AI to extend our cognitive capabilities&#xA;&#xA;Yet there is a strong case of an equal and commensurate danger that use of LLMs can reduce our cognitive capabilities.&#xA;&#xA;Learning more formal content and skills, like what we learn in school or in a job, requires deliberate effort until we develop an unconscious fluency. If students learning new concepts and skills externally automate their practice of new learning (such as writing or math) to an LLM, then they will not–ironically–be able to develop the automatized internal knowledge and capacity they need to wield powerful tools like AI more effectively.&#xA;&#xA;When “experts” use tools like AI, they know where the gaps are in the output and are able to use it strategically to enhance their own production and output. A few examples of this:&#xA;&#xA;Simon Willison, a programmer who is also a great communicator, uses different LLMs to support his projects, and writes and speaks about how he does so. Here’s a podcast, for example, where he explains how he uses them.&#xA;&#xA;Nicholas Carlini, a research scientist at Google DeepMind, similarly wrote about how he uses AI to support his work.  &#xA;&#xA;Cal Newport, who writes extensively about how to do “deep work” in a world of distractions, recently wrote in The New Yorker how he has found ChatGPT useful to his writing.&#xA;&#xA;All the people above are highly skilled at what they do – so when they explore and then figure out how to use AI to support their work, they do so in a way that does not diminish their own hard-earned ability, but rather enhances and extends their capabilities.&#xA;&#xA;On the other hand, for students–who are by definition novices in the skills and knowledge they are learning–an over-reliance on AI tools may limit their ability to develop skills such as literacy, critical thinking, problem-solving, and creativity. &#xA;&#xA;Recent reports on AI in education, such as from Cognitive Resonance, Center for American Progress, and Bellwether, have rightfully raised this concern.&#xA;&#xA;And all educators, whether K-12 or in higher ed, are seeing an increasing use of AI by students to complete homework assignments, so this danger of truncating the development of internal capacity is real.&#xA;&#xA;I think the steps we can to take to address this are two-fold:&#xA;&#xA;limit the use of digital technology for learners at the earliest stages of learning, whether learners are preK-3 or learners being introduced to a new concept&#xA;&#xA;move practice of essential skills directly into the classroom as much as possible, while considering how AI could be used to extend, rather than diminish, any practice and feedback outside of the classroom&#xA;&#xA;In a post on ethical use of AI, Jacob Kaplan-Moss argues that fully automated AI is unethical in the public sector due to its inherent biases and potential for unfairness in high-stakes situations. In contrast, the assistive use of AI can enhance human decision-making.&#xA;&#xA;This assistive vs. automated use of AI may be a useful frame for thinking of how AI can be used most ethically and effectively in education. We want AI to be used to assist the learning process, rather than simply automating the solving of math problems or writing essays. This view aligns with Ethan Mollick’s idea of “co-intelligence” as well.&#xA;&#xA;So far, I find the most powerful and interesting assistive applications for AI are more focused on educators (“the experts”), rather than on students (“the novices”). Teachers can leverage AI to support administrative tasks, analysis of student data, and consider additional enhancements of their instruction based on student data.&#xA;&#xA;That said, I don’t think the assistive use cases of AI are only limited to “experts” in a domain. AI can also help to equip those without knowledge and expertise in a specific area with the language they need to navigate learning or real-world communications more effectively. And there are some really interesting use cases of AI for feedback on student thinking and writing, when structured with specific guidelines and criteria and with the teacher in the loop.&#xA;&#xA;But in the context of classroom learning, such uses must be very strategically designed and cautiously incorporated. For example, see this explanation from professor Michael Brenner on how he has begun incorporating AI into his pedagogy. But note this example is from a graduate level math class, so again, that novice vs. expert dynamic is quite different from what we would need to consider at a preK-8 level. But even at that graduate level, you can see there is quite a bit of complexity the instructor needed to consider and think through to design his course to leverage LLMs so strategically.&#xA;&#xA;There’s a lot more to unpack here on all sides of the equation. I’ll leave this one here for now, accepting non-closure, and I hope to dig further into these tensions and opportunities in both this space and in my professional work.&#xA;&#xA;6) I claim that LLMs have shown that language can be learned without any innate programming or structure – therefore demonstrating the power of statistical, implicit learning&#xA;&#xA;I’d moved into the “Chomsky is wrong” camp for a while now, but I happened to listen to an interview of Jean-Rémi King recently, a scientist at Meta AI, by Stephen Wilson on The Language Neuroscience podcast (did I tell you I’m a nerd?). Towards the end of the conversation, King warns against writing off Chomsky too readily, and that there is something intrinsic to the human brain in its readiness for language.&#xA;&#xA;I uploaded the relevant portion of the transcript from the interview, and asked Claude AI for a concise summary of King&#39;s main claims, which it willingly obliged (while I’m sure it drew upon an unconscionable amount of energy):&#xA;&#xA;  King argues that human brains likely don&#39;t use the same &#34;next word prediction&#34; principle as large language models for language acquisition, primarily because humans are exposed to far less linguistic data than these models. &#xA;&#xA;  He contends that while language models have shown impressive capabilities, they are extremely inefficient compared to human language learning, suggesting that we&#39;re missing some fundamental principles of how humans acquire language so efficiently.&#xA;&#xA;While I think I’ve tried to temper most of my pronouncements throughout this series, I think it’s important to acknowledge that the fact that LLMs can learn language from statistical associations of word tokens alone does not mean that is exactly how we humans must also learn language.&#xA;&#xA;It is rather a proof of concept that language can be learned in this way (without any innate grammar or teaching of rules). But as King points out, this is via a scale of input that is ridiculously and exponentially larger than that of any child.&#xA;&#xA;That said, there are other Artificial Neural Networks (ANNs), such as in the research of Gašper Beguš, that learn from raw speech in an unsupervised manner, more closely mimicking human language acquisition. His lab has found interesting similarities between these ANNs and the human brain in processing language sounds – a parallel to King’s own research, which has found that LLM models can generate brain-like representations when predicting words from context. &#xA;&#xA;And there will continue to be research into tinier models trained off sparser, and potentially richer, data.&#xA;&#xA;But as King points to, there’s just so much more we need to learn. And this is exactly where I find all of this the most exciting.&#xA;&#xA;Where I may be most rightfully critiqued in my last post, and perhaps in other posts, may be in extrapolating from the theoretical demonstration of LLMs to implications for classrooms. &#xA;&#xA;So let me state my position a bit more clearly in case there was any confusion that I am falling onto the side of the Goodmans or something. Children need consistency, stability, clarity, and coherency in their learning experiences, and teaching what is most important to know for a given subject directly and explicitly is critical. For children at the earliest stages of learning abstract skills and content, such as learning to read, explicit and well-structured teaching is essential. At the same time, however, we need to ensure that students have abundant structured opportunities to apply and practice what they are learning – and this is where ensuring they are spending more time reading, writing, and talking–connected to the content of what we are teaching–is essential.&#xA;&#xA;If you have more critiques that I am missing in any of the above, please do share!&#xA;&#xA;Egads, I think I may actually have ANOTHER post left in me after all of this. Who knew LLMs would be such an interesting topic?!&#xA;&#xA;#language #literacy #AI #LLMs #cognition #research #computation #models&#xA;]]&gt;</description>
      <content:encoded><![CDATA[<p><img src="https://i.snap.as/2AqWHRP3.jpeg" alt="Novice bunny and expert bunny on bikes"/>
When I typically begin a series of blogs to conduct nerdy inquiry into an abstract topic, I don&#39;t generally know where I&#39;m going to end up. This series on LLMs was unusual in that in <a href="https://languageandliteracy.blog/language-and-llms">our first post</a>, I outlined pretty much the exact topics I would go on to cover.</p>

<p>Here&#39;s where I had spitballed we might go:</p>
<ul><li>The surprisingly inseparable interconnection between form and meaning</li>
<li>Blundering our way to computational precision through human communication; Or, the generative tension between regularity and randomness</li>
<li>The human (and now, machine) capacity for learning and using language may simply be a matter of scale</li>
<li>Is language as separable from thought (and, for that matter, from the world) as Cormac McCarthy said?</li>
<li>Implicit vs. explicit learning of language and literacy</li></ul>

<p>Indeed, we then went on to explore each of these areas, in that order. Cool!
</p>

<h2 id="some-hypotheses-from-this-series" id="some-hypotheses-from-this-series">Some Hypotheses from This Series</h2>

<p>What theories have we raised through this exploration?</p>

<p>1) LLMs gain their uncanny powers from <a href="https://languageandliteracy.blog/language-and-llms">the statistical nature of language itself</a>;
2) the meaning and experiences of our world are <a href="https://languageandliteracy.blog/the-algebra-of-language-unveiling-the-statistical-tapestry-of-form-and-meaning">more deeply entwined with the form and structure</a> of our language than we previously imagined;
3) LLMs may offer us an opportunity to further the <a href="https://languageandliteracy.blog/the-pathway-of-human-language-towards-computational-precision-in-llms">convergence between human and machine language</a>;
4) AI can potentially <a href="https://languageandliteracy.blog/scaling-our-capacity-for-processing-information">extend our cognitive abilities</a>, enabling us to process and understand far more information;
5) Both human and machine learning progresses <a href="https://languageandliteracy.blog/the-interplay-of-language-cognition-and-llms-where-fuzziness-meets-precision">from fuzzy, imprecise representations to higher precision</a>, and the greater the precision, the greater the effort and practice (or “compute”) that is required; and
6) LLMs challenge Chomsykan notions of innateness and suggest that <a href="https://write.as/manderson/llms-statistical-learning-and-explicit-teaching">implicit, statistical learning</a> alone can lead to gaining the grammatical structure and meaning of a language.</p>

<p>While I’ve been mostly positive and excited about the potential of AI (aside from pointing out how it is <a href="https://languageandliteracy.blog/scaling-our-capacity-for-processing-information">accelerating the looming ecological catastrophe</a> that seems to be our trajectory) I should probably pause here to acknowledge that there may be important counterpoints to many of these (perhaps somewhat starry-eyed) hypotheses.</p>

<h2 id="onto-the-counterclaims" id="onto-the-counterclaims">Onto the Counterclaims</h2>

<p>Let&#39;s take a more critical look at some of my claims:</p>

<h3 id="1-i-claim-that-language-is-fundamental-to-the-generative-powers-of-llms" id="1-i-claim-that-language-is-fundamental-to-the-generative-powers-of-llms">1) I claim that language is fundamental to the generative powers of LLMs.</h3>

<p>Yet Andrej Karpathy, who is no stranger to LLM development, <a href="https://x.com/karpathy/status/1835024197506187617">tweeted</a>:</p>

<blockquote><p>It&#39;s a bit sad and confusing that LLMs (“Large Language Models”) have little to do with language; It&#39;s just historical. They are highly general purpose technology for statistical modeling of token streams. A better name would be Autoregressive Transformers or something.</p>

<p>They don&#39;t care if the tokens happen to represent little text chunks. It could just as well be little image patches, audio chunks, action choices, molecules, or whatever. If you can reduce your problem to that of modeling token streams (for any arbitrary vocabulary of some set of discrete tokens), you can “throw an LLM at it.”</p></blockquote>

<p>I agree that LLMs are performing “statistical modeling of token streams,” and that “for any arbitrary vocabulary of some set of discrete tokens, you can ‘throw an LLM at it.’”</p>

<p>We now have multimodal LLMs that are modeling out of token streams of audio, visual, and text, and will no doubt have ones feeding from additional streams of sensory data as they are increasingly paired with cameras on humans, objects, and robots.</p>

<p>Yet I also think Karpathy undersells that when LLMs suddenly exploded into general public awareness and fascination, it was merely a “historical” fact that they were trained upon vast amounts of human generated text and were able to reproduce and generate human language. As we’ve explored in this and <a href="https://languageandliteracy.blog/innate-vs">a previous series</a>, there is something about human language itself uniquely adapted for our brain circuitry and the propagation of our culture within social interaction in our world. And being able to communicate with a powerful computational model through the medium of conversational human language has been a revolutionary advent. We are just in the beginning stages of grokking it.</p>

<p>As I <a href="https://x.com/mandercorn/status/1835279679650971667">tweeted in response</a> to Karpathy, token streams may be applied to anything, but human language seems to be uniquely suited to the advancement of combined human and machine learning. Not only because we rely on it for communication – but furthermore due to the <a href="https://languageandliteracy.blog/the-algebra-of-language-unveiling-the-statistical-tapestry-of-form-and-meaning">algebraic and statistical nature of our language</a>.</p>

<p>Recent case in point: the viral attention currently on NotebookLM’s Audio Overview. Listening to a conversation, however artificial, resonates with us, because that&#39;s what&#39;s in our social nature. And, surprisingly, it does a fairly good job of surfacing information from across multiple multimodal sources (and soon, across languages) that we find interesting, relevant, and meaningful.</p>

<p>Speaking of NotebookLM Audio Overview. . . here’s one derived from all the blog posts (except this one) from this series, as well as the sources–<a href="https://languageandliteracy.blog/language-and-llms">outlined in post 1</a>–that inspired them all: <a href="https://notebooklm.google.com/notebook/a4f35399-e288-4293-b2d2-0489e6b1f037/audio">https://notebooklm.google.com/notebook/a4f35399-e288-4293-b2d2-0489e6b1f037/audio</a></p>

<h3 id="4-i-claim-there-is-great-potential-for-ai-to-extend-our-cognitive-capabilities" id="4-i-claim-there-is-great-potential-for-ai-to-extend-our-cognitive-capabilities">4) I claim there is great potential for AI to extend our cognitive capabilities</h3>

<p>Yet there is a strong case of an equal and commensurate danger that use of LLMs can reduce our cognitive capabilities.</p>

<p>Learning more formal content and skills, like what we learn in school or in a job, requires deliberate effort until we develop an unconscious fluency. If students learning new concepts and skills externally automate their practice of new learning (such as writing or math) to an LLM, then they will not–ironically–be able to develop the automatized internal knowledge and capacity they need to wield powerful tools like AI more effectively.</p>

<p>When “experts” use tools like AI, they know where the gaps are in the output and are able to use it strategically to enhance their own production and output. A few examples of this:</p>
<ul><li><p>Simon Willison, a programmer who is <a href="https://simonwillison.net/">also a great communicator</a>, uses different LLMs to support his projects, and writes and speaks about how he does so. <a href="https://newsletter.pragmaticengineer.com/p/ai-tools-for-software-engineers-simon-willison">Here’s a podcast</a>, for example, where he explains how he uses them.</p></li>

<li><p>Nicholas Carlini, a research scientist at Google DeepMind, similarly <a href="https://nicholas.carlini.com/writing/2024/how-i-use-ai.html">wrote about how he uses AI</a> to support his work.</p></li>

<li><p>Cal Newport, who writes extensively about how to do “deep work” in a world of distractions, recently <a href="https://www.newyorker.com/culture/annals-of-inquiry/what-kind-of-writer-is-chatgpt">wrote in The New Yorker</a> how he has found ChatGPT useful to his writing.</p></li></ul>

<p>All the people above are highly skilled at what they do – so when they explore and then figure out how to use AI to support their work, they do so in a way that does not diminish their own hard-earned ability, but rather enhances and extends their capabilities.</p>

<p>On the other hand, for students–who are by definition <strong>novices</strong> in the skills and knowledge they are learning–an over-reliance on AI tools may limit their ability to develop skills such as literacy, critical thinking, problem-solving, and creativity.</p>

<p>Recent reports on AI in education, such as from <a href="https://cognitiveresonance.net/resources.html">Cognitive Resonance</a>, <a href="https://www.americanprogress.org/article/using-learning-science-to-analyze-the-risks-and-benefits-of-ai-in-k-12-education/">Center for American Progress</a>, and <a href="https://bellwether.org/publications/learning-systems/">Bellwether</a>, have rightfully raised this concern.</p>

<p>And all educators, whether K-12 or in higher ed, are seeing an increasing use of AI by students to complete homework assignments, so this danger of truncating the development of internal capacity is real.</p>

<p>I think the steps we can to take to address this are two-fold:</p>
<ul><li><p>limit the use of digital technology for learners at the earliest stages of learning, whether learners are preK-3 or learners being introduced to a new concept</p></li>

<li><p>move practice of essential skills directly into the classroom as much as possible, while considering how AI could be used to extend, rather than diminish, any practice and feedback outside of the classroom</p></li></ul>

<p>In <a href="https://jacobian.org/2024/oct/1/ethical-public-sector-ai/">a post on ethical use of AI</a>, Jacob Kaplan-Moss argues that fully automated AI is unethical in the public sector due to its inherent biases and potential for unfairness in high-stakes situations. In contrast, the assistive use of AI can enhance human decision-making.</p>

<p>This assistive vs. automated use of AI may be a useful frame for thinking of how AI can be used most ethically and effectively in education. We want AI to be used to assist the learning process, rather than simply automating the solving of math problems or writing essays. This view aligns with <a href="https://english.elpais.com/technology/2024-10-03/ethan-mollick-analyst-students-who-use-ai-as-a-crutch-dont-learn-anything.html">Ethan Mollick’s idea of “co-intelligence”</a> as well.</p>

<p>So far, I find the most powerful and interesting assistive applications for AI are more focused on educators (“the experts”), rather than on students (“the novices”). Teachers can leverage AI to support administrative tasks, analysis of student data, and consider additional enhancements of their instruction based on student data.</p>

<p>That said, I don’t think the assistive use cases of AI are only limited to “experts” in a domain. AI can also help to equip those without knowledge and expertise in a specific area with the language they need to navigate learning or real-world communications more effectively. And there are some really interesting use cases of AI for feedback on student thinking and writing, when structured with specific guidelines and criteria and with the teacher in the loop.</p>

<p>But in the context of classroom learning, such uses must be very strategically designed and cautiously incorporated. For example, <a href="https://x.com/SebastienBubeck/status/1829701643925151757">see this explanation from professor Michael Brenner</a> on how he has begun incorporating AI into his pedagogy. But note this example is from a graduate level math class, so again, that novice vs. expert dynamic is quite different from what we would need to consider at a preK-8 level. But even at that graduate level, you can see there is quite a bit of complexity the instructor needed to consider and think through to design his course to leverage LLMs so strategically.</p>

<p>There’s a lot more to unpack here on all sides of the equation. I’ll leave this one here for now, accepting non-closure, and I hope to dig further into these tensions and opportunities in both this space and in my professional work.</p>

<h3 id="6-i-claim-that-llms-have-shown-that-language-can-be-learned-without-any-innate-programming-or-structure-therefore-demonstrating-the-power-of-statistical-implicit-learning" id="6-i-claim-that-llms-have-shown-that-language-can-be-learned-without-any-innate-programming-or-structure-therefore-demonstrating-the-power-of-statistical-implicit-learning">6) I claim that LLMs have shown that language can be learned without any innate programming or structure – therefore demonstrating the power of statistical, implicit learning</h3>

<p>I’d moved into the “Chomsky is wrong” camp for a while now, but I happened to listen to an interview of Jean-Rémi King recently, a scientist at Meta AI, by Stephen Wilson on <a href="https://langneurosci.org/podcast/#ep27">The Language Neuroscience podcast</a> (did I tell you I’m a nerd?). Towards the end of the conversation, King warns against writing off Chomsky too readily, and that there is something intrinsic to the human brain in its readiness for language.</p>

<p>I uploaded the relevant portion of the transcript from the interview, and asked Claude AI for a concise summary of King&#39;s main claims, which it willingly obliged (while I’m sure it drew upon an unconscionable amount of energy):</p>

<blockquote><p>King argues that human brains likely don&#39;t use the same “next word prediction” principle as large language models for language acquisition, primarily because humans are exposed to far less linguistic data than these models.</p>

<p>He contends that while language models have shown impressive capabilities, they are extremely inefficient compared to human language learning, suggesting that we&#39;re missing some fundamental principles of how humans acquire language so efficiently.</p></blockquote>

<p>While I think I’ve tried to temper most of my pronouncements throughout this series, I think it’s important to acknowledge that the fact that LLMs can learn language from statistical associations of word tokens alone does not mean that is exactly how we humans must also learn language.</p>

<p>It is rather a proof of concept that language <em>can</em> be learned in this way (without any innate grammar or teaching of rules). But as King points out, this is via a scale of input that is ridiculously and exponentially larger than that of any child.</p>

<p>That said, there are other Artificial Neural Networks (ANNs), such as in <a href="https://medium.com/@begus.gasper/artificial-and-biological-intelligence-humans-animals-and-machines-142bc3c4b304">the research of Gašper Beguš</a>, that learn from raw speech in an unsupervised manner, more closely mimicking human language acquisition. His lab has found interesting similarities between these ANNs and the human brain in processing language sounds – a parallel to King’s own research, which has found that LLM models can generate brain-like representations when predicting words from context.</p>

<p>And there will continue to be research into tinier models trained off sparser, and potentially richer, data.</p>

<p>But as King points to, there’s just so much more we need to learn. And this is exactly where I find all of this the most exciting.</p>

<p>Where I may be most rightfully critiqued <a href="https://write.as/manderson/llms-statistical-learning-and-explicit-teaching">in my last post</a>, and perhaps in other posts, may be in extrapolating from the theoretical demonstration of LLMs to implications for classrooms.</p>

<p>So let me state my position a bit more clearly in case there was any confusion that I am falling onto the side of <a href="https://write.as/manderson/learning-to-read-an-unnatural-act">the Goodmans</a> or something. Children need consistency, stability, clarity, and coherency in their learning experiences, and teaching what is most important to know for a given subject directly and explicitly is critical. For children at the earliest stages of learning abstract skills and content, such as learning to read, explicit and well-structured teaching is essential. At the same time, however, we need to ensure that students have abundant structured opportunities to apply and practice what they are learning – and this is where ensuring they are spending more time reading, writing, and talking–connected to the content of what we are teaching–is essential.</p>

<p>If you have more critiques that I am missing in any of the above, please do share!</p>

<p>Egads, I think I may actually have ANOTHER post left in me after all of this. Who knew LLMs would be such an interesting topic?!</p>

<p><a href="https://languageandliteracy.blog/tag:language" class="hashtag"><span>#</span><span class="p-category">language</span></a> <a href="https://languageandliteracy.blog/tag:literacy" class="hashtag"><span>#</span><span class="p-category">literacy</span></a> <a href="https://languageandliteracy.blog/tag:AI" class="hashtag"><span>#</span><span class="p-category">AI</span></a> <a href="https://languageandliteracy.blog/tag:LLMs" class="hashtag"><span>#</span><span class="p-category">LLMs</span></a> <a href="https://languageandliteracy.blog/tag:cognition" class="hashtag"><span>#</span><span class="p-category">cognition</span></a> <a href="https://languageandliteracy.blog/tag:research" class="hashtag"><span>#</span><span class="p-category">research</span></a> <a href="https://languageandliteracy.blog/tag:computation" class="hashtag"><span>#</span><span class="p-category">computation</span></a> <a href="https://languageandliteracy.blog/tag:models" class="hashtag"><span>#</span><span class="p-category">models</span></a></p>
]]></content:encoded>
      <guid>https://languageandliteracy.blog/reviewing-claims-ive-made-on-llms</guid>
      <pubDate>Mon, 07 Oct 2024 00:10:15 +0000</pubDate>
    </item>
    <item>
      <title>The Pathway of Human Language Towards Computational Precision in LLMs</title>
      <link>https://languageandliteracy.blog/the-pathway-of-human-language-towards-computational-precision-in-llms?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[Natural digital&#xA;&#xA;Regularity and irregularity. Decodable and tricky words. Learnability and surprisal. Predictability and randomness. Low entropy and high entropy.&#xA;&#xA;Why do such tensions exist in human language? And in our AI tools developed to both create code and use natural language, how can the precision required for computation co-exist alongside this necessary complexity and messiness of our human language?&#xA;!--more--&#xA;An Algebraic Symphony of Language and Meaning&#xA;&#xA;In our last post, we examined how there is a statistical and algebraic nature to language that drives the power of LLMs, and that the form and meaning of a language may be much more intertwined than we assume, given just how much meaning LLMs are able to approximate via computation of statistical arrays alone.&#xA;&#xA;This interlacement of form and meaning is in relation to where and how words show up repeatedly in sentences and texts, not simply in the form of words themselves. Because all languages contain words that have the same form but different meanings. Some words that share the same form have entirely unrelated meanings (homophony), while other words with the same form have closely related meanings (polysemy). Yet LLMs are able to use them in a more or less “natural” manner due to the high dimensional mappings of word parts in statistical relation to one another – such that word analogies can be calculated mathematically:&#xA;&#xA;  “For example, Google researchers took the vector for biggest, subtracted big, and added small. The word closest to the resulting vector was smallest.”&#xA;  –Large language models, explained with a minimum of math and jargon, Timothy Lee &amp; Sean Trott&#xA;&#xA;That the algebraic and statistical relationship of words in natural language can drive computational models&#39; generative capabilities suggests that language itself may reflect the limitations and potential of AI. And the thing with natural, human language is that while it is endlessly generative, it also tends to be imprecise. For our human usage, gestures and the context of our social interaction, who and when we are speaking to, plays a big role. As long as we get our main message across, we’re good.&#xA;&#xA;Human language is fundamentally communicative and social, and there’s feelings involved.&#xA;&#xA;The Imprecision of Human Expression&#xA;&#xA;Imagine yourself in a bustling restaurant in an international airport, surrounded by people from diverse linguistic backgrounds. You&#39;re trying to communicate with a traveller whose language you don&#39;t speak. What do you do?&#xA;&#xA;You resort to body language. You gesture hyperbolically and make exaggerated facial expressions. You point to objects, mime actions, and mouth simple words you hope the other person might use as a basis for basic understanding.&#xA;&#xA;Depth, nuance, and complexity are not possible (beyond each individual’s imagination) in this most elemental of interactions.&#xA;&#xA;So what is required for depth, nuance, and complexity?&#xA;&#xA;A shared language, whether spoken, written, or signed. In which a small set of sounds, letters, or signs are concatenated in a wide assortment of ways, both commonplace and surprising, to convey a wide assortment of ideas and feelings.&#xA;&#xA;Yet a shared language, while providing a platform for greater depth, may still remain imprecise. What is meant to be conveyed is not always exactly what is understood.&#xA;&#xA;There are furthermore gradations of precision in language, beginning with the ephemeral and contextual nature of spoken and signed language, moving into the more ossified form of written language, in which spelling must be exact and word selection must be more intentional. There is also a movement from the language we use with our family, with frequent, commonly used words, to the language we use when writing an academic paper, with domain-specific, rarer words. In education, we often refer to this type of language as Tier 2 or 3 vocabulary.&#xA;&#xA;Tiered Vocabulary&#xA;&#xA;If a person is equipped with more of that academic, domain-specific language, then greater precision in communication can be achieved. Yet the challenge of whether the listener hears and interprets what is intended remains. For example, in this article in Scientific American, “People Have Very Different Understandings of Even the Simplest Words”, they discuss how the more abstract a word, the more it can be tied to an emotional valence and someone’s identity and experiences, rather than a precise meaning.&#xA;&#xA;The Computational Imperative&#xA;&#xA;But in some ways, this inherent fuzziness of our language may be a feature, rather than a bug. It gives us a complex adaptive system for navigating, creating, and communicating in a world of complex adaptive systems.&#xA;&#xA;For computers and computations, however, exactness and precision in language is required – either a line of code input runs the correct function as an output or it doesn’t. So it’s quite interesting that one of the most immediately powerful use cases so far of LLMs seems to be as a natural language interface to develop and review code.&#xA;&#xA;Stephen Wolfram, in a long and interesting explainer on how LLMs work, [“What Is ChatGPT Doing … and Why Does It Work?”](&#xA;https://writings.stephenwolfram.com/2023/02/what-is-chatgpt-doing-and-why-does-it-work/), explores some of this tension between computational and natural language. &#xA;&#xA;  “Human language is fundamentally imprecise, not least because it isn’t “tethered” to a specific computational implementation, and its meaning is basically defined just by a “social contract” between its users. But computational language, by its nature, has a certain fundamental precision—because in the end what it specifies can always be “unambiguously executed on a computer”. Human language can usually get away with a certain vagueness. (When we say “planet” does it include exoplanets or not, etc.?) But in computational language we have to be precise and clear about all the distinctions we’re making.”&#xA;&#xA;Computational Irreducibility and the Limits of Predictability and Learning&#xA;&#xA;One of the limitations Wolfram raises between human and computational language is what he terms “computational irreducibility,” a term he uses to describe the difficulty in making accurate predictions for a highly complex system, such as for weather or climate systems. For such systems, it would require performing step-by-step computation based on an initial state, and thus can’t be swiftly calculated by compressing data.&#xA;&#xA;In some ways, this “compression” of information is what we are doing with language as we use more “Tier 2” and “Tier 3” – or academic – words in our speech or writing. There is a greater density of information provided in academic speech and writing, in which more abstract words are used to convey complex concepts, and our sentences tend to become more compound and complex. The simpler, more frequent words, phrases, and sentences we use in our everyday speech are more regular and thus, more learnable.&#xA;&#xA;  . . . there’s just a fundamental tension between learnability and computational irreducibility. Learning involves in effect compressing data by leveraging regularities. But computational irreducibility implies that ultimately there’s a limit to what regularities there may be.&#xA;&#xA;  . . . there’s an ultimate tradeoff between capability and trainability: the more you want a system to make “true use” of its computational capabilities, the more it’s going to show computational irreducibility, and the less it’s going to be trainable. And the more it’s fundamentally trainable, the less it’s going to be able to do sophisticated computation.&#xA;&#xA;Irregularity and Regularity in Language&#xA;&#xA;What’s interesting to note here is that all languages have constructive tensions between regularity and irregularity. This tension may be a process of language being honed over time to be more learnable within our cognitive constraints. We’ve explored some of this before in our post, Irregularity Enhances Learning (Maybe), in which we examined a paper by Michael Ramscar that suggested there is some level of tension between language forms that show up again and again, and the language forms that are more infrequent, but thus inherently gain more of our attention. This relates to the theory of “statistical learning” with which we not only learn language, but also when we map a language to its written form.&#xA;&#xA;For Wolfram, that LLMs are as powerful as they are suggests that human language is actually more statistically regular than we may have thought: &#xA;&#xA;  “my strong suspicion is that the success of ChatGPT implicitly reveals an important “scientific” fact: that there’s actually a lot more structure and simplicity to meaningful human language than we ever knew—and that in the end there may be even fairly simple rules that describe how such language can be put together.”&#xA;&#xA;  And instead what we should conclude is that tasks—like writing essays—that we humans could do, but we didn’t think computers could do, are actually in some sense computationally easier than we thought.&#xA;&#xA;  In other words, the reason a neural net can be successful in writing an essay is because writing an essay turns out to be a “computationally shallower” problem than we thought. And in a sense this takes us closer to “having a theory” of how we humans manage to do things like writing essays, or in general deal with language.&#xA;&#xA;And so thus far the unrealized potential, for Wolfram, is that with a greater underlying capability in AI for computational language, it may be able to bridge our more “computationally shallow” human language with the precision required for more complex computations:&#xA;&#xA;  ”its very success gives us a reason to think that it’s going to be feasible to construct something more complete in computational language form. And, unlike what we’ve so far figured out about the innards of ChatGPT, we can expect to design the computational language so that it’s readily understandable to humans.”&#xA;&#xA;Decontextualized Language: The Pathway to Precision&#xA;&#xA;On this pathway towards integration of human language and computational language, it’s interesting to consider how in our own language development,  we are able to better “compress information” and develop greater precision in our thinking and communication as we learn and incorporate rarer and more abstract language into our own. We’ve spoken before about “decontextualized language” – the language that takes us beyond the immediate context and moment, and how such language can take us beyond our own delimited feelings and experiences, and into a realm of interpersonal and cultural thought, knowledge, and perspectives. This is the language of storybooks, of science, and – at it’s greatest extreme – of code. We begin teaching this form of language when we engage in storytelling with our children and reading with them and talking to them about books. It becomes increasingly dense and complex as we move into disciplinary study.&#xA;&#xA;There is some evidence that training LLMs on this specific form of language is more powerful – such as this study training a “tiny LLM” on children’s stories. And if you think about what LLMs have been getting trained on thus far – it’s a corpus of written language, not training on conversations using everyday language. As we’ve explored in depth on this blog, written language is not synonymous with oral language – by nature of it being written, it is already more “decontextualized,” and requires more inference and perspective-taking. That LLMs are trained on such a corpus may be, in fact, why their algebraic and statistical magic can be so surprisingly powerful. There is a greater density of information in the written forms of our languages.&#xA;&#xA;Implications for Teaching and Learning&#xA;&#xA;What might all of this say about teaching and learning? Well, so far, one of the facets we’ve highlighted from LLMs is that the statistical nature of language alone can take us pretty far, which suggests that alongside of social interaction and peer engagement and communication, we want to increase the volume of that language exposure and use. And in terms of the nature of the language we want to increase: the more that the form of that language combines precision with abstraction, the greater computational power it can provide. Turning up the dial on decontextualized language use and exposure – in other words – providing our children with “textual feasts,” to use Alfred Tatum’s term, may be the key to enhanced learning.&#xA;&#xA;Sources for Further Exploration&#xA;&#xA;If you are interested in further exploring some of the tensions we began this post with – between regularity and irregularity in language, here’s some further interesting reads to geek out on:&#xA;&#xA;“Source codes in human communication” by Michael Ramscar&#xA;“Expectation-based syntactic comprehension” by Roger Levy&#xA;“Cognitive approaches to uniformity and variability in morphology” by Petar Milin, Neil Bermel, and James Blevins&#xA;&#xA;#language #computation #algorithms #learning #LLMs #cognition]]&gt;</description>
      <content:encoded><![CDATA[<p><img src="https://i.snap.as/PmjZi6Hw.png" alt="Natural digital"/></p>

<p>Regularity and irregularity. Decodable and tricky words. Learnability and <a href="https://www.sciencedirect.com/science/article/abs/pii/S0010027707001436?via%3Dihub">surprisal</a>. Predictability and randomness. Low entropy and high entropy.</p>

<p>Why do such tensions exist in human language? And in our AI tools developed to both create code and use natural language, how can the precision required for computation co-exist alongside this necessary complexity and messiness of our human language?
</p>

<h2 id="an-algebraic-symphony-of-language-and-meaning" id="an-algebraic-symphony-of-language-and-meaning">An Algebraic Symphony of Language and Meaning</h2>

<p><a href="https://write.as/manderson/the-algebra-of-language-unveiling-the-statistical-tapestry-of-form-and-meaning">In our last post</a>, we examined how there is a statistical and algebraic nature to language that drives the power of LLMs, and that the form and meaning of a language may be much more intertwined than we assume, given just how much meaning LLMs are able to approximate via computation of statistical arrays alone.</p>

<p>This interlacement of form and meaning is in relation to where and how words show up repeatedly in sentences and texts, not simply in the form of words themselves. Because all languages contain words that have the same form but different meanings. Some words that share the same form have entirely unrelated meanings (homophony), while other words with the same form have closely related meanings (polysemy). Yet LLMs are able to use them in a more or less “natural” manner due to the high dimensional mappings of word parts in statistical relation to one another – such that word analogies can be calculated mathematically:</p>

<blockquote><p>“For example, Google researchers took the vector for <strong>biggest</strong>, subtracted <strong>big</strong>, and added <strong>small</strong>. The word closest to the resulting vector was <strong>smallest</strong>.”
–<a href="https://www.understandingai.org/p/large-language-models-explained-with">Large language models, explained with a minimum of math and jargon</a>, Timothy Lee &amp; Sean Trott</p></blockquote>

<p>That the algebraic and statistical relationship of words in natural language can drive computational models&#39; generative capabilities suggests that language itself may reflect the limitations and potential of AI. And the thing with natural, human language is that while it is endlessly generative, it also tends to be imprecise. For our human usage, gestures and the context of our social interaction, who and when we are speaking to, plays a big role. As long as we get our main message across, we’re good.</p>

<p>Human language is fundamentally communicative and social, and there’s feelings involved.</p>

<h2 id="the-imprecision-of-human-expression" id="the-imprecision-of-human-expression">The Imprecision of Human Expression</h2>

<p>Imagine yourself in a bustling restaurant in an international airport, surrounded by people from diverse linguistic backgrounds. You&#39;re trying to communicate with a traveller whose language you don&#39;t speak. What do you do?</p>

<p>You resort to body language. You gesture hyperbolically and make exaggerated facial expressions. You point to objects, mime actions, and mouth simple words you hope the other person might use as a basis for basic understanding.</p>

<p>Depth, nuance, and complexity are not possible (beyond each individual’s imagination) in this most elemental of interactions.</p>

<p>So what <em>is</em> required for depth, nuance, and complexity?</p>

<p>A shared language, whether spoken, written, or signed. In which a small set of sounds, letters, or signs are concatenated in a wide assortment of ways, both commonplace and surprising, to convey a wide assortment of ideas and feelings.</p>

<p>Yet a shared language, while providing a platform for greater depth, may still remain imprecise. What is meant to be conveyed is not always exactly what is understood.</p>

<p>There are furthermore gradations of precision in language, beginning with the ephemeral and contextual nature of spoken and signed language, moving into the more ossified form of written language, in which spelling must be exact and word selection must be more intentional. There is also a movement from the language we use with our family, with frequent, commonly used words, to the language we use when writing an academic paper, with domain-specific, rarer words. In education, we often refer to this type of language as Tier 2 or 3 vocabulary.</p>

<p><img src="https://i.snap.as/I36of2l3.png" alt="Tiered Vocabulary"/></p>

<p>If a person is equipped with more of that academic, domain-specific language, then greater precision in communication can be achieved. Yet the challenge of whether the listener hears and interprets what is intended remains. For example, in this article in Scientific American, <a href="https://www.scientificamerican.com/article/people-have-very-different-understandings-of-even-the-simplest-words/">“People Have Very Different Understandings of Even the Simplest Words”</a>, they discuss how the more abstract a word, the more it can be tied to an emotional valence and someone’s identity and experiences, rather than a precise meaning.</p>

<h2 id="the-computational-imperative" id="the-computational-imperative">The Computational Imperative</h2>

<p>But in some ways, this inherent fuzziness of our language may be a feature, rather than a bug. It gives us a complex adaptive system for navigating, creating, and communicating in a world of complex adaptive systems.</p>

<p>For computers and computations, however, exactness and precision in language is required – either a line of code input runs the correct function as an output or it doesn’t. So it’s quite interesting that one of the most immediately powerful use cases so far of LLMs seems to be as a natural language interface to develop and review code.</p>

<p>Stephen Wolfram, in a long and interesting explainer on how LLMs work, <a href="https://writings.stephenwolfram.com/2023/02/what-is-chatgpt-doing-and-why-does-it-work/">“What Is ChatGPT Doing … and Why Does It Work?”</a>, explores some of this tension between computational and natural language.</p>

<blockquote><p>“Human language is fundamentally imprecise, not least because it isn’t “tethered” to a specific computational implementation, and its meaning is basically defined just by a “social contract” between its users. But computational language, by its nature, has a certain fundamental precision—because in the end what it specifies can always be “unambiguously executed on a computer”. Human language can usually get away with a certain vagueness. (When we say “planet” does it include exoplanets or not, etc.?) But in computational language we have to be precise and clear about all the distinctions we’re making.”</p></blockquote>

<h2 id="computational-irreducibility-and-the-limits-of-predictability-and-learning" id="computational-irreducibility-and-the-limits-of-predictability-and-learning">Computational Irreducibility and the Limits of Predictability and Learning</h2>

<p>One of the limitations Wolfram raises between human and computational language is what he terms “computational irreducibility,” a term he uses to describe the difficulty in making accurate predictions for a highly complex system, such as for weather or climate systems. For such systems, it would require performing step-by-step computation based on an initial state, and thus can’t be swiftly calculated by compressing data.</p>

<p>In some ways, this “compression” of information is what we are doing with language as we use more “Tier 2” and “Tier 3” – or academic – words in our speech or writing. There is a greater density of information provided in academic speech and writing, in which more abstract words are used to convey complex concepts, and our sentences tend to become more compound and complex. The simpler, more frequent words, phrases, and sentences we use in our everyday speech are more regular and thus, more learnable.</p>

<blockquote><p>. . . there’s just a fundamental tension between learnability and computational irreducibility. Learning involves in effect compressing data by leveraging regularities. But computational irreducibility implies that ultimately there’s a limit to what regularities there may be.</p>

<p>. . . there’s an ultimate tradeoff between capability and trainability: the more you want a system to make “true use” of its computational capabilities, the more it’s going to show computational irreducibility, and the less it’s going to be trainable. And the more it’s fundamentally trainable, the less it’s going to be able to do sophisticated computation.</p></blockquote>

<h2 id="irregularity-and-regularity-in-language" id="irregularity-and-regularity-in-language">Irregularity and Regularity in Language</h2>

<p>What’s interesting to note here is that all languages have constructive tensions between regularity and irregularity. This tension may be a process of language being honed over time to be more learnable within our cognitive constraints. We’ve explored some of this before in our post, <a href="https://languageandliteracy.blog/irregularity-enhances-learning-maybe">Irregularity Enhances Learning (Maybe)</a>, in which we examined a paper by Michael Ramscar that suggested there is some level of tension between language forms that show up again and again, and the language forms that are more infrequent, but thus inherently gain more of our attention. This relates to the theory of “statistical learning” with which we not only learn language, but also when we map a language to its written form.</p>

<p>For Wolfram, that LLMs are as powerful as they are suggests that human language is actually more statistically regular than we may have thought:</p>

<blockquote><p>“my strong suspicion is that the success of ChatGPT implicitly reveals an important “scientific” fact: that there’s actually a lot more structure and simplicity to meaningful human language than we ever knew—and that in the end there may be even fairly simple rules that describe how such language can be put together.”</p>

<p>And instead what we should conclude is that tasks—like writing essays—that we humans could do, but we didn’t think computers could do, are actually in some sense computationally easier than we thought.</p>

<p>In other words, the reason a neural net can be successful in writing an essay is because writing an essay turns out to be a “computationally shallower” problem than we thought. And in a sense this takes us closer to “having a theory” of how we humans manage to do things like writing essays, or in general deal with language.</p></blockquote>

<p>And so thus far the unrealized potential, for Wolfram, is that with a greater underlying capability in AI for computational language, it may be able to bridge our more “computationally shallow” human language with the precision required for more complex computations:</p>

<blockquote><p>”its very success gives us a reason to think that it’s going to be feasible to construct something more complete in computational language form. And, unlike what we’ve so far figured out about the innards of ChatGPT, we can expect to design the computational language so that it’s readily understandable to humans.”</p></blockquote>

<h2 id="decontextualized-language-the-pathway-to-precision" id="decontextualized-language-the-pathway-to-precision">Decontextualized Language: The Pathway to Precision</h2>

<p>On this pathway towards integration of human language and computational language, it’s interesting to consider how in our own language development,  we are able to better “compress information” and develop greater precision in our thinking and communication as we learn and incorporate rarer and more abstract language into our own. We’ve spoken before about <a href="https://write.as/manderson/research-highlight-2-the-language-teachers-use-influences-the-language">“decontextualized language”</a> – the language that takes us beyond the immediate context and moment, and how such language can take us beyond our own delimited feelings and experiences, and into a realm of interpersonal and cultural thought, knowledge, and perspectives. This is the language of storybooks, of science, and – at it’s greatest extreme – of code. We begin teaching this form of language when we engage in storytelling with our children and reading with them and talking to them about books. It becomes increasingly dense and complex as we move into disciplinary study.</p>

<p>There is some evidence that training LLMs on this specific form of language is more powerful – <a href="https://www.quantamagazine.org/tiny-language-models-thrive-with-gpt-4-as-a-teacher-20231005/">such as this study</a> training a “tiny LLM” on children’s stories. And if you think about what LLMs have been getting trained on thus far – it’s a corpus of written language, not training on conversations using everyday language. As we’ve explored in depth on this blog, written language is not synonymous with oral language – by nature of it being written, it is already more “decontextualized,” and requires more inference and perspective-taking. That LLMs are trained on such a corpus may be, in fact, why their algebraic and statistical magic can be so surprisingly powerful. There is a greater density of information in the written forms of our languages.</p>

<h2 id="implications-for-teaching-and-learning" id="implications-for-teaching-and-learning">Implications for Teaching and Learning</h2>

<p>What might all of this say about teaching and learning? Well, so far, one of the facets we’ve highlighted from LLMs is that the statistical nature of language alone can take us pretty far, which suggests that alongside of social interaction and peer engagement and communication, we want to increase the volume of that language exposure and use. And in terms of the nature of the language we want to increase: the more that the form of that language combines precision with abstraction, the greater computational power it can provide. Turning up the dial on decontextualized language use and exposure – in other words – providing our children with <a href="https://languageandliteracy.blog/provide-our-students-with-textual-feasts">“textual feasts,”</a> to use Alfred Tatum’s term, may be the key to enhanced learning.</p>

<h3 id="sources-for-further-exploration" id="sources-for-further-exploration">Sources for Further Exploration</h3>

<p>If you are interested in further exploring some of the tensions we began this post with – between regularity and irregularity in language, here’s some further interesting reads to geek out on:</p>
<ul><li><a href="https://arxiv.org/pdf/1904.03991">“Source codes in human communication”</a> by Michael Ramscar</li>
<li><a href="https://www.sciencedirect.com/science/article/abs/pii/S0010027707001436?via%3Dihub">“Expectation-based syntactic comprehension”</a> by Roger Levy</li>
<li><a href="https://www.degruyter.com/document/doi/10.1515/cog-2024-0027/html">“Cognitive approaches to uniformity and variability in morphology”</a> by Petar Milin, Neil Bermel, and James Blevins</li></ul>

<p><a href="https://languageandliteracy.blog/tag:language" class="hashtag"><span>#</span><span class="p-category">language</span></a> <a href="https://languageandliteracy.blog/tag:computation" class="hashtag"><span>#</span><span class="p-category">computation</span></a> <a href="https://languageandliteracy.blog/tag:algorithms" class="hashtag"><span>#</span><span class="p-category">algorithms</span></a> <a href="https://languageandliteracy.blog/tag:learning" class="hashtag"><span>#</span><span class="p-category">learning</span></a> <a href="https://languageandliteracy.blog/tag:LLMs" class="hashtag"><span>#</span><span class="p-category">LLMs</span></a> <a href="https://languageandliteracy.blog/tag:cognition" class="hashtag"><span>#</span><span class="p-category">cognition</span></a></p>
]]></content:encoded>
      <guid>https://languageandliteracy.blog/the-pathway-of-human-language-towards-computational-precision-in-llms</guid>
      <pubDate>Sun, 19 May 2024 15:05:11 +0000</pubDate>
    </item>
    <item>
      <title>Language, Cognition, and LLMs</title>
      <link>https://languageandliteracy.blog/language-and-llms?pk_campaign=rss-feed</link>
      <description>&lt;![CDATA[“Semantic gradients,” are a tool used by teachers to broaden and deepen students&#39; understanding of related words by plotting them in relation to one another. They often begin with antonyms at each end of the continuum. Here are two basic examples:&#xA;&#xA;Semantic gradient examples&#xA;&#xA;Now imagine taking this approach and quantifying the relationships between words by adding numbers to the line graph. Now imagine adding another axis to this graph, so that words are plotted in a three dimensional space in their relationships. Then add another dimension, and another . . . heck, make it tens of thousands more dimensions, relating all the words available in your lexicon across a high dimensional space. . .&#xA;&#xA;. . . and you may begin to envision one of the fundamental powers of Large Language Models (LLMs).&#xA;&#xA;!--more--&#xA;&#xA;LLMs Are Powered by Language: Or, Words as a Vast Sea of Interrelated Statistical Arrays of Tokens&#xA;&#xA;At root, the most powerful current forms of AI derive their capacities from decomposing human language into vast arrays of numbers based on their high dimensional statistical relationships and then predicting probabilistically what the next tokens are most likely to be. &#xA;&#xA;There’s a kind of alchemical transformation that occurs that seems to maintain the meaning in the generative pronouncements of the frontier LLMs, all the more amazing because so far the very engineers who have designed the structure for these operations to occur do not fully understand what the models are doing to arrive at their seemingly oracular destinations.&#xA;&#xA;In other words – the power of LLMs seemingly derives from the statistical power of language. There is something in the nature of language itself that seems to provide these computations of vast arrays of numbers with a lattice of our world, enabling LLMs to gain uncanny abilities from superpowered next word prediction. That LLMs have the generative powers they have—and that they have them without any consciousness or social interaction whatsoever—bolsters the argument that there is something about language itself, not just our brains, that is powerful.&#xA;&#xA;An Aside on Power Law Scaling&#xA;&#xA;One of the interesting features of human language is that it exhibits power scaling laws, as with other complex adaptive systems such as animals, cities, or businesses, as I recently examined in this post about Geoffrey West&#39;s fascinating book, Scale. The frequency of word usage, the length of sentences and texts, and the number of words in a language all follow power law distributions. This means that a small number of words are used frequently, while most words are used infrequently, and long sentences and texts are less common than shorter ones. As an interesting parallel, power law scaling is exhibited not only by language itself and through its generative manifestations in LLMs, but furthermore through the data--and the data centers and energy--required for training and using LLMs. Thus far, there is no apparent ceiling for LLM advancement in capability beyond that of the ceiling on the scalability of computer chips, data centers, and training data.&#xA;&#xA;Innate vs. Developed Language: A Review of Our Path Traversed Thus Far&#xA;&#xA;In our series “Innate vs. Developed”, we have explored the nature of language, challenging a widely held view that language is completely and innately hardwired in the human brain. Drawing upon “The Language Game” and &#34;Rethinking Innateness” as sources of inspiration, we have considered the notion that language is an emergent, culturally-evolved phenomenon that mounts atop an “inner scaffold” that exists within our brains and further refines and specializes our neural networks through simple repeated social interactions over time. &#xA;&#xA;We also considered how developing proficiency in reading and writing yet further extends and reinforces these channels across our brains – and how developing proficiency in multiple languages and literacies makes those networks even yet more robust.&#xA;&#xA;We went further afield and investigated Cormac McCarthy’s ponderings on a seeming division between language and the ancient parts of our brain that exist before and beyond language. We also investigated the paradoxical nature of language, in that it can both enhance and potentially occlude our connection to our unconscious selves and to our natural world.&#xA;&#xA;I promised at the end of the first post in this series that I would “maybe dig into the relation of cognition and language and literacy a little, and riff on the implications for AI, ANNs, and LLMs.” It’s taken me some time to let all of this ripen, especially given the rapid pace at which LLMs are developing. I think I’m finally starting to gain some perspective on LLMs that may allow me to indulge in a little riffing.&#xA;&#xA;Sources for Spelunking&#xA;&#xA;Before said indulgence in my next post, I’ll first outline a few sources I will draw upon at the outset so you can go off and explore on your own before being further biased by my own rambling.&#xA;&#xA;First, if you are interested in learning more about that analogy of a high dimensional semantic gradient and gaining insight into how LLMs kinda work, I recommend three sources shared by Ethan Mollick (he himself is also an excellent source):&#xA;&#xA;But what is a GPT? Visual intro to transformers&#xA;Large language models, explained with a minimum of math and jargon&#xA;What Is ChatGPT Doing … and Why Does It Work?&#xA;&#xA;Second, if you want to explore some interesting aspects of language itself that are related to LLMs, check out the following:&#xA;&#xA;Uniquely human intelligence arose from expanded information capacity&#xA;Stream of Search (SoS): Learning to Search in Language&#xA;The Structure of Meaning in Language: Parallel Narratives in Linear Algebra and Category Theory&#xA;&#xA;An Anticipation of Where We May Go From Here&#xA;&#xA;From these and other sources, including dabbling with Copilot and Claude and Gemini, I will ponder some of the following points on what computational neural networks may be able to tell us about language and what language may be able to tell us about LLMs – and, ultimately, perhaps, what this all may be able to tell us about teaching and learning:&#xA;&#xA;The surprisingly inseparable interconnection between form and meaning&#xA;Blundering our way to computational precision through human communication; Or, the generative tension between regularity and randomness&#xA;The human (and now, machine) capacity for learning and using language may simply be a matter of scale&#xA;Is language as separable from thought (and, for that matter, from the world) as Cormac McCarthy said? . . . which actually ended up becoming more about fuzziness and precision in language, but hey!&#xA;Implicit vs. explicit learning of language and literacy&#xA;&#xA;#language #literacy #LLMs #computation #statistical #learning #ai&#xA;]]&gt;</description>
      <content:encoded><![CDATA[<p><a href="https://www.readingrockets.org/classroom/classroom-strategies/semantic-gradients">“Semantic gradients,”</a> are a tool used by teachers to broaden and deepen students&#39; understanding of related words by plotting them in relation to one another. They often begin with antonyms at each end of the continuum. Here are two basic examples:</p>

<p><img src="https://i.snap.as/eXGuugvn.png" alt="Semantic gradient examples"/></p>

<p>Now imagine taking this approach and quantifying the relationships between words by adding numbers to the line graph. Now imagine adding another axis to this graph, so that words are plotted in a three dimensional space in their relationships. Then add another dimension, and another . . . heck, make it tens of thousands more dimensions, relating all the words available in your lexicon across a high dimensional space. . .</p>

<p>. . . and you may begin to envision one of the fundamental powers of Large Language Models (LLMs).</p>



<h2 id="llms-are-powered-by-language-or-words-as-a-vast-sea-of-interrelated-statistical-arrays-of-tokens" id="llms-are-powered-by-language-or-words-as-a-vast-sea-of-interrelated-statistical-arrays-of-tokens">LLMs Are Powered by Language: Or, Words as a Vast Sea of Interrelated Statistical Arrays of Tokens</h2>

<p>At root, the most powerful current forms of AI derive their capacities from decomposing human language into vast arrays of numbers based on their high dimensional statistical relationships and then predicting probabilistically what the next tokens are most likely to be.</p>

<p>There’s a kind of alchemical transformation that occurs that seems to maintain the meaning in the generative pronouncements of the frontier LLMs, all the more amazing because so far the very engineers who have designed the structure for these operations to occur do not fully understand what the models are doing to arrive at their seemingly oracular destinations.</p>

<p>In other words – the power of LLMs seemingly derives from the statistical power of language. There is something in the nature of language itself that seems to provide these computations of vast arrays of numbers with a lattice of our world, enabling LLMs to gain uncanny abilities from superpowered next word prediction. That LLMs have the generative powers they have—and that they have them without any consciousness or social interaction whatsoever—bolsters the argument that there is something about language itself, not just our brains, that is powerful.</p>

<h3 id="an-aside-on-power-law-scaling" id="an-aside-on-power-law-scaling">An Aside on Power Law Scaling</h3>

<p>One of the interesting features of human language is that it exhibits power scaling laws, as with other complex adaptive systems such as animals, cities, or businesses, as I recently examined in <a href="https://schoolecosystem.wordpress.com/2024/03/17/power-law-scaling-and-schools/">this post</a> about Geoffrey West&#39;s fascinating book, <em>Scale</em>. The frequency of word usage, the length of sentences and texts, and the number of words in a language all follow power law distributions. This means that a small number of words are used frequently, while most words are used infrequently, and long sentences and texts are less common than shorter ones. As an interesting parallel, power law scaling is exhibited not only by language itself and through its generative manifestations in LLMs, but furthermore through the data—and the data centers and energy—required for training and using LLMs. Thus far, there is no apparent ceiling for LLM advancement in capability beyond that of the ceiling on the scalability of computer chips, data centers, and training data.</p>

<h2 id="innate-vs-developed-language-a-review-of-our-path-traversed-thus-far" id="innate-vs-developed-language-a-review-of-our-path-traversed-thus-far">Innate vs. Developed Language: A Review of Our Path Traversed Thus Far</h2>

<p>In our series <a href="https://languageandliteracy.blog/innate-vs">“Innate vs. Developed”</a>, we have explored the nature of language, challenging a widely held view that language is completely and innately hardwired in the human brain. <a href="https://languageandliteracy.blog/language-like-reading-may-not-be-innate">Drawing upon “The Language Game” and “Rethinking Innateness”</a> as sources of inspiration, we have considered the notion that language is an emergent, culturally-evolved phenomenon that mounts atop <a href="https://languageandliteracy.blog/the-inner-scaffold-for-language-and-literacy">an “inner scaffold”</a> that exists within our brains and further refines and specializes our neural networks through simple repeated social interactions over time.</p>

<p>We also considered how developing proficiency in reading and writing yet further extends and reinforces these channels across our brains – and how developing proficiency in multiple languages and literacies makes those networks <a href="https://languageandliteracy.blog/accelerating-the-inner-scaffold-across-modalities-and-languages">even yet more robust</a>.</p>

<p>We went further afield and <a href="https://languageandliteracy.blog/thinking-inside-and-outside-of-language">investigated Cormac McCarthy’s ponderings</a> on a seeming division between language and the ancient parts of our brain that exist before and beyond language. We also investigated the paradoxical nature of language, in that it can both enhance and <a href="https://languageandliteracy.blog/speaking-ourselves-into-being-and-others-into-silence-the-power-of-language">potentially occlude</a> our connection to our unconscious selves and to our natural world.</p>

<p>I promised at the end of <a href="https://languageandliteracy.blog/language-like-reading-may-not-be-innate">the first post in this series</a> that I would “maybe dig into the relation of cognition and language and literacy a little, and riff on the implications for AI, ANNs, and LLMs.” It’s taken me some time to let all of this ripen, especially given the rapid pace at which LLMs are developing. I think I’m finally starting to gain some perspective on LLMs that may allow me to indulge in a little riffing.</p>

<h2 id="sources-for-spelunking" id="sources-for-spelunking">Sources for Spelunking</h2>

<p>Before said indulgence in my next post, I’ll first outline a few sources I will draw upon at the outset so you can go off and explore on your own before being further biased by my own rambling.</p>

<p>First, if you are interested in learning more about that analogy of a high dimensional semantic gradient and gaining insight into how LLMs kinda work, I recommend three sources <a href="https://x.com/emollick/status/1775355910761681177">shared by Ethan Mollick</a> (he himself is also <a href="https://www.oneusefulthing.org/">an excellent source</a>):</p>
<ul><li><a href="https://www.youtube.com/watch?v=wjZofJX0v4M">But what is a GPT? Visual intro to transformers</a></li>
<li><a href="https://www.understandingai.org/p/large-language-models-explained-with">Large language models, explained with a minimum of math and jargon</a></li>
<li><a href="https://writings.stephenwolfram.com/2023/02/what-is-chatgpt-doing-and-why-does-it-work/">What Is ChatGPT Doing … and Why Does It Work?</a></li></ul>

<p>Second, if you want to explore some interesting aspects of language itself that are related to LLMs, check out the following:</p>
<ul><li><a href="https://www.nature.com/articles/s44159-024-00283-3.epdf?sharing_token=dc9WtYt3C_FN2N5q5mmKatRgN0jAjWel9jnR3ZoTv0PIvBIKEnJUrpLA70zYn0mjSaDkgiBUb43hOoUEou9xdgynS0nAWob7QAH5X7gROQMoz5n9acglkBUa_86OzUA1B-Wg9_p5hHRLFUQ95SWsfFXtU8jHuxKnM8_fWZKCoAA%3D">Uniquely human intelligence arose from expanded information capacity</a></li>
<li><a href="https://arxiv.org/abs/2404.03683">Stream of Search (SoS): Learning to Search in Language</a></li>
<li><a href="http://ams.org/journals/notices/202402/rnoti-p174.pdf">The Structure of Meaning in Language: Parallel Narratives in Linear Algebra and Category Theory</a></li></ul>

<h2 id="an-anticipation-of-where-we-may-go-from-here" id="an-anticipation-of-where-we-may-go-from-here">An Anticipation of Where We May Go From Here</h2>

<p>From these and other sources, including dabbling with Copilot and Claude and Gemini, I will ponder some of the following points on what computational neural networks may be able to tell us about language and what language may be able to tell us about LLMs – and, ultimately, perhaps, what this all may be able to tell us about teaching and learning:</p>
<ul><li><a href="https://languageandliteracy.blog/the-algebra-of-language-unveiling-the-statistical-tapestry-of-form-and-meaning">The surprisingly inseparable interconnection between form and meaning</a></li>
<li><a href="https://languageandliteracy.blog/the-pathway-of-human-language-towards-computational-precision-in-llms">Blundering our way to computational precision through human communication; Or, the generative tension between regularity and randomness</a></li>
<li><a href="https://write.as/manderson/scaling-our-capacity-for-processing-information">The human (and now, machine) capacity for learning and using language may simply be a matter of scale</a></li>
<li><a href="https://write.as/manderson/the-interplay-of-language-cognition-and-llms-where-fuzziness-meets-precision">Is language as separable from thought (and, for that matter, from the world) as Cormac McCarthy said?</a> <em>. . . which actually ended up becoming more about fuzziness and precision in language, but hey!</em></li>
<li><a href="https://write.as/manderson/llms-statistical-learning-and-explicit-teaching">Implicit vs. explicit learning of language and literacy</a></li></ul>

<p><a href="https://languageandliteracy.blog/tag:language" class="hashtag"><span>#</span><span class="p-category">language</span></a> <a href="https://languageandliteracy.blog/tag:literacy" class="hashtag"><span>#</span><span class="p-category">literacy</span></a> <a href="https://languageandliteracy.blog/tag:LLMs" class="hashtag"><span>#</span><span class="p-category">LLMs</span></a> <a href="https://languageandliteracy.blog/tag:computation" class="hashtag"><span>#</span><span class="p-category">computation</span></a> <a href="https://languageandliteracy.blog/tag:statistical" class="hashtag"><span>#</span><span class="p-category">statistical</span></a> <a href="https://languageandliteracy.blog/tag:learning" class="hashtag"><span>#</span><span class="p-category">learning</span></a> <a href="https://languageandliteracy.blog/tag:ai" class="hashtag"><span>#</span><span class="p-category">ai</span></a></p>
]]></content:encoded>
      <guid>https://languageandliteracy.blog/language-and-llms</guid>
      <pubDate>Tue, 23 Apr 2024 14:48:41 +0000</pubDate>
    </item>
  </channel>
</rss>