While my bandwidth to peruse research has diminished this year (work has been busy, and I like spending time with my children) I have still encountered a fair number of compelling studies. In keeping with the tradition begun in 2023, and building on last year’s review, I am endeavoring to round up the research that has crossed my radar over the last 12 months.
This year presents a difficult juncture for research. Political aggression against academic institutions, the immigrants who power their PhD programs, and the federal contracts essential to their survival has disrupted research. Despite this, strong research continues to be published. Because research is a slow-moving endeavor, I suspect the full effects of these disruptions will manifest increasingly in future roundups; for now, the good work persists.
The research landscape of 2025 highlights a continued shift toward experience-dependent plasticity. This view treats the human mind as a dynamic ecosystem shaped by biological rhythms, cultural “software,” and technological catalysts. Learning is no longer seen as a linear accumulation of skills, but as a sophisticated orchestration of “statistical” internal models and external social and cultural and technological attunements.
Longtime readers will recognize this “ecosystem” view from my other blog on Schools as Ecosystems. It is validating to see the field increasingly adopting this ecological lens—viewing the learner not as an isolated machine, but as an organism deeply embedded in a biological and cultural context.
Our “big buckets” for this year have ended up mirroring the 2024 roundup, which means, methinks, that we have settled upon a perennial organizational structure:
The Science of Reading and Writing
Content Knowledge as an Anchor to Literacy
Studies on Language Development
Multilinguals and Multilingualism
Rhythm, Attention, and Memory
School, Social-Emotional, and Contextual Effects
The Frontier of Artificial Intelligence and Neural Modeling
In the typical Hollywood action movie, a hero acquires master-level skill in a specialized art, such as Kung Fu, in a few power ballad-backed minutes of a training montage.
In real life, it may seem self-evident that gaining mastery takes years of intense, deliberate, and guided work. Yet the perennial optimism of students cramming the night before an exam tells us that the pursuit of a cognitive shortcut may be an enduring human impulse.
It is unsurprising, then, that students—and many adults—increasingly use the swiftly advancing tools of AI and Large Language Models (LLMs) as a shortcut around deeper, more effortful cognitive work.
Learning new information in L2 is more effortful than in L1. We found different functional connectivity networks of naturalistic learning through speech among adolescents, confirming this prevalent observation
Does learning language require effort? Does it require more effort when learning a new language later in our lives? Why?
Today, we will highlight a study that shows the additional neurological networks that adolescents activate when learning in a second language – a key insight for all educators to consider.
Language Learning: Effortless for Babies, Effortful for Adults
Babies learn language with such ease that they have already begun to recognize the unique patterns of a language–even to distinguish between the unique patterns of multiple languages–while still in the womb.
We therefore tend to assume there is something wholly innate or natural to learning language.
Yet as we’ve explored previously in a series on this blog, even learning our first languages may not be as innate or natural as it can appear. Human language reflects a unique synchrony between our biological and cultural evolution, finely attuned to the social environment in which we interact.
When I typically begin a series of blogs to conduct nerdy inquiry into an abstract topic, I don't generally know where I'm going to end up. This series on LLMs was unusual in that in our first post, I outlined pretty much the exact topics I would go on to cover.
Here's where I had spitballed we might go:
The surprisingly inseparable interconnection between form and meaning
Blundering our way to computational precision through human communication; Or, the generative tension between regularity and randomness
The human (and now, machine) capacity for learning and using language may simply be a matter of scale
Is language as separable from thought (and, for that matter, from the world) as Cormac McCarthy said?
Implicit vs. explicit learning of language and literacy
Indeed, we then went on to explore each of these areas, in that order. Cool!
In a previous series, “Innate vs. Developed,” we’ve also challenged the idea that language is entirely hardwired in our brains, highlighting the tension between our more recent linguistic innovations and our more ancient brain structures. Cormac McCarthy, the famed author of some of the most powerful literature ever written, did some fascinating pontificating on this very issue.
In this post, we’ll continue picking away at these tensions, considering implications for AI and LLMs.
“Over cultural evolution, the human species was so pressured for increased information capacity that they invented writing, a revolutionary leap forward in the development of our species that enables information capacity to be externalized, frees up internal processing and affords the development of more complex concepts. In other words, writing enabled humans to think more abstractly and logically by increasing information capacity. Today, humans have gone to even greater lengths: the Internet, computers and smartphones are testaments to the substantial pressure humans currently face — and probably faced in the past — to increase information capacity.”
According to the perspectives of the authors in the paper quoted above, the capacity to process and manage vast quantities of information is a defining characteristic of human intelligence. This ability has been extended over time through the development of tools and techniques for externalizing information, such as via language, writing, and digital technology. These advancements have, in turn, allowed for increasingly abstract and complex thought and technologies.
The paper by Jessica Cantlon & Steven Piantadosi further proposes that the power of scaling is what lies behind human intelligence, and that this power of scaling is what further lies behind the remarkable results achieved by artificial neural networks in areas such as speech recognition, LLMs, and computer vision, and that these accomplishments have not been achieved through specialized representations and domain-specific development, but rather through the use of simpler techniques combined with increased computational power and data capacity.
Regularity and irregularity. Decodable and tricky words. Learnability and surprisal. Predictability and randomness. Low entropy and high entropy.
Why do such tensions exist in human language? And in our AI tools developed to both create code and use natural language, how can the precision required for computation co-exist alongside this necessary complexity and messiness of our human language?
”. . . the fact, as suggested by these findings, that semantic properties can be extracted from the formal manipulation of pure syntactic properties – that meaning can emerge from pure form – is undoubtedly one of the most stimulating ideas of our time.”
In our last post, we began exploring what Large Language Models (LLMs) and their uncanny abilities might tell us about language itself. I posited that the power of LLMs stems from the statistical nature of language.
In a previous post, Thinking Inside and Outside of Language, we channelled Cormac McCarthy and explored the tension between language and cognition. We dug in even further and considered Plato's long ago fears of the deceptive and distancing power of written language in Speaking Ourselves into Being and Others into Silence: The Power of Language, and how bringing a critical consciousness to our use of language could temper unconscious biases and power dynamics.
If you find any of that interesting, I recommend reading this short interview, How to Quiet Your Mind Chatter in Nautilus Magazine with Ethan Kross, an experimental psychologist and neuroscientist at the University of Michigan.
Two relevant quotes:
“What we’ve learned is that language provides us with a tool for coaching ourselves through our problems like we were talking to another person. It involves using your name and other non-first person pronouns, like “you” or “he” or “she.” That’s distanced self-talk.”
“The message behind mindfulness is sometimes taken too far in the sense of 'you should always be in the moment.' The human mind didn’t evolve to always be in the moment, and we can derive enormous benefit from traveling in time, thinking about the past and future.”