Back
other Jan 5, 2024

Cognitive Science of Learning: How the Brain Works

Cognition involves the flow of information through sensory, working, and long-term memory banks in the brain. Sensory memory temporarily holds raw data, working memory manipulates and organizes information, and long-term memory stores it indefinitely by creating strategic electrical wiring between neurons. Learning amounts to increasing the quantity, depth, retrievability, and generalizability of concepts and skills in a student’s long-term memory. Limited working memory capacity creates a bottleneck in the transfer of information into long-term memory, but cognitive learning strategies can be used to mitigate the effects of this bottleneck.

by Justin Skycak (@justinskycak) justinmath.com 3,251 words
View original

Cognition involves the flow of information through sensory, working, and long-term memory banks in the brain. Sensory memory temporarily holds raw data, working memory manipulates and organizes information, and long-term memory stores it indefinitely by creating strategic electrical wiring between neurons. Learning amounts to increasing the quantity, depth, retrievability, and generalizability of concepts and skills in a student’s long-term memory. Limited working memory capacity creates a bottleneck in the transfer of information into long-term memory, but cognitive learning strategies can be used to mitigate the effects of this bottleneck.

This post is part of the book The Math Academy Way (Working Draft, Jan 2024). Suggested citation: Skycak, J., advised by Roberts, J. (2024). Cognitive Science of Learning: How the Brain Works. In The Math Academy Way (Working Draft, Jan 2024). https://justinmath.com/cognitive-science-of-learning-how-the-brain-works/

Want to get notified about new posts? Join the mailing list and follow on X/Twitter.


At the most fundamental level, learning is the creation of strategic electrical wiring between neurons (“brain cells”) that improves the brain’s ability to perform a task.

When the brain thinks about objects, concepts, associations, etc, it represents these things by activating different patterns of neurons with electrical impulses. Whenever a neuron is activated with electrical impulses, the impulses naturally travel through its outward connections to reach other neurons, potentially causing those other neurons to activate as well. By creating strategic connections between neurons, the brain can more easily, quickly, accurately, and reliably activate more intricate patterns of neurons.

As one might expect, it is extraordinarily complicated to understand what these specific brain patterns are, how they interact, and how the brain identifies strategic ways to improve its connectivity. However, to some extent, these are just nature’s way of implementing cognition – and the overarching cognitive processes of the brain are much better understood.

Sensory, Working, and Long-Term Memory

At a high level, human cognition is characterized by the flow of information across three memory banks:

  1. Sensory memory temporarily holds a large amount of raw data observed through the senses (sight, hearing, taste, smell, and touch), only for several seconds at most, while relevant data is transferred to short-term memory for more sophisticated processing.
  2. Short-term memory, and more generally, working memory, has a much lower capacity than sensory memory, but it can store the information about ten times longer. Working memory consists of short-term memory along with capabilities for organizing, manipulating, and generally “working” with the information stored in short-term memory. The brain’s working memory capacity represents the degree to which it can focus activation on relevant neural patterns and persistently maintain their simultaneous activation, a process known as rehearsal.
  3. Long-term memory effortlessly holds indefinitely many facts, experiences, concepts, and procedures, for indefinitely long, in the form of strategic electrical wiring between neurons. Wiring induces a “domino effect” by which entire patterns of neurons are automatically activated as a result of initially activating a much smaller number of neurons in the pattern. The process of storing new information in long-term memory is known as consolidation. At a cognitive level, learning can be described as a positive change in long-term memory.

These memory banks work together to form the following pipeline for processing information:

  1. Sensory memory receives a stimulus from the environment and passes on important details to working memory.
  2. Working memory holds and manipulates those details, often augmenting or substituting them with related information that was previously stored in long-term memory.
  3. Long-term memory curates important information as though it were writing a “reference book” for the working memory.

Note, however, that there is a crucial conceptual difference between long-term memory and a reference textbook: long-term memory can be forgotten. The text in a reference book remains there forever, accessible as always, regardless of whether you read it – but the representations in long-term memory gradually, over time, become harder to retrieve if they are not used, resulting in forgetting. The phenomenon of forgetting in long-term memory has been widely researched and can be characterized as follows (Hardt, Nader, & Nadel, 2013):

However, the lower-level mechanisms underlying forgetting in long-term memory are not yet well understood.

image

There are two complementary perspectives by which we can think about this pipeline.

In the context of mathematical talent development, once a student is beyond the stage of learning how to read and count, we are less concerned with their sensory memory and more concerned with their long-term memory. The goal of instruction is to increase the quantity, depth, retrievability, and generalizability of mathematical concepts and skills in the student’s long-term memory.

The student’s working memory capacity is a bottleneck in the transfer of information into their long-term memory. However, by leveraging cognitive learning strategies and properly scaffolding and adapting instruction to the student’s individual needs, we can minimize the degree to which their working memory capacity limits their learning, thereby maximizing the transfer of new information and the retention of existing information in long-term memory.

Design Constraints

The brain’s information-processing pipeline is designed to be incredibly efficient. However, even the most efficient designs have limitations. Design is all about balancing trade-offs to achieve the best possible outcome in the face of constraints. To understand the constraints and the rationale behind a design, it can be helpful to attempt some naive critiques.

Critique: Why is long-term memory needed? Why can’t the brain just hold everything in working memory forever through rehearsal?

Rationale: Rehearsal requires a lot of effort. It is very taxing on the brain. When the brain engages in rehearsal, it’s like a muscle that is lifting a weight.

Just like a muscle has a limit to the amount of weight it can hold, the brain has a limit to the amount of new information it can hold in working memory via rehearsal. Most people can only hold about 7 digits (or more generally 4 chunks of coherently grouped items) simultaneously and only for about 20 seconds (Miller, 1956; Cowan, 2001; Brown, 1958). And that assumes they aren’t needing to perform any mental manipulation of those items – if they do, then fewer items can be held due to competition for limited processing resources (Wright, 1981).

Long-term memory solves this problem by providing a place where the brain can store lots of information for a long time without requiring much effort.

Critique: Why doesn’t the brain just store everything it encounters in long-term memory? That way, it would never forget anything.

Rationale: When it comes to information storage, more is not always better. In order for it to be worthwhile to store a piece of information, the benefit must offset the cost. Creating connections between neurons is costly in the sense that it requires biological resources – the connections are physical growths between cells, which means they have to be actively constructed and maintained by the body.

To illustrate with a concrete example, suppose that you want to buy a biography book that will help you understand somebody’s background and their impact on society. One book contains 300 pages, costs $20, and covers formative experiences in their childhood, their career arc, and occasional anecdotes to illustrate key points and themes. Another book contains 10,000 pages, costs $1,000, covers all of the information in the first book, and also includes a description of every single meal the person ate throughout their life. Unless you have a specific, intense interest in this person’s dietary habits (which you probably don’t), it’s easy to see that the first option is superior.

Case Study: Information Flow During a Computation

To illustrate how information flows through these memory banks when solving a math problem, let’s analyze what happens as we compute $4^3$ using typical arithmetic strategies while writing down some intermediate steps. (Remember that exponentiation is just repeated multiplication: $4^3$ means to take three 4’s and multiply them together, that is, $4^3$ = 4 × 4 × 4 = 64.)

First, let’s get a sense of how each memory bank will help us solve the problem:

  1. Sensory memory will capture visual data that lets us read the problem or any intermediate work that we’ve written down, thereby allowing the written information to be loaded into working memory. It will also filter out any distractions (e.g. background noise) as we solve the problem.
  2. Working memory will hold the relevant pieces of the problem, request additional information from long-term memory, and apply that information to incrementally transform the pieces of the problem into the solution. Our problem-solving narrative will take place within the working memory.
  3. Long-term memory will, upon request from working memory, produce definitions, facts, and procedures that we learned previously. It is like an internal “reference book” that we can use to look up additional information that would be helpful while solving the current problem.

It’s worth re-emphasizing that the problem-solving narrative will take place within the working memory. Sensory and long-term memory will supply working memory with information, which working memory will combine, transform, and use to guide our behavior to solve the problem. As researchers elaborate (Roth & Courtney, 2007):

Now, let’s walk through the specific steps needed to solve the problem while observing what happens in each memory bank.

image
image
image
image
image
image
image
image

What learning resulted from this computation? Remember that learning occurs when the wiring of long-term memory is changed in a positive way that increases a student’s ability to perform a task. This can involve any combination of wiring up new information, wiring up connections between existing pieces of information, reorganizing existing wiring so that the information can be retrieved more efficiently, etc.

With this in mind, let’s take inventory of the processes that occurred within long-term memory in the example above:

All of these pieces of information will become further consolidated in long-term memory, and there will be additional wiring connecting these component skills as part of a larger procedure for computing exponents.

Additionally, the fact $4^3 = 64$ will also begin consolidating in long-term memory (though it will soon be forgotten unless it is repeatedly reviewed into the future). Indeed, many people who frequently perform mental math with exponents know their cubes from $1^3$ to $6^3$ by heart and can simply retrieve their values as opposed to computing them.

Neuroscience of Working Memory

Recall that when the brain thinks about objects, concepts, associations, etc, it represents these things by activating different patterns of neurons with electrical impulses. Loosely speaking, the brain’s working memory capacity represents the degree to which it can focus activation on relevant neural patterns and persistently maintain their simultaneous activation. The cognitive load of a task represents the level of exertion that the brain would experience while completing the task.

(More strictly speaking: higher working memory capacity refers not to the ability to sustain more neural activity in the energy sense, but rather, the ability to sustain relevant neural activity while suppressing interference from irrelevant neural activity. At a biological level, hitting a working memory capacity limit does not entail exhausting one’s ability to maintain more neural activity, but rather exhausting one’s ability to maintain focus and attention, that is, appropriate concentration or allocation of one’s neural activity.)

As summarized by D’Esposito (2007):

Long-term learning is represented by the creation of strategic electrical wiring between neurons. Whenever a neuron is activated with electrical impulses, the impulses naturally travel through its outward connections to reach other neurons, potentially causing those other neurons to activate as well. By creating strategic connections between neurons, the brain can more easily, quickly, accurately, and reliably activate more intricate patterns of neurons.

Talcott (2021) summarizes this process as follows:

Wiring induces a “domino effect” by which entire patterns of neurons are automatically activated as a result of initially activating a much smaller number of neurons in the pattern. However, when the brain is initially learning something, the corresponding neural pattern has not been “wired up” yet, which means that the brain has to devote effort to activating each neuron in the pattern. In other words, because the dominos have not been set up yet, each one has to be toppled in a separate stroke of effort. This imposes severe limitations on how much new information the brain can hold simultaneously in working memory.

References

Brown, J. (1958). Some tests of the decay theory of immediate memory. Quarterly journal of experimental psychology, 10(1), 12-21.

Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and brain sciences, 24(1), 87-114.

D’Esposito, M. (2007). From cognitive to neural models of working memory. Philosophical Transactions of the Royal Society B: Biological Sciences, 362 (1481), 761-772.

Hardt, O., Nader, K., & Nadel, L. (2013). Decay happens: the role of active forgetting in memory. Trends in cognitive sciences, 17(3), 111-120.

Miller, G. A. (1956). The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychological review, 63 (2), 81.

Roth, J. K., & Courtney, S. M. (2007). Neural system for updating object working memory from different sources: sensory stimuli or long-term memory. Neuroimage, 38 (3), 617-630.

Talcott, J. B. (2021). The neurodevelopmental underpinnings of children’s learning: Connectivity is key.

Wright, R. E. (1981). Aging, divided attention, and processing capacity. Journal of Gerontology, 36 (5), 605-614.


This post is part of the book The Math Academy Way (Working Draft, Jan 2024). Suggested citation: Skycak, J., advised by Roberts, J. (2024). Cognitive Science of Learning: How the Brain Works. In The Math Academy Way (Working Draft, Jan 2024). https://justinmath.com/cognitive-science-of-learning-how-the-brain-works/

Want to get notified about new posts? Join the mailing list and follow on X/Twitter.