Abstract
Metacognitive calibration—the capacity to accurately self-assess one’s performance—forms the basis for error detection and self-monitoring and is a potential catalyst for conceptual change. Limited brain imaging research on authentic learning tasks implicates the lateral prefrontal and anterior cingulate brain regions in expert scientific reasoning. This study aimed to determine how variation in undergraduate life sciences students’ metacognitive calibration relates to their brain activity when evaluating the accuracy of biological models. Fifty undergraduate students enrolled in an introductory life sciences course completed a biology model error detection task during fMRI. Students with higher metacognitive calibration recruited lateral prefrontal regions linked in prior research to expert STEM reasoning to a greater extent than those with lower metacognitive calibration. Findings suggest that metacognition relates to important individual differences in undergraduate students’ use of neural resources during an authentic educational task and underscore the importance of fostering metacognitive calibration in the classroom.
Similar content being viewed by others
Introduction
Self-regulated learning is the extent to which students are engaged, motivated, and behaviorally active in their learning1. Self-regulated learners accomplish tasks or goals by continuously monitoring and correcting the effectiveness of their strategies and past behaviors to better control future learning2. Metacognitive calibration, the match between a student’s objective performance and their subjective self-assessment of that performance3,4, is fundamental to self-regulated learning because it forms the basis for performance monitoring5. Students with more accurate calibration presumably have more reliable information on which to base their deployment of effortful, strategic, cognitive resources to optimize task performance6. Metacognitive calibration has been linked to dorsomedial prefrontal systems thought to be involved in monitoring and detecting errors in performance, as well as to lateral PFC regions linked to effortful cognitive control7. However, neuroimaging studies of metacognitive calibration generally have used experimental tasks rather than tasks reflective of authentic educational experiences, leaving a gap in understanding how metacognition relates to students’ neural processes in classroom-relevant learning. Modeling and model evaluation are authentic tasks in scientific practice and education8, and expert scientists are skilled at detecting inaccuracies in models9. This study aimed to determine the relation of metacognitive calibration to student performance and neural activity when evaluating biological models.
Although a general definition of metacognition is “thinking about thinking”10, contemporary conceptualizations typically divide it into two distinct processes: knowledge about cognition, including self-monitoring, and self-regulatory mechanisms10,11,12. Self-monitoring involves knowing how well one is performing and recognizing the likelihood of accuracy or inaccuracy in one’s judgments or behaviors. Conversely, metacognitive self-regulation concerns the process of organizing one’s cognition, planning, being aware of one’s comprehension, and evaluating the efficacy of strategies during task performance13,14. Learners who develop high self-monitoring and self-regulation skills revise and reconstruct concepts to advance their knowledge and achieve high self-efficacy, persistence, and self-discipline in performing tasks15,16,17. Well-calibrated self-assessments also have been linked to students’ use of more effective study strategies and to higher levels of academic growth over time18,19,20.
Student ratings of confidence in their own performance offer one means of assessing self-awareness and calibration21. For well-calibrated learners, confidence ratings should correspond with actual performance22. That is, students should have high confidence when they are accurate and low confidence when not. Although confidence estimates are higher for correct relative to incorrect trials23, lower-performing students tend to be more overconfident in their performance predictions, while higher-performing learners are accurate or underconfident in their self-assessments12,24,25. Additionally, learners tend to be overconfident when evaluating complex or difficult information and underconfident about relatively easier information26.
Metacognition has repeatedly been linked to the prefrontal cortex (PFC), with greater activity in the dorsolateral and anterior medial PFC being associated with higher levels of self-awareness and metacognitive accuracy27,28,29. Several metacognition studies highlight the ventromedial PFC and posterior medial frontal cortex as central brain areas related to confidence estimates30,31 and error detection processes32,33,34, while both the frontopolar cortex and the lateral PFC are involved in explicit metacognitive judgments23,35, metacognitive control, and subsequent behavioral regulation36,37. More generally, areas in the dorsal anterior cingulate cortex (ACC) are thought to monitor for conflict or errors in performance, whereas the lateral PFC presumably uses these inputs to bias behavior toward more adaptive cognitive strategies7,35,38. From these studies, it is reasonable to hypothesize that students with higher levels of metacognitive calibration will show higher levels of activity in medial and lateral PFC regions during academic tasks, reflecting optimal use of self-monitoring and regulatory processes to guide performance. However, given that these studies have often involved experimental tasks, the extent to which these same neural regions are linked to metacognition during more authentic educational tasks remains unclear.
Model-based reasoning, a core emphasis area in STEM education, offers a particularly relevant context for understanding how metacognitive calibration relates to students’ brain activity during authentic learning experiences39. From simple flowcharts and graphs to complex computer simulations, models pervade all areas of science and are how scientists reason, evaluate hypotheses, and convey ideas40,41. Undergraduate biology students encounter models through textbooks, lectures, note-taking, and classroom activities that convey foundational introductory principles regarding genetic, ecological, and cellular systems42. Each time they encounter a model, students must determine whether their prior knowledge of the topic is in agreement with the observed model. Measuring biology students’ capacity to detect errors in models constitutes a powerful means to assess the relation of their metacognition to their performance of this critical scientific skill.
Linking metacognitive calibration to model error detection also offers the potential for understanding how to elicit conceptual change to ultimately foster a deeper understanding of fundamental scientific concepts43. To learn scientific concepts, students must be able to actively refine their learning process and integrate new concepts and relationships with their prior knowledge44. Learners must first become dissatisfied with their existing conceptions, embracing new conceptions as plausible45. However, conceptual alterations prove difficult when learners hold misconceptions46,47 because these errors continue to be recalled and reinforced until scientifically accurate knowledge outcompetes them48. Absent or poor metacognition may predict the extent to which learners both ignore new information and resist changing their minds, even when new information indicates errors in their original beliefs27. For learners who hold misconceptions, conceptual change is more predictable when an alternative scientific viewpoint initiates cognitive conflict, allowing error detection in the misconception49.
Metacognitive calibration may be a critical initial step in the process of conceptual change50. Students who are more metacognitively aware of misconceptions about a given topic are more likely to detect and remedy those misconceptions to develop a deeper understanding of a concept or topic, decreasing the likelihood of entrenching misconceptions51. Conversely, students who are unaware of their misconceptions may become overly secure in those misconceptions (i.e., they do not know what they do not know)51,52 and may consequently miss opportunities for knowledge revision53. Evaluating how students’ metacognitive calibration links to their error detection and ultimate learning may offer clues as to how metacognition supports conceptual change.
Limited literature on the neural regions involved in expert science cognition highlights its association with prefrontal regions linked to error detection, conflict monitoring, and inhibition54. Additionally, experts evaluating the accuracy of scientific circuits activate anterior cingulate cortex (ACC), ventrolateral PFC (VLPFC), and dorsolateral PFC (DLPFC) to a greater extent than novices, who tend to activate more limited prefrontal regions55,56. These brain regions are thought to work in concert to support scientific reasoning. The lateral PFC provides the cognitive control necessary for logical thinking, flexible reasoning, and inhibition of inappropriate responses or attentional stimuli32,57; the ventromedial PFC integrates emotional and cognitive factors to guide decision-making56,58; and the ACC contributes to inhibitory processes and error monitoring, helping individuals stay on track and correct or adjust their reasoning when necessary59.
With respect to metacognition, novice students exhibited greater dorsomedial frontal activity when they were confident vs. not confident in their responses to physics electric circuit diagrams60. Taken together, these studies implicate ACC and lateral PFC regions in the expert processing and resolution of science misconceptions. Assuming that students with higher metacognitive calibration are better equipped to engage in ‘expert’-like error detection and associated engagement of control, we expected such students to use these regions to a greater extent than their less calibrated peers when evaluating biological models.
Conceptual change theories identify metacognitive calibration as a critical process that enables students to detect and resolve gaps in their knowledge. However, neuroimaging studies typically examine the neural bases of metacognitive processing and calibration accuracy in the context of novel or self-reflection-based tasks, rather than educationally authentic tasks like those involving scientific reasoning61,62. Further understanding of the neural bases of metacognitive monitoring during an authentic task of model error detection will offer new insights into the mechanisms by which this construct supports students’ development as scientists. The limited neuroimaging research on scientific cognition indicates that ‘expert’ scientists activate DLPFC, VLPFC, and ACC brain regions to a greater extent than ‘novice’ undergraduates when reasoning about scientific phenomena, particularly when accurate reasoning involves detecting and inhibiting misconception55,63. The current study’s goal was to examine the relation of students’ metacognitive calibration to their neural activity during a biological model evaluation task. Specifically, we aimed to determine how individual differences in students’ calibration accuracy64,65 related to their activity in lateral prefrontal and ACC regions linked both to metacognition and to scientific expertise (see “Methods” section). Based on the theory that accurate calibration serves as a foundation for error detection and conceptual change, we hypothesized that students with higher metacognitive calibration would show higher levels of activity in prefrontal regions linked to these scientific processes.
Results
Behavioral results
Table 1 presents descriptive statistics and correlations among the different behavioral variables. Students on average were correct on 65% of the model evaluation task trials, whereas they were confident ~75% of the time. Figure 1 illustrates the significant difference in student confidence levels for accurate vs. inaccurate trials of the model evaluation task (t(49) = 7.72, p < 0.001, d = 1.09), although students still responded that they were confident more often than not for trials where they made an inaccurate response. The moderate correlation between total accurate responses and total confident responses (r = 0.403, p = 0.004) indicated that as participants’ accuracy increased, their overall confidence levels also increased (for further behavioral findings highlighting the importance of self-monitoring sees1,2).
The different metrics for metacognitive calibration, sensitivity, and efficiency, as calculated from students’ confidence ratings and accuracy in the fMRI model evaluation task, were robustly inter-correlated (p’s < 0.01). Neither student’s ϕ scores nor their metacognitive efficiency scores correlated with their model evaluation accuracy, although meta-d’ scores did, as expected (p < 0.001). Final course grade in the introductory life science course correlated slightly with model evaluation task accuracy (p = 0.008), but not with confidence, p = 0.39. Measures of metacognition showed minimal correlations with final course grades, except for meta-d’, p = 0.05.
Neuroimaging results from the model evaluation task
Overall patterns of brain activity for confident and non-confident responses
Table 2 and Fig. 2 show that during trials where students indicated they were Confident > Not confident, widespread neural activation occurred across occipital regions, and in the parietal cortex and lateral and medial PFC (see Supplementary materials for the same analysis using a more conservative statistical threshold). We performed a further contrast of Error > No error models only for trials where students were confident, which indicated that students showed increased activity in a medial frontal cluster, in bilateral inferior prefrontal/insular regions, and in the lingual gyrus (Table 2, Supplementary Fig. 2). Finally, we contrasted trials where students were confident and accurate in their response > confident but inaccurate (i.e., overconfident) in their response. This analysis revealed that students exhibited more activity in a single cluster in the left lingual gyrus when they were confident and accurate relative to confident and inaccurate (Table 2, Supplementary Fig. 3). Overall when examined by trial type, students exhibited greater activity in lateral and medial PFC regions when they were confident, although PFC activity did not diverge for trials where students were overconfident vs. calibrated in their responses.
Relation of individual differences in student confidence to neural response patterns
Focusing on the Error > No error model contrast, we evaluated the relation of overall accuracy and confidence to students’ neural responses by including these variables as covariates of interest in the group design matrix. There was no association of students’ model evaluation accuracy with their neural response patterns for this contrast. Likewise, students’ overall confidence was not linked to their brain activity. However, students with lower confidence bias showed more activity in two clusters centered in the left inferior frontal gyrus (Table 3, Fig. 3).
Relation of student metacognitive calibration to neural response patterns
Controlling for accuracy on the model evaluation task, students’ ϕ values were positively associated with and their level of activity in the right middle frontal gyrus for the Error > No error model contrast (see Table 3). Students with higher ϕ scores showed a greater increase in activity in this cluster, which spanned BA 8 and 9 (Fig. 4). Comparable results emerged when we instead used the metacognitive efficiency score (i.e., d’–meta d’), which was positively linked to activation in a right middle frontal cluster that overlapped with the cluster identified for ϕ. Controlling for accuracy, students’ meta-d’ scores were associated with activity in the left inferior parietal, superior occipital and postcentral gyri (Fig. 5).
Discussion
Metacognition is central to science education: it is critical to conceptual change and to cultivating a deep understanding of scientific concepts66. Specifically, in biology education, students with a higher awareness of the learning process and a stronger ability to monitor, regulate, and control learning manifest a more meaningful understanding of targeted biology concepts and better scientific inquiry skills67,68. The current study evaluated the relation of students’ metacognitive calibration to their neural activity when evaluating errors in biology models. A major finding is that students with higher metacognitive calibration and efficiency recruit lateral prefrontal regions to a greater extent than their peers with lower metacognitive calibration when evaluating error-containing models. This finding has implications for understanding the role of metacognitive monitoring in students’ learning behavior and for approaches to STEM instruction.
Controlling for task accuracy, higher metacognitive calibration, as measured using phi, was linked to higher levels of activity in the right dorsolateral PFC. The same effect was when we instead used students’ metacognitive efficiency scores, which more rigorously parse metacognitive calibration from confidence response biases and task performance65. Those few neuroimaging studies that have focused on STEM learning have linked the recruitment of lateral prefrontal and ACC brain regions to ‘expert’ scientific reasoning, with the lateral PFC being especially linked to accurate error detection within STEM experts’ domains of expertise9,59. Therefore, the students in our study with higher levels of calibration may be more ‘expert-like’ in the regions they deploy when evaluating error-containing models. Brault Foisy and colleagues63 suggested that lateral prefrontal activity may be especially relevant for resolving interference when viewing scientific errors or misconceptions. Given that the DLPFC forms a core part of the fronto-parietal network associated with executive control69, findings may indicate that these students are more effectively deploying executive resources to resolve the errors present in these models. Against a growing literature showing that error detection and the associated recruitment of lateral PFC is a marker of STEM expertise, our study links the lateral PFC recruitment to students’ metacognitive calibration. Effective metacognitive calibration may act as a bridge to conceptual change by facilitating students’ use of higher-level conflict resolution and executive resources associated with the lateral PFC.
Surprisingly, students’ behavioral phi, meta-efficiency, and confidence bias scores were not linked significantly to their ultimate course grades, although both model evaluation accuracy and meta-d’ scores were significantly correlated with course grades. Meta-d’ reflects the predicted d’ of an individual, given their proportions of calibrated vs. non-calibrated confidence ratings70. The more meaningful indicator of a student’s metacognitive calibration relative to other students is thus the metacognitive efficiency score. The lack of correlation between meta-efficiency and course grades may indicate that, once we consider the individual’s error detection performance, calibration is less related to overall class grades. Thus, although higher calibration may be tied to the use of prefrontal brain regions associated with effortful resource deployment and, in previous studies, to students’ self-regulated learning and academic success71,72,73, it is students’ mastery of the concepts in these models and their ability to recognize accurate relative to inaccurate models that ultimately is reflected in their course performance. In summary, we discovered that the neural results were distinct from our behavioral results, indicating that there may be neural implications of metacognitive calibration that cannot be inferred only from behavioral results.
Students’ mean confidence was slightly higher than their accuracy (27 vs. 23) on the model task and students in general responded that they were confident more often than not for inaccurate trials. Indeed, some students had a biased score that exceeded 0, which means that they had more confident, inaccurate trials than confident, accurate trials. These descriptive findings support the notion that students tend to overestimate their performance, particularly when performance is low25. This overconfidence in one’s abilities suggests a lack of metacognitive awareness of their deficits, which may lead to ineffective self-regulation skills12. For example, an overconfident student, believing they know the material, might decide not to study for a test, thus increasing that student’s probability of doing poorly due to inadequate preparation.
Although we did not find an association between confidence bias and performance or course grades, we did find that confidence bias correlated with lower metacognitive efficiency and with less left inferior frontal signal change when students viewed error-containing models. Bellon et al.22 found that children showed activation in the left inferior frontal gyrus when rating their confidence in arithmetic problem-solving and that the extent of children’s activity in this region was linked to their mathematics performance. Unlike Bellon et al., we examined neural activity during model evaluation as opposed to during the interval when students were making confidence ratings. However, the overlapping link to the left inferior frontal gyrus perhaps suggests that this region is important for modulating links between confidence and achievement.
Interestingly, differences in neural activity for accurate and confident relative to inaccurate and confident trials were confined to the lingual gyrus; there were no differences in prefrontal activity for this contrast. Potvin et al.60 similarly found that calibrated vs. overconfident trials in an electric circuit validation task were linked to parietal, premotor, and fusiform regions rather than prefrontal regions. It is possible that this activity in posterior regions for correct and calibrated trials reflects the use of visual processing resources to support accurate processing and evaluation of the model.
A significant study limitation is that one model evaluation task is not necessarily synonymous with overall academic performance or general metacognitive abilities. Additional assessments would be beneficial for determining and evaluating the neural and behavioral effects of metacognitive processes and of different forms of instruction. It is also important to note that the relationship between brain activity and metacognitive calibration is complex and multifaceted. Different aspects of metacognition, such as monitoring, control, and evaluation, may involve distinct brain regions and networks. Individual differences in brain structure and functioning, as well as task-specific factors, can influence the relationship between brain activity and metacognition. There is continued debate, for example, as to whether metacognitive calibration is task-specific or a general, trait-like characteristic74,75.
Another limitation of the study was that we had insufficient trial numbers to fully cross accuracy and confidence at the trial level to examine implications for brain activity: high levels of confidence meant that there were small numbers of trials where students were underconfident. Contrasts of confident vs. non-confident trials or calibrated vs. overconfident trials may also be less reliable because of unbalanced trial numbers for these conditions. Moreover, given the complexity of the task stimuli, each error-containing model was equated in complexity to non-error-containing models. Because the control for stimulus complexity was lost when we removed inaccurate trials, our primary contrast was based on all trials, regardless of students’ accuracy. Future studies should examine how accuracy modulates these relationships. Other limitations included a disproportionate representation of females and Euro-American Whites, as well as a small range in GPAs, which may have limited our power to capture the full extent of variation in students’ metacognitive monitoring performance.
Future research is needed to determine how these individual differences in brain activity relate to ongoing academic achievement. One especially intriguing question is whether and how manipulations to instructional design, such as immediate feedback following errors, may impact students' brain activity and, ultimately, help students become more self-regulated learners76. Likewise, future research should also examine how students might be encouraged to reflect on their calibration accuracy and whether such reflection might aid in the process of conceptual change in science. Developing a more nuanced confidence scale might be beneficial to examine confidence judgments in relation to metacognitive calibration. Lastly, additional research is needed to determine how metacognitive calibration affects students’ learning and academic performance throughout their educational trajectory to understand learning acquisition for longer-term retention.
The present study’s finding that students with higher metacognitive monitoring and less confidence bias deployed more ‘expert-like’ lateral prefrontal activity when they encountered error-containing models underscores a need to foster and nurture metacognition and self-awareness in the classroom. Conceptual change is theorized to drive science learning and hinges on students’ willingness to engage with or experience conflict from new ideas45,47,55,77,78,79. Error detection is crucial for effective learning, helping students to identify knowledge gaps, facilitating deeper processing and reflection, and ultimately leading to improved mastery. Embracing errors as valuable learning opportunities can significantly enhance the learning process and foster continuous growth and improvement80,81,82. Students who persistently learn from their own errors through self-reflection may be more motivated to continue to learn after failure, to correct their errors, and to recognize misconceptions83,84. Instructors might provide opportunities for students to reflect on their coursework by focusing on the learning process rather than on the content itself85, thereby engaging students in reflexive and adaptive thinking. They may also provide direct and immediate feedback, a core mechanism of self-regulated learning, which can positively impact academic achievement86. Finally, instructors may need to teach effective study habits to help students act on their metacognitive judgements, as students may lack this knowledge even when they are prepared to alter their strategies87. These instructional strategies may represent effective pathways for helping STEM students to be active, self-regulated learners and for moving them toward more expert model evaluation capacities.
Method
Participants
Fifty-one undergraduate students were recruited through class announcements from five separate sections of one introductory life sciences course taught by two different instructors in two consecutive academic years at a large Midwestern university. First, we recruited from the sections taught by a professor who heavily used model-based instruction (n = 35 students). To maximize our sample size from the following academic year’s sections we recruited more students from one section taught by the professor who also used model-based instruction (n = 6), and students from sections of the same course taught by an instructor who used less model-based instruction (n = 10). Both instructors have taught the introductory life sciences course for over nine years and used models regularly in their pedagogy. While specific models in this study were never seen or used in the course, the concepts are foundational to biology. While students may have been exposed to textbooks or classes using foundational biology models, students were not familiar with the models as presented in this study. Therefore, students needed to draw on their knowledge and understanding of these biological systems, as opposed to visual recognition of the model, to determine whether each model included an error (see supplementary Fig. 4 for an example of a presented model).
Students were screened to ensure that none had a learning disability, Attention-Deficit/Hyperactivity Disorder, experience of concussion, or other neurological diagnosis that might impact neural response patterns and that none had contraindications to MRI. One student was excluded from analyses because they consistently gave the same response to every task trial. Of the final analytic sample (Valid N = 50, Mage = 19.62, SDage = .0.90), 35 (70%) were first-year freshmen, 12 (24%) sophomores, and three (6%) juniors. Seven (14%) were first-generation college students. Forty-three (86%) were European American/White, three (6%) Hispanic, three (6%) Asian, and one (2%) identified as both European American/White and Hispanic. All but two students were native English speakers, 38 (76%) were female, 10 (20%) male, and two (4%) another gender. The average grade-point average (GPA), on a 4.0 scale and collected through a third party from the Registrar’s Office, was 3.54 (SD = 0.53). The average final course grade from the introductory life sciences course, also on a 4.0 scale, was 3.55 (SD = 0.62). For analyses, final course grades are used, rather than cumulative GPAs, because final course grades are a more concise measure of student biology knowledge. On average, participants’ combined aggregated standardized score on the Kaufman Brief Intelligence Test, Second Edition (KBIT-2)88 was 103.37 (SD = 11.07), with all students falling within two standard deviations of the normative mean.
Procedure
All procedures were approved by the University of Nebraska Human Subjects Protection Committee, and participants provided written informed consent to participation. Scans were also sent to a radiologist for review and students were informed of any incidental findings, with all of these being limited to sinus congestion. Students were compensated with $50 cash after attending a 2-h appointment at the university’s imaging center, where they completed the fMRI task. Study appointments occurred after the semester’s fourth week to allow students to become familiar with core course concepts and with the process of evaluating models or diagrams in biology.
Students underwent MRI in a 3 T Siemens Skyra scanner using a 32-channel head coil. After being fitted with ear protection, students reclined on the scanning table. First, a T1-weighted MPRAGE was acquired (TR = 1, TE = 2.95 ms, voxel size = 1 mm3, flip angle = 9, field of view = 270,176 sagittal slices) for registration purposes. This scan was followed by T2*-weighted echoplanar images (TR = 1 s, TE = 25 ms, 3 mm voxels, flip angle = 90°, FOV = 224 mm) collected during the model evaluation task.
Model evaluation task
In the scanner, participants evaluated a series of models, formatted as flow charts, diagrams, or textbook-like images, that captured a breadth of content from the introduction to the life sciences course (e.g., human evolution, central dogma, genetic mutation). Stimuli were designed to examine the participants’ ability to detect errors and inhibit misconceptions (Fig. 6). To provide context for the model, students first saw a 2-s prompt, e.g., “Is there an error in the relationships?” They then viewed a model and the same prompt for 10 s, after which they could take up to 30 s to indicate via a response pad whether the model was correct or incorrect. Lastly, participants were cued to indicate whether they were confident in their response (yes or no). Participants completed three separate, randomly ordered runs lasting approximately 5 min each. Each run was composed of 12 randomly ordered trials. Generally, there were two incorrect versions for every correct version of a model, with 14 total correct models and 22 error-containing trials. Trials were followed by a baseline rest period with a jittered interval of 2–10 s, during which students saw random figures extracted from the model stimuli but without any words. Because students took varying times to respond, the number of volumes collected for each student varied between 231 and 424 volumes per run.
Data analysis
Collecting multiple confidence ratings affords calculation of a simple phi coefficient (ϕ), the correlation between students’ accuracy and confidence ratings across all trials of a task65. Although widely used in the education literature, this method is susceptible to bias based on the participant’s task performance and does not adequately parse students’ confidence bias—their general tendency to make high or low confidence ratings—from their capacity to discriminate correct from inaccurate responses35,70. Consequently, cognitive neuroscience researchers recommend using model-based Signal Detection Theory metrics to provide response-bias-free measures of how precisely confidence ratings track task accuracy64,65. These measures include meta-d’, a measure of Student’s metacognition that is conditioned on their task performance distribution, and metacognitive efficiency, which reflects the difference between a student’s sensitivity to their performance and their actual task performance. An ‘ideal metacognitive observer’ should exhibit little difference between their meta-d and their task performance65. Such signal detection metrics parse students’ confidence biases and task performance from their capacity to discriminate optimally from less optimal performance, enabling a deeper understanding of the implications of different aspects of metacognition for learning.
Using SPSS89, we calculated our primary metric for calibration accuracy, the phi coefficient (ϕ), which reflects the correlation between a student’s accuracy and confidence scores. Two students were excluded from this analysis because they reported that they were confident on every trial; therefore, we could not calculate their ϕ. We also calculated students’ total confidence (i.e., the total number of times they responded ‘yes’ to the confidence prompt) and total accuracy across all trials and runs of the task, as well as creating a biased score based on the proportion of inaccurate trials where students reported they were confident—the proportion of accurate trials on which they were confident89. We used MATLAB scripts developed by Fleming64 to calculate a type 1 d’ score, a type 2 meta-d’ score, and a metacognitive efficiency score based on Bayesian estimation using Markov Chain Monte Carlo simulation implemented with the JAGS package. The advantage of this Bayesian approach to estimating these metacognitive metrics is that it is more robust with small trial numbers and can be used even when individuals have 0 responses in a particular cell. We used d’—meta d’ as our estimate of metacognitive efficiency because the alternative, ratio-based metacognitive efficiency score resulted in some extreme values.
Functional MRI data were processed and analyzed using the FMRIB Software Library90. Images were skull stripped, high pass filtered at 0.01 Hz, realigned, spatially smoothed with a 5 mm Gaussian kernel, registered to the T1 image using boundary-based registration, and normalized to the MNI 2 mm template. In the first level models for each run, we regressed the fMRI signal on task onsets convolved with a double gamma hemodynamic response function. The interval of focus was between when participants viewed the model and prompt to when they responded whether the model was correct or incorrect, i.e., the ‘view model’ and ‘respond’ phases in Fig. 1. The ‘confidence’ and prompt slides were treated as nuisance regressors in the design matrix, along with regressors for motion and temporal derivatives. Contrasts were performed for trials reflecting Error > No Error models. Furthermore, we contrasted trials where students responded they were Confident > Not confident and trials where they were Confident and Accurate > Confident and Inaccurate in their responses. Parameters for each participant were averaged in fixed effects models and passed to a group mixed effects analysis. In separate models, we added students’ metacognitive scores as covariates of interest in the design matrix for the Error > No error contrast. All contrasts were performed with a Z threshold of 3.1, p < 0.001 and a cluster-corrected p < 0.05.
Reporting summary
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Data availability
Group statistical maps are available on NeuroVault at https://neurovault.org/collections/13902/. Student behavioral data is available at the UNL data repository: https://doi.org/10.32873/unl.dr.20230419. Raw MRI data will be made available to researchers upon reasonable request and is not publicly available because participants did not agree to public data access during informed consent.
Code availability
Analytic code is available from the authors upon request.
References
Zimmerman, B. J. & Martinez-Pons, M. Construct validation of a strategy model of student self-regulated learning. J. Educ. Psychol. 80, 284–290 (1988).
Roebers, C. M. Executive function and metacognition: towards a unifying framework of cognitive self-regulation. Dev. Rev. 45, 31–51 (2017).
Andrade, H. L. A critical review of research on student self-assessment. Front. Educ. 4, 1–13 (2019).
Stone, N. J. Exploring the relationship between calibration and self-regulated learning. Educ. Psychol. Rev. 12, 437–475 (2000).
Crane, N., Zusho, A., Ding, Y. & Cancelli, A. Domain-specific metacognitive calibration in children with learning disabilities. Contemp. Educ. Psychol. 50, 72–79 (2017).
Desender, K., Van Opstal, F. & Van Den Bussche, E. Feeling the conflict: the crucial role of conflict experience in adaptation. Psychol. Sci. 25, 675–683 (2014).
Boldt, A. & Gilbert, S. J. Partially overlapping neural correlates of metacognitive monitoring and metacognitive control. J. Neurosci. 42, 3622–3635 (2022).
Windschitl, M., Thompson, J. & Braaten, M. Beyond the scientific method: model-based inquiry as a new paradigm of preference for school science investigations. Sci. Educ. 92, 941–967 (2008).
Allaire-Duquette, G. et al. An fMRI study of scientists with a Ph.D. in physics confronted with naive ideas in science. Npj Sci. Learn. 6, 1–12 (2021).
Flavell, J. H. Metacognition and cognitive monitoring: a new area of cognitive-developmental inquiry. Am. Psychol. 34, 906–911 (1979).
Cross, D. R. Developmental and instructional analyses of children’s metacognition and reading comprehension. J. Educ. Psychol. 80, 131–142 (1988).
Kruger, J. & Dunning, D. Unskilled and unaware of it: how difficulties in recognizing one’s own incompetence lead to inflated self-assessments. J. Pers. Soc. Psychol. 77, 1121–1143 (1999).
Roebers, C. M., Krebs, S. S. & Roderer, T. Metacognitive monitoring and control in elementary school children: their interrelations and their role for test performance. Learn. Individ. Differ. 29, 141–149 (2014).
Spiess, M. A., Meier, B. & Roebers, C. M. Development and longitudinal relationships between children’s executive functions, prospective memory, and metacognition. Cogn. Dev. 38, 99–113 (2016).
Bandura, A. Social cognitive theory: an agentic perspective. Asian J. Soc. Psychol. 2, 21–41 (1999).
Paris, S. G. & Newman, R. S. Development aspects of self-regulated learning. Educ. Psychol. 25, 87–102 (1990).
Winne, P. H. Improving measurements of self-regulated learning. Educ. Psychol. 45, 267–276 (2010).
Bellon, E., Fias, W. & De Smedt, B. More than number sense: the additional role of executive functions and metacognition in arithmetic. J. Exp. Child Psychol. 182, 38–60 (2019).
Dunlosky, J. & Rawson, K. A. Overconfidence produces underachievement: inaccurate self evaluations undermine students’ learning and retention. Learn. Instr. 22, 271–280 (2012).
Rinne, L. F. & Mazzocco, M. M. M. Knowing right from wrong in mental arithmetic judgments: calibration of confidence predicts the development of accuracy. PLoS ONE 9, 1–11 (2014).
Wollenschläger, M., Hattie, J., Machts, N., Möller, J. & Harms, U. What makes rubrics effective in teacher-feedback? Transparency of learning goals is not enough. Contemp. Educ. Psychol. 44-45, 1–11 (2016).
Bellon, E., Fias, W., Ansari, D. & De Smedt, B. The neural basis of metacognitive monitoring during arithmetic in the developing brain. Hum. Brain Mapp. 41, 4562–4573 (2020).
Fleming, S. M., Weil, R. S., Nagy, Z., Dolan, R. J. & Rees, G. Relating introspective accuracy to individual differences in brain structure. Science 329, 1541–1543 (2010).
Dunning, D., Johnson, K., Ehrlinger, J. & Kruger, J. Why people fail to recognize their own incompetence. Curr. Dir. Psychol. Sci. 12, 83–87 (2003).
Hacker, D. J., Bol, L., Horgan, D. D. & Rakow, E. A. Test prediction and performance in a classroom context. J. Educ. Psychol. 92, 160–170 (2000).
Gigerenzer, G., Hoffrage, U. & Kleinbfilting, H. Probabilistic mental models: a brunswikian theory of confidence. Psychol. Rev. 98, 506–528 (1991).
Fleming, S. Know Thyself: The Science of Self-Awareness (Basic Books, 2021).
Mansouri, F. A., Koechlin, E., Rosa, M. G. P. & Buckley, M. J. Managing competing goals —a key role for the frontopolar cortex. Nat. Rev. Neurosci. 18, 645–657 (2017).
Molenberghs, P., Trautwein, F.-M., Böckler, A., Singer, T. & Kanske, P. Neural correlates of metacognitive ability and of feeling confident: a large-scale fMRI study. Soc. Cogn. Affect. Neurosci. 11, 1942–1951 (2016).
Lebreton, M., Abitbol, R., Daunizeau, J. & Pessiglione, M. Automatic integration of confidence in the brain valuation signal. Nat. Neurosci. 18, 1159–1167 (2015).
Morales, J., Lau, H. & Fleming, S. M. Domain-general and domain-specific patterns of activity supporting metacognition in human prefrontal cortex. J. Neurosci. 38, 3534–3546 (2018).
Ridderinkhof, K. R., Ullsperger, M., Crone, E. A. & Nieuwenhuis, S. The role of the medial frontal cortex in cognitive control. Science 306, 443–447 (2004).
Taylor, S. F., Stern, E. R. & Gehring, W. J. Neural systems for error monitoring: recent findings and theoretical perspectives. Neuroscientist 13, 160–172 (2007).
Ullsperger, M., Danielmeier, C. & Jocham, G. Neurophysiology of performance monitoring and adaptive behavior. Physiol. Rev. 94, 35–79 (2014).
Fleur, D. S., Bredeweg, B. & Van Den Bos, W. Metacognition: Ideas and insights from neuro- and educational sciences. Npj Sci. Learn. 6, 1–11 (2021).
Qiu, L. et al. The neural system of metacognition accompanying decision-making in the prefrontal cortex. PLoS Biol. 16, 1–27 (2018).
Shekhar, M. & Rahnev, D. Distinguishing the roles of dorsolateral and anterior PFC in visual metacognition. J. Neurosci. 38, 5078–5087 (2018).
Seow, T. X. F., Rouault, M., Gillan, C. M. & Fleming, S. M. How local and global metacognition shape mental health. Biol. Psychiatry 90, 436–446 (2021).
National Academy of Science. Next Generation Science Standards: For States, By States (National Academies Press, 2013).
Brewe, E. & Sawtelle, V. Modelling instruction for university physics: examining the theory in practice. Eur. J. Phys. 39, 1–25 (2018).
Bryce, C. M. et al. Exploring models in the biology classroom. Am. Biol. Teach. 78, 35–42 (2016).
Dauer, J. T. et al. (eds) Fostering Understanding of Complex Systems in Biology Education. Contributions from Biology Education Research (Springer, Cham, 2022).
Brault Foisy, L., Matejko, A. A., Ansari, D. & Masson, S. Teachers as orchestrators of neuronal plasticity: effects of teaching practices on the brain. Mind Brain Educ. 14, 415–428 (2020).
Dong, A., Jong, M. S.-Y. & King, R. B. How does prior knowledge influence learning engagement? The mediating roles of cognitive load and help-seeking. Front. Psychol. 11, 1–10 (2020).
Posner, G. J., Strike, K. A., Hewson, P. W. & Gertzog, W. A. Accommodation of a scientific conception: toward a theory of conceptual change. Sci. Educ. 66, 211–227 (1982).
Dole, J. A. & Sinatra, G. M. Reconceptalizing change in the cognitive construction of knowledge. Educ. Psychol. 33, 109–128 (1998).
Nadelson, L. S., Heddy, B. C., Jones, S., Taasoobshirazi, G. & Johnson, M. Conceptual change in science teaching and learning: introducing the dynamic model of conceptual change. Int. J. Educ. Psychol. 7, 151–195 (2018).
Kendeou, P., Butterfuss, R., Kim, J. & Van Boekel, M. Knowledge revision through the lenses of the three-pronged approach. Mem. Cogn. 47, 33–46 (2019).
Broughton, S. H., Sinatra, G. M. & Reynolds, R. E. The nature of the refutation text effect: an investigation of attention allocation. J. Educ. Res. 103, 407–423 (2010).
Danielson, R. W., Sinatra, G. M. & Kendeou, P. Augmenting the refutation text effect with analogies and graphics. Discourse Process 53, 392–414 (2016).
Pieschl, S., Budd, J., Thomm, E. & Archer, J. Effects of raising student teachers’ metacognitive awareness of their educational psychological misconceptions. Psychol. Learn. Teach. 20, 214–235 (2021).
Ehrlinger, J., Johnson, K., Banner, M., Dunning, D. & Kruger, J. Why the unskilled are unaware: further explorations of (absent) self-insight among the incompetent. Organ. Behav. Hum. Decis. Process. 105, 98–121 (2008).
Kendeou, P., Walsh, E. K., Smith, E. R. & O’Brien, E. J. Knowledge revision processes in refutation texts. Discourse Process 51, 374–397 (2014).
Vaughn, A. R., Brown, R. D. & Johnson, M. L. Understanding conceptual change and science learning through educational neuroscience. Mind Brain Educ. 14, 82–93 (2020).
Masson, S., Potvin, P., Riopel, M. & Foisy, L.-M. B. Differences in brain activation between novices and experts in science during a task involving a common misconception in electricity: brain activation related to scientific expertise. Mind Brain Educ. 8, 44–55 (2014).
Potvin, P., Malenfant-Robichaud, G., Cormier, C. & Masson, S. Coexistence of misconceptions and scientific conceptions in chemistry professors: a mental chronometry and fMRI study. Front. Educ. 5, 1–16 (2020).
Fleming, S. M. & Dolan, R. J. The neural basis of metacognitive ability. Philos. Trans. R. Soc. B Biol. Sci. 367, 1338–1349 (2012).
Bogacz, R., Wagenmakers, E.-J., Forstmann, B. U. & Nieuwenhuis, S. The neural basis of the speed-accuracy tradeoff. Trends Neurosci. 33, 10–16 (2010).
Nenciovici, L., Allaire-Duquette, G. & Masson, S. Brain activations associated with scientific reasoning: a literature review. Cogn. Process. 20, 139–161 (2019).
Potvin, P., Turmel, É. & Masson, S. Linking neuroscientific research on decision making to the educational context of novice students assigned to a multiple-choice scientific task involving common misconceptions about electrical circuits. Front. Hum. Neurosci. 8, 1–13 (2014).
Fleming, S. M. & Frith, C. D. The Cognitive Neuroscience of Metacognition (Springer, Berlin, Heidelberg, 2014).
Weil, L. G. et al. The development of metacognitive ability in adolescence. Conscious. Cogn. 22, 264–271 (2013).
Brault Foisy, L.-M., Potvin, P., Riopel, M. & Masson, S. Is inhibition involved in overcoming a common physics misconception in mechanics? Trends Neurosci. Educ. 4, 26–36 (2015).
Fleming, S. M. HMeta-d: hierarchical Bayesian estimation of metacognitive efficiency from confidence ratings. Neurosci. Conscious. 1, 1–14 (2017).
Fleming, S. M. & Lau, H. C. How to measure metacognition. Front. Hum. Neurosci. 8, 1–9 (2014).
Nielsen, W. S., Nashon, S. & Anderson, D. Metacognitive engagement during field‐trip experiences: a case study of students in an amusement park physics program. J. Res. Sci. Teach. 46, 265–288 (2009).
Eilam, B. & Reiter, S. Long-term self-regulation of biology learning using standard junior high school science curriculum: long-term self-regulated biology curriculum. Sci. Educ. 98, 705–737 (2014).
Martin, B. L., Mintzes, J. J. & Clavijo, I. E. Restructuring knowledge in biology: Cognitive processes and metacognitive reflections. Int. J. Sci. Educ. 22, 303–323 (2000).
Niendam, T. A. et al. Meta-analytic evidence for a superordinate cognitive control network subserving diverse executive functions. Cogn. Affect. Behav. Neurosci. 12, 241–268 (2012).
Maniscalco, B. & Lau, H. Signal detection theory analysis of type 1 and type 2 data: meta-d′, response-specific meta-d’, and the unequal variance SDT model. In The Cognitive Neuroscience of Metacognition (eds Fleming, S. M. & Frith, C. D.) 25–66 (Springer, Berlin, Heidelberg, 2014).
Carretti, B., Caldarola, N., Tencati, C. & Cornoldi, C. Improving reading comprehension in reading and listening settings: the effect of two training programmes focusing on metacognition and working memory. Br. J. Educ. Psychol. 84, 194–210 (2014).
Ohtani, K. & Hisasaka, T. Beyond intelligence: A meta-analytic review of the relationship among metacognition, intelligence, and academic performance. Metacogn. Learn. 13, 179–212 (2018).
Zepeda, C. D., Richey, J. E., Ronevich, P. & Nokes-Malach, T. J. Direct instruction of metacognition benefits adolescent science learning, transfer, and motivation: an in vivo study. J. Educ. Psychol. 107, 954–970 (2015).
Pieschl, S. Metacognitive calibration—an extended conceptualization and potential applications. Metacogn. Learn. 4, 3–31 (2009).
Winne, P. H. & Perry, N. E. Measuring self-regulated learning. In Handbook of Self-Regulation (eds Boekaerts, M., Pintrich, P. R. & Zeidner, M.) 531–566 (Orlando, FL, Academic Press, 2000).
Mathan, S. A. & Koedinger, K. R. Fostering the intelligent novice: learning from errors with metacognitive tutoring. Educ. Psychol. 40, 257–265 (2005).
diSessa, A. A. A “Theory bite” on the meaning of scientific inquiry: a companion to Kuhn and Pease. Cogn. Instr. 26, 560–566 (2008).
diSessa, A. A. & Sherin, B. L. What changes in conceptual change? Int. J. Sci. Educ. 20, 1155–1191 (1998).
Novak, J. D. Meaningful learning: the essential factor for conceptual change in limited or inappropriate propositional hierarchies leading to empowerment of learners. Sci. Educ. 86, 548–571 (2002).
Cooper, K. M., Downing, V. R. & Brownell, S. E. The influence of active learning practices on student anxiety in large-enrollment college science classrooms. Int. J. STEM Educ. 5, 1–18 (2018).
Steele-Johnson, D. & Kalinoski, Z. T. Error framing effects on performance: cognitive, motivational, and affective pathways. J. Psychol. 148, 93–111 (2014).
Zamora, Á., Suárez, J. M. & Ardura, D. Error detection and self-assessment as mechanisms to promote self-regulation of learning among secondary education students. J. Educ. Res. 111, 175–185 (2018).
Tulis, M., Steuer, G. & Dresel, M. Learning from errors: a model of individual processes. Frontline Learn. Res. 4, 12–26 (2016).
Zimmerman, B. J. Self-regulation involves more than metacognition: A social cognitive perspective. Educ. Psychol. 30, 217–221 (1995).
Upmeier Zu Belzen, A., Engelschalt, P. & Krüger, D. Modeling as scientific reasoning—the role of abductive reasoning for modeling competence. Educ. Sci. 11, 1–11 (2021).
Butler, D. L. & Winne, P. H. Feedback and self-regulated learning: a theoretical synthesis. Rev. Educ. Res. 65, 245–281 (1995).
Stanton, J. D., Neider, X. N., Gallegos, I. J. & Clark, N. C. Differences in metacognitive regulation in introductory biology students: when prompts are not enough. CBE—Life Sci. Educ. 14, 1–12 (2015).
Kaufman, A. S., & Kaufman, N. L. (with Breaux, K. C.) Technical & Interpretive Manual. Kaufman Test of Educational Achievement Brief Form 3rd edn (NCS Pearson, Bloomington, MN, 2014).
IBM Corp. Released 2020. IBM SPSS Statistics for Macintosh, Version 27.0 (IBM Corp., Armonk, NY, 2020).
Jenkinson, M., Beckmann, C. F., Behrens, T. E. J., Woolrich, M. W. & Smith, S. M. FSL. NeuroImage 62, 782–790 (2012).
Acknowledgements
This research is supported by NSF grant 2000549. We are grateful to the students who dedicated their time to participate in this study, as well as Eric Brewe, Nathan Faivre, Jonathan Fugelsang, and Cindy Hmelo-Silver for their expert feedback and assistance throughout this project.
Author information
Authors and Affiliations
Contributions
J.D. and C.C. designed and directed the project; J.D., C.C., M.G.B. and M.E. performed the experiments; C.C. analyzed the neural data; C.C. and M.G.B. developed the theoretical framework; M.G.B. wrote the manuscript, with significant input from C.C. and J.D. All authors provided critical feedback and helped shape the research, analyses, and manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Behrendt, M.G., Clark, C., Elliott, M. et al. Relation of life sciences students’ metacognitive monitoring to neural activity during biology error detection. npj Sci. Learn. 9, 16 (2024). https://doi.org/10.1038/s41539-024-00231-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41539-024-00231-z