Cognitively Informed Cross-Lingual Curriculum Learning for Adaptive Reading Comprehension
DOI:
https://doi.org/10.64748/0pcg4t12Keywords:
curriculum learning, cognitive load, eye tracking, pupil dilation, surprisal, multilingual NLP, reading comprehension, fairnessAbstract
Adaptive educational systems rarely incorporate direct psychophysiological evidence about learners while also handling multilingual content with rigor. We propose NeuroCL, a cognitively informed curriculum learning framework that uses moment-to-moment estimates of cognitive load (derived from pupil dilation and eye movement features) to steer a cross-lingual reading comprehension curriculum. The framework combines (i) neurocognitive signals, (ii) linguistically grounded difficulty features (lexical overlap, morphological complexity, dependency length), and (iii) model uncertainty (surprisal/entropy from multilingual encoders) to adaptively select texts and questions in multiple languages. We outline theory, algorithms, and an empirical protocol, and we report results from a privacy-preserving pilot with synthetic-but-plausible data that demonstrate improved comprehension, faster reading times, and reduced cross-language performance gaps relative to static or heuristic curricula. We discuss implications for fair and transparent multilingual learning technologies.
Dr. Helena Varga is an Associate Professor specializing in memory, learning, and brain plasticity. Her research focuses on how neural networks adapt during skill acquisition, particularly in multilingual individuals. She integrates experimental psychology with neuroimaging techniques, aiming to bridge cognitive theory and applied educational practice. Dr. Varga has published extensively on working memory models and leads several EU-funded projects on bilingual education and neuroplasticity. She serves on the editorial boards of Cognitive Science and Frontiers in Psychology.
Downloads
Posted
How to Cite
License
Copyright (c) 2025 Marcello Conti, Helena Varga (Author)

This work is licensed under a Creative Commons Attribution 4.0 International License.