Current Projects

The Attentional Curve of Forgetting
Forgetting is a fundamental part of our lives, but we know very little about how and why it happens. Most importantly, past research has mostly neglected the critical role played by visual attention when memories are stored in the first place. Using a novel combination of eye-tracking and advanced statistical modelling, this research seeks to explicitly link overt attention to the forgetting of naturalistic visual information. We will develop advanced statistical models that predict the rate of forgetting at the level of individuals integrating measures of recognition performance, overt attention and intrinsic properties of the images. Critically, this research is framed within a cognitive aging framework to provide key insights into which memory mechanisms change or are preserved as we age. The project’s findings will generate timely theoretical insights and offer technological advances in predictive modelling with key implications for developing diagnostic tools for the early detection of neuropathology yet in the prodromal stage. The project is funded by NextGeneration EU under Piano di Ripresa e di Resilienza (PRIN 2022) with grant agreement (2022APAFFN; CUP B53D23014480001).
Early Discrimination of Dementia Types in Synchronised Brain and Eye Activity

Lewy Body Dementia (DLB) is the second most common type of degenerative dementia after Alzheimer’s, yet it remains poorly understood and is frequently misdiagnosed. By the time clear symptoms appear, the disease is often advanced, limiting treatment options.  This project aims to identify hidden cognitive markers by looking at how brain activity responds when individuals in prodromal dementia stages deploy their attention during simple visual tasks. Our key objectives are to pinpoint specific, abnormal patterns in eye movements and brainwave signals in patients in the earliest stages of dementia, link these patterns directly to the core symptoms of the disease, like visual hallucinations, and so be able to discriminate DLB type of dementia from others (e.g., Alzheimer’s disease). By focusing on the visual system, which is known to be affected early in DLB, and using non-invasive, inexpensive, and highly sensitive technology (eye-tracking and EEG), we aim to catch the disease in its tracks while opening promising new avenues for early detection. This project is funded by NextGeneration EU under Piano di Ripresa e di Resilienza (PRIN/PNRR 2022) with grant agreement (P2022WPNKZ).

Enhancing Digital Neuropsychological Assessments

Healthy ageing has become a critical challenge with the global rise in life expectancy. While traditional neuropsychological tests are valuable, they are often administered too late and lack temporally sensitive measures, e.g., eye movements, which can reveal more subtle markers highlight dementia yet in an early stage. This project combines empirical research with software engineering to revolutionise how we monitor and assist healthy ageing. By building on digitalised neuropsychological tests, we are developing a new system that incorporates fine-grained data like eye movements and motor responses, tracks individuals over time, and uses machine learning to create personalised, adaptive assessments and training routines. The key objective of this project is to develop a robust, affordable, scalable and cost-effective system that capitalises on temporally sensitive eye movement responses, harvested through webcams, to improve the early detection of prodromal dementia and allow earlier interventions to slow or delay the progression of neurodegeneration. This project has been funded by Sapienza, Università di Roma, with grant agreement (RG123188B3F9299C). 

Past Projects and Research Lines

Understanding Long-term Memory Mechanisms of Naturalistic Visual Information in Healthy and Neuro-Degenerate Ageing

The format of memory representations for visual information has always been of central interest in vision science. A key debate regards the role played by semantic knowledge in scaffolding such representations, and how it interfaces with perceptual (sensorial) information of independent episodic instances stored in memory. Moreover, as we age, the way we form, maintain and access these representations can significantly change, and so it is critical to assess what mechanisms are preserved (or impaired) in later life and especially in those individuals that may be experiencing early signs of neurodegeneration (e.g., mild cognitive impairment, MCI). By combining experimentation and computational modelling, this project has demonstrated that eye movements can reveal the effects of semantic interference during the learning of naturalistic scenes (Mikhailova et al., 2021, Psychonomic Bulleting and Review), and this mechanism is still operating in healthy aged individuals, as well as people suffering from MCI (Coco, et al., 2021, Neuropsychology). Moreover, we uncovered that semantic interference does not automatically imply a detriment to memory representations, as our ability to discriminate between similar items may benefit from it (Delhaye et al., 2024, Memory and Cognition). Through computational analyses of the representational layers of visual information using algorithms based on low- and high-level visual information, as well as more complex deep neural network models, we distinguished between perceptual and semantic interference on similarity judgment (Mikhailova et al., 2022, Lecture Notes in Computer Science) and recognition memory (Mikhailova et. al., 2024, Cognitive Processing), while developing predictive models of MCI with results comparable to state of the art (Rocha, et al., 2025, to appear). The achievements of this project, funded by Fundação para a Ciência e a Tecnologia (PT) and the EU through the European Regional Development Fund (ERDF) with grant agreement (PTDC/PSI-ESP/30958/2017), collectively advance our knowledge of visual memory, its degradation, and the potential for new diagnostic markers, contributing significantly to the fields of cognitive neuroscience and neuropsychology. 

Semantic Influences on Gaze Control and Short-Term Visual Memory in Healthy and Neurodegenerative Aging

How high-level semantic knowledge versus low-level visual features are used to direct our gaze, and particularly what information can be processed outside the direct line of sight (in extra-foveal vision) has been extensively debated in vision science. Furthermore, understanding how these attentional and memory-binding processes are maintained or decline with age has often been neglected despite their critical role in identifying early signs of neurodegeneration. By combining eye-tracking, electrophysiology, and computational modelling, this project investigated how object and scene semantics shape attention and memory in healthy younger and older adults, as well as in individuals with Mild Cognitive Impairment (MCI). We demonstrated that the meaning of objects is processed in extra-foveal vision and can guide attention from the very first eye movement (Cimminella et al., 2020, Attention, Perception & Psychophysics), also in healthy older adults (Borges et al., 2020, Aging, Neuropsychology and Cognition) as well as in people suffering from Alzheimer’s disease (Cimminella, et al., 2022, Journal of geriatric psychiatry and neurology), challenging models assuming that extra-foveal guidance is based purely on low-level information. Further, we were the first to show, to our knowledge, that an object’s consistency with its surrounding scene is detected in brain responses timelocked to the fixation preceding foveation during natural vision (Coco et al., 2020, Journal of Cognitive Neuroscience). When examining these processes across the lifespan, we found that while healthy older adults and individuals with MCI can both use scene context to guide attention and form successful short-term representations (D’Innocenzo et al, 2022, Scientific Reports; Allegretti et al., 2025, Cortex), they employ compensatory strategies, such as longer fixations, to form stable memories (Coco et al., 2023, Neuropsychology). This highlights a close link between overt attention and memory formation in naturalistic settings. To facilitate this and future research, we produced the VISIONS database, a large, richly-annotated collection of naturalistic scenes with normed data on their attentional, conceptual, and perceptual features (Allegretti, et al., 2025, Behaviour Research Methods). The achievements of this project, funded by The Leverhulme Trust with grant agreement (ECF-014-205), advance our knowledge of object-scene integration by clarifying the interplay between visual and short-term memory mechanisms and uncovering how it is affected by healthy and pathological ageing. 

Action-Dynamics and Neural Correlates of Cross-Modal Integration

A fundamental question of human cognition is how it integrates information from different senses, like vision and language, to build coherent contextual representations of our surrounding world.  The project explored what happens when we encounter a “context violation”, for example, when the information depicted in a visual scene doesn’t match what a sentence refers to. Our aim was to understand how our cognitive system processes and resolves unexpected information, such as implausible stimuli and consequentially incongruent messages. We demonstrate that the interaction of expectancy mechanisms manifests in action-dynamics responses (Coco and Duran, 2017, Psychonomic Bulletin and Review), and their temporal dynamics are differentially expressed in neural signatures with messages evaluated first by their sequential congruency and later by their semantic plausibility (Coco, et al, 2017, Neuropsychologia). This research, funded by  Fundação para a Ciência e a Tecnologia with grant agreement (SFRH/BPD/88374/2012), provided a new approach to study cross-modal integration, offering a window into how our minds make sense of the world and the context we experience.

Interpersonal Synchrony: towards a Multidimensional Understanding of Collective Cognitive Dynamics

Humans live in a very interactive context, in which the coordination of behaviours between people to exchange information with conspecifics occurs through temporally sensitive mechanisms of interpersonal synchrony of linguistic and non-linguistic responses. Alignment mechanisms, such as imitation, entrainment, or priming, are fundamental to achieve shared understanding and, consequently, optimise joint action during cooperative tasks. By investigating,  the role played by feedback and alignment in achieving successful collaborations, we demonstrated that dyads unable to exchange feedback showed increased gaze alignment, but this actually led to a decrease in performance during a joint visual search task, suggesting that “too much” alignment can be detrimental if it prevents a diversified search strategy (Coco, et al., 2018, Topics in Cognitive Science). We also demonstrated that motor coordination between partners engaged in a cooperative block-stacking task is different across body parts (e.g., head vs. wrist) and influenced by task role (e.g., leader vs. follower; Coco et al, 2016, IEEE: Transactions on Cognitive and Developmental Systems). Moreover, we revealed that attentional coordination between a learner and a demonstrator’s actions is a key predictor of learning success, supporting the idea of “hand-eye coordination” as an alternative route to joint attention, challenging the more conventional view that gaze-following is the primary mechanism (Pagnotta, et al., 2022, Cognition). Finally, as interpersonal dynamics generate complex responses, our work has contributed to advancing the statistical methodologies available by implementing software packages capable of providing integrated analyses of multidimensional time series (Coco et al., 2021, The R-journal)

The Interplay of Vision and Language in Situated Contexts
Among the most fascinating aspects of human cognition is our ability to communicate. Language, though, never happens in a vacuum, and it occurs concurrently with other cognitive modalities, e.g., vision. This research line explores the intricate relationship between visual attention and language mechanisms of comprehension and production, demonstrating that the two are not separate processes but are tightly coordinated. By analysing eye-movement and speech data concurrently produced (or comprehended), we first demonstrated their intimate co-dependence by showing that the temporal sequences of where we look in a scene (i.e., a scan pattern) are predictive of the sentences we produce (Coco & Keller, 2012, Cognitive Science), within and across languages with different word orders (Coco et al., under review). These findings show, for the first time, the existence of a pre-linguistic universal grammar of perception, measurable in eye-movement responses, which transcends from the syntactic form. Further, when looking at the process of anchoring a linguistic referent as a visual object, we found that various mechanisms (perceptual, conceptual, and structural) have to be synchronized (Coco & Keller 2025, Cognitive Processing), and informed by both low-level visual features (like saliency) and high-level, knowledge-based information (like contextual fit; Coco, et al., 2014, Quarterly Journal of Experimental Psychology). Likewise, the comprehension of sentences in situated contexts is guided by suprasegmental properties of speech (Coco & Keller, 2015, Quarterly Journal of Experimental Psychology) as well as predictions attached to contextual expectations (Coco et al., 2016, Cognitive Science). Finally, attentional signatures subsumed in eye-movement responses can be revealing of the type of task a person is performing (Coco & Keller, 2014, Journal of Vision). Overall, this research line advances an integrated framework to understand situated cognition, and the fundamental mechanisms that coordinate how we look while we hear (or talk) about the world around us.