Cognitive Hearing Science: Three Memory Systems, Two Approaches, and the Ease of Language Understanding Model
Abstract
Purpose
The purpose of this study was to conceptualize the subtle balancing act between language input and prediction (cognitive priming of future input) to achieve understanding of communicated content. When understanding fails, reconstructive postdiction is initiated. Three memory systems play important roles: working memory (WM), episodic long-term memory (ELTM), and semantic long-term memory (SLTM). The axiom of the Ease of Language Understanding (ELU) model is that explicit WM resources are invoked by a mismatch between language input—in the form of rapid automatic multimodal binding of phonology—and multimodal phonological and lexical representations in SLTM. However, if there is a match between rapid automatic multimodal binding of phonology output and SLTM/ELTM representations, language processing continues rapidly and implicitly.
Method and Results
In our first ELU approach, we focused on experimental manipulations of signal processing in hearing aids and background noise to cause a mismatch with LTM representations; both resulted in increased dependence on WM. Our second—and main approach relevant for this review article—focuses on the relative effects of age-related hearing loss on the three memory systems. According to the ELU, WM is predicted to be frequently occupied with reconstruction of what was actually heard, resulting in a relative disuse of phonological/lexical representations in the ELTM and SLTM systems. The prediction and results do not depend on test modality per se but rather on the particular memory system. This will be further discussed.
Conclusions
Related to the literature on ELTM decline as precursors of dementia and the fact that the risk for Alzheimer's disease increases substantially over time due to hearing loss, there is a possibility that lowered ELTM due to hearing loss and disuse may be part of the causal chain linking hearing loss and dementia. Future ELU research will focus on this possibility.

Over the last 2 decades, the hearing research community has increasingly accepted that cognitive factors play an important role in models of hearing and language processing, all the way from early postcochlear processing of the speech signal to cortical understanding of gist. This applies especially to listening under adverse conditions (Mattys et al., 2012). Several ways of understanding this top-down–bottom-up interaction have been proposed (e.g., Akeroyd, 2008; Amichetti et al., 2013; Anderson et al., 2013; Arehart et al., 2013, 2015; Besser et al., 2013; Holmer et al., 2016a; Humes et al., 2013; Luce & Pisoni 1998; Pichora-Fuller et al., 2016; Rudner, 2018; Signoret & Rudner, 2019; Stenfelt & Rönnberg, 2009; Wingfield et al., 2015). Memory functions, especially working memory (WM), have been at the focus of our research on mechanisms behind communicative abilities in persons with hearing loss.
However, a more comprehensive version of the ELU model has to address how long-term memory systems contribute to online decoding, encoding, and inference-making in WM during communication. Therefore, our research has focused on Ease of Language Understanding (ELU) that depends on three interacting memory systems: WM, episodic long-term memory (ELTM), and semantic long-term memory (SLTM; see, e.g., Classon et al., 2013; Ng & Rönnberg, 2019; Rönnberg, 2003; Rönnberg et al., 2011, 2019, 2013, 2010). The works of Humes (e.g., Humes, 2007; Humes et al., 2013) and Gatehouse (Gatehouse et al., 2003, 2006) have been particularly inspirational in our contribution to the development of the field we have dubbed Cognitive Hearing Science. In Arlinger et al. (2009), a broader historical view on the emergence of Cognitive Hearing Science is presented.
Three Memory Systems and the ELU Model
The ELU model (Rönnberg, 2003; Rönnberg et al., 2019, 2016, 2013, 2008, 2010) builds on the interplay between the three memory systems (WM, ELTM, and SLTM) relevant for language understanding as well as their interface with a module for sensing language input. This module is assumed to operate Rapidly, Automatically, and Multimodally, when Binding PHOnological (RAMBPHO) information into a coherent percept, irrespective of source of the phonological information. This integration and binding of the different sources of phonological information is assumed to be very rapid, around 180–200 ms (Stenfelt & Rönnberg, 2009). A slower binding process would definitively slow down the implicit, predictive part of the model. Postdiction involves the explicit engagement of WM and its interactions with LTM systems (see more details and examples under Approach I: Memory System Interactions in Online Communication section).
The model was originally (Rönnberg, 2003; Rönnberg et al., 2008) postulated to cover different speech communication modes such as speechreading by means of tactile-visual (e.g., Rönnberg, 1993), visual–only speechreading (e.g., Lyxell & Rönnberg, 1987, 1989; Rönnberg, 1990), audiovisual facilitation in gating and speech perception (e.g., Moradi et al., 2017, 2019, 2014), and for manipulations of auditory target-auditory background stimuli (Lunner et al., 2009). The general cognitive observation was that, irrespective of communication mode, we were struck by the fact that complex WM capacity (WMC, especially measured by reading span), LTM access speed and the fidelity of phonological representations in LTM re-appeared as good predictor variables in study after study.
These observations prompted the formulating of a general and comprehensive ELU model, focused on interactions with the three memory systems (Rönnberg, 2003) and the communicative task (see Rönnberg et al., 2019 for thorough discussions of these matters, and under Approach I, this review article, for a more detailed model description).
Although most research has been focused on speech in one form or another, we have successively noted some constraints when comparing WM for sign and speech. For example, in Rönnberg et al. (2004), WM for sign showed specific activations of parietal areas but a communality with audiovisual speech for prefrontal/frontal areas of the cortex, typically associated with WM processing (Eriksson et al., 2015), was still true. This language modality specificity has also been shown recently in a study on WM for sign compared to moving visual nonsense objects (point-light displays; Cardin et al., 2018). While frontal activations were very similar in an n-back task for deaf native signers, compared to hearing native signers and hearing nonsigners, it turned out that superior temporal sulcus was specifically active in the deaf native signing group for both kinds of visual stimuli, suggesting early cortical and cognitive plasticity due to lack of sensory input (Cardin et al., 2018; see also Cardin et al., 2013, for cognitive plasticity; Rudner et al., 2007, for sign specificity regarding the episodic buffer).
Constraints on the ELU model with respect to hearing status have also been addressed in one meta-analytic study by Füllgrabe and Rosen (2016), demonstrating that, for hearing-impaired and older participants, WM accounts for significant portions of variance in explaining speech-in-noise (SPIN) performance, whereas this was not the case for normal-hearing young participants. However, later research has shown that even for subclinical/normal-hearing individuals, small variations in hearing acuity is associated with atrophy of the brain in predicted auditory and cognitive cortical sites (Rudner et al., 2019). The same message is true for a paper by Ayasse et al. (2019), where variations within the limits of normal hearing show that even minimal hearing loss has effects on listening to sentences in noise, when sentence grammar is more complex and WM-demanding. Thus, there are interesting findings on hearing status as a constraint on the ELU model such that it can be argued that the model also applies to so called normal or subclinical variability, even though these latter studies did not involve WMC as such (see reference to WMC and context dependence in normal-hearing participants in Rönnberg et al., 2019).
WM refers to an individual's ability to hold and manipulate a set of items or linguistic fragments currently in mind, for example, in the form of predictions or guesswork (Baddeley, 2012). Our use of the concept materialized after having read the seminal paper by Daneman and Carpenter (1980), in which they emphasized that the manipulation component of WM launched by Baddeley and Hitch (1974) was particularly important when it came to parsing sentences. In other words, there was a need for both storage and a processing function when more complex linguistic materials than the individual words used in much of Baddeley's early work were to be manipulated and understood (Baddeley et al., 2019). Apart from assisting in grammatical processing, the manipulation component of WM is engaged in semantic processing, inference-making (Hannon & Daneman, 2001), and keeping steadfast attention on the gist of a conversation, including turn-taking behavior (Rönnberg et al., 2013). Unlike Baddeley and Hitch (1974), our notion of WM is based on a central pool of resources that can be allocated flexibly to storage of sensory/semantic information, or to semantic, and grammatical processing. If processing of some kind takes most of the resources, less will be available for storage, and vice versa, if storage demands increase, then they will dampen or inhibit processing activities (Sörqvist et al., 2016). The only component of our WM model that is relatively encapsulated from the task-dependent and dynamic storage and processing functions is RAMBPHO. RAMBPHO must by necessity be fast, it can be primed and contextually framed, but before contact is made with LTM, only implicit less time-demanding operations can occur. In addition, the ELU model also differs to other models/frameworks of WM with respect to the communicative focus of WM. For detailed comparisons with other models, we refer to the discussion in Rönnberg et al. (2013).
We have typically used the visual reading span task (RST) introduced by Daneman and Carpenter (1980) as an index of WMC because then we avoided confounding cognitive measurement with audibility issues when investigating populations with hearing loss (see also Daneman & Merikle, 1996). The advantage of the RST test in the field of Cognitive Hearing Science compared to other tests of WMC is that it taps into the key dual-task storage-processing interaction in speech understanding mentioned above. Specifically, the RST taxes both storage of a set of sentences and semantic processing of each sentence in the set. In our use of the RST, we have instructed the participants to verify if a sentence is absurd or not (e.g., “The train sang a song”). After two or up to six presented sentences, the participant is asked to recall, in the correct sentence-wise order, the first or last words. Thus, in our version of the RST, you cannot be strategic in the sense of focusing on, for example, the last word of each sentence, simply because you do not know which words to recall until the whole set has been presented. The storage-processing interaction in the RST presumably reveals a “raw” real-life dynamic WMC and is therefore a better predictor of recall performance of targets in SPIN tasks than traditional simple span or letter updating/monitoring tasks (Rönnberg et al., 2016, 2013; Rudner et al., 2009). However, RST mimics what we actually do when we listen to speech in noise or at the cocktail party: We try to remember the gist, take turns in a dialogue, and execute semantic verification judgments more or less at the same time. Generally speaking, the dual demands of a WM task seem to be an even more crucial aspect than the actual presentation modality of the task (i.e., a visuospatial vs. a text-based task) in terms of predictive power of SPIN performance, and the tests load equally high on a latent WM factor in a large test battery of 200 participants (see Rönnberg et al., 2016).
The sensitivity of the storage-processing interaction is revealed in another way when we consider another type of WM-related experiments (Ng et al., 2013, 2015). We developed a SPIN test (i.e., the Sentence-final Word Identification and Recall [SWIR] test), in which each target word is audible (as evidenced by immediate recall of each individual sentence-final word), but where the delayed recall of the final target words of a set of sentences is still facilitated by hearing aid signal processing (i.e., noise reduction). This is true especially against a background of four talkers (4T, i.e., two men and two women reading from four different paragraphs of a newspaper; Micula et al., 2020) and for native speaker babble compared to 4T babble in a foreign language (Ng et al., 2015), implying a difference in the engagement of SLTM, and, hence, in the amount of distraction caused by the masker. Kilman et al. (2014) shows the same SLTM engagement in the case of bilingual experiments with maskers in the native versus nonnative language.
The proposal is that hearing aid signal processing of maskers can reduce distraction even for audible targets, hence supporting WM storage. In other words, when listening takes place under challenging conditions, hearing aid signal processing can relieve pressure on SLTM processing, rendering more storage capacity. Micula et al. (2020) has demonstrated that binary masking of the noise is particularly important when recall conditions are less predictable in the SWIR test (cf. Ng et al., 2015). Thus, storage and processing interact all the time (see more under postdiction) and, again, this is probably the main reason why, for example, the RST in many instances is a better predictor of sentence-based SPIN performance than simple span tests such as digit or word span (Rönnberg et al., 2013). In Rönnberg et al. (2016; Supplementary Materials, n = 200), we observed that the RST, semantic word pair test, and visuospatial WM (all being dual storage and processing tasks) loaded on the same WM factor (.57 to .68), all three loading higher than a nonword span task (only storage), which loaded .52 on that factor, which validates that the other three tasks are tapping into the processing aspect of WM as well. These dual tasks also were more effective predictors of, for example, recall of Hagerman matrix sentences in 4T babble. We can also make the inference that it is the dual task demands of the cognitive task, not the sensory modality that is critical here.
ELTM is a memory system of personally experienced events, or episodes, tagged by time, place, space, emotions, and context (Tulving, 1983). As we experience an episode, memory traces of multimodal sensory information, intertwined with semantic associations, related to the objects and context of the episode, are encoded as a personal episodic memory (Rugg et al., 2015). The retrieval process of ELTM is constructive, and reminiscence of a specific event is triggered and supported by episodic and semantic cues (Renoult et al., 2019; Rugg et al., 2015). That is, when an individual is trying to remember a specific episode, the person does this in an active manner, trying to use relevant sensory-perceptual traces and semantic associations to reconstruct the event, rather than simply accessing a stored video clip from LTM.
Thus, ELTM can be assessed in many ways, with varying contextual support (from, e.g., cued recognition, recognition, cued recall, to free recall), which in turn demand different levels of self-initiated memory search (Craik, 1983). A typical episodic everyday memory question is: “What did you have for breakfast this morning?” ELTM interacts with SLTM in the sense that ELTM always relies on preexisting knowledge structures (e.g., your mental lexicon). However, although ELTM depends on SLTM, neurocognitive evidence suggests partly nonoverlapping systems (Renoult et al., 2019). The notion of what constitutes “long term” varies with experimental paradigms, but most researchers would agree that from 30 min and beyond is acceptable as long term.
An important interaction between WM and ELTM was observed with a new kind of WM test. In Sörqvist and Rönnberg (2012), we measured WM by a dual task that first had the participants to compare the size of objects or animals (e.g., “is a zebra larger than a mouse”), and then the to-be-remembered (TBR) word appeared (e.g., “elephant”) in a list of comparison and TBR words. Serial recall of the TBR items is required, and the crucial aspect of this SIze Comparison span task (SIC span) is that both the TBR and comparison words belong to the same category in the same list. This can cause confusion at recall, that is, whether the recalled item was a TBR item or a comparison word. WMC measured this way demands the regular storage function and but also an inhibition processing component. In the experiment, the participant was instructed to focus on a target speech about a fictitious culture (masked by another spoken fictitious story). Data show that participants scoring high on the SIC span test predicted higher immediate recall and crucially performed better in delayed recall of the story. No similar correlation was found for RST. This example implies that ELTM is best promoted by a WM function that has the power to focus on the relevant semantic information while disregarding competing semantic information from SLTM.
SLTM refers to general knowledge, without personal reference, for example, vocabulary tests by means of fluency or by lexical access speed (e.g., Rönnberg et al., 2011, 2016), grammar (e.g., tested by means of comprehension of embedded clauses; Ayasse et al., 2019), phonology (e.g., tested by means of the Cross-Modal Phonological Awareness Test; Holmer et al., 2016b), or world knowledge, for example, in the form of scripts and knowledge about objects and people (Samuelsson & Rönnberg, 1993). An everyday example here would be that most people do not remember when they learned that Paris is the capital of France or, in the case of phonology, knowing that the speech sounds and associated visual cues of a specific word might overlap with sounds of other words in that language. Thus, in those cases, the information belongs to the general knowledge that we carry around in our minds or SLTM, not to personal ELTM traces.
Approach I: Memory System Interactions in Online Communication
When we first proposed the ELU model (Rönnberg, 2003; Rönnberg et al., 1998, 2008; Rudner et al., 2009, 2008), we were interested in describing a mechanism that explains why language understanding in some conditions demands extra allocation of cognitive resources, while in other conditions language processing takes place smoothly and effortlessly. To do this, we relied on the three memory systems briefly described above. In this online processing approach, we proposed that the mechanism in question was that of phonological mismatch between phonological information contained in the input signal (picked up by an RAMBPHO input buffer) and phonological representations in SLTM. The original hypothesis was that, especially, the syllable was an important linguistic unit for unlocking the lexicon (Rönnberg, 2003; Rönnberg et al., 2011). If the syllabic information perceived by the listener was distorted or blurred beyond a hypothetical threshold, then a mismatch would trigger WM to aid explicit reconstruction, or so called postdiction (Rönnberg et al., 2013, 2019). The ELU assumption is that explicit use of WM is involved to some degree in increasing effort (Rudner et al., 2012): more missed encoding and retrievals from ELTM and SLTM (Rönnberg et al., 2013) will inevitably cause increased perceived effort to overcome the obstacles of communication (Pichora-Fuller et al., 2016).
Postdiction
Several experimental methods have been developed to trigger the putative mismatch function between RAMBPHO output and existing SLTM representations. For example, experimental acclimatization to a nonhabitual kind of signal processing (e.g., Wide Range Dynamic Compression) in the hearing aid (e.g., FAST or SLOW Wide Range Dynamic Compression; Rudner et al., 2008, 2009), with subsequent testing in an acclimatized/familiarized (or nonacclimatized/nonfamiliarized) mode of signal processing, produced strong reliance on WM in mismatched conditions (i.e., FAST–SLOW or SLOW–FAST conditions). For reviews of these and other kinds of data supporting this kind of mismatch mechanism, see Rönnberg et al., 2019; Souza & Sirow, 2014; and Souza et al., 2015, 2019).
Another example is the manipulation of background noise where the use of speech babble maskers, engaging SLTM, produced the most pronounced distractions (e.g., Kilman et al., 2014; Mattys et al., 2012; Sörqvist & Rönnberg, 2012). It should be noted that the original data of WM dependence (using “speechlike” maskers) had already been observed and discussed (Lunner, 2003; Lunner & Sundewall-Thorén, 2007; see a review by Rönnberg et al., 2010). WMC is also an important predictor of ELTM in such circumstances of initial speech-in-speech maskers (Ng & Rönnberg, 2019; Sörqvist & Rönnberg, 2012). The SIC span emphasizes semantic inhibition rather than semantic verification (as in the RST) and seems like a better predictor of ELTM than the RST in that case (Sörqvist & Rönnberg, 2012; see a previous more detailed explanation).
As a further example, the ELU model predicts that when individuals become accustomed to the sounds transmitted by their hearing aids, they will automatize speech processing and will be less dependent on WM to understand speech since presumably representations build up over time in SLTM that more closely match the signal perceived (Holmer & Rudner, 2020; Rönnberg et al., 2019). Thus, less reconstructive postdiction processing is needed to disambiguate the speech signal. In line with this reasoning, Ng et al. (2014) demonstrated that after a period of up to 6 months with new hearing aid settings, initial associations with WMC during speech recognition in noise seemed to vanish. However, the original ELU model did not appropriately cover such developmental effects.
Therefore, in the context of sign language imitation, Holmer et al. (2016a) proposed the Developmental ELU (D-ELU) model to account for the importance of pre-existing cognitive representations that influences further development of new representations. The model assumes that mismatch-induced postdictive processes push the system toward appropriate adjustment of SLTM and that the formation of novel representations is supported and, at the same time, constrained by existing representations in the lexicon (Holmer & Rudner, 2020). The notion that WM dependence for hearing aid users becomes weaker over time (Ng et al., 2014) is thus in line with the D-ELU, which proposes that new representations are formed that are adapted for changed hearing conditions.
However, in a recent paper by Ng and Rönnberg (2019), we have been able to show that especially for speech maskers (4T) and for mild-to-moderate hearing impairment, there can be a much more prolonged (up to 10 years) WM dependence. This may imply that in some conditions, it is impossible to acclimatize to masking of target stimuli when speech distractors are dynamic and hard to form effective representations for. Interestingly, Han et al. (2019) reported worse word learning and weaker influence of existing representations on learning in the context of broadband white noise, suggesting that the mechanism proposed by the D-ELU is disrupted by noise. Particular combinations of speech and speech distractors may amplify mismatch between phonological representations and the semantic content of a person's SLTM. The SLTM component of speech maskers will always interfere with processing of target sentences to some degree (Ng & Rönnberg, 2019). It may actually be the general case that increased WM dependence becomes an everyday rule rather than an exception, even with advanced signal processing in the hearing aid. This, obviously, has clinical implications in the sense that the hearing impairment exacerbates effort and fatigue during the day, as well as in the long term (Rudner et al., 2011). We, of course, note that we are always more or less dependent on WM for language processing in real-life discourse, but what is meant here is the extra load on WM that may still come about in complicated RAMBPHO-LTM interactions.
Finally, recent studies on children show that WMC and vocabulary (i.e., SLTM) constitute important cognitive predictors when listening to speech in adverse conditions (Walker et al., 2019). This applies to children with normal hearing as well as children with hearing impairment (McCreery et al., 2019) and other hearing difficulties (Torkildsen et al., 2019). Furthermore, WMC predicts language development in children with hearing impairment, whereas vocabulary predicts reading comprehension in this group (Wass et al., 2019). This only goes to show that the ELU model has a certain amount of generality across the life span and the importance of vocabulary development gives support to D-ELU (Holmer et al., 2016a).
Prediction
So far, we have discussed the postdictive aspect of WM in the ELU model and its capacity to support reconstruction of misperceived information. However, as we have emphasized in Rönnberg et al. (2013, 2019), WM is also involved in the pretuning of the cognitive system and priming of to-be-understood sentences, albeit in a different functional role that demands less elaborative processing but is purposively related to identification and detection of targets. Examples of pretuning or priming do not necessarily build on explicit and elaborative processes. We have shown correlations with WM in different paradigms that rely on cognitive processes operating prior to actual stimulus presentations, for example, in repetition priming paradigms (Signoret & Rudner, 2019), or in cued sentence perception (semantically matching vs. mismatching cues of upcoming target sentences; Zekveld et al., 2011, 2012, 2013). A further remarkable example is when WM load in a visual letter-based n-back task seems to dampen the postcochlear, olivary complex responses (Wave V5) to tones in an odd-ball paradigm (Sörqvist et al., 2012; see also Kraus & White-Schwoch, 2015; Lehmann & Skoe, 2015; Molloy et al., 2015). Dampening of the W5 response was even further reinforced by the WMC of the participant, obviously having an early, top-down, inhibitory effect on brainstem processing and attention.
Thus, this last example represents resource allocation due to a clearly explicit involvement of WM, whereas in the two preceding examples, information is kept in mind in a way that could be either explicit or implicit depending on the participants' task strategy. In a more recent experiment, building on Sörqvist et al. (2012), we employed the same visual n-back—auditory odd-ball paradigm—and predictably so, inhibition and dampening of cortical activity of the superior temporal lobe was observed especially with high WM load (Sörqvist et al., 2016; see also Rosemann & Thiel, 2018; Sharma & Glick, 2016).
Our demonstrations cited above show that while some items will be held explicitly in WM (prestimulus presentation), other paradigms use an implicit side of WM when it comes to prediction (e.g., Davis et al., 2005). This implies that the explicit/implicit distinction is not as crucial for predictive—compared to postdictive—processes as we previously had assumed.
APPROACH II: ARHL and Long-Term Interactions Among Memory Systems
A second approach focuses on the long-term effects of age-related hearing loss (ARHL). The ELU model builds on a memory systems view of the long-term consequences of hearing impairment, unlike a common cause account (e.g., Baltes & Lindenberger, 1997; Humes et al., 2013), which assumes some common neural degeneration that is responsible for a general cognitive decline. The ELU model takes sides with a view that assumes that hearing loss may cause cognitive decline (Rönnberg et al., 2014) and that gray matter volume is proportional to audiometric hearing loss, which in turn is correlated with brain activity during sentence comprehension (Peelle et al., 2011; Peelle & Wingfield, 2016; Rudner et al., 2019). Lin (2011) and Lin et al. (2011, 2014) show that over times, it may be the case that the hearing loss is driving brain atrophy, which in turn will undermine cognitive integrity and ultimately lead to cognitive decline and dementia (Livingston et al., 2017). Indeed, we showed that even subclinical levels of poorer hearing in a middle-aged population are associated with smaller brain volumes in auditory and cognitive processing regions of the brain (Rudner et al., 2019). An even more recent study by Ayasse et al. (2019) shows that grammatical complexity is enough to tax the resources of participants with very small (within “normal”) hearing impairments.
Although not proving causality, independent data from the Betula database (Nilsson et al., 1997; Rönnberg et al., 2011; n = 160 hearing aid wearers) employing structural equation modelling (SEM), we obtain satisfactory model fits and significant links between variables only for hearing loss, not for vision loss (Rönnberg et al., 2011). Note also that using hearing-impaired participants who wear hearing aids is a conservative test of the hypothesis. Nevertheless, the hearing loss effect is manifested for two memory systems: ELTM and SLTM, not short-term memory, or WM. Finally, included in the latent construct of ELTM, we used three different tasks: oral recall of auditory presented word list (hearing aids on), oral recall of textually/auditorily presented sentences, and oral recall of motorically executed imperatives like “comb your hair” or “tie your shoe laces.” All ELTM tests (free verbal recall but with different encoding instructions) were affected by hearing loss, and if anything, the highest simple correlation with hearing loss was the motorically encoded imperatives, also called Subject Performed Tasks. This tells us that representations in ELTM must be multimodal (as is RAMBPHO) and that impaired hearing can drive such counterintuitive results as the SPT data (e.g., Rugg et al., 2015).
This kind of pattern of results, that is, selectivity for memory systems and sensory modality/generality for encoding instruction, is not easily accounted for by a common cause model. However, since the study was cross-sectional, we still cannot be sure about causality. Nevertheless, it is in line with the ELU prediction of disuse of a multimodal ELTM system that it is driven by hearing loss. Furthermore, the data suggest that the memory system selectivity we observed is neither due to information degradation (e.g., Schneider et al., 2002) nor to consumption of attention due to a degraded auditory stimulus (e.g., Verhaegen et al., 2014) because then the auditory-only encoding would have suffered more relative to the other encoding conditions (Rönnberg et al., 2011). The effect is presumably not dependent on changes in the phonological/lexical structures of SLTM, because the hearing loss memory system selectivity remained after partialling out for age in the SEM models (Rönnberg et al., 2011). SLTM structures like phonological neighborhoods would be expected to deteriorate with age (Neighborhood Activation Model; Luce & Pisoni, 1998; Sommers, 1996), but, again, the selective hearing loss effect on memory systems survived in the SEM analysis as a significant predictor variable (Rönnberg et al., 2011).
Furthermore, it was also the case in Rönnberg et al. (2014), using a very large sample of participants from the UK Biobank, that hearing loss was related to visual ELTM, yielding an effect size in the moderate range. This is a further argument with respect to the multimodality issue. Finally, a further study by Armstrong et al. (2020), using data from the Baltimore Longitudinal Study of Aging, demonstrated that hearing thresholds 2 years prior to testing predicted performance on the California Verbal Learning test (delayed audio-verbal recall), which is still another suggestion of a possible causal relationship between ARHL and ELTM.
However, that is the overall picture, and the more specific underlying mechanism as to why hearing loss may lead to dementia is still unclear (Hewitt, 2017; Livingston et al., 2017; Roberts & Allen, 2016; Wayne & Johnsrude, 2015). When it comes to functional ELU mechanisms, brain plasticity, and the behavioral antecedents of the effects of hearing loss, we advocate the following rationale: a relative use/disuse of memory systems view. In this context, this view claims that postdiction/reconstruction of misunderstood or misheard words is heavily dependent on daily WM activity (Rönnberg et al., 2014). WM is involved in reconstruction and repair of misheard words or sentences, day in, day out. However, even though WM is used many times during postdiction, WM reconstruction will not always be successful. Unsuccessful reconstruction will reduce the number of times that events, communicated words, and meanings can be encoded into ELTM. This will, as a consequence, reduce the number of times that ELTM will be used; the number of encodings of episodic traces and subsequent retrievals will be blocked or reduced when WM is in error. Therefore, ELTM with ARHL will deteriorate faster than ELTM without ARHL because of less usage and practice.
While WM is reconstructing and inferring knowledge from the bits and pieces of information decoded and stored for processing, SLTM will be called on for providing, for example, phonological constraints, meanings, and word knowledge to narrow down the intelligent guesswork that is needed to postdictively reconstruct what was said, for example, in the gating paradigm (Moradi et al. 2017, 2019, 2014). This means that WM will be compared to SLTM—in addition to retrieving stored representations of phonology, semantic meanings of words, knowledge of grammar, and objects—have to use the perceived items currently in mind, and combine those with the SLTM contributions retrieved online. Thus, WM has this inherent dual purpose of combining and inferring, while SLTM provides semantic support, and ELTM encoding is a consequence of the WM-SLTM interaction. So, on a use–disuse dimension, the general prediction is that the degree of deterioration of memory systems due to hearing loss is as follows, starting with the highest degree: ELTM > SLTM > WM.
One caveat here is about how old the person is and how well developed the cognitive system and its SLTM representations actually are. A mature and richly interconnected semantic representational system will probably interact with WM more and be less disused; a rich ELTM will probably have more SLTM connections in its memory representations, and so the use–disuse rank order of memory systems may be affected. However, in general, we submit that the order of memory systems decline is the one suggested above.
It is important to note that the basic prediction of the ELU model regarding the effect of hearing loss on memory systems is not dependent on the encoding modality per se but rather on the memory system as such. It seems to be the particular memory system that suffers, independent of encoding modality (e.g., motor, visual, or auditory). This has made us conclude that hearing impairment affects multimodal memory systems, not just single modality-specific, for example, auditory or visual short-term memory systems.
These results are illustrated by the recent studies on the effects of ARHL (see Rönnberg et al., 2011, 2019, 2014, 2013). While ELTM is the most fragile memory system, being susceptible to brain damage of different kinds, it is also the most advanced system and is the latest memory system to mature in ontogeny of a person (Tulving, 1983, 1985). It seems that memory systems obey the principle of last in–first out. We know that ELTM deficits may be indicative of mild cognitive impairment (e.g., Farias et al., 2017; Fortunato et al., 2016), sooner or later leading to dementia (Gallagher & Koh, 2011). We also know that hearing impairment increases the risk of dementia over a period of around 10 years (Lin et al., 2011, 2014; Livingston et al., 2017).
In addition, it seems to be the case that the source of, for example, Alzheimer's disease is connected to encoding problems generally (which of course affects the number of retrievals per day), with negative effects for LTM systems (Stamate et al., 2020). This would be in line with the ELU disuse prediction of LTM systems. It is also possible that at least some of the patients had hearing impairments, although the authors stated that the profoundly hearing-impaired patients were excluded. As we have seen, only minor impairments can lead to effects on memory and comprehension. This is of course speculative but would fit with the overall argument that hearing loss drives problems with encoding and subsequent retrievals from ELTM and SLTM systems, hence resulting in disused LTM systems.
If we combine these facts—and to assess a more stringent causality of the prediction—then one must conduct a study where we employ hearing impairment variables, cognitive and memory systems variables, as well as outcome variables. To be able to do this, and to causally model the impairment–cognition–outcome relationships, the study must also be a longitudinal one, with a sufficient number of years between test occasions and a sufficient number of participants enrolled. This is exactly what we are doing in Linköping right now, in the so-called n200 study (see Rönnberg et al., 2016, for a description of the test battery).
Some Final Comments About Future Research and Clinical Applications
-
We know by now that different kinds of signal processing in hearing instruments tax the cognitive system in different ways (e.g., Arehart et al., 2013, 2015; Foo et al., 2007; Lunner et al., 2009; Rudner et al., 2009; Zekveld et al., 2013). Some individuals gain less than they win from more advanced signal processing algorithms. Typically, individuals with high WMC tolerate and benefit from that kind of signal processing (Lunner et al., 2009; Lunner & Sundewall-Thorén, 2007). What is relatively new is that even for high positive signal-to-noise ratios, WMC modulates short-term retention of spoken materials, operationalized by, for example, the SWIR test (Ng et al., 2013; Souza et al., 2015, 2019). Future research should focus on task difficulty of the SWIR (see Micula et al., 2020) to investigate under what conditions signal processing off-loads WM. It is interesting to know in this context that even “within normal” levels of hearing acuity, cognitive and speech understanding performance may be negatively affected by small upward shifts in hearing thresholds, for example, comprehension of syntactically complex sentences will be exacerbated by such slight hearing losses (Ayasse et al., 2019; cf. Rudner et al., 2019).
-
As we have shown, WM in a visual n-back task dampens “early” subcortical “attention” mechanisms as a function of WM load, as early as at the brainstem level (Sörqvist et al., 2012; cf. Kraus & Chandrasekaran, 2010; Kraus et al., 2012). This illustrates the power of cognitive hearing. In the same vein, we have also demonstrated that attention at the cortical level, with the same n-back task, inhibits auditory temporal lobe functions, hence reflecting the brainstem data (Sörqvist et al., 2016). A further future experiment that builds on Sörqvist et al. (2016) would be to test for a double dissociation by having participants engage in an auditory n-back task, while as the visuospatial task one could, for example, count the number of capital (deviant) letters in a stream of presented letters, one by one, analogous to counting of deviant tones. The prediction would be a dampening of occipito-parietal parts of the brain. In both cases it is easy to imagine, for example, traffic situations that would be dangerous if one of the senses were to be dampened by cross-modal loading of the other cognitive sense modality.
-
Current attention research is investigating the possibilities of steering signal processing of hearing aids through capture of the electroencephalogram signals, not least through eye movements (i.e., watching the speaker you are talking to; see Alickovic et al., 2019). The basic problem is to how to decode appropriate signals from the brain while attending to a speaker. Future research aims at capturing different listening intentions with electroencephalogram signals (e.g., listening to comprehend, listening to remember, or selective listening for a particular cue).
-
In the original ELU model, no mechanism was assumed that could account for developing new SLTM representations. The D-ELU (Holmer et al., 2016a) is a step in the direction of accounting for developmental effects, but several questions remain. Future experimental research will focus on what can be represented and what cannot in SLTM. Which dynamic auditory events will be hard to develop new representations for (Ng & Rönnberg, 2019), and which are possible to develop? In addition, how do preexisting representations, contextual factors, and WMC interact in the establishment of new representations?
-
One major goal of our WM research is to develop a test that is predictive of the capacity to deal with online, adverse listening conditions, and at the same time measures the capacity to optimize transfer to ELTM. We submit that ELTM relating to the contents of a conversation held in noise is an important clinical marker of being able to focus on the contents of the conversation “here and now.”
Many other topics remain like, for example, a longitudinal study of the relationship between hearing parameters, cognitive abilities and different kinds of outcome variables, as well as WM intervention studies. These and other topics will be presented in papers to come.
Acknowledgments
This research was supported by Grant 2017-06092 from the Swedish Research Council as well as by the Linnaeus Centre HEAD, financed by the Swedish Research Council, awarded to Jerker Rönnberg as PI, which also funded the research reported for the Humes Keynote Opening Lecture, at the Aging and Speech Communication 2019 Conference, Tampa, Florida, and by FORTE: Swedish Research Council for Health, Working Life, and Welfare.
References
-
Akeroyd, M. A. (2008). Are individual differences in speech reception related to individual differences in cognitive ability? A survey of twenty experimental studies with normal and hearing-impaired adults.International Journal of Audiology, 47(Suppl. 2), S53–S71. https://doi.org/10.1080/14992020802301142 -
Alickovic, E., Lunner, T., Gustafsson, F., & Ljung, L. (2019). A tutorial on auditory attention identification methods.Frontiers in Neuroscience, 13. https://doi.org/10.3389/fnins.2019.00153 -
Amichetti, N. M., Stanley, R. S., White, A. G., & Wingfield, A. (2013). Monitoring the capacity of working memory: Executive control and listening effort.Memory & Cognition, 41, 839–849. https://doi.org/10.3758/s13421-013-0302-0 -
Anderson, S., T., White-Schwoch, T.,Parbery-Clark, A., & Kraus, N. (2013). A dynamic auditory-cognitive system supports speech-in-noise perception in older adults.Hearing Research, 300, 18–32. https://doi.org/10.1016/j.heares.2013.03.006 -
Arehart, K. H., Souza, P., Baca, R., & Kates, J. M. (2013). Working memory, age, and hearing loss: Susceptibility to hearing aid distortion.Ear and Hearing, 34(3), 251–260. https://doi.org/10.1097/AUD.0b013e318271aa5e -
Arehart, K. H., Souza, P., Kates, J., Lunner, T., & Pedersen, M. S. (2015). Relationship among signal fidelity, hearing loss, and working memory for digital noise Suppression.Ear and Hearing, 36(5), 505–516. https://doi.org/10.1097/AUD.0000000000000173 -
Arlinger, S., Lunner, T., Lyxell, B., & Pichora-Fuller, M. K. (2009). The emergence of cognitive hearing science.Scandinavian Journal of Psychology, 50(5), 371–384. https://doi.org/10.1111/j.1467-9450.2009.00753.x -
Armstrong, N. M., An, Y., Ferrucci, L., Deal, J. A., Lin, F. R., & Resmick, S. M. (2020). Temporal sequence of hearing impairment and cognition in the Baltimore longitudinal study of aging.The Journals of Gerontology: Series A, 75(3), 574–580. https://doi.org/10.1093/gerona/gly268 -
Ayasse, N., Penn, L., & Wingfield, A. (2019). Variations within normal hearing acuity andspeech comprehension: An exploratory study.American Journal of Audiology, 28(2), 369–375. https://doi.org/10.1044/2019_AJA-18-0173 -
Baddeley, A. (2012). Working memory: Theories, models, and controversies.Annual Review of Psychology, 63, 1–29. https://doi.org/10.1146/annurev-psych-120710-100422 -
Baddeley, A. D., & Hitch, G. J. (1974). Working memory.The Psychology of Learning and Motivation, 8, 47–89. https://doi.org/10.1016/S0079-7421(08)60452-1 -
Baddeley, A. D., Hitch, G. J., & Allen, R. J. (2019). From short-term store to multicomponent working memory: The role of the modal model.Memory & Cognition, 47, 575–588. https://doi.org/10.3758/s13421-018-0878-5 -
Baltes, P. B., & Lindenberger, U. (1997). Emergence of a powerful connection between sensory and cognitive functions across the adult lifespan: A new window to the study of cognitive aging?.Psychology and Aging, 12(1), 12–21. https://doi.org/10.1037/0882-7974.12.1.12 -
Besser, J., Koelewijn, T., Zekveld, A. A., Kramer, S. E., & Festen, J. M. (2013). How linguistic closure and verbal working memory relate to speech recognition in noise—A review.Trends in Amplification, 17(2), 75–93. https://doi.org/10.1177/1084713813495459 -
Cardin, V., Orfanidou, E., Rönnberg, J., Capek, C. M., Rudner, M., & Woll, B. (2013). Dissociating cognitive and sensory neural plasticity in human superior temporal cortex.Nature Communications, 4. Article number:1473. https://doi.org/10.1038/ncomms2463 -
Cardin, V., Rudner, M., De Oliveira, R. F., Andin, J., Su, M. T., Beese, L., Woll, B., & Rönnberg, J. (2018). The organization of working memory networks is shaped by early sensory experience.Cerebral Cortex, 28(10), 3540–3554. https://doi.org/10.1093/cercor/bhx222 -
Classon, E., Rudner, M., & Rönnberg, J. (2013). Working memory compensates for hearing related phonological processing deficit.Journal of Communication Disorders, 46(1), 17–29. https://doi.org/10.1016/j.jcomdis.2012.10.001 -
Craik, F. I. M. (1983). On the transfer of information from temporary to permanent memory.Philosophical Transactions of the Rocial Society B: Biological Sciences, 302(1110), 341–359. https://doi.org/10.1098/rstb.1983.0059 -
Davis, M. H., Johnsrude, I. S., Hervais-Adelman, A., Taylor, K., & McGettigan, C. (2005). Lexical information drives perceptual learning of distorted speech: Evidence from the comprehension of noise-vocoded sentences.Journal of Experimental Psychology: General, 134(2), 222–241. https://doi.org/10.1037/0096-3445.134.2.222 -
Daneman, M., & Carpenter, P. A. (1980). Individual differences in working memory and reading.Journal of Verbal Learning and Verbal Behavior, 19(4), 450–466. https://doi.org/10.1016/S0022-5371(80)90312-6 -
Daneman, M., & Merikle, P. M. (1996). Working memory and language comprehension: A meta-analysis.Psychonomic Bulletin and Review, 3(4), 422–433. https://doi.org/10.3758/BF03214546 -
Eriksson, J., Vogel, E. K., Lansner, A., Bergström, F., & Nyberg, L. (2015). Neurocognitive architecture of working memory.Neuron, 88(1), 33–46. https://doi.org/10.1016/j.neuron.2015.09.020 -
Farias, S. T., Lau, K., Harvey, D. J., Denny, K. G., Barba, C., & Mefford, A. N. (2017). Early functional limitations in cognitively normal older adults predict diagnostic conversion to mild cognitive impairment.Journal of the American Geriatrics Society, 65(6), 1152–1158. https://doi.org/10.1111/jgs.14835 -
Foo, C., Rudner, M., Rönnberg, J., & Lunner, T. (2007). Recognition of speech in noise with new hearing instrument compression release settings requires explicit cognitive storage and processing capacity.Journal of the American Academy of Audiology, 18(7), 618–631. https://doi.org/10.3766/jaaa.18.7.8 -
Fortunato, S., Forli, F., Guglielmi, V., Corso, D., E., Paludetti, G., Berrettini, S., & Fetoni, A. R. (2016). A review of new insights on the association between hearing loss and cognitive decline in ageing.Acta Otorhinolaryngologica Italica, 36, 155–166. https://doi.org/10.14639/0392-100X-993 -
Füllgrabe, C., & Rosen, S. (2016). On the (un)importance of working memory in speech-in-noise processing for listeners with normal hearing thresholds.Frontiers in Psychology, 7, 1268. https://doi.org/10.3389/fpsyg.2016.01268 -
Gallagher, M., & Koh, M. T. (2011). Episodic memory on the path to Alzheimer's disease.Current Opinion in Neurobiology, 21(6), 929–934. https://doi.org/10.1016/j.conb.2011.10.021 -
Gatehouse, S., Naylor, G., & Elberling, C. (2003). Benefits from hearing aids in relation to the interaction between the user and the environment.International Journal of Audiology, 42 (Suppl. 1), S77–S85. https://doi.org/10.3109/14992020309074627 -
Gatehouse, S., Naylor, G., & Elberling, C. (2006). Linear and nonlinear hearing aid fittings–1: Patterns of benefit.International Journal of Audiology, 45(3), 130–152. https://doi.org/10.1080/14992020500429518 -
Han, M. K., Storkel, H., & Bontempo, D. E. (2019). The effect of neighborhood density on children's word learning in noise.Journal of Child Language, 46(1), 153–169. https://doi.org/10.1017/S0305000918000284 -
Hannon, B., & Daneman, M. (2001). A new tool for measuring and understanding individual differences in the component processes of reading comprehension.Journal of Educational Psychology, 93(1), 103–128. https://doi.org/10.1037/0022-0663.93.1.103 -
Hewitt, D. (2017). Age-related hearing loss and cognitive decline: You haven't heard the half of it.Frontiers in Aging Neuroscience, 9, 112. https://doi.org/10.3389/fnagi.2017.00112 -
Holmer, E., Heimann, M., & Rudner, M. (2016a). Imitation, sign language skill and the developmental Ease of Language Understanding (D-ELU) model.Frontiers in Psychology, 7, 107. https://doi.org/10.3389/fpsyg.2016.00107 -
Holmer, E., Heimann, M., & Rudner, M. (2016b). Evidence of an association between sign language phonological awareness and word reading in deaf and hard-of-hearing children.Research in Developmental Disabilities, 48, 145–159. https://doi.org/10.1016/j.ridd.2015.10.008 -
Holmer, E., & Rudner, M. (2020). Developmental ease of language understanding model and literacy acquisition: Evidence from deaf and hard-of-hearing signing children.InQ. Y. Wang & J. F. Andrews (Eds.), Multiple paths to become literate: International perspective in deaf education. Gallaudet University Press. -
Humes, L. E. (2007). The contributions of audibility and cognitive factors to the benefit provided by amplified speech to older adults.Journal of the American Academy of Audiology, 18(7), 590–603. https://doi.org/10.3766/jaaa.18.7.6 -
Humes, L. E., Busey, T. A., Craig, J., & Kewley-Port, D. (2013). Are age-related changes in cognitive function driven by age-related changes in sensory processing?.Attention, Perception, & Psychophysics, 75, 508–524. https://doi.org/10.3758/s13414-012-0406-9 -
Kilman, L., Zekveld, A., Hällgren, M., & Rönnberg, J. (2014). The influence of non-native language proficiency on speech perception performance.Frontiers in Psychology, 5, 651. https://doi.org/10.3389/fpsyg.2014.00651 -
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills.Nature Reviews Neuroscience, 11, 599–605. https://doi.org/10.1038/nrn2882 -
Kraus, N., & White-Schwoch, T. (2015). Unraveling the biology of auditory learning: A cognitive-sensorimotor-reward framework.Trends in Cognitive Science, 19(11), 642–654. https://doi.org/10.1016/j.tics.2015.08.017 -
Kraus, N., Parbery-Clark, A., & Strait, D. L. (2012). Cognitive factors shape brain networks for auditory skills: Spotlight on auditory working memory.Annals of the New York Academy of Sciences, 1252(1), 100–107. https://doi.org/10.1111/j.1749-6632.2012.06463.x -
Lehmann, A., & Skoe, E. (2015). Robust encoding in the human auditory brainstem: Use it or lose it?.Frontiers in Neuroscience, 9, 451. https://doi.org/10.3389/fnins.2015.00451 -
Lin, F. R. (2011). Hearing loss and cognition among older adults in the United States.The Journals of Gerontology: Series A, 66A(10), 1131–1136. https://doi.org/10.1093/gerona/glr115 -
Lin, F. R., Ferrucci, L., An, Y., Goh, J. O., Doshi, J., Metter, E. J., Davatzikos, C., & Resnick, S. M. (2014). Association of hearing impairment with brain volume changes in older adults.NeuroImage, 90, 84–92. https://doi.org/10.1016/j.neuroimage.2013.12.059 -
Lin, F. R., Metter, E. J., O'Brien, R. J., Resnick, S. M., Zonderman, A. B., & Ferrucci, L. (2011). Hearing loss and incident dementia.Archives of Neurology, 68(2), 214–220. https://doi.org/10.1001/archneurol.2010.362 -
Livingston, G., Sommerlad, A., Orgeta, V., Costafreda, S. G., Huntley, J., Ames, D., Costafreda, S. G., Huntley, J., Ames, D., Ballard, C., Banerjee, S., Burns, A., Cohen-Mansfield, J., Cooper, C., Fox, N., Gitlin, L. N., Howard, R., Kales, H. C., Larson, E. B., Ritchie, K., Rockwood, K., Sampson, E. L., … Mukadam, N. (2017). Dementia prevention, intervention, and care.Lancet, 390(10113), 2673–2734. https://doi.org/10.1016/S0140-6736(17)31363-6 -
Luce, P. A., & Pisoni, D. B. (1998). Recognizing spoken words: The neighborhood activation model.Ear and Hearing, 19(1), 1–36. https://doi.org/10.1097/00003446-199802000-00001 -
Lunner, T. (2003). Cognitive function in relation to hearing aid use.International Journal of Audiology, 42(Suppl. 1), 49–58. https://doi.org/10.3109/14992020309074624 -
Lunner, T., Rudner, M., & Rönnberg, J. (2009). Cognition and hearing aids.Scandinavian Journal of Psychology, 50(5), 395–403. https://doi.org/10.1111/j.1467-9450.2009.00742.x -
Lunner, T., & Sundewall-Thorén, E. (2007). Interactions between cognition, compression, and listening conditions: Effects on speech-in-noise performance in a 2-channel hearing aid.Journal of the American Academy of Audiology, 18(7), 604–617. https://doi.org/10.3766/jaaa.18.7.7 -
Lyxell, B., & Rönnberg, J. (1987). Guessing and speechreading.British Journal of Audiology, 21(1), 13–20. https://doi.org/10.3109/03005368709077769 -
Lyxell, B., & Rönnberg, J. (1989). Information-processing skills and speechreading.British Journal of Audiology, 23(4), 339–347. https://doi.org/10.3109/03005368909076523 -
Mattys, S. L., Davis, M. H., Bradlow, A. R., & Scott, S. (2012). Speech recognition in adverse conditions: A review.Language and Cognitive Processes, 27(7–8), 953–978. https://doi.org/10.1080/01690965.2012.705006 -
McCreery, R. W., Walker, E. A., Spratford, M., Lewis, D., & Brennan, M. (2019). Auditory, cognitive, and linguistic factors predict speech recognition in adverse listening conditions for children with hearing loss.Frontiers in Neuroscience, 13, 1093. https://doi.org/10.3389/fnins.2019.01093 -
Micula, A., Ng, E. H., El-Azm, F., & Rönnberg, J. (2020). The effects of task difficulty, background noise and noise reduction on recall.International Journal of Audiology. https://doi.org/10.1080/14992027.2020.1771441 -
Molloy, K. T., Griffiths, D., Chait, M., & Lavie, N. (2015). Inattentional deafness: Visual load leads to time-specific suppression of auditory evoked responses.Journal of Neuroscience, 35(49), 16046–16054. https://doi.org/10.1523/JNEUROSCI.2931-15.2015 -
Moradi, S., Lidestam, B., Ng, E., Danielsson, H., & Rönnberg, J. (2017). Visual cues contribute differentially in audiovisual perception of consonants and vowels in improving recognition and reducing cognitive demands in listeners with hearing impairment using hearing aids.Journal of Speech, Language, and Hearing Research, 60(9), 2687–2703. https://doi.org/10.1044/2016_JSLHR-H-16-0160 -
Moradi, S., Lidestam, B., Ng, E. H. N., Danielsson, H., & Rönnberg, J. (2019). Perceptual doping: An audiovisual facilitation effect on auditory speech processing, from phonetic feature extraction to sentence identification in noise.Ear and Hearing, 40(2), 312–327. https://doi.org/10.1097/AUD.0000000000000616 -
Moradi, S., Lidestam, B., Saremi, A., & Rönnberg, J. (2014). Gated auditory speech perception: Effects of listening conditions and cognitive capacity.Frontiers in Psychology, 5, 531. https://doi.org/10.3389/fpsyg.2014.00531 -
Ng, E. H. N., Classon, E., Larsby, B., Arlinger, S., Lunner, T., Rudner, M., & Rönnberg, J. (2014). Dynamic relation between working memory capacity and speech recognition in noise during the first six months of hearing aid use.Trends in Hearing, 18. https://doi.org/10.1177/2331216514558688 -
Ng, E. H. N., & Rönnberg, J. (2019). Hearing aid experience and background noise affect the robust relationship between working memory and speech recognition in noise.International Journal of Audiology, 59(3), 208–218. https://doi.org/10.1080/14992027.2019.1677951 -
Ng, E. H. N., Rudner, M., Lunner, T., Pedersen, M. S., & Rönnberg, J. (2013). Effects of noise and working memory capacity on memory processing of speech for hearing-aid users.International Journal of Audiology, 52(7), 433–441. https://doi.org/10.3109/14992027.2013.776181 -
Ng, E. H. N., Rudner, M., Lunner, T., & Rönnberg, J. (2015). Noise reduction improves memory for target language speech in competing native but not foreign language speech.Ear and Hearing, 36(1), 82–91. https://doi.org/10.1097/AUD.0000000000000080 -
Nilsson, L.-G., Bäckman, L., Erngrund, K., Nyberg, L., Adolfsson, R., Bucht, Karlsson, S., Widing, M., & Winblad, B. (1997). The Betula prospective cohort study: Memory, health, and aging.Aging, Neuropsychology, and Cognition, 4(1), 1–32. https://doi.org/10.1080/13825589708256633 -
Peelle, J. E., Troiani, V., Grossman, M., & Wingfield, A. (2011). Hearing loss in older adults affects neural systems supporting speech comprehension.Journal of Neuroscience, 31(35), 12638–12643. https://doi.org/10.1523/JNEUROSCI.2559-11.2011 -
Peelle, J. E., & Wingfield, A. (2016). The neural consequences of age-related hearing loss.Trends in Neuroscience, 39(7), 486–497. https://doi.org/10.1016/j.tins.2016.05.001 -
Pichora-Fuller, M. K., Kramer, S. E., Eckert, M. A., Edwards, B., Hornsby, B. W. Y., Humes, L., Lemke, U., Lunner, T., Matthen, M., Mackersie, C. L., Naylor, G., Phillips, N. A., Richter, M., Rudner, M., Sommers, M., Tremblay, K. L., & Wingfield, A. (2016). Hearing impairment and cognitive energy: The Framework for Understanding Effortful Listening (FUEL).Ear and Hearing, 37(Suppl. 1), S5–S27. https://doi.org/10.1097/AUD.0000000000000312 -
Renoult, L., Irish, M., Moscovitch, M., & Rugg, M. D. (2019). From knowing to remembering: The semantic-episodic distinction.Trends in Cognitive Sciences, 23(12), 1041–1057. https://doi.org/10.1016/j.tics.2019.09.008 -
Roberts, K. L., & Allen, H. A. (2016). Perception and cognition in the ageing brain: A brief review of the short- and long-term links between perceptual and cognitive decline.Frontiers in Aging Neuroscience, 8, 39. https://doi.org/10.3389/fnagi.2016.00039 -
Rosemann, S., & Thiel, C. M. (2018). Audio-visual speech processing in age-related hearing loss: Stronger integration and increased frontal lobe recruitment.NeuroImage, 175, 425–437. https://doi.org/10.1016/j.neuroimage.2018.04.023 -
Rönnberg, M. (1990). Cognitive and communicative function: The effects of chronological age and “handicap age.”.European Journal of Cognitive Psychology, 2(3), 253–273. https://doi.org/10.1080/09541449008406207 -
Rönnberg, M. (1993). Cognitive characteristics of skilled tactiling: The case of GS.European Journal of Psychology, 5(1), 19–33. https://doi.org/10.1080/09541449308406512 -
Rönnberg, J. (2003). Cognition in the hearing impaired and deaf as a bridge between signal and dialogue: A framework and a model.International Journal of Audiology, 42(Suppl. 1), S68–S76. https://doi.org/10.3109/14992020309074626 -
Rönnberg, J., Andersson, J., Andersson, U., Johansson, K., Lyxell, B., & Samuelsson, S. (1998). Cognition as a bridge between signal and dialogue: Communication in the hearing impaired and deaf.Scandinavian Audiology, 27(4), 101–108. https://doi.org/10.1080/010503998420720 -
Rönnberg, J., Danielsson, H., Rudner, M., Arlinger, S., Sternäng, O., Wahlin, Å., & Nilsson, L.-G. (2011). Hearing loss is negatively related to episodic and semantic long-term memory but not to short-term memory.Journal of Speech, Language, and Hearing Research, 54(2), 705–726. https://doi.org/10.1044/1092-4388(2010/09-0088) -
Rönnberg, J., Holmer, E., & Rudner, M. (2019). Cognitive hearing science and ease of language understanding.International Journal of Audiology, 58(5), 247–261. https://doi.org/10.1080/14992027.2018.1551631 -
Rönnberg, J., Hygge, S., Keidser, G., & Rudner, M. (2014). The effect of functional hearing loss and age on long- and short-term visuospatial memory: Evidence from the UK Biobank resource.Frontiers in Aging Neuroscience, 6, 326. https://doi.org/10.3389/fnagi.2014.00326 -
Rönnberg, J., Lunner, T., Ng, E. H. N., Lidestam, B., Zekveld, A. A., Sörqvist, P., Lyxell, B., Träff, U., Yumba, W., Classon, E., Hällgren, M., Larsby, B., Signoret, C., Pichora-Fuller, M. K., Danielsson, H., & Stenfelt, S. (2016). Hearing impairment, cognition and speech understanding: Exploratory factor analyses of a comprehensive test battery for a group of hearing aid users, the n200 study.International Journal of Audiology, 55(11), 623–642. https://doi.org/10.1080/14992027.2016.1219775 -
Rönnberg, J., Lunner, T., Zekveld, A. A., Sörqvist, P., Danielsson, H., Lyxell, B., Dahlström, Ö., Signoret, C., Stenfelt, S., Pichora-Fuller, M. K., & Rudner, M. (2013). The Ease of Language Understanding (ELU) model: Theoretical, empirical, and clinical advances.Frontiers in Systems Neuroscience, 7, 31. https://doi.org/10.3389/fnsys.2013.00031 -
Rönnberg, J., Rudner, M., Foo, C., & Lunner, T. (2008). Cognition counts: A working memory system for Ease of Language Understanding (ELU).International Journal of Audiology, 47(Suppl. 2), S99–S105. https://doi.org/10.1080/14992020802301167 -
Rönnberg, J., Rudner, M., & Ingvar, M. (2004). Neural correlates of working memory for sign language.Cognitive Brain Research, 20(2), 165–182. https://doi.org/10.1016/j.cogbrainres.2004.03.002 -
Rönnberg, J., Rudner, M., Lunner, T., & Zekveld, A. A. (2010). When cognition kicks in: Working memory and speech understanding in noise.Noise and Health, 12(49), 263–269. https://doi.org/10.4103/1463-1741.70505 -
Rudner, M. (2018). Working memory for linguistic and non-linguistic manual gestures: Evidence, theory, and application.Frontiers in Psychology, 9, 679. https://doi.org/10.3389/fpsyg.2018.00679 -
Rudner, M., Foo, C., Rönnberg, J., & Lunner, T. (2009). Cognition and aided speech recognition in noise: Specific role for cognitive factors following nine-week experience with adjusted compression settings in hearing aids.Scandinavian Journal of Psychology, 50(5), 405–418. https://doi.org/10.1111/j.1467-9450.2009.00745.x -
Rudner, M., Foo, C., Sundewall-Thorén, E., Lunner, T., & Rönnberg, J. (2008). Phonological mismatch and explicit cognitive processing in a sample of 102 hearing aid users.International Journal of Audiology, 47(Suppl. 2), S163–S170. https://doi.org/ -
Rudner, M., Fransson, P., Ingvar, M., Nyberg, L., & Rönnberg, J. (2007). Neural representation of binding lexical signs and words in the episodic buffer of working memory.Neuropsychologia, 45(10), 2258–2276. https://doi.org/10.1016/j.neuropsychologia.2007.02.017 -
Rudner, M., Lunner, T., Behrens, T., Sundewall Thorén, E., & Rönnberg, J. (2012). Working memory capacity may influence perceived effort during aided speech recognition in noise.Journal of the American Academy of Audiology, 23(8), 577–589. https://doi.org/10.3766/jaaa.23.7.7 -
Rudner, M., Rönnberg, J., & Lunner, T. (2011). Working memory supports listening in noise for persons with hearing impairment.Journal of the American Academy of Audiology, 22(3), 156–167. https://doi.org/10.3766/jaaa.22.3.4 -
Rudner, M., Seeto, M., Keidser, G., Johnson, B., & Rönnberg, J. (2019). Poorer speech reception threshold in noise is associated with reduced brain volume in auditory and cognitive processing regions.Journal of Speech, Language, and Hearing Research, 62(4S), 1117–1130. https://doi.org/10.1044/2018_JSLHR-H-ASCC7-18-0142 -
Rugg, M. D., Johnson, J. D., & Uncapher, M. R. (2015). Encoding and retrieval in episodic memory: Insights from fMRI.InD. R. Addis, M. Barense, & A. Duarte (Eds.), The Wiley Handbook on the Cognitive Neuroscience of Memory (pp. 84–107). Wiley. https://doi.org/10.1002/9781118332634.ch5 -
Samuelsson, S., & Rönnberg, J. (1993). Implicit and explicit use of scripted constraints in lipreading.European Journal of Cognitive Psychology, 5(2), 201–233. https://doi.org/10.1080/09541449308520116 -
Schneider, B. A., Daneman, M., & Pichora-Fuller, M. K. (2002). Listening in aging adults: From discourse comprehension to psychoacoustics.Canadian Journal of Experimental Psychology, 56(3), 139–152. https://doi.org/10.1037/h0087392 -
Sharma, A., & Glick, H. (2016). Cross-modal re-organization in clinical populations with hearing loss.Brain Sciences, 6(1), 4. https://doi.org/10.3390/brainsci6010004 -
Signoret, C., & Rudner, M. (2019). Hearing impairment and perceived clarity of predictable speech.Ear and Hearing, 40(5), 1140–1148. https://doi.org/10.1097/AUD.0000000000000689 -
Sommers, M. S. (1996). The structural organization of the mental lexicon and its contribution to age-related declines in spoken-word recognition.Psychology and Aging, 11(2), 333–341. https://doi.org/10.1037//0882-7974.11.2.333 -
Souza, P., Arehart, K. H., Schoof, T., Anderson, M., Strori, D., & Balmert, L. (2019). Understanding variability in individual response to hearing aid signal processing in wearable hearing aids.Ear and Hearing, 40(6), 1280–1292. https://doi.org/10.1097/AUD.0000000000000717 -
Souza, P., Arehart, K. H., Shen, J., Anderson, M., & Kates, J. M. (2015). Working memory and intelligibility of hearing-aid processed speech.Frontiers in Psychology, 6, 526. https://doi.org/10.3389/fpsyg.2015.00526 -
Souza, P., & Sirow, L. (2014). Relating working memory to compression parameters in clinically fit hearing aids.American Journal of Audiology, 23(4), 394–401. https://doi.org/10.1044/2014_AJA-14-0006 -
Sörqvist, P., & Rönnberg, J. (2012). Episodic long-term memory of spoken discourse masked by speech: What is the role for working memory capacity?.Journal of Speech, Language, and Hearing Research, 55(1), 210–218. https://doi.org/10.1044/1092-4388(2011/10-0353) -
Sörqvist, P., Stenfelt, S., & Rönnberg, J. (2012). Working memory capacity and visual-verbal cognitive load modulate auditory-sensory gating in the brainstem: Toward a unified view of attention.Journal of Cognitive Neuroscience, 24(11), 2147–2154. https://doi.org/10.1162/jocn_a_00275 -
Sörqvist, P., Dahlström, Ö., Karlsson, T., & Rönnberg, T. (2016). Concentration: The neural underpinnings of how cognitive load shields against distraction.Frontiers in Human Neuroscience, 10, 221. https://doi.org/10.3389/fnhum.2016.00221 -
Stamate, A., Logie, R. H., Baddeley, A. D., & Sala, S. D. (2020). Forgetting in Alzheimer's disease: Is it fast? Is it affected by repeated retrieval?.Neuropsychologia, 138, 107351. https://doi.org/10.1016/j.neuropsychologia.2020.107351 -
Stenfelt, S., & Rönnberg, J. (2009). The signal-cognition interface: Interactions between degraded auditory signals and cognitive processes.Scandinavian Journal of Psychology, 50(5), 385–393. https://doi.org/10.1111/j.1467-9450.2009.00748.x -
Torkildsen, J. V. K., Hitchins, A., Myhrum, M., & Wie, O. B. (2019). Speech-in-noise perception in children with cochlear implants, hearing aids, developmental language disorder and typical development: The effects of linguistic and cognitive abilities.Frontiers in Psychology, 10, 2530. https://doi.org/10.3389/fpsyg.2019.02530 -
Tulving, E. (1983). Elements of episodic memory. Oxford University Press. -
Tulving, E. (1985). Memory and consciousness.Canadian Psychology/Psychologie Canadienne, 26(1), 1–12. https://doi.org/10.1037/h0080017 -
Verhaegen, C., Collette, F., & Majerus, S. (2014). The impact of aging and hearing status on verbal short-term memory. Neuropsychology, Development, and Cognition.Aging, Neuropsychology, and Cognition, 21, 464–482. https://doi.org/10.1080/13825585.2013.832725 -
Walker, E. A., Sapp, C., Oleson, J. J., & McCreery, R. W. (2019). Longitudinal speech recognition in noise in children: Effects of hearing status and vocabulary.Frontiers in Psychology, 10, 2421. https://doi.org/10.3389/fpsyg.2019.02421 -
Wass, M., Anmyr, L., Lyxell, B., Östlund, E., Karltorp, E., & Löfkvist, U. (2019). Predictors of reading comprehension in children with cochlear implants.Frontiers in Psychology, 10, 2155. https://doi.org/10.3389/fpsyg.2019.02155 -
Wayne, R. V., & Johnsrude, I. S. (2015). A review of causal mechanisms underlying the link between age-related hearing loss and cognitive decline.Ageing Research Reviews, 23(Pt B), 154–166. https://doi.org/10.1016/j.arr.2015.06.002 -
Wingfield, A., Amichetti, N. M., & Lash, A. (2015). Cognitive aging and hearing acuity: Modeling spoken language comprehension.Frontiers in Psychology, 6, 684. https://doi.org/10.3389/fpsyg.2015.00684 -
Zekveld, A. A., Rudner, M., Johnsrude, I. S., Festen, J. M., van Beek, J. H. M., & Rönnberg, J. (2011). The influence of semantically related and unrelated text cues on the intelligibility of sentences in noise.Ear and Hearing, 32(6), e16–e25. https://doi.org/10.1097/AUD.0b013e318228036a -
Zekveld, A. A., Rudner, M., Johnsrude, I. S., Heslenfeld, D., & Rönnberg, J. (2012). Behavioral and fMRI evidence that cognitive ability modulates the effect of semantic context on speech intelligibility.Brain and Language, 122(2), 103–113. https://doi.org/10.1016/j.bandl.2012.05.006 -
Zekveld, A. A., Rudner, M., Johnsrude, I. S., & Rönnberg, J. (2013). The effects of working memory capacity and semantic cues on the intelligibility of speech in noise.Journal of the Acoustical Society of America, 134(3), 2225–2234. https://doi.org/10.1121/1.4817926


Add to your Mendeley library