Keywords
Repeat-Recall Test - semantic context use - directional microphone - realistic signal-to-noise ratios
In a previous paper, we reported on the use of the Repeat-Recall Test (RRT) as an integrative tool to examine the efficacy of a directional microphone (DIRM) and a noise reduction (NR) algorithm.[1] We examined how the signal-to-noise ratios (SNRs) of the environment and the working memory capacities (WMCs) of the listeners affected the efficacy of these two features on outcome measures of speech intelligibility (repeat), word/sentence retention (recall), and ratings of listening effort and tolerable time. We showed that the noted efficacy interacted with SNR, WMC, and passage context. Specifically, while all participants benefited from the use of the DIRM on the repeat task, participants in the good WMC group received more DIRM benefit at the poorer SNRs and no benefit at the SNR of 15 dB while those in the poorer WMC group showed slightly less benefit at the poorer SNR but such benefit continued to a SNR of 15 dB for the low context (LC) materials. Furthermore, those in the poorer WMC group benefited from NR on rating of listening effort. Space limitations did not permit us to explore how the use of context was affected by the study parameters. We report on how context use (CU) is affected in this article.
Speech comprehension involves both bottom-up and top-down processes. Bottom-up processes include factors that affect stimulus audibility such as room acoustics, SNRs, and hearing losses. Top-down processes include factors that facilitate stimulus comprehension. They could include the cognitive capacity of the listener,[2]
[3]
[4] knowledge of the language, or the listeners' ability to use context, among others.
Contextual cues refer to any and all social, physical, visual, tactile, linguistic, and/or semantic information that a listener might use to gain communication success. Contextual cues might increase the speed and/or accuracy of speech identification and free up cognitive resources for storage and processing of the intended communication. In turn, this may decrease the perceived effort associated with communication. Although some have suggested that context contribution increases with the difficulty of the listening situation,[5]
[6] it is not immediately clear how SNRs representative of real-world conditions affect CU on different tasks.
Early studies on the effects of semantic context [5]
[7]
[8] used the Speech Perception in Noise (SPIN) test.[9]
[10]
[11] The SPIN test quantifies context effects by comparing intelligibility scores for sentence-final words between sentences where said words are either predictable (e.g., He is sleeping on the bed) or unpredictable (e.g., He is going to buy the bed) based on the sentence context. Other methods and materials have also been used to study context effects. For example, Boothroyd and Nittrouer[12] created their own high and low probability sentences. Helfer and Freyman[13] reported that providing knowledge of sentence topic improved the perception in noise. Zekveld et al[4] asked subjects to generate text cues and examined their effects on the intelligibility of natural sentences in noise. Guediche et al[14] manipulated voice onset time to produce ambiguous and unambiguous target word stimuli (goat and coat) and investigated the effects of prior sentence context on phonetic perception of the target words. Indeed, many manipulations that may affect the top-down processing of target stimuli could serve as context.
Study of semantic context effect is not limited to speech intelligibility task. Many researchers have shown that semantic context also improved sentence retention and recall.[3]
[4]
[15]
[16]
[17] Holmes et al[18] reported that listeners rated semantically congruent sentences on the Connected Speech Test[19] as less effortful compared with semantically incongruent sentences. Similarly, Winn[20] reported that semantic context reduced listening effort as evaluated by pupillometry.
Together, these studies support the benefits that context adds to speech understanding, recall, and listening effort. However, speech understanding tasks have different functional and cognitive requirements than recall tasks which are yet different than those required for rating of listening effort. Hence, the SNR(s) at which CU is maximal may differ across the different evaluative criteria or outcome measures. Furthermore, differences in the cognitive requirements of each measure suggest that the WMC of listeners may also modulate CU. For example, a hearing aid (HA) feature that provides slight SNR improvements might improve speech-in-noise performance but no improvement in listening effort. In such a scenario, wearers may still be dissatisfied with the performance of the HAs. A study that systematically examines how CU changes with SNR for different evaluative criteria may provide a better understanding of factors affecting CU in listeners with different cognitive capacities and offer guidance for the future design and selection of HAs as well as patient counseling.
For the average listener, the overall sound level and the SNRs of the listening environment determine the listening difficulty. For a hearing-impaired listener, the degree of hearing loss and HA status could also affect the difficulty of the listening situation. For example, a hearing-impaired listener would hopefully (but not always) find the listening situation less difficult when aided than unaided. The type of technology within the HA, including the use of NR and DIRM, could further affect the listener's difficulty in the listening situation. This could, in turn, affect the degree to which context might alleviate listening difficulty. For example, DIRMs reportedly improved the SNRs of the listening environment by 1 to 6 dB.[21] While there is limited evidence to support that NR algorithms improve SNR, they have been shown to reduce listening effort.[1]
[17] Thus, the use of processing features on a HA could change CU across realistic SNRs.
When studying context effects, it is important that the chosen speech materials minimize variables that may bias the observed effect. For example, Zekveld et al[4] criticized the SPIN in that the SNR of the sentence context was the same as the target word. Thus, the audibility of the context cues was not assured at SNRs where the audibility of the target words was questionable. Other lexical factors such as word frequency and familiarity, phonological similarity, or age of word acquisition could affect the use of context.[22] Moulin and Richard[23] reported that spondees that occurred more frequently provided more contextual information than spondees that occurred less frequently. Thus, to best reflect true context effects, high (HC) and LC speech materials should match their frequency of occurrence and word difficulty. Furthermore, the syntactic structure of the materials with and without context should also be similar (if not identical) to minimize bias. These issues complicated the use of the SPIN to study context effects because high and low probability SPIN sentences are not identical in syntactical structure or word familiarity.
We developed the RRT as an integrated speech test that allows for study of context effects across several outcome measures.[24] This includes the listener's ability to (1) repeat sentences in quiet and in noise, (2) retain and recall those sentences, (3) rate the perceived listening effort required of the test conditions, and (4) judge their willingness to stay engaged in conversation (referred to as “tolerable time”). The test uses high and LC sentences presented under SNRs of 0, 5, 10, 15 dB, and quiet that are representative of real-world communication conditions.[25]
[26] HC sentences are short meaningful 6 to 8 word sentences each with 3 to 4 target words. These are grouped into 6-sentence lists, where sentences within a list are related to a theme (e.g., food). LC sentences are then made by rearranging target words within a HC list such that the resulting list of 6 sentences are still syntactically similar or identical to the original HC sentences, but semantically meaningless. Thus, the meaningfulness of the sentences is used as a context to help listeners identify the target words. CU is calculated as a difference in target word scores between the HC and LC sentences. An example of a list of complementary HC and LC sentences is shown in [Appendix A].
Appendix A
Example of a list of high context (HC) and low context (LC) sentences
High context (HC) sentences
|
Keep the ice cream in the freezer
|
The chef cooks food in a restaurant
|
The barbecue grill used hickory wood
|
Wash the fruit in the sink.
|
The tart pie had too much lemon
|
He tried new foods in different countries
|
Low context (LC) sentences
|
Keep the ice foods in the lemon
|
The cream cooks food in a country
|
The barbecue chef used hickory freezer
|
Wash the grill in the restaurant
|
The tart fruit had too much wood
|
He tried new pie in different sinks
|
Note: The bolded words are the target words.
This manner of defining context has several advantages. First, the same words are used in both HC and LC materials; this minimizes any issues with word familiarity, frequency of occurrence, and/or difficulty. Second, because both versions use the same words, the long-term spectra of complementary HC and LC materials are similar. This helps to control possible confounds related to word audibility. Third, the same sentence structure is used for both HC and LC materials, which minimizes syntactical biases. On the other hand, because the target words and the rest of the sentence are presented at the same SNR, the audibility of the contextual cues is necessarily tied to the audibility of the target words.
In this study, we wanted to use the different outcome measures (i.e., repeat, recall, listening effort, and tolerable time) on the RRT to study changes in CU across a range of realistic SNRs. In addition, we wanted to examine if CU depends on HA features such as DIRMs and NR and/or on the WMC of the listener. Answers to these questions will allow one to know (1) if the RRT can be used to study context effects, (2) how CU changes for each RRT outcome measure, (3) how HA technology influences CU, and (4) if WMC affects how much context is used.
Methods
The readers are referred to Kuk et al[1] for a detailed description of the Methods. A brief summary of the study details is reported here.
Participants
Nineteen hearing-impaired adults (average age of 73.6 years) with a bilaterally symmetrical mild-to-moderately severe sensorineural hearing loss and normal cognition participated ([Fig. 1]). Informed consent was obtained from all participants in accordance with protocols approved by an external institutional review board.
Fig. 1 Average pure-tone thresholds for the left (black exes) and right (gray circles) ears of 19 hearing-impaired listeners. Error bars represent 1 standard deviation.
Hearing Aid Conditions
Participants completed all testing in the aided mode with bilaterally fitted receiver-in-canal HAs coupled to fully occluding “double-dome” instant-fit ear-tips. The NAL-NL2[27] fitting target was used and all fittings were verified for adequate audibility using the SoundTracker feature of the fitting software.[28] The fully adaptive beamformer was set to a fixed hypercardioid mode during testing. When activated, the modulation-based NR algorithm reshaped the frequency response to optimize the speech intelligibility index with a maximum gain reduction of 12 dB and maximum gain increase of 4 dB in the mid frequencies. Four combinations of microphone and NR conditions were evaluated: omni-DIRM with NR enabled (OMNI.NR.ON); omni-DIRM with NR disabled (OMNI.NR.OFF); DIRM with NR enabled (DIRM.NR.ON); and DIRM with NR disabled (DIRM.NR.OFF).
Test Materials and Procedure
The study followed a double-blind within-subjects design. Subject performance on the RRT was evaluated using a different list for each HA condition. Listeners first repeated the sentence that they heard. After all 6 sentences within a list were repeated, listeners recalled as many of the sentences (or target words) as they could recall. Afterwards, listeners rated the amount of perceived listening effort using a 1- to 10-point scale with “1” being “not effortful” and “10” being “extremely effortful.” Listeners then estimated the amount of time (in minutes with a minimum of less than 1 minute and a maximum of 2 hours) that they were willing to spend listening under the specific SNR condition. A practice trial at a SNR of 10 dB was completed. The LC sentences were always presented prior to the HC sentences.
Speech stimuli were delivered in the free-field at a fixed peak level of 75 dB sound pressure level 1 m from the front. A spectrally matched, continuous speech-shaped noise was presented 1 m directly behind the listener so both the DIRM and NR algorithm may be activated. Background noise was presented at fixed levels to produce SNRs of 0, 5, 10, and 15 dB in a random order.
Results
The recall score for HC sentences presented at a SNR of 15 dB was used to group listeners into good and poor WMC categories. This test condition was used because the repeat scores were ≥ 95% in all participants to assure audibility. Because two peaks (at 35 and 50%) were noted in the distribution of the recall scores, listeners with recall performance ≥ 43% were placed into the “good” WMC group and those with recall performance < 43% were placed into the “poor” WMC group. There were 10 participants in the good WMC group and 9 in the poor WMC group. Participants in both groups (good vs. poor) were similar in their ages (73 vs. 74 years), pure-tone averages (47 vs. 51 dB hearing level), and Montreal Cognitive Assessment scores [29] (27 vs. 26). The absolute scores for different test conditions were reported in the previous report [1] and detailed in the “Discussion” section. In this report, CU was the dependent variable and it was calculated as the difference between HC and LC sentence scores with the restriction that CU is not smaller than 0.
Repeated measures analyses of variance (ANOVAs) were conducted to assess the within-subjects factors of Microphone (2 levels, DIRM and OMNI), NR (2 levels, NR.ON and NR.OFF), SNR (4 levels, 0, 5, 10, and 15 dB), and WMC group (2 levels, good WMC and poor WMC) on CU separately for repeat, recall, listening effort, and tolerable time. Analyses assessed all interactions of these factors. Degrees of freedom were adjusted using Greenhouse–Geisser correction wherever the assumption of sphericity was violated. All ANOVAs were calculated using Type III sums of squares. The value of η
2 is reported to allow judgment of effect size. It has been suggested that η
2 values of 0.01, 0.09, and 0.25 may reflect small, medium, and large effect sizes, respectively.[30]
Repeat Performance
For repeat performance, CU was significantly affected by the main effect of SNR (F
(3,51) = 11.93, p < 0.001, η
2 = 0.12). This main effect was qualified by a two-way interaction between SNR and Microphone (F
(3,51) = 24.07, p < 0.001, η
2 = 0.17). These were medium size effects. [Fig. 2] compares CU between WMC groups for each Microphone condition. With OMNI processing, maximum CU was observed at SNR = 10 dB. With DIRM processing, maximum CU was observed at SNR = 5 dB, decreasing as SNR increased beyond that level. CU was also higher in the OMNI mode than in the DIRM mode at SNR ≥ 10 dB. There was no significant effect of the WMC group or NR.
Fig. 2 Context use (CU) for repeat performance across signal-to-noise ratios (SNRs) for people with good (solid line) and poor (dotted line) working memory capacities (WMCs) in the omnidirectional (OMNI, top) and directional mode (DIRM, bottom) mode.
The amount of CU exceeded 30% in some test conditions. When collapsed across all SNR and microphone conditions, CU was estimated to be approximately 4.5 dB at a speech reception threshold criterion of 75%. This is less than the 6.5 dB improvement offered by the use of the DIRM.[1] It is difficult to compare the magnitude of the context effect measured in this study with those of others[18] because of the differences in test materials.
Recall Performance
CU during recall was significantly affected by the main effects of WMC group (F
(1,17) = 24.39, p = 0.001, η
2 = 0.14), Microphone (F
(1,17) = 6.03, p = 0.025, η
2 = 0.02), and SNR (F
(3,51) = 16.70, p < 0.001, η
2 = 0.17). The effect size of WMC group and SNR was medium while that of Microphone was small. The effect of NR was not significant. These main effects were qualified by a significant Microphone × SNR interaction (F
(3,51) = 22.00, p < 0.001, η
2 = 0.14) with a medium effect size. [Fig. 3] compares CU for recall between participants in each microphone mode. The good WMC group benefitted more from context than the poor WMC group at all SNRs for both microphone modes. Listeners made more use of context in the DIRM versus the OMNI mode at poorer SNRs (0 and 5 dB). This pattern reversed at SNR = 10 dB. In addition, CU was relatively stable across SNRs in the DIRM mode but increased as SNR increased in the OMNI mode. These results support previous observations that context improves recall[4] and that people with better WMC show more CU than those with a poorer WMC.[3]
Fig. 3 Context use (CU) for recall performance across signal-to-noise ratios (SNRs) for people with good (solid line) and poor (dotted line) working memory capacities (WMCs) in the omnidirectional (OMNI, top) and directional mode (DIRM, bottom) mode.
Ratings of Listening Effort
CU in rating of perceived listening effort was affected by WMC group (F
(1,17) = 9.64, p = 0.006, η
2 = 0.11) and SNR (F
(3,51) = 22.65, p < 0.001, η
2 = 0.14). These effect sizes were medium. The effect of NR was not significant. These effects were further qualified by significant WMC group × SNR (F
(3,51) = 4.02, p < 0.012, η
2 = 0.03) and Microphone × SNR (F
(3,51) = 9.78, p < 0.001, η
2 = 0.05) interactions with a small effect size. [Fig. 4] compares CU between participants in each microphone mode. In general, good WMC listeners reported a greater reduction in perceived listening effort when processing HC versus LC sentences than did poor WMC listeners, except at SNR = 0 dB where the two groups did not differ. In addition, all listeners benefited more from context in the DIRM than in the OMNI mode at SNRs of 0 and 5 dB, but not at SNRs of 10 and 15 dB where CU was similar between the two microphone modes. For the DIRM, CU was relatively constant across SNRs, whereas for the OMNI, CU increased as SNR increased.
Fig. 4 Context use (CU) for perceived listening effort ratings across signal-to-noise ratios (SNRs) for people with good (solid line) and poor (dotted line) working memory capacities (WMCs) for the omnidirectional (OMNI, top) and directional mode (DIRM, bottom) mode.
Estimates of Tolerable Time
CU for tolerable time was affected by SNR (F
(3,51) = 8.81, p < 0.001, η
2 = 0.06) with significant two-way interactions between SNR and WMC group (F
(3,51) = 2.95, p = 0.041, η
2 = 0.02) and between WMC group and Microphone (F
(1.17) = 5.61, p = 0.030, η
2 = 0.02). These effect sizes were small. The effect of NR was not significant. [Fig. 5] compares CU between participants in each microphone mode. Context improved estimates of tolerable time as SNR increased; however, this effect was stronger in listeners with good WMC than those with poor WMC. Listeners with good WMC reported longer tolerable time from context in the DIRM (vs. OMNI) mode, whereas CU was unaffected by microphone mode in listeners with poor WMC.
Fig. 5 Context use (CU) for tolerable time ratings across signal-to-noise ratios (SNRs) for people with good (solid line) and poor (dotted line) working memory capacities (WMCs) for the omnidirectional (OMNI, top) and directional (DIRM, bottom) mode.
Discussion
The current study shows that the degree to which listeners use context depends on the interaction between the SNR of the environment, availability of a DIRM on the HA, and the WMC of the listeners. In addition, the pattern of CU may be different among the four outcome measures used on the RRT. Medium effect size was observed in most of the comparisons.
Starting at a poor SNR (i.e., 0 dB), CU increases with SNR until it reaches a maximum and then it either levels off (as observed for recall, listening effort, and tolerable time) or decreases (as observed of repeat) as SNR increases. This suggests that at SNRs of 0 or 5 dB, inaudibility of the speech signal limits the usability of any semantic cues. As SNR improves, some of the semantic cues become audible and contribute to improving performance for target words. Beyond a particular SNR, the audibility of the speech material is sufficient for target word identification even in the absence of semantic cues. Thus, CU decreases for the repeat task when audibility is the determining factor. On the other hand, CU remains the same when task performance is not solely dependent on audibility as observed for recall, listening effort, and tolerable time measures when the HAs were in the DIRM mode.
HA technology influenced the SNR at which CU was maximal. Use of a DIRM improved audibility and thus usability of the semantic cues even at a SNR of 0 dB. Conversely, CU was not observed in the OMNI condition until a SNR of 5 dB. Maximum CU occurred at SNR of 5 dB in the DIRM mode and 10 dB in the OMNI mode. A DIRM alters the effective SNR at the listener's ears, which increases the availability of usable context cues to the listener. The NR algorithm used in the current study, while improving listening effort,[1] did not influence CU on any of the RRT measures.
Knowing the lowest SNR where CU occurred may provide an estimate of the minimum internal SNR required for optimal performance. For a repeat task, this was the SNR favorable enough to make the available semantic cues audible and maximally usable. When the HA was in the OMNI mode, maximum CU was observed at SNR = 10 dB, suggesting this SNR may meet the minimal internal SNR requirement. The observation of maximum CU at SNR = 5 dB in the DIRM mode reinforced this speculation. This is because the 5 dB input SNR in the DIRM mode, when added to the 6.5 dB benefit from DIRM,[1] was equal to an effective input SNR of 11.5 dB (5 + 6.5 dB). Thus, the use of a DIRM is mandatory if listeners are to benefit from semantic context at input SNRs ≤ 5 dB. Otherwise, the input SNR must be > 10 dB to fully utilize context. Interestingly, Smeds et al[25] and Wu et al[26] observed that the realistic SNRs of listeners with a mild-to-moderate hearing loss peaked around 10 dB. The results of the current study raise the possibility that these listeners might have chosen environments where they can fully utilize semantic contextual cues.
Patterns of CU varied depending on the outcome measure. On the repeat measure, CU reached a maximum and then decreased as SNR increased. For the other tasks (recall, listening effort, and tolerable time), CU stayed at similar levels in the DIRM mode and increased in the OMNI mode as SNR increased. The WMC of the listeners did not affect CU on the repeat measure, whereas listeners with better WMC were able to use more context on measures including recall, listening effort, and tolerable time. This difference in CU patterns across outcome measures may have implications for the test conditions under which we examine context effects in aided hearing-impaired listeners. Maximal context effects were noted on the repeat measure at SNRs between 5 and 10 dB. However, on the recall, listening effort, and tolerable time measures, similar CU was seen across SNRs ≥ 5 dB in the DIRM mode and at SNRs ≥ 10 dB in the OMNI mode. This suggests that the SNR where an aided hearing-impaired listener makes the most use of context depends on the outcome measure. If speech intelligibility is used to examine CU, then SNRs should be < 10 dB. On the other hand, if the listener's tasks involve recall, rating of effort, or willingness to stay in noise, the required SNR would be higher. If used with a DIRM, an SNR between 5 and 10 dB may be adequate. However, if it were an OMNI mic, then an SNR between 10 and 15 dB may be required to observe any effects.
Previous studies have suggested that CU depends on the listeners' WMCs.[4] In this study, we observed CU during the repeat task to be similar between the good and poor WMC groups. However, listeners in the good WMC group showed more CU on recall, listening effort, and tolerable time measures. One possible explanation is that the cognitive capacities of all listeners in our sample were good enough to make use of contextual cues during the repeat task, but those in the good WMC group had additional cognitive spare capacity that could be directed at using contextual cues for encoding strategies and later retrieval. Spare capacity might also explain why listeners in the good WMC group found HC materials to be less effortful and more tolerable than LC materials at certain SNRs.
A Source of Difference in Context Use between WMC Groups
Because CU is a difference score between HC and LC sentences, a review of the absolute scores may provide additional insights into how listeners with good and poor WMC differed across measures. [Fig. 6] summarizes the absolute scores for each measure reported in Kuk et al,[1] averaged across SNRs and participants in each WMC group. For repeat, the good WMC listeners scored higher for both the HC and LC sentences than the poor WMC listeners; however, both groups were equally effective in utilizing context to help in speech understanding. Thus, CU was not different between WMC groups on the repeat task.
Fig. 6 High context (HC, solid line) and low context (LC, dotted line) sentence scores for good working memory capacities (WMC) and poor WMC listeners for the four outcome measures (repeat, recall, listening effort, and tolerable time) used on the Repeat-Recall Test (RRT).
The good WMC listeners again scored higher than the poor WMC listeners on both the HC and LC sentences on the recall measure. However, the difference between WMC groups was less with the LC sentences than the HC sentences. Thus, the good WMC listeners benefitted more from context than the poor WMC listeners to facilitate recall.
A different pattern emerged on the listening effort measure. Listeners in the good WMC group rated the HC sentences less effortful (8.5 for the good WMC vs. 9 for the poor WMC) and LC sentences more effortful (10.5 for the good WMC vs. 10 for the poor WMC) than the poor WMC listeners. Thus, a smaller difference in effort ratings between HC and LC materials was seen in the poor WMC listeners than the good WMC listeners. This resulted in greater CU in the good WMC group than the poor WMC group (2 vs. 1, a medium effect size). Intuitively, one would expect that the poor WMC listeners to rate the test conditions as more effortful than the good WMC listeners. This was indeed true for the HC sentences but not for the LC sentences.
Observations on the tolerable time (willingness to stay in noise) trended similarly as the listening effort ratings. Listeners in the good WMC group were willing to stay longer than the poor WMC listeners when HC materials were used (7 vs. 5.5 minutes) but less when LC materials were used (3 vs. 3.2 minutes). This resulted in greater CU in the good WMC group than the poor WMC group (4 vs. 2.3 minutes, a small effect size). This means that the meaningfulness of the message could increase the willingness of good WMC listeners to stay in a noisy situation but less so for those with poor WMC.
This finding may be related to the motivation of the listeners.[31] In a challenging condition, some listeners may have given up on the task and rated their effort for all test conditions similarly. Thus, the subjective ratings sampled at these conditions did not solely reflect the true difficulty of the task but were biased by the motivation, or lack thereof, of the listeners. Listeners in the poor WMC group may have perceived greater difficulties with the task and became more easily demotivated under some of the same test conditions than their good WMC peers. If so, this would suggest that listeners with poor WMC may be at a higher risk (than listeners with better WMC) of giving up in a communication task when it becomes difficult. The narrower range of effort ratings between HC and LC materials (i.e., CU) in the poor WMC listeners may suggest that these listeners have a smaller range of listening conditions where they may remain motivated. Kochkin[32] reported that HA wearers' satisfaction for their HAs correlated with the number of listening situations in which they were successful. Thus, it is not unreasonable to speculate that listeners with a poorer WMC are more likely to be dissatisfied with their HAs. For these listeners, it is important that they are provided HA technology that can expand the range of listening situations they are engaged in. Technologies such as DIRMs (which improve SNR and effort rating), adaptive sound classifiers (which adapt HA processing automatically based on acoustic analysis), or multiple programs (fixed set of different frequency gain characteristics) may be beneficial.
In summary, CU was similar between WMC groups on the repeat task, but smaller for the poor WMC group on the recall, listening effort, and tolerable time tasks. On the recall task, the smaller CU (in the poor WMC group vs. the good WMC group) was the result of a lower score on both the LC and HC sentences. On the other hand, the smaller CU in the poor WMC group on the listening effort and tolerable time tasks was the result of an “inflated” score on the LC sentences and a “deflated” score on the HC sentences (compare with the good WMC).
Conclusion
The current study supports the use of the RRT to evaluate CU. The study demonstrated that the amount of CU was a result of the interaction between the SNR of the test environment, the processing features on the HA, and the WMC of the listeners. In addition, the interaction of these factors on CU depends on the outcome measures used.