Impact of Mild to Moderate U-shape Hearing Loss on Listening Effort in School-aged Children with Alport Syndrome

DOI: https://doi.org/10.21203/rs.3.rs-732585/v1

Abstract

In this study, we aimed to investigate the effect of mild to moderate U-shape hearing loss on listening effort in Alport syndrome (AS) children by pupillometry. Subjects were required to answer questions after listening to conversations that simulate real scenes in daily life. We recorded the accuracy rate and pupil data of two conditions: SNR = + 15 dB and SNR = -2 dB. A mixed effect model was established to analyze the influence of SNR, mid-frequency energy proportion and hearing status on the accuracy and pupil response. The results showed SNR had a main effect on the accuracy. The baseline pupil diameter of AS children was always smaller than normal hearing children. When analyzing the time window including the stages of listening to the conversation, listening to the question and thinking, SNR and the hearing status had main effect on mean pupil dilation. We concluded that AS children with hearing loss were often in a state of low arousal before auditory task. Both hearing status and task difficulty have impact on listening effort of AS children. The effort of AS children with U-shape hearing loss might come mainly from consequent cognitive processing (as a consequence of effortful listening) instead of passive listening during speech communication.

1. Introduction

Alport syndrome (AS) is a rare hereditary disorder that is characterized by chronic kidney disease, sensorineural hearing loss and ocular abnormalities, resulting from mutations in the collagen IV genes[1]. Majority of the X-linked AS male patients experience mild to moderate sensorineural hearing loss, which is manifested in the middle frequency range[2]. Hearing loss mainly occurs at school age and gradually deteriorates with increasing age[2, 3]. Therefore, school-aged children with AS typically show a bilateral, symmetrical, mild to moderate, U-shape hearing loss (Fig-1). Their audiogram configuration almost located at the range of ‘Speech Banana’, which was widely applied to hearing aid fitting and rehabilitation in clinics to ensure speech information reception. Accordingly their speech recognition scores are usually good[3].

For school-aged children, successful communicating and learning in the educational environment requires a combination of perceiving, comprehending and inferring the message conveyed by speaker, which involves both speech reception and cognitive processing of the received information. These steps require constant allocation of attentional and cognitive resources[4]. However, classrooms are frequently noisy environments and not always optimal for listening. It is generally accepted that people with hearing impairment have difficulties in speech processing under acoustically challenging conditions, which may increase reliance on top-down cognitive processes for compensation and result in the experience of “effortful listening”[5, 6]. Thus, it is reasonable to speculate that AS children with hearing loss are at high risk of sustained effortful listening in classroom compared to their normally hearing peers.

This could lead further to significant negative impacts in school-aged children[7]. Children develop key cognitive, linguistic, and academic skills in the classroom, however, the increased mental demands of effortful listening may exert deleterious effects on the educational outcomes[8, 9]. Furthermore, it was shown in previous literatures that key aspects of auditory cognitive functions are impaired in school-aged children even though with mild hearing loss[10].Therefore, how and when to intervene is a significant concern for hearing-impaired children with AS. A better understanding of auditory cognitive ability in these population would be valuable to reduce listening effort and optimize learning performance.

Currently assessment and intervention options of these AS children always depend on auditory test. However, conventional audiological batteries that utilize pure-tone audiometry coupled with speech intelligibility tests seem to be insensitive to AS children with mild to moderate U-shape hearing loss due to ceiling effects. Moreover, traditional speech recognition tasks mainly provide information about the correct reception of what was heard, but failed to provide a clear evaluation of the “effort aspect” of a patient’s hearing experience[11]. Consequently, reliable clinical measure of listening effort could be a more comprehensive means of indexing this important dimension [12, 13].

Owing to a surge of interest in the study of auditory cognitive ability of children with hearing loss (CHL), an increasing number of studies have sought to measure listening effort of school-aged children by behavioral tests, physiological measures or self-report scales[7, 14, 15]. To date, however, none have used any methods including pupillometry to evaluate listening effort in school-aged children with a typical U-Shape hearing loss. For this group of hearing loss children, their audiograms and speech recognition scores all suggest successful speech communication. But it remains unclear whether they still experience high listening effort without hearing aids intervention.

Pupillometry, a new approach for analyzing changes in the size of the eye’s pupil, could provide a reliable and sensitive indication of listening effort[16]. The tight relationship between pupil diameter and mental effort has been well documented for over a century[1719]. Pupillometry enables the continuous recording of the pupil diameter, which could monitor moment-to-moment changes of cognitive load during listening task[20]. Changes in pupil diameter were more sensitivity and reliability than skin conductance, reaction time or subjective rating scales [21]. Peak pupil dilation (PPD) have repeatedly been verified to be sensitive to speech intelligibility level and masking type[22]. Wendt et al., found that PPD was highest when speech recognition performed in the presence of 1-talker masker in comparison to 4-talker babble and fluctuating noise masker[23]. Additionally, PPD changed in a nonlinear way with decreasing SNRs, reaching a maximum value until speech recognition performance was around 40–50%[24]. Mean pupil dilation (MPD) could be used to evaluate how cognitive load is sustained across a given auditory trial, it is more robust compared to PPD in longer stimuli designs and reflects the changes of whole-duration auditory performance more comprehensively [25].

Because pupil measurement requires stabilized-head position, sustained attention and patience, it is relatively hard to access and test children subjects with pupillometry, there are only a few available literatures using pupillometry to evaluate listening effort in school-aged children. Johnson and colleagues measured pupil dilation in children aged 7.5 to 14 years to investigate the extent to which age differences in short-term memory (STM) capacity were related to task engagement during encoding[26]. Steel et al., examined whether binaural fusion reduces listening effort in 11 to 15-year-old children who use bilateral cochlear implants[27]. McGarrigle et al., measured listening effort and fatigue in normal-hearing children (8–11 years old)[28]. These studies showed that it is feasible to use pupillometry to quantify the elevated listening effort in hearing-impaired school-aged children.

Therefore, we conducted our study to evaluate the impact of mild to moderate U-shape hearing loss on listening effort in a group of school-aged children with AS, who warrant more research attention. To approximate classroom-like communication pattern, a listening-answer task was used in the current research. Two listening conditions were created: SNR = + 15 dB and SNR= -2 dB, which represented quiet environment and classroom environment respectively[29]. We hypothesised similar answer performance in both conditions because speech recognition performance were shown to be. Particularly, hearing loss of AS children is characterized as apparent drop in mid-frequency, and energy in the 1k–3k Hz range of targeted sentence were found to be positively correlated with intelligibility[30]. Therefore we also hypothesised that the adverse effect of hearing loss on the listening-answer task and listening effort could depend not only on the level of background noise, but also on the spectral distribution of target sentences. The present study revealed that both hearing status and cognitive ability have impact on the listening effort of AS children. The effort of AS children with U-shape hearing loss might come mainly from the consequent cognitive processing of the degraded speech inputs instead of simply speech reception in daily life.

2. Results

2.1 audio stimulus material analysis

 
Table 1

The average value of all items of stimulus material

Items

Mean value

average word number of stimulus materials

40.48 ± 4.132

MD of dialogue

9.023 ± 1.23 s

MD of question

2.765 ± 0.43 s

MER13 of dialogue

41.04 ± 6.5%

MER13 of question

39.15 ± 5.7%

Table 1 describes the average value of all items of stimulus material, and Fig-2 shows correlation results of each item. The regression analysis showed that MER13 of the stimulus material and MER13 of dialogue were not related to the total number of words and the duration of the stimulus (P > 0.05), which indicates that although the stimulus materials are not consistent in terms of duration and total number of words, they have no effect on the spectrum distribution of sentences.

2.2 behavioural data analysis

Fig-3 shows accuracy of answering questions across different SNRs and MER13 for NH and HI children. A mixed-model ANOVA revealed significant main effects of SNR on the accuracy of answering questions (Chisq = 114.8, df = 1, p < 0.001). The overall accuracy of answering questions was higher when SNR = + 15dB, regardless of the level of hearing status. A significant interaction between SNR and hearing status was revealed, (Chisq = 22.22, df = 1, p < 0.001), suggesting that HI children had significantly lower answer correct at difficult SNR than NH children () MER13 had no main effect on the accuracy of answering questions (Chisq = 0.001, df = 1, p > 0.05). Although there was no significant interaction between SNR and hearing status and MER13, we found a trend that high MER13 had a negative influence on accuracy for HI children but not for NH children (Chisq = 3.73, df = 1, p = 0.05)

2.3 baseline pupil diameter (BPD)

There was significant difference in baseline pupil level between NH children and HI children. The NH children always showed larger baseline pupil level before listening task, regardless of SNR = + 15 dB or SNR = -2 dB (Chisq = 4.67, df = 1, p = 0.04) (Fig-4).

2.4 time window of analysis containing listening to question and dialogues, thinking

Fig-5 shows visualization of pupil trace when aligned by the onset of response. Therefore, this time window of analysis contains the whole process of listening to dialogue, listening to questions andfurther processing of the information to yield an answer. It can be seen that the pupil size of HI children had no significant differences with that of NH children when SNR = + 15 dB. However, pupil size of HI children increase remarkably at -2 dB SNR, especially in the time range from 0 to -5s, which was the stage of hearing and thinking.

3.4.1 peak pupil dilation (PPD)

AS Fig-6 showed, no significant main effect of hearing status (Chisq = 1.88, df = 1, p = 0.17), SNR (Chisq = 0.78, df = 1, p = 0.37) or MER13 (Chisq = 1.03, df = 1, p = 0.31) on PPD was found when performing mixed-model ANOVA. Similarly, there was no pairwise interaction between hearing status, SNR or MER13 (Chisq = 0.64, df = 1, p = 0.42; Chisq = 0.07, df = 1, p = 0.79; Chisq = 1.35, DF = 1, p = 0.25), there was no interaction between the three (Chisq = 0.07, DF = 1, p = 0.79), indicating that adding any of the factors did not affect PPD..

2.4.2 mean pupil dilation (MPD)

A mixed-model ANOVA revealed a significant main effect of SNR (Chisq = 2.05, df = 1, p < 0.05) and hearing status (Chisq = 0.29, df = 1, p < 0.05) on the MPD (Fig-7). The analysis also revealed a significant interaction effect between hearing status (HI versus NH) and SNR (Chisq = 5.74, DF = 1, p = 0.02), indicating that the response across SNRs varied between listener groups. MPD of HI children was significantly increased at -2 dB SNR, compared with that of NH children. However, when SNR = + 15 dB, there was no significant difference in MPD between the two groups. There was no significant effect of MER13 on MPD (Chisq = 0.44, df = 1, p = 0.51), and no significant interaction effect was observed between MER13 and SNR or hearing status.

2.5 time window of analysis containing listening to dialogue

Fig-8 shows visualization of pupil size when aligned by the onset of question in two conditions. In this part, time window of analysis only contains listening to dialogue, without the process of listening to questions and thinking. There is no significant difference in pupil size between HI children and NH children when they simply listen to dialogue. The effects of SNR, hearing status and MER13 on PPD and MPD were analyzed in the same way. The results showed that there was no main effect of SNR, hearing status and MER13 on PPD or MPD. Additionally, the two-way interactions between hearing status x SNR, hearing status x MER13, MER13 x SNR and the three-way interaction between hearing status x SNR x MER13 were not significant.

2.6 Correlation analysis

To investigate the relationship between PPD, MPD and other variables, we analyzed PPD, MPD, age, PTA1k − 2k, PTA0.5k−4k, and the accuracy of answering questions of all subjects under different listening conditions through correlation analysis using Pearson’s r. The results show that when SNR = -2 dB, PPD was correlated with the hearing status (Table 2). The higher PTA1k − 2k and PTA0.5k−4k, the greater PPD (R2 = 0.41; R2 = 0.40). PPD was negatively correlated with the accuracy rate (R2 = 0.40). When SNR = + 15 dB, PPD was not significantly correlated with hearing status, but was still negatively correlated with accuracy (Table 3).

Table 2

When SNR= -2 dB, the correlation between age, PPD, MPD and PTA1k − 2k, PTA0.5k−4k, the accuracy of answering questions. (*: p < 0.05)

SNR= -2 dB

Age

PTA1k − 2k

PTA0.5k−4k

accuracy

PPD

MPD

Age

           

PTA1k − 2k

0.6914

         

PTA0.5k−4k

0.6628

0.0000

       

accuracy

0.5488

0.0000

0.0000

     

PPD

0.3601

0.0355 (*)

0.0442 (*)

0.0438 (*)

   

MPD

0.9506

0.1441

0.1413

0.1761

0.0061

 


Table 3

When SNR = + 15 dB, the correlation between age, PPD, MPD and PTA1k − 2k, PTA0.5k−4k, the accuracy of answering questions. (*: p < 0.05)

SNR = + 15 dB

Age

PTA1k − 2k

PTA0.5k−4k

accuracy

PPD

MPD

Age

           

PTA1k − 2k

0.6914

         

PTA0.5k−4k

0.6628

0.0000

       

accuracy

0.3415

0.0000

0.0000

     

PPD

0.8821

0.6398

0.7010

0.0324 (*)

   

MPD

0.2898

0.7101

0.7375

0.9705

0.0040

 

3. Discussion

Hearing loss or background noise will increase the cognitive resources required for decoding auditory inputs, leading to effortful listening. [31]. AS children with hearing loss may often experience effortful listening, which could seriously affects the quality of learning and work. In the present study, we acquired the pupil dilation response (BPD, PPD and MPD) during a listening-answer task in two conditions (SNR= -2 dB and SNR = + 15 dB) for NH and AS children. We showed the influence of SNR and spectral distribution of the test materials on listening effort (as indexed by pupil response) in school-aged children with Alport syndrome. Additionally, the present results also corroborate a close link between cognitive demands imposed by a listen-answer task and task-evoked pupil dilation, which was consistent with research in adults. This provides further evidence of the negative effects of long-term hearing loss on the auditory cognitive ability.

Our findings extend prior works in two ways. First, we provide evidence that AS children with mild to moderate U-shape hearing loss need more cognitive resources mainly in the stage of thinking not just listening compared to NH children under current experimental conditions. This is the first study to evaluate listening effort on school-aged children with mild to moderate U-shape hearing loss by pupillometry. This is a special group because according to traditional audiometry battery, this group has audiograms that don’t suggest hearing loss in speech-related information, and speech communication is successful. Second, our longer-term speech material and listening-answer task format reflect better ecologically realistic scenarios, which regularly experienced by school-aged children in everyday life. We have also shown that the task could elicit task-evoked pupillary response reliably in HI and NH children. Therefore, this task could reveal the process of cognitive resource allocation in real situation.

Our hypothesis was to explore the influence of SNR and spectrum distribution on listening effort, however, we found significant effect of SNR instead of MER13 on accuracy or pupil response. There was a trend that AS children may have worse performance when listening to auditory materials with more energy in 1k-3k Hz. Note that speech materials used were not processed in the spectral aspect, therefore the difference in spectral concentration between high and low group is natural and small. Although this allows us to use pre-validated speech materials (especially important when this task format is new), it might explain the insignificant trend observed in the effect of MER13. We assumed that AS children might have significantly larger pupil size and worse performance if manipulate the spectral features of the speech materials in a more causal experimental design.

In our study, we used two kinds of time window to analyze pupil responses. In the time window containing only listening to dialogue, no matter SNR = -2 dB or SNR = + 15 dB, there was no significant difference of pupil size between AS children and NH children. The results of the mixed effect model also suggested that SNR, hearing status and MER13 had no significant effects on PPD and MPD, indicating that AS children might not need to provide much extra cognitive resources for the expenditure of intense effort at this stage. This is consistent with the results from clinical audiograms and speech recognition tasks that AS children are generally successful in speech communication.

However, when looking at the time window including listening dialogues, questions and thinking, larger pupil size were observed in AS children at -2 dB SNR compared to their NH peers, suggesting that extra effort might appear especially at the stage of thinking for AS children when the task is difficult. In this time window, the subjects need to re-extract the dialogue content they heard, connect it with the questions and make decisions after reasoning or calculating at this stage. The results show significant main effect of SNR and hearing status on MPD, which is more robust compared to PPD in longer stimuli designs[32]. MPD gives further information for the sustained cognitive load in a given time window[25]. Larger MPD in this time window represents constant allocation of cognitive resources in AS children. Increasing SNR will result in reduction of MPD, which reminds us that intervention might release effortful listening.

The framework of ELU (Ease of Language Understanding Model) may propose an explanation to our results[33]. In time window including listening dialogues, questions and thinking, the listeners needed to store the sentence and reconstruct information that were either missed or distorted. When there is a mismatch between the input signal and the long-term memory, explicit processing needs to be initiated, which depends on cognitive abilities such as working memory capacity and efficiency[34]. This explicit processing requires extra cognitive resources from the limited pool of resources that would otherwise be used for other mental tasks (i.e., calculation, deduction etc). Therefore we see a lower score of AS children in answer correct at difficult SNR conditions because they could not invest as many cognitive resources to perform the mental processing to get the correct answer AS children also need to allocate more cognitive resources in general, which is reflected by larger pupil size.

Note that without utilizing this new behavioral task that involves both listening and associated mental processing, we might not observe this difference between AS and NH children in both behavioral and physiological responses. Ecologically, it is rare to find a speech communication scenario where only passive listening and repetitions are required, as in some previous listening effort studies. Typically, listeners need to decode the speech inputs, process the information and prepare a response accordingly to either extract more information or giving more information. All of these require cognitive resources from a limited pool or resources. Extra resource needs at the stage of speech comprehension would impair the accuracy and efficiency of later stages, leading to poorer social communication, learning and development. This highlights the importance to investigate ecological speech communication scenarios in order to identify groups at higher risk of cognitive development.

It is widely accepted that pupil response have a close relationship with cognitive function[17], but the exact relationship between the ability to allocate cognitive resources and cognitive function remains unclear. According to the hypothesis proposed by Van der[35], our results are more consistent with the second hypothesis, namely, people with better cognitive function utilize cognitive resources more efficiently, so they only need to invest relatively a few cognitive resources when executing a task, it is suggested that long-term hearing loss lead a negative effect on the efficiency of utilizing cognitive resource. Therefore, AS children need to devote more cognitive resources to cope with the current task, and show larger mean pupil dilation due to low efficiency of utilizing cognitive resources.

As indicated, some researchers considered that BPD can reflect the arousal state of the cognitive system[36]. Gilzenrat believed that larger BPD before the experiment represents an over-arousal state of the cognitive system, indicating that the subjects are anxiety or stressful[37]. Moreover, Alhanbali found that there was a significant correlation between BPD and fatigue scales[38], smaller BPD represents low arousal state of the cognitive system, indicating great fatigue of the subjects. According to the Yerkes-Dodson Law, both over arousal and low arousal are associated with poor performance of cognitive task, while arousal levels between these two extremes are expected to have the best performance[39] [40]. Animal studies have also revealed an inverted U-shaped relationship between task performance and arousal levels. McGinley found that optimal signal detection behavior and sound-evoked responses of mice occurred only at intermediate arousal when membrane potentials were stably hyperpolarized, instead of low arousal or over arousal [41]. In our study, BPD of AS children are smaller in all listening conditions compared to NH children, suggesting they suffered from auditory fatigue when processing auditory cognitive. The negative impact of long-term hearing loss was their cognitive system are often in a state of low arousal. Recent findings indicated that long-term hearing loss of AS children affects cognition in two aspects, one is efficiency of utilizing cognitive resource, the other is arousal state, and both of them further affects listening effort (Fig-9).

There are several limitations of the present study that need to be mentioned. First, we only measured objective listening effort but no specialized test for cognitive abilities such as working memory capacity and efficiency. We might have gained more insight into the associations between listening effort and cognitive function if we had also included variables controlling for individual differences in cognitive abilities. Second, we did not manipulate experimentally the spectral distribution of the speech materials but rather choose to use pre-validated and more natural stimuli. The results of the study only proved that the MER13 had an influence on the behavioral test, but no significant results were found. Strictly modulating of spectral distribution might be improved in future studies.

4. Methods

4.1 Participants

A total of 33 boys aged 8–16 years were included in this study: 22 were AS children (mean age = 12.21 ± 0.66 years) with a U-shape hearing loss (Fig-10, PTA0.5k−4k=45.12 ± 8.774 dB HL, PTA1k − 2k=50.18 ± 10.25 dB HL) and 11 were age-matched normal hearing (NH) children (mean age = 11.64 ± 0.52 years) with pure tone audiometric thresholds confirmed to be lower than 20 dB HL at 500, 1000, 2000 and 4000 Hz. All children were native Mandarin speakers and had normal vision. Written parental consent and verbal assent from children were obtained prior to testing. Peking University First Hospital Research Ethics Committee approved study (reference number: 2020-068).

4.2 Equipment

The participant was seated 60 cm away from a 19-inch flat-screen computer monitor at 0 azimuth, which displayed the visual reminder of stimulus onset and response prompt. Stimulus presentation was programmed using the Matlab software. We presented auditory stimuli through around-ear headsets (HDA200). Pupil size was recorded using an eye camera, which has a sampling rate of 100 Hz (https://pupil-labs.com/products).

4.3 Speech stimuli

A total of 50 listening-focused and age-appropriate speech conversations were created for experimental trials. These short conversations were taken from the Youth Chinese Test (YCT), which was developed and validated by Confucius Institute Headquarters to assess primary and middle school students' abilities to use Chinese in their daily lives (http://www.hanban.org).

The comprehension difficulty level of these short conversations was equivalent. Each conversation contained three sentences consisting of 30–50 words in total. Each conversation lasted between 9s and 13s. The first two sentences of the conversation introduced the scene, which were recorded respectively from a native male or female speaker of standard mandarin. The third sentence raised a question referred to the conversation. (e.g. Female: Hurry up, it is half past eight, the train is about to leave. Male: Don't worry, there are still half one hour left. Question: What time does the train leave?) Therefore, different from speech materials used in past listening effort studies that requires only listening and repeating the sentences, this listening-answer task requires participants to both understand the conversation and perform further mental processing of the received information to provide the correct answer.

All sentences were recorded in an anechoic chamber and at a natural pace, from native male and female Mandarin speakers. To evaluate the impact of energy in mid-frequency on listening effort of AS children, Praat was used to divide these conversations into two categories according to the proportion of mid-frequency: Group A with more energy in mid-frequency (1000 Hz to 3000 Hz), Group B with less (Fig. 11). The performance results obtained by participants in pilot study further suggests that these short conversations were suitable for children aged 8–16 years.

To investigate the impact of SNR on listening effort, a background noise file consisting of multi-talker babble was digitally mixed with the speech stimuli to create two listening conditions: easy and difficult. The SNR for the easy condition was + 15 dB, while for the difficult condition was − 2 dB. Multi-talker babble was used as the masker because informational masking is believed to have consequences of increased cognitive load[42]. Overall output level for both listening conditions was fixed at 65 dB SPL.

4.4 Design and procedure

On arrival, participants were seated comfortably in the soundproof booth. The luminance (206 lux) of the visual field was controlled by adjusting room lighting and screen brightness levels. Pupil calibration was then performed to ensure that raw eye data could be accurately mapped. Then, participants were given the following instructions: ‘‘During the listening task, you will hear some short conversations. Please fixate on the cross in middle of the screen and try not to blink or move suddenly. After each conversation, you will also hear a question. Shortly, a “o” sign will appear on the screen. Please try your best to answer as quickly and accurately as you can at the sight of the “o” sign. Try to look straight ahead at the cross in the middle of the screen while you listen to each conversation. It’s important that you must pay attention to the whole conversation.” Participants were given the opportunity to perform a practice session before beginning the recorded experiment, which ensure the children were adequately familiar with the task and were able to perform the task accurately.

Overall, each participant was presented with 2 easy blocks and 2 difficult blocks, each block consisting of 10 conversations (40 conversations in total). The order of easy and difficult blocks was random. Participants began each trial by fixating on a “+” sign shown in the center of the screen. Background noise (multi-talker babble) was then presented. After 2s of noise-alone presentation, the conversation began. Each conversation was also followed by 1s of noise-alone presentation. Following the end of the conversation was a question recorded from a native female speaker. Then, after a 2s period for thinking, a “o” sign appeared on the screen and children answer the question. (Fig-12)

4.5 Analysis

4.5.1 speech materials feature summary

To check whether the word number and duration of the stimulus material have influence on the spectrum distribution, we conducted a correlation analysis on some items of stimulus materials, including (1)mean word number of stimulus materials, (2)mean duration of short dialogue (MD of dialogue), (3)mean duration of the question (MD of question), (4)mean energy ratio of short dialogue at 1k Hz to 3k Hz (MER13 of dialogue), (5) mean energy ratio of question at 1k Hz to 3k Hz (MER13 of question). This is to identify potential confounds in speech materials features that could impact listening-answer task performance and listening effort.

4.5.2 pupil data

The analyses of pupil data were consistent with those described in the literature by Yue Zhang[43] and Matthew B[44]. The pupil diameter measured from the dialogue onset to the end of the trial was divided by that baseline level to obtain relative changes in pupil diameter elicited by the task.

Pupil diameter data were pre-processed in Matlab to detect and remove blinks and gaze displacement. Sample points were coded as blinks when pupil diameter values were below 3 standard deviation (SD) of the mean of the unprocessed trace or when gazing positions were 3 SD away from the center of the fixation., because they can affect the validity pupil diameter measurement. Traces between 10 data points (0.1s) before the start and after the end of blink were interpolated cubically in Matlab, to further decrease the impact of the obscured pupil from blinks. Trials where blinks and missing data exceeded 20% of the total samples were considered invalid and excluded from all analyses. No participants were excluded from the analysis. All valid traces were then smoothed using a first-order Butterworth filter and downsampled from 120 Hz to 30 Hz. Processed traces were then aligned by 1) the onset of the response prompt (the display of circle to signal participants to answer the question) or 2) aligned by the offset the dialogue. These two alignment measures provide two window of analysis that contains: 1) listening to the dialogue and question with extra time for preparing the answer and 2) listening to the dialogue. The aligned traces were then aggregated per listener by each condition.

All statistical analyses and plotting were performed in R 4.0 (https://www.rproject.org). We applied linear mixed models (LMM) to analyze the data as LMMs tolerate missing values, while repeated measures ANOVA tests only use complete cases contrary to multilevel analyses. Moreover, mixed-effects models are more flexible in processing the multilevel structure of the data. To examine the effect of SNR, hearing status and MER13 on PPD, a mixed-effect model was fitted on listeners’ word recognition, using SNR, hearing status and MER13 as fixed effect factors, with LISTENER and material list as random effect factors. Mixed effect models allow for controlling the variance associated with random factors without data aggregation. Therefore, by using LISTENER and WORD LIST used for stimuli as random effect factors in the model, we controlled for the variance in overall performance (random intercept) and dependency on other fixed factors (random slope) that were associated with LISTENER and material LIST. Models were constructed using the lme4 package in R, and figures were produced using the ggplot2 package. Fixed and random effect factors entered the model, and remained in the model only if they significantly improved the model fitting, using Chi-squared tests based on changes in deviance (p < 0.05). Differences between levels of each factor and interactions were examined with post-hoc Wald test. p values were estimated using the z distribution in the test as an approximation for the t distribution. In the same way, linear mixed models were used to analyze the effect on MPD and accuracy. Finally, packages lm was used to analyze the correlation between age, PPD, MPD and PTA1k-2k, PTA0.5k-4k, the accuracy of answering questions of all subjects.

Declarations

Statement

Peking University First Hospital Research Ethics Committee approved study (reference number: 2020-068). Informed consent was obtained from all participants and their legal guardians. Our study and all the protocols used were in accordance with the Declaration of Helsinki.

References

  1. Flinter, F. A. et al. Genetics of classic Alport's syndrome. Lancet (London, England), 2 (8618), 1005–1007 (1988).
  2. Zhang, X. et al. X-linked Alport syndrome: pathogenic variant features and further auditory genotype-phenotype correlations in males. Orphanet journal of rare diseases, 13 (1), 229 (2018).
  3. Barozzi, S. et al. Vestibular and audiological findings in the Alport syndrome.American journal of medical genetics Part A. 2020.
  4. Bronkhorst, A. W. The cocktail-party problem revisited: early processing and selection of multi-talker speech. Attention, perception & psychophysics, 77 (5), 1465–1487 (2015).
  5. Wendt, D., Kollmeier, B. & Brand, T. How hearing impairment affects sentence comprehension: using eye fixations to investigate the duration of speech processing.Trends in hearing. 2015;19.
  6. Hornsby, B. W. The effects of hearing aid use on listening effort and mental fatigue associated with sustained speech processing demands. Ear and hearing, 34 (5), 523–534 (2013).
  7. Hornsby, B. W. et al. Subjective fatigue in children with hearing loss: some preliminary findings. American journal of audiology, 23 (1), 129–134 (2014).
  8. Bess, F. H. & Hornsby, B. W. Commentary: listening can be exhausting–fatigue in children and adults with hearing loss. Ear and hearing, 35 (6), 592–599 (2014).
  9. Klatte, M., Lachmann, T. & Meis, M. Effects of noise and reverberation on speech perception and listening comprehension of children and adults in a classroom-like setting. Noise & health, 12 (49), 270–282 (2010).
  10. Moore, D. R., Zobay, O. & Ferguson, M. A. Minimal and Mild Hearing Loss in Children: Association with Auditory Perception, Cognition, and Communication Problems. Ear and hearing, 41 (4), 720–732 (2020).
  11. Sarampalis, A. et al. Objective measures of listening effort: effects of background noise and noise reduction. Journal of speech, language, and hearing research: JSLHR, 52 (5), 1230–1240 (2009).
  12. McGarrigle, R. et al. Listening effort and fatigue: what exactly are we measuring? A British Society of Audiology Cognition in Hearing Special Interest Group 'white paper'. International journal of audiology, 53 (7), 433–440 (2014).
  13. Peelle, J. E. Listening Effort: How the Cognitive Consequences of Acoustic Challenge Are Reflected in Brain and Behavior. Ear and hearing, 39 (2), 204–214 (2018).
  14. Hsu, B. C., Vanpoucke, F. & van Wieringen, A. Listening Effort Through Depth of Processing in School-Age Children. Ear and hearing, 38 (5), 568–576 (2017).
  15. Bess, F. H. et al. Salivary Cortisol Profiles of Children with Hearing Loss. Ear and hearing, 37 (3), 334–344 (2016).
  16. Zekveld, A. A., Kramer, S. E. & Festen, J. M. Pupil response as an indication of effortful listening: the influence of sentence intelligibility. Ear and hearing, 31 (4), 480–490 (2010).
  17. Beatty, J. Task-evoked pupillary responses, processing load, and the structure of processing resources. Psychological bulletin, 91 (2), 276–292 (1982).
  18. Hess, E. H. & Polt, J. M. Pupil Size in Relation to Mental Activity during Simple Problem-Solving 1431190–2(Science, New York, NY, 1964). 3611
  19. Zekveld, A. A. et al. The eye as a window to the listening brain: neural correlates of pupil size as a measure of cognitive listening load. NeuroImage, 101, 76–86 (2014).
  20. Van Engen, K. J. & McLaughlin, D. J. Eyes and ears: Using eye tracking and pupillometry to understand challenges to speech recognition. Hearing research, 369, 56–66 (2018).
  21. Giuliani, N. P., Brown, C. J. & Wu, Y. H. Comparisons of the Sensitivity and Reliability of Multiple Measures of Listening Effort. Ear and hearing. 2020.
  22. Ohlenforst, B. et al. Impact of SNR, masker type and noise reduction processing on sentence recognition performance and listening effort as indicated by the pupil dilation response. Hearing research, 365, 90–99 (2018).
  23. Dorothea, W. et al. Patrycja, Ksi??Ek, Toward a more comprehensive understanding of the impact of masker type and signal-to-noise ratio on the pupillary response while performing a speech-in-noise test.Hearing Research. 2018:S0378595517305294-.
  24. Ohlenforst, B. et al. Impact of stimulus-related factors and hearing impairment on listening effort as indicated by pupil dilation. Hearing research, 351, 68–79 (2017).
  25. Seifi Ala, T. et al. An exploratory Study of EEG Alpha Oscillation and Pupil Dilation in Hearing-Aid Users During Effortful listening to Continuous Speech. PloS one, 15 (7), e0235782 (2020).
  26. Johnson, E. L. et al. Task-evoked pupillometry provides a window into the development of short-term memory capacity. Frontiers in psychology, 5, 218 (2014).
  27. Steel, M. M., Papsin, B. C. & Gordon, K. A. Binaural fusion and listening effort in children who use bilateral cochlear implants: a psychoacoustic and pupillometric study. PloS one, 10 (2), e0117611 (2015).
  28. McGarrigle, R. et al. Measuring listening-related effort and fatigue in school-aged children using pupillometry. Journal of experimental child psychology, 161, 95–112 (2017).
  29. Picard, M. & Bradley, J. S. Revisiting speech interference in classrooms. Audiology Official Organ of the International Society of Audiology, 40 (5), 221 (2001).
  30. Green, T. et al. Talker intelligibility differences in cochlear implant listeners. The Journal of the Acoustical Society of America, 121 (6), EL223–9 (2007).
  31. Wayne, R. V. & Johnsrude, I. S. A review of causal mechanisms underlying the link between age-related hearing loss and cognitive decline. Ageing research reviews, 23, 154–166 (2015).
  32. Karatekin, C. Development of attentional allocation in the dual task paradigm. International journal of psychophysiology: official journal of the International Organization of Psychophysiology, 52 (1), 7–21 (2004).
  33. Rönnberg, J. et al. Cognition counts: a working memory system for ease of language understanding (ELU).International journal of audiology. 2008:S99-105.
  34. Wingfield, A. Evolution of Models of Working Memory and Cognitive Resources.Ear and hearing. 2016:35S-43S.
  35. van der Meer, E. et al. Resource allocation and fluid intelligence: insights from pupillometry., 47 (1), 158–169 (2010).
  36. Granholm, E. & Steinhauer, S. R. Pupillometric measures of cognitive and emotional processes. International journal of psychophysiology: official journal of the International Organization of Psychophysiology, 52 (1), 1–6 (2004).
  37. Gilzenrat, M. S. et al. Pupil diameter tracks changes in control state predicted by the adaptive gain theory of locus coeruleus function. Cognitive, affective & behavioral neuroscience, 10 (2), 252–269 (2010).
  38. Alhanbali, S. et al. Dimensions of self-reported listening effort and fatigue on a digits-in-noise task, and association with baseline pupil size and performance accuracy.International journal of audiology. 2020:1–11.
  39. Ayasse, N. D. & Wingfield, A. Anticipatory Baseline Pupil Diameter Is Sensitive to Differences in Hearing Thresholds. Frontiers in psychology, 10, 2947 (2019).
  40. Yerkes, R. M. & Dodson, J. D. The relation of strength of stimulus to rapidity of habit-formation.Journal of Comparative Neurology and Psychology. 1908;18(5).
  41. McGinley, M. J., David, S. V. & McCormick, D. A. Cortical Membrane Potential Signature of Optimal States for Sensory Signal Detection., 87 (1), 179–192 (2015).
  42. Prodi, N. et al. Noise, Age, and Gender Effects on Speech Intelligibility and Sentence Comprehension for 11- to 13-Year-Old Children in Real Classrooms. Frontiers in psychology, 10, 2166 (2019).
  43. Zhang, Y., Lehmann, A. & Deroche, M. Disentangling listening effort and memory load beyond behavioural evidence: Pupillary response to listening effort during a concurrent memory task. PloS one, 16 (3), e0233251 (2021).
  44. Winn, M. B. et al. Best Practices and Advice for Using Pupillometry to Measure Listening Effort: An Introduction for Those Who Want to Get Started. Trends in hearing, 22, 2331216518800869 (2018).