Abstract
Health-care service delivery models have evolved from a practitioner-centered approach toward a patient-centered ideal. Concurrently, increasing emphasis has been placed on the use of empirical evidence in decision-making to increase clinical accountability. The way in which clinicians use empirical evidence and client preferences to inform decision-making provides an insight into health-care delivery models utilized in clinical practice. The present study aimed to investigate the sources of information audiologists use when discussing rehabilitation choices with clients, and discuss the findings within the context of evidence-based practice and patient-centered care. To assess the changes that may have occurred over time, this study uses a questionnaire based on one of the few studies of decision-making behavior in audiologists, published in 1989. The present questionnaire was completed by 96 audiologists who attended the World Congress of Audiology in 2014. The responses were analyzed using qualitative and quantitative approaches. Results suggest that audiologists rank clinical test results and client preferences as the most important factors for decision-making. Discussion with colleagues or experts was also frequently reported as an important source influencing decision-making. Approximately 20% of audiologists mentioned utilizing research evidence to inform decision-making when no clear solution was available. Information shared at conferences was ranked low in terms of importance and reliability. This study highlights an increase in awareness of concepts associated with evidence-based practice and patient-centered care within audiology settings, consistent with current research-to-practice dissemination pathways. It also highlights that these pathways may not be sufficient for an effective clinical implementation of these practices.
Introduction
The type of information used by various health-care professionals in their clinical decision-making has been widely studied (Heiwe et al., 2011; Metcalfe et al., 2001; Sosnowy, Weiss, Maylahn, Pirani, & Katagiri, 2013; Straus, Tetroe, & Graham, 2011). There is, however, a paucity of current research relating to decision-making in audiology. Clinical decision-making is a contextual, continuous, and evolving process, where data are gathered, interpreted, and evaluated in order to select a course of action (Tiffen, Corbridge, & Slimmer, 2014). The way that clinicians make decisions about patient care provides insights into their style of practice. A study conducted by Doyle in 1989 surveyed Australian audiologists’ breadth of decision-making, including the frequency and topics of decisions relating to diagnostic (e.g., which test to use), rehabilitative (e.g., which type of hearing device to recommend), and procedural decisions (e.g., how to manage waiting lists). That study also surveyed the sources of information used when making decisions (e.g., client’s preferences, journals, or colleagues); the perceived degree of difficulty in reaching decisions; and the perceived confidence in decisions. The findings suggested that audiologists considered published literature a low priority for assisting clinical decisions, as it was seldom mentioned and consistently ranked below other sources of information. Interestingly, audiologists perceived some degree of difficulty in making decisions due to complex and sometimes conflicting information sources; however, they generally felt confident in their decisions (Doyle, 1989).
Over the last 30 years, health-care service delivery models have evolved away from the long-standing practitioner-centered framework, where decisions are unilateral and formed primarily from expert opinion, toward a model that is more patient-centered, placing greater emphasis on patient engagement and shared decision-making (Richards, Montori, Godlee, Lapsley, & Paul, 2013). As described by Berwick (2016), there has been a transition from an era in which clinicians had the authority to judge the quality of their own work, toward an era promoting outcome measurement and accountability to patients, third-party funding bodies, and society. This transition resulted from unexplained variations between practices and patient outcomes, high rates of injuries from errors, and increased intervention costs (Brennan et al., 1991; McIntyre & Popper, 1983; Vincent, Neale, & Woloshynowych, 2001; Wilson et al., 1995). Clinical practice guidelines (CPG), including performance measures (Williams, Schmaltz, Morton, Koss, & Loeb, 2005) and checklists (Cohen, Littenberg, Wetzel, & Neuhauser, 1982; Haynes et al., 2009), were introduced to minimize practice variability and error rates. Ideally, these guidelines are developed based on principles of evidence-based practice (EBP; Haines & Jones, 1994; Woolf, 1992). That is, guidelines should develop from systematic identification and synthesis of the best available scientific evidence, in conjunction with clinical expertise, and patient values (National Health and Medical Research Council, 1999; Sackett, Rosenberg, Gray, Haynes, & Richardson, 1996; Wolf, 1999). Clinicians’ imperative to align their practice with EBP is reflected in professional practice standards across many professional bodies.
While the important role of patient input is recognized in EBP, the specific concept of patient-centered care (PCC) has also been given greater attention in health care in recent years. In its most basic form, PCC refers to patient–clinician interactions, emphasizes the importance of relationship building, and sharing of input and control in information exchange and decision-making. That is, PCC advocates for a more biopsychosocial and mutualistic approach (Roter, 2000), whereby the patient’s beliefs, goals, and perspectives are taken into account in the development of practice guidelines and delivery of health care (Montori, Brito, & Murad, 2013).
EBP and PCC models may be more difficult to achieve within health-care professions that also function as a business practice, where financial interests might be in conflict with the patient’s best interest. For example, audiologists in many countries can both assess hearing and dispense hearing aids to the general public (Metz, 2000), and recent reports suggest that audiology has a focus on product sales (Lowy, 2013; Shaw, 2016). Audiology therefore provides an interesting field to evaluate the adoption of EBP and PCC practices.
Over the last decade, several authors have studied or promoted the benefits of using EBP (Cox, 2005; Moodie et al., 2011; Wong & Hickson, 2011) and PCC in audiology (Ekberg, Grenness, & Hickson, 2014; Grenness, Hickson, Laplante-Lévesque, & Davidson, 2014; Grenness, Hickson, Laplante-Lévesque, Meyer, & Davidson, 2015a; Grenness, Hickson, Laplante-Lévesque, Meyer, & Davidson, 2015b; Laplante-Lévesque, Hickson, & Grenness, 2014; Laplante-Lévesque, Hickson, & Worrall, 2010; Poost-Foroosh, Jennings, Shaw, Meston, & Cheesman, 2011). As shown in Figure 1, among all articles ( Percentage of articles over time, in audiology journals, that mention evidence-based practice (EBP) and patient-centered care (PCC).
Due to the increased focus on EBP and PCC, in combination with indications of notable discrepancies between recommended best practice and current practice, there is a need to reevaluate how clinical decisions are made in audiology. This questionnaire-based study aimed to: (a) identify how audiologists rate the importance and reliability of different sources of information; (b) identify which sources of information audiologists mention accessing in difficult clinical scenarios; (c) assess perceived difficulty and confidence in decision-making; (d) evaluate how these aspects of decision-making (a–c) have changed relative to Doyle (1989); and (e) assess how compatible current clinical decision-making behaviors are with EBP and PCC.
Methodology
Participants
Demographic Characteristics of Participants.
Questionnaire
The questionnaire was based on Doyle’s (1989) survey of Australian audiologists’ decision-making. While Doyle (1989) covered three types of decisions: (a)
Questionnaire items included rating scales, ranking questions, and open-ended questions relating to decision-making behaviors. The topics covered in the questionnaire included decision-making difficulties, confidence in recommendations, and information used when making decisions. Ranking questions required participants to order common sources of information used in decision-making both in terms of importance and reliability. The questionnaire was piloted with 10 audiologists and feedback regarding the questions was incorporated before distribution. The complete questionnaire is found in the Appendix.
Analysis
Wilcoxon signed-rank tests with Bonferroni correction for multiple testing were used to compare the ranking questions, while a multiple linear regression analysis was used to determine whether age, experience, or practice funding may influence respondents’ self-reported use of CPG. The use of CPG was measured on a Likert-type scale, which can be argued to be ordinal and therefore would formally break the assumptions for multivariate regressions. However, as extensively discussed in the literature, the gains in power outweighs the small biases that they may cause (cf. Knapp, 1990; Labovitz, 1970). A
The responses to the two open-ended questions were initially analyzed using a qualitative deductive approach to content analysis (Elo & Kyngäs, 2008; Graneheim & Lundman, 2004). The questions were as follows: (a) In situations where the optimal solution for your client is not straightforward, how do you identify an appropriate recommendation? (b) When you have to make a difficult clinical decision and your sources of information are conflicting or contraindicating, what would you do? In other words, the questions were similar but the first question was placed at the beginning of the questionnaire and intended to focus on difficult scenarios where more than one option is possible. The second question was placed at the end of the questionnaire and focused on conflicting information. Responses to the second question may therefore have been more influenced by the content of the questionnaire itself, in comparison to the first question. Responses to the open-ended questions were read several times and recurring ideas were identified, then compared and contrasted between participant responses to extract meaning (Graneheim & Lundman, 2004). The meaning units were then condensed within codes and categorized. The initial categorization followed a structured matrix based on the information used in decision-making from Doyle’s (1989) study (i.e., client history; client characteristic; information from client’s family and friends; referral or file notes; test results; texts, journals, or other publications). Aspects of the data that did not fit the matrix categorization were further analyzed to build new categories based on an inductive approach (Elo & Kyngäs, 2008). Two investigators reviewed the survey responses, first independently and then together until a consensus was achieved in condensing the final categories. The final categories identified were: client goals and preferences; discussion with colleagues or experts; trial and error; texts, journals, or other publications; further test results; and clinical experience. The categories were ultimately analyzed quantitatively to identify how frequently each category had been used. This study was approved by Macquarie University’s Human Sciences Research Ethics Sub-Committee.
Results
Demographic Profile of Respondents
The majority of participants who completed the questionnaire were from Australia (
Conceptualization of Difficult Clinical Situations
After reading examples of audiology rehabilitation scenarios where the intervention was not straightforward (see scenarios in Appendix), respondents were asked to write an example of their own. This question was included to help respondents to further conceptualize this type of situation before answering the rest of the questionnaire. Respondents noted multiple examples of difficult situations which involved factors such as: selecting the most appropriate intervention for particular types of hearing losses (
Ranking of Information Sources for Decision-Making
Importance and Reliability Rankings of Information Sources for Decision-Making. a
rating scale: 1 = most important or reliable; and 11 = least important or reliable.
The pattern of results was different for ratings of the information source’s
Similarly, respondents in Doyle (1989) ranked information from clients and test results as the most important source of information for the majority of respondents. In addition, information from texts, journals, and other publications were ranked lower than all the other categories. A Bonferroni-corrected Wilcoxon signed-ranks test identified that four sources were ranked significantly different (
Self-Reported Use of Information in Decision-Making
Categorization of Open-Ended Responses for the Question “In situations where the optimal solution for your client is not straightforward, how do you identify an appropriate recommendation?”
Multiple responses possible; mean number of sources reported = 1.45.
Results indicated that 54% of the 96 respondents considered client’s goals and preferences in their decision-making and 42% reported using a combination of multiple sources of information (as illustrated in Figure 2). For example, 12% mentioned the combination of client factors and use of trial and error: “I ask clients how they feel about options or what they would like to try first.” A further 28% mentioned client factors as their only source of information. For example: “By speaking with the client and identifying their wants and needs.” Interestingly, in these difficult cases, 10% reported referring to colleagues or experts only. In comparison, while 5% reported using only texts, journals, or other publications when a decision was not straightforward, 21% mentioned a combination of publications with other sources of information: “I would make a decision based on my clinical experience and peer reviewed literature. Also I would consult my colleague.” And: “[I would] weight everything from results, practice guidelines, textbooks and colleagues opinion.”
Illustration of the combination of sources of information identified by audiologists to the question: “In situations where the optimal solution for your client is not straightforward, how do you identify an appropriate recommendation?” Note that an exact proportional scaling and overlapping of six variables is not possible and this illustration is therefore a close-approximation.
Categorization of Open-Ended Responses for the Question “When you have to make a difficult clinical decision and your sources of information are conflicting or contraindicating, what would you do?”
Multiple responses possible; mean number of sources reported = 1.48.

Illustration of the combination of sources of information identified by audiologists to the question: “When you have to make a difficult clinical decision and your sources of information are conflicting or contraindicating, what would you do?” Note that an exact proportional scaling and overlapping of six variables is not possible and this illustration is therefore a close-approximation.
Because EBP is based on using a combination of different sources of information (empirical evidence, patient factors, and clinical expertise), it was not identified as a specific category in Tables 3 and 4. Across both open-ended questions, however, 13 clinicians mentioned EBP directly or indirectly in situations where there was no clear solution for the client as demonstrated by the following quotes: “All decisions are made on some-what evidence-based practice,” “I base my recommendations […] on my professional knowledge of the scientific evidence plus my clinical experience of similar situations […] fitting the needs of the patient.”
Multiple Linear Regression of How Often Clinical Guidelines Are Followed.
Coded 1 female and 0 for male.
Comparison group is the group with a Master’s degree.
Comparison group is those fully privately funded.
In addition, respondents in publically funded clinics considered the guidelines more important in decision-making and deemed them more reliable as compared with respondents in privately funded clinics (independent-samples median test:
Perceived Difficulty and Confidence in Decision-Making
In this study, participants reported moderate or slight difficulty in making recommendations when there was not one clear solution (Figure 4). In Doyle (1989), when respondents were asked to rate their degree of difficulty in the decision-making in general (not just when there was not one clear solution), they mainly rated a slight degree of difficulty.
Level of perceived difficulty in making recommendations. The X-axis represents the percentage of audiologists who responded the specified level of perceived difficulty.
Interestingly, when asked to rate their confidence in decision-making in these situations, the vast majority of respondents reported they were moderately to completely confident in their decision (98% and 99%, respectively for Doyle’s [1989] and this study; Figure 5). Only two respondents reported that they were “not very” or “not at all” confident in their decision, despite respondents reporting a relative degree of difficulty. As in Doyle (1989), a negative correlation was found between confidence of decision and perceived difficulty ( Level of perceived confidence in appropriateness of recommendations. The X-axis represents the percentage of audiologists who responded the specified level of perceived confidence.
Discussion
The aims of this study were to (a) identify how audiologists rate different sources of information that influence their decision-making, in order of importance and reliability; (b) identify which sources of information audiologists mention accessing in difficult clinical scenarios; (c) assess perceived difficulty and confidence in decision-making; (d) evaluate how these aspects of decision-making (a–c) have changed over the past several decades (Doyle, 1989); and (e) to assess how compatible the current clinical decision-making behavior is with EBP and PCC. The results indicated that client factors and test results are considered the most important by audiologists for their clinical decision-making. In comparison, information from peer-reviewed literature and textbooks were ranked lower than test results and client factors, in both this study and Doyle (1989). In the present study, respondents’ ranking suggested that while client’s opinion was an important factor in decision-making, it was considered somewhat less reliable. The opposite was observed with peer-reviewed literature, which was ranked as being more reliable than important for decision-making when the recommendation was not straightforward. This pattern of response suggests that clinicians distinguish between the importance and the reliability of different information sources. However, it is not clear how audiologists use this information in their daily practice, to incorporate best available evidence with patient’s preferences and values as part of EBP. Clinical decision-making requires a balance between identifying information that is important, and understanding its reliability and applicability to the clinical issue. This is typically not a simple task for complex clinical scenarios. As such, clinicians appear to adopt different styles of service delivery, as suggested by the two following quotes: “I will be firm on my findings” indicating a greater weight on the test results
The opinion of colleagues or experts took a predominant place when clinicians described how they came to a recommendation in more complex scenarios. It was, however, ranked mid-range in terms of importance and reliability. Specifically, in difficult clinical decision-making situations, results suggested that clinicians were more likely to refer to an expert or colleague over the search for published empirical evidence. A preference for oral sources over written sources, such as books, guidelines, and scientific papers, is in line with knowledge seeking behaviors of other professionals. For example, many health professionals would rely more on their peers than empirical research evidence during clinical decision-making (Lyons, Brown, Tseng, Casey, & McDonald, 2011; Mayer & Piterman, 1999; McAlister, Graham, Karr, & Laupacis, 1999; Rappolt & Tassone, 2002; Vallino-Napoli & Reilly, 2004). Evidence from outside health care reveals that information seeking behaviors of professionals follow a similar pattern, that is, employees prefer contacting colleagues directly before consulting written sources even when these colleagues are in other, including competing, organizations (Allen, James, & Gamlen, 2007; Ibrahim, Fallah, & Reilly, 2008; Lundmark & Klofsten, 2014; Schrader, 1991). Information technologies may also have expanded options for accessing the collective clinical expertise of audiologists, through media such as social media, distribution lists, or online forums. This collective clinical expertise may be particularly helpful in cases where there is no existing research evidence or practitioners encounter difficulties implementing CPGs. However, as with all elements of EBP, they should not be used in isolation (Sackett et al., 1996).
CPG enable clinicians to make informed decisions about clinical care (see Woolf, 1992). While CPG may be developed without following rigorous guidelines or engaging patients, and are often based on consensus within a specific clinic (Shaneyfelt, Mayo-Smith, & Rothwangl, 1999), they have the potential benefit of decreasing variability in practice and reducing error rates. In this study, among audiologists who reported having CPG at their workplace, those who reported using them often were more likely to be in a publically funded, as compared with a privately funded, organization. This may be partly due to differences in organizational culture (Dodek, Cahill, & Heyland, 2010; Hung, Leidig, & Shelley, 2014) as well as the standardized requirements for assessment for public funding in Australia (Australian Government Department of Health, 2015). Hung et al. (2014) demonstrated that primary care providers in clinics with stronger “group, hierarchical, and rational” culture reported a greater adherence to clinical guidelines. Certainly, in Australia, Australian Hearing, established as a federal agency in 1947, is the largest provider of Government-funded services. One of the objectives of this program is to maintain consistent standards of clinical service across their centers, through the provision of regular education and support to clinical staff, facilitating a strong and cohesive group culture.
The results also indicated that clinicians with longer clinical experience tend to report following CPG less often, but that older clinicians tend to report following guidelines more often. Interestingly, this result is in opposition to Chan et al. (2002) who evaluated compliance of nurse practitioners in Hong Kong with Universal Precautions (UP; i.e., preventing or minimizing exposure to blood and bodily fluids). While they found a positive relationship between age and compliance, they also found a significant positive relationship between years of experience and compliance. Another study, however, assessing hospital-based physicians in the United States have shown that compliance with Universal Precautions decreases with age (Michalsen et al., 1997). Because both these studies only provide results of univariate analyses, it is conceptually plausible that older respondents on average are also more experienced, a point raised by Michalsen et al. (1997). Our results indicated that while age and experience are correlated, their effects on compliance with guidelines may go in opposite directions. If so, this can explain contradicting findings in previous research. There are theoretical explanations for a differentiated effect of experience and increasing age such that as experience increases, it is possible that clinicians have developed clinical practice routines and their own “evidence-base” that might reduce the perceived need for guidelines (Harrison, Légaré, Graham, & Fervers, 2010). In addition, age in general may be positively associated with compliance based on generational differences in organizational behavior (Becton, Walker, & Jones-Farmer, 2014) and well-established age-related changes in personality traits such as agreeableness and conscientiousness (Soto, John, Gosling, & Potter, 2011).
With regard to the self-reported degree of difficulty and confidence in decision-making, the results for the current study are also similar to Doyle (1989). In particular, while audiologists identified a relative degree of difficulty, they still reported being confident in their decisions. The possible reasons for this remain to be explored; however, a number of clinicians stated that they would provide all the options to the client and then let the client decide. This behavior, however, appears to shift decision-making control and responsibility to the client (Say, Murtagh, & Thomson, 2006) and does not acknowledge the importance of providing decisional support to patients throughout the decision-making process (Elwyn et al., 2012; Stiggelbout et al., 2012) . Further, in some cases, it was unclear whether clinicians understood how to effectively involve the patient in shared decision-making. Shared decision-making is specific to the patient and takes into account the research evidence, clinician expertise, and patient preferences and values (Légaré & Witteman, 2013). Not all patients, however, want to be fully involved in the decision-making process, and many factors influence this preference (Say et al., 2006). For example, patients with higher socioeconomic status prefer an active role in the decision-making process as compared with those with lower socioeconomic status (Murray, Pollack, White, & Lo, 2007). Although the client makes the final choice, the clinician can facilitate the process by narrowing the options, framing the information in a certain way toward the patient’s needs, or choosing to include or omit information based on their discretion (Hibbard, Slovic, & Jewett, 1997). On the other hand, it is suggested that audiologists often play the dominant role in decision-making (Grenness et al., 2014). This is worthy of further exploration, as multiple barriers exist to the uptake of shared decision-making by health-care professionals, including motivation and time constraints, as well as the perception that it may not lead to improved patient outcomes or health-care processes (see Gravel, Légaré, & Graham, 2006). Therefore, it is unclear whether shared decision-making is not effectively utilized within audiology practice because of a poor understanding of how to implement it within clinical practice or whether the barriers for implementation have not yet been effectively addressed.
Over half of the surveyed audiologists responded that they would use multiple sources of information when their sources of information conflicted or contradicted each other. Of these, a reference to further research, peer-reviewed literature, or CPG was frequently mentioned. This finding suggests that there is a greater self-reported use of research in current clinical decision-making compared with the 1989 study by Doyle. Client factors were also commonly mentioned as being used as part of multiple sources, which is in line with the EBP and PCC models, and some clinicians mentioned EBP directly. While responses varied and suggested a range of practice styles, the overall results suggest a considerable increase in awareness of EBP and PCC for decision-making in audiology, which is aligned with the increased mention of EBP and PCC in the literature over the past few decades. Despite this, little is known about the extent to which EPB is implemented in audiological practice; moreover, of the evidence that exists regarding implementation of PCC, there appears to be significant scope for improvement (Grenness et al., 2015a, 2015b). Assessing the quality, reliability, and applicability of empirical research evidence to a clinical question requires skills and time (Mullen & Streiner, 2006; Straus & McAlister, 2000). If clinicians do not have adequate critical skills, it is unlikely that the EBP model will be implemented (Rambur, 1999) as little guidance is given when various forms of information are conflicting or vague. In addition, developing the psychosocial skills required for a truly patient-centered approach is complex and goes beyond the theoretical knowledge of PCC (Levinson, Lesser, & Epstein, 2010).
It is important to note that the results of the current study are based on questionnaire responses from 96 audiologists who attended the XXXII World Congress of Audiology, with 81 practicing in Australia. This number represents only a small proportion of the total membership of Audiology Australia, the main professional body of audiologists in the country (over 2,000 members). It is therefore possible to assume that respondents had certain characteristics that could bias the results toward more reports of EPB and PCC, than what may have been reported in a more representative sample of practicing audiologists who did not attend the conference or who attended the conference but chose not to respond to the questionnaire. This self-selection bias limits the generalizability of the findings. For example, respondents attending the conference and accepting to complete the questionnaire may have been more likely to be interested in or involved in research, leading to greater reports of using EBP and PCC.
Implications for Knowledge Translation and Professional Development in Audiology
Typical knowledge translation approaches in audiology include publication in peer-reviewed literature and presentations at conferences and seminars. These translation pathways may facilitate the dissemination of knowledge about EBP and PCC, but may not be sufficient for effective implementation of these practices in clinical settings (i.e., a measurable change in clinical behavior toward an increased use of EBP and PCC). In this study, peer-reviewed literature and information from conferences were not ranked as important nor reliable sources of information by the audiologists. This finding is not unique to audiology; clinical implications from research are seldom highlighted in publications and presentations, making it difficult for clinicians to grasp the every-day applicability to their patients (Heiwe et al., 2011; Metcalfe et al., 2001; Moodie et al., 2011). Moreover, as raised in multiple fields, knowing about an evidence-based intervention is not sufficient to adopt or implement this intervention in a scalable and sustainable manner within clinics (Damschroder et al., 2009). As suggested by Moodie et al. (2011), the audiology profession would benefit from working toward an integrated model of knowledge translation where front line clinicians are integrated in the research process, rather than a research driven hierarchical model, where the researchers identify the empirical evidence and inform the clinics about this evidence. An integrated model of knowledge translation would involve an “active collaboration between researchers and research users (clinicians) in […] designing the research questions, shared decision-making regarding methodology, data collection and tools development involvement, interpretation of the findings, and dissemination and implementation of the research results” (Moodie et al., 2011, p.11). The literature relating to knowledge translation and implementation in health care supports that interactive sessions with practitioners can be effective in changing behavior toward EBP and shared decision-making (Davis et al., 1999; Grol & Grimshaw, 2003). Further, frameworks such as the Consolidated Framework for Implementation Research can be used to guide teams of clinicians and researchers toward the most effective implementation strategies (Kirk et al., 2016).
Conclusion
This study suggests that many audiologists value patient preferences, as part of multiple sources of information, when making rehabilitative clinical decisions, which is aligned with PCC. While research evidence was reported to be used less often than expert opinion in complex decision cases, the reported use of empirical research evidence appears to have increased since 1989. Overall, while positive changes toward EBP and PCC appear to have occurred, the audiology profession may further benefit from reassessing its knowledge translation and implementation pathways.
Footnotes
Appendix
Declaration of Conflicting Interests
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This study was supported by the HEARing Cooperative Research Centre, established and supported under the Business Cooperative Research Centres Programme of the Australian Government. The authors acknowledge the support of the Organizing Committee of the XXXII World Congress of Audiology and Audiology Australia for supporting data collection. In addition, the Australian Hearing Hub, an initiative of Macquarie University and the Australian Government, provided financial support for the publication of this research article.
