Abstract
Background
Policy evaluations often focus on ex post estimation of causal effects on short-term surrogate outcomes. The value of such information is limited for decision making, as the failure to reflect policy-relevant outcomes and disregard for opportunity costs prohibits the assessment of value for money. Further, these evaluations do not always consider all relevant evidence, other courses of action, or decision uncertainty.
Methods
In this article, we explore how policy evaluation could better meet the needs of decision making. We begin by defining the evidence required to inform decision making. We then conduct a literature review of challenges in evaluating policies. Finally, we highlight potential methods available to help address these challenges.
Results
The evidence required to inform decision making includes the impacts on the policy-relevant outcomes, the costs and associated opportunity costs, and the consequences of uncertainty. Challenges in evaluating health policies are described using 8 categories: 1) valuation space; 2) comparators; 3) time of evaluation; 4) mechanisms of action; 5) effects; 6) resources, constraints, and opportunity costs; 7) fidelity, adaptation, and level of implementation; and 8) generalizability and external validity. Methods from a broad set of disciplines are available to improve policy evaluation, relating to causal inference, decision-analytic modeling, theory of change, realist evaluation, and structured expert elicitation.
Limitations
The targeted review may not identify all possible challenges, and the methods covered are not exhaustive.
Conclusions
Evaluations should provide appropriate evidence to inform decision making. There are challenges in evaluating policies, but methods from multiple disciplines are available to address these challenges.
Implications
Evaluators need to carefully consider the decision being informed, the necessary evidence to inform it, and the appropriate methods.
Evaluating policies by estimating their causal effects on short-term surrogate outcomes is, in isolation, of limited value for decision making. Evidence for informing decision making needs to link policies to relevant outcomes, costs, and associated opportunity costs and reflect the magnitude and consequences of uncertainty in these estimations. Challenges in program evaluation range across defining the valuation space and comparators, understanding mechanisms of action and effects, estimating opportunity costs, and external validity. Methods from multiple disciplines are available to address these challenges.
Much of the applied health economic research of population- and system-level health policies has focused on the ex post estimation of the causal effects on short-term surrogate outcomes using observational data.1,2 This literature has been supported by methodological developments that facilitate causal inference. 3 However, the information produced by these studies is usually insufficient for decision making. The failure to reflect the impacts on policy-relevant outcomes, for which decision makers can be held accountable (e.g., population health) and the disregard for opportunity costs prohibits the assessment of value for money.1,4 Further, the ex post nature limits their value for decision making, since the information they provide is necessarily available only after the decision to introduce the policy has been made, although it may be useful to inform subsequent decisions (e.g., policy redesign). Studies that focus their conclusions on what previous decisions achieved often display limited consideration of future decisions study results could inform (for example, the introduction of similar policies in other countries). Decision makers need information on a policy’s expected impact at different points in its life cycle; for example, whether to introduce it, maintain it, scale it up, or withdraw it. Further, evidence-informed policy choices are not concerned only with whether a policy should be implemented but also with the timing of when it should be implemented in the face of uncertainty or whether there is value in investing in further research implementing a policy.
Economic evaluation methods, often using the framework of decision analysis, have been developed to inform a range of decisions, including whether the interventions should be introduced, continued with, or disinvested from by health care services and also to consider the value of collecting additional evidence.5,6 However, these methods have not been widely used to assess health policies. Kreif et al 7 found that of 2419 health-related impact evaluations identified from 2010 to 2016, only 42 (2%) included an economic evaluation, and those were generally of poor quality. In contrast to evaluations of health policies, the value for money of clinical interventions (e.g., medicines, procedures, and diagnostics) is routinely assessed in many countries, 8 often through health technology assessment processes, although debates continue about what are suitable methods. 9
Until recently, there has been little attempt to integrate program (or impact) evaluation with economic evaluation methods. However, recent research has started to bring the two together and to integrate learning from other fields such as epidemiology, 10 to consider the value for money of health policies.1,4 This article aims to develop this further by exploring how evaluations could better inform policy choices. We begin by defining the evidence required in terms of the impacts on the policy-relevant outcomes, the costs and associated opportunity costs, and the magnitude and consequences of uncertainty. Then we identify and categorize the key challenges encountered when evaluating policies. Finally, we highlight approaches that have been applied and a series of examples as signposts for readers who wish to explore these methods further.
What Decision Is Being Informed and What Evidence Is Required?
Economic evaluation informs policy choices by providing evidence on the benefits, costs, and opportunity costs of alternative courses of action. An intervention would be considered value for money if its benefits exceed its opportunity costs (i.e., its benefits exceed the benefits that could be generated by alternative use of the same resources). 11 Economic evaluation has been used within health care to consider the value of clinical interventions and by governments to consider the value of a wide range of policies.8,12
To be informative, an economic evaluation needs to report on outcome(s) of relevance to the decision maker(s) involved, based on their objectives and responsibilities. 13 The outcomes not only refer to an intervention’s planned benefits but also include any unintended consequences that can result from policies. 14 To facilitate consistency in decisions and assessment of opportunity cost, it is preferable that the outcomes reflect what is considered relevant to other potential uses of the same resources (e.g., what outcomes would be considered if the resources were instead used for the provision of particular treatments). An economic evaluation should then proceed as the evidential assessment of the impact of the policies on those outcomes, which can be used to compare the intervention’s outcomes with its opportunity costs to determine if the former exceed the latter.
To evaluate value for money, it is necessary to compare an intervention against relevant alternative courses of action. This may constitute a defined set of mutually exclusive options available to pursue specific objective(s), which goes further than just comparing one policy specification with the status quo. 11 For example, an evaluation of a sugar-sweetened beverage tax might warrant consideration of different ways of implementing a tax on sugar. Similarly, a decision may not always be a dichotomy of whether to introduce a policy or not, and other options may be considered; for example, whether to scale up a policy or whether to invest in further research on a policy’s impacts.
Evaluation is inevitably uncertain, reflecting incomplete evidence and knowledge. This uncertainty imposes costs, as choosing the suboptimal policy has negative implications for objectives. Here, we refer not to statistical significance but to decision uncertainty, the probability and consequences of incorrect decisions. Regardless of the risk of error, decisions will have to be made about what course of action to pursue at a given time. By assessing the implications of the decision uncertainty, the decision maker can assess the value of acquiring further evidence to reduce it. On this basis, evidence-informed decision making can expand on the range of decision options beyond the “yes” or “no” to a given policy to include those such as whether to delay an adoption decision until additional evidence becomes available.5,15
Challenges in Evaluating Population and Systems Interventions: A Targeted Literature Review and Classification
We conducted a targeted literature review to identify key challenges in undertaking economic evaluation of policies, many of which apply to any form of evaluation. We identified articles by asking members of an Expert Advisory Group (see supplementary material) for key papers, both methodological and applied, on the evaluation of population- and system-level policies. We reviewed the reference list for these articles to identify further relevant publications. In total, 40 methods articles and 2 reviews of applied papers were identified (see supplementary material).
Each article was reviewed by one member of the research team responsible for reviewing (S.W., A.C., J.A., and M.D.), and the details of any methods challenges raised were extracted. Following the review of all articles, the entire research team (all coauthors) considered the methods challenges raised and categorized them under 8 themes: 1) the valuation space; 2) comparators; 3) timing of the evaluation; 4) mechanisms of action; 5) effects; 6) resources, constraints, and opportunity costs; 7) fidelity, adaptation, and level of implementation; and 8) generalizability and external validity. Supplementary Table S1 sets out a brief description of each theme, the associated challenges, the problem(s) they entail, and key quotes. Below, we summarize the challenges grouped under each theme.
Valuation Space
We define the valuation space as the set of policy-relevant outcomes. When evaluating health care treatments, the focus is often limited to their impact on the health of patients, typically measured using generic units of health such as the quality-adjusted life-year (QALY). This may also be true when considering some population- and system-level policies. However, broader impacts may be relevant for some policies, including those related to health (e.g., mortality) and health care processes (e.g., access to care) but also non–health related outcomes (e.g., labor market participation, financial protection).13,16–19 The unit of analysis for the outcomes of interest may be at the individual level or higher (e.g., firm, hospital).
Delineating and defining the relevant outcomes for evaluation can be challenging. This is particularly the case where there are different decision makers involved with interests in varying outcomes.16–18 Expanding the valuation space to multiple outcomes raises challenges related to the measurement and valuation of the outcomes (e.g., double counting) and to comparing them with the outcomes generated by other policies. 13 The valuation space also needs to be broad enough to capture any unintended consequences. 1 Economic evaluation requires evidence of not only the direct effects of the policies on all outcome(s) of interest but also of its opportunity costs, which ideally would be expressed in terms of these outcomes. Where there are improvements in some outcomes and deterioration in others, some method of aggregating across outcomes is required to decide whether the policy is beneficial (whether done informally or formally). 13
Decision makers may also care about reducing unfair inequalities in the outcomes.16,20 To consider these in an evaluation, evidence on both the baseline distribution of the outcomes and distribution of the policy effects (both direct and opportunity costs) on the outcomes across equity-relevant groups of interest is required. 21 Further information will be needed on the degree of inequality aversion for the outcomes.22,23
Comparators
We define the comparators as the set of alternative courses of actions subject to evaluation. To establish whether a policy is of value requires it to be compared with other potential courses of action that could be used to pursue the same objective(s). Evaluations of health care interventions often focus on alternative options for a single component within a treatment pathway, for example, whether to provide drug A or B. However, health care policies are typically more complex, potentially consisting of multiple components that have an impact across different points in multiple pathways.24,25 As such, there are, in principle, a large number of potential comparators, and establishing the expected impacts of each on the relevant outcomes is analytically and computationally challenging, if not impossible. One potential comparator is the status quo, although this may differ between settings and over time. 25
Time of the Evaluation and Decisions
We refer to the time of the evaluation in respect to the policy’s life cycle, which runs from its conception prior to implementation, through initial roll out, to full implementation, and finally refinement and maturation. Different decisions are required at different time points in its life cycle. Dependent on the timing of the evaluation, there may be different evidential challenges to address. For example, early on, the challenges may stem from limited understanding of the possible outcomes. 1 Later, challenges may stem from lack of data on what the outcomes would be without the policy (counterfactual). The value of a policy may also change over time. For example, if evaluating a policy with large upfront (potentially sunk) costs, these would need to be considered when evaluating whether to introduce it, but following its introduction, these costs may be irrecoverable and do not result in further opportunity costs when considering whether to continue with the policy. The level of decision uncertainty will also vary over time, and decision makers may need to consider whether they require further information to help inform the current decision and whether a decision to implement the policy would prevent further evidence generation or alternatively lead to further learning about its effects. 6
Mechanisms of Action
The mechanisms of action are the causal chains and processes on which a policy acts to produces its effects. Health systems are complex with multiple interacting constituent parts,26,27 many of which may be affected by the introduction of a policy. The various potential mechanisms of action are unlikely to be independent of each other, and the overall impact of a policy will reflect their joint effect, including any interactions or spillovers. To evaluate the overall effect of a policy, it may be necessary to understand how the changes resulting from its introduction affect each constituent part and their interactions, with the failure to do so potentially leading to incorrect estimation of effects. For example, the introduction of co-payments intended to reduce unnecessary care was shown to reduce demand for health care for those patients incurring the co-payments that, if considered in isolation, suggested it had the desired effect. 28 However, further research demonstrated that physicians responded by increasing care and costs to other patients at the same practice, thus increasing unnecessary care. 29
Estimating Effects
Establishing a causal relationship between a policy and policy-relevant outcomes is necessary to determine its value; however, it is challenging. 1 For example, as noted above, the effects of a policy are potentially dependent on its impact and interaction with numerous constituent parts, 1 or it may not be easy to separate the impacts of the policy from other changes occurring contemporaneously.4,30 This is further complicated by the effects of the policy evolving over time, for example, due to learning or bedding in processes.30,31 Health care policies may also have important unintended consequences that will need to be captured.1,4,25,29,30,32 The above all raise challenges around the appropriate methods to estimate a policy’s impacts, including estimation problems associated with causal inference, understanding the processes by which a policy affects outcomes, capturing any changes over time, and having an appropriately broad perspective to capture both intended and unintended impacts.
Resources, Constraints, and Opportunity Costs
For clinical interventions, economic evaluation has used cost-effectiveness thresholds to represent the maximum a health care system might pay for additional health outcomes. The threshold can represent an estimate at the margin of how much health is lost elsewhere for a given amount of expenditure being reallocated (i.e., it represents the opportunity cost). 33 Using this approach, the monetary cost of an intervention can be converted into opportunity costs measured in the same metric as the benefits (i.e., policy-relevant outcomes) to see whether the benefits exceed the opportunity costs. Resource constraints are often depicted as a singular monetary constraint, for example, the health care budget. However, population- and system-level policies may affect multiple budgets (e.g., primary and secondary care) and other nonfinancial constraints (e.g., staff, capital equipment, capacity), which can result in differential opportunity costs per unit of expenditure depending on the particular limits of each constraint.13,34–36
Because of their scale, some policies may have nonmarginal impacts, requiring very large amount of resources (e.g., a new policy requiring a significant percentage of the health care budget) and/or inducing a change in the overall production function (e.g., a new technology that significantly affects the overall health produced by the health care system) such that estimates of the marginal productivity of health care expenditure no longer reflect opportunity costs. 37 Estimating the monetary costs of a policy may itself be challenging, particularly if they are subject to economics of scale (where the marginal resource requirements of a policy decrease as its scale increases) or economies of scope (where marginal resource requirements are dependent on the other services being provided). Finally, the extent of the valuation space may require understanding of opportunity costs on a wide set of outcomes and that an estimate of opportunity cost related to health and health care resources may not be sufficient. 38
Fidelity, Adaptation, and Level of Implementation
A health care policy’s impacts reflect the way in which it is applied and may be altered by factors such as the fidelity of its application, the intensity of effort to apply it, the comprehensiveness with which it is applied, the level of uptake, and the scale of implementation.30,39,40 These may all differ over the life cycle of the policy. For example, experience from the early stages of implementation of a policy is frequently used to inform implementation in the later stages, policies are not always clearly defined and may evolve over time, and staff may adapt their behavior to them. 4 Evaluations should strive to take these factors into consideration when assessing a policy.
Generalizability and External Validity
Generalizability refers to the utility of evidence outside the time or setting in which it was generated. Program evaluation methods that are highly focused on establishing the internal validity of the evaluation are tied to the setting in which the observations were made. However, for decision making, external validity is vitally important, particularly when decision makers are interested in applying policies to alternative settings. 41 The effects of policies are likely to vary across settings and over time. For example, one might expect financial incentives for improving care to have different impacts in the United Kingdom as compared with the United States, given differences in the remuneration structure for clinicians. Careful consideration needs to be given to whether the evidence produced by an evaluation is applicable to the decision being addressed.
What Are the Methods Available to Address These Issues?
Many, if not all, of the challenges outlined are not unique to population- or system-level interventions and arise also in the evaluation of other interventions. Different evaluations will be affected by these challenges to differing degrees, and there are no universal approaches or methods to address all of them. In this section, we highlight some potential approaches and methods available to tackle these challenges and provide some examples in which these approaches have been applied as signposts for readers who wish to see further details.
What Outcomes to Capture and How Should They Be Valued? Defining the Valuation Space
The choice of outcomes to capture represents a value judgment about the key issues of consequence, and this judgment should reflect the views of the relevant decision makers, not the analyst. 13 When considering which outcomes to include, one potential approach is to determine a set of outcomes and to estimate the impact on each of them. 13 When multiple outcomes are considered, an aggregation function (a further value judgment) will need to be imposed to determine if a policy is beneficial overall if there are winners and losers. 13 A cost-benefit analysis or social return on investment approach would aggregate outcomes by attaching monetary values, often derived from individuals’ preferences, to each. 11 Multicriteria decision analysis can also be used to help reach consensus on the impacts on multiple outcomes; however, consideration needs to be given to consistency across decisions and whether opportunity costs are appropriately considered.42–44 An alternative approach is to define a single (universal) outcome measure; examples include approaches such at the QALY, capability approaches, or extended QALYs.45–48 However, such an approach imposes a value judgment that all issues of consequence and tradeoffs between them are appropriately captured within the measure.
If a broad set of outcomes are relevant for decision making, the approach of capturing the impacts on multiple outcomes may be preferable. Further, by capturing the outcomes individually, alternative value judgments around the appropriate methods of aggregation can be considered. Recent examples, looking at policies for alcohol use disorder and air pollution policies, have shown how the impacts on multiple outcomes can be estimated, and different value judgments in their aggregation considered.13,49–51
What Decision Are We Trying to Inform?
Determining the decision options is likely to involve significant scoping work and interactions with the decision makers and other relevant stakeholders to determine the range of options available. 30
Deferring a decision on the implementation of a policy to conduct further research to inform a later decision may be possible.5,52 The ability of a decision maker to consider further research will depend on his or her responsibilities and the nature of the policy being considered.5,6,53 Any evaluation will be undertaken in the face of uncertainty, and reducing this uncertainty generates benefits in terms of better decisions. 54 Value-of-information methods consider the value of further research to reduce decision uncertainty. Research into the adoption of new interventions expanded decision options away from only “approve” or “reject” to include options such as “only in research,” in which interventions are funded only for those included in research studies or coverage with evidence development, whereby an intervention is funded for all while research is conducted.5,6,52 These approaches could be used to inform policy decisions, for example, before a policy’s introduction, a program of further research including only a pilot or partial roll out may be feasible to generate additional evidence.5,6
Vehicles for Evaluation
A key challenge in evaluating policies is establishing their effects on relevant outcomes. This may require bringing together evidence from a range of sources. There is a continuum of approaches for evaluation, ranging from empirical analysis of 1 or more data sets to decision analytic modeling approaches that synthesize evidence from multiple sources.
Empirical approaches rely on having data available, which includes measurements that can be used to establish the impacts of a policy over an appropriate time horizon. The analysis by nature will be a retrospective assessment of the value of the policy over that period in that setting. As such, careful consideration needs to be given to the relevance of the evidence for any subsequent decisions within that or a different setting. In addition, there may be several data sets, each of which has potential relevance to a decision problem and should be reflected in the evaluation.
Decision analytic modeling brings together evidence from the range of available sources to estimate the impacts of policies,11,55 often in unison with statistical synthesis methods.56,57 Modeling approaches to valuing policies can be broadly categorized in 2 ways: 1) linking short-term impacts of the policies from empirical analyses to longer-term outcomes modeled by the impact on individual(s) and 2) modeling the impacts on the systems and individuals jointly. The first of these involves no explicit modeling of the impact of the policy on different constituent parts of the system, instead taking an estimate of the overall impact of these system effects on a surrogate outcome (e.g., on mortality, length of stay) from empirical or other studies and extrapolating the change in the surrogate to the overall impacts on individuals (e.g., QALYs). Recent evaluations using such an approach include assessment of payment for performance mechanisms with mortality impacts linked to lifetime health outcomes and costs.58–60 The second approach models the system and its constituent parts. Approaches under this range from mathematical programming looking at aggregating the independent impacts on the costs and outcomes resulting from different constituent parts (e.g., treatments received), to complex system modeling reflecting the interactions between multiple constituent parts.34,61–63 The mathematical programming approach to the allocation of health care resources has been used to consider budgetary policies, health system strengthening, and investment in resource constraints.34,35,63–66 Dynamic simulation modeling, such as system dynamic modeling and agent-based modeling, have been used to simulate the impacts of health care policies on health care systems and their multiple constituent parts.67,68 A recent review identified 39 studies that have used such approaches to examine the impacts of health care policies on targets such as overstretched resources, length of stay, and undesirable patient outcomes. 68
What Are the Expected Mechanisms of Action of the Policy?
A key challenge in evaluation is understanding the way a policy is expected to affect the relevant outcomes. Theory-based approaches can be used to build this understanding by defining the hypothesized causal chains.69,70 Approaches such as mechanism mapping, in which potential mechanism-context interactions are identified, or group model building, in which a diverse set of practitioners collaborate to develop causal models, can help to establish the expected effects and to provide suitable context-specific adaptations.71,72 Realist evaluation can also play a role, focusing on the context-specific nature of mechanisms of action to understand “what works, for whom, under what circumstances.”73,74 Artificial intelligence methods may also have a role in generating information on the mechanisms of action. 75
Estimating Effects
Establishing a causal relationship between a policy and the outcomes is necessary to determine its value. The estimate of the impact could be on a singular overall measure of effect, which could be a key driver of value or relate to the impacts across different constituent parts of the system. Before a policy is introduced, evidence on its impacts may have to be taken from the literature or other settings, with careful consideration given to its transferability to the context being considered. Methods are available to help adapt estimates in a transparent way (e.g., midrange theory 76 ). If no relevant empirical evidence exists, methods such as structured expert elicitation are available. 77
If a program of research is planned to be conducted as part of the roll out of a policy to inform a future decision, randomized controlled trials (RCTs) or similar approaches may be feasible. 53 Pandya et al. 60 provided a recent example of the use of an RCT to evaluate financial incentives to manage cholesterol levels. However, RCTs have been subject to criticism for the evaluation of policies, notably regarding limits on their external validity and whether the causal effects estimated in the experimental setting generalize to “routine” contexts.78–80 If a randomized comparison is not feasible, an organized program of data collection using a study design that provides an opportunity for rigorous causal inference methods could be considered, although such methods are also subject to issues of external validity. 81 It may be desirable to improve the external validity by considering the results of different causal inference evaluations relevant to different contexts using an analytical modeling approach. Choice of study design can be informed by an explicit consideration of the value of the evidence generated using value-of-information methods.82,83
For analyses following the introduction of a policy, the use of quasi-experimental approaches to establishing causality has been a key focus of the program evaluation literature. 3 These rely on the ability to identify a suitable control against which to judge the policy. A recent example used causal inference methods (synthetic controls) to estimate a mortality impact and a decision analytic model to extrapolate the results to estimate the cost-effectiveness of the UK Quality and Outcomes framework. 59
Generating Evidence on Costs and Opportunity Costs
To establish value for money of a health policy, it is essential to estimate the opportunity costs. Challenges arising from multiple budgets, nonfinancial constraints, and nonmarginal impacts do not require significant deviations from standard cost-effectiveness methods. For multiple budgets or nonfinancial constraints, estimates of the opportunity costs can be captured through marginal productivities for the outcome(s) of interest for each budget or constraint.13,33,84,85 An evaluation of air pollution strategies considered budgets in the National Health Service (NHS), public health, and social care, whereas nonfinancial constraints have been considered in evaluations of eye care services in Zambia and viral load testing in sub-Saharan Africa.34,35,54 When there are nonmarginal impacts, evidence is required on the scale of the impact and the change in the opportunity costs per unit of expenditure. 37 Lomas et al. considered the implications of nonmarginal budget impacts on value for money of new treatments for hepatitis C, which had been estimated to cost the NHS more than £700m per annum (0.7% of total NHS budget).37,86
The challenges raised here around financial costs being used to estimate opportunity costs also raise issues with the use of cost-benefit analysis methods (whereby the benefits measured in monetary terms are directly compared with the monetary costs), 11 which typically assume that opportunity cost is equal to monetary cost. Recent research has considered the extension of cost-benefit analysis approaches for evaluating public expenditure by considering the marginal value of public funds, which is equal to the ratio of beneficiaries’ willingness to pay and the net cost to the public sector. 87
Reflecting System Complexity
Where there are significant interactions within the system, estimating financial cost implications and the effectiveness of policies in isolation may not be sufficient for establishing value for money. Previous work has shown how extensions to the mathematical programming approach on which cost-effectiveness analysis is based can be used to evaluate different types of system-strengthening interventions.63,66 For example, Morton et al. 66 considered the optimal spend on system strengthening, which increased the effectiveness of treatments in an HIV prevention program. However, these approaches are informationally expensive, requiring estimates of the costs and effects of all the independent interventions in the system and the impact that the policies have on each of these. Alternatively, others have highlighted the need to develop whole-system models to evaluate policy outcomes and opportunity costs. The Thanzi La Onse project is developing an individual-based comprehensive whole system and an all-disease model of the Malawian health care system, which will be capable of examining the impacts with and without population- and system-level health policies. 62
Considerations for the Analyst
In this section, we have signposted a range of approaches of which we are aware to tackle challenges in evaluating population- and system-level health policies. Many of the approaches are not mutually exclusive and can be used in combination, and the choice of methods requires considering their pros and cons in light of the specific evaluation being undertaken. Further, the choice of methods may be limited by the agency of the decision maker and analyst and by the structural circumstances of the evaluation. For example, where the decision maker lacks agency in research, producing information on the value of further research may be unwarranted. Evaluations are not costless, and analysts will have to use their scientific judgment to select approaches that are both appropriate and feasible within the resources available. Regardless of the approach taken, the analyst should engage with the decision makers involved to ensure the evaluation will provide appropriate evidence.
Conclusions
When evaluating health care policies, it is essential to consider both their effects and their opportunity costs to establish whether they represent value for money. However, program evaluation of population- and system-level health policies has often focused on estimating causal effects on short-term surrogate outcomes. These evaluations are of limited value for decision making as they fail to reflect the policy-relevant outcomes and disregards opportunity costs. This article has aimed to show how the evaluations of such policies could better inform decision making by defining the evidence required to inform policy choices in terms of the impacts on the policy-relevant outcomes of interest, the costs and associated opportunity costs, and the magnitude and consequences of uncertainty. The article also identified key challenges described in evaluating population- and system-level policies and examined methods that can address those challenges. We would advocate that further method development is necessary but that a multidisciplinary approach bringing together health economics and adjacent fields such as epidemiology and mathematical modeling will improve the evaluation of population- and system-level policies.
Supplemental Material
sj-docx-1-mdm-10.1177_0272989X211016427 – Supplemental material for Program Evaluation of Population- and System-Level Policies: Evidence for Decision Making
Supplemental material, sj-docx-1-mdm-10.1177_0272989X211016427 for Program Evaluation of Population- and System-Level Policies: Evidence for Decision Making by Simon Walker, Aimee Fox, James Altunkaya, Tim Colbourn, Mike Drummond, Susan Griffin, Nils Gutacker, Paul Revill and Mark Sculpher in Medical Decision Making
Footnotes
Acknowledgements
We would like to acknowledge the members of our Expert Advisory Group for providing suggestions for the articles on which our targeted review was based and for providing comments on the article. We would also like to acknowledge Anna Heard (Poverty Action) for providing comments on the article and Angela Bates (Northumbria University) and members of the audience for discussing an earlier draft of the article at the Health Economics Study Group in Newcastle in 2020.
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Financial support for this study was provided by the National Institute for Health Research (NIHR; NIHR Global Health Econometrics & Economics Group) using UK aid from the UK government. The funding agreement ensured the authors’ independence in designing the study, interpreting the data, writing, and publishing the report. The views expressed in this publication are those of the authors and not necessarily those of the NIHR or the Department of Health and Social Care.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
