Hostname: page-component-78c5997874-j824f Total loading time: 0 Render date: 2024-11-10T10:41:25.780Z Has data issue: false hasContentIssue false

Behavioural Insights and Public Policy in Canada

Published online by Cambridge University Press:  28 April 2023

Vincent Hopkins
Affiliation:
Department of Political Science, University of British Columbia, Buchanan Building, Block C425, 1866 Main Mall, Vancouver, BC V6T 1Z1, Canada Johnson Shoyama Graduate School of Public Policy, University of Saskatchewan, 101 Diefenbaker Place, Saskatoon, SK S7N 5B8, Canada
Andrea Lawlor*
Affiliation:
Department of Politics and International Relations, King's University College, Western University, 266 Epworth Ave., London, ON N6A 2M3, Canada
*
*Corresponding author. E-mail: andrea.lawlor@uwo.ca
Rights & Permissions [Opens in a new window]

Abstract

Much of political science rests on assumptions about how policy makers and citizens behave. However, questions remain about how public policy can improve the government–citizen relationship. In this research note, we present behavioural insights (BI) as one way to address this gap. First, we argue that BI can be strategically used both to alleviate administrative burdens and to enhance citizen experience. Second, we argue that BI interventions can assist in several stages of the policy process, strengthening causal inferences about policy efficacy. Third, we present original data from Canada's ongoing experimentation with BI across multiple jurisdictions and areas of public policy. We conclude by acknowledging the myriad pathways through which BI research can engage with public policy to support the enhancement of citizen-oriented service delivery.

Résumé

Résumé

Une grande partie de la science politique repose sur des hypothèses concernant le comportement des décideurs politiques et des citoyens. Cependant, des questions subsistent sur la manière dont les politiques publiques peuvent améliorer la relation entre le gouvernement et les citoyens. Dans cette note de recherche, nous présentons l'introspection comportementale (IC) comme un moyen de combler cette lacune. Tout d'abord, nous soutenons que les IC peuvent être utilisées stratégiquement pour alléger les charges administratives et améliorer l'expérience des citoyens. Deuxièmement, nous faisons valoir que les interventions de IC peuvent aider à plusieurs étapes du processus politique, renforçant les inférences causales sur l'efficacité des politiques. En troisième lieu, nous présentons des données originales tirées de l'expérimentation continue des IC au Canada dans de multiples territoires de compétence et domaines de politique publique. Nous concluons en reconnaissant la myriade de voies par lesquelles la recherche sur les IC peut s'engager dans la politique publique pour soutenir l'amélioration de la prestation de services axés sur les citoyens.

Type
Research Note/Note de recherche
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s), 2023. Published by Cambridge University Press on behalf of the Canadian Political Science Association (l’Association canadienne de science politique) and/et la Société québécoise de science politique

In recent years, political scientists have shed new light on the political consequences of seemingly irrational decision making by various actors in the political realm. Research on political behaviour shows voters often use mental shortcuts—that is, heuristics—when thinking about public policy. These include elite partisan cues (see, for example, Pickup et al., Reference Pickup, Stecula and Linden2020), a disproportionate focus on short-term policy impacts (Jacobs and Matthews, Reference Jacobs and Matthews2012), opinion polls (Singh et al., Reference Singh, Roy, Fournier, Blais, Laslier and Van der Straeten2016) and ideology (Caruana et al., Reference Caruana, McGregor and Stephenson2015). And voters are not the only ones. Public servants and politicians exhibit similar “fast thinking” about policy, leading to predictable cognitive biases (Kahneman, Reference Kahneman2013). These patterns are consistent with classic concepts from behavioural science, including the Asian disease experiment (Sheffer and Loewen, Reference Sheffer and J. Loewen2019), anchoring (Bellé et al., Reference Bellé, Cantarelli and Belardinelli2018), escalation of commitment (Sheffer et al., Reference Sheffer, Loewen, Soroka, Walgrave and Sheafer2018), biased information processing (Doberstein, Reference Doberstein2017b) and framing effects (Walgrave et al., Reference Walgrave, Sevenans, Van Camp and Loewen2018). Much of this work uses the experimental method, which is an increasingly popular method in public administration research and practice (Doberstein, Reference Doberstein2017a).

As behavioural science has become more popular in political science, so has the applied discipline of behavioural insights (BI) gained prominence in government. The rapid growth of BI units within the public sector has been well documented in countries such as the United Kingdom (Halpern, Reference Halpern2015; John, Reference John2014), the United States (Congdon and Shankar, Reference Congdon and Shankar2015; DellaVigna and Linos, Reference DellaVigna and Linos2022) and Australia (Jones et al., Reference Jones, Head and Ferguson2021). Yet the situation in Canada remains underexplored, with little empirical evidence about the systematic use of BI across different policy contexts (though see Lindquist and Buttazzoni, Reference Lindquist and Buttazzoni2021, for an important exception).Footnote 1 This gap is surprising given Canada's leading role in BI and the emerging work on policy labs (Evans and Cheng, Reference Evans and Cheng2021). The country was an early adopter of the technique and has pioneered its application (Castelo et al., Reference Castelo, Hardy, House, Mazar, Tsai and Zhao2015; Robitaille et al., Reference Robitaille, House and Mazar2021). In this research note, we ask to what degree, in what policy areas and to what effect has BI been applied in Canada? To answer these questions, we provide an environmental scan of BI in Canada using a new, original dataset of all publicly facing BI projects run by the governments of Canada, Ontario and British Columbia since 2015. We also provide original data from interviews with staff in BI units across three jurisdictions. Our research note begins with a definition of BI, situated within the broader field of behavioural public policy. Next, we highlight how BI interventions can support the policy process and improve policy efficacy. Third, we present empirical data on the size and scope of the existing BI infrastructure in Canada, the number of BI projects by policy area, the efficacy of those projects, and staffing resources over time. We conclude by acknowledging the myriad pathways that BI can support the enhancement of citizen-oriented service delivery and by pointing to the ways in which political science as a discipline can engage with BI in discussions on policy tools, design and efficacy. In Canada, as in many advanced industrial democracies, many administrative processes do not work well for priority populations. We advocate for greater use of BI, especially as a way to reduce administrative burden and improve the lived experience of Canadians.

What Are Behavioural Insights?

Legislative actors advance policy goals, but public servants handle the specifics of policy design and implementation, using research and evidence (French, Reference French2018). Policy success depends on balancing competing considerations, such as financial inputs and policy goals, as well as citizen behaviour. As citizens access public services, they confront technical and cognitive factors that may produce suboptimal decisions (for themselves and/or the government). Citizens can also face cognitive overload (Ly and Soman, Reference Ly and Soman2013), which overwhelms their ability to effectively access public services (McFadden, Reference McFadden2006). For example, consider a municipality that institutes a small fine for violating a new bylaw. The city spends considerable effort designing the optimal fine amount. However, the online portal for paying the fine is clunky and unintuitive. A citizen who wants to make such a payment tries repeatedly, only to get frustrated and abandon the task. The citizen is no better off (having to engage in a more laborious trip to city hall to pay the fine) and the city is no better off (delaying payment and having to engage a more costly and inefficient mode of processing the payment).

One innovative solution to this problem is behavioural public policy (BPP). BPP applies a behavioural lens to the policy process, with special focus on design, development, implementation and evaluation (Loewenstein and Chater, Reference Loewenstein and Chater2017; Oliver, Reference Oliver2017).Footnote 2 BPP's theoretical underpinnings are driven by the rationalist and behavioural sciences at the nexus of policy, economic and psychological research. To this, policy makers apply experimental and data-driven methodologies to make design recommendations, improve implementation and evaluate policy effectiveness. BPP focuses on perceived utility and incorporates behaviourally motivated techniques, including BI. These techniques often simplify services and improve outcomes for citizens and government (Chen et al., Reference Chen, Bendle and Soman2017; Ewert, Reference Ewert2020).Footnote 3 The term BI, as we apply it here, refers to the mechanism that underlies policy design and implementation.Footnote 4 BI may incorporate incentives (explicit inducements) and what Thaler and Sunstein (Reference Thaler and Sunstein2008: 6) call a “nudge” or “any aspect of the choice architecture that alters people's behavior in a predictable way without forbidding any options or significantly changing their economic incentives.”Footnote 5 Common behaviours include applying for public services/benefits, saving money, improving personal health, and adherence with laws or regulations (for example, paying taxes). BI can help encourage or discourage behaviours. BI can help promote pro-self and pro-social behaviours and encourage citizens to avoid “regret loss” from choices based on intuition and emotion rather than deliberation and reasoned analysis (Hagman et al., Reference Hagman, Andersson, Västfjäll and Tinghög2015). Above all, BI avoids rigid commitments that disengage citizens by removing their choices.

BI is unique because it integrates psychology, political and economic theory, and rigorous testing through randomized control trials. BI has its intellectual roots in behavioural research in Daniel Kahneman and Amos Tversky's work in the 1970s and borrows from their research on prospect theory (Reference Kahneman and Tversky1979) and mental shortcuts (Tversky and Kahneman, Reference Tversky and Kahneman1974). With the recognition that BI is not the solution to every policy problem (Richard Thaler (quoted in Nesterak Reference Nesterak2021) argues that “nudge is part of the solution to almost any problem but is not the solution to any problem”) and that policy makers must judiciously decide when to apply BI, the technique has recently become mainstream in public policy.Footnote 6 For example, McDavid and Henderson (Reference McDavid and Henderson2021) persuasively argue that the time has come for greater professionalization of BI, drawing insights from the field of evaluation to suggest a path forward. Mainstreaming BI is possible in part due to recent progress on ethics and ethical frameworks (for example, Delaney, Reference Delaney2018; Lepenies and Malecka, Reference Lepenies, Małecka, Poama and Lever2019; Lades and Delaney, Reference Lades and Delaney2022). These have allowed the field to more confidently advance practical discussions of how and when to apply BI (Ewert, Reference Ewert2020: 338).

As BI gains popularity, scholars are investigating its effectiveness, focusing on replication and publication bias. Some BI-style interventions have failed to replicate, even in controlled settings. Prominent examples include priming effects (Cesario, Reference Cesario2014) and inconsistencies in the ego depletion effect (for example, Carter et al., Reference Carter, Kofler, Forster and McCullough2015, Dang et al., Reference Dang, Barker, Baumert, Bentvelzen, Berkman and Buchholz2021). One challenge is the potential for heterogeneous effects across interventions, populations and field settings. Rather than generalize about its overall effectiveness, researchers are increasingly asking when and where BI can be effective (Bryan et al., Reference Bryan, Tipton and Yeager2021). A related concern involves systematic biases in the reporting of BI trials. Consistent with other fields (see, for example, Franco et al., Reference Franco, Malhotra and Simonovits2014; Esarey and Wu, Reference Esarey and Wu2016), meta-analysis shows evidence of publication bias in peer-reviewed behavioural science (Hummel and Maedche, Reference Hummel and Maedche2019). Using internal data from two large BI units in the United States, DellaVigna and Linos (Reference DellaVigna and Linos2022) show publication bias tends to exaggerate the effectiveness of BI interventions. In short, the evidence suggests that BI can be effective, that its effectiveness varies by context, and that academic publications involving BI may overstate the magnitude of these effects.

Situating BI in the Public Policy Literature and Process

The traditional policy cycle includes agenda setting, policy formulation, decision making, implementation and evaluation (Howlett et al., Reference Howlett, Ramesh and Perl2020). Classically, the cycle envisions senior policy makers and their staff at the centre of a top-down policy process (Wegrich and Jann, Reference Jann, Wegrich, Fischer, Miller and Sidney2006). However, the rise of behaviourally informed policy implementation brings a renewed emphasis on citizens as end-users of public services (Pal et al., Reference Pal, Auld and Mallett2021: 235–36). Brock similarly notes the pivot to the “deliverology” model of public service that is meant to transcend “the traditional obstacles to change in the public sector that [include] bureaucratic lethargy, lack of expertise, intransigence and risk aversion” (Reference Brock2021: 230). In one sense, this behavioural turn expands on previous studies of implementation that focus on citizen experiences, such as “street- level” bureaucracy (Brodkin, Reference Brodkin2012; Lipsky, Reference Lipsky2010), “bottom-up” implementation (Sabatier, Reference Sabatier1986; Thomann et al., Reference Thomann, van Engen and Tummers2018) and “backward mapping” (Elmore, Reference Elmore1979). But even here, the perspective is usually that of a policy maker or public manager, rather than the citizen. In short, there is a gap between prominent theories of public policy (which focus on decision making) and questions of citizen engagement (which focus on service-level interactions with government). Underlying this gap is the 40-year decline in institutional trust and confidence in government.Footnote 7

With the rapid expansion of the state in the postwar era, citizens are presented with a great number of areas and ways in which government can provide services. But this wealth of choice has come with a problem: undersubscription and unequal access to public services. Low uptake is particularly concerning because it disproportionately impacts the least resourced populations, such as seniors, immigrants, refugees, Indigenous peoples, and youth. Compounding this complexity is the digital divide and the transition to digital government (Longo, Reference Longo2017), particularly in the pandemic and post-pandemic environment. Research by the Citizens First initiative suggests a growing challenge for service delivery in Canada. Their 2020 report uses polling data to show that overall satisfaction with public services is at a two-decade low (Citizen First, 2021).

In this context, BI has proven a timely resource for policy makers. As Howlett notes, “conceptual work on procedural policy instruments [has tended to lag] behind administrative practice” (Howlett, Reference Howlett2013: 119). BI is already firmly present in Canadian public policy making. As a diagnostic tool, BI helps policy makers focus on policy goals, anticipate delivery gaps and increase service uptake. One of the most effective ways to increase uptake is to reduce administrative burden (Herd and Moynihan, Reference Herd and Moynihan2019; Moynihan et al., Reference Moynihan, Herd and Harvey2015). Burdens arise when an individual experiences program or service enrolment or engagement as onerous (Burden et al., Reference Burden, Canon, Mayer and Moynihan2012). Burdens result from costs imposed on the user, such as the hassle of figuring out eligibility requirements (learning costs), potential stigma in accessing public services (psychological costs) and the time/effort required to fill out documentation (compliance costs) (Herd and Moynihan, Reference Herd and Moynihan2019). In this way, BI can be considered a type of procedural policy instrument capable of reducing burdens and narrowing this divide. What appears on the surface as an operational challenge (improving policy uptake) is really an interconnected chain of policy design and service delivery choices that influence how citizens experience government.

Policy makers are increasingly optimistic about the potential for BI to improve public policy and service delivery. In a recent survey of 719 Canadian public servants, Doberstein and Charbonneau (Reference Doberstein and Charbonneau2020) find BI teams (and innovation labs more generally) are now seen as the single most important source of public sector innovation. However, it would be naive to suggest that more innovative service delivery will remedy all problems of citizen engagement. Failing to engage with citizens can be viewed as the product of choices by politicians—the result of purposefully underinvesting in the welfare state. It is therefore no accident when the most underserved are those with the fewest resources to access the system (Banting, Reference Banting, Banting, Hoberg and Simeon1997, Reference Banting2010). In this way, policy design and implementation are deeply political.

Taking into account the literature on administrative burden, policy makers must be careful not to use BI solely as a solution to “middle-class” problems—such as saving for retirement, buying energy-conserving appliances, and so on (Benartzi et al., Reference Benartzi, Beshears, Milkman, Sunstein, Thaler, Shankar, Tucker-Ray, Congdon and Galing2017). Policy makers who apply BI have demonstrated its potential to yield high returns at low cost—but, in many cases, these successes are concentrated among those who are already capable of effectively interacting with government. If BI works well for the middle class but does not reduce administrative burden among those with fewer resources, it will not realize its potential to advance democratic, pro-social outcomes. It is equally important that policy makers understand the potential for heterogeneous results, such as differences by race, gender and income. BI should not be applied without consideration for resource disparities, lest the public view it simply as “policymaking by other means” (Herd and Moynihan, Reference Herd and Moynihan2019).

For these reasons, an ethical lens is crucial when using BI, especially when selecting targeting behaviours and considering downstream impacts. For example, a Canadian government experiment aimed to increase women's enrolment in the military using Facebook ads with different appeals (Privy Council Office, 2017). But MacDonald and Paterson (Reference MacDonald and Paterson2021) argue that the evidence is mixed about whether increasing women's participation in the Armed Forces will increase equity. They cite research on the Royal Canadian Mounted Police (Bastarache, Reference Bastarache2020) to point out that when women enter traditionally male-dominant professions, it may accompany increased reports of sexual assault and harassment. We do not suggest it is inherently unethical to nudge women into the Armed Forces; however, we believe an ethical lens is required both in the scoping process and in post hoc analysis.

Data, Methods and Analysis

Our analysis is a four-pronged environmental scan. An environmental scan is a method to gather information and identify opportunities and risks for future decisions. The method involves various data sources, including both published and unpublished information (Porterfield et al., Reference Porterfield, Hinnant, Kane, Horne, McAleer and Roussel2012). Environmental scans are not exhaustive but rather give a general picture of a field (Wheatley and Hervieux, Reference Wheatley and Hervieux2020). Our environmental scan considers (1) the number and location of existing BI units in government, (2) BI projects from public-facing reports spanning 2015 (the first published results from a publicly reported BI project by government in Canada) to 2022 by policy area, (3) BI projects by methodology and (4) BI projects by rate of success. While we are relying on self-reported government data, and recognize the associated concerns around governments only presenting “successes,” we observe that governments are increasingly following an “open science” model of BI. For example, the government of British Columbia has preregistered several field trials, which limits the degree to which public reporting may represent (only) positive outcomes. Preregistration is an encouraging trend given the concerns over replication and publication bias described earlier in this research note.

Finally, we supplement our environmental scan with interview data conducted with BI personnel across five BI units in three jurisdictions. Semistructured interviews took place remotely between July 2020 and July 2021. The thematic foci of these conversations included project design, responses to project successes and failures, and methodological approaches, as well as partnerships across government units. While we do not present a full thematic analysis of the interview data, we do use the observations that were noted across multiple offices to complement our empirical findings. Since Canada's BI field is comparably small and there are only three stand-alone units at present (though several mini units within government departments), we report interview comments anonymously without geographic or jurisdictional identification.

BI units in Canada

The nudge movement was an important catalyst in the development and institutionalization of BI in Canada. The template for BI teams is largely found in the 2010 British initiative under David Cameron's Cabinet Office. One of the authors of Nudge (2008), Richard Thaler, served on its advisory board, with social and political scientist David Halpern serving as its chief executive. In 2015, Barack Obama established the Social and Behavioral Sciences Team in the White House Office of Science and Technology Policy. This followed the appointment of Nudge co-author Cass Sunstein as administrator of the Office of Information and Regulatory Affairs in 2009.

Canada's BI experience reflects policy learning from these international efforts. In 2017, the federal Impact and Innovation Unit (IIU), previously known as the Innovation Hub, was founded as part of the federal government's Innovation Agenda. The IIU is embedded within the Privy Council Office. The interdisciplinary team consists of approximately 15 individuals with a variety of educational backgrounds, including education and neuroscience. The stated goal of the IIU is to reduce barriers to innovation within government and to “leverage the benefits of impact measurement to support evidence-based decision-making” (Impact and Innovation Unit, 2021). Examples of completed projects include an effort to increase charitable giving (2017) and help low-income families access the Canada Learning Bond for post-secondary education (2018). The IIU mostly runs projects through partnerships with other government departments such as the Canada Revenue Agency and Employment and Social Development Canada. Typically, these partnerships result in randomized controlled trials (RCTs) to test potential solutions and improve policy outcomes.

Several provinces have their own BI teams (full listing in the supplementary material). Designs vary by province. Ontario's BI unit was developed in 2015 within the Treasury Board Secretariat. Its policy successes include an increase in the uptake of the human papillomavirus (HPV) vaccine, increasing cervical cancer screening rates for women, and enhanced engagement with provincial and municipal online services. Using RCTs, the BI unit reported a 143 per cent increase in organ and tissue donor registrations, an 82 per cent increase in correct recycling of organics, and increased use of the province's online portal to complete administrative tasks such as renewing driver's licence plate stickers (Government of Ontario, 2018). Much like the federal unit, the Ontario team works with other sectors of government. Correspondingly, its focus is jurisdictionally tailored to problems that fall under the province's purview. It also has partnerships with academia—specifically, the Behavioural Economics in Action at Rotman and the Southern Ontario Behavioural Decision Research Group, which support RCT design and analysis.

British Columbia's Behavioural Insights Group (BIG) started in 2016 and is situated within the Public Service Agency. Similar to the federal and Ontario BI teams, the British Columbia team partners with other government ministries to design and test policy solutions. In its first four years, the unit designed 19 RCTs and 35 behavioural “lenses,” which offer advice to client ministries about how BI might improve a policy, program or service. Starting in 2017, Alberta's CoLab engaged in BI work but was disbanded under the United Conservative government in 2020. Nova Scotia is in the process of expanding its BI capacity and has already reported notable success with an organ donation campaign in 2019. Several other provinces support behavioural labs, which may support community partnerships or industry, and their focus is not necessarily on BI alone (Evans and Cheng, Reference Evans and Cheng2021). However, they often feature BI approaches and may work in tandem with BI units. And while no municipality has a stand-alone BI team at present, several larger cities have ongoing funding for projects, such as CityStudio (Vancouver) or Civic Innovation (Toronto) that focus on enhancement of citizen engagement and improvements to service delivery. Outside of government, there is a Canadian office of the Behavioural Insights Team, which is a social purpose company that emerged from the original BI unit in the UK government. Established in 2019, the office follows a consultancy model, working with government actors and not-for-profit organizations to support BI projects.

BI projects

We next address our data collected from public reports published by the three active BI units in Canada (the federal IIU and the BI units in Ontario and British Columbia). Our dataset includes a total of 59 projects across 13 policy areas (see supplementary material for more information on coding). Our data is supplemented by email correspondence and interviews with individuals in five BI units (federally and provincially) to address discrepancies or data limitations in the reports and to gain insight into project efficacy and rollout. Where appropriate, we supplement the data with observations from our interview data. Given the small number of projects, the analysis is necessarily limited. However, we are able to provide top-line information in policy coverage, methodological design, and goal-meeting or project efficacy.

Figure 1 illustrates the number of projects by policy area (full dataset available in supplementary materials). A clear majority of projects (N = 30) were initiated in the areas of health (N = 17) and government operations (N = 13; for example, administrative fines, documentation renewals), though we do see a variety of projects across social and administrative policy areas. Health projects range from the Ontario BI unit's “Supporting Ontario's COVID-19 Response” to helping physicians improve prescribing behaviour to reduce the prescription of opioids. A total of 21 per cent of these projects were initiated through the federal IIU, while the remaining 79 per cent come from provincial offices (68 per cent from Ontario and 11 per cent from British Columbia). Interview participants R1 and R2 both noted the suitability of administrative operations to BI trials because they often entail fewer ethical concerns (compared with projects involving social policy) and because they lend well to experimentation using digital touchpoints (that is, sending notices, online platform-based delivery). During the early days of the COVID-19 pandemic, these factors facilitated the continuation of BI work.

Figure 1 Number of Projects by Policy Area.

We also evaluate the form of these projects, specifically the methodological approaches taken. The majority of publicly released BI projects are field studies (N = 32)—that is, interventions or initiatives that apply BI principles in real-world settings. Another 22 projects focused on user-centred design; four focused on using survey experiments to test BI assumptions. All of these projects are evidence-based responses to policy challenges, with RCTs as the dominant methodological design (49 per cent) and the remainder using survey research or qualitative designs such as focus group analysis and post hoc reviews by experts in the field.Footnote 8 Interviews with R2 and R3 yielded a preference for a multi- or mixed-method approach to validate findings with supplementary stakeholder interviews, or to engage priority communities with requests for feedback. Respondents demonstrated a responsiveness to broad critiques of BI as being too quantitatively focused, and they accentuated the importance of consultation with program beneficiaries to reduce the risk of losing the “human” aspect of their goal.

Project efficacy is measured through a very simple metric: whether the BI project yielded a significant result on its stated goal. Importantly, the scale of BI results is moderate. Some projects hope to increase uptake by a modest 3 to 5 per cent; however, given the scale of these interventions, this can translate into tens—or even hundreds—of thousands of citizens. In every interview, practitioners noted the challenges of articulating the significance of what appeared to be modest effects when demands on government resources are so large. However, three respondents also observed that government partners would quickly embrace these findings when they viewed them as sustainable ways to enhance participation at a low cost to the budget unit. Many projects that we tracked are still ongoing and therefore may have only reported preliminary data. However, of those that are registered as complete, 80 per cent were successful at attaining their goal. These outcomes ranged from reducing delinquency on tax collection payments by 35 per cent in British Columbia to increasing applications for the Ontario Electricity Support program for low-income Ontarians.

One might reasonably ask how confident we can be that the BI interventions were the cause of these impacts. We suggest confidence depends on research design. When the project involves an RCT, randomization to treatment and control ensures an unbiased estimate of the impact (for example, intent-to-treat effect, average treatment effect). When we restrict our analysis to only those completed BI projects involving an RCT (N = 25), we find a similar efficacy as the full sample: 80 per cent of BI RCTs in Canada were successful at attaining their goal.

Finally, we use Canada's Government Electronic Directory Services to estimate changes in the number of BI personnel over time. We collect cross-sectional directory data across seven timepoints, spanning 2016 to 2022 (see supplementary material for more information). For each timepoint, we identified employees who either worked in units containing the phrase “behaviour” or those who worked at the IIU. We then manually reviewed each case, excluding observations where the employee did not appear to work in BI. In Figure 2, we tally the number of personnel by year. We find that the number of BI staff has increased tenfold over six years, with an average annual increase of approximately 48 per cent. In 2016, six federal employees worked in BI. By 2022, there were 63.Footnote 9 Further, our data allow us to map the spread of BI across departments. For example, in 2016, BI staff worked in just two departments: the Canada Revenue Agency (N = 1) and Privy Council Office (N = 5). By 2022, we find BI staff in 10 departments, including the Public Health Agency of Canada (N = 16), Immigration, Refugees and Citizenship (N = 5), Treasury Board Secretariat (N = 3) and Environment and Climate Change (N = 2). Growth has been especially strong in the Public Health Agency. In 2019, prior to the COVID-19 pandemic, the agency reported just one BI employee. By 2022, the BI team had grown to 16.

Figure 2 Growth of BI staff in the Government of Canada, 2016–2022.

Note: Data collected from the Government Electronic Directory Services.

Going Forward

Behavioural science is changing public policy, and, with it, political science. This research note provides four empirical contributions about BI in Canada. First, we observe that governments are actively investing in BI as a policy-making tool.Footnote 10 Governments invest in BI to increase program success and more efficiently use financial resources. Second, governments are expanding the scope of BI, moving beyond central agencies to line departments. The number of BI practitioners across government has increased dramatically—by our estimates, a tenfold increase in just six years. Third, our data show that, in general, BI initiatives are working. A critic might observe that governments are only likely to advertise successful projects, leaving failed projects firmly within the private reports of civil servants. However, our data show that governments have acknowledged several instances where BI projects have come up short. In part, this is because the methodological approach to BI embraces failure as a part of design and evaluation. Finally, as demonstrated in our reporting and data analysis, Canadian governments are using BI across policy domains, with a clear emphasis on improving government operations and service delivery, health, and social welfare.

From a comparative perspective, our findings dovetail with recent research on the application and methodology of BI across contexts. Like others (for example, Hummel and Maedche, Reference Hummel and Maedche2019: 52), we find that BI projects in Canada are most common in the domain of health. Similarly, we find a growing acceptance for non-experimental BI work in Canada, which parallels recent research from Australia (for example, Ball and Head, Reference Ball and Head2021: 114–16). Nonetheless, we recognize several limitations of our Canada-focused environmental scan. First, limited data availability mean we cannot report on some projects. For example, we find few instances where Canadian governments discuss using BI to improve tax collection. Yet media reports (Beeby, Reference Beeby2018) and our own data on the number of BI personnel suggest the Canada Revenue Agency is actively using BI to improve tax collection. Additionally, interview data that we have collected indicates numerous projects in progress across multiple policy domains that are not yet ready for disclosure. We argue that even without the full universe of BI data, our study includes many policy initiatives that reflect a particular approach to policy making—one that should be recognized in political science. Additionally, no policy scholar has a complete list of government projects to evaluate, owing to the same data considerations we face here (for example, access to information, privacy, secrecy). A second limitation involves how to define a BI project. As we have noted above, distinguishing between what the government might formally call a BI project and instances where BI tools or approaches have been used informally means that we may have undercounted our observations of BI in government.Footnote 11

In closing, we reflect on Michael Atkinson's 2013 presidential address to the Canadian Political Science Association. Atkinson (Reference Atkinson2013) urged the discipline to embrace its roots in social and behavioural science. He framed this as an empirical and normative calling: to improve governance and, by extension, the lives of Canadians. In many ways, the discipline has taken up the charge. So too have some parts of government. In this note, we reflect on BI as a potentially powerful tool that can help public policy practitioners and scholars live up to Atkinson's ideal and strengthen citizen–government engagement in Canada. This can include external evaluation of BI policy design and implementation on its own merits but could also extend to collaboration between BI practitioners and political scientists, who bring a rich skillset around research questions and methodological design, while simultaneously taking a critical lens to governments’ policy design choices.

This research note highlights the fertile ground that exists to explore a robust research agenda on BI or with BI. Two streams clearly emerge: (1) evaluating the BI projects conducted by government and (2) using BI either with or independent of government as a tool to generate insights around program delivery. We make no claim that BI is the only or best tool to resolve any policy problem. We merely suggest it is an underused tool for resolving policy challenges in novel and cost-effective ways. We hope political scientists will take up this call.

Supplementary Material

To view supplementary material for this article, please visit https://doi.org/10.1017/S0008423923000100

Footnotes

1 For a recent discussion about BI within the federal government, see a recent chapter by Hardy et al. (Reference Hardy, Khan, Paleja and Ruggeri2021).

2 In differentiating BPP from BI, Ewert (Reference Ewert2020) describes BPP as a “multi-purpose approach that allows the application of Behavioural Insights ‘throughout the policy process’ and in combination with regulatory policies.” Thus, BPP involves a behavioural lens to think about policy problems and solutions, while BI focuses specifically on choice architecture and individual decision making, with special attention to the design and implementation process.

3 Here, we think of outcomes not solely in terms of financial maximization but also in terms of improving policy compliance or simply improving the government–citizen relationship.

4 We situate this conception alongside other popular definitions. For example, Hallsworth and Kirkman (Reference Hallsworth and Kirkman2020: 1) defined “the behavioral insights approach” as one that “applies evidence about human behavior to practical problems.” The Organisation for Economic Co-operation and Development (2017) defines BI as “an inductive approach to policy making that combines insights from psychology, cognitive science, and social science with empirically-tested results to discover how humans actually make choices.”

5 Underscoring the idea of nudging is a broader philosophical approach of libertarian paternalism. This term combines the freedom of choice valued by libertarians and the paternalistic “self-conscious efforts, by institutions in the private sector and also by government, to steer people's choices in directions that will improve their lives” (Thaler and Sunstein, Reference Thaler and Sunstein2008: 8). Key to this idea is the maintenance of a set of choices for the citizen, such as making it easy to opt out of a policy or program.

6 Such mainstreaming is especially impressive given the “adhocracy” that defines the establishment of open innovation units in government, including BI teams (Lindquist and Buttazzoni, Reference Lindquist and Buttazzoni2021).

7 Consider the Canadian Election Study (Reference Stephenson, Harell, Rubenson and Loewen2020), which reports that over 55 per cent of Canadians trust the public service “a great deal’ or “quite a lot,” compared with only approximately 35 per cent who trust the federal or provincial governments to the same degree.

8 We recognize that there may be limits to RCTs around recognizing subgroup differences in a context of reporting high-level aggregate responses. To mitigate this issue, we point to the ways in which BI teams have often supplemented RCTs with additional qualitative tools, such as focus groups, interviews or expert reports, to look at the ways in which the efficacy of the intervention may vary across subgroups, particularly those that the researcher would have pre-existing cause to suspect differentiated effects (for example, gender, income status, newcomers / Canadian-born citizens, and so on).

9 See supplementary material SM1.

10 See supplementary material SM2.

11 Consider one example. In 2009, the Mental Health Commission of Canada launched the At Home/Chez Soi initiative. The project involved a large RCT to help Canadians experiencing homelessness or precarious housing. Individuals were randomly assigned to either the status quo or a Housing First treatment that made it easier to access housing supports (Mental Health Commission of Canada, 2014). This treatment, which was shown to be more effective than the status quo condition, includes several behaviourally informed elements. For example, the intervention includes a significant reduction in administrative burden, which presumably reduces cognitive load. The intervention is also “choice-preserving” in that it emphasizes an individual's freedom to accept or decline the intervention. Similarly, the use of an RCT echoes the strong emphasis on testing that is common to BI. However, we do not include this initiative in our dataset, as it does not describe itself as a BI project, it does not explicitly incorporate insights from psychology (for example, heuristic-based decision making), and the nature of the intervention is sufficiently different from most definitions of a light-touch nudge (Thaler and Sunstein, Reference Thaler and Sunstein2008).

References

Atkinson, Michael M. 2013. “Policy, Politics and Political Science: Presidential Address to the Canadian Political Science Association, Victoria, June 5, 2013.” Canadian Journal of Political Science 46 (4): 751–72.10.1017/S000842391300084XCrossRefGoogle Scholar
Ball, Sarah and Head, Brian W.. 2021. “Behavioural Insights Teams in Practice: Nudge Missions and Methods on Trial.” Policy and Politics 49 (1): 105–20.10.1332/030557320X15840777045205CrossRefGoogle Scholar
Banting, Keith G. 1997. “The Social Policy Divide: The Welfare State in Canada and the United States.” In Degrees of Freedom: Canada and the United States in a Changing World, ed. Banting, Keith G., Hoberg, George and Simeon, Richard. Montreal and Kingston: McGill-Queen's University Press.CrossRefGoogle Scholar
Banting, Keith G. 2010. “Is There a Progressive's Dilemma in Canada? Immigration, Multiculturalism and the Welfare State: Presidential Address to the Canadian Political Science Association, Montreal, June 2, 2010.” Canadian Journal of Political Science 43 (4): 797820.CrossRefGoogle Scholar
Bastarache, Michel. 2020. Broken Lives, Broken Dreams: The Devastating Effects of Sexual Harassment On Women in the RCMP Final Report on the Implementation of the Merlo Davidson Settlement Agreement. Ottawa: Royal Canadian Mounted Police.Google Scholar
Beeby, Dean. 2018. “Nudge Resistant: Canadians Seem to Ignore Taxman's Environmental Message.” CBC News, July 22. https://www.cbc.ca/news/politics/nudge-online-cra-experiment-tax-filing-environment-trees-letters-1.4751769.Google Scholar
Bellé, Nicola, Cantarelli, Paola and Belardinelli, Paolo. 2018. “Prospect Theory Goes Public: Experimental Evidence on Cognitive Biases in Public Policy and Management Decisions.” Public Administration Review 78 (6): 828–40.CrossRefGoogle Scholar
Benartzi, Shlomo, Beshears, John, Milkman, Kay L., Sunstein, Cass R., Thaler, Richard H., Shankar, Maya, Tucker-Ray, Will, Congdon, William J. and Galing, Steven. 2017. “Should Governments Invest More in Nudging?” Psychological Science 28 (8): 1041–55.CrossRefGoogle ScholarPubMed
Brock, Kathy L. 2021. “Policy Labs, Partners and Policy Effectiveness in Canada.” Policy Design and Practice 4 (2): 228–41.10.1080/25741292.2021.1880063CrossRefGoogle Scholar
Brodkin, Evelyn Z. 2012. “Reflections on Street-Level Bureaucracy: Past, Present, and Future.” Public Administration Review 72 (6): 940–49.10.1111/j.1540-6210.2012.02657.xCrossRefGoogle Scholar
Bryan, Christopher J., Tipton, Elizabeth and Yeager, David S.. 2021. “Behavioural Science is Unlikely to Change the World without a Heterogeneity Revolution.Nature Human Behaviour 5 (8): 980–9.10.1038/s41562-021-01143-3CrossRefGoogle ScholarPubMed
Burden, Barry C., Canon, David T., Mayer, Kenneth R. and Moynihan, Donald P.. 2012. “The Effect of Administrative Burden on Bureaucratic Perception of Policies: Evidence from Election Administration.” Public Administration Review 72 (5): 741–51.CrossRefGoogle Scholar
Carter, Evan C., Kofler, Lilly M., Forster, Daniel E. and McCullough, Michael E.. 2015. “A Series of Meta-analytic Tests of the Depletion Effect: Self-Control Does Not Seem to Rely on a Limited Resource.” Journal of Experimental Psychology: General 144: 796815.10.1037/xge0000083CrossRefGoogle ScholarPubMed
Caruana, Nicholas J., McGregor, R. Michael and Stephenson, Laura B.. 2015. “The Power of the Dark Side: Negative Partisanship and Political Behaviour in Canada.” Canadian Journal of Political Science 48 (4): 771–89.10.1017/S0008423914000882CrossRefGoogle Scholar
Castelo, Noah, Hardy, Elizabeth, House, Julian, Mazar, Nina, Tsai, Claire and Zhao, Min. 2015. “Moving Citizens Online: Using Salience and Message Framing to Motivate Behavior Change.” Behavioral Science and Policy 1 (2): 5768.10.1353/bsp.2015.0018CrossRefGoogle Scholar
Cesario, Joseph. 2014. “Priming, Replication, and the Hardest Science.” Perspectives on Psychological Science 9 (1): 4048.CrossRefGoogle ScholarPubMed
Chen, Katie, Bendle, Neil and Soman, Dilip. 2017. Policy by Design: The Dawn of Behaviourally-Informed Government. Toronto: Rotman School of Management.Google Scholar
Congdon, William and Shankar, Maya. 2015. “The White House Social & Behavioral Sciences Team: Lessons Learned from Year One.” Behavioral Science & Policy 1 (2): 7786.10.1353/bsp.2015.0010CrossRefGoogle Scholar
Dang, Junhua, Barker, Paul, Baumert, Anna, Bentvelzen, Margriet, Berkman, Elliot, Buchholz, Nita, et al. 2021. “A Multilab Replication of the Ego Depletion Effect.” Social Psychological and Personality Science 12 (1): 1424.CrossRefGoogle ScholarPubMed
Delaney, Liam. 2018. “Behavioural Insights Team: Ethical, Professional and Historical Considerations.” Behavioural Public Policy 2 (2): 183–89.10.1017/bpp.2018.19CrossRefGoogle Scholar
DellaVigna, Stefano and Linos, Elizabeth. 2022. “RCTs to Scale: Comprehensive Evidence from Two Nudge Units.” Econometrica 90 (1): 81116.10.3982/ECTA18709CrossRefGoogle Scholar
Doberstein, Carey. 2017a. “An Emerging Experimental Approach to Public Administration Research in Canada.” Canadian Public Administration 60 (1): 135.10.1111/capa.12208CrossRefGoogle Scholar
Doberstein, Carey. 2017b. “Whom Do Bureaucrats Believe? A Randomized Controlled Experiment Testing Perceptions of Credibility of Policy Research: Perceptions of Credibility of Policy Research.” Policy Studies Journal 45 (2): 384405.CrossRefGoogle Scholar
Doberstein, Carey and Charbonneau, Étienne. 2020. “Experimenting with Public Sector Innovation: Revisiting Gow for the Digital Era.” Canadian Public Administration 63 (1): 733.10.1111/capa.12353CrossRefGoogle Scholar
Elmore, Richard F. 1979. “Backward Mapping: Implementation Research and Policy Decisions.” Political Science Quarterly 94 (4): 601–16.10.2307/2149628CrossRefGoogle Scholar
Esarey, Justin and Wu, Ahra. 2016. “Measuring the Effects of Publication Bias in Political Science.” Research and Politics 3 (3): 2053168016665856.CrossRefGoogle Scholar
Evans, Bryan and Cheng, Siu Mee. 2021. “Canadian Government Policy Innovation Labs: An Experimental Turn in Policy Work?” Canadian Public Administration 64 (4): 587610.CrossRefGoogle Scholar
Ewert, Benjamin. 2020. “Moving beyond the Obsession with Nudging Individual Behaviour: Towards a Broader Understanding of Behavioural Public Policy.” Public Policy and Administration 35 (3): 337–60.10.1177/0952076719889090CrossRefGoogle Scholar
Citizen First. 2021. “A Highlight of the Citizens First 2020 Study.” Institute for Citizen-Centred Service. https://citizenfirst.ca/assets/uploads/research-repository/JC-Report-August-2021-EN.pdf.Google Scholar
Franco, Annie, Malhotra, Neil and Simonovits, Gabor. 2014. “Publication Bias in the Social Sciences: Unlocking the File Drawer.” Science 345 (6203): 1502–5.10.1126/science.1255484CrossRefGoogle ScholarPubMed
French, Richard D. 2018. “Lessons from the Evidence on Evidence-Based Policy.” Canadian Public Administration 61 (3): 425–42.CrossRefGoogle Scholar
Government of Ontario. 2018. Behavioural Insights in Ontario: Update Report 2018. Treasury Board Secretariat. https://www.ontario.ca/page/behavioural-insights-ontario-update-report-2018.Google Scholar
Hagman, William, Andersson, David, Västfjäll, Daniel and Tinghög, Gustav. 2015. “Public Views on Policies Involving Nudges.” Review of Philosophy and Psychology 6 (3): 439–53.10.1007/s13164-015-0263-2CrossRefGoogle Scholar
Hallsworth, Michael and Kirkman, Elspeth. 2020. Behavioral Insights. Cambridge: MIT Press.10.7551/mitpress/12806.001.0001CrossRefGoogle Scholar
Halpern, David. 2015. Inside the Nudge Unit: How Small Changes Can Make a Big Difference. London: Random House.Google Scholar
Hardy, Elizabeth, Khan, Haris and Paleja, Meera. 2021. “Behavioral Insights – a Government of Canada Perspective.” In Psychology and Behavioral Economics: Applications for Public Policy, ed. Ruggeri, Kai. New York: Routledge.Google Scholar
Herd, Pamela and Moynihan, Donald P.. 2019. Administrative Burden: Policymaking by Other Means. New York: Russell Sage Foundation.10.7758/9781610448789CrossRefGoogle Scholar
Howlett, Michael. 2013. Canadian Public Policy: Selected Studies in Process and Style. Toronto: University of Toronto Press.Google Scholar
Howlett, Michael, Ramesh, M. and Perl, Anthony. 2020. Studying Public Policy: Principles and Processes. Oxford: Oxford University Press.Google Scholar
Hummel, Dennis and Maedche, Alexander. 2019. “How Effective Is Nudging? A Quantitative Review on the Effect Sizes and Limits of Empirical Nudging Studies.” Journal of Behavioral and Experimental Economics 80: 4758.10.1016/j.socec.2019.03.005CrossRefGoogle Scholar
Impact and Innovation Unit. 2021. “About.” https://impact.canada.ca/index.php/en/about.Google Scholar
Jacobs, Alan M. and Matthews, J. Scott. 2012. “Why Do Citizens Discount the Future? Public Opinion and the Timing of Policy Consequences.” British Journal of Political Science 42 (4): 903–35.10.1017/S0007123412000117CrossRefGoogle Scholar
Jann, Werner and Wegrich, Kai. 2006. “Theories of the Policy Cycle.” In Handbook of Public Policy Analysis: Theory, Politics, and Methods, ed. Fischer, Frank, Miller, Gerald J. and Sidney, Mara S.. Boca Raton, Florida: CRC Press.Google Scholar
John, Peter. 2014. “Policy Entrepreneurship in UK Central Government: The Behavioural Insights Team and the Use of Randomized Controlled Trials.” Public Policy and Administration 29 (3): 257–67.CrossRefGoogle ScholarPubMed
Jones, Stephen, Head, Brian and Ferguson, Michele. 2021. “In Search of Policy Innovation: Behavioural Insights Teams in Australia and New Zealand.” Australian Journal of Public Administration 80 (3): 435–52.CrossRefGoogle Scholar
Kahneman, Daniel. 2013. Thinking, Fast and Slow. New York: Farrar, Straus and Giroux.Google Scholar
Kahneman, Daniel and Tversky, Amos. 1979. “Prospect Theory: An Analysis of Decision Under Risk.” Econometrica 47: 263–91.CrossRefGoogle Scholar
Lades, Leonhard K. and Delaney, Liam. 2022. “Nudge FORGOOD.” Behavioural Public Policy 6 (1): 7594.10.1017/bpp.2019.53CrossRefGoogle Scholar
Lepenies, Robert and Małecka, Magdalena. 2019. “The Ethics of Behavioural Public Policy.” In Routledge Handbook of Ethics and Public Policy, ed. Poama, A. and Lever, A.. New York: Routledge.Google Scholar
Lindquist, Evert A. and Buttazzoni, Michael. 2021. “The Ecology of Open Innovation Units: Adhocracy and Competing Values in Public Service Systems.” Policy Design and Practice 4 (2): 212–27.CrossRefGoogle Scholar
Lipsky, Michael. 2010. Street-Level Bureaucracy: Dilemmas of the Individual in Public Services. 30th anniversary expanded ed. Russell Sage Foundation.Google Scholar
Loewenstein, George and Chater, Nick. 2017. “Putting Nudges in Perspective.” Behavioural Public Policy 1 (1): 2653.10.1017/bpp.2016.7CrossRefGoogle Scholar
Longo, Justin. 2017. “The Evolution of Citizen and Stakeholder Engagement in Canada, from Spicer to #Hashtags.” Canadian Public Administration 60 (4): 517–37.10.1111/capa.12229CrossRefGoogle Scholar
Ly, Kim and Soman, Diliip. 2013. Nudging around the World. Rotman School of Management.Google Scholar
MacDonald, Fiona and Paterson, Stephanie. 2021. “In the Military and Beyond, More Women Doesn't Mean More Equity.” The Conversation, April 18. https://theconversation.com/in-the-military-and-beyond-more-women-doesnt-mean-more-equity-157287.Google Scholar
McDavid, Jim and Henderson, Kathryn. 2021. “Hard Won Lessons Learned in the Evaluation Field: Implications for Behavioural Insights Practice.” Canadian Public Administration 64 (1): 725.10.1111/capa.12401CrossRefGoogle Scholar
McFadden, Daniel. 2006. “Free Markets and Fettered Consumers.” American Economic Review 96 (1): 529.10.1257/000282806776157542CrossRefGoogle Scholar
Mental Health Commission of Canada. 2014. “National Final Report: Cross-Site At Home/Chez Soi Project.” Ottawa.Google Scholar
Moynihan, Donald P., Herd, Pamela and Harvey, Hope. 2015. “Administrative Burden: Learning, Psychological, and Compliance Costs in Citizen-State Interactions.” Journal of Public Administration Research and Theory 25 (1): 4369.10.1093/jopart/muu009CrossRefGoogle Scholar
Nesterak, Evan. 2021. “Five Takeaways from Our Conversation with Richard Thaler about the Past, Present, and Future of Nudge.” The Behavioral Scientist. September 21. https://behavioralscientist.org/five-takeaways-from-our-conversation-with-richard-thaler-about-the-past-present-and-future-of-nudge/.Google Scholar
Oliver, Adam. 2017. The Origins of Behavioural Public Policy. Cambridge: Cambridge University Press.CrossRefGoogle Scholar
Organisation for Economic Co-operation and Development. 2017. Behavioural Insights and Public Policy: Lessons from Around the World. https://www.oecd.org/gov/regulatory-policy/behavioural-insights.htm.Google Scholar
Pal, Leslie, Auld, Graeme and Mallett, Alexandra. 2021. Beyond Policy Analysis: Public Issue Management in Turbulent Times. Toronto: Thomson Nelson.Google Scholar
Pickup, Mark, Stecula, Dominik and Linden, Clifton van der. 2020. “Novel Coronavirus, Old Partisanship: COVID-19 Attitudes and Behaviours in the United States and Canada.” Canadian Journal of Political Science 53 (2): 357–64.10.1017/S0008423920000463CrossRefGoogle Scholar
Porterfield, Deborah S., Hinnant, Laurie W., Kane, Heather, Horne, Joseph, McAleer, Kelly and Roussel, Amy. 2012. “Linkages between Clinical Practices and Community Organizations for Prevention.” American Journal of Preventive Medicine 42 (6): S163–71.10.1016/j.amepre.2012.03.018CrossRefGoogle ScholarPubMed
Privy Council Office. 2017. Increasing Recruitment of Women into the Canadian Armed Forces. Ottawa: Government of Canada.Google Scholar
Robitaille, Nicole, House, Julian and Mazar, Nina. 2021. “Effectiveness of Planning Prompts on Organizations’ Likelihood to File Their Overdue Taxes: A Multi-Wave Field Experiment.” Management Science 67 (7): 4327–40.10.1287/mnsc.2020.3744CrossRefGoogle Scholar
Sabatier, Paul A. 1986. “Top-Down and Bottom-Up Approaches to Implementation Research: A Critical Analysis and Suggested Synthesis.” Journal of Public Policy 6 (1): 2148.10.1017/S0143814X00003846CrossRefGoogle Scholar
Sheffer, Lior and J. Loewen, Peter. 2019. “Electoral Confidence, Overconfidence, and Risky Behavior: Evidence from a Study with Elected Politicians.” Political Behavior 41 (1): 3151.10.1007/s11109-017-9438-0CrossRefGoogle Scholar
Sheffer, Lior, Loewen, Peter J., Soroka, Stuart, Walgrave, Stefaan and Sheafer, Tamir. 2018. “Nonrepresentative Representatives: An Experimental Study of the Decision Making of Elected Politicians.” American Political Science Review 112 (2): 302–21.CrossRefGoogle Scholar
Singh, Shane P., Roy, Jason and Fournier, Patrick. 2016. “Polls, Partisanship, and Voter Decision-Making: An Experimental Analysis.” In Voting Experiments, ed. Blais, A., Laslier, J.-F. and Van der Straeten, K.. Cham, Switzerland: Springer International.Google Scholar
Stephenson, Laura B., Harell, Allison, Rubenson, Daniel and Loewen, Peter J.. 2020. 2019 Canadian Election Study, Harvard Dataverse, V1. https://doi.org/10.7910/DVN/DUS88V.CrossRefGoogle Scholar
Thaler, Richard H. and Sunstein, Cass R.. 2008. Nudge: Improving Decisions about Health, Wealth, and Happiness. New Haven: Yale University Press.Google Scholar
Thomann, Eva, van Engen, Nadine and Tummers, Lars. 2018. “The Necessity of Discretion: A Behavioral Evaluation of Bottom-Up Implementation Theory.” Journal of Public Administration Research and Theory 28 (4): 583601.CrossRefGoogle Scholar
Tversky, Amos and Kahneman, Daniel. 1974. “Judgment under Uncertainty: Heuristics and Biases: Biases in Judgments Reveal Some Heuristics of Thinking under Uncertainty. Science 185 (4157): 1124–31.10.1126/science.185.4157.1124CrossRefGoogle Scholar
Walgrave, Stefaan, Sevenans, Julie, Van Camp, Kirsten and Loewen, Peter J.. 2018. “What Draws Politicians’ Attention? An Experimental Study of Issue Framing and Its Effect on Individual Political Elites.” Political Behavior 40 (3): 547–69.CrossRefGoogle Scholar
Wheatley, Amanda and Hervieux, Sandy. 2020. “Artificial Intelligence in Academic Libraries: An Environmental Scan.” Information Services & Use 39 (4): 347–56.10.3233/ISU-190065CrossRefGoogle Scholar
Figure 0

Figure 1 Number of Projects by Policy Area.

Figure 1

Figure 2 Growth of BI staff in the Government of Canada, 2016–2022.Note: Data collected from the Government Electronic Directory Services.

Supplementary material: File

Hopkins and Lawlor supplementary material

Hopkins and Lawlor supplementary material 1

Download Hopkins and Lawlor supplementary material(File)
File 21.3 KB
Supplementary material: File

Hopkins and Lawlor supplementary material

Hopkins and Lawlor supplementary material 2

Download Hopkins and Lawlor supplementary material(File)
File 22.3 KB