Hostname: page-component-cd9895bd7-dzt6s Total loading time: 0 Render date: 2024-12-26T04:59:01.218Z Has data issue: false hasContentIssue false

Discretion in the Automated Administrative State

Published online by Cambridge University Press:  11 January 2023

Sancho McCann*
Affiliation:
Peter A. Allard School of Law, University of British Columbia, Vancouver, Canada

Abstract

Automated decision-making takes up an increasingly significant place in the administrative state. This article presents a conception of discretion that is helpful for evaluating the proper place of algorithms in public decision-making. I argue that the algorithm itself is not a site of discretion. The threat is that automated decision-making alters the relationships between traditional actors in a way that can cut down discretion and human commitment. Algorithmic decision-makers can serve to fetter the discretion that the legislature and the populace expect to be exercised. We must strive to maintain discretion, moral agency, deliberative ideals, and human commitment through the system that surrounds the use of an algorithm and to develop a new expertise that can retain and exercise the expected discretion. Backing this argument are traditional legal constraints, public expectations, and administrative law principles, tied together through the organizing principle of discretion.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press on behalf of University of Western Ontario (Faculty of Law)

Introduction

In this project, I seek to present discretion as an organizing principle in public law that will be useful for evaluating the legal and political legitimacy of algorithmic decision-making.Footnote 1 An algorithm may be used in public decision-making in two distinct ways: as a tool of a decision-maker or as a new decision-maker itself. The duties and expectations that we place on these two modes of use are contrasted through this organizing principle of discretion.

I will develop a conception of discretion that I propose is helpful for understanding the role of algorithms across a spectrum of delegation. I am not interested merely in what choices have been left in the hands of administrative decision-makers, but rather, what does it mean for those decision-makers to make those decisions with discretion? Especially interesting is the seemingly unconstrained margin: when the legislature has conferred broad discretion on an administrative decision-maker, what does law require in that setting? What is expected from the decision-maker to fulfill their purpose(s) and to maintain legitimacy in a political system? Answers to these questions will inform where it is appropriate to deploy algorithms in public law decision-making and the system design that is necessary to ensure discretion is being properly exercised.

I argue that sub-delegation to an algorithm is appropriate only in limited circumstances. When legislated delegation contains minimal discretion—where we can break down an algorithm into discrete, expressible rules or operations reflecting the direction of the legislature—the use of an algorithm is unproblematic. However, even when used as a tool, algorithms get between actors that have traditionally been in contact; this presents the real risk of displacing human commitment and moral agency behind the decisions that are ultimately made. My novel claim is that the algorithm itself is not a site of discretion, so when discretion is called for, it must be maintained in the social structures and organizational design surrounding the use of the algorithm.

I begin with an abridged history of the administrative state in Canada and the ontology that has traditionally existed. I then present a contrast between that traditional ontology and the new ontology of an automated administrative state. Finally, I present a collection of expectations and principles that I have extracted from case law and secondary sources, related together by the organizing principle of discretion.

The Old Ontology

An ontology of the administrative state is crucial to understanding algorithms as “analytically distinct” forces and “tools in the governance process.”Footnote 2 In order to understand the place of algorithms in today’s ontology and the changes they bring, it is helpful to review the traditional ontology and the motivations for its structure.Footnote 3 In this review of the old ontology, I recognize that algorithms have been used throughout history to give effect to legislative intent. For example, the Income War Tax Act of 1917 necessarily included a formula that prescribed the amount of income tax to be assessed.Footnote 4 In the most general sense, an algorithm is simply an “encoded procedure[] for transforming input data into a desired output, based upon specified calculations”;Footnote 5 these legislated taxation formulae fit that mold. However, much delegation has also included discretion. For example, in that same Act, discretion is involved in determining values for the inputs to the formulae, “reasonable allowance … by the Minister for depreciation,” allowing the Minister to exclude income on a discretionary basis, etc.Footnote 6 This discretion permeates the administrative state.

The Canadian administrative state finds its roots in the Privy Council and in the constitutional conventions of responsible government and Parliamentary sovereignty.Footnote 7 Under these traditions, the Privy Council (in practice, Cabinet)—on behalf of the Crown—exercises executive-branch decision-making as delegated by Parliament.Footnote 8 In the early 1900s, Parliament recognized the need for administrative bodies beyond the literal Cabinet. Parliament created boards of dedicated experts that would handle issues relating to, for instance, railways, grain supply, and international water boundaries.Footnote 9 Cabinet had neither the capacity nor the expertise to handle these questions and administration efficiently. Throughout the twentieth century, the administrative state has grown to touch nearly every aspect of our lives.

As mentioned, some of this delegated or sub-delegated power is imperative and non-discretionary. Examples of this kind of delegation are found in the front-line administration of large-scale logistics,Footnote 10 registration,Footnote 11 or licensing systems.Footnote 12 Vancouver’s current dog-licencing scheme is one such example.Footnote 13 Once a dog owner provides the required information and payment, a licence will be issued. With delegation like this—with minimal discretion and where we can break down an algorithm into discrete, expressible rules or operations—the use of an algorithm is unproblematic. These rule-like statutory duties “must be performed without unreasonable delay, and this may be enforced by mandamus.”Footnote 14

Of course, in much delegation, the only non-discretionary aspect is the requirement that a decision be made; aside from that element, delegation often allows and expects discretion. Vancouver, for example, has muzzle and security requirements for aggressive dogs: dogs with “a known propensity … to attack without provocation.”Footnote 15 This determination cannot be made without discretion and expertise. Courts supervise this discretion through judicial review. The approach taken by courts in supervising this discretion has evolved over the years. Through judicial review, courts both affirm and shape the ontology of the administrative state. In judicial review, the government is forced to self-identify its motivations for delegation and its requests for deference, and courts announce factual and legal realities regarding this ontology. By following this evolution, we can extract a traditional ontology of the administrative state—one centered on expertise.

Early on, courts were hyper-focused on ensuring the administrative actors had jurisdiction to do what they were doing—that the actors had interpreted their powers and duties correctly and were acting within that sphere of responsibility.Footnote 16 That view eventually gave way to one that recognized that “law making and legal interpretation are shared enterprises in the administrative state”Footnote 17 and that recognized the relative expertise inherent in the tribunals as institutions. One of the most oft-cited justifications for the delegated discretion is “the need for greater specialization and technical or subject-matter expertise.”Footnote 18 The Supreme Court has previously recognized the relative expertise of such decision-makers as a reason to defer to them, even on some questions of law.Footnote 19

Another early feature of Canadian administrative law was a specialized approach to reviewing discretionary decisions.Footnote 20 Under that approach, discretionary decisions were only reviewable on grounds of bad faith, improper purpose, or the use of irrelevant considerations.Footnote 21 But even the reasoning of Justice Rand, in Roncarelli v. Duplessis, Footnote 22 spoke about review of discretionary decisions more broadly: “there is always a perspective within which a statute is intended to operate; and any clear departure from its lines or objects is just as objectionable as fraud or corruption.”Footnote 23 And the Supreme Court in Baker explicitly folded the review of discretionary decision-making into the general framework of substantive review.Footnote 24

Underlying judicial review in Canadian administrative law is a tension between parliamentary sovereignty and the rule of law.Footnote 25 Courts must toe a fine line to both uphold the rule of law and avoid “undue interference … in respect of the matters delegated to administrative bodies by Parliament and legislatures.”Footnote 26 On the other hand, judicial review can also be seen as furthering parliamentary sovereignty when courts ensure that a decision-maker’s jurisdiction is “narrowly circumscribed and defined according to the intent of the legislature.”Footnote 27

In resolving these tensions, courts have assumed a particular ontology of the administrative state. Differing opinions within the Supreme Court of Canada’s membership can be explained through differing ontologies. These differences pre-Vavilov are best displayed through Alberta (Information and Privacy Commissioner) v. Alberta Teachers Association Footnote 28 and Edmonton (City) v. Edmonton East (Capilano) Shopping Centres Ltd.Footnote 29

Alberta Teachers displayed a full spectrum of perspectives regarding the role of expertise. Justice Rothstein, writing for the majority, held that when a decision-maker is interpreting or applying their “home statute,” the presumption is that “the Court should afford a measure of deference under the standard of reasonableness.”Footnote 30 Under this view, decision makers benefit from assumed expertise when working within their home statute. Justice Cromwell, writing for himself in concurrence, presented a view of judicial review centered on determining legislative intent through a multi-factorial analysis.Footnote 31 He wrote, “The touchstone of judicial review is legislative intent.”Footnote 32 This is a view that is skeptical of whether a decision-maker possesses the expertise that the majority assumes, or alternatively, skeptical of whether the legislature intended any such expertise to be granted deference. Justice Binnie, writing for himself and Justice Deschamps, sets out a compromise position. For Binnie J., deference should be presumed for a decision-maker interpreting or applying their home statute only when it is “within [their] expertise” and when the issue is not of “general legal importance.”Footnote 33 This position keeps expertise central to justifying deference, without assuming it exists simply because a decision-maker is working within their home statute.

By the time of Edmonton East, in 2016, the entire membership of the Court seemed to accept that expertise played a central role in determining whether deference to the administrative decision-maker is warranted. The majority wrote that “the presumption of reasonableness is grounded in the legislature’s choice to give a specialized tribunal responsibility for administering the statutory provisions, and the expertise of the tribunal in so doing.”Footnote 34 Under this view, “expertise is something that inheres in a tribunal itself as an institution.”Footnote 35 In this one paragraph, the majority both elevates and eliminates the importance of expertise. It becomes the core of the justification of deference to administrative tribunals but also becomes meaningless in practice given that it inheres in the tribunal itself. It is shorthand for the entire justification of deference. The dissent also focused on expertise but left open the possibility that certain tribunals lack expertise in certain areas before them. Their view was that the Board had a “lack of relative expertise in interpreting the law”Footnote 36 and that the “legislature … has a role to play in designating and delimiting the presumed expertise of an administrative tribunal.”Footnote 37 Even the dissent’s view seemingly premises deference on expertise, albeit expertise that needs to be demonstrated.

So, until very recently, Canada has presented, and the courts have recognized, an ontology of the administrative state centered on the legislature’s considered decision to delegate matters to institutions or people with relative expertise compared to the legislators and the courts. Where the legislature has left uncertainty to be filled in with meaning at the front lines or through expert bureaucracy, the court has been willing to defer. Courts are not the only site of moral agency that might “vindicate private autonomy.”Footnote 38 Moving into the present, I show the changing position of expertise in the court’s view of the ontology and the invisibility of an old actor with new roles: the algorithm.

The New Ontology

Recently, in Canada (Minister of Citizenship and Immigration) v. Vavilov,Footnote 39 the Supreme Court set out a new vision of the administrative state that relegates expertise to consideration only within reasonableness review: expertise is no longer a factor in selecting the standard of review. While expertise remains part of the justification for reasonableness review as the default, expertise no longer plays any analytical role when selecting the standard of review. The majority in Vavilov viewed an administrative body’s inherent expertise as an unnecessary proxy for legislative intent. In eschewing a contextual analysis for selection of the standard of review, perhaps it was inevitable that expertise would be sidelined in this analysis.

But the majority went further. One of the circumstances that will displace the default standard of reasonableness is when the decision-maker has decided a “general question[] of law of central importance to the legal system as a whole.”Footnote 40 Previously, this ground for correctness review had required more: that the question also be “outside the adjudicator’s specialized area of expertise.”Footnote 41 The majority says that because consideration of expertise is “folded into the new starting point … namely, the presumption of reasonableness review,”Footnote 42 it is no longer relevant on this path to correctness review. The logical implication is that there are now questions that formerly would not have attracted correctness review (because they were not outside the adjudicator’s specialized area of expertise) that will now attract correctness review, even though the new starting point includes the assumption that the decision-maker has inherent expertise.

The majority in Vavilov did not merely sidestep the use of expertise as an unnecessary proxy; it also eliminated the possibility that relative expertise (either inherent or demonstrated) could prevent correctness review of questions of law that are of central importance to the legal system as a whole.

This vision set out by Vavilov reflects the Court’s current understanding of the ontology of the administrative state. The actors in this ontology are still the legislature, delegated decision-makers (either located in or associated with the executive branch), subjects of the decisions, and the courts. One primary relationship in this ontology is a reciprocal exchange of deference for justification. The delegated decision-maker obtains legitimacy and deference only when it justifies its decision to the subject of the decision (but reviewed by the courts). The practical expression of expertise in this new ontology is found in the “form and content” of the reasons.Footnote 43

Some long-standing aspects of Canada’s administrative ontology are somewhat elided in the decision in Vavilov. Vavilov does recognize a spectrum of decision-makers and functions previously acknowledged by the Court.Footnote 44 For example, Cabinet’s Orders in Council are to be struck down only on “an egregious case” of “jurisdictional or other compelling grounds.”Footnote 45 Some actions of the Governor in Council are even considered “legislative action in its purest form.”Footnote 46 Regulations created by government agencies benefit from a “presumption of validity,” only to be set aside if “irrelevant, extraneous, or completely unrelated to the statutory purpose.”Footnote 47 The decision of a municipal council to pass a by-law will stand unless it is one that “no reasonable body” would have passed, informed by the wide variety of factors including broad “social, economic, and political factors.”Footnote 48 But this variation is mentioned only in passing in Vavilov with the apparent goal that it be captured within a single standard of reasonableness review.

A clearer blind spot is the novel place of algorithms in the automated administrative state. The use of algorithms has a causal effect that shapes this new ontology in a way that is potentially problematic, and this has gone unrecognized by the courts. A complete description of the modern ontology cannot ignore how algorithms change relationships between existing actors. We must not only account for the algorithms themselves, but a collection of other entities they bring along with them: algorithm designers, procurers, and the algorithmic output.Footnote 49 Algorithms get between human actors, displace human commitment, and are jurispathic in that they stymie sensitivity and creativity.

First, algorithms alter the relationship between front-line bureaucrats, the ultimate decision-maker (even when it is that same front-line bureaucrat), and the people subject to the decision. Algorithms get between actors that would have traditionally been in contact.Footnote 50 Instead of a front-line worker reading someone’s application, an algorithm might pre-process the application, compressing and transforming that information. Those transformations encode policy choices that are often only implicitly made.Footnote 51

Algorithms also displace human commitment.Footnote 52 The discretion granted to delegated decision-makers is an unavoidable and often intentional choice by the legislature to defer committing to meaning. This meaning is instead intended to be built up at the front lines: through caseworkers in social-assistance programsFootnote 53 or immigration officers,Footnote 54 for example. The resultant meaning—this “legal world”—is created “only to the extent that there are commitments that place bodies on the line.”Footnote 55 For these decisions and interpretations to create thick meaning, they must be backed by human commitment to the effects they will have on people. Algorithms avoid developing such commitment on the part of the state. The procurers and designers, who are crystalizing interpretation and policy,Footnote 56 don’t see the front-line effects. The front-line workers may abdicate responsibility to the algorithmic decisions, especially if in an environment that does not foster a sense of moral agency.Footnote 57 Decisions will be made but without intentional commitment.

The resulting algorithmic decisions likewise do not develop tradition. They cannot, in and of themselves, be a site of jurisgenesis.Footnote 58 Jurisgenesis (in the sense developed by Robert M. Cover) has a “social basis”Footnote 59 that is lacking at the site of algorithmic decisions. It requires meaning to be “essentially contested,”Footnote 60 even if it may be ultimately articulated by an authoritative institution. This front-line flexibility is an essential quality of delegated discretion.Footnote 61 Bernardo Zacka presents this ideal: “Street-level bureaucrats must … resist the pull toward moral dispositions that are overly narrow … they must strive, as a group, to retain a range of dispositions that are sufficiently diverse.”Footnote 62 I am not arguing that there can be no jurisgenesis in a broader system that includes algorithms as a mere component, but where such jurisgenesis exists, it is found in the other relationships surrounding the algorithm. An algorithm itself contains pre-programmed meaning, killing off all alternatives. The use of an algorithm is thus jurispathic. Once deployed, they constrain creativity, variation, sensitivity, and justifiable divergence from past decisions.

What is discretion?

There are spaces in which the legislature can well specify the decision to be made and the manner in which it should be made: for example, British Columbia’s moose lottery.Footnote 63 In such circumstances, where the legislature itself has sufficiently constrained meaning and flexibility, we may be losing nothing by using an algorithm to execute that legislative will. These would be sites where the legislature itself has been jurispathic, leaving no need or jurisdiction for creativity at the front lines. Recognizing this, another question arises: how should we decide which decisions can unproblematically be made by an algorithm? To answer that question, we must first understand the nature of discretion: what does it mean to exercise discretion and what are our expectations on that exercise? These questions motivate the remainder of this piece.

The conception I develop here considers discretion as part of law, contrary to the aphorism: “[w]here law ends, discretion begins.”Footnote 64 While this conception is largely complementary to Kenneth Culp Davis’s project, which was focused on confining discretion, there are portions of his text that clearly recognize the importance of preserving discretion.Footnote 65 As a whole, Davis was interested in finding the sweet-spot on the control-discretion spectrum.Footnote 66 He also emphasized the important role for discretion in the administrative state—that “creativity is impossible without discretion” and that “when discretion shrinks too much, affirmative action is needed to recreate it.”Footnote 67 His 1969 text was written at a time when computing was in its infancy, and Davis only mentions computers once.Footnote 68

Decision-making has always been delegated along a spectrum of constraints.Footnote 69 On one end of the spectrum, there is in effect no decision-making at all: a rule has been successfully specified. To give an example, front-line auto insurance providers in British Columbia have no discretion regarding the basic insurance tariff as it is dictated by the Insurance Corporation of British Columbia (ICBC).Footnote 70 Other examples include the taxation formulae and the moose lottery mentioned above. At the other end of the spectrum, the legislature seems to have delegated near-unconstrained discretion. For example, under Newfoundland and Labrador’s Fish Inspection Act, the minister “may refuse to issue a licence required under this Act or the regulations without assigning a reason for the refusal.”Footnote 71 Much delegated decision-making falls somewhere in the middle. Consider, for example, the many decisions made by immigration officials when they determine a person’s eligibility for various types of immigration into CanadaFootnote 72 or ICBC’s determination of the formula for the basic insurance tariff.Footnote 73 This section aims to describe what is happening when exercising that discretion.

Discretion is not mere choice. It has meaning as a legal term of art. First, there is a sense of the word which connotes jurisdiction, capacity, or power.Footnote 74 Courts often even use the terms discretion and power interchangeably or as part of the doublet, ‘discretion or power.’Footnote 75 In this sense, discretion can range from strong discretion to weak discretion. But in public law, “[t]here is no such thing as absolute and untrammelled ‘discretion.’”Footnote 76 Discretion as a public-law power is always coupled with duties.Footnote 77 These duties arise because the people subject to the decision are not mere objects with liabilities to the decision-maker—they also have rights. These rights correlate with the duties of the delegated decision-maker.Footnote 78

The duty of discretion is not to deliver a particular outcome, but to exercise the discretion in a particular manner. While no outcome is guaranteed, discretion is. It is this second sense, the duty of discretion, that I wish to develop further. Here, discretion also becomes the “name of an intellectual virtue.”Footnote 79 This discretion does not require legally-trained interpretation of statutes, but it does require informal, personal judgment conducted in good faith.Footnote 80

I do not adopt Dworkin’s conception of strong discretion, which would mean the decision-maker has “the right to make any decision he wishes” and which would “deny any other participant the right to claim a particular decision.”Footnote 81 Discretion in our administrative state rarely possesses this quality. Dworkin himself argues that discretion in that sense is undesirable. He is notably arguing against that conception of strong discretion in the judicial context, where it has been advanced by others.Footnote 82

Dworkin says that it would be “strange to say that a person who seeks to decide a troublesome question of conduct in terms of the moral standards of his community” had discretion.Footnote 83 I do not think it so strange, given the conception of discretion that I develop here. Discretion is not a shedding of judicial obligation.Footnote 84 When exercising discretion, the public decision-maker must “intend[] his reasoning to be based on public, not private, standards of good argument.”Footnote 85

When an administrative decision-maker faces hard cases of interpretation or application, it may be the case that the ultimate decision is controversial—that reasonable people will disagree about the reasoning, justification, or outcome. It will also be often wrong to say that any party had a right to one outcome versus another. It is nonetheless helpful to recognize this as a site of discretion. “[F]ew, if any, legal rules admit of no element of discretion in their interpretation and application.”Footnote 86 But it is not unfettered discretion. Only those decisions that are “based on an internally coherent and rational chain of analysis and that [are] justified in relation to the facts and law that constrain the decision maker” are reasonable.Footnote 87 Canadian law even recognizes that sometimes, this discretion collapses to a single reasonable outcome.Footnote 88 This is still called discretion and there are still public-law expectations for how that discretion is exercised.

Ultimately, discretion in public law is a process of reasoned decision-making, attentive to relevant factors as directed by the legislature and revealed by context. The exercise of discretion must establish anew a commitment to the validity, interpretation, and application of law in the often-unique circumstances of the subject before the decision-maker.

Where is discretion?

I will now identify several new sites of discretion in the ontology of the automated administrative state.

When an algorithm is deployed in the process of making a decision, there are many preceding and subsequent decisions to consider, aside from the decision or recommendation that might be exercised by the algorithm itself. There is the decision to employ an algorithm in the first place. There is the decision to select a particular algorithm. On the other side of algorithm selection is algorithm design, often performed by engineers and programmers wholly unaccountable to the people affected by the algorithm’s operation.Footnote 89 Front-line bureaucrats may play a role in shaping the input that is delivered to the algorithm.Footnote 90 And those same or other bureaucrats will have to interpret or give effect to the algorithmic output. All of these are sites of discretion to be constrained in various degrees by law.

Recognizing that these are sites of discretion is important for assessing the legitimacy and reasonableness of the outcome produced using the algorithm. First, just what is happening when an agency decides to employ an algorithm in its decision-making? If this is a delegation of decision-making power, moving the decision to the algorithm itself, the power to sub-delegate must generally be located in statutory authority.Footnote 91 Without such authority, this sub-delegation would be akin to fettering.Footnote 92 And even if the authority to sub-delegate can be located in statute or common law, I propose that sub-delegation to an algorithm is appropriate only in limited circumstances.

Specifically, when the constraints presented in the following section require discretion to be exercised, this cannot be delegated to an algorithm because the algorithm itself is not a site of discretion. Sure, algorithms may produce results that are unpredictable ahead of time or results that vary even when presented with the same input data.Footnote 93 But this unpredictability or randomness is not discretion. Such variation does not flow from a “reflexive[] refine[ment] of decision criteria.”Footnote 94 Any variation is limited by “decision boundaries effectively established ex ante.”Footnote 95 The dispersion of outputs may look from the outside like the dispersion of outputs produced by a person, but the process and reasoning that lead the algorithm to that place are nothing like what we would call discretion. Some people even question whether such disembodied reasoning can ever replicate human cognition.Footnote 96 Human thoughts, and discretion, are “inextricably associated with perception, action, and emotion, and … our brain and body work together to have cognition.”Footnote 97

But an algorithm may be used as a tool of the decision-maker rather than being deployed as a decision-maker itself. In such a case, the institutional relationships described above (design, procurement, use) involving the algorithm must be structured in a way that preserves the discretion in the decision-maker, lest the relationship become a site of jurispathos.

Up to this point, I have argued that the algorithm itself should be the decider only in limited circumstances where the legislature has sufficiently constrained discretion. But in all other circumstances, algorithms must be carefully controlled, to be used only as a tool by the ultimate human decision-maker(s). And no matter whether delegating to an algorithm or merely using an algorithm as a tool, the law has recognized constraints on this decision-making. The next section considers those.

What are the constraints?

I have alluded to various expectations placed on the exercise of discretion. In this section, I present the legal constraints that are placed on delegated administrative decision-makers in Canada when exercising discretion. Discretion can be viewed as an organizing principle that ties together various constraints. Violating the duty of non-arbitrariness, conveying a reasonable apprehension of bias, approaching a decision with a closed mind, or fettering one’s decision—these are all different ways in which a decision-maker may fail to make a decision with discretion. This section maps to Kenneth Culp Davis’s question of structuring discretion.Footnote 98 How must the decision be made in order to be made with discretion? As a starting point, discretion is “never absolute and beyond legal control.”Footnote 99 Discretion comes with duties.Footnote 100 These duties reflect the typical grounds of review within administrative law.

One duty is a duty to stay within the limits of delegated power. A somewhat outdated view is that the decision-maker has a preliminary duty to stay within their jurisdiction—within the scope of the delegated authority.Footnote 101 Under today’s framing, the question of the “limits and contours of [a] decision maker’s authority” is folded into the reasonableness review: “a decision that strays beyond the limits set by the statutory language it is interpreting” will be impossible to justify.Footnote 102

A second duty is non-arbitrariness. The decision cannot be based on capricious or irrelevant purposes or on bad faith.Footnote 103 Discretion is not choice based on “fancy or mere whim.”Footnote 104 It is this duty that precludes randomness as an aspect of discretion.

A third duty is that the decision-maker must not convey a reasonable apprehension of bias. Baker established that this duty is a part of procedural fairness.Footnote 105 At the highest level of generality, any person with a significant role in decision-making—no matter whether they are the ultimate decision maker or a subordinate officer—must act in an impartial manner.Footnote 106 The test is usually phrased in terms of what would give rise to an impermissible apprehension of bias:

[T]he apprehension of bias must be a reasonable one, held by reasonable and right minded persons, applying themselves to the question and obtaining thereon the required information.… [T]hat test is “what would an informed person, viewing the matter realistically and practically—and having thought the matter through—conclude. Would he think that it is more likely than not that [the decision-maker], whether consciously or unconsciously, would not decide fairly.”Footnote 107

This duty against a reasonable apprehension of bias has been most developed in the context of internal tribunal consultations with the ostensible goal of consistency.Footnote 108 Consistency in the sense of treating like cases alike is a legitimate goal for an administrative decision maker.Footnote 109 It is reasonable to “foster coherence” and “avoid … conflicting results.”Footnote 110 However, this consistency cannot come at the expense of discretion. The jurisprudence in this area has revealed a set of restrictions that tend to protect the decision-maker’s discretion and avoid the imposition of external bias: full board meetings regarding a decision must not be mandatory—they cannot be imposed on a decision-maker by a superior; no attendance or minutes are to be taken; the discussion must remain focused on law and policy; the ultimate decision must be left to the decision-maker(s) who heard the dispute.Footnote 111

A subset of that duty against bias—or, a relaxation of that duty in certain contexts—is the duty to approach the decision with an open mind. This duty is rooted in the nemo judex principle that applies to decision makers.Footnote 112 Some of the case law presents this duty as a prohibition on making statements that would indicate a closed mind.Footnote 113 The reviewing court will ask whether the statements “are the expression of a final opinion on the matter, which cannot be dislodged.”Footnote 114 Other case law states this duty more strongly: “the decision-maker must approach the issue submitted to him or her with an open mind, not influenced by personal interests or outside pressure.”Footnote 115 A unifying question is whether any “submissions would be futile.”Footnote 116

Fourth, there is the basic duty to actually exercise discretion, a breach of which is considered fettering. While fettering is often discussed alongside the duty to maintain an open mind, they are generally considered separate grounds of review.Footnote 117 Fettering has been placed under the umbrella of “abuse of discretion.”Footnote 118 It has in the past been unclear whether fettering is an independent, nominate ground for judicial review.Footnote 119 But it is now clear that Canadian courts conceive of fettering as undermining substantive acceptability and it is encompassed within reasonableness review.Footnote 120 When the legislature has granted a sphere of discretion, the decision maker “cannot, in a binding way, cut down that scope.”Footnote 121 This issue often arises from the use of guidelines, jurisprudential guides, or leading cases.Footnote 122 Decision makers, especially agencies or tribunals comprising many distinct individuals, will naturally adopt coherence and consistency as a legitimate goal.Footnote 123 But strategies adopted to foster consistency and reduce discord must “not operate to fetter decision making.”Footnote 124 Fettering is an abdication of discretion. It is refusal to apply the principles of natural justice to the case in front of the decision maker. The exercise of discretion requires one to develop their own opinion on the basis of the particular facts of the case.Footnote 125 When a decision maker’s discretion is fettered, this precludes the development of internal commitment.

Here is a hypothetical scenario that I suggest complies with the constraints just canvassed. The British Columbia Society for the Prevention of Cruelty to Animals (SPCA) has the power to sell or give abandoned dogs to new owners.Footnote 126 Interested new owners complete an adoption form that includes biographical information, description of the intended home, how the dog will be cared for, details about previous pets, and willingness to have an SPCA representative do a home visit, among other things. The data from the form could very well be input to an algorithm that has been trained to produce a recommendation to an SPCA employee: approve, reject, or flag for further review. The employee sees this recommendation, turns their attention to any aspects of the adoption form that the algorithm has highlighted, and surveys the remainder of the input for anything suggesting deviation from the algorithm’s recommendation. Seeing none, the employee follows the recommendation of the algorithm. In such a scenario, the employee retained and exercised discretion. The algorithm was an efficiency tool, not a decision-maker.

The analysis changes if an algorithm is left to exercise discretion itself. Certainly, there are several of the duties just presented that the algorithm would meet by design. As a preliminary matter, barring a physical failure, the algorithm would not fail to produce an outcome; that is, it would at least exercise the power to make a decision, often even considering the factors required by statute. And regarding the above duties, the algorithm cannot act in bad faith. But that does not imply that it would be acting in good faith or non-arbitrarily.

The Consolidated-Bathurst/Tremblay/Ellis-Don/Shuttleworth Footnote 127 arc further suggests that the use of algorithmic tools in discretionary decision-making should not be made mandatory by an agency absent the direction of the legislature. That line of cases establishes that internal agency mechanisms intending to foster consistency must not turn into constraints with outsized influence. Consultative proceedings cannot be imposed on the decision maker. For example, in Tremblay, the Court found that “the referral process [to plenary meetings] … circumvents the will of the legislature by seeking to establish a prior consensus by persons not responsible for deciding the case.”Footnote 128 Likewise, when an algorithmic tool is provided to a decision-maker, an administrative body should question whether it has maintained an environment in which decision-makers feel free to refuse to use the tool. If an administrative body were to demand that decision-makers achieve a degree of throughput that is unachievable without resort to algorithmic assistance, this may “compel or induce”Footnote 129 decision-makers to decide against their own opinions, or rather, fail to even form their own opinions.Footnote 130 The risk is that the algorithm will serve to import and instill a normative world that will “tower in importance over the others.”Footnote 131

Especially widely researched today is the risk of bias within an algorithm. Here, I refer to a conception of bias that is akin to prejudice or discrimination.Footnote 132 This could potentially run afoul of human rights codes or the Canadian Charter of Rights and Freedoms Footnote 133 and may give rise to a reasonable apprehension of bias. However, this is not the focus of this paper. Even if it were possible to craft an algorithm with ‘equitable’ error rates—for example, an algorithm with the same false-rejection rate across all demographics of interestFootnote 134 —my argument is that if that decision were made without an exercise of discretion, that lack of discretion would be an independent ground of criticism.

Again, although the algorithm may produce decisions, even if predictively accurate when compared to what a human decision-maker may have decided (or at least indistinguishable from the distribution of decisions that a human decision-maker would produce), the reasoning process will “bear little resemblance”Footnote 135 to that which would be exercised by a human. To achieve such a façade, an algorithm will either be following deterministic instructions, which is a fettering of discretion, or it will incorporate randomness, and that also isn’t what we expect from discretion.

Principles behind the constraints

The above duties are those that we have traditionally placed on human administrative decision-makers. Why do I argue that the constraints I extracted above are the appropriate administrative-law constraints to apply to the use of algorithmic decision-making? I propose this is justified by reference to principles behind the duties, namely: the expectations of the legislature, deliberative ideals, and the central role that expertise has played in our ontology of the administrative state.Footnote 136 This particular cross-section of principles is tied together by the overarching organizing principle of discretion as a prerequisite to legitimacy for certain decisions. While not exhaustive, this family of principles is especially useful when focusing the lens of judicial review on algorithmic decisions. And I admit that these may merely be motivating principles rather than principles that normatively justify the above constraints.

The decision to deploy an algorithm or otherwise eliminate discretion in public-law decision-making must grapple with a trade-off. On the one hand, as an instance of a decision-making system that necessarily treats its rules as non-defeasible, an algorithm may provide “certainty, predictability, settlement, and stability for stability’s sake.”Footnote 137 But at the same time, we miss out on the potential advantages of “fairness, equity, and, in theory, reaching the correct result in every instance.”Footnote 138 It is therefore a political decision, sometimes constitutionally constrained, to determine where the former values of certainty and the like might overtake the latter values of equity, correctness, and the like. When I suggest constitutional constraints, I am talking about, for example, decisions for which the over- or under-inclusiveness produced by an algorithmic decision-maker might deprive a person of a section 7Footnote 139 Charter interest—the right to life, liberty and security of person—in a manner that is overbroad or arbitrary;Footnote 140 or, as I suggest above, decisions that could discriminate in a manner contrary to the equality rights as outlined in section 15Footnote 141 of the Charter.Footnote 142 I do not propose to establish where those constitutional constraints lie. But within those constraints, the process of determining where the political and normative lines lie requires us to give effect to the expectations of the legislature, deliberative ideals, and expertise at the front lines. These principles are part of a procedural theory, rather than a substantive theory about precisely where the legislature might be justified in allowing or requiring an algorithm to be used for public decision-making.

These principles are important today as governments and legislatures attempt to deploy algorithms in a way that is consistent with the existing judicial constraints. The European Union has released draft regulation on artificial intelligence which would regulate the entire life-cycle of an AI system.Footnote 143 The Treasury Board of Canada has published a directive controlling how automated decision-making is used within much of the Canadian government.Footnote 144 And the Law Commission of Ontario is providing recommendations about what concerns regulation should address and some recommendations about how to go about that.Footnote 145 Whether any such regulation adequately constrains the use of algorithms will be up to the populace and the courts. Courts must consider whether the judicial constraints need to evolve in order to give effect to these principles within the new ontology of the automated administrative state. We should give effect to the expectations of the legislators, foster deliberative ideals at the front-line decision sites, and re-capture a pragmatic role for human expertise in the use of algorithms as tools.

First, legislators often expect discretion to be exercised. The nature of the freedom granted to administrative decision-makers is not due merely to the open texture of law and language (which of course, also exists),Footnote 146 but is often due to “avowed discretion.”Footnote 147 Legislatures delegate to administrative decision-makers the power to make rules, determine facts, and announce the rights of parties in particular disputes because of known unknowns. When the legislature is aware of its lack of foresight and aware that any rules it crafts would fall short of the task, it often will grant discretion.Footnote 148 “Interpreting a statute in a way that promotes effective public policy and administration may depend … upon the understanding and insights of the front-line agency.… Administration and interpretation go hand in glove.”Footnote 149 But along with this grant are expectations. The legislature may hope that rules and principles will “ultimately evolve in the course of the discretionary authority’s experience.”Footnote 150 And statutes are not a “one-way projection of authority originating with government and imposing itself on the citizen.”Footnote 151 So, the legislature may very well expect to share interpretative responsibility with institutions that are in contact with the subject.Footnote 152 When the legislature expects this work to happen at the front lines, executive agencies should not “cut down that scope”Footnote 153 of discretion by using automated decision-making.Footnote 154

Deliberative constitutionalism likewise expresses some collective expectations of delegated decision-makers. One premise is that participants be “willing to revise preferences in light of discussion, new information, and claims made by fellow participants.”Footnote 155 Under this theory, decision-makers must take a hard look at the unique situation of the individual affected by the decision. Even frontline decision-makers should be performing self-reflective inquiry about the proper scope and legitimacy of their actions, deciding to what degree they are ready to commit to state action.Footnote 156 The interactions between decision-makers and subjects should be “sites of deliberative empowerment.”Footnote 157 This is reflected in the courts’ emphasis in past decades on duties of procedural fairness, like the duty to be heard, and more recently on reason-giving. Geneviève Cartier suggests that the “centre of legal justification” be found in “collaborative practices of accountability and justification between courts, executive decision-makers and citizens.”Footnote 158 I would emphasize the importance of contingent commitment by a decision maker who can feel the stakes. “Legal interpretation takes place in a field of pain and death.”Footnote 159 But an algorithm does not cry; an algorithm does not die.

This view of governance can be undermined when algorithmic systems are deployed in the decision-making process because they separate humans from action and because they displace human commitment and fail to foster moral agency. But a third principle—a re-centering of human expertise—may operate to help further the previously-mentioned principles.Footnote 160 If systems are designed to retain and foster moral agency in human decision-makers throughout the development, deployment, and use of an automated decision-making tool, the threat to legislative intent and deliberative ideals can be minimized. It may be useful to adopt the keyword ‘tool’ to describe the role of an algorithm in public decision-making.Footnote 161

Organizational design that views an algorithm as a tool would place the algorithm in a relationship of feedback with humans.Footnote 162 A human decision-maker can retain and exercise discretion to adopt or reject any suggestion made by the algorithm. They can retain discretion to alter the algorithm’s interactions with subjects, to retrain the algorithm on improved data, to extract additional features from input data, or to discard the use of the algorithm altogether. The human managers of automated decision-making systems should be new experts capable of skeptical interaction with the automated decision-making tools they are using and should retain the agency to alter the way that those tools operate in practice.Footnote 163 I would caution, though, that skepticism should not become cynicism.

This new expertise will have to be different in content from the expertise that is assumed in the accepted ontology of the administrative state, but not different in purpose. Today, the presumptive posture of courts on judicial review is deferential, in part because of the expertise that is part of the ‘starting point’ the legislature had in mind when delegating certain decisions to an administrative decision-maker. This assumption of expertise lies behind the legislature’s delegation. For this assumption to actually be realized on the ground and continue to justify reasonableness as the default standard of review, decision-makers must not let themselves be displaced from their central position in this ontology.

The expertise required, in addition to the assumed subject-matter expertise, is a meta-expertise: an ongoing, self-reflective assessment of one’s own position in the organizational system that leads to a decision.

Naïve approaches to maintaining human-in-the-loop decision-making often fail to account for the effects that an algorithm can have on human decision-making.Footnote 164 As just one example, empirical research has revealed that even if a risk-assessment algorithm helps a decision-maker come to more accurate risk estimates, it may at the same time induce risk aversion and lead to more negative decisions.Footnote 165 Lorne Sossin has highlighted the important role of system design and training in fostering proper application of Charter values in administrative decision-making.Footnote 166 Such system design and training is just as important for preserving discretion in the decision-maker.

Toward these ends, a great deal of practical work is left to be done on the ground: empirical research to uncover the tensions, pressures, and “impossible situations” faced by front-line decision makers,Footnote 167 identifying the informal but essential tactics that these workers deploy in order to develop and maintain their moral agency.Footnote 168 Algorithmic tools must be designed and evaluated with these goals in mind.Footnote 169

Conclusion

Algorithms and automated decision-making are here to stay. They play a role in automating routine, non-discretionary administrative tasks and assist in discretionary decision-making. The threat is that automated decision-making alters the relationships between traditional actors in a way that cuts down discretion and human commitment. Algorithmic decision makers can serve to fetter the discretion that the legislature and the populace expect to be exercised. And algorithms can undercut ideals of deliberative democracy and moral agency that are hoped to exist in front line decision-making.

Recognizing the above, there are only limited circumstances when delegation to an algorithm is unproblematic: sites with no expected discretion. Otherwise, given that the algorithm itself is not a site of discretion, we must strive to maintain discretion, moral agency, deliberative ideals, and human commitment throughout the system and the organizational design surrounding the algorithm. The courts have not yet updated their understanding of the ontology of this automated administrative state. But this understanding will have to evolve in order to scrutinize the new relationships and the new kinds of expertise that are needed to give effect to legislative intent and to protect individual rights.

Acknowledgments

I would like to thank Professor Mary Liston for supervising this research project and for our many engaging discussions. I also thank the participants of the Algorithms and Rule of Law reading group convened at the Peter A Allard School of Law.

References

1. Throughout this article, I use the terms algorithms, algorithmic decision-making, and automated decision-making. Not all algorithms are used to make decisions. They can be used to recommend or to direct attention, for example. And historically, non-automated algorithms have been used (e.g. tax formulae). However, when an algorithm implemented on a computer makes an ultimate decision, that becomes automated decision-making. Today, the terms algorithmic decision-making and automated decision-making can generally be used interchangeably. While the focus of this paper is on Canadian public law, such automated decision-making has implications for most Western legal systems, as automated decision-making begins to permeate the administrative state.

2. Daria Gritsenko & Matthew Wood, “Algorithmic Governance: A Modes of Governance Approach” (2020) 16:1 Regulation & Governance 1 at 3.

3. To be clear, this is not independent historical research on my part. I have relied heavily on secondary sources and reviews from case authority—for example, the reasons of Abella and Karakatsanis JJ in Canada (Minister of Citizenship and Immigration) v Vavilov, 2019 SCC 65 [Vavilov].

4. See Income War Tax Act, SC 1917, c 28.

5. Davide Panagia, “On the Possibilities of a Political Theory of Algorithms” (2021) 49:1 Political Theory 109 at 113.

6. Income War Tax Act, supra note 4, s 3(1)(a).

7. See Colleen M Flood & Jennifer Dolling, “A Historical Map for Administrative Law: There Be Dragons” in Colleen M Flood & Lorne Sossin, eds, Administrative Law in Context, 3rd ed (Emond, 2018) 1 at 6; Constitution Act, 1867 (UK), 30 & 31 Vict, c 3, preamble, reprinted in RSC 1985, Appendix II, No 5 (establishing “a Constitution similar in Principle to that of the United Kingdom”).

8. See generally Mary Liston, “The Most Opaque Branch? The (Un)accountable Growth of Executive Power in Modern Canadian Government” in Paul Daly, Richard Albert & Vanessa MacDonnell, eds, The Canadian Constitution in Transition (University of Toronto Press, 2018) 26 (also discussing the aggregation of executive power in the prime minister’s office, ibid at 32).

9. See Flood & Dolling, supra note 7 at 7.

10. See e.g. the Canada Post Corporation, Canada Post Corporation Act, RSC 1985, c C-10. See also Letter Mail Regulations, SOR/88-430; Undeliverable and Redirected Mail Regulations, CRC, c 1298.

11. See e.g. Insurance (Vehicle) Act, RSBC 1996, c 231, s 36(1).

12. See e.g. City of Vancouver, by-law No 9150, Animal Control By-law (19 October 2021) [Animal Control By-law].

13. Ibid.

14. Blencoe v British Columbia (Human Rights Commission), 2000 SCC 44 at para 151 (Lebel J, dissenting in part), citing William Wade & Christopher Forsyth, Administrative Law, 7th ed (Clarendon Press, 1994) at 649.

15. Animal Control By-law, supra note 12, s 1.2.

16. See e.g. Anisminic Ltd v Foreign Compensation Commission, [1969] 2 AC 147; Metropolitan Life Insurance Co v International Union of Operating Engineers, [1970] SCR 425; Bell v Ontario Human Rights Commission, [1971] SCR 756. These cases adopt a formalistic sense of ‘jurisdiction,’ requiring the decision-maker to stay within its jurisdiction, both in embarking on an inquiry and in the questions it asks during its work.

17. Kevin M Stack, “Overcoming Dicey in Administrative Law” (2018) 68:2 UTLJ 293 at 310, cited in Vavilov, supra note 3 at para 211 (Abella and Karakatsanis JJ, in concurrence but dissenting on the reasons).

18. See Flood & Dolling, supra note 7 at 11.

19. See Pushpanathan v Canada (Minister of Citizenship and Immigration), [1998] 1 SCR 982 at paras 33-34 [Pushpanathan].

20. See Baker v Canada (Minister of Citizenship and Immigration), [1999] 2 SCR 817 at para 53 [Baker].

21. Ibid .

22. See Roncarelli v Duplessis, [1959] SCR 121 at 140 (Rand J, concurring) [Roncarelli].

23. Ibid at 140 (Rand J, concurring).

24. See Baker, supra note 20 at paras 54-56, that there is no “rigid dichotomy of ‘discretionary’ or ‘non-discretionary’ decisions” and that the “standard of review of the substantive aspects of discretionary decisions is best approached within this framework”—at the time, the framework being the pragmatic and functional approach to selecting a standard of review developed in Pezim, Southam, and Pushpanathan. See Pezim v British Columbia (Superintendent of Brokers), [1994] 2 SCR 557; Canada (Director of Investigation and Research) v Southam Inc, [1997] 1 SCR 748; Pushpanathan, supra note 19.

25. See Wilson v Atomic Energy of Canada Ltd, 2015 FCA 17 at para 51 (Stratas J); Dunsmuir v New Brunswick, 2008 SCC 9 at paras 27–31 [Dunsmuir]; Vavilov, supra note 3 at para 30.

26. Wilson v Atomic Energy of Canada Ltd, 2016 SCC 29 at para 28, quoting Dunsmuir, supra note 25 at paras 27-28.

27. Dunsmuir, supra note 25 at para 30.

28. 2011 SCC 61 [Alberta Teachers].

29. 2016 SCC 47 [Edmonton East].

30. Alberta Teachers, supra note 28 at para 45. Just two years earlier, in Canada (Citizenship and Immigration) v Khosa, 2009 SCC 12, Rothstein J held the opposite position in dissent: that it is legislative intent that can justify deference and that absent a privative clause, extractable questions of law should be reviewed under a correctness standard.

31. See Alberta Teachers, supra note 28 at paras 95-101.

32. Ibid at para 96.

33. Ibid at para 83.

34. Edmonton East, supra note 29 at para 33 (Karakatsanis J, for the majority).

35. Ibid .

36. Ibid at para 66 (Côté and Brown JJ, for the dissent).

37. Ibid at para 85.

38. National Corn Growers Assn v Canada (Import Tribunal), [1990] 2 SCR 1324 at 1334 (Wilson J, in concurrence) [National Corn Growers], citing PP Craig, “Dicey: Unitary, Self-Correcting Democracy and Public Law” in PP Craig, Public Law and Democracy in the United Kingdom and the United States of America (Clarendon Press, 1990) at 118-19.

39. See Vavilov, supra note 3.

40. Vavilov, supra note 3 at para 53.

41. Alberta Teachers, supra note 28 at para 30.

42. Vavilov, supra note 3 at para 58.

43. Ibid at para 92.

44. See ibid at para 88: “The administrative decision makers whose decisions may be subject to judicial review include specialized tribunals exercising adjudicative functions, independent regulatory bodies, ministers, front-line decision makers, and more. Their decisions vary in complexity and importance, ranging from the routine to the life-altering. These include matters of ‘high policy’ on the one hand and ‘pure law’ on the other.”

45. Thorne’s Hardware Ltd v The Queen, [1983] 1 SCR 106 at 111.

46. Attorney General of Canada v Inuit Tapirisat et al, [1980] 2 SCR 735 at 754.

47. Katz Group Canada Inc v Ontario (Health and Long-Term Care), 2013 SCC 64 at paras 25, 28.

48. Catalyst Paper Corp v North Cowichan (District), 2012 SCC 2 at paras 24, 30.

49. See e.g. Gritsenko & Wood, supra note 2 at 14 (noting that “algorithmic tools have an impact on the ways actors communicate, build, and maintain relationships”). See also Deirdre K Mulligan & Kenneth A Bamberger, “Procurement as Policy: Administrative Process for Machine Learning” (2019) 34:3 BTLJ 773.

50. See generally Gritsenko & Wood, supra note 2 at 13-14. And while algorithms get between actors, they do not actively nurture a positive relationship. Algorithms mediate only in the sense that they “occupy a middle position.” Raymond Williams, Keywords: A Vocabulary of Culture and Society (Oxford University Press, 1976) at 170 (under the entry ‘mediation’). Algorithms do not participate in a “political sense of mediation as reconciliation” ( ibid ).

51. See Mulligan & Bamberger, supra note 49 at 798 (presenting the various ways algorithm design embeds policy).

52. See Gritsenko & Wood, supra note 2 at 12-14.

53. See e.g. Jennifer Raso, “Unity in the Eye of the Beholder? Reasons for Decision in Theory and Practice in the Ontario Works Program” (2020) 70:1 UTLJ 1.

54. See e.g. Petra Molnar & Lex Gill, “Bots at the Gate: A Human Rights Analysis of Automated Decision-Making in Canada’s Immigration and Refugee System,” (2018) Citizen Lab & International Human Rights Program (Faculty of Law, University of Toronto) Research Report No 114.

55. Robert M Cover, “Violence and the Word” (1986) 95:8 Yale LJ 1601 at 1605.

56. See Gritsenko & Wood, supra note 2 at 4 (“choices made by algorithm designers translate their values into the system”).

57. See Bernardo Zacka, When the State Meets the Street: Public Service and Moral Agency (Harvard University Press, 2017) at 11 (“public agencies rely for their proper functioning on the moral agency of street-level bureaucrats”). See also Molnar & Gill, supra note 54 at 54 (commenting on the “real risk that human decision-makers will sometimes behave in a highly deferential manner toward outcomes rendered by automated decision systems” because of a “cognitive bias that presumes technical systems will behave ‘scientifically,’ fairly, and accurately”).

58. See Robert M Cover, “Forward: Nomos and Narrative” (1983) 97:1 Harv L Rev 4 at 11.

59. Ibid at 11.

60. Ibid at 17.

61. See Zacka, supra note 57 at 13.

62. Ibid at 13. See also Ben Shneiderman, “Design Lessons from AI’s Two Grand Goals: Human Emulation and Useful Applications” (2020) 1:2 IEEE Transactions on Technology & Society 73 at 77 (“[t]he human is always the creative force—for discovery, innovation, art, music, etc.”; to this list, I would add interpretation and discretion as well).

63. See British Columbia’s Wildlife Act, RSBC 1996, c 488. It permits the minister to create a regulation that “provide for limited entry hunting authorizations to be issued by means of a lottery or other method of random selection among applicants.” Ibid , s 16(1)(b).

64. Kenneth Culp Davis, Discretionary Justice: A Preliminary Inquiry (Louisiana State University Press, 1969) at 13.

65. See generally ibid . Davis summarizes his framework: “[t]he vast quantities of unnecessary discretionary power that have grown up in our system should be cut back, and the discretionary power that is found to be necessary should be properly confined, structured, and checked” (ibid at 175).

66. See ibid at 28-29 (“[w]hat we need to do is to work … not to minimize discretion or to maximize its control, but to eliminate unnecessary discretion and to find the optimum degree of control”).

67. Ibid at 26-27.

68. See ibid at 16 (“[t]he answers are highly crystallized for old age and survivors claims, and computers do most of the work”).

69. See Baker, supra note 20 at para 54, citing Donald JM Brown & John M Evans, Judicial Review of Administrative Action in Canada (Canvasback, 1998) at 14-47 (“[t]he degree of discretion in a grant of power can range from one where the decision-maker is constrained only by the purposes and objects of the legislation, to one where it is so specific that there is almost no discretion involved”). See also Davis, supra note 64 at 22 (“[a] standard, principle, or rule can be so vague as to be meaningless, it can have slight meaning or considerable meaning, it can have some degree of controlling effect, or it can be so clear and compelling as to leave little or no room for discretion”).

70. Insurance Corporation of British Columbia, “Basic Insurance Tariff”, 6th revision (effective 1 May 2021), online (pdf): www.icbc.com/about-icbc/company-info/documents/bcuc/basic-tariff.pdf.

71. Fish Inspection Act, RSNL 1990, c F-12, s 5(1).

72. See Molnar & Gill, supra note 54.

73. See Special Direction IC2 to the British Columbia Utilities Commission, BC Reg 307/2004.

74. See HLA Hart, “Discretion” (2013) 127:2 Harv L Rev 652 at 657-58 (“a discretion” being the “authority to choose”; “discretion” meaning a “certain kind of wisdom or deliberation” that imbues the choice with laudable qualities [emphasis in original]). See also Geneviève Cartier, “Deliberative Ideals and Constitutionalism in the Administrative State” in Ron Levy et al, eds, The Cambridge Handbook of Deliberative Constitutionalism (Cambridge University Press, 2018) 57 at 60 (discretion as a “power” vs discretion as the manner in which a choice is made).

75. Especially in relation to fiduciary and trust law. See e.g. Hodgkinson v Simms, [1994] 3 SCR 377.

76. Roncarelli, supra note 22 at 140 (Rand J concurring).

77. See JH Grey, “Discretion in Administrative Law” (1979) 17:1 Osgoode Hall LJ 107 at 108, 132. See also Wesley Newcomb Hohfeld, “Some Fundamental Legal Conceptions as Applied in Judicial Reasoning” (1913) 23:1 Yale LJ 16 at 30.

78. See generally ibid .

79. Hart, supra note 74 at 656.

80. See Paul Daly, “The Inevitability of Discretion and Judgement in Front-Line Decision-Making in the Administrative State” (2020) 2:1 Journal of Commonwealth Law 99 at 129: “[w]hilst it is unrealistic and inappropriate to expect front-line officials to exercise discretion and judgement as lawyers would, it is nonetheless entirely realistic and wholly appropriate to expect them to make good faith efforts to remain within and to further the objectives of the applicable legal framework” [emphasis in original].

81. Ronald Dworkin, “Judicial Discretion” (1963) 60:21 Journal of Philosophy 624 at 631 [emphasis in original].

82. Ibid at 625 n 2.

83. Ibid at 633.

84. Ibid at 637.

85. Ibid at 632.

86. Thamotharem v Canada (Minister of Citizenship and Immigration), 2007 FCA 198 at para 58 [Thamotharem].

87. Vavilov, supra note 3 at para 85.

88. Ibid at para 124.

89. But see Proposal for a Regulation of the European Parliament and of the Council: Laying Down Harmonised Rules on Artificial Intelligence (Artificial Intelligence Act) and Amending Certain Union Legislative Acts, European Commission, 22 April 2021, 2021/0106(COD), online: eur-lex.europa.eu/procedure/FI/2021_106 [EU Draft Regulations] (which puts obligations on providers and distributors).

90. See e.g. Raso, supra note 53.

91. See RA MacDonald & M Paskell-Mede, “Annual Survey of Canadian Law: Administrative Law” (1981) 13:3 Ottawa L Rev 671 at 686 (this requirement has some exceptions: e.g. implied delegation authority from Ministers to Deputy Ministers).

92. See e.g. Haghighi v Canada (Minister of Citizenship and Immigration), [2000] 4 FC 407 (CA); Muliadi v Canada (Minister of Employment & Immigration), [1986] 2 FC 205 at 218 (CA): a visa officer adopted the decision of an improper delegate as his own; “though he was entitled to receive information on that subject from that source it remained his duty to decide the matter” [emphasis added].

93. See generally Thomas H Cormen et al, Introduction to Algorithms, 3rd ed (MIT Press, 2009) at ch 5. See also Lehman et al, “The Surprising Creativity of Digital Evolution: A Collection of Anecdotes from the Evolutionary Computation and Artificial Life Research Communities” (2020) 26:2 Artificial Life 274; Claudio Gallicchio & Simone Scardapane, “Deep Randomized Neural Networks” (2021) [unpublished, online (pdf): arxiv.org/pdf/2002.12287.pdf].

94. Gritsenko & Wood, supra note 2 at 13.

95. Ibid .

96. See Melanie Mitchell, “Why AI is Harder Than We Think” (2021) [unpublished, online (pdf): arxiv.org/pdf/2104.12871].

97. Ibid at 7, citing Rebecca Fincher-Kiefer, How the Body Shapes Knowledge: Empirical Support for Embodied Cognition (American Psychological Association, 2019) at Preface.

98. See Davis, supra note 64 at 85: “[s]tructuring discretionary power is different from confining it.… The purpose of structuring is to control the manner of the exercise of discretionary power.”

99. Thamotharem, supra note 86 at para 58, citing Roncarelli, supra note 22 at 140 (Rand J, concurring).

100. See Grey, supra note 77 at 108-09.

101. The ‘jurisdictional error’ approach. See supra note 16. But the Supreme Court of Canada has pulled away from this approach starting with CUPE v NB Liquor Corporation, [1979] 2 SCR 227, and Vavilov, supra note 3 almost completely avoids such a framing.

102. Vavilov, supra note 3 at para 110.

103. See Roncarelli, supra note 22 at 140; Baker, supra note 20 at para 53.

104. Hart, supra note 74 at 657.

105. See Baker, supra note 20 at para 45.

106. Ibid .

107. Ibid at para 46, quoting de Grandpré J, in dissent, in Committee for Justice and Liberty v. National Energy Board, [1978] 1 SCR 369 at 394.

108. See e.g. IWA v Consolidated-Bathurst Packaging Ltd, [1990] 1 SCR 282 [Consolidated-Bathurst]; Tremblay v Quebec, [1992] 1 SCR 952 [Tremblay]; Ellis-Don Ltd v Ontario (Labour Relations Board), 2001 SCC 4 [Ellis-Don]; Shuttleworth v Ontario (Safety, Licensing Appeals and Standards Tribunals), 2019 ONCA 518 [Shuttleworth].

109. See Vavilov, supra note 3 at paras 129-132.

110. Ibid at para 130, citing Consolidated-Bathurst, supra note 108 at 324-28.

111. See Consolidated-Bathurst, supra note 108; Ellis-Don, supra note 108.

112. But see Old St Boniface Residents Assn Inc v Winnipeg (City), [1990] 3 SCR 1170 at 1190 (Sopinka J, “[t]he rules which require a tribunal to maintain an open mind and to be free of bias, actual or perceived, are part of the audi alteram partem principle which applies to decision-makers”) [Old St Boniface].

113. See e.g. ibid ; Newfoundland Telephone Co v Newfoundland (Board of Commissioners of Public Utilities), [1992] 1 SCR 623 [Newfoundland Telephone].

114. Old St Boniface, supra note 112 at 1197.

115. Imperial Oil Ltd v Quebec (Minister of the Environment), 2003 SCC 58 at para 28.

116. Newfoundland Telephone, supra note 113 at 642.

117. See e.g. Elson v Canada (Attorney General), 2019 FCA 27 (separates the questions of whether the Minister fettered his discretion from whether he had an open mind) [Elson]. But see Canada (National Revenue) v JP Morgan Asset Management (Canada) Inc, 2013 FCA 250 at para 90 [JP Morgan], which summarizes Stemijon Investments Ltd v Canada (Attorney General), 2011 FCA 299 [Stemijon] to say, “the Minister must have an open mind and cannot fetter her discretion.” Stemijon, however, only went on to discuss only fettering, not the open mind standard, suggesting that the analysis of fettering covered off both issues.

118. JP Morgan, supra note 117 at para 72, citing David J Mullan, Administrative Law (Irwin Law, 2001) at 100-13.

119. See e.g. Stemijon, supra note 117 at paras 21-23.

120. See Vavilov, supra note 3 at 108 “where a decision maker is given wide discretion, it would be unreasonable for it to fetter that discretion.” See also JP Morgan, supra note 117 at para 74 (“the current view is that these are not nominate categories of review, but rather matters falling for consideration under … reasonableness review”); Elson, supra note 117 at para 28 (which asks “Did Minister Tootoo fetter his discretion which would make his decision unreasonable?”).

121. Stemijon, supra note 117 at para 22.

122. I see no principled reason to treat discretion hemmed in by guidelines or leading cases as fettering, while treating discretion hemmed in by pressure through board meetings as a reasonable apprehension of bias. In my view, fettering is best seen as a special instance of actual bias, but it lies at the end of a spectrum of improper influence that cuts into the sphere of delegated discretion. However, this extreme endpoint of the spectrum of bias is particularly relevant in the context of algorithmic decision-making, as I discuss next.

123. See Canadian Association of Refugee Lawyers v Canada (Immigration, Refugees and Citizenship), 2020 FCA 196 at para 66 [CARL]. See also Vavilov, supra note 3 at paras 129-32.

124. Vavilov, supra note 3 at para 130. See also Davis, supra note 64 at 53 (“administrative rules which are too rigid sometimes grow out of administrative adjudication”). See also Australian Administrative Review Council, “Automated Assistance in Administrative Decision Making: Report to the Attorney-General (Report No 46)” (2004), online (pdf): Australian Government Attorney-General’s Department https://www.ag.gov.au/sites/default/files/2020-03/report-46.pdf [Australian Administrative Review Council] (expressing concerns that automated decision-making might “incorporat[e] policy that inappropriately narrows the available discretion,” ibid at 33; advocating that “[e]xpert systems that make a decision—as opposed to helping a decision maker make a decision—would generally be suitable only for decisions involving non-discretionary elements,” ibid at 16).

125. See CARL, supra note 123 at para 69.

126. See Prevention of Cruelty to Animals Act, RSBC 1996, s 17(a). The algorithmic procedure I describe next in this paragraph is purely hypothetical. It is not the mechanism used by the BC SPCA.

127. See supra note 108.

128. Tremblay, supra note 108 at 974.

129. Shuttleworth, supra note 108 at para 27.

130. See e.g. Zacka, supra note 57 at 208-209, 232-233 (discussing how “impossible” caseload mandates result in undesirable “detachment and indifference”).

131. Zacka, supra note 57 at 246. See also Cover, supra note 55 at 1615 (discussing how pronouncement of the law by a judge is a mechanism through which “a substantial part of their audience loses its capacity to think and act autonomously”).

132. See e.g. Vivek Krishnamurthy, “AI and Human Rights” in Florian Martin-Bariteau & Teresa Scassa, eds, Artificial Intelligence and the Law in Canada (LexisNexis, 2021); Anya Prince & Daniel Schwarcz, “Proxy Discrimination in the Age of Artificial Intelligence and Big Data” (2020) 105 Iowa L Rev 1257.

133. See Canadian Charter of Rights and Freedoms, Part I of the Constitution Act, 1982, being Schedule B to the Canada Act 1982 (UK), 1982, c 11 [Charter].

134. This is, of course, a naïve notion of equality in assessing algorithmic error rates. Even if an algorithm achieves such demographic parity, false rejections can have heightened effects on people that are experiencing overlapping disadvantages. See generally Kimberle Crenshaw, “Mapping the Margins: Intersectionality, Identity, and Violence Against Women of Color” (1991) 43:6 Stan L Rev 1241.

135. Mulligan & Bamberger, supra note 49 at 823.

136. This is just one possible cross-section of principles. Paul Daly provides an alternative but not inconsistent catalogue of values in administrative law: the rule of law, good administration, democracy, and separation of powers. See Paul Daly, “Administrative Law: A Values-Based Approach” in John Bell et al, eds, Public Law Adjudication in Common Law Systems: Process and Substance (Hart, 2016) 23.

137. Frederick Schauer, “On the Open Texture of Law” (2013) 87 Grazer Philosophische Studien 197 at 212.

138. Ibid .

139. See Charter, supra note 133, s 7.

140. See generally Hamish Stewart, “Bedford and the Structure of Section 7” (2015) 60:3 McGill LJ 575.

141. See Charter, supra note 133, s 15.

142. See the text accompanying supra notes 132, 134.

143. See EU Draft Regulations, supra note 89.

144. See Government of Canada, “Directive on Automated Decision-Making” (1 April 2021), online: Government of Canada, www.tbs-sct.gc.ca/pol/doc-eng.aspx?id=32592.

145. See Law Commission of Ontario, “LCO Issue Paper, Regulating AI: Critical Issues and Choices” (2021), online: LCO https://www.lco-cdo.org/wp-content/uploads/2021/04/LCO-Regulating-AI-Critical-Issues-and-Choices-Toronto-April-2021-1.pdf.

146. See generally HLA Hart, The Concept of Law, 3rd ed (Oxford University Press, 2012) at ch VII s 1.

147. Hart, supra note 74 at 656.

148. See National Corn Growers, supra note 38 at 1336 (Wilson J, in concurrence, remarking on the “growing acceptance by the courts that statutory provisions often do not yield a single, uniquely correct interpretation, but can be ambiguous or silent on a particular question, or couched in language that obviously invites the exercise of discretion”), citing John M Evans et al, Administrative Law, 3rd ed (Emond, 1989) at 114.

149. Ibid [emphasis removed].

150. Hart, supra note 74 at 656.

151. David Dyzenhaus, “Deliberative Constitutionalism Through the Lens of the Administrative State” in Ron Levy et al, supra note 74, 44 at 50, citing Lon L Fuller, The Morality of Law (Yale University Press, 1969) at 207.

152. See Dyzenhaus, supra note 151.

153. Stemijon, supra note 117 at para 22. See also Australian Administrative Review Council, supra note 124.

154. See e.g. Tremblay, supra note 108 at 974: “The tribunal hearing a new question may thus render a number of contradictory judgments before a consensus naturally emerges. This of course is a longer process; but there is no indication that the legislature intended it to be otherwise.”

155. Cartier, supra note 74 at 59. See also Dyzenhaus, supra note 151 at 44.

156. Here, as I present above, I am using commitment in the sense developed by Robert M Cover. See text accompanying supra notes 52-57.

157. Cartier, supra note 74 at 61 (presenting discretion as dialogue).

158. Ibid at 64.

159. Cover, supra note 55 at 1601.

160. See e.g. Lorne Sossin, “The Unfinished Project of Roncarelli v Duplessis: Justiciability, Discretion, and the Limits of the Rule of Law” (2010) 55:3 McGill LJ 661 at 665 (“[d]iscretion is also bound up with the principle of deference to the experience and expertise of specialized administrative decision-makers”).

161. See Raymond Williams, Keywords: A Vocabulary of Culture and Society (Oxford University Press, 1976). A ‘keyword’ is a “significant, indicative word[] in certain forms of thought.” Ibid at 13. While Williams does not include ‘tool’ as one of his listed keywords, I take inspiration from his listing of keywords and hope to elevate ‘tool’ to the status of keyword in this context. I would hope that this helps us identify our assumptions behind the use of algorithms as tools rather than as decision-makers and that this further helps define our relationships with the algorithms and with each other. Mark Antaki takes a similar turn with the keyword ‘imagination.’ See Mark Antaki, “The Turn to Imagination in Legal Theory” (2012) 23:1 Law Critique 1.

162. See Panagia, supra note 5 at 126.

163. See generally Ryan Calo & Danielle Keats Citron, “The Automated Administrative State: A Crisis of Legitimacy” (2021) 70 Emory LJ 797 at 835. See also Mulligan & Bamberger, supra note 49 at 857 (with the aspiration that these new tools will be “aligned with values chosen based on reason, expertise, transparency, and robust and ongoing deliberation and oversight”); Shneiderman, supra note 62 at 78 (“humans and machines are embedded in complex organizational and social systems, making interdependence an important goal” and “[s]ince humans remain as responsible actors (legally, morally, and ethically), should not computers be designed in ways that assure user control?”); Sossin, supra note 160 at 664 (discussing potential elements of system design that would foster a desired culture: “published guidelines, ministerial supervision, to the training, expertise, and professionalism of the public service”).

164. See e.g. Ben Green, “The Flaws of Policies Requiring Human Oversight of Government Algorithms” (2022) 45 Computer Law & Security Review at 15-16.

165. See Ben Green & Yiling Chen, “Algorithmic Risk Assessments Can Alter Human Decision-Making Processes in High-Stakes Government Contexts” (2021) 5:CSW2 Proceedings of the ACM on Human-Computer Interaction 1.

166. See Lorne Sossin, “Discretion Unbound: Reconciling the Charter and Soft Law” (2002) 45:4 Canadian Public Administration 465 at 467 (emphasizing the role played by “policy guidelines, technical manuals, rules, codes, operational memoranda, training materials, interpretive bulletins, or, more informally, through oral directive or simply as a matter of ingrained administrative culture”). See also Lorne Sossin & Mark Friedman, “Charter Values and Administrative Justice” (2014) 67 SCLR (2d) 391.

167. Zacka, supra note 57 at 200 (describing the incapacitating effect of exposure to such “impossible situations”).

168. See e.g. Raso, supra note 53 (discussing the creative ways in which front-line workers find ways to retain agency in light of divergence between legislation and software).

169. See Shneiderman, supra note 62 at 81: “[d]esign compromises, which combine AI with [human-computer interaction] methods, need to be further shaped by the contextual needs of each application domain and thoroughly tested with real users.”