Skip to main content Accessibility help
×
Hostname: page-component-586b7cd67f-vdxz6 Total loading time: 0 Render date: 2024-11-22T21:57:01.016Z Has data issue: false hasContentIssue false

8 - Democratic Obligations and Technological Threats to Legitimacy

from Part IV - The Responsibilities of Agents

Published online by Cambridge University Press:  17 May 2021

Alan Rubel
Affiliation:
University of Wisconsin, Madison
Clinton Castro
Affiliation:
Florida International University
Adam Pham
Affiliation:
California Institute of Technology

Summary

This chapter addresses autonomy’s role in democratic governance. Political authority may be justifiable or not. Whether it is justified and how it can come to be justified is a question of political legitimacy, which is in turn a function of autonomy. We begin, in section 8.1, by describing two uses of technology: crime predicting technology used to drive policing practices and social media technology used to influence elections (including by Cambridge Analytica and by the Internet Research Agency). In section 8.2 we consider several views of legitimacy and argue for a hybrid version of normative legitimacy based on one recently offered by Fabienne Peter. In section 8.3 we explain that the connection between political legitimacy and autonomy is that legitimacy is grounded in legitimating processes, which are in turn based on autonomy. Algorithmic systems—among them PredPol and the Cambridge Analytica-Facebook-Internet Research Agency amalgam—can hinder that legitimation process and conflict with democratic legitimacy, as we argue in section 8.4. We conclude by returning to several cases that serve as through-lines to the book: Loomis, Wagner, and Houston Schools.

Type
Chapter
Information
Algorithms and Autonomy
The Ethics of Automated Decision Systems
, pp. 163 - 183
Publisher: Cambridge University Press
Print publication year: 2021
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - ND
This content is Open Access and distributed under the terms of the Creative Commons Attribution licence CC-BY-NC-ND 4.0 https://creativecommons.org/cclicenses/

8.1 Two New Technologies

In 2011, UCLA anthropology professor Jeff Brantingham launched PredPol, which is an algorithmic system for predicting and preventing crime. The idea at the heart of PredPol is that crimes follow predictable, geographic patterns.Footnote 1 Thus, it may be possible to reduce crime by deploying police resources to places where predictive analytics suggests crime will occur. To facilitate this, PredPol has a graphical interface that overlays a red box (indicating a predicted crime “hotspot”) on top of a city map. This allows police to concentrate on those areas in the hopes of deterring crime.

There is evidence that PredPol is effective in some respects, especially in addressing property crime. And it has a degree of support. Time magazine called predictive policingFootnote 2 one of the “50 Best Inventions” of 2011.Footnote 3 Today, PredPol is a national leader in predictive policing technology and dozens of cities across the United States use it.Footnote 4

There are, however, several important criticisms of the technology. One is that its ability to predict property crimes may not translate well to predicting (much less preventing) other crimes. Another concern is that predictive policing in general simply recreates biases in policing practices and that predictive tools will be deployed in ways that harm communities of color, much in the same way that stop-and-frisk policies, pretrial detention, and sentencing do.Footnote 5

Not long after PredPol was getting off the ground, a different set of algorithmic systems was developing in a way that would have profound implications for electoral politics. The political data analytics firm Cambridge Analytica began building specialized personal profiles and using large datasets collected from social media to stage political influence operations at scale. While use of profiling and data to influence politics is nothing new, Cambridge Analytica’s actions are novel in several ways. The data on which their operation was built was massive and obtained in intrusive ways (including by gaining access via friends’ permissions in Facebook). In addition, the data was collected under the imprimatur of a prestigious academic institution (Cambridge University). Finally, the kinds of information pushed to users were difficult for others to see, as it was targeted on social media. This left fewer obvious trails as it was happening than, say, television or mail ads. In parallel with the Cambridge Analytica efforts was a Russia-sponsored disinformation campaign, also making use of social media and recommendation systems. This was carried out by an organization called the Internet Research Agency (IRA).

Although PredPol and Cambridge Analytica/IRA address fundamentally different parts of modern life, they both have important implications for democratic governance. PredPol is a tool that helps structure how governments exercise their enormous power to investigate and sanction crime. The ability of the state to visit harms upon its constituents is at the heart of liberal theorists’ commitments to limit state power.Footnote 6 This concern about state-sanctioned harm is sharpened by concerns about unsanctioned state violence that is often unaccountable.Footnote 7 Cambridge Analytica, in contrast, is not primarily about how a state deploys its power. Rather, it is about the connection (or disconnection) between democratic processes and governance. Put differently, it is about the necessary conditions for democratic processes to provide grounds for governments to hold power and implement policy decisions.Footnote 8

Questions about justifying exercise of power via policing and questions about democratic processes and justifying power are, at root, about political legitimacy. So far in this book, we have examined algorithmic decision systems from three autonomy-based perspectives: in terms of what we owe autonomous agents (Chapters 3 and 4), in terms of the conditions required for people to act autonomously (Chapters 5 and 6), and in terms of the responsibilities of agents (Chapter 7). Political legitimacy is another way in which autonomy and responsibility are linked. This relationship is the basis of the current chapter, and it is important in understanding the moral salience of algorithmic systems. We will draw the connection as follows: In Section 8.2 we will outline a conception of political legitimacy. In Section 8.3 we will explain that the connection between political legitimacy and autonomy is that legitimacy is grounded in legitimating processes, which are in turn based on autonomy. Algorithmic systems – among them PredPol and the Cambridge Analytica-Facebook amalgam – can hinder that legitimation process and conflict with democratic legitimacy, as we argue in Section 8.4. We will conclude by returning to our old friends, Loomis, Wagner, and Houston Schools.

8.2 POLITICAL LEGITIMACY: THREE CONCEPTIONS AND A HYBRID VIEW

Governments exercise enormous power over their denizens and expend vast resources promoting a range of policies (education, transportation infrastructure, defense, public health, commercial regulation, scientific research, information collection, parklands, safety and policing, and on and on). Those powers can be exerted in ways that are justifiable or unjustifiable, that are good or bad, and that are useful and not useful. The exercise of political authority can, in other words, be legitimate or not. In its broadest sense, political legitimacy refers to the justification of political authority, where “authority” just means having a certain political right to act coercively. But what constitutes such a right, and how could anyone ever acquire it? In this section, we discuss three conceptions of legitimacy – descriptive, democratic (or “will” based), and epistemic (or “normative authority” based) – and then offer our own, hybrid view based on a recent account from Fabienne Peter.

8.2.1 Legitimacy and Descriptive Criteria

One family of views distinguishes legitimate from illegitimate authority in terms of purely descriptive (or empirical), as opposed to normative, criteria. This sort of view, which finds its roots in Max Weber’s work,Footnote 9 is often associated with the social sciences. Such descriptive accounts examine how authority (understood as people recognizing a body’s ability to exert power) can be established and maintained.Footnote 10

Weber himself offered three methods of establishing and maintaining the ability to exert power: through tradition, through charisma, and through rational appeal. In other words, the continuity of traditional leadership, the transformative personal characteristics of charismatic rulers, and the inherent rationality of law can each, for better or worse, contribute to establishing and maintaining a political regime. Later theorists have refined Weber’s list, distinguishing, for example, claims of traditional authority staked on “convention” from those staked on “sacredness” and distinguishing claims of charismatic authority based on “personal ties” from those based on “personal qualities.”Footnote 11 In the context of algorithmic systems, Ari Waldman provides an account of this descriptive sense of legitimacy as “the socially constructed propriety of authority to make decisions for others.”Footnote 12 Given that “[t]he managerial ethos inside corporations operating in a permissive, neoliberal regulatory environment will twist process to serve corporate ends,” legitimacy requires that regulators “go beyond process to rebalance the structures of power.”Footnote 13

Work on this process of establishing recognition of a right to exert authority is important (and we will return to it in Section 8.3). However, it addresses different questions than those about the normative foundation of authority. The mere fact (if it is) that people are willing to recognize a government’s authority to exert power on the basis of, say, a leader’s charisma does not tell us anything at all about whether the government’s actions are justifiable. To understand this, we must also understand how a right to exert authority could be justified by moral or political principles.Footnote 14

8.2.2 Legitimacy and Democratic Criteria

One normative conception of legitimacy holds that it is a function of democratic consent. Legitimacy in this sense depends on some sort of approval of authority by a state’s citizens. This can be cached out in a couple of different ways.

On one view, approval requires the kind of epistemic competence and non-alienation that are conditions of what we called psychological autonomy in Chapter 2. Most people never agree to be governed in this way, of course, so explicit consent cannot be the requirement for legitimacy. One might then consider hypothetical agreement as the grounds for legitimacy, arguing that a sufficient condition of legitimate authority is that reasonable persons could consent to it, or one might argue that it is a necessary condition of legitimate authority that people could consent to it. Indeed, the account of psychological autonomy from John Christman that we use as an exemplar in Chapter 2 links autonomy to legitimacy in this fashion. Christman writes that “the test for legitimacy of political principles should be the following: principles are legitimate only if the (reasonable) citizens to whom they apply would not be understandably alienated from them.”Footnote 15 What matters is not explicit agreement, but whether people would identify with political principles if given adequate opportunity to reflect on them in light of their values and the sources and history of those values.

Another view linking autonomy to legitimacy is based on the degree to which political bodies and decisions reflect constituent wills. What is important in legitimacy is that persons subject to authority have some say in how that authority is constituted and deployed. In this way, what Marina Oshana calls “political autonomy” more closely reflects her sense of personal autonomy than psychological autonomy.Footnote 16 Legitimacy in this sense turns on the degree to which political processes afford citizens the genuine opportunity to participate, rather than principles reflecting their wills.

Both views of autonomy ground legitimacy in democratic will. Rawls, for instance, offers a principle of legitimacy that is defined in terms of “public reason.” On this view, political power is justified “when it is exercised in accordance with a constitution the essentials of which all citizens as free and equal may reasonably be expected to endorse in the light of principles and ideals acceptable to their common human reason.”Footnote 17 Similarly, Pettit’s account is grounded in “popular control.” He argues that political power is justified when all citizens have equal access to the prevailing system of influence over the government and that system imposes a direction welcomed by all.Footnote 18 In both cases, the account of legitimacy is premised on the notion of individuals as the “self-originating sources of valid claims” in the sense that “their claims have weight apart from being derived from duties or obligations specified by the political conception of justice, for example, from duties and obligations owed to society.”Footnote 19 What gives their claims this sort of weight is their autonomous wills.

Both Rawls’s and Pettit’s views align with descriptive accounts in that both Rawls and Pettit acknowledge that legitimacy is tied to stability. However, they diverge from descriptive accounts in maintaining that legitimacy requires that political authority meet normative criteria as well. Specifically, political authority must obey certain rules and democratic procedures. Rawls describes the ideal of political legitimacy not in terms of stability per se (i.e., in terms of a “modus vivendi”), but in terms of “stability for the right reasons,” where each citizen is compelled to obey the public constitution from their own individual perspective rather than merely out of self-interest. And Pettit describes the ideal of legitimacy in terms of passing a “tough luck” test, according to which authority is legitimate when people can think that “when public structures and policies and decisions frustrate their personal preferences, that is just tough luck.”Footnote 20 Note that these normative criteria for political legitimacy are versions of the reasonable endorsement principles we outline in Chapter 3. Since these principles index the degree of legitimacy to facts about the wills of constituents, we can, following Fabienne Peter, call them “will” principles of legitimacy.

8.2.3 Legitimacy and Epistemic Criteria

A third view of legitimacy ties it to whether systems or decisions meet epistemic criteria.

To understand this set of views, it is useful to start with what Peter calls the “normative facts” view. On this conception, what matters for legitimacy turns on whether political decisions track the truth of the matter as to what is morally right to do. That is, the normative facts view assumes something akin to moral realism, thinks that propositions about what governments should do can be true or false, and pegs legitimacy to whether such decisions indeed track those normative facts. And they do so regardless of what people’s beliefs about them are.

The problem of a normative facts view is that it leaves out the link between legitimacy and autonomy. That is, legitimate decisions need not be ratified in any sense by the people subject to them. Peter explains this as a violation of what she calls the “access constraint”: Political legitimacy, whatever it is, must “be such that it can settle political deliberation (at least temporarily),” and to do this, it “must involve the attitudes of at least some citizens.” Therefore, if normative facts are relevant to our political decision-making, it will be “through our beliefs about them.”Footnote 21

In this vein (i.e., bridging normative facts and beliefs about normative facts), Raz offers a “service” conception of legitimacy. On this view political authority is legitimate only if it is made in service of people’s underlying reasons. The “normal” justification for authority, in other words, is that authority done right does nothing more than guide people according to the reasons they already have. As Raz puts it, “[T]he normal way to establish that a person has authority over another person involves showing that the alleged subject is likely better to comply with reasons which apply to him […] if he accepts the directives of the alleged authority as authoritatively binding and tries to follow them, rather than by trying to follow the reasons which apply to him directly.”Footnote 22 Later theorists following Raz have explained further how this sort of epistemic preemption might work.

David Enoch, for instance, extends Raz’s notion of a duty as a special sort of reason, describing authority as giving rise to obligation when there is some “normative structure in the background” allowing the authority to give its subject those sorts of reasons.Footnote 23 On wildlife tours, the tour guide often serves (and expects to serve) as this sort of preemptive decision-maker. The guide is the authority on, for instance, when the group needs to leave a group of animals behind rather than continue to take pictures – because of the dangerous nature of the circumstances and the special knowledge the guide has acquired through experience over time.

In any case, when the normal justification is present, Raz argues, the duties prescribed by the authority can override even the dictates of the democratic will and thus can offer grounds for “a ready embrace of various paternalistic measures,” which would not pass democratic muster.Footnote 24 Here, we might find the justification for such policies as smoking bans, bans on trans fats, helmet laws, and mask mandates.

8.2.4 The Disjunctive Conception of Legitimacy

As Peter points out, neither the democratic nor the epistemic criteria seem on their own to fully encapsulate the normative grounds of political legitimacy.Footnote 25 If we accept what she calls a “will-based” conception of political legitimacy, exemplified by Rawls’s and Pettit’s democratic criteria, we run into what Peter calls the arbitrariness objection: that “[t]he validity of at least some practical claims depends on third-personal sources of validity,” and thus that these will-based conceptions “support undue arbitrariness in political decision-making.” We cannot hold the wills of all citizens to be politically relevant; individuals are not self-originating sources of valid claims but are instead often deeply irrational or unreasonable. Hence, democratic processes (via voting, consensus, public reason, or the like) are not sufficient on their own to justify exercise of political authority. There are, in other words, exercises of authority that are illegitimate even if they genuinely reflect citizen will. They may be so ill-conceived and so harmful, or they may be so detrimental to democratic processes and structures themselves, that they can not be legitimated by democratic processes.

A purely “belief-based” conception of political legitimacy (corresponding to Raz’s and Enoch’s epistemic criteria) is also inadequate. Peters argues that these accounts run the problem of “epistemic underdetermination.” Specifically, there are few (if any) political decisions for which there is sufficient evidence to ensure that they are optimal. As Peter puts it, “[T]he epistemic circumstances of politics are such that for most political decisions, there will not be a decisive normative authority.”Footnote 26 So even if we reject democratic adjudication as the univocal source of normative authority, the epistemic circumstances are often so complex and uncertain that normative authority remains difficult or impossible to establish. The main tension is that persons’ wills and epistemic criteria both seem to be key conditions of justifiable exercise of political authority. Thus, the determinative question seems to be how to incorporate the autonomous wills and beliefs of citizens, given the need for policy that is appropriately responsive to facts.Footnote 27

To address these issues, Peter defends a hybrid account of the normative grounds of political legitimacy. This account holds that legitimacy includes “both responsiveness to normative authority and adjudication between valid, but conflicting, claims as grounds of legitimacy.” The conception itself is “disjunctive,” which is to say that legitimacy can derive from either will or belief. Moreover, it favors normative authority over democratic processes; legitimacy is tied to “how the decision reflects normative authority, when normative authority can be established, or, when normative authority cannot be established, […] how it adjudicates between the conflicting, but valid claims made by the citizens.” In other words, democratic adjudication is secondary; it is a source of legitimacy only “as a response to difficulties with establishing normative authority.”

Peter’s hybrid account of the grounds of legitimacy is on the right track, but we want to refine it slightly. We cannot define our way out of the threats to legitimacy presented by our most difficult and contentious cases simply through disjunctive addition. It is not as if the arbitrariness problem somehow disappears in cases where normative authority cannot be established. (Indeed, it seems more likely that the two problems are self-reinforcing and thus run together.) It may well be possible to balance power among parties to a conflict in some cases where normative authority is epistemically underdetermined, but we should not mistake this modus vivendi for an arrangement that is genuinely politically legitimate.

Peter’s concern with securing a secondary ground of legitimacy beyond normative authority is connected to her concern with the function of political legitimacy. She writes that “[a] plausible conception of political legitimacy should be able to settle which decisions are and which are not legitimate even under unfavourable epistemic circumstances,” and this is something that can often only be settled by the balancing of wills. Nor, she argues, can we “claim that all decisions made in the absence of decisive normative authority would be illegitimate,” because this “would itself amount to a political decision about which we could ask whether it is legitimate,” leading to a regress.

Peter considers the secondary role of democratic adjudication in the context of a well-entrenched topic of disagreement: abortion. If we grant for the sake of argument that there is no decisive normative authority on that issue, it is not clear that any political method of adjudicating the conflicting claims on the issue will ever be able to provide the sort of consolation that might soothe parties who have been disenfranchised by “tough luck,” to say nothing of providing stability for the right reasons. This seems like precisely the sort of issue that democratic adjudication can settle, but not in the “right” way.

To see the grounds and function of political legitimacy correctly, we need to keep the secondary notion of democratic adjudication separate from the primary notion of normative authority. This is because there are political decisions, systems, or regimes that might be democratically ratified, but that are illegitimate in the epistemic, normative authority sense. Thus, there will be some policies that appear legitimate in the democratic sense but are not legitimate in the normative authority sense. Our view is similar to Peter’s hybrid model, with the caveat that policies can be legitimate via a combination of normative authority and democratic will. As evidence of good policy increases (i.e., as it becomes increasingly evident that some policy, law, or action is justified via normative authority), then the stringency of the democratic facet of legitimacy is weaker.

8.3 Legitimating Processes

As we explained in the previous section, there are two facets to normative legitimacy. One, characterized by “will” theories generally, is vulnerable to arbitrariness. The other is characterized by normative authority, which is in turn a function of justified belief in structures, policies, and actions of states that are in fact good. This second facet is not vulnerable to arbitrariness in the way that will theories are. That is because justification is baked into the definition. Any policy, structure, or action that is both correct and justifiably believed to be correct is, ipso facto, legitimate. The primary limitation of the normative authority facet of legitimacy is epistemic; determining what is in fact the right policy or action is difficult, to say the least (cf., Section 1.5).

When we consider technological threats to legitimacy – gaming of algorithms in social media, use of algorithms in predictive policing, for example – it is tempting to move straight from a conception of legitimacy to a claim that some set of institutions, policies, or structures are themselves illegitimate, all things considered. It will no doubt be true that some political actions will fail to meet both the will conditions and the normative authority conditions for legitimacy and those actions will thus be unjustifiable exercises of authority. However, for any particular policy taking place within a complex set of social and political structures against background facts that are themselves challenging to interpret, all-things-considered judgments about legitimacy will be difficult. Hence, it is a far larger task than we can address here. Moreover, making that kind of final judgment about legitimacy itself would take us too far afield from the subject of this book, which is about autonomy. Finally, legitimacy (like autonomy) is not a binary, success concept. It is not something that an institution, polity, law, or action simply has or lacks. There is a spectrum of legitimacy, and limitations based on will or based on normative authority will often diminish legitimacy without warranting a conclusion that a policy decision crosses some floor below which it is illegitimate.

For these reasons, our focus here is on legitimation.Footnote 28 Will theories, normative authority theories, Peter’s disjunctive theory, and our hybrid interpretation of Peter’s view all rely on exercises of persons’ autonomy as a foundation for legitimacy of governments, laws, policies, and the like. That process of exercising autonomy, either through will or through justified beliefs about normative claims (or both), is the mechanism by which use of political authority is justified. And in turn, members of a polity have the responsibility to use their autonomy to understand and support (or not support) actions by their government.Footnote 29 It is that process of legitimation that algorithmic systems can disrupt. In other words, algorithmic systems may in some cases hinder people from fulfilling their responsibilities of legitimating government policies, actions, laws, and the like.

Our argument for the centrality of autonomy to legitimacy begins with the conception of legitimacy outlined earlier. For political authority to be justified, it must be based on either democratic will or normative authority (i.e., where the exercise of political authority is good and there is justified belief that it is good), or a degree of both. As normative authority increases – that is, where policy is good and there is excellent epistemic justification for it – the degree to which it must be supported by mere will decreases.

Both paths to legitimate authority are based on a process of legitimation, each of which is itself grounded in autonomy. We explained in Section 8.2.2 that will-based theories are based on some version of agreement, public reason, or hypothetical consent and that Christman and Oshana explicitly link legitimacy to their conceptions of autonomy. The idea is straightforward. For one’s agreement to a system of governance and to the actions, policies, and decisions of a government to carry force, the agreement must be based on conditions that meet the criteria of autonomy and freedom. One must be epistemically competent, the reasoning upon which one’s agreement is based must satisfy authenticity conditions (as Christman puts it, one must not be understandably alienated from them). One should have the room to determine whether to support political decisions by having a degree of substantive independence. Moreover, people performing their legitimating responsibilities should be free of substantial impediments to the quality of their agency. As we explained in Chapter 5, one’s agency may be impaired by affective, deliberative, and social challenges. Where the quality of their agency is sufficiently challenged in those ways, their ability to fulfill the responsibility of legitimating government policies is limited.

The link between autonomy and legitimation by normative authority is not as intuitively clear as between autonomy and will theories. Normative authority requires, first, that a policy, law, or action by a government be a good one. But the mere fact that a policy is good cannot confer legitimacy because of what Peter calls the “access constraint.” Legitimacy demands that people believe that the policy is good and that they do so for sound reasons. Put another way, the legitimation process requires that people have sufficient information and understanding to assess a policy and believe that it is justified on its merits. The ability to do that is based on autonomy. To form justified beliefs about whether a policy is good, one must be epistemically competent, they must have some sense of value from which they are not alienated, and their reasoning must be substantively independent enough that their ability to reason is not compromised. Finally, the quality of their agency should not be so diminished that it undermines the degree to which their beliefs and actions are indeed their own.

8.4 Technological Threats to Legitimacy

Before we apply our arguments about legitimacy, legitimation, and autonomy to algorithmic systems, let’s summarize the conception of legitimacy that we have developed.

  1. 1. For a government policy, law, or action to be normatively politically legitimate, requires either (a) normative authority, (b) democratic will (within certain bounds, dictated by (a)), or (c) some combination of both (a) and (b).

  2. 2. For a policy, law, or action to be legitimate via normative authority requires that the policy, law, or action in fact be good, and meet the access constraint.

  3. 3. The access constraint functions to ensure that autonomy enters the process by having belief-forming mechanisms in place under sufficient quality of agency.

  4. 4. For a policy, law, or action to be legitimate via democratic will requires some process of voting or public reason.

  5. 5. For a process of voting or public reason to confer legitimacy requires that persons engaged in that process do so autonomously.

With that in mind, let’s return to PredPol, Cambridge Analytica, and the IRA.

8.4.1 PredPol, Normative Authority, and Legitimation

Our analysis of PredPol and legitimation is centered on legitimacy through normative authority. Recall that legitimacy via normative authority involves both normative claims (i.e., that some policy is in fact a good one) and the access constraint. The access constraint requires that the reason some policy is implemented is that people implementing it have epistemically sound reason to believe that the policy is in fact good. Based on these criteria, there are four key questions for determining whether a policy is legitimate based on the normative authority view: (1) Is the policy a good one? (2) Is there ample evidence that the policy is a good one? (3) Does the evidence that the policy is good drive the beliefs about the policy? and (4) Do the beliefs about the policy reflect the actual implementation of the policy?

We will consider the first two questions together, since we cannot address the question of whether the policy is good independent of evidence for it. So, is use of PredPol in fact good policy? Certainly, reduction in crime is a worthwhile objective. Hence, whether use of PredPol is good policy will turn on whether it is effective, whether it has bad consequences that offset any good ones, or whether it impinges other claims. There is some evidence PredPol is effective. In a July 2011 pilot test, the Santa Cruz, CA police department piloted a crime-deterrence program using PredPol. Zach Friend, the department’s crime analyst, reported encouraging results. Burglaries in 2011 were down 27 percent that July, compared with July in the previous year. Speaking to the question of effectiveness, Friend added, “The worst-case scenario is that it doesn’t work and we’re no worse off” (though who the “we” refers to is unclear).Footnote 30

However, the evidence of PredPol’s efficacy is equivocal. Santa Cruz’s drop in burglaries in 2011 is not enough to conclude that PredPol was the cause; after all, cities are complex entities and a one-year drop in one category of crime does not warrant a conclusion about a single cause and tells us nothing about other crimes. There have been only a few academic studies on predictive systems like PredPol. Those that exist yield mixed results, with some showing predictive methods having some crime-reducing effects, but others showing it has no effect at all.Footnote 31 For example, a study by the RAND Corporation testing a place-based system like PredPol found that the predictive system had no statistically significant effect on crime overall.Footnote 32

There is also reason to think that use of PredPol has negative effects that offset whatever advantages in crime reduction it confers. Several cities using predictive policing technology have found that police simply stay in the locations indicated by the algorithm. That is, they stayed in the “red box” that PredPol overlays on a map. Susan Merritt, San Francisco Police Department’s chief information officer, remarked that “[i]n L.A. I heard that many officers were only patrolling the red boxes, not other areas […] People became too focused on the boxes, and they had to come up with a slogan, ‘Think outside the box’.”Footnote 33 And there is evidence that this has deleterious effects. Patrol officers stayed in the “red boxed” areas and engaged in “intelligence gathering,” for example questioning people, running people’s records, investigating for potential (though unreported) narcotics use, and so forth. This, in turn, had the effect of slowing response times to emergencies.Footnote 34

A further consequence that offsets whatever benefits PredPol has is disproportionate use. The Los Angeles Police Department (LAPD) has used predictive methods as a reason to deploy helicopters (as opposed to patrol cars) to the 500 foot by 500 foot boxes that denote “hot spots.” These practices outrun the design of the predictive tool for several reasons. For one, it is difficult to discern unspecified criminal activity from a helicopter. And helicopters’ deterrent effect is fleeting. Criminologist Geoffrey Alpert – a specialist in the use of helicopters in policing – says that helicopters, which are loud and obvious, will at best deter criminals while they are overhead but that criminals will simply return when the helicopters are not around.Footnote 35 Moreover, accommodating helicopter flight paths, the search area had to be expanded well beyond the “red box,” conflicting with the purported specificity of PredPol.Footnote 36 Add to this the annoyance residents experience having loud helicopters overhead, and there are several negative consequences offsetting PredPol’s advantages.

Finally, there are potential rights impingements relevant in determining whether PredPol is good policy. Predictive models indicating a “high crime area” can be a factor in determining whether police have legal authority to stop and frisk a person.Footnote 37 Hence, people may be stopped and frisked more often in a PredPol-designated area, and such stops are more likely to be permissible under the Fourth Amendment. However, it is another question whether merely being in a high-crime area can help morally justify a stop. And because of the source of data, the patterns of high-crime areas are skewed toward neighborhoods in which residents are predominantly from racial and ethnic minorities.

So, even if use of PredPol has some positive effects on a justifiable policy goal, it is not warranted to conclude that it is good policy overall and that there is ample evidence for it being good policy. Let’s assume that the answer to the third question – whether evidence that the policy is good drives beliefs about the policy – is yes. That is not to say that the inference is correct; the access constraint merely requires a connection between evidence and belief on the assumption that the policy is good and justifiable.

The legitimation problem for use of PredPol comes in steps (3) and (4). It is possible that what evidence there is that PredPol can help reduce some types of crime grounds municipalities’ use of the technology. However, the actual implementation of PredPol does not appear to reflect the evidence of its efficacy. Even if PredPol is well suited for predicting some kinds of crimes, the evidence for that does not warrant belief about PredPol’s use as a general law enforcement tool, much less as a reason to park police officers in red-boxed areas or to act as general investigators (and slow responses to emergencies). It likewise does not support belief that deploying helicopters to fly over red-boxed and surrounding areas is warranted. It also does not provide grounds for thinking that individual persons who happen to be in red-boxed areas are themselves engaged in criminal activity. To reiterate, the access constraint is the requirement that belief in a policy’s justification be based on the fact (if it is a fact) that the policy is a good one. Here, the fact that PredPol is effective in one area instead grounds the apparent belief that it is a good tool for intensive policing of discrete areas and conducting stops of individuals. In this way, the evidence for PredPol’s efficacy in one area actually circumvents an important legitimating process. The evidence for its efficacy in one area stands in for evidence of efficacy for the tasks in which it is actually used. This is a barrier for citizens in being able to form beliefs about whether a policy is a good one. In other words, deploying PredPol well outside the range of circumstances for which there is evidence of its efficacy undermines citizens’ abilities to fulfill their responsibility of exercising their autonomy in legitimating policy.

Here we want to be very clear about the limitations of this discussion. The mere fact that a policy is not good does not render it illegitimate. Any kind of policymaking (broadly construed) is complex, takes place under epistemic limitations, and will affect lots of people with considerable interests in the policy. Moreover, for a policy to be “good” in the strong sense necessary for it to be legitimate via normative authority is a very tall order. And the fact that use of PredPol does not seem to clear that threshold is therefore unsurprising. Use of PredPol might instead be legitimate based on the democratic, will-based view. Hence, our argument is emphatically not an attempt to make an all-things-considered judgment that using PredPol is not democratically legitimate (perhaps it is, perhaps it isn’t). Rather, as we emphasized earlier, our purpose here is to examine autonomy and the process of legitimation. And on that question, PredPol is an informative case because it precludes citizens’ abilities to legitimate predictive policing by undercutting their ability to assess its policy rationale.

8.4.2 Cambridge Analytica, the IRA, and Will-Based Authority

What are often referred to as the “Cambridge Analytica scandal” and “Russian hacking case” are in fact an interconnected, overlapping, and confusing welter of activities, organizations, tools, and technologies that grew up before and became widely known in the wake of the 2016 UK Brexit campaign and the 2016 US presidential election. The events are well documented in that there have been multiple tell-all books,Footnote 38 government investigations,Footnote 39 academic reports,Footnote 40 and news stories.Footnote 41 However, the interwoven stories are obscured by their very complexity, redactions in official reports and indictments, disinformation and deflection campaigns, and failures of public institutions to follow through with further investigations. Given the scope of the stories, we cannot possibly do them justice here. However, we can pull out a few important threads to ground our arguments about legitimation.

Cambridge Analytica is a data analytics and political consultancy firm based in London, UK. It is an offshoot of the SCL Group. In the years before the Brexit campaign and the 2016 US presidential election, the firm became interested in using the gigantic, rich datasets generated on social media platforms – particularly Facebook – to build ever more sophisticated models of voters and behavior. In order to pursue this work, they partnered with Cambridge University psychologist Aleksandr Kogan, who was already collecting social media information for his own research. Kogan had developed (along with other researchers) a personality app called “This Is Your Digital Life.” The app, which included a “lengthy psychological questionnaire,”Footnote 42 was actually only downloaded by around 270,000 users, but it ultimately harvested user data from those users and from their friends, a group that might number up to 87 million.Footnote 43 This data was shared with Cambridge Analytica. On the basis of psychological profiles and data about Facebook users, Cambridge Analytica offered micro-targeted advertising based on their data-driven psychometrics. They did this first for Republican Party nomination seeker Ted Cruz.Footnote 44 When Cruz dropped out of the nomination race, they shifted their operation to candidate Donald Trump.

The targeted advertising and social media campaign spearheaded by Cambridge Analytica is of particular note. Christopher Wylie is a data scientist and former employee of Cambridge Analytica who eventually became a whistleblower and author of a book about the company. He describes several key features of the company’s strategies. For example, he describes a project targeting African American voters with messages designed to decrease their motivation for voting. He also describes using the techniques of using its influence to create in-person meetings of angry citizens.Footnote 45 As Wylie put things, in the original report published by The Guardian: “We exploited Facebook to harvest millions of people’s profiles. And built models to exploit what we knew about them and target their inner demons. That was the basis the entire company was built on.”Footnote 46

The Russian Internet Research Agency (IRA) is a Russian state-supported organization that was established in 2013. It operates “like a sophisticated marketing agency,” has trained “over a thousand people to engage in round-the-clock influence operations,” and has targeted citizens in a variety of places, including Russia, Ukraine, and the United States.Footnote 47 In February 2018, the U.S. Department of Justice indicted the IRA and a number of Russian nationals who worked with the IRA for conspiracy to commit fraud, wire fraud, and bank fraud.Footnote 48

The charges were based in part on disinformation “active measures” carried out on social media.Footnote 49 These measures included creation of fictitious groups that engaged social media users with inflammatory and divisive content. The IRA created fake profiles that followed other IRA sites and groups, thereby increasing their footprint and visibility to real people on social media platforms. To reach larger audiences, the IRA purchased Facebook ads that were then placed in potential followers’ Facebook newsfeeds.Footnote 50 The inflammatory, divisive content promoted by the IRA sought to engage with social media from a range of political affiliations. The U.S. Senate Select Committee on Intelligence found that the IRA’s targeting was particularly intensive toward African Americans. It writes, “By far, race and related issues were the preferred target of the information warfare campaign designed to divide the country in 2016.”Footnote 51 It did this by creating pages and groups that sought to engage users that the social media platform identified as African American and by targeting geographic areas that are predominantly African American. These include pages such as “Blacktivist” and social media posts related to police shootings and NFL player kneeling protests of police shootings.Footnote 52 Other groups the IRA targeted with incendiary messaging include groups with names that imply a range of political dispositions (“Being Patriotic,” “Stop All Immigrants,” “Secured Borders,” “Tea Party News,” “Black Matters,” “Don’t Shoot Us,” “LGBT United,” and “United Muslims of America”).Footnote 53 The IRA’s social media accounts “reached tens of millions of U.S. persons” and had “hundreds of thousands of followers.”Footnote 54

These accounts were used to organize in-person rallies, recruit activists to perform political tasks, and promote IRA content.Footnote 55 The IRA’s activity is linked to social media users being exposed to a great deal of misinformation. A study of Twitter by the Oxford Internet Institute found that in the run-up to the 2016 presidential election, “users got more misinformation; polarizing, and conspiratorial content than professionally produced news.”Footnote 56

The breadth of activities by the IRA is striking. Its activities range across social media platforms, including Facebook, YouTube, Instagram, and Twitter. It targets groups from a range of social and political perspectives. Its tactics include aiming to suppress votes, aiming to direct voters toward third parties, aiming to depress turnout, encouraging secessionist movements (including in Texas and in California, comparing them to Brexit), and denigrating a range of leaders across parties.Footnote 57 They also aimed both to seed news media with content generated by Russian disinformation campaigns and to undermine trust in traditional news media.Footnote 58

There is no official recognition or definitive documentation connecting Cambridge Analytica and Russian active measures. In his book, Wylie recounts Cambridge Analytica’s involvement with Russian actors, message-testing about Vladimir Putin, close connections to people involved with pro-Russia factions in Ukraine politics,Footnote 59 and involvement with Russian oil firm Lukoil.Footnote 60 SCL’s promotional materials indicate that they did work for Russia. However, SCL denies being involved in political campaigns at the behest of Russia, insisting that their work in Russia was for “private” interests.Footnote 61 It is worth noting here that the distinction between state and nominally private interests in kleptocracy is blurry.

So the existence and nature of the connection between the IRA and Cambridge Analytica are unclear. However, what matters here is not whether there was a single group of people acting in concert to sow anger and seed disinformation in the United States and elsewhere. Rather, our focus here is on the nexus of intrusive data collection (itself possible based on Facebook’s weak privacy protections and permissive terms of service (themselves difficult to comprehend and rarely read), and the imprimatur of Cambridge University), predictive analytics using that data to better target influence, and algorithmic systems that suggest advertising to clients (see Section 7.4) and promote content to users. That is, our concern here is a socio-technical, big-data, and algorithmically aided group of systems that affect the attitudes, beliefs, dispositions, and actions of people within democratic states. Whether Cambridge Analytica entered into agreements or acted in explicit concert with the IRA is neither here nor there for our purposes in this chapter.

We should also emphasize that the effect of Cambridge Analytica’s and the IRA’s actions on the outcomes of particular elections is unclear, and no single set of events could ever be said to be the sole cause of complex phenomena like election outcomes and broad policy-making. It is disputed just how well these kinds of tactics actually work. Kogan argued that the “accuracy” of the data was exaggerated in media reports and that “[w]hat Cambridge has tried to sell is magic.”Footnote 62 At first, the Cruz campaign credited the data-driven approach with its win in the 2016 Iowa caucus,Footnote 63 but NPR reports that, as the election cycle moved forward, the campaign grew more skeptical and eventually phased out the psychological profiling after later losing the South Carolina primary.Footnote 64

Nonetheless, the tactics are important from the standpoint of political legitimacy. As a team of psychologists have recently shown, there is an increasing amount of evidence for “the effectiveness of psychological targeting in the context of real-life digital mass persuasion,” and “tailoring persuasive appeals to the psychological profiles of large groups of people allowed us to influence their actual behaviors and choices.”Footnote 65 The Cambridge Analytica case, they comment, “illustrates clearly how psychological mass persuasion could be abused to manipulate people to behave in ways that are neither in their best interest nor in the best interest of society.”Footnote 66

This creates a problem of legitimacy. Whereas in the PredPol case we consider the normative authority path to legitimacy, here our concern is the will or democratic path. Recall that this path relies on persons’ agreement (whether tacit, hypothetical, public reason, or voting processes) to be governed and to the actions, policies, and decisions of a government, and for citizens’ agreement to carry force, it must be based on conditions that meet the criteria of autonomy. And, again, our concern here is not an all-things-considered judgment about legitimacy, but about the ability of citizens to fulfill their responsibilities of legitimation.

There are several ways in which the exploitation of data, psychological profiles, social media, and algorithmic systems creates a drag on legitimation. Consider first Christman’s test for legitimacy, which requires that political principles be such that reasonable citizens would not be understandably alienated from them.Footnote 67 The fake groups that the IRA created and advertised were promoted by fake user profiles but eventually were followed by real people, referenced real issues, and mimicked names and language of genuine social, religious, and political groups (e.g., Stop All Immigrants, Blacktivist, LGBTQ United). The language, the titles, and linking to real issues and groups drew in people who came to express their views and wills on topics with real stakes. However, by joining, engaging with, and thereby amplifying those groups and messages (and in some cases populating in-person eventsFootnote 68), it is likely that their wills were not autonomous because the expression is one from which people would be alienated. That is, if people were to “engage in sustained critical reflection,” which requires reflecting on how they came to their commitment, the expressions would be incompatible with their sense of self and practical identity over time.Footnote 69 And, if a person earnestly engaged in political debates about policing, rights of LGBTQ+ persons, and immigration were to reflect on the fact that Facebook groups like Stop All Immigrants, Blacktivist, and LGBTQ United were ersatz communities built by Russian agents to exploit and enrage them, they would in all likelihood be alienated from their support. In other words, the social media campaigns undermine the legitimation process by undermining the authenticity condition for autonomy.

Consider next the ways in which psychological profiling and targeting on social media works. It works, first, to engage people by keeping them scrolling on social media feeds.Footnote 70 It also seeks to create sustained engagement. One mechanism for creating such engagement is provoking strong emotions – including affinity for a cause and anger at opponents. The tactics draw people in and exploit their emotions. These tactics, for reasons we described in Chapter 5.2, are instances of the affective, deliberative, and social challenges to agency. They thus undermine the quality of people’s agency and thereby conflict with their autonomy. This diminishes people’s ability to perform their legitimating responsibilities.

The fake accounts also promoted falsehoods. First, they promoted falsehoods about the nature of the groups themselves. But they also promulgated false claims. This circumvents persons’ autonomy by undermining its epistemic requirements. And it creates yet another drag on the legitimation process.

Finally, there is an issue that goes back to Feinberg’s understanding of “ideal” autonomy and to social/relational conceptions of autonomy.Footnote 71 People are not isolated, atomistic individuals. They are parts of families, communities, and social groups. Such groups are vital to autonomy both in the sense that they are important in causing people to flourish and to develop their autonomy and in the sense that autonomy involves the ability to act and realize one’s values within communities. The tactics of Cambridge Analytica and the IRA undermine that. They specifically seek to create ersatz communities that displace genuine ones and to rend broader communities into factions that neither trust nor communicate effectively with each other. Deep disputes are an important and ever-present facet of democratic polities. However, encouraging disparate groups to further distrust others and undermining the ability of those groups to communicate conflicts with the social facet of autonomy and, hence, with the legitimation process.

8.5 Once More Past the Pole

The issue of legitimation in the cases of PredPol, Cambridge Analytica, and the IRA is momentous. How police exercise their power to put others in harm’s way, efforts to affect an election via social profiling, and interference of a hostile nation in a US presidential election cut to the quick of democratic legitimacy. But issues of political authority and legitimation are not limited to matters at the heart of democratic procedures and at the far reaches of state-sanctioned exercise of power. They extend also to ordinary administrative tasks as well, among them risk assessment and teacher evaluation.

COMPAS raises similar issues to PredPol on the normative authority arm of political legitimacy. It is no doubt true that efficient allocation of scarce resources for supervising people in the criminal justice system is good policy. There are substantial open questions about that system overall, of course. But if we accept the premises that the state should sanction some kinds of actions with a range of penalties and that supervision and resources (e.g., drug and alcohol treatment, job programs, counseling, housing) are appropriate responses, assessing persons for placement within those programs would appear to be a reasonable approach. But, as we explained in Section 3.4.3, the use for which COMPAS is designed and suited is different than the use to which it is put in Loomis. Hence, even if there is evidence that COMPAS is well calibrated in assessing risk, that same evidence does not underwrite use for sentencing purposes. As with PredPol, there is a disconnect between evidence and tool use, and that creates a drag on the epistemic facet of legitimation via normative authority.

The VAM cases are a bit simpler in that the tool itself has substantial flaws, as we discussed in Sections 1.1 and 3.4. Evaluating teachers is a reasonable policy goal. However, the fact that the tools to do so have such big flaws precludes use of EVAAS and TVAAS from legitimation via normative authority. There is not a similar impediment to legitimation via democratic will. Note two things, though. First, that some institution, government, law, or policy is legitimate – which is to say based on a justified exercise of political power – is not the same as it being morally justified. Political legitimacy is in that sense a lower bar. Second, the fact that the algorithmic systems in these cases do not place a drag on the legitimation process is not a sufficient condition to ensure legitimacy. As we made clear in Section 8.4, our concern is that some kinds of algorithmic systems hinder citizens’ abilities to exercise their responsibility to legitimate policies. VAMs do not obviously hinder fulfilling that responsibility.

8.6 Conclusion

Autonomy is foundational for political legitimacy. It grounds each of the accounts we canvassed in Section 8.2, including the hybrid account we endorse. We argued that one component part of legitimacy is the process of legitimation. That is, citizens have a responsibility to exercise their autonomy to legitimate governments, policies, laws, and actions. Algorithmic systems can impede citizens from fulfilling those responsibilities. PredPol (and to an extent COMPAS) is an example of an impediment to legitimation via normative authority. Whatever evidence there is of PredPol advancing a good policy goal (and there is some), that deflects from assessing how PredPol is actually used. That conflicts with the access condition, whereby belief in a policy’s value has to be linked to evidence for its actual value.

The cases of Cambridge Analytica and the IRA undermine the second arm of political legitimacy. They create conditions where citizens would be alienated from the source of their beliefs and attitudes; they substitute ersatz involvement for genuine involvement and generate false beliefs.

But we should reiterate here that political legitimacy goes much further than the process of legitimation. For one, there are questions about different levels or targets of legitimacy. The fact, if it is, that policies using particular technologies create a drag on legitimation or, worse, are not legitimate all-things-considered does not tell us whether the agencies using the technologies, the municipalities, states, or nations using those technologies are legitimate or much else. It is important not to overinterpret drags on legitimation to make broader conclusions and legitimacy per se. Related is that legitimate authority can get things wrong and still retain legitimacy to exercise authority. That breadth of legitimacy is why our conclusions here are limited to the ability of citizens to exercise their responsibilities.

Footnotes

1 The analogies advocates use to explain PredPol are varied and tend to liken crime to phenomena outside of complex, modern society. One analogy is with earthquakes. After an earthquake, it is likely that there will be another one in the same area, an aftershock. Similarly for crime on this theory, certain “place-based” crimes (such as a burglary) are followed by crimes in the same area. See Goode, “Sending the Police before There’s a Crime.” Another analogy is to hunter-gatherers, in part because the PredPol grew out of Brantingham’s work using computer models to understand hunter-gatherers. Brantingham states that “[c]riminals are effectively foragers … [c]hoosing what car to steal is like choosing which animal to hunt. The same decision-making processes go into both of these choices.” See Hoff, “Professor Helps Develop Predictive Policing by Using Trends to Predict, Prevent Crimes.” We are not convinced that these analogies are helpful beyond the fact that some kinds of crime exhibit geographic patterns.

2 Note that in this chapter we will use “predictive policing” to refer to “place-based” predictive policing, which focuses on forecasting where crimes will occur. This is often contrasted with “person-based” predictive policing. Person-based predictive policing raises similar concerns, but we will not focus on it here. For an excellent overview of these issues, see Ferguson, The Rise of Big Data Policing, chapter 3.

3 Grossman et al., “The 50 Best Inventions.”

4 Ferguson, The Rise of Big Data Policing, 65–67.

5 Edwards, “Predictive Policing Software Is More Accurate at Predicting Policing than Predicting Crime”; Ferguson, The Rise of Big Data Policing, 73–74.

6 Shklar, “The Liberalism of Fear.”

7 Schwartz, “After Qualified Immunity.”

8 For the purposes of this chapter we won’t draw sharp distinctions between government actions, decisions, policies, and laws. While they can be distinguished (both legally and philosophically) and different justificatory burdens may be appropriate to each, our concern here is about legitimacy and legitimation at a general level that is applicable to each.

9 Weber, Economy and Society.

10 This is a question of internal legitimacy or legitimacy within a state. There may be questions as to whether a state can justify itself internationally. See Peter, Democratic Legitimacy, chapter 1.

11 Matheson, “Weber and the Classification of Forms of Legitimacy.”

12 Waldman, “Power, Process, and Automated Decision-Making,” 614.

13 Waldman, 616.

14 For a helpful and recent overview of the topic of the grounds of political legitimacy, see Peter, “The Grounds of Political Legitimacy.”

15 Christman, The Politics of Persons: Individual Autonomy and Socio-Historical Selves, 239 (emphasis in original).

16 Oshana, Personal Autonomy in Society, 97–100.

17 Rawls, Political Liberalism, 137.

18 Pettit, Just Freedom, chapter 5.

19 Rawls, “Justice as Fairness,” 242.

20 Pettit, Just Freedom, 112.

21 Peter, “The Grounds of Political Legitimacy,” 377.

22 Raz, The Morality of Freedom, 53.

23 Enoch, “Authority and Reason-Giving,” 31.

24 Raz, The Morality of Freedom, 422.

25 Peter, “The Grounds of Political Legitimacy.”

26 For a longer discussion of this problem, see Peter, “The Epistemic Circumstances of Democracy.”

27 Estlund, Democratic Authority: A Philosophical Framework.

28 Note that “legitimation” in this normative sense is different from, though related to, the descriptive sense of legitimation in descriptive accounts of legitimacy. The descriptive sense of legitimation refers to the processes by which a governing entity establishes and maintains power, regardless of whether those processes justify that power. Descriptive legitimation may be necessary for normatively legitimate policies to function.

29 To be clear, it is neither possible nor desirable that every member of a polity understand every kind of government institution, law, and action. Rather, there are some broad states of affairs that most any member of society has a responsibility to understand and support (or not): some conception of fair representation and enfranchisement, basic human rights. Others will have responsibilities based on their communities, experiences, and expertise. The responsibility is to exercise autonomy (properly understood as social and relational) in some facets of social life.

30 Goode, “Sending the Police before There’s a Crime.”

31 Ferguson, The Rise of Big Data Policing, 72; Meijer and Wessels, “Predictive Policing: Review of Benefits and Drawbacks.”

32 Hunt, Saunders, and Hollywood, “Evaluation of the Shreveport Predictive Policing Experiment.”

33 Bond-Graham, “All Tomorrow’s Crimes”; Ferguson, The Rise of Big Data Policing, 79.

34 Ferguson, The Rise of Big Data Policing, 79; Hunt, Saunders, and Hollywood, “Evaluation of the Shreveport Predictive Policing Experiment.”

35 Mather and Winton, “LAPD Uses Its Helicopters to Stop Crimes before They Start.”

36 Ferguson, The Rise of Big Data Policing, 81; Mather and Winton, “LAPD Uses Its Helicopters to Stop Crimes before They Start.”

37 Ferguson, The Rise of Big Data Policing. For example, in Illinois v. Wardlow, 528 U.S. 119, the Supreme Court determined that simply being in high-crime area did not suffice to justify a Terry stop, but it can be relevant in determining whether other facts are sufficiently suspicious for a stop.

38 Wylie, Mindf*ck: Cambridge Analytica and the Plot to Break America; Kaiser, Targeted: The Cambridge Analytica Whistleblower’s Inside Story of How Big Data, Trump, and Facebook Broke Democracy and How It Can Happen Again.

39 U.S. Department of Justice, “Report on the Investigation into Russian Interference in the 2016 Presidential Election, Volume I (‘Mueller Report’)”; Select Committee on Intelligence, United States Senate, “Report of the Select Committee on Intelligence, United States Senate, on Russian Active Measures Campaigns and Interference in the 2016 U.S. Election, Volume I: Russian Efforts against Election Infrastructure with Additional Views.”

40 Howard et al., “Social Media, News and Political Information during the US Election: Was Polarizing Content Concentrated in Swing States?”; DiResta et al., “The Tactics & Tropes of the Internet Research Agency.”

41 Cadwalladr and Graham-Harrison, “Revealed: 50 Million Facebook Profiles Harvested for Cambridge Analytica in Major Data Breach”; Kang and Frenkel, “Facebook Says Cambridge Analytica Harvested Data of up to 87 Million Users.”

42 Confessore, “Cambridge Analytica and Facebook.”

43 Kang and Frenkel, “Facebook Says Cambridge Analytica Harvested Data of up to 87 Million Users.”

44 Davies, “Ted Cruz Campaign Using Firm That Harvested Data on Millions of Unwitting Facebook Users.”

45 Wylie, Mindf*ck: Cambridge Analytica and the Plot to Break America, 127–129.

46 Cadwalladr and Graham-Harrison, “Revealed: 50 Million Facebook Profiles Harvested for Cambridge Analytica in Major Data Breach.”

47 DiResta et al., “The Tactics & Tropes of the Internet Research Agency,” 6.

48 U.S. v. Internet Research Agency, LLC (Indictment, February 16, 2018).

49 U.S. v. Internet Research Agency, LLC (Indictment, February 16, 2018).

50 U.S. Department of Justice, “Report on the Investigation into Russian Interference in the 2016 Presidential Election, Volume I (‘Mueller Report’),” 25.

51 Select Committee on Intelligence, United States Senate, “Report of the Select Committee on Intelligence, United States Senate, on Russian Active Measures Campaigns and Interference in the 2016 U.S. Election, Volume II: Russia’s Use of Social Media and Additional Views,” 6.

52 Select Committee on Intelligence, United States Senate, 6–7.

53 U.S. Department of Justice, “Report on the Investigation into Russian Interference in the 2016 Presidential Election, Volume I (‘Mueller Report’),” 24–25.

54 U.S. Department of Justice, 26.

55 U.S. Department of Justice, 31–32.

56 Howard et al., “Social Media, News and Political Information during the US Election: Was Polarizing Content Concentrated in Swing States?” 1.

57 DiResta et al., “The Tactics & Tropes of the Internet Research Agency,” 8–10.

58 DiResta et al., 65–66.

59 Wylie, Mindf*ck: Cambridge Analytica and the Plot to Break America, 142.

60 Wylie, 141, 155–156.

61 Hakim and Rosenberg, “Data Firm Tied to Trump Campaign Talked Business with Russians”; Wylie, Mindf*ck: Cambridge Analytica and the Plot to Break America, 139–159.

62 Weaver, “Facebook Scandal: I Am Being Used as Scapegoat – Academic Who Mined Data.”

63 Hamburger, “Cruz Campaign Credits Psychological Data and Analytics for Its Rising Success.”

64 Detrow, “What Did Cambridge Analytica Do during the 2016 Election?”

65 Matz et al., “Psychological Targeting as an Effective Approach to Digital Mass Persuasion,” 12717.

66 Matz et al., 12717.

67 Christman, The Politics of Persons: Individual Autonomy and Socio-Historical Selves, 239 (emphasis in original).

68 Wylie recounts that in summer 2014, Cambridge Analytica began creating fake Facebook and other social media pages and groups with politically charged identities that were tailored to be fed into the news feeds of susceptible users by recommendation algorithms. As people joined the groups, Cambridge Analytica would set up meetings in small spaces (in order to make the group feel big). Because of the nature of the groups, and the nature of the content fed into the groups, “[p]eople would show up and find a fellowship of anger and paranoia.” Wylie, Mindf*ck: Cambridge Analytica and the Plot to Break America, 128. These groups would, when large enough and prompted by Cambridge Analytica, meld with similar groups, creating a network of engaged, angry people who felt they were under siege. Wylie, 127–129.

69 Christman, The Politics of Persons: Individual Autonomy and Socio-Historical Selves, 155.

70 Wylie, Mindf*ck: Cambridge Analytica and the Plot to Break America, 127; Eyal, Hooked: How to Build Habit-Forming Products.

71 See Sections 2.2.2, 2.6, and 3.3.

Save book to Kindle

To save this book to your Kindle, first ensure [email protected] is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

Available formats
×

Save book to Dropbox

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.

Available formats
×

Save book to Google Drive

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.

Available formats
×