Introduction
Philosophy, perhaps in particular moral philosophy, differs from other academic disciplines in its profuse and unabashed use of examples that are empirically implausible, sometimes even physically impossible. As Philip Abbot noted:
“But what examples! The world of the philosopher is filled with people spores, child missile launchers, Martians, talking robots, talking dogs, kittens, chimps, jig [s]aw cells that form human beings, transparent wombs, and cool hands – everything in fact but fetuses growing in wombs and infants cradled in parents’ arms… Philosophers have moved into the world of fantasy in the same way and with the same verve that social scientists moved into the world of quantifiable facts… It is very questionable whether the moral dimensions of our lives can be clarified in circumstances in which the very basis for morality is no longer present.”Footnote 1
This is a tendency that seems to have escalated in recent decades. As several authors have noted, it appears to be particularly strong in neuroethics.Footnote 2 However, it has also repeatedly been subject to critical comments.Footnote 3 It is the purpose of this article to clarify what types of criteria we should apply in distinguishing between irrelevant speculative scenarios and useful thought experiments. In doing so, we need to distinguish between the two major areas of ethical discourse, namely, on the one hand, foundational moral philosophy and on the other, applied ethical disciplines such as neuroethics.
The word “speculate” entered the English language in the late 16th century as a synonym of “contemplate” or “theorize.” Today, it is usually employed in a derogatory sense about thoughts or discussions “of a conjectural or theoretical nature.”Footnote 4 In discussions on philosophical style, it refers to claims or arguments that are based on implausible or unrealistic assumptions. Perhaps a less pejorative terminology, referring to “unrealistic,” “fictional,” or “imaginative,” rather than “speculative” thoughts and examples, would have been preferable, but no attempt at terminological reform will be made here.
We will now turn to an analysis of the uses of empirically unsubstantiated speculation in foundational moral philosophy. It concludes with recommendations on the sound use of such speculation. After that, the use of speculative examples in applied ethics will be briefly introduced, followed by a critical analysis of the major argumentative patterns that have been offered in defense of such examples. Finally, this will all be brought to bear on current developments in neuroethics, which is one of the areas of applied ethics that are most affected by speculative examples and scenarios.
Foundational Moral Philosophy
General or foundational moral philosophy, as it is taught and researched primarily in departments of philosophy, is usually devoted to discussions on ethical theories. By an ethical theory (in a wide sense), we mean a general approach to moral assessment that is intended to serve as a framework guiding moral decisions in the various situations that moral agents may encounter.
There is an obvious, and basically sound, answer to the question what types of examples or scenarios are relevant for the evaluation or development of a theory: an example is relevant if and only if it concerns events or situations falling within the scope that the theory is intended to cover. We can call this the scope criterion. It applies to all kinds of theories: empirical, metaphysical, normative, evaluative, etc. theories. For instance, a theory about the flight of birds cannot be confirmed or disconfirmed by investigations of flight patterns in butterflies (but, of course, such investigations are relevant for possible extensions of the theory to cover these creatures as well). For the same obvious reasons, an ethical example concerning potential beings that differ from us humans in ways that are crucial for the example cannot be used to confirm or disconfirm an ethical theory whose scope is limited to human beings.
Although the scope criterion is simple and arguably trivial, its application to ethical theories is often difficult. Instead, it is common to implicitly assume that moral theories are limitless in the sense of applying to all kinds of creatures and situations that we can at all imagine. Consequently, there is no principled way to exclude the use even of examples that are so far-fetched that it is difficult to see how we can at all have intuitions about them. This limitless approach to the scope of moral theories is closely connected with two other, important but seldom discussed, assumptions.
The first and most fundamental of these assumptions is that moral philosophy deals with moral principles and theories that hold not only in human societies as we know them, but also under virtually all imaginable conditions where intelligent entities make choices. Moral theories such as utilitarianism are commonly taken to have such a wide scope of application, but this is a grand ambition rather than a reasoned conclusion, and it can certainly be questioned.Footnote 5 With a more restricted scope for our moral principles and theories, many of the speculative examples will become irrelevant. For instance, if our notion of personal identity is supposed to cover all conceivable conscious entities, then it can be swayed by examples involving uploaded, duplicated, and teletransported minds. But if we restrict its scope to creatures of the types known to us, then such counterexamples will have no bearing on the concept.
The second assumption is that our methods for analyzing moral examples are sufficiently reliable also when applied to this kind of examples. Since the analysis of moral examples ultimately relies on our moral intuitions, this amounts to an assumption that these intuitions are sufficiently reliable also when exposed to bizarre and unheard of examples. This is questionable, not least since our moral concepts have been shaped to deal with the types of moral choices and standpoints that we may encounter in real life. W.V. Quine said:
“To seek what is ‘logically required’ for sameness of person under unprecedented circumstances is to suggest that words have some logical force beyond what our past needs have invested them with.”Footnote 6
Similarly, Ludwig Wittgenstein warned: “If you imagine certain facts otherwise, describe them otherwise, than the way they are, then you can no longer imagine the application of certain concepts, because the rules for their application have no analogue in the new circumstances.”Footnote 7
There is of course nothing wrong with conducting investigations based on these two assumptions, but it is problematic that they are almost universally taken implicitly for granted, rather than being carefully articulated and argued for. It is far from self-evident that we can learn what is morally right and wrong in real life by reflecting on what would be right to do under quite different conditions that can never obtain. The following four tentative rules are proposed as a starting-point for a more active discussion on the scope of moral theories:
-
1) It is a respectable intellectual activity to explore the implications of various moral theories under the assumption that these theories as well as our moral intuitions cover all imaginable counterfactual circumstances.
-
2) It is also a respectable intellectual activity to explore the implications of such moral theories under various restrictions of their scope, not least their restriction to agents who are human (in the common sense of that word) and who act under circumstances that we do not have strong reasons to consider impossible.
-
3) The advantages of a broad scope for moral theorizing have to be weighed against the potential disadvantages. These include the risk that features incorporated in a moral theory to make it more plausible in unrealistic examples will also make it less plausible in real life.
-
4) The scope of a moral principle or theory needs to be explicitly stated and should preferably also be explicitly justified.
Applied Ethics
Applied ethics originated largely as the ethics of professions. The ethics of the medical profession goes back to antiquity. Engineering ethics has been discussed in professional societies since the 19th century. Research ethics also has a long tradition, but it did not gain much momentum until after World War II, in response to the nuclear bomb and atrocious experiments in Nazi concentration camps. In the 1970s, several other branches of applied ethics got off the ground, including business ethics, agricultural ethics, and computer ethics. Still today, most of applied ethics concerns the ethical responsibilities of a relatively small number of professions. Many other social areas with important ethical issues still lack a developed ethical discourse. This applies for instance to traffic safety, welfare provision, insurance, building and architecture, and foreign aid.Footnote 8
Beginning around 1970, moral philosophers became increasingly engaged in applied ethics. The largest influx of philosophers took place in medical ethics, which has also been more affected by philosophical theory than most other areas of applied ethics. Philosophers working in this area often prefer the term “bioethics” (coined by Van Rensselaer Potter II in 1970) to the older term “medical ethics.”Footnote 9 Philosophers brought with them a more extensive use of moral theory. Some of them also introduced issues and examples that healthcare professionals have reacted against as too remote from actual clinical practices, such as whether a physician should cause a patient to die in order to acquire organs for transplantations that would save two other patients.Footnote 10
In the early 1990s, considerable resources for applied ethics were created through the research program Ethical Legal and Social Implications (ELSI) that was part of the Human Genome Project (HGP). ELSI was tasked to “anticipate the social consequences of the project’s research and to develop policies to guide the use of the knowledge it produces.”Footnote 11 Apparently, many philosophers interpreted this as an opening for speculative and futuristic scenarios. As observed by Ari Schick:
“ELSI effectively inaugurated speculative bioethics in the regulatory mode… attempting at once to be both future-oriented and practical; not merely exploring in general terms the possibilities generated by future technologies or treating them as material for philosophical thought experiments, but pre-empting them and offering an image of the future in which these technologies would produce benefits and few harms through the implementation of proper guidelines and oversight.”Footnote 12
Similar funding initiatives in other areas, not least nanotechnology, provided additional opportunities to apply ethical theories to futuristic scenarios such as the use of nanotechnological devices to upload a brain on a computer. In this way, “a type of immortality could be achieved,” and “such humans could travel at the speed of light and communicate directly from mind to mind.”Footnote 13 Speculative ethics has also been developed by philosophers working in other areas closely connected with human biology, such as cryonics, human enhancement, and transhumanism.Footnote 14 In these areas, speculative examples are routinely offered as arguments on healthcare ethics and public policy, in contradistinction to the restrictedly theoretical use of similarly unrealistic examples in foundational moral theory, as discussed in the previous section.
Mere Possibility Arguments
A common argument for the use of unrealistic examples in applied ethics is that they may not be that unrealistic after all. Then their investigation can be subsumed under the common, and quite sound, argument that we need to investigate future technological developments before they actually take place.
“[C]ritics may believe that it is inappropriate or premature to consider such issues now. But we do not need to resolve that question here in order to take seriously the ethical and social issues advanced nanotechnology might raise. Even if advanced nanotechnology is a remote possibility, its scenarios appear so disruptive that they merit consideration… [I]f a political course had even a bare possibility to leading to a devastating war, costing the lives of millions, it seems that we are morally obligated to seriously consider that possibility, no matter how remote… [I]f history is any guide, most of our mid- and long-term predictions about technology will be overly optimistic or pessimistic. Many things we have today were once believed to be impossible or impractical—such as gas streetlights, residential electricity, telephones, highways, radio, airplanes, rockets, and even today’s ubiquitous personal computer—so perhaps the prudent course is to treat most of these possibilities as reasonable until proven otherwise.”Footnote 15
“Some highly unlikely future events nevertheless have consequences that are so bad that the course of action with the best net benefits involves taking measures to prepare for such events. For example, the extremely slim chance that the Large Hadron Collider—the world’s biggest particle collider, currently under construction—will result in the extremely bad consequence of destroying the earth has led to a lawsuit to halt its creation.”Footnote 16
The arguments referred to here are mere possibility arguments. By this is meant arguments “in which a conclusion is drawn from the mere possibility that the choice of an option, behavior, or course of action may lead to, or be followed by, certain consequences.”Footnote 17 There are two major variants of mere possibility arguments:
A negative mere possibility argument:
A can lead to B.
B should not be realized.
Thus, A should not be realized.
A positive mere possibility argument:
A can lead to B.
B should be realized.
Thus, A should be realized.
There are situations when a mere possibility argument forms an adequate basis for a decision. For instance, suppose that a member of a group of visitors to a weapons factory takes up a just finished pistol and puts it against his head, indicating that he will pull the trigger, just for the fun of it. Another member of the group says: “Don’t! You can’t know, it may be loaded.” This is a mere possibility argument, which it seems reasonable to act upon.Footnote 18 However, in other situations reliance on mere possibility arguments may lead to injudicious decisions.
A major reason for this is that in almost any decision, an unlimited number of mere possibility arguments can be constructed. Due to the ramified and often chaotic nature of causation, almost any action can possibly lead to extremely positive or extremely negative consequences. For instance, any new medical treatment can have unpredicted lethal side effects. Any new technological device can turn out to be useful for terrorists in unforeseen ways. And almost any action that we take can conceivably give rise to social conflicts that escalate, go out of control, and ultimately end up in war. Obviously, we cannot take all remote possibilities seriously. The mere possibility arguments that are referred to in ethical and political discussions are but a small selection of the arguments that can be constructed. It is not advisable to act upon the mere possibility arguments that have been put forward, without first considering whether other mere possibility arguments can be made. Perhaps there are even more plausible arguments pointing in the opposite direction.
The risk of a biased selection of mere possibility arguments is particularly great if resourceful or passionate stakeholders are mobilized on one side of the issue but not on the other. This can result in an unbalanced discussion, based on the false impression that all mere possibility arguments point in one and the same direction.
Unfortunately, once fears or hopes have been set in motion in this way, they are not easily extinguished, even if they are later shown to be utterly unfounded. One example of this is the debate on genetically modified organisms. In July 1974, soon after the first successful experimental modifications of DNA had been reported, the research community voluntarily deferred experiments with biologically active recombinant DNA molecules.Footnote 19 After intense studies and discussions, it was agreed on a conference in February 1975 that the potential dangers were manageable. The moratorium was lifted and experiments were resumed, with safeguards that had been agreed upon. We now have massive evidence confirming that it was right to lift the moratorium. This technology can be used safely to achieve important gains in agriculture and, not least, in medicine.Footnote 20 But still today, 44 years after the moratorium was lifted, there are activists who, contrary to the scientific consensus, uphold the picture of genetic modification as a new technology whose effects are largely unknown and potentially disastrous. This perception of the technology also continues to hold sway in a considerable part of the ethical literature on biotechnology. Scientific uncertainties that were overcome and laid to rest several decades ago still have a ghost life in the ethical literature.Footnote 21
Transfusion medicine offers another example of the difficulties involved in reversing precaution. In the early 1980s, countries all over the world introduced lifetime deferrals of blood donations from men who have sex with men in order to avoid the use of HIV-contaminated blood in blood transfusions. Since then, new methods based on testing of the blood have been developed.Footnote 22 A deferral period is still considered necessary to protect against recently acquired infections that will not be detected in antibody tests, but lifelong deferral is not needed. Quite a few countries have therefore shortened the deferral period to 12 months, and initiatives have been taken to reduce it to 3 months.Footnote 23 However, many countries still apply a lifetime deferral period, although it does not provide any additional protection. Such policies contribute to the scarcity of donors, and they may also be conducive to homophobic sentiments and discrimination.Footnote 24
“Ideally, precautionary policies would be applied when an uncertainty of risk exists, and its application modified or removed as new evidence is brought to light. Such an approach is rarely taken, however. In transfusion medicine, it has proven politically challenging to reverse a decision that was introduced for the purposes of protecting the public.”Footnote 25
Similar tendencies can be seen in other areas. For instance, as Ari Schick has pointed out, the ethical discourse on human enhancement has largely been detached from scientific and technological research. This puts “speculative bioethics at even greater risk of becoming a self-perpetuating program, unresponsive to the unpredictability that is at the heart of the scientific enterprise, and largely unaccountable for the role it plays within various ‘communities of promise [and peril]’.”Footnote 26
The growth of speculative ethics may seem to be unstoppable. Once mere possibility arguments have been put forward, are they not impossible to ignore? To the contrary, I will argue that it is possible to write off many mere possibility arguments for the simple reason that what is possible may nevertheless be irrelevant for practical decisionmaking. For an example of a speculation that was rather promptly dismissed we can consider the “polywater” scare that circulated in 1969. Polywater, or polymerized water, was an alleged substance that would supposedly “grow at the expense of normal water under any conditions found in the environment,” thus replacing all natural water on earth and destroying all life on this planet.Footnote 27 Proposals were made to take precautions against this substance, but its alleged properties were highly implausible from a chemical point of view, and it was soon shown not to exist. And of course, polywater is just one of many constructible chemical doomsday stories. As Michael King and coworkers pointed out, “there may be a nonzero possibility that a chemical synthesized in a laboratory may initiate a chain-reaction that obliterates the ozone layer, destroying all life on earth.”Footnote 28 But as these authors also say, “prohibiting all chemical synthesis based on this possibility would be ridiculous.”Footnote 29 (Such a prohibition would put an end to all developments of new pharmaceuticals.) There is no obvious reason why polywater should be singled out for action among all the logically possible but scientifically highly implausible doomsday scenarios that can be constructed.
Another example, with considerable public health implications, is the supposition that MMR vaccine causes autism. This claim was put forward by Andrew Wakefield in a 1998 article, which is now known to be based on falsified data.Footnote 30 But in spite of that information, antivaccination activists continue to claim that the vaccine may possibly give rise to autism.Footnote 31 Concededly, science cannot prove with absolute certainty that this vaccine will never, in any person, causally contribute to autism. However, there are two other things that science can do, and has in fact already done. First, science can find out whether autism is more common among vaccinated than unvaccinated persons. Such studies have shown no difference between vaccinated and unvaccinated individuals.Footnote 32 Second, science can tell us whether there are any valid reasons to suspect that the vaccine, rather than anything else that happens to a child, can lead to autism. There is in fact no such reason, or in other words: No one has been able to present a plausible mechanism for such an effect of the vaccine. Obviously, this does not prove that there is no such mechanism. There might be some potential mechanism that we are not aware of. However, the same is true about any other of the many events and influences in a young person’s life. In the same sense that it is a (mere) possibility that the vaccine causes autism, it is also a (mere) possibility that autism is caused by riding the merry-go-round, playing with a skipping rope, or eating ice cream with strawberries. Most importantly, the same applies to the supposition that the vaccine has a protective effect against autism.Footnote 33 Therefore, even if science cannot tell us with the absolute certainty of 1 + 1 = 2 that there is no vaccine-autism connection, it has already provided sufficient reason to disregard that (mere) possibility in all practical deliberations.
The gist of these examples is that with the help of science and argumentative tools, we can disarm and defeat pointless mere possibility arguments.Footnote 34 An important reason to do this work, and thereby remove unwarranted doomsday scenarios from the agenda, is that this place should be reserved for real problems to which our attention needs to be directed.
Neuroethics
Neuroethics is one of the areas of applied ethics that has been most affected by unrealistic scenarios. There is, unfortunately, some truth in the scathing criticism of the whole field that was launched a decade ago by a clinical neurologist and a medical anthropologist:
“The moral scope of the new field of neuroethics is, in our view, limited… Neuroethicists are excessively enamored of the moral dilemmas they believe to be associated with the use of unproven functional imaging technologies to read minds and the use of yet-to-be-developed cognitive enhancement drugs to improve them. Moreover, they ignore issues of public health and social justice, such as the effects of lifelong exposure to toxins on our brains. Also ignored are the ethical challenges posed by traumatic brain injuries that are incurred in warfare.”Footnote 35
Similarly, Gilbert et al.Footnote 36 describe neuroethics as “engaged in a speculative bubble, which may need to be deflated and rectified by grounding it in empirical evidence as available through the scholarly neuroscientific and medical literature.”
Cognitive enhancement, mind reading, and remote control by means of implanted devices are among the most common topics in the speculative parts of neuroethics.Footnote 37 In a paper critical of such speculations, Alfred Nordmann and Arie Rip wrote:
“Current discussions about ethics and nanotechnology take considerable interest in nano-enabled brain implants, mind–machine interfaces and related developments—such as the privacy issues that arise when thoughts can be read routinely from brain activity.”Footnote 38
After briefly criticizing the assumptions behind these concerns, they continued in the same paragraph:
“At the same time, other developments that demand ethical attention receive much less attention. For instance, nano-enabled advances in deep-brain stimulation can produce tremendous benefits for patients with Parkinson’s disease, but they can also be used to alter moods and even personalities—but these issues are mostly overlooked by the nanoethics community.”Footnote 39
Thus, in the view of these authors, personality changes induced by deep brain stimulation are among the practically relevant issues that neuroethicists should turn to instead of other, more fanciful scenarios. Similarly, in 2014 Gilbert and Goddard criticized ethicists’ preoccupation with “the fictional scenario of whether it is permissible to control human minds,” arguing that this focus induces ethicists to “ignore the practical everyday postoperative significance of implanted patient selfhood, that is, who someone is, their sense of identity, and risk minimizing the importance of changes in who they are.”Footnote 40 In another article, Gilbert and coworkers criticized neuroethical discussions on potential mind control via brain implants for optogenetic stimulation:
“Such speculation detracts from focusing on more immediate ethical needs related to application possibilities such as optogenetic restoration of lost vision, irreversibility of the procedure, and postoperative adverse effects on the patient’s personality.”Footnote 41
These authors present personality changes due to implanted devices as an example of realistic and legitimate neuroethical concerns, in contradistinction to scenarios involving mind reading or mind control. As I see it, this is a reasonable appraisal. Brain activity takes place in exceptionally compact three-dimensional structures. The construction of a nondestructive brain-machine interface that interacts with more than a minuscule fraction of the brain’s activity is a daunting task, and no credible way to do it seems to have been proposed.Footnote 42 Even the more science-fictional scenarios, involving multiple sensors (somehow) inserted into the brainFootnote 43 are very far from the complete reading of the mind that is presupposed in the more futuristic parts of neuroethics. As aptly pointed out by Berger, many of these scenarios are based on a primitive understanding of the brain as some sort of digital hardware, which is a gross misrepresentation of its biological nature.Footnote 44
On the other hand, changes in personality and self-image are well-known side effects from interventions in the brain, such as tumour resection,Footnote 45 brain artery aneurysm surgery,Footnote 46 and surgical treatments of psychiatric disorders.Footnote 47 Therefore, although undesired psychological changes seem to be rare in current applications of deep brain stimulation,Footnote 48 concern for potential such effects is warranted when stimulation of other parts of the brain are considered, for instance in treatment of psychiatric conditions.Footnote 49 The same applies when new and more advanced implants are introduced, such as bidirectional implants, which monitor brain activity as a means to optimize the stimulation.Footnote 50 These risks have to be weighed against the benefits obtainable with deep brain stimulation, which include that many patients report becoming “more themselves in comparison to their previous life that was dominated by their obsessive–compulsive disorder.”Footnote 51 Notably, even when speculative exaggerations have been put aside—or perhaps in particular when this has been done—psychological effects of this and other treatments that affect brain function give rise to interesting philosophical issues. For instance, some psychiatric patients “regard elements of their illness as part of their identity.”Footnote 52 In such cases, it is no easy matter to determine which personality changes are desirable and which should be avoided. Both philosophical and clinical investigations are needed to make progress in this issue.
Conclusions
To sum up, I have proposed that the use of unrealistic examples in ethics:
-
1) is justified in foundational ethics, but only under the (controversial) assumption that we need ethical principles and theories that would be applicable even under the extreme circumstances referred to in these examples,
-
2) is positively harmful in applied ethics whenever it leads us away from issues and aspects that are in urgent need of our attention,
-
3) is often engendered by mere possibility arguments that seem unconquerable but can in fact be written off with proper use of science and argument analysis, and
-
4) currently have an inordinate role in neuroethics, where they should yield place to problems that are relevant for the treatment and care of patients and for the development of new methods for diagnosis and treatment.