First Principles for Medical Artificial Intelligence

Download PDF: PDF File Image

Abstract

Medical artificial intelligence (AI) technologies, by their capacity to decipher enormous data sets, identify meaningful patterns beyond what human intelligence can recognize, and in some cases render decisions without human assistance, are poised to transform healthcare. As with any powerful technology, careful ethical analysis is needed if we are to realize the benefits of AI while avoiding its perils. Four available perspectives are recognized. One perspective is technological sentimentalism, which resists novel technologies that seem to displace a more natural way of inhabiting the world. A second perspective is technological messianism, which uncritically welcomes novel technology as intrinsically good and the answer to all human problems. A third perspective, common today, is technological pragmatism, which weighs benefits and risks in a utilitarian framework that emphasizes empirical facts but disregards moral values, considering them to be opinions without consequence or validity. A fourth and preferred perspective is technological responsibilism, which considers not only outcomes but also the moral values laden in the design and implementation of technology. Technological responsibilism respects the deeply human attributes of voluntary responsibility, moral agency, and character. Morally responsible use of AI is needed if healthcare professionals are to sustain their focus, not on technology, but on patients.

Keywords: Artificial Intelligence, Christian Bioethics, Ethics of Technology, Medical Technology, Moral Responsibility, Pragmatism, Virtue

The history of medicine is punctuated by novel innovations. Consider the impact of anesthesia, antibiotics, the hypodermic syringe, the stethoscope, x-ray imaging, MRI, artificial ventilators, pagers, pacemakers, organ transplants, gene sequencing—the list goes on. The first use of each of these inventions revolutionised diagnosis and treatment of the sick. Technology turns the pages of the book of medical progress.

Prophets of the technosphere predict that the next wave of extraordinary progress lies with artificial intelligence, or “AI” for short. AI refers to the simulation of human intelligence by machines that are programmed to mimic human thought patterns such as learning, adapting to new data, problem-solving, executing complex tasks, even social interaction—without being assisted by humans. Enthusiasm for its transformative potential has led to predictions that AI will usher in a fourth technological revolution comparable to the neolithic transition to agriculture, the industrial revolution utilising mechanised production, and the digital revolution based on computer processing of digital information.

Medical AI[1]

What AI might mean for medicine is a subject of exciting speculation. Inspired by accelerating change, the CEO of Google has predicted that, within the next five years, “we will shift to a world that is AI-first.”[2] The challenge for physicians will be to calibrate medical AI in such a way as to place the interests of patients first. With that in mind, a responsible approach to making decisions about medical AI must not be limited to technique but should explore thoughtfully the ethics of human interaction with technology.

Ways to Think About Technology

The bioethicist Michael Sleasman offers a helpful ethical framework for thinking through decisions about the use of technology.[3] He proposes four categories, which he labels as sentimentalism, messianism, pragmatism, and responsibilism. How one navigates these categories and balances their concerns draws from prior assumptions about what it means to be human and to be a thinking, morally responsible being.

Technological Sentimentalism

In one corner are those who instinctively question or resist technological innovation. They may feel threatened by technological trends that erode kindness in communication or that seem to treat people as things to be manipulated. Finding it increasingly difficult to unplug from electronic connectivity, they feel that technology deprives them of a more natural way of inhabiting the world. They see technology as the source of a host of social problems and, thus, novel technologies such as AI a Pandora’s box of potential new problems. Recognising that AI has the potential to magnify biases in data, they worry that AI could worsen unjust disparities and social inequalities, placing the poor and vulnerable at greater disadvantage. Sentimentalists value being satisfied with what one has and lament the loss of what is nostalgically imagined as a simpler, less technological age. Sentimental rejection of medical AI, however, would deny us its benefits.

Technological Messianism

At another corner are those who regard technology as intrinsically good and desirable. The potential useful benefits of AI stagger the imagination, especially as medical progress relies increasingly on digital technologies. The enormous amount of data in every person’s genome and the accumulating scientific content in medical journals exceed the capacity of the human brain to recall, interpret, or synthesise all there is to know relevant to health and disease. AI promises to augment our ability to assess megadata and bridge the information gap. An already familiar application of AI is natural language processing to generate clinical notes efficiently from speech. In medical research, machine-learning algorithms are mining vast clinical data sets to detect previously unrecognised disease patterns. Artificial neural networks utilising deep learning are beginning to analyse radiographic, histologic, and morphologic images with stunning diagnostic accuracy. Such advances may lead to earlier detection of disease when it is more easily treatable.

Optimism in technological progress traces to the European Enlightenment, which sought to improve the human condition by applying rational knowledge and invention to fulfill material needs. Optimism in technology can overlook unintended bad consequences, and it flies off the rails if taken to the extreme by those who would place ultimate hope in machinery as the saviour of society.

The uncritical adoption of technology sometimes appeals to the technological imperative, which is the philosophical claim that technology in all its achievable forms is inevitable. Bowing to supposed inevitability, the technological imperative reduces to the belief that, if it can be done, it should be done. Such fatalism is seriously mistaken because it abandons ethics by denying the role of human decision and responsibility.

Technological Pragmatism

The prevailing attitude toward technology today is pragmatism, which regards usefulness as the ultimate measure of truth and meaning. From a pragmatic perspective, AI is a transactional tool for solving human problems by plugging predicted benefits and risks into a mathematical equation, digesting more and more data to calculate potential consequences far into the future. Would such hyperpragmatism ultimately deliver human flourishing?

Thinking about thinking technology brings this question to the forefront. Projects to construct a machine intelligence superior to human intelligence presuppose that thought is reducible to matter and its physical causes and random collisions. If intelligence were nothing more than neurons firing and neurochemicals churning in a particular pattern, then, in principle, intelligence equivalent to or surpassing the human mind could be replicated in electronic circuits. Some philosophers go so far as to suggest that, if machines can mimic human intelligence, then more powerful machines will eventually exceed it and potentially replace it. The Oxford bioethicist Julian Savulescu speculates that, “Humans may become extinct. . . . We might have reason to save or create such vastly superior lives, rather than continue the human line.”[4] One rightly shudders to consider the implications of such a view for medicine.

Whereas pragmatism seems objective, its blind spot is the assumption that empirically verifiable facts are the only valid knowledge. If that were true, then ethical problems could be resolved by acquiring a sufficient amount of measurable factual data and subjecting that data to intense machine analysis. In reality, facts are necessary but insufficient for ethical analysis. As the volume of clinical data increases, the variables that can interact increase, and the number of potential outcomes enlarges exponentially. Having more facts is not enough. Ethical decisions depend on considering what is meant by “useful,” which requires assigning to facts not just numerical value, but moral value. Lacking or ignoring knowledge of moral right and wrong, pragmatism and computer programs that emulate it descend into moral relativism.[5] C. S. Lewis understood this when he warned that recognising facts while denying the realm of value undermines the authority of the ethical principles needed to justify limits on human manipulation.[6] Unconstrained pragmatism makes it possible to justify committing any evil, provided a greater benefit is predicted for society at large.

Powerful computational technology, even if supplied with unimaginable quantities of data and outfitted with the superfast computer processing speed of AI, operates within the realm of fact. Technology, while useful, is not alone the final answer to human problems. Also needed is virtue, an aspect of human intelligence that surpasses the realm of factual knowledge where AI operates, not by degree but in kind.

Technological Responsibilism

Technological responsibilism realistically accepts both the promise and peril of technology while engaging with it critically and with an awareness of what it means to be human. This perspective recognises both fact and value and distinguishes between the real and the artificial.

Technology itself is morally ambiguous and laden with the values of its designers and implementers. AI is a tool and not an agent, a machine made of matter and not a living being. Although AI might be programmed to imitate human emotion and in some ways to deliver care, it cannot truly care about the patient. An AI might be programmed to render responsible decisions that conform with humanly decided values, but it cannot be programmed to be responsible. Applying the attribute of trust to AI, according to physicians Matthew DeCamp and Jon Tilburt, “is a category error, mistakenly assuming that AI belongs to a category of things that can be trusted.”[7] AI cannot be said to have the human attributes of voluntary responsibility, moral agency, motives, or character.

Some uses of AI blur the line of moral responsibility. A signature feature of deep learning AI is what is known as its “black box” problem, in which hidden layers within artificial neural networks are inaccessible to human users, who see only input and output and cannot know how the AI is analysing data. One might accept an AI diagnosis of cancer based on a proven track record for accurate analysis of images, but if an unsupervised AI supplied with clinical data were to advise an elderly patient to forego life-prolonging treatment, without giving reasons, one would be justified in querying whether the machine decision was ethically valid.

Technological responsibilism entails a more comprehensive perspective on human nature than is available within the paradigm of pragmatism. Christian healthcare professionals affirm that humans are much more than the sum of their cells. Every patient is a precious unity of body and spirit who is loved by God (John 3:16–17). All people are endowed with unique dignity as image-bearers of the Creator (Genesis 1:27). Responsible use of technology respects this innate dignity in others, a dignity affirmed by Jesus Christ, who, though being in the form of God, took on humanity (Philippians 2:5–7). Unlike AI, which is incapable of compassion, Jesus loves us (John 15:9, Romans 5:8) and came that we might have life abundantly (John 10:10).

Moral responsibility for medical AI lies with those who design and implement its programming, for which reason it matters what moral vision is embedded in the technology. For medicine to retain its moral integrity, AI must remain our tool and not become our master. We must ensure the prerogative to override an AI-generated healthcare decision we believe to be morally wrong and harmful to patients. In harnessing the power of AI, we must be careful not to become so enchanted by its technical charm that we lose sight of the special dignity of our patients, who are bearers of God’s image.

Responsible use of AI extends to awareness of how its use may influence implicit attitudes toward others. Once we habitually converse with AI, a mere machine imitating human emotion, we must guard against thinking of other persons as objects. The philosopher Jay Richards observes, “The greatest delusion of our age is the paradoxical penchant to deny our own agency while attributing agency to the machines we create.”[8]

Conclusion

The potential medical benefits of AI are promising. If used wisely, AI promises to be a powerful tool to assist Christian healthcare professionals in their calling to heal the sick and to love and serve those in need (Mark 12:31, Luke 10:24–37, Galatians 5:13). Caution also is needed to fulfill the Hippocratic imperative to first do no harm.[9]

The eventual entry of superintelligent AI into medicine will challenge healthcare professionals in new ways to choose whom we will serve. When we reach the limit of what we as finite humans can accomplish, will we place our ultimate trust in supercomputers or in Christ? Will we look to and empower technology, potentially allowing it to master us, or will we strive to align our use of technology with God’s good purposes? Christians understand that God’s thoughts are infinitely higher than anything possible by machine intelligence (Isaiah 55:8). Christians believe that Christ, not technology, is the true saviour, and His teachings our first principles for providing medical care. Christians know that Christ, not machine intelligence, is the way, the truth, and the life (John 14:6), our only reliable source of wisdom, our rescue from disease and death, and our everlasting hope.

 

References

[1] This essay is co-published in collaboration with Triple Helix, the magazine of the Christian Medical Fellowship of the British Isles (UK and Ireland), edition 83, “New Technologies,” https://cmf.li/40z8d1z.

[2] Scot Schultz, “Artificial Intelligence: The Next Industrial Revolution,” InsideHPC, February 5, 2018, https://insidehpc.com/2018/02/artificial-intelligence-industrial-revolution.

[3] Michael J. Sleasman, “Thinking Through Technology,” The Center for Bioethics & Human Dignity (podcast), July 25, 2008, https://www.cbhd.org/podcast/thinking-through-technology-part-iii.

[4] Julian Savulescu, “The Human Prejudice and the Moral Status of Enhanced Beings: What Do We Owe the Gods?” in Human Enhancement, ed. Julian Savulescu and Nick Bostrom (Oxford, UK: Oxford University Press, 2009), 244.

[5] Austin L. Hughes, “The Folly of Scientism,” The New Atlantis 37 (Fall 2012): 32–50, https://www.thenewatlantis.com/publications/the-folly-of-scientism.

[6] John G. West, ed. The Magician’s Twin: C. S. Lewis on Science, Scientism, and Society (Seattle: Discovery Institute Press, 2012).

[7] Matthew DeCamp and Jon C. Tilburt, “Why We Cannot Trust Artificial Intelligence in Medicine,” The Lancet Digital Health 1, no. 8 (2019): E390, https://doi.org/10.1016/S2589-7500(19)30197-9.

[8] Jay W. Richards, The Human Advantage: The Future of American Work in an Age of Smart Machines (New York: Crown Forum, 2018), 195.

[9] Hippocrates, Epidemics (Circa 400 BC), Book I, Section XI, http://www.perseus.tufts.edu/hopper/text?doc=Perseus%3Atext%3A1999.01.0251%3Atext%3DEpid.%3Abook%3D1%3Achapter%3D2%3Asection%3D11.

 

Cite as: William P. Cheshire, Jr., “First Principles for Medical Artificial Intelligence,” Ethics & Medicine 37, no. 2–3 (2021): 82–86.

Download PDF: PDF File Image

About the Author

William P. Cheshire, Jr., MD
+ posts

William P. Cheshire, Jr., MD, MA, is Professor of Neurology at the Mayo Clinic in Jacksonville, Florida. He is also Senior Fellow in The Center for Bioethics & Human Dignity’s Academy of Fellows. In 2019, the Christian Medical & Dental Associations awarded him Educator of the Year. The views expressed herein are his own and do not necessarily reflect the positions of the professional organizations with which he is affiliated.

Posted in Commentary and tagged , , .