In the business world, AI advocates tell companies and workers that they will fall behind if they fail to integrate generative AI into their operations. In the sciences, AI advocates promise that AI will aid in curing hitherto intractable diseases.
In higher education, AI promoters admonish teachers that students must learn how to use AI or risk becoming uncompetitive when it comes time to find a job.
In national security, AI's champions say that either the nation invests heavily in AI weaponry, or it will be at a disadvantage against the Chinese and Russians, who are already doing so.
The argument across these different domains is essentially the same: The time for AI skepticism has come and gone. The
In the past few years, my colleagues and I (Nir Eisikovits, Professor of Philosophy and Director, Applied Ethics Center, UMass Boston) have been studying the ethical questions raised by the widespread adoption of AI, and I believe the inevitability argument is misleading.
History and hindsight
This claim is the most recent version of a deterministic view of technological development. It's the belief that innovations are unstoppable once people start working on them. In other words, some genies don't go back in their bottles. The best you can do is harness them for your good purposes.This deterministic approach to tech has a long history. It's been applied to the influence of the printing press as well as to the rise of automobiles and the infrastructure they require, among other developments.
But I believe that when it comes to AI, the technological determinism argument is both exaggerated and oversimplified.
AI in the field(s)
Consider the contention that businesses can't afford to stay out of the AI game. The case has yet to be made that AI is delivering significant productivity gains to the firms that use it. A report in TheAI's role in higher education is also still very much an open question. Though universities have, in the past two years, invested heavily in AI-related initiatives, evidence suggests they may have jumped the gun.
The technology can serve as an interesting pedagogical tool. For example, creating a Plato chatbot that lets students have a text conversation with a bot posing as Plato is a cool gimmick.
However, AI is already starting to displace some of the best tools teachers have for assessment and for developing critical thinking, such as writing assignments. The college essay is going the way of the dinosaurs as more teachers give up on the ability to tell whether their students are writing their papers themselves. What's the cost-benefit argument for giving up on writing, an important and useful traditional skill?
In the sciences and medicine, the use of AI seems promising. Its role in understanding the structure of proteins, for example, will likely be significant for curing diseases. The technology is also transforming medical imaging and has been helpful in accelerating the drug discovery process.
But the excitement can become exaggerated. AI-based predictions about which cases of COVID-19 would become severe have roundly failed, and doctors rely excessively on the technology's diagnostic ability, often against their own better clinical judgment. And so, even in this area, where the potential is great, AI's ultimate impact is unclear.
In national security, the argument for investing in AI development is compelling. Since the stakes can be high, the argument that if the Chinese and the Russians are developing AI-driven autonomous weapons, the United States can't afford to fall behind has real purchase.
But a complete surrender to this form of reasoning, though tempting, is likely to lead the
One step at a time
Surveying the potential significance and risks of AI in these different domains merits some skepticism about the technology. I believe that AI should be adopted piecemeal and with a nuanced approach rather than subject to sweeping claims of inevitability. In developing this careful take, there are two things to keep in mind: -First, companies and entrepreneurs working on artificial intelligence have an obvious interest in the technology being perceived as inevitable and necessary, since they make a living from its adoption. It's important to pay attention to who is making claims of inevitability, and why.
-Second, it's worth taking a lesson from recent history. Over the past 15 years, smartphones and the social media apps that run on them have come to be seen as a fact of life - a technology as transformative as it is inevitable. Then data started emerging about the
After a long experiment with the mental health of kids, facilitated by claims of technological determinism, Americans changed course. What seemed fixed turned out to be alterable. There is still time to avoid repeating the same mistake with artificial intelligence, which potentially could have larger consequences for society.