AI’s dominance is not inevitable, according to a tech ethicist

Anyone following the rhetoric around artificial intelligence in recent years has heard one version or another of the claim that AI is inevitable. Common themes are that AI is already here, it is indispensable, and people who are bearish on it harm themselves.

In the business world, AI advocates tell companies and workers that they will fall behind if they fail to integrate generative AI into their operations. In the sciences, AI advocates promise that AI will aid in curing hitherto intractable diseases.

In higher education, AI promoters admonish teachers that students must learn how to use AI or risk becoming uncompetitive when the time comes to find a job.

And, in national security, AI’s champions say that either the nation invests heavily in AI weaponry, or it will be at a disadvantage vis-à-vis the Chinese and the Russians, who are already doing so.

The argument across these different domains is essentially the same: The time for AI skepticism has come and gone. The technology will shape the future, whether you like it or not. You have the choice to learn how to use it or be left out of that future. Anyone trying to stand in the technology’s way is as hopeless as the manual weavers who resisted the mechanical looms in the early 19th century.

In the past few years, my colleagues and I at UMass Boston’s Applied Ethics Center have been studying the ethical questions raised by the widespread adoption of AI, and I believe the inevitability argument is misleading.

History and hindsight

In fact, this claim is the most recent version of a deterministic view of technological development. It’s the belief that innovations are unstoppable once people start working on them. In other words, some genies don’t go back in their bottles. The best you can do is harness them to your good purposes.

This deterministic approach to tech has a long history. It’s been applied to the influence of the printing press, as well as to the rise of automobiles and the infrastructure they require, among other developments.

But I believe that when it comes to AI, the technological determinism argument is both exaggerated and oversimplified.

AI in the field(s)

Consider the contention that businesses can’t afford to stay out of the AI game. In fact, the case has yet to be made that AI is delivering significant productivity gains to the firms that use it. A report in The Economist in July 2024 suggests that so far, the technology has had almost no economic impact.

AI’s role in higher education is also still very much an open question. Though universities have, in the past two years, invested heavily in AI-related initiatives, evidence suggests they may have jumped the gun.

The technology can serve as an interesting pedagogical tool. For example, creating a Plato chatbot that lets students have a text conversation with a bot posing as Plato is a cool gimmick.

But AI is already starting to displace some of the best tools teachers have for assessment and for developing critical thinking, such as writing assignments. The college essay is going the way of the dinosaurs as more teachers give up on the ability to tell whether their students are writing their papers themselves. What’s the cost-benefit argument for giving up on writing, an important and useful traditional skill?

In the sciences and in medicine, the use of AI seems promising. Its role in understanding the structure of proteins, for example, will likely be significant for curing diseases. The technology is also transforming medical imaging and has been helpful in accelerating the drug discovery process.

But the excitement can become exaggerated. AI-based predictions about which cases of COVID-19 would become severe have roundly failed, and doctors rely excessively on the technology’s diagnostic ability, often against their own better clinical judgment. And so, even in this area, where the potential is great, AI’s ultimate impact is unclear.

In national security, the argument for investing in AI development is compelling. Since the stakes can be high, the argument that if the Chinese and the Russians are developing AI-driven autonomous weapons, the United States can’t afford to fall behind, has real purchase.

But a complete surrender to this form of reasoning, though tempting, is likely to lead the U.S. to overlook the disproportionate impact of these systems on nations that are too poor to participate in the AI arms race. The major powers could deploy the technology in conflicts in these nations. And, just as significantly, this argument de-emphasizes the possibility of collaborating with adversaries on limiting military AI systems, favoring arms race over arms control.

One step at a time

Surveying the potential significance and risks of AI in these different domains merits some skepticism about the technology. I believe that AI should be adopted piecemeal and with a nuanced approach rather than subject to sweeping claims of inevitability. In developing this careful take, there are two things to keep in mind:

First, companies and entrepreneurs working on artificial intelligence have an obvious interest in the technology being perceived as inevitable and necessary, since they make a living from its adoption. It’s important to pay attention to who is making claims of inevitability, and why.

Second, it’s worth taking a lesson from recent history. Over the past 15 years, smartphones and the social media apps that run on them came to be seen as a fact of life—a technology as transformative as it is inevitable. Then data started emerging about the mental health harms they cause teens, especially young girls. School districts across the U.S. started to ban phones to protect the attention spans and mental health of their students. And some people have reverted to using flip phones as a quality of life change to avoid smartphones.

After a long experiment with the mental health of kids, facilitated by claims of technological determinism, Americans changed course. What seemed fixed turned out to be alterable. There is still time to avoid repeating the same mistake with artificial intelligence, which potentially could have larger consequences for society.

<hr class=“wp-block-separator is-style-wide”/>

Nir Eisikovits is a professor of philosophy and director of the Applied Ethics Center at UMass Boston.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

https://www.fastcompany.com/91225614/artificial-intelligence-dominance-not-inevitable-ethics?partner=rss&amp;utm_source=rss&amp;utm_medium=feed&amp;utm_campaign=rss+fastcompany&amp;utm_content=rss

Creato 3mo | 9 nov 2024, 12:40:03


Accedi per aggiungere un commento

Altri post in questo gruppo

The American woman who went viral in Pakistan has a crypto coin

The “American woman in Pakistan” now has a crypto coin. 

If you don’t know who that is, American Onijah Andrew Robinson recently went viral after claiming she flew to Pa

11 feb 2025, 19:20:09 | Fast company - tech
DOGE has disregarded data protection and privacy norms. The consequences will be felt years down the line

It has been a tumultuous few weeks since Donald Trump took office for the second time as president of the United States, While Trump has garnered headlines for his outlandish executive orders aime

11 feb 2025, 17:10:05 | Fast company - tech
Workplace Wellness: Calm CEO’s guide to prioritizing mental health

David Ko, CEO of Calm, speaks with Brendan Vaughan about the state of mental health solutions in the workplace.

https://www.fastcompany.com/91276663/workplace-wellness-calm-ceos-guide-to-prio

11 feb 2025, 17:10:04 | Fast company - tech
3 ways Tesla stands to win from Elon Musk’s war on the U.S. government

Elon Musk has long railed against the U.S. government, saying a crushing number of

11 feb 2025, 17:10:03 | Fast company - tech
Will my social media posts really help my career?

There are certain social media rules we can all agree on: Ghosting a conversation is impolite, and replying “k” to a text is the equivalent of a backhand slap (violent, wrong, and rude). But what

11 feb 2025, 12:20:12 | Fast company - tech
This Google Maps ‘safety’ feature is actually making roads more dangerous

Picture this: You’re driving on a crowded highway, preparing to change lanes and pass a tractor-trailer. As you check your mirrors, a loud chime on your car’s infotainment screen rings out.

11 feb 2025, 12:20:10 | Fast company - tech
How SoftBank’s Masayoshi Son plans to win the AI wars

Masayoshi Son is back on top. On January 22, President Donald Trump announced a joint venture from Son’s investment holding company, SoftBank, along with OpenAI and Oracle, to

11 feb 2025, 12:20:08 | Fast company - tech