Lohn at al. ‘18 (Andrew J. has Ph.D. in electrical engineering, University of California Santa Cruz, and is an engineer at the RAND Corporation and a professor of public policy at the Pardee RAND Graduate School, 4/30/18, “Will artificial intelligence undermine nuclear stability?” https://thebulletin.org/2018/04/will-artificial-intelligence-undermine-nuclear-stability/)(Shiv)
Artificial intelligence and nuclear war have been fiction clichés for decades. Today’s AI is impressive to be sure, but specialized, and remains afar cry from computers that become self-aware and turn against their creators. At the same time, popular culture does not do justice to the threats that modern AI indeed presents, such as its potential to make nuclear war more likely even if it never exerts direct control over nuclear weapons.¶ Russian President Vladimir Putin recognized the military significance of AI when he declared in September that the country that leads in artificial intelligence will eventually rule the world. He may be the only leader to have put it so bluntly, but other world powers appear to be thinking similarly. Both China and the United States have announced ambitious efforts to harness AI for military applications, stoking fears of an incipient arms race.¶ In the same September speech, Putin said that AI comes with “colossal opportunities” as well as “threats that are difficult to predict.” The gravest of those threats may involve nuclear stability—as we describe in a new RAND publication that outlines a few of the ways in which stability could be strained.¶ Strategic stability exists when governments aren’t tempted to use nuclear threats or coercion against their adversaries. It involves more than just maintaining a credible ability to retaliate after an enemy attack. In addition to that deterrent, nuclear stability requires assurance and reassurance. When a nation extends a nuclear security guarantee to allies, the allies must be assured that nukes will be launched in their defense even if the nation extending the guarantee must put its own cities at risk. Adversaries need to be reassured that forces built up for deterrence and to protect allies will not be used without provocation. Deterrence, assurance, and reassurance are often at odds with each other, making nuclear stability difficult to maintain even when governments have no interest in attacking each other.¶ In a world where increasing numbers of rival states are nuclear-armed, the situation becomes almost unmanageable. In the 1970s, four of the five declared nuclear powers primarily targeted their weapons on the fifth, the Soviet Union (Beijing, after its 1969 border clashes with the Soviet Union, feared Moscow much more than Washington). It was a relatively simple bilateral stand-off between the Bolsheviks and their many adversaries. Today, nine nuclear powers are entangled in overlapping strategic rivalries—including Israel, which has not declared the nuclear arsenal that it is widely believed to possess. While the United States, the United Kingdom, and France still worry about Russia, they also fret about an increasingly potent China. Beijing’s rivals include not just the United States and Russia but India as well. India fears China too, but primarily frets about Pakistan. And everyone is worried about North Korea.¶ In such a complex and dynamic environment, teams of strategists are required to navigate conflict situations—to identify options and understand their ramifications. Could AI make this job easier? With AI now beating human professionals in the ancient Chinese strategy game Go, as well as in games of bluffing such as poker, countries may be tempted to build machines that could “sit” at the table amid nuclear conflicts and act as strategists.¶ Artificially intelligent machines may prove to be less error-prone than humans in many contexts.But for tasks such as navigating conflict situations, that moment is still far off in the future. Much effort must be expended before machines can—or should—be relied on for consistent performance of the extraordinary task of helping the world avoid nuclear war. Recent research suggests that it is surprisingly simple to trick an AI system into reaching incorrect conclusions when an adversary gets to control some of the inputs, such as how a vehicle is painted before it is photographed.