A new RAND report says ideas like mutually assured destruction and minimal deterrence strategy offer a lot less assurance in the age of intelligent software.
Artificial intelligence could destabilize the delicate balance of nuclear deterrence, inching the world closer to catastrophe, according to a working group of experts convened by RAND. New smarter, faster intelligence analysis from AI agents, combined with more sensor and open-source data, could convince countries that their nuclear capability is increasingly vulnerable. That may cause them to take more drastic steps to keep up with the U.S. Another worrying scenario: commanders could make decisions to launch strikes based on advice from AI assistants that have been fed wrong information.
Read more ....
WNU Editor: The RAND report is here .... How Artificial Intelligence Could Increase the Risk of Nuclear War (RAND).
Military And Intelligence News Briefs -- April 25, 2018
Doomsday AI will cause a nuclear war by 2040 that could DESTROY humanity, and there may be no way to prevent it -- Daily Mail
Pentagon AI center progressing, but hypersonics and lasers may not get same treatment -- Defense News
China’s AI Talent 'Arms Race' -- Elsa B. Kania, The Strategist
US Defense intelligence chief: ‘A lot of technology remains untapped’ -- Space News
Why Syria may be the most aggressive electronic warfare environment on Earth -- Mark Pomerleau, C4ISRNet
Exclusive: Russian civilians helping Assad use military base back home - witnesses -- Reuters
Russia Starts Production of S-400 Missile Systems for Turkey -- Sputnik
Russia Parades War-‘Bots for the First Time -- War Is Boring
Sounding The Alarm About A New Russian Cyber Threat -- NPR
Russian and Chinese Hypersonic Glide Vehicles: Closing the Gap -- RCD
Chinese cyberspies allegedly targeted Japanese defense firms for North Korea secrets: report -- Japan Times/Bloomberg
Indian Army reduces ammunition purchases as cost-saving measure -- Jane's 360
South Korea to open bid for anti-submarine helicopters -- Defense News
Turkey awards multibillion-dollar contract for indigenous Altay tank -- Defense News
Majority of British F-35B fleet to arrive in UK this summer -- UK Defense Journal
Afghan military trainees in the US keep going AWOL -- Military Times
US House Lawmakers Close Pentagon Budget Hearing to the Public -- Defense One
The Army Reserve Simply Doesn’t Have Enough People Willing to Fill Command Slots -- Task & Purpose
Navy Accepts Partial Delivery of Zumwalt-class DDG Michael Monsoor -- USNI News
Marines add Stinger missiles, lasers to vehicles to make up for lagging air defense -- Marine Times
US Air Force, Boeing still clash over KC-46 delivery timeline -- Defense News
F-16 from Luke Air Force Base crashes -- Air Force Times
How 1 Revolutionary New Technology Is Making the F-35B Even Deadlier -- National Interest
US Air Force to Put Sensors on Allies’ Satellites -- Defense One
As Space War Looms, Air Force's Biggest Weakness May Be How It Buys Satellites -- Loren Thompson, Forbes
Marines Zero In On Requirements for Future MUX Unmanned Aerial Vehicle -- USNI News
Drone Technology Is Now Dangerous Enough to Kill For -- NextGov
How to Run a Cyber War…Game -- Michael Sulmeyer & Dmitri Alperovitch, The Cipher Brief
CIA rethinks how to protect operatives’ digital lives -- C4ISRNet
Senate confirms new head of Cyber Command, NSA -- Fifth Domain
The Army just dumped a bunch of mandatory training to free up soldiers’ time -- Army Times
Rise in aviation mishaps signals serious readiness, manning challenges -- Military Times editorial
Do new Trump arms export rules live up to the hype? -- Aaron Mehta, Defense News
Countering China’s Militarization of the Indo-Pacific -- Michael J. Green and Andrew Shearer, War on the Rocks
America’s Endangered Nuclear Deterrent: The Case for Funding Two Critical Capabilities -- Mark Gunzinger, Carl Rehberg, and Gillian Evans, War On The Rocks
Families of former POWs angry loved ones can't receive full honors at Arlington -- Stars and Stripes
1 comment:
The bigger problem is this: If an AI tells you to kill your enemy because of reason X, generals won't listen to you or anyone else. There will more and more be an automatism in decision making. While you of course still will have people in important decision making roles for decades to come, they will - due to the speed and complexity of AI - more and more just rubber-stamp kill and other decisions. And, if an AI tells you that you MUST nuke your enemy and gives you XYZ reasons, you will likely do it, if that machine wasn't wrong before....It's like in the movie Fail Safe, just with AI
Post a Comment