Written by 8:21 pm AI policies, Latest news

### Pentagon’s Alleged Development of Lethal AI Prompts Stringent Regulations

The Pentagon updated its 2012 directive on autonomous weapons systems to put an oversight and appro…

Fears that the Pentagon has been ‘constructing lethal robots in the basement’ may have prompted stricter regulations on AI, requiring approval for all systems before deployment.

The Department of Defense (DoD) recently revised its AI guidelines amidst uncertainty surrounding the utilization of autonomous decision-making machines in combat, as stated by the deputy assistant defense secretary.

Michael Horowitz clarified at a recent event that the ‘directive does not hinder the advancement of any systems’ but aims to ‘explicitly outline permissible boundaries’ while upholding a ‘commitment to ethical conduct’ in the development of lethal autonomous systems.

Despite the Pentagon’s belief that these adjustments will reassure the public, some remain skeptical of the efforts.

Fears the Pentagon has been ‘building killer robots in the basement’ may have led to stricter AI rules that mandated all systems must be approved before deployment. Pictured is the MAARS unmanned vehicle unveiled in in 2015

The news of the Pentagon’s update to the 2012 ‘Autonomy in Weapon Systems’ has triggered online discussions, with many questioning, ‘If the Pentagon denies it, are they actually doing it?’

Dailymail.com has contacted the DoD for a statement.

The DoD has been actively pursuing modernization of its arsenal with autonomous drones, tanks, and other weapons capable of selecting and engaging targets without human intervention.

Mark Brakel, director of the advocacy group Future of Life Institute (FLI), expressed concerns to DailyMail.com, stating: ‘These weapons pose a significant risk of unintended escalation.’

He elaborated on the potential dangers of AI-powered weapons misinterpreting benign stimuli, such as sunlight, as threats, leading to unwarranted attacks on foreign entities.

Brakel warned that without effective human oversight, AI-driven weapons could amplify the likelihood of accidents in volatile regions like the Taiwan Strait, resembling the Norwegian rocket incident, but with heightened consequences.

The Department of Defense (DoD) recently updated its AI rules among ‘a lot of confusion about’ how it plans to use self-decision-making machines on the battlefield

The DoD has advocated for global oversight of AI weaponry by urging nations to endorse the Political Declaration on Responsible Military Use of Artificial Intelligence and Autonomy, with 47 countries having pledged their support as of November.

During a panel discussion on January 9, Horowitz emphasized the DoD’s dedication to fostering public trust and confidence in technology, along with its commitment to upholding international humanitarian laws.

Brakel emphasized that the efficacy of any new directive should be measured by the restrictions it imposes or significant modifications it initiates on weapon development processes.

He noted a lack of evidence showcasing tangible impacts from the directive changes on weapon development practices.

Apart from the U.S., China and Russia are also actively deploying their AI-equipped military systems in ongoing conflicts.

In a bold move, the Pentagon announced plans in November to deploy thousands of AI-enabled autonomous vehicles by 2026 to match the capabilities of adversaries.

Humanitarian groups worry that the AI-powered weapons could fall into the wrong hands or unintentionally spark a war by mistaking something like sunlight for an attack

The ambitious initiative, known as Replicator, aims to accelerate advancements in the military’s transition towards agile, intelligent, cost-effective, and numerous platforms, as highlighted by Deputy Secretary of Defense Kathleen Hicks in August.

Horowitz described Replicator as a strategic approach to swiftly and effectively field critical capabilities aligned with the national defense strategy.

He reiterated the DoD’s unwavering commitment to building public trust in technology and upholding international humanitarian laws.

However, some members of FLI remain unconvinced about the efficacy of the stringent regulations and the DoD’s ability to enforce them.

Anna Hehir, leading autonomous weapons system researcher at FLI, cautioned against the potential consequences of embracing AI in military applications without a thorough understanding of its implications.

She likened the current scenario to the dawn of the nuclear era, emphasizing the need for cautious consideration to avert global disasters resulting from an arms race mentality.

Visited 3 times, 1 visit(s) today
Last modified: January 24, 2024
Close Search Window
Close