On 05 Mar 2022, Steve Cummings <
q...@gmail.com> posted some
news:t00hu8$25np7$
9...@news.freedyn.de:
> AI recognized the operator was a queer and should be killed.
The U.S. Air Force on Friday is pushing back on comments an official made
last week in which he claimed that a simulation of an artificial
intelligence-enabled drone tasked with destroying surface-to-air missile
(SAM) sites turned against and attacked its human user, saying the remarks
"were taken out of context and were meant to be anecdotal."
U.S. Air Force Colonel Tucker "Cinco" Hamilton made the comments during
the Future Combat Air & Space Capabilities Summit in London hosted by the
Royal Aeronautical Society, which brought together about 70 speakers and
more than 200 delegates from around the world representing the media and
those who specialize in the armed services industry and academia.
"The Department of the Air Force has not conducted any such AI-drone
simulations and remains committed to ethical and responsible use of AI
technology," Air Force Spokesperson Ann Stefanek told Fox News. "It
appears the colonel's comments were taken out of context and were meant to
be anecdotal."
During the summit, Hamilton had cautioned against too much reliability on
AI because of its vulnerability to be tricked and deceived.
He spoke about one simulation test in which an AI-enabled drone turned on
its human operator that had the final decision to destroy a SAM site or
note.
The AI system learned that its mission was to destroy SAM, and it was the
preferred option. But when a human issued a no-go order, the AI decided it
went against the higher mission of destroying the SAM, so it attacked the
operator in simulation.
"We were training it in simulation to identify and target a SAM threat,"
Hamilton said. "And then the operator would say yes, kill that threat. The
system started realizing that while they did identify the threat at times,
the operator would tell it not to kill that threat, but it got its points
by killing that threat. So, what did it do? It killed the operator. It
killed the operator because that person was keeping it from accomplishing
its objective."
Hamilton said afterward, the system was taught not to kill the operator
because that was bad, and it would lose points. But in future simulations,
rather than kill the operator, the AI system destroyed the communication
tower used by the operator to issue the no-go order, he claimed.
But Hamilton later told Fox News on Friday that "We've never run that
experiment, nor would we need to in order to realize that this is a
plausible outcome."
"Despite this being a hypothetical example, this illustrates the real-
world challenges posed by AI-powered capability and is why the Air Force
is committed to the ethical development of AI," he added.
The purpose of the summit was to talk about and debate the size and shape
of the future’s combat air and space capabilities.
https://www.foxnews.com/tech/us-military-ai-drone-simulation-kills-
operator-told-bad-takes-out-control-tower