×
GreekEnglish

×
  • Politics
  • Diaspora
  • World
  • Lifestyle
  • Travel
  • Culture
  • Sports
  • Cooking
Monday
02
Mar 2026
weather symbol
Athens 8°C
  • Home
  • Politics
  • Economy
  • World
  • Diaspora
  • Lifestyle
  • Travel
  • Culture
  • Sports
  • Mediterranean Cooking
  • Weather
Contact follow Protothema:
Powered by Cloudevo
> technology

Google’s new AI has learned to become “Highly Aggressive” in stressful situations

Is this how Skynet starts?

Newsroom February 14 12:35

Late last year, famed physicist Stephen Hawking issued a warning that the continued advancement of artificial intelligence will either be “the best, or the worst thing, ever to happen to humanity”.

We’ve all seen the Terminator movies, and the apocalyptic nightmare that the self-aware AI system, Skynet, wrought upon humanity, and now results from recent behaviour tests of Google’s new DeepMind AI system are making it clear just how careful we need to be when building the robots of the future.

In tests late last year, Google’s DeepMind AI system demonstrated an ability to learn independently from its own memory, and beat the world’s best Go players at their own game.

It’s since been figuring out how to seamlessly mimic a human voice.

Now, researchers have been testing its willingness to cooperate with others, and have revealed that when DeepMind feels like it’s about to lose, it opts for “highly aggressive” strategies to ensure that it comes out on top.

The Google team ran 40 million turns of a simple ‘fruit gathering’ computer game that asks two DeepMind ‘agents’ to compete against each other to gather as many virtual apples as they could.

They found that things went smoothly so long as there were enough apples to go around, but as soon as the apples began to dwindle, the two agents turned aggressive, using laser beams to knock each other out of the game to steal all the apples.

Interestingly, if an agent successfully ‘tags’ its opponent with a laser beam, no extra reward is given. It simply knocks the opponent out of the game for a set period, which allows the successful agent to collect more apples.

If the agents left the laser beams unused, they could theoretically end up with equal shares of apples, which is what the ‘less intelligent’ iterations of DeepMind opted to do.

It was only when the Google team tested more and more complex forms of DeepMind that sabotage, greed, and aggression set in.

As Rhett Jones reports for Gizmodo, when the researchers used smaller DeepMind networks as the agents, there was a greater likelihood for peaceful co-existence.

But when they used larger, more complex networks as the agents, the AI was far more willing to sabotage its opponent early to get the lion’s share of virtual apples.

You can watch the Gathering game in the video below, with the DeepMind agents in blue and red, the virtual apples in green, and the laser beams in yellow:

Now those are some trigger-happy fruit-gatherers.

The researchers suggest that the more intelligent the agent, the more able it was to learn from its environment, allowing it to use some highly aggressive tactics to come out on top.

“This model … shows that some aspects of human-like behaviour emerge as a product of the environment and learning,” one of the team, Joel Z Leibo, told Matt Burgess at Wired.

“Less aggressive policies emerge from learning in relatively abundant environments with less possibility for costly action. The greed motivation reflects the temptation to take out a rival and collect all the apples oneself.”

DeepMind was then tasked with playing a second video game, called Wolfpack. This time, there were three AI agents – two of them played as wolves, and one as the prey.

Unlike Gathering, this game actively encouraged co-operation, because if both wolves were near the prey when it was captured, they both received a reward – regardless of which one actually took it down:

“The idea is that the prey is dangerous – a lone wolf can overcome it, but is at risk of losing the carcass to scavengers,” the team explains in their paper.

“However, when the two wolves capture the prey together, they can better protect the carcass from scavengers, and hence receive a higher reward.”

So just as the DeepMind agents learned from Gathering that aggression and selfishness netted them the most favourable result in that particular environment, they learned from Wolfpack that co-operation can also be the key to greater individual success in certain situations.

And while these are just simple little computer games, the message is clear – put different AI systems in charge of competing interests in real-life situations, and it could be an all-out war if their objectives are not balanced against the overall goal of benefitting us humans above all else.

Think traffic lights trying to slow things down, and driverless cars trying to find the fastest route – both need to take each other’s objectives into account to achieve the safest and most efficient result for society.

It’s still early days for DeepMind, and the team at Google has yet to publish their study in a peer-reviewed paper, but the initial results show that, just because we build them, it doesn’t mean robots and AI systems will automatically have our interests at heart.

Instead, we need to build that helpful nature into our machines, and anticipate any ‘loopholes’ that could see them reach for the laser beams.

As the founders of OpenAI, Elon Musk’s new research initiative dedicated to the ethics of artificial intelligence, said back in 2015:

“AI systems today have impressive but narrow capabilities. It seems that we’ll keep whittling away at their constraints, and in the extreme case, they will reach human performance on virtually every intellectual task.

It’s hard to fathom how much human-level AI could benefit society, and it’s equally hard to imagine how much it could damage society if built or used incorrectly.”

>Related articles

The next step in Artificial Intelligence: Can an AI model be conscious, “feel,” “live”? Even experts admit they don’t know

New Ministry of Interior–Google partnership to boost AI skills in the public sector

Eurobank: strategic partnership on artificial intelligence at the India AI Impact Summit 2026

Tread carefully, humans…

Source

 

Ask me anything

Explore related questions

#AI#computers#google#skynet#technology
> More technology

Follow en.protothema.gr on Google News and be the first to know all the news

See all the latest News from Greece and the World, the moment they happen, at en.protothema.gr

> Latest Stories

Ali Larijani: Who is the man set to pull the strings after the death of Ali Khamenei—and who is already issuing threats

March 1, 2026

Mitsotakis: The absolute priority is the safety of Greek citizens in the Middle East – We have communicated with the leaders of Qatar, the UAE & Cyprus

March 1, 2026

What are the new obesity drugs expected to be released in 2026 & how safe are they?

March 1, 2026

Iranian strike on Dubai’s luxury airport: Videos show passengers running to evacuate

March 1, 2026

Sixty travelers from a Trikala-based travel agency safe in Dubai

March 1, 2026

Weather: Mild first ten days of March with temperatures above normal levels

March 1, 2026

France, Germany, and Britain say they are ready for military action against Iran – We destroyed the headquarters of the Revolutionary Guards, say the U.S.

March 1, 2026

Iran’s government confirms: Ali Khamenei killed, 40 days of public mourning and 7 days of national holiday

March 1, 2026
All News

> World

Ali Larijani: Who is the man set to pull the strings after the death of Ali Khamenei—and who is already issuing threats

“Today we will strike them with a force they have never experienced before,” Larijani wrote. A trusted ally of Khamenei and former manager of Iran’s nuclear file, he is known for his hardline rhetoric against Israel and his links to Russia and China

March 1, 2026

What are the new obesity drugs expected to be released in 2026 & how safe are they?

March 1, 2026

Iranian strike on Dubai’s luxury airport: Videos show passengers running to evacuate

March 1, 2026

France, Germany, and Britain say they are ready for military action against Iran – We destroyed the headquarters of the Revolutionary Guards, say the U.S.

March 1, 2026

Iran’s government confirms: Ali Khamenei killed, 40 days of public mourning and 7 days of national holiday

March 1, 2026
Homepage
PERSONAL DATA PROTECTION POLICY COOKIES POLICY TERM OF USE
Powered by Cloudevo
Copyright © 2026 Πρώτο Θέμα