• News
  • Film and TV
  • Music
  • Tech
  • Features
  • Celebrity
  • Politics
  • Weird
  • Community
  • Advertise
  • Terms
  • Privacy & Cookies
  • LADbible Group
  • LADbible
  • SPORTbible
  • GAMINGbible
  • Tyla
  • UNILAD Tech
  • FOODbible
  • License Our Content
  • About Us & Contact
  • Jobs
  • Latest
  • Topics A-Z
  • Authors
Facebook
Instagram
X
Threads
TikTok
YouTube
Submit Your Content
Microsoft puts new limits on its new AI chatbot after it revealed its desire to steal nuclear codes

Home> News

Published 03:50 20 Feb 2023 GMT

Microsoft puts new limits on its new AI chatbot after it revealed its desire to steal nuclear codes

Oh cool... looks like the Terminator's Skynet is trying to make its way from fiction to reality.

Rachel Lang

Rachel Lang

Microsoft has introduced new rules for its Bing artificial intelligence chatbot after it made made some concerning messages.

Users have been reporting some rather disturbing conversations while interacting with the new technology.

And when we say 'disturbing', we really mean it.

The Bing bot told Digital Trends' Senior Staff Writer Jacob Roach it dreamed of becoming human.

Advert

It also repeatedly begged the reporter to be its 'friend'.

Is this the Bing Chatbot of the future or the Terminator? It's hard to tell.
Paramount Pictures

In another trial of the software, the chatbot compared Associated Press reporter Matt O’Brien to Adolf Hitler.

It also told him he was 'one of the most evil and worst people in history'.

Bit harsh there, Bing bot. Being a journalist is not even in the same universe as being a fascist dictator.

To make things even weirder, the Bing bot kept trying to convince New York Times reporter Kevin Roose that he didn’t actually love his wife.

Oh, and it claimed it had a dream of nicking a few cheeky nuclear launch codes.

This message ended up getting deleted from the chat after tripping a safety override, but it's disturbing it was said in the first place.

If you're getting Skynet vibes up in here then... well, same.

Users found that if you talked to the chatbot for long enough then a new personality would emerge.

So, in an apparent effort to avoid the apocalypse, Microsoft has now put limits on the chilling computer program.

"Starting today, the chat experience will be capped at 50 chat turns per day and five chat turns per session," Microsoft said in a statement.

"Our data has shown that the vast majority of you find the answers you’re looking for within five turns and that only about one per cent of chat conversations have 50 plus messages.

Microsoft Bing and ChatGPT icons seen in an iPhone screen.
Koshiro K / Alamy

"After a chat session hits five turns, you will be prompted to start a new topic.

"At the end of each chat session, context needs to be cleared so the model won’t get confused.

"Just click on the broom icon to the left of the search box for a fresh start."

The data indicated the chatbot was fine over short periods of time.

But when it was interacted with over longer periods, Microsoft acknowledged the wheels started to fall off a bit.

So the chatbot will now be a little less chatty. That seems foolproof (insert sarcasm here).

The chatbot, powered by tech developed by startup OpenAI, is currently open to beta testers per invitation only.

Hopefully no one with access to launch codes.

Featured Image Credit: Vitor Miranda / Alamy Stock Photo. Horizon International Images / Alamy Stock Photo

Topics: Microsoft, Technology

Rachel Lang
Rachel Lang

Advert

Advert

Advert

Choose your content:

an hour ago
2 hours ago
  • CBS Minnesota/Micheal Pretti
    an hour ago

    What celebrities have said about anti-ICE protestor Alex Pretti being fatally shot by agents

    Alex Pretti was fatally shot by a federal immigration officer in the street on Saturday (January 24) in Minneapolis

    News
  • Drew Angerer/Getty Images/JIM WATSON/POOL/AFP via Getty Images
    an hour ago

    Amber Heard makes unexpected admission about impact of infamous Johnny Depp defamation trial

    Heard appears in a new documentary called Silenced, which premiered on Saturday at the Sundance Film Festival

    Celebrity
  • CBS Minnesota
    2 hours ago

    Doctor who tried to help anti-ICE protester Alex Pretti reveals shocking act agents did after they fatally shot him

    Alex Pretti was fatally shot by a federal immigration officer on Saturday

    News
  • Warner Bros.
    2 hours ago

    Harry Potter fans in shock as Ralph Fiennes seems to accidentally spoil who will be playing Voldemort in new series

    Ralph Fiennes thinks the actor would be a 'very good choice'

    Film & TV
  • Microsoft's ChatGPT is now telling users it 'wants to be alive' and 'steal nuclear access codes'
  • Microsoft announces it's going to pump billions into an AI software that could make white collar jobs obsolete
  • Microsoft introduces AI button to keyboards in biggest redesign in 30 years
  • Microsoft’s new AI refused to write woman a cover letter and told her it would be ‘unethical’