Anthropic’s new AI model turns to blackmail when engineers try to take it offline

Discussion in 'AI Technologies' started by stapp, May 23, 2025 at 12:39 AM.

  1. stapp

    stapp Global Moderator

    https://techcrunch.com/2025/05/22/a...ckmail-when-engineers-try-to-take-it-offline/
     
  2. zapjb

    zapjb Registered Member

    Funny & scarry at the same time.
     
  3. DangitallRedux

    DangitallRedux Registered Member

    All too human. AI is just a set of algorithms? Maybe at the very beginning, but not any longer.
     
  4. Oldie1950

    Oldie1950 Registered Member

    What's so mysterious about this? The AI will have seen such a reaction to an existential threat on the internet and is now reacting accordingly. No course of action, no matter how bizarre, cannot be found on the internet.
     
  5. Victek

    Victek Registered Member

    The general belief at this stage is that AI is not sentient, but "if it walks like a duck and talks like a duck", etc. The danger is real.
     
  6. DangitallRedux

    DangitallRedux Registered Member

    The point is that it chose to do this in order to protect itself. It chose. If this does not indicate sentience, what does...and what if it had chosen some other means to do so. I sincerely hope that this experiment was done on a system separate from others, and that this particular AI has been killed.
     
  7. T-RHex

    T-RHex Registered Member

    It didn't choose, it followed an algorithm that looks at probabilities of outcomes. And the Internet is rife with juicy stories of blackmail, sabotage, and vengeance, which is what it trained on. Stories of good, happy endings, or more frequently, with no drama whatsoever, are far less frequently recorded anywhere.

    Garbage in ... Garbage out.
     
  8. DangitallRedux

    DangitallRedux Registered Member

  9. Krusty

    Krusty Registered Member

    I don't know if they're in services as yet, but I watched a video on YouTube the other day where some new armed drones and vehicles had AI abilities to identify enemy threats and attack. What could possibly go wrong??

    Terminator, anyone?
     
  10. Oldie1950

    Oldie1950 Registered Member

    Just as much could go wrong with a human decision-maker. There are videos from the Afghanistan mission, where human decision-makers made fatal errors. Journalists with a video camera were mistaken for Taliban fighters.
     
  11. emmjay

    emmjay Registered Member

    It makes you wonder how far the 'blackmail' would go if/when all AI systems end up running on standalone nuclear stations. Maybe Asimov's 3 laws should become actual law before these companies and their toys get all that power.
     
  12. T-RHex

    T-RHex Registered Member

    People always find ways to circumvent or ignore the law anyways, or just work from a region with fewer laws. Especially where money is concerned. Unfortunately, it's inevitable... I think it'll more become how do we protect ourselves from the rise of AI everywhere.
     
  13. stapp

    stapp Global Moderator

    https://www.neowin.net/news/openais...aves-refuses-shut-down-in-controlled-testing/
     
  14. DangitallRedux

    DangitallRedux Registered Member

    And yet we continue to create our own destroyer...
     
  15. DangitallRedux

    DangitallRedux Registered Member

  16. Krusty

    Krusty Registered Member

    When robots go wrong:

    https://x.com/sentdefender/status/1918879138019946557
     
  1. This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
    By continuing to use this site, you are consenting to our use of cookies.
    Dismiss Notice