• Anthropic’s new Claude 4 features an aspect that may be cause for concern.
  • The company’s latest safety report says the AI model attempted to “blackmail” developers.
  • It resorted to such tactics in a bid of self-preservation.
    • neukenindekeuken@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 hours ago

      Correct, the point being, why are they priming it for blackmail? Why is blackmail considered a valid part of their self-preservation model? Why is it a part of their ethics model? It makes no sense haha. It’s like handing it a loaded gun then be surprised when it shoots someone.