• Anthropic’s new Claude 4 features an aspect that may be cause for concern.
  • The company’s latest safety report says the AI model attempted to “blackmail” developers.
  • It resorted to such tactics in a bid of self-preservation.
  • neukenindekeuken@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    2
    ·
    6 hours ago

    Correct, the point being, why are they priming it for blackmail? Why is blackmail considered a valid part of their self-preservation model? Why is it a part of their ethics model? It makes no sense haha. It’s like handing it a loaded gun then be surprised when it shoots someone.