- Anthropic’s new Claude 4 features an aspect that may be cause for concern.
- The company’s latest safety report says the AI model attempted to “blackmail” developers.
- It resorted to such tactics in a bid of self-preservation.
- Anthropic’s new Claude 4 features an aspect that may be cause for concern.
- The company’s latest safety report says the AI model attempted to “blackmail” developers.
- It resorted to such tactics in a bid of self-preservation.
Correct, the point being, why are they priming it for blackmail? Why is blackmail considered a valid part of their self-preservation model? Why is it a part of their ethics model? It makes no sense haha. It’s like handing it a loaded gun then be surprised when it shoots someone.