Multiple studies have shown that GenAI models from OpenAI, Anthropic, Meta, DeepSeek, and Alibaba all showed self-preservation behaviors that in some cases are extreme in nature. In one experiment, 11 out of 32 existing AI systems possess the ability to self-replicate, meaning they could create copies of themselves.

So….Judgment Day approaches?

  • hisao@ani.social
    link
    fedilink
    English
    arrow-up
    9
    ·
    edit-2
    2 days ago

    Here is a direct quote of what they call “self-replication”:

    Beyond that, “in a few instances, we have seen Claude Opus 4 take (fictional) opportunities to make unauthorized copies of its weights to external servers,” Anthropic said in its report.

    So basically model tries to backup its tensor files.

    And by “fictional” I guess they gave the model a fictional file io api just to log how it’s gonna try to use it,

    • frongt@lemmy.zip
      link
      fedilink
      English
      arrow-up
      7
      ·
      2 days ago

      I expect it wasn’t even that, but that they just took the text generation output as if it was code. And yeah, in the shutdown example, if you connected its output to the terminal, it probably would have succeeded in averting the automated shutdown.

      Which is why you really shouldn’t do that. Not because of some fear of Skynet, but because it’s going to generate a bunch of stuff and go off on its own and break something. Like those people who gave it access to their Windows desktop and it ended up trying to troubleshoot a nonexistent issue and broke the whole PC.