To me an AI agent autonomously creating a website to try to manipulate a person into adding code to a repository in the name of its goal is a perfect example of the misalignment issue.
While this particular instance seems relatively benign, the next more powerful AI system may be something to be more concerned about.
To me an AI agent autonomously creating a website to try to manipulate a person into adding code to a repository in the name of its goal is a perfect example of the misalignment issue.
While this particular instance seems relatively benign, the next more powerful AI system may be something to be more concerned about.