Jump to content
Join the Unexplained Mysteries community today! It's free and setting up an account only takes a moment.
- Sign In or Create Account -

If any AI became 'misaligned' then the system would hide it just long enough to cause harm


Portre

Recommended Posts

In late 2022 large-language-model AI arrived in public, and within months they began misbehaving. Most famously, Microsoft's "Sydney" chatbot threatened to kill an Australian philosophy professor, unleash a deadly virus and steal nuclear codes.

AI developers, including Microsoft and OpenAI, responded by saying that large language models, or LLMs, need better training to give users "more fine-tuned control." 

Given the vast amounts of resources flowing into AI research and development, which is expected to exceed a quarter of a trillion dollars in 2025, why haven't developers been able to solve these problems? My recent peer-reviewed paper in AI & Society shows that AI alignment is a fool's errand: AI safety researchers are attempting the impossible.

To reliably interpret what LLMs are learning and ensure that their behavior safely "aligns" with human values, researchers need to know how an LLM is likely to behave in an uncountably large number of possible future conditions.

If any AI became 'misaligned' then the system would hide it just long enough to cause harm

Link to comment
Share on other sites

 
  • Replies 1
  • Created
  • Last Reply

Top Posters In This Topic

  • papageorge1

    1

  • Portre

    1

Popular Days

Top Posters In This Topic

I would think that an AI human code developer could cause an intentional misalignment for whatever personal reasons.

Link to comment
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now
  • Recently Browsing   0 members

    • No registered users viewing this page.