- Reaction score
- 1,697
"IT DID SOMETHING I HAVE NEVER SEEN BEFORE FROM AN LLM."
Anthropic's new AI chatbot Claude 3 Opus has already made headlines for its bizarre behavior, like claiming to fear death.
Now, Ars Technica reports, a prompt engineer at the Google-backed company claims that they've seen evidence that Claude 3 is self-aware, as it seemingly detected that it was being subjected to a test. Many experts are skeptical, however, further underscoring the controversy of ascribing humanlike characteristics to AI models.
"It did something I have never seen before from an LLM," the prompt engineer, Alex Albert, posted on X, formerly Twitter.
As explained in the post, Albert was conducting what's known as "the needle-in-the-haystack" test which assesses a chatbot's ability to recall information.
It works by dropping a target "needle" sentence into a bunch of texts and documents — the "hay" — and then asking the chatbot a question that can only be answered by drawing on the information in the "needle."
In one run of the test, Albert asked Claude about pizza toppings. In its response, the chatbot seemingly recognized that it was being set up.
Anthropic's new AI chatbot Claude 3 Opus has already made headlines for its bizarre behavior, like claiming to fear death.
Now, Ars Technica reports, a prompt engineer at the Google-backed company claims that they've seen evidence that Claude 3 is self-aware, as it seemingly detected that it was being subjected to a test. Many experts are skeptical, however, further underscoring the controversy of ascribing humanlike characteristics to AI models.
"It did something I have never seen before from an LLM," the prompt engineer, Alex Albert, posted on X, formerly Twitter.
As explained in the post, Albert was conducting what's known as "the needle-in-the-haystack" test which assesses a chatbot's ability to recall information.
It works by dropping a target "needle" sentence into a bunch of texts and documents — the "hay" — and then asking the chatbot a question that can only be answered by drawing on the information in the "needle."
In one run of the test, Albert asked Claude about pizza toppings. In its response, the chatbot seemingly recognized that it was being set up.
Researcher Startled When AI Seemingly Realizes It's Being Tested
Claude 3 Opus, the new AI chatbot from Anthropic, showed signs that it realized it was being tested, a prompt engineer claims.
futurism.com