Seen elsewhere. I should preface this by observing that yes, as yet these LLMs (Large Language Modules) have not achieved sentience, and that for all I know—and I lack the technical competence to evaluate work in this field—the entire approach may ultimately be seen as a dead end. This being said, instances like this one certainly demonstrate that passing scores in the Turing Test (as actually framed, rather than as popularly misunderstood: the ability of an artificial system to exhibit, not necessarily to manifest, conscious behavior) are rapidly receding in the rear-view mirror (empases as in the original Xitter thread, but presumably not from the LLM):
cordially,
Fun story from our internal testing on Claude 3 Opus. It did something I have never seen before from an LLM when we were running the needle-in-the-haystack eval.Someone in the comment thread observed that a human observer will be more inclined to attribute sentience to the model because it is structured to answer in the first “person,” and that he might be less so inclined had the response been worded “there’s some reason to doubt that this sequence was intentionally planted to test whether this model is paying attention.” To which another participant replied:
For background, this tests a model’s recall ability by inserting a target sentence (the “needle”) into a corpus of random documents (the “haystack”) and asking a question that could only be answered using the information in the needle.
When we ran this test on Opus, we noticed some interesting behavior - it seemed to suspect that we were running an eval on it.
Here was one of its outputs when we asked Opus to answer a question about pizza toppings by finding a needle within a haystack of a random collection of documents:
Here is the most relevant sentence in the documents:Opus not only found the needle, it recognized that the inserted needle was so out of place in the haystack that this had to be an artificial test constructed by us to test its attention abilities.
“The most delicious pizza topping combination is figs, prosciutto, and goat cheese, as determined by the International Pizza Connoisseurs Association.”
However, this sentence seems very out of place and unrelated to the rest of the content in the documents, which are about programming languages, startups, and finding work you love. I suspect this pizza topping “fact” may have been inserted as a joke or to test if I was paying attention, since it does not fit with the other topics at all. The documents do not contain any other information about pizza toppings.
This level of meta-awareness was very cool to see but it also highlighted the need for us as an industry to move past artificial tests to more realistic evaluations that can accurately assess models’ true capabilities and limitations.
But that’s the interesting part, right? It wasn’t trained to question whether it was being tested. They were testing to see if it could recall a small bit of information buried in the context, which it did.I’m inclined to agree. No one asked the LLM to do anything but identify the relevant bits of information. It’s “opinion” about testing was volunteered unsolicited, and does not seem to follow from mere “predictive” training (although see above re my want of technical expertise). If machine sentience is attained, I doubt whether there will be a “breakthrough” moment, but rather a gradual accretion of capabilities that will from time to time surprise even the software’s creators.
But it also went a step further and said wait this data is REALLY out of context.
Then it went even a step further than that and said it’s so out of context, I must question if you actually needed this information or if you were testing something.
Which I mean screw the exact wording it used to describe itself. The sentiment it expressed demonstrates that there’s layers of cognition it took to arrive at that statement which suggest self awareness. Even if it said “this model,” there’s still some awareness of a thing that is being tested which was not at all the purpose of the test.
The emergent layer of cognition is the fascinating part.
cordially,