Claude 3 notices when a sentence about pizza toppings doesn't fit with its surrounding text. Whole internet including Tim Sweeney and Margaret Mitchell concludes that it's probably self-aware now.
me when the machine specifically designed to pass the turing test passes the turing test
If you can design a model that spits out self-aware-sounding things after not having been trained on a large corpus of human text, then I'll bite. Until then, it's crazy that anybody who knows anything about how current models are trained accepts the idea that it's anything other than a stochastic parrot.
Despite the hype, from my admittedly limited experience I haven't seen a chatbot that is anywhere near passing the turing test. It can seemingly fool people who want to be fooled but throw some non-sequiturs or anything cryptic and context-dependent at it and it will fail miserably.
I don't think a computer program has passed the Turing test without interpreting the rules in a very lax way and heavily stacking the deck in the bot's favor.
I'd be impressed if a machine does something hard even if the machine is specifically designed to do that. Something like proving the Riemann hypothesis or actually passing an honest version of Turing test.
As somebody said, and im loosely paraphrasing here, most of the intelligent work done by ai is done by the person interpreting what the ai actually said.
A bit like a tarot reading. (but even those have quite a bit of structure).
Which bothers me a bit is that people look at this and go 'it is testing me' and never seem to notice that LLMs don't really seem to ask questions, sure sometimes there are related questions to the setup of the LLM, like the 'why do you want to buy a gpu from me YudAi' thing. But it never seems curious in the other side as a person. Hell, it won't even ask you about the relationship with your mother like earlier AIs would. But they do see signs of meta progression where the AI is doing 4d level chess style things.
As somebody said, and im loosely paraphrasing here, most of the intelligent work done by ai is done by the person interpreting what the ai actually said.
This is an absolutely profound take that I hadn't seen before; thank you.
It prob came from a few of the fired from various ai places ai ethicists who actually worry about real world problems like the racism/bias from ai systems btw.
The article itself also mentions ideas like this a lot btw. This: "Fan describes how reinforcement learning through human feedback (RLHF), which uses human feedback to condition the outputs of AI models, might come into play. "It's not too different from asking GPT-4 'are you self-conscious' and it gives you a sophisticated answer,"" is the same idea with extra steps.
Well, the LLM was prompted to find the odd one. Which I consider a (relatively) easy one. Reading the headline, I thought that the LLM was able to point this out by itself, like "Excuse me, but you had one sentence about pizza toppings in your text about programming. Was that intended to be there for some reason, or just a mistaken CTRL-V?"
I'm confused how this is even supposed to demonstrating "metacognition" or whatever? It's not discussing its own thought process or demonstrating awareness of its own internal state, it just said "this sentence might have been added to see if I was paying attention." Am I missing something here? Is it just that it said "I... paying attention"?
This is a thing humans already do sometimes in real life and discuss -- when I was in middle school, I'd sometimes put the word "banana" randomly into the middle of my essays to see if the teacher noticed -- so pardon me if I assume the LLM is doing this by the same means it does literally everything else, i.e. mimicking a human phrasing about a situation that occurred, rather than suddenly developing radical new capabilities that it has never demonstrated before even in situations where those would be useful.
Yup, it's 100% repeating the kind of cliché that is appropriate to the situation. Which is what the machine is designed to do. This business is getting stupider and more desperate by the day.
I'm also going from the other post which said that this is all simply 90's era algorithms scaled up. But using that form of neural net stuff, wouldn't we expect minor mistakes like this from time to time? Neural net does strange unexplained thing suddenly is an ancient tale.
it doesn't even have to do the 'are you paying attention' thing (which shows so many levels of awareness it is weird (but I guess they are just saying it is copying the test idea back at us (which is parroting, not cognition but whatever))) because it is aware, it could just be an error.
The problem is that whether or not an AI is self-aware isn't a technical question - it's a philosophical one.
And our current blinkered focus on STEM and only STEM has made it so that many (most?) of those most involved in AI R&D are woefully underequipped to make a sound judgment on such a matter.
And our current blinkered focus on STEM and only STEM has made it so that many (most?) of those most involved in AI R&D are woefully underequipped to make a sound judgment on such a matter.
who would be equipped to make a sound judgment on such a matter?
The Anthropic researcher didn’t have this take. They were just commenting that it was interesting. It’s everyone else who seemed to think it meant something more.
Doesn’t it just indicate that the concept of needle-in-a-haystack testing is included in the training set?
I think some here are grossly overestimating average human capacity. There are many humans that have difficulty discerning the context of a statement based on their experiences aka examples.
This isn't AGI, but in another couple years at this pace, it's coming. Not necessarily because it is some higher mind, but because the metric for AGI is can it perform all the tasks our minds can at our level. Not necessarily Stephen Fry or Albert Einstein, just as well as a median asshole. Have you met us?
We aren't all that, and most of us spend most of our time on a script, sapience must be exercised, many do many don't, and isn't necessary for what we will abuse these for. It would probably be kinder to restrict discussion of such topics from memory when this matures. Even humans have great difficulty wrestling with them, to the point of depression and existential dread.
nah, if anything this A.I. craze has made me appreciate how incredibly smart even the supposedly dimmest of humans are. we can use language of our own volition, to create meaning. in fact we frigging invented it!!! we're just bloody amazing, to hell with misanthropy.
We aren’t all that, and most of us spend most of our time on a script, sapience must be exercised, many do many don’t, and isn’t necessary for what we will abuse these for. It would probably be kinder to restrict discussion of such topics from memory when this matures. Even humans have great difficulty wrestling with them, to the point of depression and existential dread.
holy fuck please log off and go to therapy. I’m not fucking around. if this is actually how you see yourself and others, you are robbing yourself of the depth of the human experience by not seeking help.
the person there just commented on the average human's capacity for reasoning (not all humans, just the average one), and, in all fairness, they're sort of right, I think
don't just think of your friends and family, but about all humans. think about what makes it in the news and then how many things don't make it. religious nuts stoning people for whatever reason, gang sexual assault in the street in certain areas of the world, people showing up in ERs with weird stuff up their back ends, or finding unexploded ordnance from wars past and deciding the best course of action would be to smash it with a hammer or drill into it. this is all of course in addition to the pressing issues nowadays which do also seem to come from a place of not exercising sapience.
and for the less extreme cases, I do think the original commenter here is correct in saying people do tend to follow scripts and glide through life.
When people say stuff like this it always makes me wonder "what pace, exactly?" Truthfully, I feel like hearing someone say "well, generative AI is such a fast-moving field" at this point is enough on its own to ping my BS detector.
Maybe it was forgivable to say it in May 2023, but at this point it definitely feels like progress has slowed down/leveled off. AI doesn't really seem to me to be significantly more capable than it was a year ago -- I guess OpenAI can generate videos now, but it's been almost a year since "will smith eating spaghetti," so...
So... the idea here is that OpenAI and friends are gonna charge you N bucks a month so you can have chat conversations with the average internet user? Spoiler alert: that service is already free.
For $150 they save you the inconvenience of finding the style of twit you wish to interact with, and will dress up to whatever twit your heart desires!
In the beginning.. I can’t wait to see what happens to their pricing when they believe they’ve locked enough in and shift from vc subsidy to actual customer-carried charge. Bet it’s gonna be real popular..