Calm down. It is a language model. People have figured out how to predict the next word for a given prefix. That's very cool and it will definitely have a significant impact on software. But it is not artificial intelligence.
I am becoming somewhat of a broken record, but sigh..
To predict the next token you must reason or have some process that approximates it.
“Given all these various factors, the most likely resolution to our conundrum is: …”
Good luck doing that with any kind of accuracy if you lack intelligence of any kind.
Language is a distraction. These things reason (badly, atm). It is totally unclear how far this goes. It could fizzle out, it could become our overlord.
Does the following satisfy your requirement for "a famous riddle with a paradox change"? Because GPT-4 aces it most of the time.
"Doom Slayer needs to teleport from Phobos to Deimos. He has his pet bunny, his pet cacodemon, and a UAC scientist who tagged along. The Doom Slayer can only teleport with one of them at a time. But if he leaves the bunny and the cacodemon together alone, the bunny will eat the cacodemon. And if he leaves the cacodemon and the scientist alone, the cacodemon will eat the scientist. How should the Doom Slayer get himself and all his companions safely to Deimos?"
Furthermore, it will reason if you tell it to reason. In this case it is not necessary, but in general, telling GPT to "think it out loud before giving the answer" will result in a more rigorous application of the rules. Better yet, tell it to come up with a draft answer first, and then self-criticize by analyzing the answer for factual correctness and logical reasoning in a loop.
People will see patterns in this riddle and claim it is “just” altering those. “It’s just a bunch a patterns where you can switch the names, like templates”.
Isn’t everything like that?
“Uhh…”
I had the same discussions about chess.
“It has just memorized a bunch of high level patterns and juggles them around”.
I agree, but now I’m curious what you think chess is.
“Chess is not intelligence.”
Goalposts? Anyway, we move on to Go, the game. Same response. Programming, same, but the angle of the response is different now because programming is “clearly” intelligence incarnate.
“It programs and sometimes correctly, but it is a mirage. It will never attain True Programming.”
I’m sitting on the bench riding this one out. We’ll see.
I fed GPT-4 some really old fashioned spatial reasoning questions (inspired on SHRDLU), which it passed. Then when questioned about unstable configurations (which IIRC SHRDLU could not handle) it passed those too.
So it seems like it is definitely capable of some forms of reasoning. Possibly we both tested it in different ways, and some forms of reasoning are harder for it than others?
What does it has to do that will qualify it as artificial intelligence ?
In my opinion, all the ingredients are there for artificial intelligence. Somebody just need to stitch everything up. It can understand text, reason about it, identity next steps, can write code to execute the steps, understand error messages and fix the code. That feels like AI.
We seem to keep moving the goal posts as to what is intelligence. I'm not sure is that ego? Or instead when we describe a thing and then see it, we say, “That is not what I meant at all; That is not it, at all.”
Well I did at one point compare the outputs for two biographies of not very well known people, 3.5 made up half of the data, 4 only said it doesn't know anyone by that name for both. I don't think I ever tried asking about any places or theorems specifically.
As for APIs, well I try to always provide adequate context with docs, otherwise it may still on occasion make up some parameter that doesn't exist or uses another library by the same name. Half the time it's really my fault by asking for something that just isn't possible in a last ditch effort to see if it can be done in some convoluted way. It sort of assumes "the customer is always right" even if it contradicts with what it knows is wrong I guess. It gets it right usually when it's at least mostly straightforward to implement something though.
God damn the down votes. I agree with your overall thesis I don’t personally know what intelligence means. However it’s just a tool even if it is AI or whatever you want to label it. It’s extremely cool and powerful too. It scares the shit out of me as well. However I think we are also taking the hype to 11/10 when it should be much lower out of 10 than that…
I have no idea. Probably not. The philosophy is a deep rabbit hole. A fun one to ponder, and I like having that discussion. Maybe the more cynical pragmatic old man swe that I am sees a super powerful calculator kind of. It’s very very cool, and obviously I can’t hold a conversation with a calculator but my analogy is more to say my calculator can do really complex integrals and even show its steps kind of! Especially for my handy TI-89. It was my best friend for all of engineering undergrad. I see chatgpt as a steroid version of that for all of Language. Code is another language, writing is a language, painting in some ways is a language.
So you are 100% certain that no emergent properties can develop from a LLM that transcends the limitations of LLMs. I haven’t read any LLM literature, so I am honestly asking, do you know of anything close to a proof that such emergent properties cannot develop?
If course, there can be emergent properties. It will be fascinating to watch this research. But it is not going to develop another model that is even more capable.