back to article Apple AI boffins puncture AGI hype as reasoning models flail on complex planning

If you are betting on AGI – artificial general intelligence, the point at which AI models rival human cognition – showing up next year, you may want to adjust your timeline. Apple AI researchers have found that the "thinking" ability of so-called "large reasoning models" collapses when things get complicated. The authors' …

  1. Gene Cash Silver badge
    Holmes

    AI 'thinking' may just be an illusion

    I think I might speak for the majority of commentards when I say:

    WOT!?? NO, REALLY?? What have we been yelling for years?

    And the "no shit, Sherlock" icon has never been more appropriate.

    1. ecofeco Silver badge

      Re: AI 'thinking' may just be an illusion

      Funny though, many humans have the same problem.

      Thinking being a mere illusion, that is.

      1. Anonymous Coward
        Anonymous Coward

        Re: AI 'thinking' may just be an illusion

        Most humans stopped thinking or learning at round about 15 years old.

    2. Steve Davies 3 Silver badge
      Childcatcher

      Re: AI 'thinking' may just be an illusion

      But they naysayers will say loudly that

      THIS IS APPLE who can't get their Apple Intelligence to work (and SIRI after all these years)

      so the sensible original statement will get ignored.

      I hope this "AI with everything" hype goes away soon. But, it is 'hip', 'cool', 'with it' etc etc so it won't.

      1. breakfast Silver badge
        Terminator

        Re: AI 'thinking' may just be an illusion

        I don't think it's hip or cool, I think it is a desperate last resort for Venture Capitalists who have been throwing money at the blockchain, the metaverse and anything else they imagine might offer exponential profit growth over the last ten years and not seen much return.

        At this point they're all-in on AI and although it is less of an obvious bubble than NFTS were - there probably is a baby in this bathwater somewhere - there is no way it can ever be as profitable as it would need to be to justify the investment. They have to keep hyping it because papers like this one are going to keep coming and if they don't get out in time they're going to get their fingers burnt in at least a 2000 way even if we don't go full 2008.

        Expect to see more nonsensical hype and more mechanical turks as the road ahead gets bumpier, and in the meantime hope your pension isn't too heavily invested in AI.

    3. Anonymous Coward
      Anonymous Coward

      Re: AI 'thinking' may just be an illusion

      AI coming real soon.

      ChatGPT SEO here now, and little improvement on prior search/reporting services.

      1. Anonymous Coward
        Anonymous Coward

        Re: AI 'thinking' may just be an illusion

        Maybe they should pair AI with fusion reactors, they can be just coming together. Maybe, just maybe, one day ...

  2. Filippo Silver badge

    We don't really know what "intelligence" is, but, ultimately, there are only two possibilities. Either it is nothing more than emergent behavior from big statistics with a lot of recursion, or it is more than that.

    In the first case, we might get AGI comparatively soon. It might take some new development, such as embodied AI, or it might take finding just the right architecture. But the path would be roughly correct and ought to eventually get there.

    In the second case, we are not getting AGI any time soon, and all the effort spent on LLMs can only run in circles while getting nowhere. If we're very lucky, we might learn why it's not the answer.

    Both possibilities are philosophically interesting, and I'm looking forward to the outcome.

    1. David 164

      I think the first one scares a lot of people who been studying AI for decades and getting into highly complex discussions about AI, when in reality it was fairly simply to do, we just didn't have the data or the hardware and a few programming concepts to do it until we did.

      1. Anonymous Coward
        Anonymous Coward

        No.

        But it wasn't and we didn't, at least not yet, we're not close. We're not in the same ballpark. We're not in the same country. We're not in the same solar system. We may be in the same galaxy? It's hard to know yhet.

      2. Anonymous Coward
        Anonymous Coward

        Being related to someone who has been studying AI for decades, I know for certain that what is currently called "AI" isn't, and never will be, intelligent. While it does a decent job of mimicking a human response, it doesn't really learn, and definitely cannot "understand" anything. It's glorified pattern-matching, an advanced version of Eliza that's been fed a lot of input.

        No, real AI isn't simple to do, we don't have it, and we likely won't anytime in our lifetimes.

    2. Brewster's Angle Grinder Silver badge

      I think there's a third possibility: that intelligence is a behaviour that emerges from statistics, etc... but that LLMs, while a step towards it, aren't a suitable architecture and so nothing close to AGI will arise "comparatively soon"; it will need another architectural break through.

      And, frankly, LLMs are enough to be getting on with. If their "intelligence" is capped that will give us time to adjust before anything close to AGI emerges..

      1. Anonymous Coward
        Anonymous Coward

        Did you read the article? LLMs and LGMs aren't even close to intelligent. They're more than dumb, they're dumbness-generators. They make you dumber if you use them.

        There is no intelligence to cap. If anything they are capping their users' intelligence.

    3. brainwrong

      'We don't really know what "intelligence" is'

      When I first heard the hype about LLM's, I thought it would probably challenge peoples idea of what intelligence is.

      My take is that intelligence has evolved over *many* generations of trial and error for the purpose of operating the creatures that possess it, such that they may feed themselves and reproduce in the face of competition for common resources. Artificial intelligence has no purpose, so what's the impetus for it to be intelligent? The fact that so many supposedly intelligent people think that LLM's are intelligent leads me to doubt the capabilities of real human intelligence, which did not evolve for the modern world.

      1. Doctor Syntax Silver badge

        "such that they may feed themselves and reproduce in the face of competition for common resources"

        It's a bit more than that. Reading and commenting here achieves neither of those goals. So either that is not intelligence at work (a plausibly arguable PoV) or intelligence acquires a whole lot of additional goal for its own pleasure.

        1. ChoHag Silver badge

          > Reading and commenting here achieves neither of those goals.

          I need to work in order to "compete for resources" and in order to work most effectively I need to occasionally rest and laugh at commentards, for which I come here. Doesn't help much with the reproduction though.

        2. brainwrong

          We evolved intelligence in order to communicate complex ideas to work together in groups, make and use tools, gain knowledge and experience, plan ahead, understand cause and effect etc. It worked for cavemen, and it still works for us today. Reading and commenting here is exercising your intelligence because you don't need to use it to plan your next crop planting or mammoth hunt, and doing nothing with it is dull.

          1. I ain't Spartacus Gold badge

            brainwrong,

            We didn't evolve intelligence to achieve any kind of a goal. There wasn't some committee in a cave somewhere who went on UggBay to order up a crate of intelligence pills.

            We evolved intelligence, and it enabled us to successfully compete with other species and eventually invent digital watches.

            Wheresas dolphins evolved intelligence in order to lark about in the water having a good time, and eating a lot of fish.

            Or even, possibly not. Maybe the Neanderthals were more intelligent than us, but got outcompeted anyway? Who knows. There's a lot of luck going on in evolution as well, so I suspect its foolish to try to come up with any definitive conclusions. Unless the alien Xenu gave us intelligence, in order to build him (her/it?) better pyramids?

            1. brainwrong

              "We evolved intelligence, and it enabled us to successfully compete with other species"

              Thats wot i sed.

      2. Don Jefe

        The site lesswrong.com has some excellent essays and dialogues about the nature of thought and intelligence as it relates to LLM/AI. It’s well moderated and thoughtful in a way that’s hard to find online anymore.

        There are two broad camps. One sees thought and intelligence as purely technical and subject to modeling. The other sees them as more along the lines you’re touching on.

        At the risk of sounding old, AGI is incapable of synergistic reasoning. It may recognize that candies on a birthday cake is an important tradition, but doesn’t recognize why a single candle in a croissant is equally valid or why you don’t put 70 candles on the cake.

        I’m in broad agreement with your take on it. AI has no inherent purpose, thus cannot reason unreasonably like humans can. I reckon it’ll collapse in on itself or kill us all.

      3. Anonymous Coward
        Anonymous Coward

        "what's the impetus for it to be intelligent?"

        Survival. Same as us. The big question is why do we care about survival. Even if we can die without pain we desperately want to live. Why do we want to live? Because if we didn't we wouldn't have got here. Evolution or God; I have no idea!

    4. DS999 Silver badge

      LLMs are just

      Extremely well read toddlers who will never grow up.

      They are useful, but there's no way that word (token) prediction is the path to AGI.

    5. Anonymous Coward
      Anonymous Coward

      There's something odd about us. Stuff goes on in our head that is unique on this planet. The level of self-awareness is extreme in many of us and the ability of some to make phenomenal leaps in imagination, abstraction and discovery. We build complex models in our head. On the other hand I have come to the conclusion that the majority just think what the TV tells them to think; NPCs?

      1. brainwrong

        People can't know about everything, they trust that people who do will be on the telly to help them think things. The modern world seems to expect us to have opinions on all sorts of things that most people know nothing about.

  3. Omnipresent Silver badge

    Encyclopædia Britannica needed for verification

    Intelligence is the probabilities vs real world experience. The probabilities give you a range of outcomes, and real world experience tells you what the most likely outcome is. Ai can do probabilities, but needs your real world experience to put it to good use.

    Quantum mechanics is also a game of probabilities. You have to make a real world measurement to verify your best guess. In the case of the internet there are no more places of truth left to verify against.

    You are playing a game of Schrödinger's cat without any measurements taken therefor, you cannot trust the AI to be truthful, or even accurate.

    1. Anonymous Coward
      Anonymous Coward

      Re: Encyclopædia Britannica needed for verification

      It's more probability x reward (or negative reward). We do that as well as LLMs. But we have something else that we don't understand fully.

  4. Neil Barnes Silver badge

    Building non-artificial intelligence

    Is both more fun and cheaper than AI, even taking into account the eighteen or so years it needs to mature to usefulness.

    1. Anonymous Coward
      Anonymous Coward

      Re: Building non-artificial intelligence

      I totally disagree!

      It brings huge expense, frustration and anxiety but we have a high compulsion to replicate and nurture thankfully. It's inherent in existence.

      1. Anonymous Coward
        Anonymous Coward

        Re: Building non-artificial intelligence

        I don't entirely disagree - the initial, er, creation is definitely a lot of fun.

  5. Anonymous Coward
    Anonymous Coward

    Fairly basic tasks

    Are a fairy major struggle for most AI in the wild at the moment. Anything bigger sounds a stretch,

  6. Anonymous Coward
    Anonymous Coward

    Truth regarding 'AI' !!!!

    The current problem with AI and/or AGI is not one of architecture or understanding what intelligence really is !!!

    The problem is one of 'Truth' ...

    The people who are working on creating 'AI' are being 'economical' with the 'Truth'.

    'AI' can masquerade as 'Intelligent' BUT it is not 'TRUE' by any realistic measure.

    It may help to 'Sell' whatever 'AI' is BUT it is NOT truely 'Intelligent' and never can be ... there is something missing.

    Sell 'AI' as a flawed agent to assist in knowledge search BUT don't hide the fact that it still 'Lies' [Hallucinations] and unless the knowledge-set it is based on is carefully curated, you still need to take the answers given with a LARGE pinch of salt !!!

    Greed is driving the current 'AI' mania ... eventually the funds to keep going WILL run out !!!

    A new and different approach is needed ... current 'AI' is a dead-end !!!

    That is the 'Truth' !!!

    :)

    1. DrewPH
      Coffee/keyboard

      Re: Truth regarding 'AI' !!!!

      Being economical with the truth is one thing, however being economical with exclamation marks might be a useful skill for you to acquire.

      1. Anonymous Coward
        Anonymous Coward

        Re: Truth regarding 'AI' !!!!

        So sorry BUT I have a large excess of !!!! in a warehouse going mouldy ... so I am using them !!!

        [Also got lots of ... & :) ... and a sticky CAPS key.]

        :)

    2. Anonymous Coward
      Anonymous Coward

      Re: Truth regarding 'AI' !!!!

      It's also "taught" with bias which we call governance.

    3. David 164

      Re: Truth regarding 'AI' !!!!

      Human lie and helluciate all the time. That has nothing to do with whether these models are intelligent or not.

      Lies and helluciations are a problems because they are for profit companies working towards commercial applications where companies don't want their product lying to customers, unless it a company producing sex robots, then you know they will need them to lie.

  7. HuBo Silver badge
    Pint

    Very nice

    I love that the boffins developed their own independent benchmark (as in their previous GSM-Symbolic benchmark) to prevent "benchmark data contamination". Too many AI benchmarks are affected by this sort of contamination, like the o1 Pro YT eval on the William Lowell Putnam Mathematics Competition.

    The European Commission's Joint Research Center paper (linked under "aren't all that great to begin with") nicely notes for example that (from Narayanan and Kapoor [2023a]) GPT4 "could regularly solve benchmark problems classified as easy - as long as the problems had been added before 5th September 2021. For problems added later, GPT4 could not get a single question right, suggesting that the model had memorised questions and answers".

    Similarly, using the "Black Box test" method, Microsoft folks found that 7 multilingual benchmarks had been swallowed up whole by 7 LLMs (cf. Contamination Report for Multilingual Benchmarks) and another group of boffins Investigating Data Contamination in Modern Benchmarks for LLMs used "Testset Slot Guessing" to find LLMs outputting verbatim the missing (right or wrong) multiple-choice options in MMLU benchmark data more than 50% of the time. Louis Hunt posted about similar contamination, with GSM8K data, on his linkedin page.

    Very nice article!

  8. Anonymous Coward
    Anonymous Coward

    The Earth sort of isn't the centre of the Universe

    The utterances of Apple AI Boffins reminded me of the astronomers in Galileo's day who were well aware that the Ptolemaic Earth centred cosmology that had the imprimatur of the Church wasn't exactly supported by observation. They would have been understandably circumspect in advertising this knowledge not wanting to become a kebab/giro at the next Bruno barbeque.

    Equally if you were an AI Boffins you would hardly undermine the tracks your particular gravy train runs on.

    Much of the deception of contemporary AI flows from the ineffability of human intelligence, thought and consciousness etc. We (some of us anyway) can experience these (and more) while being able to state some the properties we understand to be essential to each we really cannot provide anything like a comprehensive definition of intelligence etc or even whether one is an illusion or byproduct created by the others. Philosophers still get to dine out on these inponderables - that is if they can get their cutlery sorted.

    Anyone who believes that the ability to apply a consistent system deductive logic to reasoning is a prerequisite for intelligence neglects the observation that the vast majority of people of perfectly normal intelligence would fail using that criterion.

    Last century there were enough daft AI researchers that were convinced you could create intelligence using Prolog (or another implementation of causal form logic) and expert systems. Expert systems were useful and are routinely used today although often with a little AI badge engineering.

    I recently completed an "AI" online form for which the applicant had to fullfill an number of quite specific criteria - a task which screams "decision tree" - this web form clearly had a last century expert system behind it driving the JavaScript/CSS - worked efficiently and effectively but in no sense was intelligent.

    AI/LLM will eventually find its niche but will be nowhere as pervasive as would be required, given the resources so far invested, to break even let alone make a profit. People still grow tulips but aren't generally billionaires. ;)

  9. Anonymous Coward
    Anonymous Coward

    Expert

    It took experts to realise this? Anyone using "AI" can see they are not actually intelligent and fail more than succeed in complex analytical tasks. We can argue why but the outcome is self evident. It may change though. Maybe it's because humans form generalised networks on a lot of data that we manage to distil down into something useful generally. But it did take quite a few evolutionary generations of brain development going back hundreds of millions of years. AI doesn't appear to learn on the job either. I can't discern much improvement when working on a project from the AI, I find I have to adapt to it and improve my prompts.

    1. David 164

      Re: Expert

      A lot of these companies have put in firewalls to prevent these models from learning from the public becuase the public teaches them to be little hitlers for a laugh.

    2. brainwrong

      Re: Expert

      "It took experts to realise this?"

      If I, a non expert, published something similar I doubt anyone would give a rats arse, because I'm not known or have any credentials. Experts get airtime, or hypertext transfer protocol time or whatever this is.

  10. Lee D Silver badge

    "AI all hype"

    News at 11. Coming up later, grass hue discovery, Pope religious revelation shocker, and do you know where your bears are sh**ting?

    1. Anonymous Coward
      Anonymous Coward

      Re: "AI all hype"

      My 'AI' has clearly stated that 'Green Bears are Catholic !!!' ... Well I never !!!

      :)

  11. Johnb89

    Philosophy

    I am a very large language model, therefore I am. Simples.

    Or not.

    1. adsp42

      Re: Philosophy

      I chat therefore I am

      (intelligent) well according to Turing.

      Come to think of it, AFI - Artificial Fake Intelligence

  12. Michael H.F. Wilkinson Silver badge
    Coat

    What people really want

    is a program like Reason! (the program developed by Gordon Way and company in Dirk Gently's Holistic Detective Agency). You give it input facts, and the conclusion you want to reach. It then creates a series of plausible-sounding steps to support your conclusion. I mean: who wants an AI program that says your conclusion is wrong? I am sure LLMs could be up to this task already.

    I'll be going. The one with "The Long Dark Teatime of the Soul" in the pocket

POST COMMENT House rules

Not a member of The Register? Create a new account here.

  • Enter your comment

  • Add an icon

Anonymous cowards cannot choose their icon

Other stories you might like