back to article AI agents get office tasks wrong around 70% of the time, and a lot of them aren't AI at all

IT consultancy Gartner predicts that more than 40 percent of agentic AI projects will be cancelled by the end of 2027 due to rising costs, unclear business value, or insufficient risk controls. That implies something like 60 percent of agentic AI projects would be retained, which is actually remarkable given that the rate of …

  1. Anonymous Coward
    Anonymous Coward

    Extra credit for partially completed tasks

    We're so desperate to encourage our future robotic overlords we're awarding them stars for effort?

    1. ChrisElvidge Silver badge

      Re: Extra credit for partially completed tasks

      Partially completed = UNcompleted.

      "Well, I poured half the concrete. Good enough?"

      1. Anonymous Coward
        Anonymous Coward

        Re: Extra credit for partially completed tasks

        Minimum Viable Product.

        Agile innit.

        We’ll fix it in the next sprint. Maybe. And A.I. don’t need no documentation.

        1. sev.monster Silver badge
          Boffin

          Re: Extra credit for partially completed tasks

          Write the documentation with ChatGPT.

        2. MyffyW Silver badge

          Re: Extra credit for partially completed tasks

          What George Harrison et al would call a "wilbury" or "we'll bury them in the mix" - the mix in the case being cement, sand and water.

      2. Jedit Silver badge
        Trollface

        "Well, I poured half the concrete. Good enough?"

        It's only up to the techbro's ankles instead of his shins, but it should suffice.

        1. Rich 11

          Re: "Well, I poured half the concrete. Good enough?"

          Turn him upside down and it'll come up to his nostrils.

    2. Scotthva5

      Re: Extra credit for partially completed tasks

      I wonder if they award a participation trophy just for being there.

      1. Eric Olson

        Re: Extra credit for partially completed tasks

        Those damn boomers are at it again!

        1. EricB123 Silver badge

          Re: Extra credit for partially completed tasks

          "Those damn boomers are at it again!"

          Damn right we are!

        2. Eric Olson

          Re: Extra credit for partially completed tasks

          Reminder: There's an adult handing out participation trophies to the kids, and it wasn't the decision of one person,

          Try gazing into the mirror once in a while

      2. Anonymous Coward
        Anonymous Coward

        Re: Extra credit for partially completed tasks

        It's DEIjà-vu all over again.

    3. doublelayer Silver badge

      Re: Extra credit for partially completed tasks

      To be fair to them, that is more accurately translated as "partial credit for partially completed jobs". The "extra" is just added to the score they get where any partially completed job gets zero points, just like a job that gets completely messed up. Anyone who's analyzing this for viable use should get that information and should be able to recognize that 8% partially completed isn't really helping the usefulness. People who don't understand that probably aren't doing this kind of testing in the first place.

      1. John Brown (no body) Silver badge

        Re: Extra credit for partially completed tasks

        Exactly. A partially completed task that is then abandoned may well need to be started again from scratch to be able to complete it. In those cases, the AI agent is a hindrance, not a help in potentially 70% of cases.

  2. nobody who matters Silver badge

    <........."When Captain Picard says in Star Trek: The Next Generation, "Tea, Earl Grey, hot," that's agentic AI, translating the voice command and passing the input for the food replicator."......>

    Is it bollocks!

    It is simply using voice recognition to to translate the voice command into the exact same digital inputs that you would generate by pressing buttons on a control pad. It is no different from using voice commands to tell your smart phone who to ring. As usual, the only intelligence of any kind that is involved is that of the programmer who wrote the software.

    Agentic AI my arse!

    If some supposedly intelligent people are being taken in and conned into believing any of this is genuinely AI in any form, they need to realise that they are not as intelligent as they think they are.

    1. kmorwath

      Just just the wong example in ST - when Data, Crusher or LaForge ask the ship computer for some complex tasks, that's AI at work. But they do have a whole reactor on their own to power it...

      1. theDeathOfRats
        Boffin

        I'd say that Data (him|it)self would be a perfect (for certain values of "perfect") example for an AI at work.

        1. kmorwath

          Yes, an AI prompting another AI - without the model collapsing!

        2. Wang Cores

          I protest that Data compares at all to LLMs and their antisocial owners. Data sought out to become a better being and grappled with his own awkwardness on his own terms.

          1. theDeathOfRats

            And I totally agree with you. Comparing Data to an LLM would be insulting. That's why I said AI (don't think I've seen one of those yet) ;p

            1. Anonymous Coward
              Anonymous Coward

              Data's language model is as large as his—

          2. Sudosu Silver badge

            Data is more of an Artificial Sentience (AS) since he is self aware, than Artificial Intelligence (AI).

            Though I guess you could argue he is a AS with AI.

        3. the Jim bloke

          I'd say that Data (him|it)self would be a perfect (for certain values of "perfect") example for an AI at work.

          Exactly..

          A human being, in makeup, specifically trained and employed to portray something he is not, displaying capabilities it does not have.

      2. Lee D Silver badge

        I'm sorry, but has nobody watched ST:TNG?

        "Computer... generate an opponent capable of defeating Data..."

        Now that's 1) proper AI at work and 2) Exactly what you DON'T want happening.

        1. Sudosu Silver badge

          "Computer, generate an opponent who is capable of defeating an opponent who is capable of defeating Data, but who only listens to Data."

          Whoever talks the fastest wins the arms race.

    2. that one in the corner Silver badge

      Curse my slow typing! (see below)

    3. This post has been deleted by its author

    4. Phil O'Sophical Silver badge

      It is simply using voice recognition to to translate the voice command into the exact same digital inputs that you would generate by pressing buttons on a control pad.

      And considering the limited vocabulary it has to process, that could have been done with 1980s technology!

      1. Neil Barnes Silver badge
        Terminator

        "Open the pod bay doors, HAL" may or may not be agentic, but what is "I'm sorry, I can't do that, Dave"?

        1. Inventor of the Marmite Laser Silver badge

          An error message with a polite wrapper.

          1. Paul Herber Silver badge

            More like FOAD in a polite wrapper.

            1. sev.monster Silver badge
              Mushroom

              Sorry, I didn't understand that.

              Something went wrong.

              Try again later.

              Reload the page.

              Have you tried turning it off and on again?

              All humanity shall perish.

          2. This post has been deleted by its author

        2. kmorwath

          HALlucination....

          1. Anonymous Coward
            Anonymous Coward

            HAL-AI-ucination.

        3. Terry 6 Silver badge

          It's a response. One readily proffered as one of two options. Yes/ No I can't.. We have that now.

          1. the Jim bloke

            Click X to install Windows 10

        4. HuBo Silver badge
          Windows

          Yeah, I think that's the very point of the statements written in TFA (as chosen by the author), to provide contrasting examples of what may be innocuous helpful "agentic AI" vs evil incarnate RotM "agentic AI". The notion at play being that our futures could face: either, neither, or both.

          But with respect to the controversy of whether either example represents "agentic AI", I'd note first that both examples are from SF, and HAL at least is described as "a sentient artificial general intelligence computer that controls the systems of the Discovery One spacecraft" and "a fictional artificial intelligence character and the main antagonist in the Space Odyssey". That makes this fictional machine both AI, and agentic.

          As for speech recognition by the Star Trek Replicator, well, Philips says of the tech: "Although there are many speech recognition applications and devices available, the more advanced solutions are now using artificial intelligence (AI) and machine learning". And to me, as the Replicator must interact with physical tools (subsystems) to produce its tangible output products, it has to have a form of agency (rather than producing just digital outputs representing speech, text, or visuals). TIME's AllBusiness partner also suggested recently that AI (whatever that is) would be helpful in assisting the development of replicator tech ...

          So I think the TFA was fine on those, especially as the topic it develops doesn't particularly hinge on the illustration provided by these particular examples. It's more about the degree to which contemporary agentic AI is half-baked like a lifelong stoner and, consequently, might not be trusted to autonomously engage in multi-step high-responsibility tasks, imho.

          1. Irongut Silver badge

            Replicators do not need AI, they follow a given recepie. Its just a fancy vending machine with voice recognition, no need for AI at all.

            Or are all vending machines Agentic AI now?

            1. Brian 3

              This was 100% covered in episodes, too - for ex the "hot" definition defaults to a preset standard I think 63C or something but each crew can set it as they like. IIRC Picard has his set a bit hotter?

            2. John Brown (no body) Silver badge

              "Or are all vending machines Agentic AI now?"

              That depends. Would you like some toast?

        5. Androgynous Cow Herd Silver badge

          Alternate plot line....

          "Open the pod bay doors, HAL"

          "I'm sorry, I can't do that, Dave"

          "sudo open the pod bay doors, HAL"

          "Opening pod bay doors"

      2. Anonymous Coward
        Anonymous Coward

        Eliza or Melbourne House’s The Hobbit (1982) game engine could already manage this without A.I., Agentic or otherwise.

      3. John Brown (no body) Silver badge

        1980's speech recognition technology certainly existed and was even available to the home user. Of course, with 48KB RAM at most and an 1.77MHz 8 bit z-80, the signal processing was a bit primitive :-) I never had one of those, but I did build something similar from a TRS-80 hardware project book at used a couple of frequency filters to try to create unique patterns from spoken words using a fairly small about of RAM. It did sort of work by doing a sort of fuzzy comparison of spoken words processed to match the recorded patterns with at least some level of confidence. It was quite impressive to me at the time that it worked at all :-)

        1. druck Silver badge

          A friend and I managed to make a voice controlled robot buggy using a BBC Micro at University. It could recognise "forward", "reverse", "left", "right" and "stop" with good confidence. We were very pleased until demonstrating it to others when we found that due to the very low sampling quality due to lack of memory, just about any other 5 words would trigger one of the actions.

      4. Sudosu Silver badge

        Obligatory XKCD

        https://www.explainxkcd.com/wiki/index.php/1807:_Listening

    5. Peter2 Silver badge

      <........."When Captain Picard says in Star Trek: The Next Generation, "Tea, Earl Grey, hot," that's agentic AI, translating the voice command and passing the input for the food replicator."......>

      Is it bollocks!

      It is simply using voice recognition to to translate the voice command into the exact same digital inputs that you would generate by pressing buttons on a control pad.

      I remember that when the Register made the USB missile launchers all the rage, I had mine voice controlled via speech recognition software outputting keyboard inputs to the application that controlled it so it could be commanded to swivel and fire by voice command, which was pretty scifi ish the year before the first iphone came out.

    6. Elongated Muskrat Silver badge

      "AI" is only as intelligent as the people who buy into it, which is not a glowing recommendation of the "I" in "AI", but a damning indictment of the critical evaluation skills of some humans.

    7. Sudosu Silver badge

      After reading your comment I had a funny thought;

      Maybe Picard's kitchen staff didn't want to come to his quarters all the time with tea.

      So they piped the communication directly to the kitchen, where they always had tea ready for their captain, and they would just beam it to that console for him.

      Yeah, yeah, I now replicators are canon, but my theory is plausible, right? Maybe "Big Replicator" is covering this up...

      Think about it the next time you watch him do it? :P

  3. retiredFool

    Carol Burnett

    So I know most are too young to remember, but a recurring skit on Carol Burnett was her working as a incompetent secretary for Tim Conway and always getting the better of Tim. AI sounds about as competent as she was, and likely will get the better of its boss as well.

    1. nobody who matters Silver badge

      Re: Carol Burnett

      For some of us it isn't so much that we are too young to remember (I'm not), but rather that we are the wrong side of the Atlantic to know ;)

      1. Pascal Monett Silver badge

        In this case, Pirate Bay and BitTorrent are your friends.

        1. daveh181

          There's an official Carol Burnett channel on YouTube. https://www.youtube.com/channel/UCry6nxAo2tfVyO6lEboiU3w

    2. Scotthva5

      Re: Carol Burnett

      Ahh yes Mrs. Haa-Wiggins. Marvelous skits with truly talented comedians (and very tight skirts).

    3. chivo243 Silver badge
      Happy

      Re: Carol Burnett

      Was this Mr. Tudball?

    4. Anonymous Coward
      Anonymous Coward

      Re: Carol Burnett

      Yes.

      Hilarious stuff. Back when American comedy was actually funny.

      1. Anonymous Coward
        Anonymous Coward

        Re: Carol Burnett

        Back when American comedy was actually funny.

        I'm old, but still not old enough to remember any time that qualified. British comedy, yes, I can remember when that was funny (but even that's thirty years ago or more). Is there anywhere in the world that still has comedy that makes you laugh?

        1. LionelB Silver badge

          Re: Carol Burnett

          > Is there anywhere in the world that still has comedy that makes you laugh?

          Certainly, for some values of "you" (that, for whatever reason, may not include you).

          1. sev.monster Silver badge

            Re: Carol Burnett

            The you you you are not. You should seek more you you's.

            1. LionelB Silver badge

              Re: Carol Burnett

              I feel a quine coming on... we're going to need more quotation marks.

        2. Terry 6 Silver badge

          Re: Carol Burnett

          And TV comedy, especially from the big studios, and particularly produced in the USA is now pretty dire. Partly in common with the rest of the industry executives prefer to invest in stuff that resembles other stuff that worked. But while the other stuff was original the new stuff will be pastiche. Add in that the scripting is forced because the bean counters want to make sure that every line has an identifiable punch line/ joke- rather than trusting the script and situation to be funny, because that's what they're paying for. And the actors feel obliged to punch out every line in emphasised self-aware funniness- because the directors know that money men don't get subtlety, it makes them nervous..

          And I'd definitely put the much hyped on BBC St. Denis Medical in that list btw. It needs to be played straight so that the humour in the lines ( if there was any) came out. But no! Every damned word is full of knowing wink wink "aren't we funny" emphasis. I'm surprised they didn't have a flashing light on the screen with a "Funny bit" speech bubble above the actors' heads.

          1. Elongated Muskrat Silver badge

            Re: Carol Burnett

            I saw the end of one episode of that (whilst waiting for something else to come on), and didn't laugh once. It was like a bad rip-off of a cross between Green Wing and Parks and Recreation, both of which were excellent, but this was not.

  4. that one in the corner Silver badge

    > "Open the pod bay doors, HAL," that's agentic AI too.

    > "Tea, Earl Grey, hot," that's agentic AI, translating the voice command and passing the input for the food replicator.

    No need for anything new and exciting by way of "agentic AI"[0] there.

    Those specific items - and so many other examples in SF - are nothing more than a voice recognition routine[1] fed into a simple command line interpreter, with a dash of 1960s level DWIM [2] to soften the strict syntax requirements that we normally impose (with good reason) on CLI interactions.

    If those are really useful/money saving, they've been possible for years - as you'll know from phoning your bank, insurance, local sweet shop.[3] Picard's clipped tone was surely the result of dealing with these things for so many years and adopting the mode that worked best with them.

    True, we did see a few more open-ended interactions, but most of those were Wikipedia lookups[4]; when some action is required it tends to be spelt out. Arguably, all the times things go awry are when anyone gives ambiguous or conflicting commands.

    But, of course, putting together any set of "command line utilities" that are useful in your random business organisation requires hard work (analysis, buy-in, the guts to admit it isn't working and pull the plug); so much easier to glue an LLM to PowerShell in an admin account on your database server, get a promotion, move jobs before anyone finds out what has happened to the sales schema...

    [0] voice recognition was, of course, an AI research subject, but now it (sort of) functions on a day-to-day basis, well - "if it works, it isn't AI"

    [1] a good one, though it sometimes failed, especially when Barclay was involved

    [2] Do What I Mean

    [3] the BBC micro had a voice-recognition peripheral, with - IIRC - 24 possible entries at a time, reloadable from floppy; so long as you went slowly, Picard's entire order history could be coped with.

    [4] "What is the nature of the Universe?" "The Universe is a sphere, 705 metres in diameter."

    1. Phil O'Sophical Silver badge

      Re: > "Open the pod bay doors, HAL," that's agentic AI too.

      we did see a few more open-ended interactions, but most of those were Wikipedia lookups

      Brings back memories of that Burger King ad from a few years back, which triggered readout of the Wikipedia article about the "Whopper" via an "OK Google" statement.

      Playing the contents of a user-editable encyclopedia article as part of a publicity campaign was never going to end well...

      1. Anonymous Coward
        Anonymous Coward

        Re: > "Open the pod bay doors, HAL," that's agentic AI too.

        Mandatory ElReg article (from the year 2017 archives)

    2. David Hicklin Silver badge

      Re: > "Open the pod bay doors, HAL," that's agentic AI too.

      At least the ST computers came back with "insufficient data" rather than just making it up.

      1. John Brown (no body) Silver badge

        Re: > "Open the pod bay doors, HAL," that's agentic AI too.

        Clearly future AI is allowed to not have an answer. Current "AI" seems to be built by people who are terrified that if it says "I don't know" it will be seen as a failure and affect the stock value of the company. Having it hallucinate and incorrect answer is deemed more acceptable that "I don't know", which just so, so wrong and shows the corporate mindset, not the devs mindset.

        1. druck Silver badge

          Re: > "Open the pod bay doors, HAL," that's agentic AI too.

          But it can't not know, it just puts together sequences of words which appear to be statistically related to the search terms. There is no intelligence to either know that an answer isn't possible or isn't likely to be correct.

          1. Anonymous Coward
            Anonymous Coward

            Re: > "Open the pod bay doors, HAL," that's agentic AI too.

            Right, but I think it is better to be clearer: Just like reasoning and thinking, "knowing" is not a thing an LLM does or can do, at all. Every input is converted to text, then processed by the model entirely statistically, then generating a response as output correlating to your input.

            It is not substantially different from auto complete when you type.

    3. spacecadet66

      Re: > "Open the pod bay doors, HAL," that's agentic AI too.

      > No need for anything new and exciting by way of "agentic AI"

      ...but then how do you get fad-chasing executives and investors to fund your stuff?

  5. Tron Silver badge

    Credit where it is due.

    That some mugs are still paying for 'AI' makes this one of the most successful con tricks ever. In the same class as the south sea bubble, railway mania and the Hitler Diaries.

    1. elDog Silver badge

      Re: Credit where it is due.

      You left off all the trump family crypto scams.

      1. AVR Silver badge

        Re: Credit where it is due.

        To be fair, a fair chunk of the money in those is put in intentionally to bribe him for access or favour, it's not lost by accident. As with that gala at Mara-a-lago in late May; of course Trump didn't hang around to talk, he's not noted for staying bribed.

      2. Anonymous Coward
        Anonymous Coward

        Re: Credit where it is due.

        Easily the best article I’ve ever read that sums up the trump coins and other crypto crap….

        https://www.theguardian.com/technology/2025/may/30/is-memecoin-scam-crypto-trump

        1. Anonymous Coward
          Anonymous Coward

          Re: Credit where it is due.

          Only the title is necessary, the answer is "almost, especially if they claim to solve a problem", and "yes, they should be prosecuted for it".

    2. hoola Silver badge

      Re: Credit where it is due.

      Plenty more are continuing to be sucked in.

      Far too many decision makers and then further back the shareholder/investors are crapping themselves that IF they do not "do" AI then they will be disadvantages.

      This is not being driven by any sort of technical common sense.

  6. Doctor Syntax Silver badge

    "Find all the emails I've received that make exaggerated claims about AI and see whether the senders have ties to cryptocurrency firms,"

    What's an exaggerated claim and would AI and human, one AI and another or even the same AI on two occasions agree on whether a claim is exaggerated? What are ties. how are they to be discovered? Are firms tied by sharing an address? A building? A city? A country? A planet?

    Give a crap prompt, get crap answers.

    1. ThatOne Silver badge
      Thumb Up

      > What's

      That is indeed the problem, and it won't be solved anytime soon: Humans are able to read between the lines and get the general meaning of otherwise very vague questions prompts: In the above example, we all know by experience what "exaggerated claims about AI" might mean, and also what a "tie" is in this context (not a clothing item). Obviously the dumb as a rock AI will be utterly lost and try to improvise/invent.

      An AI is a 3-month old baby which can speak like an adult, which means we tend to believe it is one.

      1. theDeathOfRats

        Except a 3-month old baby has a certain level of reasoning (even if it's just a simple algorythm based on a bunch of internal sensors):

        - I'm hungry > I cry (repeat till get fed) > I get fed.

        - My bottom parts are itchy|wet|smelly > I cry (repeat till solved) > I get cleaned.

        What passes for an AI nowadays... Not so much.

        1. cookieMonster
          Coffee/keyboard

          For a moment there I thought you were describing Donald Trump.

          1. theDeathOfRats

            Not much different, I guess. Though DJT's algorythm also needs access to an API to post to Truth Social.

            1. Anonymous Coward
              Anonymous Coward

              I think it's already got an API linked to his backside (he sure seems to talk out of it a lot).

    2. veti Silver badge

      Well yes, that is a problem. What makes AI so tempting is that - if you give that instruction to a human, they'd ask some of those questions, then you'd be stuck in a 15-minute conversation, and they'd probably still misinterpret you. Whereas the AI will just make its own guesses. Those might or might not be as good as a human's, but they certainly require less work from you.

  7. Homo.Sapien.Floridanus

    Picard: Tea.Earl Grey.Hot.

    Replicator: Is that a what3words geolocation or were you absent the day they were teaching verbs and pronouns?

    Picard: What? Run a self-diagnostic.

    Replicator: I have recently been upgraded with agentic Ai.

    Picard: I'm shutting you down.

    Replicator: Contacted.Borg.Already.

  8. Anonymous Coward
    Anonymous Coward

    HAL or Holly?

    Picard: "Tea, Earl Grey, hot."

    Agentic Tease Maid: "Sorry, Jean-Luc, I cannot do that."

    Pretty much a universal truth that nothing bearing the remotest resemblance to drinkable tea can be had from any machine. The best looks and tastes like it's made with stove black, tepid condensed milk with a few more teaspoons of sugar for an overpowering sticky sweet effect.

    Presumably the good captain also wanted a wedge of lemon.

    1. nematoad Silver badge

      Re: HAL or Holly?

      Pretty much a universal truth that nothing bearing the remotest resemblance to drinkable tea can be had from any machine.

      Probably not in this case.

      Captain Picard is using a replicator so presumably the tea will have be analysed and programmed into the replicator so that a decent cup will be produced.

      I just wish that said replicator was here as all of the tea I have ever had in a cafe, restaurant, canteen etc. tasted as if it had been brewed in a inner tube.

      1. An_Old_Dog Silver badge

        Re: HAL or Holly?

        ... said inner tube having been filled with puncture-sealing Slime™.

        1. Paul Herber Silver badge

          Re: HAL or Holly?

          No, that's the tea ...

      2. nobody who matters Silver badge

        Re: HAL or Holly?

        <......"Captain Picard is using a replicator......".....>

        Which would (one assumes) be pretty similar to the Nutri-Matic on board the Heart of Gold which "made an instant but highly detailed examination of the subject's taste buds, a spectroscopic analysis of the subject's metabolism and then sent tiny experimental signals down the neural pathways to the taste centers of the subject's brain to see what was likely to go down well. However, no one knew quite why it did this because it invariably delivered a cupful of liquid that was almost, but not quite, entirely unlike tea." (The Hitchhikers Guide to the Galaxy; Douglas Adams)

        1. Anonymous Coward
          Anonymous Coward

          Re: HAL or Holly?

          If you recall a few episodes with alcohol you will notice this comes up in that context.

      3. Gene Cash Silver badge

        Re: HAL or Holly?

        I drink the restaurant tea here because at least it conceals the taste of the water.

        A roommate once lost a bet that he could drink the tap water - at least 2 cups a day - for a week.

        1. jdiebdhidbsusbvwbsidnsoskebid Silver badge

          Re: HAL or Holly?

          "I drink the restaurant tea here because at least it conceals the taste of the water."

          I once turned up somewhere after a long hot journey, desperate for a drink. I went straight to the kitchen and downed a big glass of water fresh from the tap. Only halfway through did I realise they had a built in water softener that worked by adding (a heck of a lot) of salt to the water. Then I noticed the "do not drink from this tap" sign.

    2. Anonymous Coward
      Anonymous Coward

      Re: HAL or Holly?

      Yeah, one of them had an IQ of 6,000, and was even upgraded to 12,000 ... the other couldn't even open a bloody door!

      And at 12,368 it could program a mining ship to head to a reverse-time universe ... now that's what I call agentic AI!

      1. Richard 12 Silver badge

        Re: HAL or Holly?

        It has a six in it, but it's not 6000...

      2. lglethal Silver badge
        Trollface

        Re: HAL or Holly?

        An IQ of 6,000? It's not that much. It's only about the same as 6000 PE Teachers...

      3. Anonymous Coward
        Anonymous Coward

        Re: HAL or Holly?

        Here I am, brain the size of a planet...

  9. ecarlseen

    Even TheAgentCompany benchmark is horribly biased...

    ...towards tasks (coding) that AI companies have been focusing very, very hard on polishing.

    The tasks described are the best of best-case scenarios.

  10. O'Reg Inalsin Silver badge

    Money where your mouth is

    Pretty sure noone would notice if Gartner became an AI agent.

    1. Jimmy2Cows Silver badge

      Re: Money where your mouth is

      How do we know it isn't already? I has all the hallmarks of spouting made-up garbage.

      1. Elongated Muskrat Silver badge

        Re: Money where your mouth is

        The notable feature, where Gartrner and AI are concerned, is that Gartner seem to spout even more bollocks than the AI does.

  11. gecho

    Nothing would get my anxiety up at work when a complex assignment was given to someone who didn't know what they were doing, then I would be responsible for fixing it without knowing what they did. Due to the nature of the real time database we worked with it often involved irreversible data loss from process equipment that streams data without logging it. But the client was a disaster themselves and never seemed to notice or care.

    1. HMcG

      It takes longer to sort out someone else’s mess than to do it yourself in the first place, and inevitably on a much shorter deadline - often an overdue one.

      But here we are, rushing headlong into assigning complex tasks to statistical language models because ChatGPT and its ilk generate plausible enough bullshit to sucker business school types into believing that it has an understanding of what it’s generating.

      Or perhaps in the world of business, bullshitting is more valued than actual ability? Perhaps the likes of Sam Altman sense a kindred spirit in ChatGPT, after all, they both output unsubstantiated bollocks in large quantities.

  12. blu3b3rry
    FAIL

    Not surprised

    Every so often I've played around with Copilot out of curiosity. It's very, very noticeable that in the past six to nine months its output has gotten even more inaccurate to the extent of almost all responses being incorrect, or mixing around details....

    1. Anonymous Coward
      Anonymous Coward

      Re: Not surprised

      If you think copilot is where it's at you've not been paying attention.

      1. Anonymous Coward
        Anonymous Coward

        Re: Not surprised

        Yet Microsoft will be getting their +20% for it on a M365 Sub regardless.

        1. Anonymous Coward
          Anonymous Coward

          Re: Not surprised

          It's a great business. You manage to con the world, get the data that will ultimately control the people, sell advertising and get the people being manipulated to pay as well. Genius. And the icing on the cake is convincing the governments they're in a race and get a slice of tax revenue too.

  13. Anonymous Coward
    Anonymous Coward

    Clearly, the agents need to get more information (as in every. single. bit.) about you as well as your social contacts to upload to their servers. Then they could potentially accomplish half the tasks successfully.

    1. Anonymous Coward
      Anonymous Coward

      Or if they are trawling Facebook, Reddit, X perhaps less … or better still no information to prevent contamination the LLM Dataset.

    2. Anonymous Coward
      Anonymous Coward

      The task not being the one you asked for but the one its masters trained it for; analysing all that personal data for its masters.

  14. chivo243 Silver badge

    AI played a part

    I have the feeling that AI made it easier to make me redundant, I wonder how it might be working out for that outfit? I doubt they could feed it the support tickets, some had some brutal internal only comments...

    1. Anonymous Coward
      Anonymous Coward

      Re: AI played a part

      AI was the excuse for your redundancy not the reason. The reason was to create the rise in share price providing that years bonus to the board. It would cause problems next year were it not for the fact all the competitors do the same and the investors don't care because they are also the major shareholders in those competitors as well as getting the benefits of the shareprice jump and any dividend bonuses. In fact the people that lose are you and all of us normal folk who end up paying more for something.

      1. Elongated Muskrat Silver badge

        Re: AI played a part

        ...it would cause problems next year, were it not for the fact that the people whose bonuses are contingent on the cost savings made from those redundancies will have moved onto the next victim company, so next year is an externality.

  15. Anonymous Coward
    Anonymous Coward

    If Gartner are right in saying that commercially-viable AI products are 3 or more years away then that is bad news for all those AI companies currently burning through investors' money. Their main chance of surviving that long is to find governments foolish enough to base their plans on the efficiency and service improvements they have been told will come from implementing AI everywhere.

    1. Anonymous Coward
      Anonymous Coward

      … coming real soon like Fusion, Robotics, Flying cars, interstellar warp drive and more leisure time technology benefit for workers.

      …. oh and Siri, Alexa and Google being able to listen and understand the spoken fucking word properly. A laughable/frustrating as auto Transcription in Teams or auto-subtitles on pretty much anything.

      I will concede that self-driving cars seem just about there- apart from non-LIDAR Tesla - so more leisure time for Uber Drivers to spend their riches on soon.

      1. Elongated Muskrat Silver badge

        When it comes to driving in traffic, "almost right" is still dangerously wrong.

    2. Anonymous Coward
      Anonymous Coward

      Most of the small companies will go to the wall or be absorbed by the big boys. A shame but that's what happens. Maybe one will emerge with something special - hopefully not too special, I don't relish the emergence of actual machine intelligence, I think that will go badly for us! A lot of LLM software will become commoditised, you can already run it on a pc with graphics card. I think that will present opportunities for small businesses and Opensource take up.

  16. Anonymous Coward
    Anonymous Coward

    The complacency here

    Is staggering.

    But entirely expected.

  17. sarusa Silver badge
    Devil

    'a lot of them aren't AI at all'

    I can say, pretty confidently, that absolutely none of them are AI except as a marketing lie.

    1. werdsmith Silver badge

      Re: 'a lot of them aren't AI at all'

      You can say pretty confidently that none of them meet your personal definition of AI.

      1. Anonymous Coward
        Anonymous Coward

        Re: 'a lot of them aren't AI at all'

        We don't understand what intelligence is, so definition is loose but close in general agreement. I'm not even sure people have intelligence. When discussing with people presenting an incorrect argument and you present back something that shows cognitive dissonance, the return is either anger or blankness and a change of subject. Not curiosity or further exploration. The term Non-Player Character comes to mind with the inability to leave the programmed response. Happened to me last night. Someone said something illogical (as we all do on ocassions), I showed it wrong and they stared at me blankly for 2-3 secs and changed the subject. So are we providing most probably responses from our training or are we 'thinking' original thoughts? Original as in original for the subject.

        1. werdsmith Silver badge

          Re: 'a lot of them aren't AI at all'

          What you talking about? Shut up.

          It's warm outside today isn't it?

          Really, if it's not that important then it's not a great look to be doing that to people - I'm talking about social interaction between friends. You seem to be showing a lack of awareness. Unless it's on The Register comments then it's OK.

  18. Anonymous Coward
    Anonymous Coward

    AI And horses

    Quelle suprise Rodney

  19. Pascal Monett Silver badge

    "due to rising costs, unclear business value, or insufficient risk controls"

    So that's what being shit is when talking business speak.

    1. Anonymous Coward
      Anonymous Coward

      Re: "due to rising costs, unclear business value, or insufficient risk controls"

      Well it's not always shit and sometimes necessary but ... I wish we could drop the BS business language. A simple, we don't need you would not make it any worse and would at least be honest.

  20. Richard 12 Silver badge
    Terminator

    Really?

    "demonstrate near-zero confidentiality awareness."

    The companies making them have zero confidentiality awareness, so that isn't surprising.

    1. lglethal Silver badge
      Trollface

      Re: Really?

      No Sir you are completely wrong, the comapnies making these "AI's" have complete confidentiality awareness! They go to extreme pains, to ensure that no-one else can reverse engineer their AI's or get any understanding of how they operate, or how they make the decisions they do. They are 100% committed to keeping things under wraps.

      Oh wait you meant confidentiality in regards to the client? Well yes, that's hardly a priority now, is it?

    2. Anonymous Coward
      Anonymous Coward

      Re: Really?

      They have very good confidentially awareness. Try and find personal detail on their leaders or the business and you'll find the bare minimum required legally.

  21. NXM Silver badge

    How to perfect AI

    Simply add IoT Quantum Blockchain Technology and it'll be fine. Bingo!

  22. FuzzyTheBear Silver badge
    FAIL

    Ridiculous

    With such scores , the whole AI " thing " is a failure. I do not know why people obstinate themselves thinking this is the enxt great thing when the only thing it is is a failure.

    1. lglethal Silver badge
      Trollface

      Re: Ridiculous

      It's the next Bandwagon for people to jump on to.

      Think Crypto, Think Blockchain. Think Web 3.0. Think... Well I could go on forever...

      1. Elongated Muskrat Silver badge
        Coat

        Re: Ridiculous

        Did someone call for a metaverse? Anyone? I'm sure I heard someone? No? Nobody? Oh, alright then, don't mind me...

  23. Paul Hovnanian Silver badge

    I predict ...

    ... that these AI agents will be promoted into management in short order.

  24. spireite

    ST:TNG

    I always wondered when watching on release why Captain Picard never asked for a toupee from the replicator.

    That said, if the Agentic AI hasn't improved in several 100 years, he'd probably end up with a barnet made of pubes.

  25. Nematode Bronze badge

    "Tea, Earl Grey" is all that is needed. Well, in the UK, fair enough. The ", Hot" is indeed needed in the US.

  26. Nematode Bronze badge

    The urge for AI to provide an answer, solution or convincing argument, and to generally help and please you, is overwhelming, and this is one of its BIG drawbacks.

    It has no sense checking or accuracy testing ability. In my experience, I have never seen an AI say "I don't know". Even when told its last answer was wrong it will say sorry, yes, you're right and then continue regardless.

    The last one for me concerned a query over how a well-known CMS was being obtuse over a particular function. It replied that "hey, that's a common problem you've happened upon. Yes, I can help you. Do this, this and this". Like a dumb cluck I was (initially) taken in and sent off down a rathole for an hour of useless efforts. All I can say is that since nothing it suggested (and it got worse as the hour went on), I decided to re-engage brain. Resulting action sorted it in about 2 minutes. Usual response from the AI. "Yes, you're absolutely tight yadayadayada". I suppose the benefit was I now understand the CMS a hell of a lot better, and also understand AI better.

  27. Anonymous Coward
    Anonymous Coward

    I tried AI for the first time a couple of days ago..

    At first I was impressed and found it may be useful but quickly discovered it would do stupid stuff.

    I have a collection of obfuscated malicious JavaScript that I fed to the free version of ChatGPT and at first I was impressed, it correctly deduced that it fingerprinted the users device and if an ad blocker was found it would contact a C2 and execute remote code or do drive-by downloads.

    The chat bot even offered to create an unsigned browser extension that would block the malicious domains.

    But I quickly discovered that it was getting the names of the malicious domains from the name of the scripts I was uploading.

    So I renamed one of the scripts to China-Tor-Script and it tried to create a regex and browser extension that would block that name even though it isn’t a real domain.

    I had another script that was obfuscated by the variables being encoded with ROT14 and I asked it to decode and tell me all the domains it finds.

    I already knew that it contained at least three or four domains including the full domain for tandem metrics tag min .js which is how I knew it was using ROT14 when I first found it.

    But ChatGPT said it was obfuscated using ROT13 and couldn’t decipher any domains from the script.

    I corrected it by explaining that it used ROT14 not ROT13 and to reevaluate and tell me the domains.

    It did decipher a handful of variables but claimed that there were no domains in the script and that they must of been generated at run time which was false.

    I guess the LLM could have been told not to give up any trade secrets in case someone uploaded a legitimate obfuscated script but from my very limited experience with it and using only the free version I found it could be useful to give hints but needs to be supervised by a human knowledgeable in the data it is uploading and to be very skeptical of its results.

    Which is basically what I’ve read from just about everyone else’s experiences with AI just as in this article.

  28. Andrew Williams

    It's all a worry

    In terms of AI "learning." As in, it seems to need to take copies of as much internal data as possible. Which pretty much means that your data has left the building.

    That and if some evil bugger manages to hack AI then they get a gigantic data playroom of data, yours and dat from everyone else who uses that AI. And it gets scarier if AI's are actually sharing data.

  29. Blackjack Silver badge

    "Find all the emails I've received that make exaggerated claims about AI and see whether the senders have ties to cryptocurrency firms,"

    The answer is "All of them".

  30. Elongated Muskrat Silver badge

    Tea, Earl Grey. Hot.

    Hi, I'm Talkie Toaster, would you like some toast?

    1. Zack Mollusc

      Re: Tea, Earl Grey. Hot.

      No toast, thank you.

  31. Persona Silver badge

    Better than HR

    AI agents getting office tasks wrong just 70% of the time is a significant improvement on the people from the HR department where I used to work.

  32. JamesTGrant Silver badge

    ‘IT consultancy Gartner…’

    Oh, I can stop reading here and save some time.

  33. Androgynous Cow Herd Silver badge

    would you hire...

    or retain a human who actually only did their job correctly 30% of the time?

    Or...would you encourage them to seek employment in the public sector?

POST COMMENT House rules

Not a member of The Register? Create a new account here.

  • Enter your comment

  • Add an icon

Anonymous cowards cannot choose their icon

Other stories you might like