back to article Hugging Face puts the squeeze on Nvidia's software ambitions

Hugging Face this week announced HUGS, its answer to Nvidia's Inference Microservices (NIMs), which the AI repo claims will let customers deploy and run LLMs and models on a much wider variety of hardware. Like Nvidia's previously announced NIMs, Hugging Face Generative AI Services (HUGS) are essentially just containerized …

  1. anthonyhegedus Silver badge

    Speaking as a something of a veteran of the computer industry - I've been in IT for nearly 40 years - I've seen some pretty impressive jargon over the years. It's a reasonably comprehensible thing to IT people, but to anyone outside the game, this article has achieved Peak Jargon!

    1. Anonymous Coward
      Anonymous Coward

      Words on the page but no meaning to anyone already lacking the context. Agree 100%

    2. Anonymous Coward
      Anonymous Coward

      It's perfectly comprehensible if you're working in this space, it's a pretty big announcement. NVIDIA NIMs are becoming really very popular for corporates who want to provide GenAI to their own teams for <reasons> but rightly worry about costs spinning out of control and data security. Running your own models can be a pain, and there's so many dependencies that managing updates can be a pain in the arse, hence paying someone else to handle that donkey work for you.

      One of the big advantages here is you're not as tied into the NVIDIA ecosystem, potentially opening it up to other GPU providers in the future, making models easier to deploy on non-NVIDIA hardware.

      1. This post has been deleted by its author

      2. Anonymous Coward
        Anonymous Coward

        re: if you're working in this space, it's a pretty big announcement

        Meanwhile, for those of us on planet Earth, it's just more piggies trying to snuffle up as much of the trough as they can before the suckers realises computers are supposed to get things RIGHT.

    3. abend0c4 Silver badge

      To be fair, this isn't really a new phenomenon. It's become almost inevitable that when you visit the website of any currently fashionable project you get some marketing blurb that doesn't actually explain what the software does or why you might need it - and links that promise documentation and tutorials that in reality consist mostly of impenetrable self-referential jargon and a sample configuration file for a now-deprecated version. I suppose I'm just nostalgic for the old days when there were technical writers and they even got paid. But, ultimately, if you're expecting people to use this stuff it ultimately has to be rather more accessible.

    4. Anonymous Coward
      Anonymous Coward

      But have they successfully prevented side fumbling?

      1. ecofeco Silver badge

        The malleable logarithmic casing fixes that in such a way that the two main spurving bearings were in a direct line with the pentametric fan.

  2. NoneSuch Silver badge
    Mushroom

    Huggingface Rocks

    Would rather give them a try over some corporation who will turn on the paywall once they get market dominance.

    1. rgjnk Bronze badge
      Meh

      Re: Huggingface Rocks

      Have to invoke Poe's Law here as I genuinely can't tell if this is sincere or not.

  3. Mike007 Silver badge

    I use the ollama docker container. Just run it and you have the relevant APIs on a HTTP port.

    The issue I have is that this is the easiest part of the whole process. What is going to use this AI? An in-house application written by someone who declared running a readily avaliable preconfigured docker container to be too difficult and had to sign up to a subscription service for this?

    If a vendor is providing you with software that has functionality to plug in to an AI model, they should probably have an "install and enable AI features" button on the settings page... You know, on the screen they should already have for configuring the functionality.

    Am I missing something, or are their customers just not able to use Google?

  4. tekHedd

    ...and it's always LLMs...

    I particularly love how the solution to everything is "throw an LLM at it". It's nice to know there is more than one option in the space of "quickly get 85% of the way to a solution by throwing an LLM at it, giving investors the idea that the remaining 15% will be easy."

    1. ecofeco Silver badge

      Re: ...and it's always LLMs...

      That snake oil doesn't just sell itself, you know!

  5. pip25
    Meh

    Those prices don't seem too competitive

    I can rent a H100 SXM on-demand for $2.99 per hour on RunPod, or $1.75 if I allow my workload to be interrupted. (This is probably not the cheapest provider either, just one that comes with easy-to-setup Docker templates.) That $2.5 to $6.74 offer from DigitalOcean is rather uninspiring in comparison.

    1. IGotOut Silver badge

      Re: Those prices don't seem too competitive

      The extra is for Digital Ocean to turn a blind eye from the hacking attacks and spam originating from there servers.

      1. Teal Bee

        Re: Those prices don't seem too competitive

        I used to wonder if DO has a profit sharing arrangement with the spammers, it was that bad.

        Nowadays I get more spam from Gmail than all the other providers combined.

  6. anonymous boring coward Silver badge

    Hugging Face?

    I thought it was called a Face Hugger?

POST COMMENT House rules

Not a member of The Register? Create a new account here.

  • Enter your comment

  • Add an icon

Anonymous cowards cannot choose their icon

Other stories you might like