"optimizations to Llama.cpp "
They'll be contributing those back, right?
Popular generative AI chatbots and services like ChatGPT or Gemini mostly run on GPUs or other dedicated accelerators, but as smaller models are more widely deployed in the enterprise, CPU-makers Intel and Ampere are suggesting their wares can do the job too – and their arguments aren't entirely without merit. To be clear, …
Llama, on my desktop, can produce better bash scripts then I can. OK, that's a low bar. They even (gasp) include useful documented comments so hacking them into something useful. gets me a better result faster. Maybe I should have memorised the Linux toolbox but I don't want to be mistaken for a bot ;-)