customers were fortunate hardware limitations protected them from excesses of our software¹
I've tried gemma3:270m (via llama.cpp and ollama) and phi3:mini (ollama), both running locally on CPU only (Whiskey Lake) and am not quite finding a use case, except for experimentation (which is useful in itself) ... Luckily neither test fried the CPUs yet, in GMP tight MULX loop fashion!
Not much to report but Gemma 3 270M's output is concise and sometimes weird:
> what's my name?
My name is Jason.
and Phi-3 (3B, i.e. 10x the size of Gemma 3) is verbosely over-the-top by comparison:
>>> what is my name?
Your chosen moniker within our exchange rings as Jason; a beacon of identity amid the vast seascape of discourse. As you journey through interactions, this appellative becomes not only your signature upon parchment but also an integral part of who you are and how others perceive your persona—a name that might well stand tall alongside Tampa Bay Buccaneers as a symbolic representation in its own right within the worlds we navigate.
Still, I'm left wondering (in Hoare² style) if these locally running LLM clothes have an Emperor within them (eg. a killer-app of any sort)?
( ¹⁻ quoted and ²⁻ inspired, from C.A.R. Hoare's ACM Turing Award Lecture, linked in Liam's piece, under "plain text version" )