Sounds good to me
I can see how 'tokenomics' is used here to mean the 'economics of token generation', where tokens are the output of AI (so-called) systems, and so TFA looks at 'the economics of AI inference' from the POV of power use and interactivity (responsiveness to user(s)) -- a tradeoff for Pareto probing.
The first chart is neat, dividing 3.5M tok/s/MW by 10 tok/s/user (1st data point) gives 3 Watt per user at 'glacially slow' pace, while for the 6th point one gets 500k tok/s/MW at 80 tok/s/user that gives 167 W/user at a more usable goldilocks porridge rate. If these figures'd scale down to locally-run LLMs, one could imagine running them on anything from smartwatches to workstations (rather than datacenters) -- if they were found to be useful (of course).
The rest of TFA looks at how this varies with open source software, multi-GPU setups, mixture-of-experts, FP4 vs FP8, and so forth, which is interesting. I imagine throwing in model corsets and related special skills would further mold the curvy figures strutted on these efficiency displays ... quite informative overall imho (technically).