>The models do know they're talking nonsense, but they can't back out because they've already "promised" an answer.
That sentence may or may not be true. We don't know. Worse, that sentence is poorly defined. We have no tight definition of what "know" means in that context, or "promised" or "back out"; none of these terms feature in the science of how LLMs work. Because of this, not only we don't know whether that sentence is true, but it's not even clear whether we can know.
Compare this with the rocketry example from my OP. If the rocket company that promised a long-term ROI has a rocket explode on the pad, they might not know right away why it exploded. However, they do have exact definitions of what should have happened and what happened instead, and that provides a path towards understanding exactly what went wrong, and that understanding will provide a path towards preventing it from happening again. There will be tensile strengths to calculate, pressures to adjust, temperatures to measure. All of those things are extremely well-defined. Eventually, you'll find which numbers were wrong.
For the LLM, there are no numbers. You can't say "okay, this is a grade 11,362 hallucination, which appeared because this signal was 6,11 instead of 6,14". There are model weights, but they are utterly impossible to interpret in a causal sense. All we have is fuzzy simil-psychological terms; we don't even know whether they describe what's going on correctly, and even if we did, they provide no hard guidance on how to fix it. So the model "knows it's talking nonsense but can't back out because of a promise", right, even pretending that we know exactly what that means, is there anything in there that tells me how to adjust the model weights so that it doesn't happen again? No, there isn't. There is no hard-science path towards fixing the problem.
Basically, AI investors right now believe they are spending money to solve engineering problems, but they really aren't. I'm a strong supporter of investing in pure research, but one really ought to know what they're doing.