Nonsense from AI prompts
Had that a few times when using "AI" APIs (though as have only been one provider not sure if its a more "global issue") - have just resignedly retweaked my prompt (just like when I do not get gibberish, but do get the prompt misinterpreted, partially ignored etc.) If I reported all prompts with buggy output, would get little work done!
I typically find testing a vaguely complex prompt / rules combination needs several iterations of refinement.
Do a run, see what odd behaviour results (e.g. "AI" often goes off at an irrelevant tangent with lengthy prompts / rules, so add a rule that squishes those tangential pathway outputs)
Perform run with tweaked rules / prompts, see how good (or bad) it is, refine again etc.
Even to achieve something relatively simple, the rules / prompts setup behind the scenes can be quite large.
Once it seems OK, then add a lot mare variation to data it is to process (as lots of quirks seen are data dependent - most of my use was "AI" to analyse text / documents to extract various data & many issues only appeared with small % of documents ).
You do sometimes have to give up on prompts as after several refinements still end up with junk.
It really is only "safe" to use for automating a few tasks, with a lot of work in behind the scenes rules / prompts - just creating rules / prompts and expecting success liable to get all
I fail to get excited by people talking about "jail breaking AI" as my experience is that it is so frequently buggy that such issues are no surprise (did a few tests to see how good it was in censoring output in certain areas - not very good, far too easy to get racist, misogynist etc. content)
I think that's why there are so many job ads for "prompt engineering" - tacit admission of how bad "AI" is that it can actually be a skill getting it to do what you want.
Anon as "AI" API use work related (personally not a fan of "AI", but bills to pay, & like many companies, a big "AI" push from above as its the current trend)