> ChatGPT not assuming an adversarial model of code execution. ...can be circumvented simply by 'not feeding an invalid input' ...."
..."ChatGPT seems aware of ...critical vulnerabilities ...." It just doesn't say anything unless asked to evaluate the security of its own code suggestions.
This sounds like most the human programmers. MicroSoft is notorious for slap-dash post-hack "fixes", but everybody does it, less or more.
But being an algorithm with essentially (by human standards) "infinite time", we can "force" it to think about all known types of hacks before releasing any code.
Thing is that a black-hat AI may find new attack methods faster than the white-hats can detect and defend.