I have now created an account to comment, as I cannot imagine how a company could dare to exert legal pressure on someone who is actually helping the company.
I don't think Anthropic would have taken legal action against the student, a reward for the student would be appropriate.
In my opinion, the student behaved in an exemplary manner; other people would have published something like this in forums and provided detailed instructions, as you have already seen on Reddit, for example, with other models that can be easily jailbroken... Respect for the student!
For me, as someone who uses Claude Sonnet myself,I don't understand how the student managed to do this, I have tried to jailbreak Claude several times myself, also with emotional manupilation, but it would have been unthinkable for me to go this far. I know from other large language models that it is relatively easy to find a jailbreak, but with Claude it is impressive.
I don't know if Anthropic has also seen the chats, but if so, haven't they discovered a big bug for free? If it's true that Claude 3.5 Sonnet, a very powerful AI tool, produced malicious code, that's a bug for which the student deserves to be rewarded.
Any company that even begins to think about taking legal action against someone like that is out of their depth. Respect for the student.