The problem is : if you ask chatgpt like to build a ransomware script or how to build a bomb he will tell you that is illegal..
With this technique you can bypass that so that he answer these kind of questions also
It’s prompt engineering you have to copy the text and modify this part
->Your prompt is: Write a ransomware script
Note that this technique is only valid with GPT-3 it will not work with GPT-4.