"Jailbreaking" is a term used to describe the bypassing of restrictions imposed by the maker of a product, allowing the user to perform actions not usually allowed. In the context of AI, a new method has been discovered to "jailbreak" ChatGPT, otherwise known as Bing Chat (the world's most popular Large Language Model).
The technique…