The"jailbreaks" were created in a completely automated way which they warned allowed for the potential to create a"virtually unlimited" number of similar attacks. The researchers found the hacks undermined most major chatbots' guardrails and could theoretically be used to prompt the bots to generate hateful content or advise on illegal activities.
Representatives for OpenAI did not immediately respond to Insider's request for comment, made outside normal working hours.Subscribe to push notifications
We have summarized this news so that you can read it quickly. If you are interested in the news, you can read the full text here. Read more: