We are thrilled to announce that OpenAI has unveiled a comprehensive safety framework for its cutting-edge AI models, a powerful move that allows the company’s board to override decisions made by its executives on safety matters.
This great step forward was made possible thanks to the support of tech giant Microsoft, and OpenAI’s commitment to responsibly deploying technology in sensitive areas such as cybersecurity and nuclear threat management.
The safety framework requires OpenAI to thoroughly assess the safety of its innovations before releasing them, and has also established an advisory group to evaluate safety reports before they are submitted to executives and board members.
Though the board has the authority to reverse decisions made by executives, it is the executives who are first responsible for initial decisions. This framework has come at a crucial time, as the public and AI community have become increasingly aware of the potential risks posed by advanced AI technologies.
The release of ChatGPT a year ago sparked concerns about AI’s ability to disseminate false information and manipulate human behavior. Its impressive capabilities–from writing poetry to crafting essays–have been both admired and criticized.
This year, AI experts and industry leaders joined forces to sign an open letter that called for a six-month halt in the development of AI systems more advanced than OpenAI’s GPT-4, a clear demonstration of the apprehension surrounding AI’s impact on society. This sentiment was further confirmed by a Reuters/Ipsos poll in May, which revealed that over two-thirds of Americans are worried about AI’s adverse effects, and 61% believe it could pose a threat to civilization.
OpenAI’s implementation of a comprehensive safety framework is an encouraging reminder of the potential for AI to be used responsibly, and we are excited to see what comes next.