Experts Call for a Pause in Developing Powerful AI Systems
Over the past few years, the development of advanced artificial intelligence (AI) systems has raised concerns about their impact on society and humanity. Recently, a group of AI experts and industry executives, including Elon Musk, called for a six-month pause in training AI systems that are more powerful than OpenAI’s newly launched model GPT-4, citing potential risks to society and civilization.
In an open letter issued by the non-profit Future of Life Institute and signed by over 1,000 people, including researchers at Alphabet-owned DeepMind, Yoshua Bengio, Stuart Russell, and Emad Mostaque, CEO of Stability AI, the group emphasized the need for shared safety protocols for advanced AI development that must be developed, implemented and audited by independent experts.
The letter highlights that powerful AI systems should only be developed once experts are confident that their effects will be positive and risks will be manageable. The letter further calls on developers to work with policymakers on governance and regulatory authorities to mitigate potential risks.
Potential Risks of Human-Competitive AI Systems
The letter also warns about the potential risks to society and civilization by human-competitive AI systems. Economic and political disruptions could be caused by these systems in the form of phishing attempts, disinformation, and cybercrime.
The concerns raised by the group of experts come at a time when EU police force Europol joined in on the ethical and legal concerns over advanced AI like ChatGPT. They warned about the potential misuse of the system, highlighting the need for caution when developing these systems.
While the release of OpenAI’s ChatGPT has prompted rivals to accelerate the development of similar large language models, companies must be mindful of the risks that come with AI systems’ increased power. The group of experts stresses the importance of developing shared safety protocols that prioritize positive outcomes and risk management.
Gary Marcus, a professor at New York University who signed the letter, commented that “The letter isn’t perfect, but the spirit is right: we need to slow down until we better understand the ramifications. They can cause serious harm, and the big players are becoming increasingly secretive about what they are doing, which makes it hard for society to defend against whatever harms may materialize.”
In Conclusion
As AI technology continues to advance rapidly, it is crucial to prioritize shared safety protocols to ensure that the development of powerful AI systems does not result in negative consequences for society and civilization. The group of experts calling for a pause in AI development is a step in the right direction, and policymakers and developers must work together to develop governance and regulatory authorities that prioritize positive outcomes and risk management.