OpenAI's Strategy to Safeguard Super-Intelligent AI Revealed

OpenAI's Superalignment research team is making progress in developing methods to control super-intelligent AI systems. The team conducted experiments to allow an inferior AI model to guide the behavior of a more advanced one without diminishing its capabilities. They tested two approaches, including training progressively larger models and adding an algorithmic tweak to the stronger model. While these methods are not foolproof, they serve as a starting point for further research. OpenAI is also offering $10 million in grants and partnering with Eric Schmidt to encourage outside researchers to contribute to advancements in AI control. The company plans to hold a conference on superalignment next year.
- OpenAI's Ilya Sutskever Has a Plan for Keeping Super-Intelligent AI in Check WIRED
- OpenAI thinks superhuman AI is coming — and wants to build tools to control it TechCrunch
- Now we know what OpenAI's superalignment team has been up to MIT Technology Review
- OpenAI Demos a Control Method for Superintelligent AI IEEE Spectrum
Reading Insights
0
1
5 min
vs 6 min read
90%
1,039 → 106 words
Want the full story? Read the original article
Read on WIRED