British AI Gatekeepers with ‘Safety Guarantees’ to be Developed by DARPA

Admin

British AI Gatekeepers with ‘Safety Guarantees’ to be Developed by DARPA

AI gatekeepers, British DARPA, safety guarantees



In the world of artificial intelligence (AI), ensuring safety is paramount. With the rapid advancement of AI technology, concerns about the potential harm caused by AI systems have become increasingly prevalent. To address this issue, the Advanced Research and Invention Agency (ARIA), a British R&D unit, has proposed a revolutionary concept called “quantitative safety guarantees.”

ARIA’s vision of quantitative safety guarantees draws inspiration from the stringent safety standards implemented in nuclear power plants and passenger aviation. The idea is to establish probabilistic guarantees that no harm will result from a particular action taken by AI systems. This approach aims to provide a level of assurance and accountability in the deployment of AI technology.

At the heart of ARIA’s plan is the concept of a “gatekeeper” AI. This digital sentinel would act as a protector, ensuring that other AI agents operate within predefined boundaries set for each specific application. The gatekeeper AI would monitor and regulate the actions of AI systems, minimizing the risk of unintended consequences or harm.

To bring this vision to life, ARIA has allocated £59 million towards the project. By the end of the program, the agency aims to demonstrate a scalable proof-of-concept in one specific domain. Possible applications include electricity grid balancing and supply chain management – areas where the potential impact of AI is significant.

The success of this project could have far-reaching implications for high-stakes AI applications. Critical infrastructure improvement and optimizing clinical trials are just a few examples of the areas that could benefit from the safety guarantees provided by ARIA’s gatekeeper approach. By mitigating risks and ensuring the responsible use of AI, these applications can operate with greater confidence and trust.

Leading the initiative is David ‘davidad’ Dalrymple, who co-invented the popular cryptocurrency Filecoin. Dalrymple’s extensive research and expertise in technical AI safety have shaped his interest in the gatekeeper approach. As the program director of ARIA, Dalrymple now has the opportunity to turn his theory into practical application.

ARIA’s gatekeepers will rely on scientific world models and mathematical proofs. This combination of commercial and academic concepts is a unique approach that aims to strike a balance between cutting-edge advancements and rigorous mathematical reasoning. Dalrymple stresses the importance of avoiding an overemphasis on academic approaches, as they risk slowing down progress, while solely relying on finite samples offered by big AI companies can fall short of providing sufficient guarantees.

Deep interdisciplinary collaboration is a critical aspect of ARIA’s gatekeeper approach. ARIA facilitates this collaboration, bringing together experts from various fields to tackle the challenges of AI safety. By fostering cooperation between academia and industry, ARIA aims to leverage the best of both worlds and drive advancements in AI technology while ensuring safety and ethical considerations are at the forefront.

ARIA’s approach has drawn comparisons to DARPA, the U.S. Department of Defense’s research agency known for its “mad science” projects. Like DARPA, ARIA focuses on high-risk, high-reward research endeavors. Dalrymple likens ARIA’s project to DARPA’s HACMS program, which successfully created an unhackable quadcopter through formal verification. ARIA’s gatekeeper project aims to achieve similar success in the AI domain, providing quantitative safety guarantees for AI systems.

Yoshua Bengio, a renowned computer scientist and Turing Award recipient, has also championed the necessity of quantitative safety guarantees. In a blog post last year, Bengio lamented the lack of progress in obtaining such guarantees for current AI systems. Dalrymple’s plan aligns with Bengio’s vision, offering hope for achieving the desired level of safety assurance in AI deployments.

While ARIA’s project holds great promise, it has not been without its critics. Some lawmakers have raised concerns about the agency’s budget, questioning the £800 million funding allocated over five years. Despite this scrutiny, ARIA can also point to potential cost savings associated with its initiatives. One program launched by ARIA aims to train AI systems at only 0.1% of the current cost, suggesting that the agency is actively finding ways to maximize efficiency and drive advancements without incurring exorbitant expenses.

In conclusion, ARIA’s vision of quantitative safety guarantees for AI represents a significant step forward in addressing the safety concerns surrounding AI technology. By implementing gatekeeper AI systems and leveraging interdisciplinary collaboration, ARIA aims to provide assurances that AI applications operate within defined boundaries, minimizing the risk of harm. With David Dalrymple and his team at the helm, there is optimism that this innovative approach can pave the way for the responsible and safe deployment of AI technology across various domains.



Source link

Leave a Comment