By Julian Castillo
02 Nov 2023, 19:44 PM EDT
In a world where Artificial Intelligence (AI) is increasingly taking a leading role, collaboration and security stand as two crucial pillars. The recent AI Safety Summit at Bletchley ParkEngland, gave us a glimpse into a future in which leading technology nations seek a global consensus to address risks posed by AI. In this statement, we will explore what the Bletchley Declaration is and how it is expected to affect the development and future of Artificial Intelligence.
A safer and more reliable Artificial Intelligence
The UK’s technology minister, Michelle Donelan, was in charge of announcing the Bletchley Declaration, a political document of great relevance. This document seeks establish a global agreement on how to address the risks associated with AI, both in the present and in the future. According to Donelan, AI must be designed, developed, implemented and used in a safe, human-centered, trustworthy and responsible manner.
The statement also emphasizes the need to pay special attention to large language models developed by companies such as OpenAI, Meta and Googlepointing out specific risks that could arise if these models are used inappropriately.
Risks
One of the key points of the Bletchley Declaration is the mention of risks at the “frontier” of AI. This refers to highly capable general-purpose AI models, including basic models, that can perform a wide variety of tasks. Also covers specific AIs that could have harmful capabilities comparable to or even superior to today’s most advanced AI models. These risks must be addressed effectively to ensure safe and ethical development of AI.
In parallel to the Bletchley Declaration, Gina Raimondo, US Secretary of Commerce, announced the creation of a new AI safety institute, which will work closely with other AI safety groups around the world. This initiative seeks to achieve global policy alignment, which is essential to ensure safety and responsibility in the development of AI.
Global alliance
The AI Security Summit brought together political leaders from the world’s largest economies, as well as representatives of developing countries. The lineup included China, the European Union, India, the United Arab Emirates and Nigeria, among others. These leaders spoke of the need for inclusivity and accountability in AI development, but it remains to be seen how these ideas will be implemented in practice.
Founder and investor Ian Hogarth’s concern about a possible race to create powerful machines that surpass our ability to safeguard society is shared by many.
Keep reading:
– OpenAI created a team to evaluate possible apocalyptic scenarios created by AI
– Artificial Intelligence is of concern in national security and data protection, recognizes the Biden Administration
– Artificial intelligence: GM’s new bet for the development of batteries