A big week for AI - the political reaction
Professor Keiichi Nakata, Director of AI at the World of Work Institute, discusses the AI Safety Summit happening at Bletchley park and other international attempts to regulate the use of AI.
There was a mixture of high expectations and a feeling of fatigue in the run up to AI Safety Summit. There are some hopes that the Summit that brings together major players in tech as well as governments will lead to some sort of consensus in relation to the “safe” use of AI, or at least kick start the global conversation. On the other hand, a lot of discussions and debates have already been taking place, with varying perspectives and emphases, creating a feeling that we have been here before.
The significance of this event, however, is that it is happening.
It is no surprise that the US issued the Executive Order on Safe, Secure, and Trustworthy Artificial Intelligence just before the Summit to reaffirm their presence and their position in this debate. To some, this may feel too political and the adoption of the term “frontier models” presumably proposed by the Frontier Model Forum started by Microsoft, Anthropic, Google, and OpenAI somewhat too tech-giant-driven – but whatever the specific interest of each party might be, it demonstrates a strong political will and delivers the message that countries and businesses are taking it seriously.
The EU’s AI Act and the US Blueprint for an AI Bill of Rights issued in 2022 (among others including Google’s responsible AI principles) have already set out basic principles of responsible use of AI. In terms of the "what", a set of key principles is needed, and no one will dispute that any technology needs to be used responsibly. In terms of "how", EU AI Act has chosen the risk-based approach, which is also difficult to argue against. What is refreshing about this Executive Order is that it clearly indicates “who” should be doing “what”. There could be debates about the suitability of the agencies to take on the specific tasks but at least it is pragmatic, and some actions will be taken beyond principles and debates. Intentionally or not, in the Executive Order the word “regulate” only occurs once. Some see regulation as a hindrance to innovation, but it has many benefits including providing confidence and safety for the technology – it is the operationalisation of regulations such as “red tape” that slows down developments rather than regulation itself.
The current explosion of interest in AI research brings back memories of the ‘70s and ‘80s of the likes of DARPA in US, Alvey in UK, Esprit in Europe and Japan’s Fifth Generation Computing projects, which were heavily backed by governments. These are now often associated with the onset of an “AI Winter” by not delivering what people expected, not to mention that there were hardly considerations on responsible AI. But the business environment in which AI technologies are being used and developed is markedly different from those times. The AI Safety Summit and similar global conversations need to deliver tangible results to maximise the benefits of AI technologies as public goods.
You might also like
A very different Black Friday experience
The Good, the Bad and the Ugly Internet of Things
Twitter’s Musk-era: A defence of remote working
This site uses cookies to improve your user experience. By using this site you agree to these cookies being set. You can read more about what cookies we use here. If you do not wish to accept cookies from this site please either disable cookies or refrain from using the site.