Skip to main content

A big week for AI - the political reaction

Network 6511448 1280

There was a mixture of high expectations and a feeling of fatigue in the run up to AI Safety Summit. There are some hopes that the Summit that brings together major players in tech as well as governments will lead to some sort of consensus in relation to the “safe” use of AI, or at least kick start the global conversation. On the other hand, a lot of discussions and debates have already been taking place, with varying perspectives and emphases, creating a feeling that we have been here before.

The significance of this event, however, is that it is happening.

It is no surprise that the US issued the Executive Order on Safe, Secure, and Trustworthy Artificial Intelligence just before the Summit to reaffirm their presence and their position in this debate. To some, this may feel too political and the adoption of the term “frontier models” presumably proposed by the Frontier Model Forum started by Microsoft, Anthropic, Google, and OpenAI somewhat too tech-giant-driven – but whatever the specific interest of each party might be, it demonstrates a strong political will and delivers the message that countries and businesses are taking it seriously.

The EU’s AI Act and the US Blueprint for an AI Bill of Rights issued in 2022 (among others including Google’s responsible AI principles) have already set out basic principles of responsible use of AI. In terms of the "what", a set of key principles is needed, and no one will dispute that any technology needs to be used responsibly. In terms of "how", EU AI Act has chosen the risk-based approach, which is also difficult to argue against. What is refreshing about this Executive Order is that it clearly indicates “who” should be doing “what”. There could be debates about the suitability of the agencies to take on the specific tasks but at least it is pragmatic, and some actions will be taken beyond principles and debates. Intentionally or not, in the Executive Order the word “regulate” only occurs once. Some see regulation as a hindrance to innovation, but it has many benefits including providing confidence and safety for the technology – it is the operationalisation of regulations such as “red tape” that slows down developments rather than regulation itself.

The current explosion of interest in AI research brings back memories of the ‘70s and ‘80s of the likes of DARPA in US, Alvey in UK, Esprit in Europe and Japan’s Fifth Generation Computing projects, which were heavily backed by governments. These are now often associated with the onset of an “AI Winter” by not delivering what people expected, not to mention that there were hardly considerations on responsible AI. But the business environment in which AI technologies are being used and developed is markedly different from those times. The AI Safety Summit and similar global conversations need to deliver tangible results to maximise the benefits of AI technologies as public goods.

Professor Keiichi Nakata

Head of Business Informatics, Systems and Accounting (BISA)
Published 2 November 2023
Topics:
Leading insights AI and automation

You might also like

A very different Black Friday experience

25 November 2020
Professor Emerita Susan Rose looks at how the annual retail frenzy will be affected by the pandemic and the changing retail environment.
Leading insights

The Good, the Bad and the Ugly Internet of Things

29 July 2022
From vacuum cleaners to lights to security systems, what is the Internet of Things and how can it be used? Dr Vaughan Michell has a look at some of the benefits and pitfalls in our latest Leading Insights.
Leading insights AI and automation

Twitter’s Musk-era: A defence of remote working

15 November 2022
Dr Miriam Marra looks at the impact Elon Musk's back-to-office policy will likely have on Twitter's top talent and productivity.
Leading insights Flexible working