, , , , ,

UK’s approach to AI safety lacks credibility, report warns

The UK government has been trying to cultivate an image of itself as an international mover-and-shaker in the nascent field of AI safety in recent weeks — dropping a flashy announcement of an upcoming summit on the topic last month, along with a pledge to spend £100M on a foundational model taskforce that will do “cutting edge” AI safety research, as it tells it.

Yet the self-same government, led by UK prime minister and Silicon Valley superfan Rishi Sunak, has eschewed the need to pass new domestic legislation to regulate applications of AI — a stance its own policy paper on the topic brands “pro-innovation”.

It is also in the midst of passing a deregulatory reform of the national data protection framework that risks working against AI safety.

The latter is one of several conclusions by the independent research-focused Ada Lovelace Institute, a part of the Nuffield Foundation charitable trust, in a new report examining the UK’s approach to regulating AI that makes for diplomatic-sounding but, at times, pretty awkward reading for ministers.

The report packs a full 18 recommendations for levelling up government policy/credibility in this area — i.e. if the UK wants to be taken seriously on the topic.

The Institute advocates for an “expensive” definition of AI safety — “reflecting the wide variety of harms that are arising as AI systems become more capable and embedded in society”. So the report is concerned with how to regulate “AI systems can cause today”. Call them real-world AI harms. (Not with sci-fi-inspired theoretical possible future risks which have been puffed up by certain high profile figures in the tech industry of late, seemingly in a bid to attention-hack policymakers.)

For now, it’s fair to say Sunak’s government’s approach to regulating (real-world) AI safety has been contradictory; heavy on flashy, industry-led PR claiming it wants to champion safety but light on policy proposals for setting substantive rules to guard against the smorgasbord of risks and harms we know can flow from ill-judged applications of automation.

Here’s the Ada Lovelace Institute dropping the primary truth bomb:

The UK Government has laid out its ambition to make the UK an ‘AI superpower’, leveraging the development and proliferation of AI technologies to benefit the UK’s society and economy, and hosting a global summit in autumn 2023. This ambition will only materialise with effective domestic regulation, which will provide the platform for the UK’s future AI economy.

The report’s laundry list of recommendations goes on to make it clear the Institute sees a lot of room for improvement on the UK’s current approach to AI. 

Earlier this year the government published its preferred approach to regulating AI domestically — saying it didn’t see the need for new legislation or oversight bodies at this stage. Instead the white paper offered a set of flexible principles the government suggested existing, sector specific (and/or cross-cutting) regulators should “interpret and apply to AI within their remits”. Just without any new legal powers nor extra funding for also overseeing novel uses of AI.

The five principles set out in the white paper are: Safety, security and robustness; Appropriate transparency and explainability; Fairness; Accountability and governance; Contestability and redress. All of which sounds fine on paper — but paper alone clearly isn’t going to cut it when it comes to regulating AI safety.

The UK’s plan to let existing regulators figure out what to do about AI with just some broad-brush principles to aim for and no new resource contrasts with that of the EU where lawmakers are busy hammering out agreement on a risk-based framework which the bloc’s executive proposed back in 2021.

The UK’s shoestring budget approach of saddling existing, overworked regulators with new responsibilities for eyeing AI developments on their patch without any powers to enforce outcomes on bad actors doesn’t look very credible on AI safety to put it mildly.

It doesn’t even seem a coherent strategy if you’re shooting for being pro-innovation, either — since it will demand AI developers consider a whole patchwork of sector-specific and cross-cutting legislation, drafted long before the latest AI boom. Developers may also find themselves subject to oversight by a number of different regulatory bodies (however weaksauce their attention might be, given the lack of resource and legal firepower to enforce the aforementioned principles). So, really, it looks like a recipe for uncertainty over which existing rules may apply to AI apps. (And, most probably, a patchwork of regulatory interpretations, depending on the sector, use-case and oversight bodies involved etc. Ergo, confusion and cost, not clarity.)

Even if existing UK regulators do quickly produce guidance on how they will approach AI — as some already are or are working to — there will still be plenty of gaps, as the Ada Lovelace Institute’s report also points out — since coverage gaps are a feature of the UK’s existing regulatory landscape. So the proposal to just further stretch this approach implies regulatory inconsistency getting baked in and even amplified as usage of AI scales/explodes across all sectors. 

Here’s the Institute again:

Large swathes of the UK economy are currently unregulated or only partially regulated. It is unclear who would be responsible for implementing AI principles in these contexts, which include: sensitive practices such as recruitment and employment, which are not comprehensively monitored by regulators, even within regulated sectors; public-sector services such as education and policing, which are monitored and enforced by an uneven network of regulators; activities carried out by central government departments, which are often not directly regulated, such as benefits administration or tax fraud detection; unregulated parts of the private sector, such as retail.

“AI is being deployed and used in every sector but the UK’s diffuse legal and regulatory network for AI currently has significant gaps. Clearer rights and new institutions are needed to ensure that safeguards extend across the economy,” it also suggests.

Another growing contradiction for the government’s claimed “AI leadership” position is that its bid for the country to become a global AI safety hub is being directly undermined by in-train efforts to water down domestic protections for people’s data — such as by lowering protections when they’re subject to automated decisions with a significant and/or legal impact — via the deregulatory Data Protection and Digital Information Bill (No. 2).

While the government has so far avoided the most headbanging Brexiteer suggestions for ripping up the EU-derived data protection rulebook — such as simply deleting the entirety of Article 22 (which deals with protection for automated decisions) from the UK’s General Data Protection Regulation — it is nonetheless forging ahead with a plan to reduce the level of protection citizens enjoy under current data protection law in various ways, despite its new ambition to make the UK a global AI safety hub.

“The UK GDPR — the legal framework for data protection currently in force in the UK — provides protections that are vital to protecting individuals and communities from potential AI harms. The Data Protection and Digital Information Bill (No. 2), tabled in its current form in March 2023, significantly amends these protections,” warns the Institute, pointing for example to the Bill removing a prohibition on many types of automated decision — and instead requiring data controllers to have “safeguards in place, such as measures to enable an individual to contest the decision” — which it argues is, a lower level of protection in practice.

“The reliance of the Government’s proposed framework on existing legislation and regulators makes it even more important that underlying regulation like data protection governs AI appropriately,” it goes on. “Legal advice commissioned by the Ada Lovelace Institute… suggests that existing automated processing safeguards may not in practice provide sufficient protection to people interacting with everyday services, like applying for a loan.”

“Taken collectively, the Bill’s changes risk further undermining the Government’s regulatory proposals for AI,” the report adds.

The Institute’s first recommendation is thus for government to rethink elements of the data protection reform bill that are “likely to undermine the safe development, deployment and use of AI, such as changes to the accountability framework”. It also recommends the government widens its review to look at existing rights and protections in UK law — with a view to plugging any other legislative gaps and introducing new rights and protections for people affected by AI-informed decisions where necessary.

Other recommendations in the report include introducing a statutory duty for regulators to have regard to the aforementioned principles, including “strict transparency and accountability obligations” and providing them with more funding/resource to tackle AI-related harms; exploring the introduction of a common set of powers for regulators, including an ex ante, developer-focused regulatory capability; and that the government should look at whether an AI ombudsperson should be established to support people aversely affected by AI.

The Institute also recommends the government clarify the law around AI and liability — which is another area where the EU is already streaks ahead.

On foundational model safety — an area that’s garnered particular interest and attention from the UK government of late, thanks to the viral buzz around generative AI tools like OpenAI’s ChatGPT — the Institute also believes the government needs to go further, recommending UK-based developers of foundational models should be given mandatory reporting requirements to make it easier for regulators to stay on top of a very fast-moving tech.

It even suggests leading foundational model developers, such as OpenAI, Google DeepMind and Anthropic, should be required to provide government with notification when they (or any subprocessors they’re working with) begin large-scale training runs of new models.

“This would provide Government with an early warning of advancements in AI capabilities, allowing policymakers and regulators to prepare for the impact of these developments, rather than being caught unaware,” it suggests, adding that reporting requirements should also include information such as access to the data used to train models; results from in-house audits; and supply chain data.

Another suggestion is for the government to invest in small pilot projects to bolster its own understanding of trends in AI R&D.

Commenting on the report findings in a statement, Michael Birtwistle, associate director at the Ada Lovelace Institute, said:

The Government rightfully recognises that the UK has a unique opportunity to be a world-leader in AI regulation and the prime minister should be commended for his global leadership on this issue. However, the UK’s credibility on AI regulation rests on the Government’s ability to deliver a world-leading regulatory regime at home. Efforts towards international coordination are very welcome but they are not sufficient. The Government must strengthen its domestic proposals for regulation if it wants to be taken seriously on AI and achieve its global ambitions.

https://techcrunch.com/2023/07/17/ada-lovelace-institute-report-on-uk-regulating-ai/


January 2025
M T W T F S S
 12345
6789101112
13141516171819
20212223242526
2728293031  

About Us

Welcome to encircle News! We are a cutting-edge technology news company that is dedicated to bringing you the latest and greatest in everything tech. From automobiles to drones, software to hardware, we’ve got you covered.

At encircle News, we believe that technology is more than just a tool, it’s a way of life. And we’re here to help you stay on top of all the latest trends and developments in this ever-evolving field. We know that technology is constantly changing, and that can be overwhelming, but we’re here to make it easy for you to keep up.

We’re a team of tech enthusiasts who are passionate about everything tech and love to share our knowledge with others. We believe that technology should be accessible to everyone, and we’re here to make sure it is. Our mission is to provide you with fun, engaging, and informative content that helps you to understand and embrace the latest technologies.

From the newest cars on the road to the latest drones taking to the skies, we’ve got you covered. We also dive deep into the world of software and hardware, bringing you the latest updates on everything from operating systems to processors.

So whether you’re a tech enthusiast, a business professional, or just someone who wants to stay up-to-date on the latest advancements in technology, encircle News is the place for you. Join us on this exciting journey and be a part of shaping the future.

Podcasts

TWiT 1013: Calamari in Crisis – Touching the Sun, Fake Spotify Artists, Banished Words This Week in Tech (Audio)

Touching the Sun, Fake Spotify Artists, Banished Words AI Needs So Much Power, It's Making Yours Worse How many billions Big Tech spent on AI data centers in 2024 NASA Spacecraft 'Touches Sun' In Defining Moment For Humankind Elon Musk Calls Out NASA's Moon Ambitions: 'We're Going Straight to Mars' Elon Musk and the right's war on Wikipedia Trump Asks Supreme Court to Pause Law Threatening TikTok Ban US Treasury says Chinese hackers stole documents in 'major incident' Judge blocks parts of California bid to protect kids from social media Finland probes Russian shadow fleet oil tanker after cable-cutting incident US appeals court blocks Biden administration effort to restore net-neutrality rules The Ghosts in the Machine (fake spotify artists) Massive VW Data Leak Exposed 800,000 EV Owners' Movements, From Homes To Brothels Banished Words | Lake Superior State University 2025 Public Domain Day 2025 Happy Birthday, Bitcoin! The top cryptocurrency is old enough to drive End of the lines? QR-style codes could replace barcodes 'within two years' Host: Leo Laporte Guests: Richard Campbell, Anthony Ha, and Stacey Higginbotham Download or subscribe to This Week in Tech at https://twit.tv/shows/this-week-in-tech Get episodes ad-free with Club TWiT at https://twit.tv/clubtwit Sponsors: ZipRecruiter.com/Twit joindeleteme.com/twit promo code TWIT canary.tools/twit – use code: TWIT zscaler.com/security
  1. TWiT 1013: Calamari in Crisis – Touching the Sun, Fake Spotify Artists, Banished Words
  2. TWiT 1012: Our Best Of 2024 – The Best Moments From TWiT's 2024
  3. TWiT 1011: The Year in Review – A Look at the Top Stories of 2024
  4. TWiT 1010: The Densest State in the US – TikTok Ban, Drones Over Jersey, GM Quits Robotaxis
  5. TWiT 1009: Andy Giveth & Bill Taketh Away – Trump's Tech Titans, Crypto Boom, TikTok's US Ban, Intel CEO Exits