Close Menu
Edu Expertise Hub
    Facebook X (Twitter) Instagram
    Wednesday, July 2
    • About us
    • Contact
    • Submit Coupon
    Facebook X (Twitter) Instagram YouTube
    Edu Expertise Hub
    • Home
    • Udemy Coupons
    • Best Online Courses and Software Tools
      • Business & Investment
      • Computers & Internet
      • eBusiness and eMarketing
    • Reviews
    • Jobs
    • Latest News
    • Blog
    • Videos
    Edu Expertise Hub
    Home » Latest News » Navigating the practicalities of AI regulation and legislation
    Latest News

    Navigating the practicalities of AI regulation and legislation

    TeamBy TeamDecember 27, 2024No Comments13 Mins Read0 Views
    Facebook Twitter Pinterest LinkedIn Telegram Tumblr Email
    Navigating the practicalities of AI regulation and legislation
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Misusing artificial intelligence (AI) can have some very clear and expensive consequences. Movie studio Lionsgate recently joined a long list of organisations discovering that quotations and citations from generative AI (GenAI) systems need to be verified like any other source; Microsoft is being sued by a German journalist after Bing Copilot suggested he had committed crimes he had instead reported on; and a US telecoms service is paying a $1m fine for simply transmitting automated calls featuring a fake AI voice mimicking President Biden.

    Enterprise enthusiasm for adopting GenAI remains high, meaning organisations are busy putting various governance, risk and compliance protections in place around their usage in different jurisdictions. While the main reason for restrictions on AI usage is frequently data privacy and security concerns, regulation and copyright concerns are also high on the list.

    Part of the problem for chief information officers (CIOs), however, is knowing exactly which regulations apply to AI, from the legal basis of using personal data to train AI models, to questions of transparency and discrimination when using AI systems.   

    Many organisations focus on upcoming legislation specifically designed to set rules for those developing and deploying AI systems, alongside a mix of regulations and voluntary guidelines for AI that can be individually useful, but make up what United Nations secretary-general António Guterres rather politely called a “patchwork” of potentially inconsistent rules.

    But the impact of new laws hasn’t yet been felt, and changes in government in the UK and US make it harder to predict what future legislation will dictate, especially for UK businesses caught between the US and the European Union (EU).

    Meanwhile, existing regulations that don’t explicitly mention AI already apply – and are being applied. This summer, the Brazilian data processing authority temporarily forced Meta to stop using “publicly available” information collected from its users to train AI models – on the basis of legitimate interests allowed under Brazilian legislation similar to the General Data Protection Regulation (GDPR). The company had to notify users in advance and provide easy ways to opt out.

    To safely navigate this web of regulations and laws – both upcoming and existing – that cover different stages of the development and deployment of AI, enterprises must therefore urgently get to grips with the direction of travel and appetite for enforcement in the countries they operate in.

    Evolving UK priorities

    Although there is likely to be an AI Bill in the UK, neither of the two private member’s bills making their way through the House of Lords are a reliable guide to what future legislation might look like.

    The government seems unlikely to take exactly the same “pro-innovation” approach to AI regulation as the previous one, especially as it’s signed up to the Council of Europe’s Framework Convention on AI and Human Rights (which covers the use of AI by governments and other public bodies).

    Currently a research organisation, it’s possible the UK AI Safety Institute may get a new role as an additional regulator, alongside the Information Commissioner’s Office (ICO), Ofcom, the Financial Conduct Authority (FCA) and the Competition and Markets Authority (CMA).

    The government report on Assuring a responsible future for AI envisions a commercial ecosystem of “AI assurance” tools to guide businesses using AI to mitigate risks and harms. It promises an AI Assurance Platform with a toolkit based on standards such as ISO/IEC 42001 (covering AI management systems), the EU AI Act and the NIST AI Risk Management Framework. Indeed, existing GRC tools such as Microsoft Purview Compliance Manager are already introducing reporting templates that cover these regulations.

    As for AI providers, the minister for the future digital economy and online safety, Baroness Jones, told the World Trade Organization’s AI conference the government will soon “bring forward highly targeted binding regulation on the handful of companies developing the most powerful AI systems”.

    Similarly, Ofcom recently issued an open letter to online service providers reminding them that the Online Safety Act (which imposes additional duties on search engines and messaging services starting in December 2024 with more beginning next year) also applies to GenAI models and chatbots. That’s aimed at the big AI platforms, but businesses using services to create chatbots, for example, for customer service, will want to test them thoroughly and make sure content safety guardrails are turned on.

    Disclosure requests

    The ICO has already requested fairly detailed disclosures from large platforms such as LinkedIn, Google, Meta, Microsoft and OpenAI about the data used to train their GenAI systems. Again, that’s covered by the Data Protection Act 2018, the UK implementation of GDPR. “Substantially right now, GDPR is regulating AI,” Lilian Edwards, director at Pangloss Consulting and professor of technology law at Newcastle University, told Computer Weekly.

    While the upcoming Data (Use and Access) Bill doesn’t include the sweeping reforms to data protection rights proposed by the previous government, it also doesn’t provide any extra clarity about the impact of UK GDPR on AI, beyond the existing guidance from the ICO, which makes it clear that senior management needs to understand and address the complex data protection implications of the technology.

    “The definition of personal data is now very, very wide: data that relates to a person who can be made identifiable,” warned Edwards. “Any AI company is almost certainly, if not intentionally, processing personal data.”

    More generally, she cautioned chief information officers not to dismiss legislation that doesn’t name AI specifically, noting that all of the compliance and risk management processes already in place apply to AI. “There are plenty of laws affecting companies that have been around for donkeys’ years that people aren’t as excited about: all the normal laws that apply to businesses. Discrimination and equality: are you in some way breaking the laws that are enforced by the Equality and Human Rights Commission? Are you breaking consumer rights by putting terms and conditions into your contracts?”

    Edwards further warned that AI systems delivering hallucinations can breach health and safety laws (like Amazon selling AI-generated books that misidentify poisonous mushrooms).

    Diya Wynn, responsible AI lead at Amazon Web Services, said: “We all have been very familiar and accustomed to having an understanding of sensitivity of data, whether it’s confidential, PII or PHI. That awareness of data and protecting data still is foundational, whether you’re using AI or not, and that should underpin internal policies. If you would not share PII or confidential or sensitive information normally, then you absolutely don’t want to do that in in in AI systems that you’re building or leveraging as well.”

    International implications

    Despite (or perhaps because of) being home to the major AI suppliers, the US has no comprehensive, AI-specific federal laws. Both the result of the election and the highly fragmented legal and regulatory setup (with multiple authorities at both state and federal level) make it unclear what, if any, legislation will emerge.

    The executive order on AI issued by President Biden called for frameworks, best practices and future legislation to ensure AI tools don’t break existing laws about discrimination, worker rights or how critical infrastructure and financial institutions handle risk.

    Although ostensibly broad, it actually applied more narrowly to government services and organisations that supply the US government, including assigning responsibilities for developing guardrails to NIST, which runs the US AI Safety Institute.

    As Paula Goldman, Salesforce chief ethical and humane use officer and a member of the national AI advisory committee that advises the US president, noted: “In a policy context, there are real, legitimate conversations about some of these bigger questions, like national security risk or bad actors. In a business setting, it is a different conversation.”

    That conversation is about data controls and “a higher level of attention to good governance, hygiene and documentation that can be discussed at a board level”.

    The executive order did specifically mention GenAI models trained on very large systems, but there are no existing systems with the specified level of resources.

    Many individual states have passed laws covering AI, some implementing the principles in the executive order, others regulating decisions in critical or sensitive areas made using GenAI without significant human involvement (again, GDPR already includes similar principles). California did introduce a sweeping range of legislation covering areas such as deepfakes and AI-generated “digital replicas”, of particular interest to Hollywood.

    The Republican campaign talked about replacing the executive order with AI built on free speech principles and the somewhat mysterious “human flourishing”. It’s unclear what that would look like in practice – especially given the conflicting interests of various donors and advisors – and organisations doing business in the US will need to deal with this patchwork of regulation for some time to come.

    The EU may set the tone

    On the other hand, the EU is the first jurisdiction to pass AI-specific legislation, with the EU AI Act being both the most comprehensive regulation of AI and the one that suppliers are actively preparing for. “We are specifically building into our products, in as much as we can, compliance with the EU AI Act,” Marco Casalaina, vice-president of product at Azure AI Platform, told Computer Weekly.

    That’s something Goldman expects businesses to welcome. “We have definitely heard a desire from companies to make sure that these regimes are interoperable, so there’s an overarching set of standards that can apply globally,” she said. “Honestly, it’s a lot to keep track of.”

    The Council of Europe’s AI Convention follows similar principles, and if this leads to global legislation aligning with it in the same way that GDPR influenced legislation worldwide, that will simplify compliance for many organisations.

    That’s by no means certain (there are already AI laws in place in Australia, Brazil and China), but any business operating in the EU, or with EU customers, will certainly need to comply with it. And unlike the mostly voluntary compliance approach of the executive order, the act comes with the usual penalties – in this case, of up to 3% of global turnover.

    The first thing to remember is that the new legislation applies in addition to a wide range of existing laws covering intellectual property, data protection and privacy, financial services, security, consumer protection, and antitrust.

    “It fits into this giant pack of EU laws, so it only covers really quite a small amount,” said Edwards.

    The act doesn’t regulate search, social media, or even recommender systems, which are meant to be dealt with by legislation enforcing the Digital Services Act.

    Rather than focusing on specific technologies, the EU AI Act is about demonstrating that AI products and services available in the EU comply with product safety requirements, which include data security and user privacy, in much the same way that the CE mark acts as a passport for selling physical products in the EU market.

    The act covers both traditional and generative AI; the latter provisions are clearly a work in progress mainly intended to apply to AI providers, who have to register in an EU database. As in the US executive order, general purpose models with “systemic risks” (like disinformation or discrimination) are categorised by the capacity of the system they were trained on, but at the level of the training systems already in use by providers such as Microsoft and OpenAI, rather than future systems yet to be built.

    The most oversight (or outright bans) applies to specific higher-risk uses, which are predominantly in the public sector, but also include critical infrastructure and employment.

    “If you are building a chatbot for vetting candidates for hiring,” Edwards warned, that would fall into the high-risk category. AI systems with limited risk (which includes chatbots) require transparency (including informing users about the AI system): those with minimal or no risk, such as spam filters, aren’t regulated by the act.

    These high-risk AI systems require conformity assessments she described as “a long list of very sensible machine learning training requirements: transparency, risk mitigation, looking at your training and testing data set, and cyber security. It’s got to be trained well, it’s got to be labelled well, it’s got to be representative, it’s got to be not full of errors.”

    AI providers have to disclose at least a summary of their training set as well as the risks involved in using these AI systems, “so you can see what risks have been mitigated and which ones have been flagged as not entirely mitigated”, said Edwards.

    The original intent of the act was to regulate traditional AI systems (not models) with a specific, intended use (like training, education or judicial sentencing) and making sure they stay fit for that purpose with tracking of updates and modifications. That makes less sense for GenAI, which is rarely designed for a single purpose, but the principle remains. Organisations deploying GenAI systems have fewer responsibilities than AI providers under the act, but if they make substantial enough modifications for these to count as a new system (including putting their name or trademark on it), they will also need a conformity assessment if they fall into a high-risk category.

    Code of practice

    It’s not yet clear whether fine-tuning, RAG or even changing from one large language model to another would count as a substantial modification, as the code of practice for “general purpose” GenAI systems is still being written.

    Until that guidance clarifies the position for businesses deploying GenAI (and also the situation for open weight models), “it’s not certain how far this will apply to everyday individuals or businesses”, Edwards told Computer Weekly.

    The act comes into force gradually over the next few years, with some provisions not in force until the end of 2030. Even if the EU takes the same approach as with GDPR (of enforcing the regulations very publicly on some larger suppliers to send a message), that may take some time – although multinational businesses may already be thinking about whether they use the same AI tools in all jurisdictions.

    But one thing all organisations using AI covered by the act will need to do very quickly, is make sure staff have enough training in responsible AI and regulatory compliance to have “AI literacy” by the February 2025 deadline.

    “It’s not just for your lawyers,” Edwards warns: anyone dealing with the operation and use of AI systems needs to understand the deployment of those systems, and the opportunities and risks involved.

    That kind of awareness is important for getting the most out of AI anyway, and getting ahead of the game rather than waiting for regulation to dictate safety measures makes good business sense.

    As Forrester vice-president Brandon Purcell put it: “Most savvy legal and compliance teams know that if there’s a problem with an experience that involves your brand, you’re going to be culpable in some way, maybe not in litigation, but certainly in the court of public opinion.”

    This post is exclusively published on eduexpertisehub.com

    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Team

      Related Posts

      BenQ showcases next-gen Google-integrated education solutions

      July 2, 2025

      The road to quantum datacentres goes beyond logical qubits

      July 1, 2025

      Block by Block: The Student Skilling Journey

      July 1, 2025

      Cleverlike focuses on the real power in using games for education

      June 30, 2025

      Scattered Spider cyber gang turns fire on aviation sector

      June 30, 2025

      Forget Prestige. A New Ranking Shows Great Colleges May Be Close to Home.

      June 29, 2025
      Courses and Software Tools

      Extreme Privacy: What It Takes to Disappear

      August 24, 202453 Views

      Modern C++ Programming Cookbook: Master Modern C++ with comprehensive solutions for C++23 and all previous standards

      September 18, 202427 Views

      Meebook E-Reader M7 | 6.8′ Eink Carta Screen | 300PPI Smart Light | Android 11 | Ouad Core Processor | Out Speaker | Support Google Play Store | 3GB+32GB Storage | Micro-SD Slot | Gray

      August 19, 202422 Views

      HR from the Outside In: Six Competencies for the Future of Human Resources

      May 20, 202517 Views

      Coders at Work: Reflections on the Craft of Programming

      April 19, 202516 Views
      Reviews

      Salesforce Cloud: A Complete Guide for Beginners | Udemy Coupons 2025

      July 2, 2025

      RN Registered Nurse | Fremont 4

      July 2, 2025

      Svelte Kit Bootcamp: Basic to Advance Build 2 Real Projects | Udemy Coupons 2025

      July 2, 2025

      BenQ showcases next-gen Google-integrated education solutions

      July 2, 2025

      RN Emergency Department

      July 2, 2025
      Stay In Touch
      • Facebook
      • YouTube
      • TikTok
      • WhatsApp
      • Twitter
      • Instagram
      Latest News

      BenQ showcases next-gen Google-integrated education solutions

      July 2, 2025

      The road to quantum datacentres goes beyond logical qubits

      July 1, 2025

      Block by Block: The Student Skilling Journey

      July 1, 2025

      Cleverlike focuses on the real power in using games for education

      June 30, 2025

      Scattered Spider cyber gang turns fire on aviation sector

      June 30, 2025
      Latest Videos

      Navigate Your Marketing Career with Expert Mentorship | NIMS Academy Success Guide

      July 1, 2025

      Inside the World of Ethical Hacking in 60 Seconds | Cybersecurity Career

      June 30, 2025

      The TRUTH About Finance Jobs After MBA

      June 29, 2025

      Restart Your Digital Marketing Career in 2024 Before It’s Too Late!

      June 28, 2025

      I Break FOLTYN’S WIN STREAK in Roblox Rivals! Rage! #roblox #rivals #shorts #foltyn #gaming

      June 27, 2025
      Latest Jobs

      RN Registered Nurse | Fremont 4

      July 2, 2025

      RN Emergency Department

      July 2, 2025

      Designer – Workplace

      July 1, 2025

      Line Cook (PRN)

      July 1, 2025

      Real Estate Insights Writer

      July 1, 2025
      Legal
      • Home
      • Privacy Policy
      • Cookie Policy
      • Terms and Conditions
      • Disclaimer
      • Affiliate Disclosure
      • Amazon Affiliate Disclaimer
      Latest Udemy Coupons

      Mastering Maxon Cinema 4D 2024: Complete Tutorial Series | Udemy Coupons 2025

      August 22, 202435 Views

      Advanced Program in Human Resources Management | Udemy Coupons 2025

      April 5, 202531 Views

      Diploma in Aviation, Airlines, Air Transportation & Airports | Udemy Coupons 2025

      March 21, 202530 Views

      Python Development & Data Science: Variables and Data Types | Udemy Coupons 2025

      May 24, 202521 Views

      Time Management and Timeboxing in Business, Projects, Agile | Udemy Coupons 2025

      April 2, 202521 Views
      Blog

      Why Community Is Your Most Valuable Career Asset In 2025

      June 28, 2025

      What Employers Are Really Looking For In Job Interviews

      June 27, 2025

      The Best Way to End a Cover Letter (With 4 Winning Examples)

      June 26, 2025

      5 Job Interview Secrets To Beat The Competition

      June 25, 2025

      10 Overused LinkedIn Buzzwords (And What To Say Instead)

      June 24, 2025
      Facebook X (Twitter) Instagram Pinterest YouTube Dribbble
      © 2025 All rights reserved!

      Type above and press Enter to search. Press Esc to cancel.

      We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
      .
      SettingsAccept
      Privacy & Cookies Policy

      Privacy Overview

      This website uses cookies to improve your experience while you navigate through the website. Out of these, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may affect your browsing experience.
      Necessary
      Always Enabled
      Necessary cookies are absolutely essential for the website to function properly. This category only includes cookies that ensures basic functionalities and security features of the website. These cookies do not store any personal information.
      Non-necessary
      Any cookies that may not be particularly necessary for the website to function and is used specifically to collect user personal data via analytics, ads, other embedded contents are termed as non-necessary cookies. It is mandatory to procure user consent prior to running these cookies on your website.
      SAVE & ACCEPT