European artificial intelligence bill is about to be introduced, which may affect global technology regulation

WBOY
Release: 2023-04-11 22:40:14
forward
1127 people have browsed it

The Artificial Intelligence Bill aims to create a common regulatory and legal framework for the use of artificial intelligence, including how it is developed, what companies can use it for, and the legal consequences of not complying with the requirements. The bill could require companies to obtain approval in certain circumstances before adopting artificial intelligence, outlaw certain uses of artificial intelligence deemed too risky, and create a public list of other high-risk uses of artificial intelligence.

European artificial intelligence bill is about to be introduced, which may affect global technology regulation

In an official report on the law, the European Commission said that at a broad level, the law seeks to transform the EU into a paradigm for trustworthy artificial intelligence. Legalization. This paradigm “requires that AI be legally, ethically and technologically strong, while respecting democratic values, human rights and the rule of law”.

The proposed regulatory framework for artificial intelligence has the following objectives:

1. Ensure that artificial intelligence systems deployed and used on the EU market are safe and respect existing legal fundamental rights and EU values ;

2. Ensure legal certainty to promote investment and innovation in AI;

3. Strengthen governance and effective enforcement of existing laws on fundamental rights and security as they apply to AI systems;

4. Promote the development of a unified market for legal, safe and trustworthy artificial intelligence applications and prevent market fragmentation.

While many details of the new law are still up in the air, key being the definition of artificial intelligence, a core element appears to be its "Product Security Framework," which divides future AI products into four security level categories, Applicable to all industries.

According to a 2021 European Commission report on the new law compiled by Maruritz Kop of the Stanford-Vienna Transatlantic Technology Law Forum, the bottom of the “criticality pyramid” is the least risky system. AI applications that fall into this category do not need to comply with the transparency requirements that riskier systems must pass. In the "limited risk" category, there are some transparency requirements for AI systems, such as chatbots.

European artificial intelligence bill is about to be introduced, which may affect global technology regulation

——The European Union’s new AI bill divides AI programs into four categories, from bottom to top: low risk, limited risk, high risk and unacceptable risk.

In high-risk AI categories with strict requirements for transparency, supervision will be more stringent. According to the introduction, high-risk categories include AI applications in:

  • critical infrastructure, such as transportation, which may endanger human life or health (such as self-driving cars);
  • affects personal education or education or vocational training for career achievement (such as test scores);
  • Safety components of products (such as robot-assisted surgery);
  • Employment, employee management, self-employment (such as resumes, resume classification Software);
  • Essential private and public services (such as credit scoring);
  • Law enforcement use cases (such as evidence assessment);
  • Immigration, border management (including passport authenticity assessment);
  • Fair and democratic processes (such as applying the law to a specific set of facts);
  • Surveillance systems (such as biometric surveillance and facial recognition).

The fourth category at the top of the tower is artificial intelligence systems with "unacceptable risk". These apps are basically illegal because they pose too many risks. Examples of such applications include systems to manipulate behavior or people or "specifically vulnerable groups," social scoring, and real-time and remote biometric systems.

High-risk categories are likely to be the focus of many companies’ efforts to ensure transparency and compliance. The EU’s Artificial Intelligence Bill will advise companies to take four steps before launching AI products or services that fall into high-risk categories.

  1. Develop high-risk AI systems, preferably in-house first, using an AI impact assessment and a code of conduct overseen by an inclusive, multidisciplinary team.
  2. High-risk artificial intelligence systems must undergo conformity assessment and approval and continue to comply with the requirements of the EU AI Act throughout their life cycle. Some organizations outside the system will be involved in conformity assessment audits. This dynamic process ensures benchmarking, monitoring and validation. In addition, for dynamic systems of high-risk AI, step 2 must be repeated.
  3. The registration of independent high-risk AI systems will take place in a dedicated EU database.
  4. A declaration of conformity must be signed and high-risk AI systems must bear the CE mark.

The above is the detailed content of European artificial intelligence bill is about to be introduced, which may affect global technology regulation. For more information, please follow other related articles on the PHP Chinese website!

Related labels:
source:51cto.com
Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn
Latest Downloads
More>
Web Effects
Website Source Code
Website Materials
Front End Template
About us Disclaimer Sitemap
php.cn:Public welfare online PHP training,Help PHP learners grow quickly!