Innovation, Technology & Law

Blog over Kunstmatige Intelligentie, Quantum, Deep Learning, Blockchain en Big Data Law

Blog over juridische, sociale, ethische en policy aspecten van Kunstmatige Intelligentie, Quantum Computing, Sensing & Communication, Augmented Reality en Robotica, Big Data Wetgeving en Machine Learning Regelgeving. Kennisartikelen inzake de EU AI Act, de Data Governance Act, cloud computing, algoritmes, privacy, virtual reality, blockchain, robotlaw, smart contracts, informatierecht, ICT contracten, online platforms, apps en tools. Europese regels, auteursrecht, chipsrecht, databankrechten en juridische diensten AI recht.

Berichten met de tag human rights
EU Artificial Intelligence Act: The European Approach to AI

Stanford - Vienna Transatlantic Technology Law Forum, Transatlantic Antitrust and IPR Developments, Stanford University, Issue No. 2/2021

New Stanford tech policy research: “EU Artificial Intelligence Act: The European Approach to AI”.

Download the article here: Kop_EU AI Act: The European Approach to AI

EU regulatory framework for AI

On 21 April 2021, the European Commission presented the Artificial Intelligence Act. This Stanford Law School contribution lists the main points of the proposed regulatory framework for AI.

The Act seeks to codify the high standards of the EU trustworthy AI paradigm, which requires AI to be legally, ethically and technically robust, while respecting democratic values, human rights and the rule of law. The draft regulation sets out core horizontal rules for the development, commodification and use of AI-driven products, services and systems within the territory of the EU, that apply to all industries.

Legal sandboxes fostering innovation

The EC aims to prevent the rules from stifling innovation and hindering the creation of a flourishing AI ecosystem in Europe. This is ensured by introducing various flexibilities, including the application of legal sandboxes that afford breathing room to AI developers.

Sophisticated ‘product safety regime’

The EU AI Act introduces a sophisticated ‘product safety framework’ constructed around a set of 4 risk categories. It imposes requirements for market entrance and certification of High-Risk AI Systems through a mandatory CE-marking procedure. To ensure equitable outcomes, this pre-market conformity regime also applies to machine learning training, testing and validation datasets.

Pyramid of criticality

The AI Act draft combines a risk-based approach based on the pyramid of criticality, with a modern, layered enforcement mechanism. This means, among other things, that a lighter legal regime applies to AI applications with a negligible risk, and that applications with an unacceptable risk are banned. Stricter regulations apply as risk increases.

Enforcement at both Union and Member State level

The draft regulation provides for the installation of a new enforcement body at Union level: the European Artificial Intelligence Board (EAIB). At Member State level, the EAIB will be flanked by national supervisors, similar to the GDPR’s oversight mechanism. Fines for violation of the rules can be up to 6% of global turnover, or 30 million euros for private entities.

CE-marking for High-Risk AI Systems

In line with my recommendations, Article 49 of the Act requires high-risk AI and data-driven systems, products and services to comply with EU benchmarks, including safety and compliance assessments. This is crucial because it requires AI infused products and services to meet the high technical, legal and ethical standards that reflect the core values of trustworthy AI. Only then will they receive a CE marking that allows them to enter the European markets. This pre-market conformity mechanism works in the same manner as the existing CE marking: as safety certification for products traded in the European Economic Area (EEA).

Trustworthy AI by Design: ex ante and life-cycle auditing

Responsible, trustworthy AI by design requires awareness from all parties involved, from the first line of code. Indispensable tools to facilitate this awareness process are AI impact and conformity assessments, best practices, technology roadmaps and codes of conduct. These tools are executed by inclusive, multidisciplinary teams, that use them to monitor, validate and benchmark AI systems. It will all come down to ex ante and life-cycle auditing.

The new European rules will forever change the way AI is formed. Pursuing trustworthy AI by design seems like a sensible strategy, wherever you are in the world.

Meer lezen
Cyber Week 2021 Tel Aviv University Israel

AIRecht Director Mauritz Kop will speak at Cyber Week 2021 Tel Aviv University Israel, and participate in the Panel 'Debating Collective Cyber Defense for Democracies'. He will present his Stanford essay ‘Democratic Countries Should Form a Strategic Tech Alliance’ on July 22nd at 20:00 Israel time, see: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3814409

Cyber Week 2021 hosts a range of distinguished speakers from across the globe, including the Prime Minister of Israel Naftali Bennett, see: https://cw2021.b2b-wizard.com/expo/speakers

Debating Collective Cyber Defense for Democracies

Line-up and speakers of the ‘Debating Collective Cyber Defense for Democracies’ panel (notice the strong Dutch@Stanford representation):

Keynote: Ambassador Heli Tiirmaa-Klaar, Ambassador-at-Large for Cyber Diplomacy at the Estonian Ministry of Foreign Affairs

Lectures by:

Prof. Chris Demchak, Strategic and Operational Research Department, U.S. Naval War College

Dr. Lior Tabansky, Ph.D., (Moderator), Head of Research Development, Blavatnik Interdisciplinary Cyber Research Center, Tel Aviv University

Mauritz Kop, Stanford Law School TTLF Fellow, Founder of MusicaJuridica, and Strategic Intellectual Property Lawyer at AIRecht

Marietje Schaake, International Policy Director at the Cyber Policy Center; International Policy Fellow at the Institute for Human-Centered Artificial Intelligence, Stanford University

See the complete agenda at: https://cw2021.b2b-wizard.com/expo/agenda

Democratic Countries Should Form a Strategic Tech Alliance

Kop’s essay titled ‘Democratic Countries Should Form a Strategic Tech Alliance’ concludes that to prevent authoritarianism from gaining ground, democratic governments should do four things: (1) inaugurate a Strategic Tech Alliance, (2) set worldwide core rules, interoperability & conformity standards for key 4IR technologies such as AI, quantum, 6G and Virtual Reality (VR), (3) win the race for 4IR technology supremacy, and (4) actively embed our common democratic norms, principles and values into the architecture and infrastructure of our technology.

REGISTER for the conference following the link: https://cw2021.b2b-wizard.com/expo/home

Meer lezen