97 Amendments of Sophia IN 'T VELD related to 2021/0106(COD)
Amendment 443 #
Proposal for a regulation
Recital 17 a (new)
Recital 17 a (new)
(17 a) AI systems used by law enforcement authorities or on their behalf to predict the probability of a natural person to offend or to reoffend, based on profiling and individual risk-assessment hold a particular risk of discrimination against certain persons or groups of persons, as they violate human dignity as well as the key legal principle of presumption of innocence. Such AI systems should therefore be prohibited.
Amendment 450 #
Proposal for a regulation
Recital 18
Recital 18
(18) The use of AI systems for ‘real- time’ remote biometric identification of natural persons in publicly accessible spaces for the purpose of law enforcement is considered particularly intrusive in the rights and freedoms of the concerned persons, to the extent that it may affect the private life of a large part of the population, evoke a feeling of constant surveillance and indirectly dissuade the exercise of the freedom of assembly and other fundamental rights. In addition, the immediacy of the impact and the limited opportunities for further checks or corrections in relation to the use of such systems operating in ‘real-time’ carry heightened risks for the rights and freedoms of the persons that are concerned by law enforcement activities. The use of those systems in publicly accessible places should therefore be prohibited.
Amendment 451 #
Proposal for a regulation
Recital 18
Recital 18
(18) The use of AI systems for ‘real- time’ remote biometric identification of natural persons in publicly accessible spaces for the purpose of law enforcement is considered particularly intrusive in the rights and freedoms of the concerned persons, to the extent that it may affect the private life of a large part of the population, evoke a feeling of constant surveillance and indirectly dissuade the exercise of the freedom of assembly and other fundamental rights. In addition, the immediacy of the impact and the limited opportunities for further checks or corrections in relation to the use of such systems operating in ‘real-time’ carry heightened risks for the rights and freedoms of the persons that are concerned by law enforcement activities. The use of those systems in publicly accessible places should therefore be prohibited.
Amendment 464 #
Proposal for a regulation
Recital 19
Recital 19
Amendment 467 #
Proposal for a regulation
Recital 19
Recital 19
Amendment 473 #
Proposal for a regulation
Recital 20
Recital 20
Amendment 477 #
Proposal for a regulation
Recital 20
Recital 20
Amendment 483 #
Proposal for a regulation
Recital 21
Recital 21
Amendment 486 #
Proposal for a regulation
Recital 21
Recital 21
Amendment 490 #
Proposal for a regulation
Recital 22
Recital 22
Amendment 494 #
Proposal for a regulation
Recital 22
Recital 22
Amendment 497 #
Proposal for a regulation
Recital 23
Recital 23
Amendment 499 #
Proposal for a regulation
Recital 23
Recital 23
Amendment 515 #
Proposal for a regulation
Recital 24 a (new)
Recital 24 a (new)
Amendment 663 #
Proposal for a regulation
Recital 58
Recital 58
(58) Given the nature of AI systems and the risks to safety and fundamental rights possibly associated with their use, including as regard the need to ensure proper monitoring of the performance of an AI system in a real-life setting, it is appropriate to set specific responsibilities for users. Users should in particular use high-risk AI systems in accordance with the instructions of use and certain other obligations should be provided for with regard to monitoring of the functioning of the AI systems and with regard to record- keeping, as appropriate. Given the potential impact and the need for democratic oversight and scrutiny, users of high-risk AI systems that are public authorities or Union institutions, bodies, offices and agencies should be required to conduct a fundamental rights impact assessment prior to commencing the use of a high-risk AI system should be required to register the use of any high- risk AI systems in a public database.
Amendment 698 #
Proposal for a regulation
Recital 68
Recital 68
Amendment 741 #
Proposal for a regulation
Recital 76
Recital 76
(76) In order to facilitate a smooth, effective and harmonised implementation of this Regulation a European Artificial Intelligence Board should be established as a body of the Union and should have legal personality. The Board should be responsible for a number of advisory tasks, including issuing opinions, recommendations, advice or guidance on matters related to the implementation of this Regulation, including on technical specifications or existing standards regarding the requirements established in this Regulation and providing advice to and assisting the Commission and the national competent authorities on specific questions related to artificial intelligence.
Amendment 779 #
Proposal for a regulation
Article 1 – title
Article 1 – title
1 SAim and subject matter
Amendment 782 #
Proposal for a regulation
Article 1 – paragraph -1 (new)
Article 1 – paragraph -1 (new)
-1 The purpose of this Regulation is to ensure a high level of protection of health, safety, and fundamental rights from harmful effects of artificial intelligence systems ("AI systems") in the Union, while enhancing innovation.
Amendment 807 #
Proposal for a regulation
Article 1 – paragraph 1 a (new)
Article 1 – paragraph 1 a (new)
This Regulation is based on the principle that it is for developers, importers, distributors and users to ensure that they develop, place on the market or use AI systems that do not adversely affect health, safety, or fundamental rights. Its provisions are underpinned by the precautionary principle.
Amendment 828 #
Proposal for a regulation
Article 2 – paragraph 1 – point c a (new)
Article 2 – paragraph 1 – point c a (new)
(c a) public authorities in a third country or to international organisations where those authorities or organisations use AI systems in the framework of international agreements for law enforcement and judicial cooperation with the Union or with one or more Member States.
Amendment 868 #
Proposal for a regulation
Article 2 – paragraph 3
Article 2 – paragraph 3
3. This Regulation shall not apply to AI systems developed or used exclusively for military purposes, unless the AI system is subsequently used for non-military purposes.
Amendment 881 #
Proposal for a regulation
Article 2 – paragraph 4
Article 2 – paragraph 4
Amendment 926 #
Proposal for a regulation
Article 3 – paragraph 1 – point 1 b (new)
Article 3 – paragraph 1 – point 1 b (new)
(1 b) 'general purpose AI system’ means an AI system that is able to perform generally applicable functions for multiple potential purposes, such as image or speech recognition, audio or video generation, pattern detection, question answering, and translation, is largely customizable and often open source software;
Amendment 969 #
Proposal for a regulation
Article 3 – paragraph 1 – point 12 – point i (new)
Article 3 – paragraph 1 – point 12 – point i (new)
i) 'Reasonably foreseeable use' means the use of an AI system in a way that is or should be reasonably foreseeable and that addresses the risks to health, safety and fundamental rights that it can cause.
Amendment 1033 #
Proposal for a regulation
Article 3 – paragraph 1 – point 34
Article 3 – paragraph 1 – point 34
(34) ‘emotion recognition system’ means an AI system for the purposcapable of identifying, categorizing or inferring emotions, thoughts, states of mind (such as 'deception', 'trustworthiness', or 'trustfulness') or intentions of natural persons on the basis of their biometric data;
Amendment 1046 #
Proposal for a regulation
Article 3 – paragraph 1 – point 35
Article 3 – paragraph 1 – point 35
(35) ‘biometric categorisation system’ means an AI system for that uses biometric data, or other purposhysical, physiological or behavioral data, capable of assigning natural persons to specific categories, such as sex, age, hair colour, eye colour, tattoos, ethnic origin or sexual or political orientation, on the basis of their biometric data or inferring their characteristics and attributes;
Amendment 1047 #
Proposal for a regulation
Article 3 – paragraph 1 – point 35 a (new)
Article 3 – paragraph 1 – point 35 a (new)
(35 a) ‘remote biometric categorisation system’ means a biometric categorisation system capable of categorising natural persons at a distance;
Amendment 1061 #
Proposal for a regulation
Article 3 – paragraph 1 – point 36
Article 3 – paragraph 1 – point 36
(36) ‘remote biometric identification system’ means an AI system for the purpose of identifycapable of categorizing natural persons at a distance through the comparison of a person’s biometric data with the biometric data contained in a reference database, and without prior knowledge of the user of the AI system whether the person will be present and can be identified or other physical, physiological or behavioral data, with this data contained in a reference database;
Amendment 1128 #
Proposal for a regulation
Article 3 – paragraph 1 a (new)
Article 3 – paragraph 1 a (new)
'Social scoring' means the evaluation or categorization of an individual natural person, or a group, based on their behaviour or (personality) characteristics, where one or more of the following conditions apply: (1) the information is not reasonably relevant, necessary for, or proportionate to the evaluation or categorization; (2) the information is generated or collected in another domain than that of the evaluation or categorization; (3) the information contains or reveals special categories.
Amendment 1164 #
Proposal for a regulation
Article 5 – paragraph 1 – point a
Article 5 – paragraph 1 – point a
(a) the placing on the market, putting into service or use of an AI system that deploys subliminal techniques beyond a person’s consciousness in order to materially distort a person’s behaviour in a manner that causes or is likely to cause that person or another person physical or psychological harm;
Amendment 1172 #
Proposal for a regulation
Article 5 – paragraph 1 – point a a (new)
Article 5 – paragraph 1 – point a a (new)
Amendment 1180 #
Proposal for a regulation
Article 5 – paragraph 1 – point b
Article 5 – paragraph 1 – point b
(b) the placing on the market, putting into service or use of an AI system that exploits, intentionally or not, any of the vulnerabilities of a specificrson or group of persons due to their age, physical or mental disability, in order to materially distort the behaviour of a person pertaining to that group in a manner that causesbased on any sensitive or protected characteristic, including but not limited to age, gender and gender identity, racial or ethnic origin, health status, sexual orientation, sex characteristics, social or economic status, worker status, migration status, or dis likely to cause that person or another person physical or psychological harmability in accordance with Article 21 of the Charter of Fundamental Rights;
Amendment 1185 #
Proposal for a regulation
Article 5 – paragraph 1 – point b
Article 5 – paragraph 1 – point b
(b) the placing on the market, putting into service or use of an AI system that exploits any of the vulnerabilities of a specific group of persons due to their age, physical or mental disability, in order to materially distort the behaviour of a person pertaining to that group in a manner that causes or is likely to cause that person or another person physical or psychological harm;
Amendment 1194 #
Proposal for a regulation
Article 5 – paragraph 1 – point c – introductory part
Article 5 – paragraph 1 – point c – introductory part
(c) the placing on the market, putting into service or use of AI systems by public authorities or on their behalf for the evaluation or classification of the trustworthiness of natural persons over a certain period of time based on theirfor calculation or establishment of a 'social score' resulting from the evaluation or classification of natural persons based on their physical attributes, social behaviour or known or predicted personal or personality characteristics, with the social score leading to either or both of the following:.
Amendment 1204 #
Proposal for a regulation
Article 5 – paragraph 1 – point c – point i
Article 5 – paragraph 1 – point c – point i
Amendment 1213 #
Proposal for a regulation
Article 5 – paragraph 1 – point c – point ii
Article 5 – paragraph 1 – point c – point ii
Amendment 1223 #
Proposal for a regulation
Article 5 – paragraph 1 – point c a (new)
Article 5 – paragraph 1 – point c a (new)
(c a) the placing on the market, putting into service or use of an AI system for making individual risk assessments of natural persons in order to assess the risk of a natural person for offending or reoffending or for predicting the occurrence or reoccurrence of an actual or potential criminal offence based on profiling of a natural person or on assessing personality traits and characteristics or past criminal behaviour of natural persons or groups of natural persons;
Amendment 1235 #
Proposal for a regulation
Article 5 – paragraph 1 – point d – introductory part
Article 5 – paragraph 1 – point d – introductory part
(d) tThe use of ‘real-time’ remote biometric identification systems in publicly accessible spaces for the purpose of law enforcement, unless and in as far as such use is strictly necessary for one of the following objectives:placing on the market, putting into service or use of of AI for an automated recognition of human features in publicly accessible spaces - such as of faces but also of gait, fingerprints, DNA, voice, keystrokes and other biometric or behavioral signals - for any purpose, including law enforcement.
Amendment 1253 #
Proposal for a regulation
Article 5 – paragraph 1 – point d – point i
Article 5 – paragraph 1 – point d – point i
Amendment 1254 #
Proposal for a regulation
Article 5 – paragraph 1 – point d – point i
Article 5 – paragraph 1 – point d – point i
Amendment 1260 #
Proposal for a regulation
Article 5 – paragraph 1 – point d – point ii
Article 5 – paragraph 1 – point d – point ii
Amendment 1263 #
Proposal for a regulation
Article 5 – paragraph 1 – point d – point ii
Article 5 – paragraph 1 – point d – point ii
Amendment 1273 #
Proposal for a regulation
Article 5 – paragraph 1 – point d – point iii
Article 5 – paragraph 1 – point d – point iii
Amendment 1274 #
Proposal for a regulation
Article 5 – paragraph 1 – point d – point iii
Article 5 – paragraph 1 – point d – point iii
Amendment 1286 #
Proposal for a regulation
Article 5 – paragraph 1 – point d a (new)
Article 5 – paragraph 1 – point d a (new)
(d a) the use of an AI system for the general monitoring, detection and interpretation of private content in interpersonal communication services, including all measures that would undermine end-to-end encryption..
Amendment 1287 #
Proposal for a regulation
Article 5 – paragraph 1 – point d a (new)
Article 5 – paragraph 1 – point d a (new)
(d a) The creation or expansion of facial recognition or other biometric databases through the untargeted scraping of biometric data from social media profiles or CCTV footage or equivalent methods;
Amendment 1296 #
Proposal for a regulation
Article 5 – paragraph 1 – point d b (new)
Article 5 – paragraph 1 – point d b (new)
(d b) The use of private facial recognition or other private biometric databases for the purpose of law enforcement
Amendment 1299 #
Proposal for a regulation
Article 5 – paragraph 1 – point d c (new)
Article 5 – paragraph 1 – point d c (new)
(d c) The placing on the market, putting into service or use of 'emotion recognition systems', unless for health purposes, which would be considered high risk.Emotion recognition systems for health purposes shall be limited to their intended purpose, subject to all applicable data protection conditions and limits, and: (i) undergo strict testing to ensure scientific and clinical validity; (ii) contain clear advice to anyone that may procure or use them about the limitations of such technologies and their potential risks, including of flawed or potentially harmful outcomes; (iii) be developed with the active participation and input of the groups they are intended to benefit, as well as those with expertise in the range of fundamental rights that could be deliberately or inadvertently impacted; (iv) be developed and deployed in a manner that respects the rights of all persons likely to be affected by them; (v) be subject to an opinion of the Health Security Committee and the Fundamental Rights Agency.
Amendment 1311 #
Proposal for a regulation
Article 5 – paragraph 1 – point d d (new)
Article 5 – paragraph 1 – point d d (new)
(d d) AI systems intended to be used by law enforcement authorities as polygraphs and similar tools or to detect the emotional state of a natural person;
Amendment 1313 #
Proposal for a regulation
Article 5 – paragraph 1 – point d e (new)
Article 5 – paragraph 1 – point d e (new)
(d e) The placing on the market, putting into service or use of AI systems including, but not limited to polygraphs and similar tools to detect deception, trustworthiness or related characteristics, by or on behalf of competent authorities in migration, asylum or border control management, to profile an individual or assess a risk, including a security risk, a risk of irregular immigration, or a health risk, posed by a natural person who intends to enter or has entered the territory of a Member state, on the basis of personal or sensitive data, known or predicted, except for the sole purpose of identifying specific care and support needs;
Amendment 1317 #
Proposal for a regulation
Article 5 – paragraph 1 – point d f (new)
Article 5 – paragraph 1 – point d f (new)
(d f) the placing on the market, putting into service or use of AI systems that use psysiological, behavioural or biometric data to infer attributes or characteristics of persons or groups which are not solely determined by such data or are not externally observable or whose complexity is not possible to fully capture in data, including but not limited to gender, race, colour, ethnic or social origin, as well as political or sexual orientation, or other grounds for discrimination prohibited under Article 21 of the Charter.
Amendment 1325 #
Proposal for a regulation
Article 5 – paragraph 1 – point d g (new)
Article 5 – paragraph 1 – point d g (new)
(d g) AI systems intended to be used by public authorities or on behalf of public authorities to evaluate the eligibility of natural persons for public assistance benefits and services, as well as to grant, reduce, revoke, or reclaim such benefits and services;
Amendment 1348 #
Proposal for a regulation
Article 5 – paragraph 2
Article 5 – paragraph 2
Amendment 1354 #
Proposal for a regulation
Article 5 – paragraph 2
Article 5 – paragraph 2
Amendment 1356 #
Proposal for a regulation
Article 5 – paragraph 2 – point a
Article 5 – paragraph 2 – point a
Amendment 1357 #
Proposal for a regulation
Article 5 – paragraph 2 – point a
Article 5 – paragraph 2 – point a
Amendment 1358 #
Proposal for a regulation
Article 5 – paragraph 2 – point b
Article 5 – paragraph 2 – point b
Amendment 1359 #
Proposal for a regulation
Article 5 – paragraph 2 – point b
Article 5 – paragraph 2 – point b
Amendment 1361 #
Proposal for a regulation
Article 5 – paragraph 2 – subparagraph 1
Article 5 – paragraph 2 – subparagraph 1
Amendment 1362 #
Proposal for a regulation
Article 5 – paragraph 2 – subparagraph 1
Article 5 – paragraph 2 – subparagraph 1
Amendment 1364 #
Proposal for a regulation
Article 5 – paragraph 3
Article 5 – paragraph 3
Amendment 1367 #
Proposal for a regulation
Article 5 – paragraph 3
Article 5 – paragraph 3
Amendment 1375 #
Proposal for a regulation
Article 5 – paragraph 3 – subparagraph 1
Article 5 – paragraph 3 – subparagraph 1
Amendment 1376 #
Proposal for a regulation
Article 5 – paragraph 3 – subparagraph 1
Article 5 – paragraph 3 – subparagraph 1
Amendment 1381 #
Proposal for a regulation
Article 5 – paragraph 4
Article 5 – paragraph 4
Amendment 1387 #
Proposal for a regulation
Article 5 – paragraph 4
Article 5 – paragraph 4
Amendment 1420 #
Proposal for a regulation
Article 6 – paragraph 1 – point a
Article 6 – paragraph 1 – point a
(a) the AI system is intended to be used or reasonably foreseeable used as a safety component of a product, or is itself a product, covered by the Union harmonisation legislation listed in Annex II;
Amendment 1456 #
Proposal for a regulation
Article 6 a (new)
Article 6 a (new)
Amendment 1458 #
Proposal for a regulation
Article 6 a (new)
Article 6 a (new)
Article 6 a Risk assessment The European Artificial Intelligence Board shall develop guidance for the risk assessment.
Amendment 1491 #
Proposal for a regulation
Article 7 – paragraph 2 – introductory part
Article 7 – paragraph 2 – introductory part
2. When assessing for the purposes of paragraph 1 whether an AI system poses a risk of harm to the health and safety or a risk of adverse impact on fundamental rights that is equivalent to or greater than the risk of harm posed by the high-risk AI systems already referred to in Annex III, the Commission shall take into account, including but not limited to, the following criteria:
Amendment 1496 #
Proposal for a regulation
Article 7 – paragraph 2 – point a
Article 7 – paragraph 2 – point a
(a) the intended purpose or the reasonably foreseeable use of the AI system;
Amendment 1730 #
Proposal for a regulation
Article 10 – paragraph 4
Article 10 – paragraph 4
4. Training, validation and testing data sets shall take into account, to the extent required by the intended purpose, the characteristics or elements that are particular to the specific geographical, behavioural or functional setting within which the high-risk AI system is intended to be used.
Amendment 1884 #
Proposal for a regulation
Article 16 – paragraph 1 – point a a (new)
Article 16 – paragraph 1 – point a a (new)
(a a) ensure that, in the case of a general purpose AI system, the reasonably foreseeable uses of this system are assessed.
Amendment 2066 #
Proposal for a regulation
Article 29 – paragraph 6
Article 29 – paragraph 6
6. Users of high-risk AI systems shall use the information provided under Article 13 to comply with their obligation to carry out a data protection impact assessment under Article 35 of Regulation (EU) 2016/679 or Article 27 of Directive (EU) 2016/680, where applicable. The data protection impact assessment shall be published.
Amendment 2067 #
Proposal for a regulation
Article 29 – paragraph 6 a (new)
Article 29 – paragraph 6 a (new)
6 a. Users of high-risk AI systems shall carry out a human rights impact assessment for the different uses of the system, containing specific information on the context of use of that system, including, the intended purpose or reasonable foreseeably use, geographic and temporal scope, assessment of the legality and fundamental rights impacts of the system, any specific risk of harm likely to impact marginalised persons or those at risk of discrimination, any other negative impact on the public interest;and clear steps as to how the harms identified will be mitigated, and how effective this mitigation is likely to be. The human rights impact assessment shall be published, and be registered by the user in the database referred to under Article 60.
Amendment 2175 #
Proposal for a regulation
Article 43 – paragraph 1 – subparagraph 2
Article 43 – paragraph 1 – subparagraph 2
Amendment 2189 #
Proposal for a regulation
Article 43 – paragraph 4 – introductory part
Article 43 – paragraph 4 – introductory part
4. High-risk AI systems shall undergo a new conformity assessment procedure whenever they are substantially modified, regardless of whether the modified system is intended to be further distributed or continues to be used by the current user, or whenever a change occurs which may affect the compliance with this Regulation.
Amendment 2344 #
Proposal for a regulation
Article 54
Article 54
Amendment 2468 #
Proposal for a regulation
Article 57 – paragraph 4 a (new)
Article 57 – paragraph 4 a (new)
4 a. The Board shall take into consideration advice provided by the EDPB, particularly on new or evolving risks of high-risk AI systems processing personal data.
Amendment 2696 #
Proposal for a regulation
Article 64 – paragraph 3
Article 64 – paragraph 3
3. National public authorities or bodies which supervise or enforce the respect of obligations under Union law protecting fundamental rights in relation to the use of high-risk AI systems referred to in Annex III shall have the power to request and access any documentation created or maintained under this Regulation, including data protection impact assessments and human rights impact assessments carried out by the users of such systems, when access to that documentation is necessary for the fulfilment of the competences under their mandate within the limits of their jurisdiction. The relevant public authority or body shall inform the market surveillance authority of the Member State concerned of any such request.
Amendment 2945 #
Proposal for a regulation
Article 83 – paragraph 1 – introductory part
Article 83 – paragraph 1 – introductory part
1. This Regulation shall not apply to the AI systems which are components of the large-scale IT systems established by the legal acts listed in Annex IX that have been placed on the market or put into service before [12 mon, withs after the date of application of this Regulation referred to in Article 85(2)], unless the replacement or amendment of those legal acts leads to a significant change in the design or intended purpose of the AI system or AI systems concerned transitional period of two years after the entry into force of this Regulation.
Amendment 2950 #
Proposal for a regulation
Article 83 – paragraph 1 – subparagraph 1
Article 83 – paragraph 1 – subparagraph 1
Amendment 2961 #
Proposal for a regulation
Article 83 – paragraph 2
Article 83 – paragraph 2
2. This Regulation shall apply to the high-risk AI systems, other than the ones referred to in paragraph 1, that have been placed on the market or put into service before [date of application of this Regulation referred to in Article 85(2)], only if, from that date, those systems are subject to significant changes in their design or intended purposewith a transitional period of two years after the application of this Regulation.
Amendment 2972 #
Proposal for a regulation
Article 84 – paragraph 1 a (new)
Article 84 – paragraph 1 a (new)
Amendment 2974 #
Proposal for a regulation
Article 84 – paragraph 3 – point a
Article 84 – paragraph 3 – point a
(a) the status of the financial, technical and human resources of the national competent authorities in order to effectively perform the tasks assigned to them under this Regulation;
Amendment 2982 #
Proposal for a regulation
Article 84 – paragraph 5
Article 84 – paragraph 5
5. For the purpose of paragraphs 1 to 4 the Board, the Member States and national competent authorities shall provide the Commission with information on its request without undue delay.
Amendment 2988 #
Proposal for a regulation
Article 84 – paragraph 6
Article 84 – paragraph 6
6. In carrying out the evaluations and reviews referred to in paragraphs 1 to 4 the Commission shall take into account the positions and findings of the Board, of the European Parliament, of the Council, and of other relevant bodies or sources, which shall be attached to the report.
Amendment 2994 #
Proposal for a regulation
Article 84 – paragraph 7
Article 84 – paragraph 7
7. The Commission shall, if necessary, submit appropriate proposals to amend this Regulation, in particular taking into account the impact of this Regulation on fundamental rights, developments in technology and in the light of the state of progress in the information society.
Amendment 3121 #
Proposal for a regulation
Annex III – paragraph 1 – point 5 – point a
Annex III – paragraph 1 – point 5 – point a
Amendment 3158 #
Proposal for a regulation
Annex III – paragraph 1 – point 6 – point b
Annex III – paragraph 1 – point 6 – point b
Amendment 3177 #
Proposal for a regulation
Annex III – paragraph 1 – point 6 – point e
Annex III – paragraph 1 – point 6 – point e
Amendment 3190 #
Proposal for a regulation
Annex III – paragraph 1 – point 7 – point a
Annex III – paragraph 1 – point 7 – point a
Amendment 3199 #
Proposal for a regulation
Annex III – paragraph 1 – point 7 – point b
Annex III – paragraph 1 – point 7 – point b
Amendment 3218 #
Proposal for a regulation
Annex III – paragraph 1 – point 7 – point d a (new)
Annex III – paragraph 1 – point 7 – point d a (new)
(d a) AI systems intended to be used by or on behalf of competent authorities in migration, asylum and border control management for the forecasting or prediction of trends related to migration, movement and border crossings;
Amendment 3234 #
Proposal for a regulation
Annex III – paragraph 1 – point 8 – point a a (new)
Annex III – paragraph 1 – point 8 – point a a (new)
(a a) AI systems intended to be used by electoral constituencies for the purpose of protecting democracy and predicting the risk of a candidate for political office, in particular the position of head of government, being homophobic, sexist, dictatorial, kleptocratic and/or having other toxic personality traits;
Amendment 3304 #
Proposal for a regulation
Annex VIII – point 6 a (new)
Annex VIII – point 6 a (new)
6 a. where the user is obliged to register an AI system under Article 29, the human rights impact assessment must also be registered and publicly available;