EU: To protect fundamental rights, a ban on biometric mass surveillance is needed

Topic
Country/Region
EU

Along with 51 other organisations, Statewatch has signed an open letter calling on the European Commission to introduce measures to ban biometric mass surveillance as part of the upcoming legislative proposal on artificial intelligence (AI).

Support our work: become a Friend of Statewatch from as little as £1/€1 per month.

The letter says that "the upcoming legislative proposal on AI must take the necessary step of prohibiting applications of AI that irremediably violate fundamental rights, such as remote biometric identification technologies that enable inherently undemocratic mass surveillance."

The letter follows a previous call from civil society groups and follows the launch of a European Citizens' Initiative calling for a ban, as well as a letter from 116 MEPs demanding action.

The letter was coordinated by EDRi. See: European Commission must ban biometric mass surveillance practices, say 51 civil society groups

The full-text of the letter is below, or it is available here as a PDF.


Thursday, 01 April 2021

Re: Seeking your support for a specific ban on biometric mass surveillance practices on fundamental rights grounds

Dear Commissioner Reynders,

cc: Executive Vice-President Vestager

cc: Vice-President Jourová

cc: Commissioner Dalli

cc: Acting Director-General Saastamoinen, DG-JUST

cc: Reynders’ cabinet

We, the signatories of this letter, represent a coalition of 52 digital rights, human rights and social justice organisations working for people’s fundamental rights across Europe. We are writing to ask you to support our call for enhanced fundamental rights protections in the upcoming artificial intelligence (AI) law, in particular relating to facial recognition and other forms of biometric mass surveillance.

The European Commission has set itself the important task of carving out a European way forward with AI that puts trust, excellence, and the protection of fundamental rights at its core. To achieve that goal, the upcoming legislative proposal on AI must take the necessary step of prohibiting applications of AI that irremediably violate fundamental rights, such as remote biometric identification technologies that enable inherently undemocratic mass surveillance.

62 civil society organisations have already issued a call for regulatory limits on applications of artificial intelligence that unduly restrict human rights, such as uses of biometric technologies that enable mass surveillance, and that call was further strengthened by a letter from 116 MEPs. Research has shown that biometric mass surveillance practices unduly infringe on the rights enshrined in the Charter of Fundamental Rights of the European Union (CFEU) – in particular, rights to privacy, data protection, equality, non-discrimination, free expression and association, due process and good administration, the presumption of innocence, social rights, workers’ rights, dignity, as well as the fundamental principles of justice, democracy and the rule of law. [1]

We would thus like to call your attention to three important points:

  1. Civil society is calling for a specific ban on inherently unnecessary and disproportionate biometric mass surveillance practices because the existing general prohibitions in the EU data protection framework are not proving sufficient:
  • The processing of biometric data is prohibited under the General Data Protection Regulation (GDPR). Yet exemptions, such as on the basis of consent (Article 9(2)(a)), are being invoked by public and private actors deploying biometric systems for mass surveillance purposes. Because of these misuses of consent as a legal basis and the subsequent lack of enforcement, we strongly believe that the EU needs a specific legal ban to build on the existing general prohibition in the GDPR; [2]
  • The processing of biometric data for law enforcement purposes is restricted to ‘only where strictly necessary’ and proportionate and on the basis of Union or Member State law (Data Protection Law Enforcement Directive (LED) Article 10 and 10(a)). However, this exemption - coupled with the inadequate application of the necessity and proportionality test - has led to entities in Member States deploying some of the most harmful uses of biometric technologies for mass surveillance purposes – those by law enforcement – despite such uses contradicting the CFEU; [3]
  • This means that in reality, biometric mass surveillance is rife across the EU, and civil society and individuals are bearing the burden of trying to stop harmful and discriminatory uses of biometric data for mass surveillance purposes through a combination of investigations, campaigning, litigation and complaints to data protection authorities (DPAs). Instead, we are asking for a specific EU legal instrument to ensure that biometric mass surveillance uses are never deployed in the first place;
  • We therefore call for a ban on the indiscriminate or arbitrarily-targeted use of biometric applications in public or publicly-accessible spaces (i.e. biometric mass surveillance) without exceptions, due to the fact that the many risks and harms involved make such uses inherently unnecessary and disproportionate for the aim sought. Where uses do not inherently lead to the undue infringement of fundamental rights, but they still limit fundamental rights, they must be strictly and demonstrably necessary and proportionate.
  1. Civil society is calling for red lines specifically on the dangerous uses and applications of artificial intelligence:
  • We welcome that the Commission has chosen to focus on how AI technologies are used. For example, the use of indiscriminate or arbitrarily-targeted facial recognition in public spaces is an application of AI that our research has shown unduly restricts fundamental rights;
  • Ensuring that the EU becomes a leader in a trustworthy approach to AI development and deployment (i.e. the use of AI that is in line with the protection of fundamental rights) will mean making decisions about which applications have a place in a society committed to fundamental rights, and which do not. While other countries may pay no heed to protecting fundamental rights in their pursuit of innovation at all costs, the EU can and must lead by example by ensuring that the innovative AI developed and deployed within its borders is always developed and deployed in accordance with fundamental rights;
  • Over 43,500 EU citizens have already officially added their voices to the civil society call for a ban on biometric mass surveillance practices through the new European Citizens’ Initiative run by the Reclaim Your Face campaign. We think it is vital that their views, expressed through this powerful democratic initiative, are taken seriously in the upcoming regulatory proposal. This is especially important given the ambition set out by the Commission in the White Paper on AI for a broad public debate.
  1. AI innovation in Europe can thrive globally by respecting fundamental rights:
  • With the GDPR, the EU proved that our advantage in a digitalised world can be in ensuring that innovation always respects people’s fundamental rights, and that the EU can set standards that protect people’s rights while keeping markets competitive;
  • We have learned that both within and beyond the EU, the unfettered development and deployment of biometric technologies has severe consequences for the human rights of marginalised people and groups, who are often disproportionately subject to discriminatory deployments of such technologies whilst also seriously under-represented in EU decision-making;
  • China’s biometric mass surveillance of the Uighur population and the US’s disproportionate police surveillance of Black communities with inherently discriminatory facial recognition are not models to strive for – and US cities are increasingly taking moves to ban such uses as a result. For the EU to base its regulatory model on competition with these practices would compromise the very principles and values on which the EU is built.

We appreciate that the European Commission has so far agreed with the principle of a ban on biometric mass surveillance practices. To further protect fundamental rights in Europe, the signatories of this letter therefore call for:

  1. The legislative proposal on artificial intelligence to include an explicit ban on the indiscriminate or arbitrarily-targeted use of biometrics in public or publicly-accessible spaces which can lead to mass surveillance, on fundamental rights grounds;
  2. Legal restrictions or legislative red-lines on all uses which contravene fundamental rights;
  3. The explicit inclusion of marginalised and affected communities in the development of EU AI legislation and policy.

For a truly human-centric legislation on AI, we reiterate that there must be some uses that the EU does not allow in a democratic society. We look forward to working with you to make a ban on harmful and rights-violating biometric mass surveillance in the EU a true reality.

Sincerely,

Organisations:

  • Access Now, International
  • AlgorithmWatch, Germany
  • All Out, International
  • App Drivers and Couriers Union (ADCU), UK
  • ARTICLE 19, International
  • Associazione Luca Coscioni, Italy
  • Big Brother Watch, UK
  • Bits of Freedom (BoF), the Netherlands
  • Certi Diritti, Italy
  • Civil Liberties Union for Europe (Liberties), International
  • Chaos Computer Club (CCC), Germany
  • Chaos Computer Club Lëtzebuerg, Luxembourg
  • Defesa dos Direitos Digitais (D3), Portugal
  • Dataskydd.net, Sweden
  • Datenschutzraum e.V., Germany
  • defend digital me, UK
  • Digital Rights Ireland, Ireland
  • Digitalcourage, Germany
  • Digitale Geselleschaft, Germany
  • Digitale Gesellschaft CH, Switzerland
  • Državljan D, Slovenia
  • Electronic Frontier Finland (Effi), Finland
  • Elektroniks Forpost Norge (EFN), Norway
  • Electronic Privacy Information Center (EPIC), International
  • epicenter.works, Austria
  • EUMANS, International
  • European Center for Not-for-profit Law (ECNL), International
  • European Digital Rights (EDRi), International
  • European Digital Society (EDS), International
  • European Evangelical Alliance (EEA), International
  • European Youth Forum, International
  • FITUG e.V., Germany
  • Frënn vun der Ënn, Luxembourg
  • Hermes Center, Italy
  • Homo Digitalis, Greece
  • Human Rights League Slovakia, Slovakia
  • International Committee on the Rights of Sex Workers in Europe (ICRSE), International
  • IT Political Association of Denmark (IT-Pol), Denmark
  • Iuridicum Remedium (IuRe), Czechia
  • La Ligue des droits de l’Homme (LDH), France
  • La Quadrature du Net (LQDN), France
  • LOAD e.V., Germany
  • Panoptykon Foundation, Poland
  • PICUM, International
  • Pravno-informacijski center nevladnih organizacij (PIC), Slovenia
  • Privacy International, International
  • Privacy Network, Italy
  • Science for Democracy, International
  • Sex Workers' Rights Advocacy Network (SWAN), International
  • Statewatch, UK
  • StraLi for Strategic Litigation, Italy
  • Vrijschift.org, the Netherlands

Individuals:

  • Douwe Korff, Emeritus Professor of International Law
  • Barbora Messova, lawyer and director of Human Rights League Slovakia
  • Virginia Fiume, coordinator EUMANS

[1] 'Ban Biometric Mass Surveillance’, EDRi, May 2020: https://edri.org/wp-content/uploads/2020/05/Paper-Ban-Biometric-Mass-Surveillance.pdf
[2] To further evidence the problem with the exemptions in the current framework, we share the example of Italy, where civil society successfully challenged the legality of a use of biometric mass surveillance in Como, leading to its decommission, only for the same use to be subsequently rolled out in Turin.
[3] One example of this is in Denmark, where the state has introduced national legislation which has led to the “legal” biometric mass surveillance of people attending football matches.

Image: Ars Electronica, CC BY-NC-ND 2.0

Our work is only possible with your support.
Become a Friend of Statewatch from as little as £1/€1 per month.

 

Spotted an error? If you've spotted a problem with this page, just click once to let us know.

Report error