Why EU needs to be wary that AI will increase racial profiling
Central to predictive policing systems is the notion that risk and crime can be objectively and accurately forecasted. Not only is this presumption flawed, it demonstrates a growing commitment to the idea that data can and should be used to quantify, track and predict human behaviour. The increased use of such systems is part of a growing ideology that social issues can be solved by allocating more power, resources - and now technologies - to police.
Whether it be ‘predictive’ algorithmic systems used in the Netherlands to profile Eastern Europeans and Roma for ‘pick-pocketing’, or secretive lists of suspected criminals such as the UK’s Gang Matrix, there is a growing reliance on criminal risk-scoring technologies by police forces across Europe.
Police say these systems predict where and by whom crime happens.
These predictions and judgments can have very real impacts on the people subject to them, increasing the frequency of interactions with police.
In some cases, these risk-scores are heavily intertwined with social welfare systems, child protection watch-lists, and the classification of certain neighbourhoods as ‘ghettos’.
Central to predictive policing systems is the notion that risk and crime can be objectively and accurately forecasted. Not only is this presumption flawed, it demonstrates a growing commitment to the idea that data can and should be used to quantify, track and predict human behaviour.
The increased use of such systems is part of a growing ideology that social issues can be solved by allocating more power, resources – and now technologies – to police.
Data might be able to indicate that something may occur in the near future, but not why. As such these technologies are about controlling behaviour that has been deemed as ‘unacceptable’ by police.
Fundamentally determined by historical practice and socio-economic biases, the data underlying law enforcement systems is inherently racialised and classed.
The data reflects patterns of policing, not crime.
Based on police presumptions
Discriminatory outcomes from predictive policing are therefore not an accident. They are the result of a process which seeks to “optimise” existing policing practice.
For example, Amnesty found that 78 percent of people on the UK’s Gangs Matrix in 2017 were black, despite only 29 percent of ‘gang-related’ crimes being committed by black people.
Using indicators such as previous stop and search information, association with other ‘gang nominals’ social media activity, and even types of music listened to, the Matrix purports to predict who will commit crimes in the future.
The entire system is based on highly discriminatory police presumptions.
These tools tend to be deployed on crimes such as burglary, theft and other ‘anti-social behaviour’. Prioritising pre-emptive police intervention on activities that are constructed to be associated with working class, migrant and racialised communities hardwires value judgments about which behaviours should be prioritised by law enforcement and which should not.
In the process this further criminalises racialised and poor communities.
In another proposed legislation, the European Commission is attempting to enhance Europol’s (the EU policing agency) capacity to make use of big data – both in its operations and in the context of ‘research and innovation’.
As argued by Statewatch, the European Council is already heavily subscribed to the idea that police should make more use of artificial intelligence.
However, the EU using data “hoovered up from the member states” for research or operations presents a major concern of reinforcing historic patterns of racist policing as well as deepening police surveillance.
The EU’s tech deterministic worldview and vested interest in AI in policing casts huge doubts on whether it will take the necessary decisions to address the harms at stake.
It raises the question – will the upcoming regulation really regulate “high risk” AI or will it enable it?
Everything will depend on the obligations towards institutions that deploy these technologies. Categorising harmful uses of AI as ‘high risk’ with relatively weak procedural requirements is likely to provide a clearer roadmap for developers and states to more easily deploy technologies like predictive policing rather than deter them.
As set out in its Digital Compass, the EU will be striving to meet the goal that “three out of four companies should use cloud computing services, big data and Artificial Intelligence” by 2030.
By design, the EU’s approach is likely to enable a market of unjust AI.
At the recent EU Anti-racism summit, commissioner for Equality, Helena Dalli argued that the EU must strive to become an ‘anti-racist’ union.
In the field of digital policy, this would mean developing a regulatory model firmly based on human rights and social justice. Issues of non-discrimination and fundamental rights would have to be at the core of the approach, rather than considered after competition and industrial policy.
A truly people-centred AI regulation would take a step back and acknowledge the inherent harms AI will perpetuate if deployed for certain purposes.
As outlined by 62 human rights organisations, the EU needs to set clear limits or ‘red lines’ on the most harmful uses of AI, including predictive policing, biometric mass surveillance and applications that exacerbate structural discrimination.
The EU needs to break from its enabling approach on AI. It must commit to encouraging only those applications that can guarantee benefiting people, not just public authorities and companies.
The article was first published by the EU Observer here.
(Contribution by: Sarah Chander, Senior Policy Advisor at European Digital Rights & Fieke Jansen, Fellow of the Mozilla Foundation)
(Image credit: Lorenzo Miola/ Fine Arts)