Cressida Dick: Facial recognition critics should face up to victims of crime

Privacy concerns over live facial recognition are ‘much smaller’ than the need to protect the public from ‘a knife through the chest’, Britain’s most senior police officer has said.

Metropolitan Police Commissioner Dame Cressida Dick told a conference in Whitehall that critics of the use of such technology would need to justify to victims of crime why police should not be allowed to use these methods.

She told delegates at the Royal United Services Institute: “I and others have been making the case for the proportionate use of tech in policing, but right now the loudest voices in the debate seem to be the critics – sometimes highly inaccurate and or highly ill informed.

Cressida Dick
Dame Cressida Dick said critics of such technology would need to justify to victims of crime why police should not be allowed to use it (PA)

“It is for critics to justify to the victims of those crimes why police should not use tech lawfully and proportionally to catch criminals who caused the victims real harm.

“It is not for me and the police to decide where the boundary lies between security and privacy, though I do think it is right for us to contribute to the debate.

“But speaking as a member of public, I will be frank. In an age of Twitter and Instagram and Facebook, concern about my image and that of my fellow law-abiding citizens passing through LFR (live facial recognition) and not being stored, feels much, much, much smaller than my and the public’s vital expectation to be kept safe from a knife through the chest.”

She said that if artificial intelligence could help identify potential terrorists, rapists or killers, most members of the public would want them to use it.

Dame Cressida said: “If, as seems likely, algorithms can assist in identifying patterns of behaviour by those under authorised surveillance, that would otherwise have been missed, patterns that indicate they are radicalising others or are likely to mount a terrorist attack; if an algorithm can help identify in our criminal systems material a potential serial rapist or killer that we could not have found by human endeavour alone; if a machine can improve our ability to disclose fairly then I think almost all citizens would want us to use it.

“The only people who benefit from us not using lawfully and proportionately tech are the criminals, the rapists, the terrorists and all those who want to harm you, your family and friends.”

Recent Metropolitan Police use of facial recognition led to the arrest of eight criminals who would not otherwise have been caught, delegates heard.

LIVE FACIAL RECOGNITION | We are using the latest technology to tackle crime and keep Londoners safe.

Live Facial Recognition will assist us locate and identify those wanted by police for violent offences and serious crimes.

👉 https://t.co/ZhhNeJPHT5pic.twitter.com/0IWaCZOzln

— Metropolitan Police (@metpoliceuk) January 24, 2020

Use of the technology has been criticised as a violation of privacy.

During a deployment by the Met at Oxford Citcus last week, Silkie Carlo from civil liberties group Big Brother Watch, said: “It’s alarming to see biometric mass surveillance being rolled out in London. Never before have citizens been subjected to identity checks without suspicion, let alone on a mass scale.”

People scanned by the cameras are checked against “watchlists” – said to contain suspects wanted by police and the courts – and approached by officers if there is a match.

The Met claims that the technology has a very low failure rate, with the system only creating a false alert one in every 1,000 times.

However, using a different metric, last year research from the University of Essex said the tech only achieved eight correct matches out of 42, across six trials it evaluated.

The latest algorithm used by the Met is said to show no bias on the base of ethnicity, although it is less accurate for women than men.

Read Full Story Click here to comment

FROM OUR PARTNERS