A police force in the UK is using an algorithm to help decide which crimes are solvable and should be investigated by officers. As a result, the force trialling it now investigates roughly half as many reported assaults and public order offences.
This saves time and money, but some have raised concerns that the algorithm could bake in human biases and lead to some solvable cases being ignored. The tool is currently only used for assessing assaults and public order offences, but may be extended to other crimes in the future.
When a crime is reported to police, an officer is normally sent to the scene to find out basic facts. An arrest can be made straight away, but in the majority of cases police officers use their experience to decide whether a case is investigated further. However, due to changes in the way crimes are recorded over the past few years, police are dealing with significantly more cases.
The Evidence Based Investigation Tool (EBIT) instead uses an algorithm to produce a probability score of a crime’s solvability. Kent Police, which has previously experimented with using algorithms (see “Predicting crime”, below), has been using EBIT for a year to assess the solvability of assaults and public order offences, such as threatening someone in the street. These types of offences account for around a third of all crime in the area.
Before the force began using EBIT, officers decided to pursue around 75 per cent of cases. This has now dropped to 40 per cent as a result of the algorithm, while the number of charges and cautions has remained the same, according to Kent Police.
“Police officers naturally want to investigate everything to catch offenders. But if the solvability analysis suggests there is no chance of a successful investigation, the resources might be better used on other investigations,” says Ben Linton at the Metropolitan Police, who isn’t involved with the project.
Kent McFadzien, at the University of Cambridge, created EBIT by training the algorithm on thousands of assaults and public order cases. It identified eight factors that seemed to affect whether a case was solvable, including whether there were witnesses, CCTV footage or a named suspect.
As these factors could change over time, EBIT always recommends one or two crimes with low solvability scores for investigation each day. The officers involved aren’t aware of this score, so this is a blind test of the algorithm’s effectiveness. “It’s a permanently ongoing trial,” says McFadzien.
However, because the technology bases its predictions on past investigations, any biases contained in those decisions may be reinforced by the algorithm. For example, if there are areas that don’t have CCTV and police frequently decided not to pursue cases there, people in those places could be disadvantaged.
“When we train algorithms on the data on historical arrests or reports of crime, any biases in that data will go into the algorithm and it will learn those biases and then reinforce them,” says Joshua Loftus at Stanford University in California.
McFadzien’s blind tests are a good way to help tackle that issue, but there is a separate problem with police algorithms that can’t be so easily remedied, says Loftus.
Police forces only ever know about crimes they detect or have reported to them, but plenty of crime goes unreported, especially in communities that have less trust in the police.
This means the algorithms are making predictions based on a partial picture. While this sort of bias is hard to avoid, baking it into an algorithm may make its decisions harder to hold to account compared with an officer’s. John Phillips, superintendent at Kent Police, says that for the types of crimes that EBIT is being used for, under-reporting isn’t an issue and so shouldn’t affect the tool’s effectiveness.
Kent was the first police force in the UK to experiment with predictive policing, a technology used to suggest areas where crime is likely to occur. It used a proprietary machine-learning algorithm, provided by US firm PredPol, to predict potential crime hotspots over a five-year period.
The force spent £150,000 a year on the contract with PredPol, but stopped using the tech last year.
An internal review, published in 2014 and obtained through a freedom of information request by New Scientist, reveals that officers struggled to make use of the system’s predictions due to time constraints. The algorithm suggested some 520 hotspot boxes per day, but police only visited 86, on average. “Officers are not getting to enough of the boxes to make a significant impact on crime,” the review said.
More on these topics: