(Joe Belanger/Shutterstock)

Precognition software used to predict which prisoners will murder

Minority Report style systems might seem creepy, but they're no more flawed than human parole officers


Natasha Lennard
January 11, 2013 8:22PM (UTC)

When a prisoner goes on parole, a parole officer determines the level of supervision the individual requires based on the perceived likelihood of his committing a crime again. In a number of states, these determinations are being increasingly taken out of the flesh and blood hands of parole officers, settled instead by algorithms.

Precognition software, already in use in Baltimore and Philadelphia, determines how likely a prisoner is to commit murder and thus how much parole supervision he should receive. Wired explains how the algorithm works:

Advertisement:

To create the software, researchers assembled a dataset of more than 60,000 crimes, including homicides, then wrote an algorithm to find the people behind the crimes who were more likely to commit murder when paroled or put on probation. Berk claims the software could identify eight future murderers out of 100.

The software parses about two dozen variables, including criminal record and geographic location. The type of crime and the age at which it was committed, however, turned out to be two of the most predictive variables.

Of course, the software produces false positives and will thus condemn a number of parolees with no likelihood of committing murder to heavy supervision. This sort of Minority Report-reminiscent precognition system rightly raises concerns about criminal profiling based on all too few variables. Shawn Bushway, a professor of criminal justice at the State University of New York at Albany, noted the software could result in “punishing people who, most likely, will not commit a crime in the future.”

However, in this instance directing concerns at the technology itself is misplaced. This would assume that parole officers don't also base their decisions on which parolees remain "criminal risks" on limited, flawed and highly problematic variables, and also punish people accordingly. As the software's creator told Wired, the algorithms simply replace the ad hoc decision making done by parole officers. As long as parole boards employing the software don't view it as some perfect predictor of criminal futures, the precognition technology is no more troubling than the vagaries of human decision making in our prison systems.


Natasha Lennard

Natasha Lennard is an assistant news editor at Salon, covering non-electoral politics, general news and rabble-rousing. Follow her on Twitter @natashalennard, email nlennard@salon.com.

MORE FROM Natasha LennardFOLLOW natashalennardLIKE Natasha Lennard

Related Topics ------------------------------------------

Criminal Minority Report Murder Precognition Prison System

BROWSE SALON.COM
COMPLETELY AD FREE,
FOR THE NEXT HOUR

Read Now, Pay Later - no upfront
registration for 1-Hour Access

Click Here
7-Day Access and Monthly
Subscriptions also available
No tracking or personal data collection
beyond name and email address

•••





Fearless journalism
in your inbox every day

Sign up for our free newsletter

• • •