Justice should be blind, but the predictive policing software used in much of the US has bias and misunderstanding programmed right into it, says data scientist Cathy O’Neil.