Machine Learning Times
Machine Learning Times
Five Reasons to Go to Machine Learning Week 2020
 When deciding on a machine learning conference, why go...
Why Operationalizing Machine Learning Requires a Shrewd Business Perspective
 Originally published in Analytics Magazine For a rocket scientist,...
Sampling For Your Analysis
 So you have a mailing campaign you are about...
Accuracy Fallacy: The Media’s Coverage of AI Is Bogus
  A shorter version of this article was originally...

CONTINUE READING: Access the complete article in NPR, where it was originally published.  

3 years ago
Can Big Data Help Head Off Police Misconduct?


Big Data has been considered an essential tool for tech companies and political campaigns. Now, someone who has handled data analytics at the highest levels in both of those worlds sees promise for it in policing, education and city services.

For example, data can show that a police officer who has been under stress after responding to cases of domestic abuse or suicide may be at higher risk of a negative interaction with the public, data scientist Rayid Ghani says.

Ghani, the chief data scientist for President Obama’s re-election campaign in 2012, is now director of the Center for Data Science and Public Policy at the University of Chicago. He spoke to NPR’s Ari Shapiro about finding ways to use data analytics in fields where it’s not so common, like policing and city services.


It’s evolving to help not just predict outcomes, but change them.

On predicting police misconduct

The idea is to take the data from these police departments and help them predict which officers are at risk of these adverse incidents. Instead of the systems today — where you wait until something bad happens, and again, the only intervention at that point you have is punitive — we’re focusing on, “Can I detect these things early? And if I can detect them early, can I direct intervention to them — training, counseling.” …

Stress is a big indicator there. For example, if you are an officer and you have been … responding to a lot of domestic abuse cases or suicide cases, that is a big predictor of you being involved in an adverse incident in the near future.

On the limits of data analytics

We often work with historical data, which means if the data was collected under some sort of a biased process — so if people are giving loans and they’re biased in who they give loans to. Or if people have been collecting data on police misconduct … and if it was really hard to complain about police misconduct, then you’re not going to have the right level of data — you’ll only have data from people who really, really, really wanted to come and complain.

If you use those to build your algorithms, what happens is that the computer finds more of only those types of things. So your future predictions will be extremely biased. So we spend a lot of time thinking about, how do we detect that bias, and then how do we correct for that bias so we don’t make the wrong decisions?

CONTINUE READING: Access the complete article in NPR, where it was originally published.

By: NPR Staff, NPR Digital Media
Originally published at

Leave a Reply

Pin It on Pinterest

Share This