predictive analytics in child welfare will continue right after this important message:
A bunch of data analysts, presumably working for a firm that sells sporting goods, are spying on a woman’s recreational habits. They have amassed so much data and their algorithms are so wonderful that it’s like having a camera watching her 24/7. Not only do they know her preferences, they know exactly why she prefers one sport over another and exactly what she’ll do next.
In other words, they’re stalking her.
But this is not presented as a warning of the dangers of predictive analytics. On the contrary, virtual stalking is what they’re selling.
That’s because the commercial is not aimed at consumers – such as the woman being stalked. The target audience is potential stalkers; in this case people who want to sell her stuff.
The maker of the stalking – er, analytics – software, and maker of the commercial, is SAP – as one of the “market leaders” in predictive analytics and a potential competitor in the child welfare market.
[UPDATE: MARCH 14, 2018: When I originally posted this, I embedded the actual commercial. But SAP seems to have removed it from YouTube.]
Unlike the bland reassurances given when people raise concerns about predictive analytics, the commercial reveals the real mindset of some of the human beings pushing big data.
Apparently, no one at SAP was creeped out by the ad’s Orwellian overtones. The slogan might as well have been “Big Data Is Watching You.” That alone ought to be enough to make anyone think twice about turning these companies loose in the child welfare field.
And it ought to make anyone think twice about giving this kind of power to secretive, unaccountable child welfare bureaucracies that have almost unlimited power to take away people’s children.
But in case that’s not enough, there’s also:
§ of profound racial bias in a field far closer to child welfare than sporting goods: criminal justice.
§ The about bias in predictive analytics in child welfare itself.
§ The fact that when the highly touted predictive analytics program being tested in Los Angeles predicted a child would be the victim of a “critical incident,” it apparently was wrong .
[Links to all of this can be found in our publication Big Data is Watching You]
There’s little harm in mistakenly telling a sporting goods salesperson to promote golf clubs to someone who’s more interested in soccer. There’s far more harm in telling a child protective services caseworker that a child is in grave danger when he’s not.