“New Zealand Crunches Big Data to Prevent Child Abuse,” declared a Chronicle of Social Change headline on a 2015 story about The Chronicle’s favorite child welfare fad, predictive analytics, or as it should properly be called, data-nuking poor families.
The story quotes Paula Bennett, New Zealand’s former minister of social development, declaring at a conference: “We now have a golden opportunity in the social sector to use data analytics to transform the lives of vulnerable children.”
If implemented, the story enthuses, it would be a “world first.”
All this apparently was based on two studies that, it now turns out, used methodology so flawed that it’s depressing to think that things ever got this far. That’s revealed in a detailed analysis by Professor Emily Keddell of the University of Otago.
The studies that supposedly proved the value of predictive analytics attempted to predict which children would turn out to be the subjects of “substantiated” reports of child maltreatment.
Among the children identified by the software as being at the very highest risk, between 32 and 48 percent were, in fact, “substantiated” victims of child abuse. But that means more than half to more than two-thirds were false positives.
Think about that for a moment. A computer tells a caseworker that he or she is about to investigate a case in which the children are at the very highest level of risk. What caseworker is going to defy the computer and leave these children in their homes, even though the computer is wrong more than half the time?
But there’s an even bigger problem. Keddell concludes that “child abuse” is so ill-defined and so subjective, and caseworker decisions are so subject to bias, that “substantiation” is an unreliable measure of the predictive power of an algorithm. She writes:
How accurately the substantiation decision represents true incidence is … crucial to the effectiveness of the model. If substantiation is not consistent, or does not represent incidence, then identifying an algorithm to predict it will produce a skewed vision …
Turns out, it is not consistent, it does not represent incidence, and the vision is skewed. Keddell writes:
Substantiation data as a reflection of incidence have long been criticized by researchers in the child protection field … The primary problem is that many cases go [unreported], while some populations are subject to hypersurveillance, so that even minor incidents of abuse are identified and reported in some groups.
That problem may be compounded, Keddell says, by racial and class bias, whether a poor neighborhood is surrounded by wealthier neighborhoods (substantiation is more likely in such neighborhoods), and even the culture in a given child protective services office.
Predictive Analytics Becomes Self-fulfilling Prophecy
Algorithms don’t counter these biases, they magnify it.
Having a previous report of maltreatment typically increases the risk score. If it’s “substantiated,” the risk score is likely to be even higher. So then, when another report comes in, the caseworker, not about to overrule the computer, substantiates it again, making this family an even higher “risk” the next time. At that point, it doesn’t take a computer to tell you the children are almost certainly headed to foster care.
“prior substantiation may also make practitioners more risk averse, as it is likely to heighten perceptions of future risk to the child, as well as of the practitioner’s own liability, and lead to a substantiation decision being made.”
So the predictive analytics become a self-fulfilling prophecy.
Keddell also highlights the problems when even accurate data are misused by fallible human beings:
Several researchers note the tendency for individualised risk scores to be utilised in negative ways in practice, where actuarial approaches are prioritized over professional judgement. While statistical modellers may understand the tentative nature of statistical prediction or correlation … practitioners tend to treat statistical data, especially when stripped of its explanatory variables, as solid knowledge, which can function as a received truth.
But it turns out there may be one area where predictive analytics can be helpful. Keddell cites two studies in which variations on analytics were used to detect caseworker bias. In one, the researchers could predict which workers were more likely to recommend removing children based on questionnaires assessing the caseworkers’ personal values.
In another, the decisions could be predicted by which income level was described in hypothetical scenarios. A study using similar methodology uncovered racial bias.
So how about channeling all that energy now going into new ways to data-nuke the poor into something much more useful: algorithms to detect the racial and class biases among child welfare staff? Then we can protect children from biased decisions by “high risk” caseworkers.
Want to share your opinion or analysis with colleagues in the youth services field?
Join our one-of-a-kind Blogger Co-Op, and share in the benefits from your work!