Uploaded by Heather Franklin

Predictive Analytics Study 2

advertisement
Child Abuse Review (2019)
Published online in Wiley Online Library
(wileyonlinelibrary.com) DOI: 10.1002/car.2547
Can Predictive
Algorithms Assist
Decision-Making in Social
Work with Children and
Families?
Philip Gillingham*
School of Nursing, Midwifery and
Social Work, University of Queensland,
Brisbane, Queensland, Australia
Decision support systems (DSS) which incorporate algorithms trained using
administrative data have been promoted as the next promising development in initiatives
to improve and assist decision-making in social work with children and families. In this
article, research and grey literature about DSS designed, and in some cases
implemented, in the USA, the Netherlands, Australia and New Zealand are analysed
to assess the extent to which they are currently able to meet this promise. The
challenges of developing DSS for social work are identified and ideas are suggested
to overcome them. © 2019 John Wiley & Sons, Ltd.
KEY PRACTITIONER MESSAGES:
• Decision support systems (DSS) are being developed for social work with children
and families and they may transform how decisions are made with important
consequences for both service users and the profession.
• Current DSS are mostly not very accurate in their predictions and further work is
required in their development and on the data used to train the algorithms.
• The next stage in the development of DSS will require long-term commitment by
social workers to work in collaboration with data scientists and researchers. It will
also be expensive. Social workers need to lead the debate about whether DSS
are worth the investment.
KEY WORDS: decision support systems; big data; predictive analytics
here is a plethora of research about decision-making in social work with
children and families. Much of this research illustrates the influences on
how social workers make decisions, which may be demographic, personal
or organisational, leading to the conclusion that decision-making is a highly
subjective and potentially biased process (Buckley, 2003; Keddell, 2013;
Stanley, 2005). Decision-making tools based on blends of actuarial
and consensus-based approaches to risk assessment were introduced in
T
*Correspondence to: Dr Philip Gillingham, University of Queensland, School of Nursing, Midwifery and
Social Work, St Lucia Campus, Brisbane, Queensland 4067, Australia. E-mail p.gillingham@uq.edu.au
© 2019 John Wiley & Sons, Ltd.
Accepted: 27 October 2018
‘Decision support
systems are being
developed for social
work with children
and families and they
may transform how
decisions are made’
Gillingham
‘Big data approaches
… have been
successful in
informing decisions
about resource
allocation in
healthcare and
health promotion’
© 2019 John Wiley & Sons, Ltd.
many jurisdictions to increase accuracy and consistency in decision-making,
in response to both media and official reports about child deaths and
the perceived failings of child protection services (Reder and Duncan, 2004;
Schwalbe, 2004). While such tools may appear sound theoretically,
in practice their use may not promote consistency and practitioners may
not find them useful (Gillingham, 2009). Whether decisions should be
made ‘subjectively’ on a case-by-case basis or more ‘objectively’ using
tools is a debate that continues, linked as it is to wider debates
about the very nature of social work as a profession (Devlieghere and
Roose, 2018).
Big data approaches, which use algorithms to analyse or mine large
datasets, have been successful in informing decisions about resource
allocation in healthcare and health promotion. It has been proposed that such
approaches to assist decision-making may be the next big development in
child and family social work, following on from actuarial risk assessment
(Macchione et al., 2013). Specifically, big data approaches have been
promoted in the development of decision support systems (DSS). DSS can
be differentiated from actuarial and consensus-based decision-making tools,
such as Structured Decision Making (CRC, 2018) and the Illinois Structured
Decision Support Protocol (Kang and Poertner, 2006). DSS incorporate
algorithms that have been trained to calculate the likelihood of a particular
outcome for a service user and make recommendations to practitioners to
support decision-making. Within the fields of information science or
information technology, this may be referred to as ‘predictive analytics’.
Technological developments such as faster computing power, the
accumulation of data about service users and service delivery in electronic
information systems, and the ability to extract and match data about people
from multiple databases have, over the last ten years, made big data
approaches much more feasible in social work.
The idea of using algorithms to assist and support decision-making in social
work with children and families is not a recent development. For example,
Schoech et al. (1985) explored the potential of what they refer to as ‘artificial
intelligence’ to develop ‘expert systems’ to support professional decisionmaking. Research-based development of DSS continued in the 21st century
and identified particular challenges in their design and implementation such
as making them useful in the contexts in which they are implemented (Fitch,
2006, 2007; Foster and Stiffman, 2009), getting practitioners to use them
(Fitch, 2006, 2007; Foster and Stiffman, 2009) and accounting for the
importance of emotion and bias in human decision-making (MacFadden and
Schoech, 2010).
Despite the challenges that have been identified by research-based
development projects and the identified need for more research (MacFadden
and Schoech, 2010), significant investments have already been made in
implementing DSS in some jurisdictions. It is therefore timely to review where
developments with DSS have led and assess where they might go in future.
Using six examples from North America, Europe and Australasia, this article
provides an overview and critical appraisal of current developments. It is also
important that social workers are aware of these developments and can engage
in both debate about DSS and development processes (Fitch, 2006, 2007). One
aim of the article is therefore to inform social workers and stimulate debate
Child Abuse Rev. (2019)
DOI: 10.1002/car
Decision Support Systems and Child and Family Social Work
within the profession. DSS may have profound consequences for service users
as well as social workers. For example, both social workers and service users
may feel alienated from the decision-making process if DSS are used to
replace, rather than support, human decision-making. Explaining decisions
that are either dictated or even supported by DSS may be more difficult, and
may discourage those affected by decisions from challenging them. Big data
approaches in the social welfare sector, as the examples show, may also be
fallible and unreliable, leading to injustices for service users (see also
Gillingham, 2016a).
The other aim in this article is to use the lessons learned from the examples
provided to suggest strategies for the future development of DSSs. The
language used in this article is deliberately non-technical and aimed at a social
work audience, so for those who are more au fait with the technical and
mathematical bases of big data and machine learning, it may seem overly
simplistic.
‘DSS may have
profound
consequences for
service users as well
as social workers’
Background
The author is engaged in a four-year programme of research supported by an
Australian Research Council Future Fellowship which aims to improve how
electronic information systems are designed and used in the social welfare
sector. An important consideration in this research has been to explore what
can be done constructively with the data that social welfare agencies have
accumulated about service users and service delivery, over and above using
them for reporting purposes. Service evaluation has been one focus
(Gillingham, 2018), but, as stated above, big data approaches now offer other
possibilities.
There is, however, very little published research about DSS and their
application in the social welfare sector. As described above, there are published
articles which describe the process of designing an algorithm, but these
approaches tend to become obsolete very quickly given the speed of
technological developments in big data. DSS are a very recent and emerging
phenomenon in the social welfare sector and using specialist search engines
to identify published academic research yields very little, with the exception
of four out of the six examples analysed in this article. The author had to rely
on ‘insider knowledge’ in finding one of the other examples (Broward County)
and an online network of researchers to find the other (the Rapid Safety
Feedback (RSF) programme). The rationale for including these examples is
that they use a range of different techniques to train algorithms and each, when
analysed and considered alongside the others, provides insights into how DSS
can be improved in the future. The examples may therefore be considered as a
purposive sample (Alston and Bowles, 2012, p. 97).
‘Very little published
research about DSS
and their application
in the social welfare
sector’
Developing and Implementing DSS
Predictive Risk Modelling in New Zealand
The author has published two articles which examine the development of
predictive risk modelling (PRM) in the New Zealand child welfare system
and the following is a summary of the main points raised in them (Gillingham,
© 2019 John Wiley & Sons, Ltd.
Child Abuse Rev. (2019)
DOI: 10.1002/car
Gillingham
‘Concerns were
raised about… the
ethics of using data
collected for one
reason for a
completely different
purpose’
‘It is imperative that
the dataset used to
train an algorithm
accurately
represents what it is
being taught to
predict’
© 2019 John Wiley & Sons, Ltd.
2016a, 2017). PRM was developed by a team at the Centre for Applied
Research in Economics at the University of Auckland (CARE, 2012;
Vaithianathan et al., 2013) as part of a wide range of reforms to the child
welfare system in New Zealand. The team used data from the child welfare,
health, education and public welfare benefit systems to train an algorithm to
identify children at risk of maltreatment at the point at which their parents
began to claim public welfare benefit. Having identified these children, the
aim was to intervene to prevent maltreatment using supportive services.
Further detail about how this was achieved is provided in Gillingham (2016a,
2017) and documents later released by the Ministry of Social Development
(2014a, 2014b) in New Zealand. After much work, the team managed to
develop an algorithm that was 76 per cent accurate and plans were made to trial
and implement PRM.
PRM attracted much media and academic attention. For example, concerns
were raised about: (a) the ethics of using data collected for one reason for a
completely different purpose without the consent of the person who owned
the data; (b) the stigmatisation of children and families identified by PRM;
(c) the legal and moral basis for intervention based on the calculations of
an algorithm; and (d) a lack of transparency about how PRM had been
developed and would operate (Dare, 2013; Keddell, 2015). Eventually, plans
to trial PRM were shelved amid political concerns that children were being
treated as ‘lab rats’. In the meantime, the author had been sent some of
the publicly available documentation about the development of PRM.
Combining knowledge of machine learning with knowledge about the child
welfare system, a serious flaw in how PRM had been trained was identified.
PRM was trained to identify independent variables (child and family traits)
that were associated with the dependent variable of child maltreatment. In
this form of supervised learning, it is imperative that the dataset used to train
an algorithm accurately represents what it is being taught to predict (child
maltreatment). The team at the University of Auckland used data from the
Ministry of Social Development about children who had been in the child
welfare system and whose cases had been ‘substantiated’. The team
understood ‘substantiated’ to mean that maltreatment had been proven to
have occurred. The term ‘substantiated’ though has a different meaning in
child welfare legislation and policy in Australia and New Zealand, and is
part of the threshold criteria for further intervention by welfare services after
an investigation of alleged child maltreatment has been concluded. In New
Zealand, ‘substantiation’ may be applied to children who have been
maltreated, the siblings of those children (whether maltreated or not),
children considered to be at risk of maltreatment and their siblings,
unaccompanied minors seeking asylum and young people needing access
to mental health services. Therefore any dataset of substantiated cases
actually contains many more children who have not been maltreated (those
deemed to be at risk, siblings, young people with mental health problems,
etc.) than who have, making it entirely unsuitable for training an algorithm
to predict maltreatment. In technical language, this mistake amounts to a
‘mislabelling’ of the dataset. It also suggests that the team at the University
of Auckland misunderstood what the dataset represented, and that the
public servants who supplied it lacked knowledge about how supervised
learning works.
Child Abuse Rev. (2019)
DOI: 10.1002/car
Decision Support Systems and Child and Family Social Work
Having noticed this, the author contacted the Ministry of Social
Development in New Zealand and was invited to address senior public
officials; and, in due course, plans to trial PRM were dropped. Subsequently,
full documentation about the development of PRM was released (Ministry of
Social Development, 2014a, 2014b) and the author undertook an analysis of
both the methods and any insights that the development process may have
provided into the child welfare system and those within it. Assuming that
‘substantiation’ (rather than actual maltreatment) is of interest in terms
of prediction, PRM identified three main independent variables that
were considered predictive: poverty, single parenthood and previous
involvement with the child welfare system. None of these factors
are new insights and would be known to anyone working in the child
welfare system.
The development of PRM demonstrated that, at least in the abstract, it is
possible to use administrative data collected routinely for other purposes to
predict, with 76 per cent accuracy, a particular outcome for a group of children.
This means though that PRM would be wrong in one out of every four cases.
To put this another way, the algorithm was trained using data about cases that
were substantiated and these data were created by human decision-making
processes. The use of these data rests on the assumption that the humans
making the decisions made the correct decision every time (100%). At 76
per cent, PRM was only three-quarters as accurate as the human decisionmakers. For human decision-makers in the child protection system, such
inaccuracy in deciding which cases to investigate and at what point children
need to be removed from the care of their parents would not be acceptable. It
also demonstrated the problems that can arise through misunderstandings
about the accuracy of data and what the data actually represent.
‘The use of these
data rests on the
assumption that the
humans making the
decisions made the
correct decision
every time’
The Allegheny Family Screening Tool
The Family Screening Tool (FST) being used in Allegheny (a suburb of
Philadelphia, USA) is another example of a supervised learning algorithm.
Details about its development are scant and the following description is
paraphrased from a frequently asked questions document available on the
internet (Allegheny County, 2017). The FST is designed to help child
protection workers decide whether to screen in, or accept, a notification for
investigation of alleged child maltreatment. Based on data about the child
and his or her family, the FST calculates the likelihood that the child will be
re-notified within two years (the dependent variable) if the notification is
screened in (and investigated) and the likelihood that a child will be re-notified
within two years if the notification is screened out (and not investigated). Again
the emphasis here is on assisting practitioners to target the most vulnerable
children, but alas there is no published research about the extent to which this
has been achieved.
The algorithm is claimed to use over 100 factors about a child, drawing from
a data warehouse that contains data from 29 different sources, including child
protection, mental health, drug and alcohol services, and county court data.
The speed at which it can do this is obviously much faster than any human
could, the implication being that it is much more efficient and accurate than
human decision-making. These claims are of interest because, as mentioned
© 2019 John Wiley & Sons, Ltd.
Child Abuse Rev. (2019)
DOI: 10.1002/car
Gillingham
‘Only three
independent
variables, or factors
about children,
rather than 100 were
found to be
significantly
predictive’
with PRM in the previous section, only three independent variables, or factors
about children, rather than 100 were found to be significantly predictive. As
mentioned below, Wijenayake et al. (2018) found the same finding. While
the amount of data that the FST considers and the speed at which it does
may sound impressive, it may be misleading in that most of the factors it
considers do not contribute to its accuracy. Unfortunately, no information is
publicly available to support or refute this point, but questioning the predictive
ability of different variables, or factors about a child used by an algorithm, is
important in considering its efficacy and usability.
The RSF Programme
‘Overestimated the
likelihood of
maltreatment for
thousands of
children… and
underestimated the
risk in a few cases
where children either
died or were
seriously harmed’
The following is taken from an article by Lindsay Gloor (2017) published in
the Chicago Herald Tribune on the 6 December 2017. Various other mentions
of the RSF programme are made on the worldwide web, but all seem to have
been sourced from the Tribune article. The article reports on how the Illinois
Department of Children and Family Services decided to stop using the RSF
programme because of errors in its predictions. The RSF programme is a
supervised learning algorithm that was trained using child welfare data to
identify children most at risk of harm as they enter the child welfare system,
assigning them a score of up to 100. In short, the RSF programme
overestimated the likelihood of maltreatment for thousands of children, leading
to the department being overwhelmed, and underestimated the risk in a few
cases where children either died or were seriously harmed. For example, the
RSF programme assigned a 100 per cent chance of death or serious injury in
the next two years to 369 children, all under age nine.
The RSF programme was developed by a private company and the algorithm
and exactly how it was developed are not open to public scrutiny, hence it is
difficult to determine why it went wrong. However, given that the RSF
programme appeared to be getting worse at prediction, it would be fair to
speculate that if it used only child welfare data and was periodically (or even
continually) updated with child welfare data, then it may have fallen prey to
sample selection bias. Sample selection bias occurs when the sample of data
being used is overly biased in one or more ways in terms of its characteristics
than what would be expected in the general population (see Berk, 1983).
Within child welfare data, child maltreatment is over-represented as an
outcome and so the algorithm learns to associate all sorts of factors or
characteristics with child maltreatment. In the general population, these
characteristics may have no association with child maltreatment. Over time,
this bias increases, leading to increasingly erroneous estimations of risk.
Predicting Recidivism in Domestic Violence
Wijenayake et al. (2018) describe how they developed an algorithm using
publicly available data to predict the likelihood that men charged with crimes
related to domestic violence will offend again. This example is notable for a
number of reasons. The team used a decision tree approach to develop the
algorithm and provide a clear and accessible explanation of the process, which
is in contrast to some articles that, through their presentation of the details of
quantitative analysis, would only be accessible to academics with considerable
© 2019 John Wiley & Sons, Ltd.
Child Abuse Rev. (2019)
DOI: 10.1002/car
Decision Support Systems and Child and Family Social Work
prior knowledge. More generally, in the Australian context where the algorithm
was developed, both federal and state governments have recently made
significant investments in services and new legislation to deal with what is
perceived to be a rapidly growing, or at least increasingly public, social
problem, and so the research is very timely. In addition to demonstrating
how this can be done, a few other points about big data emerge from the article
that can be considered as problematic, or even emblematic, for the social
welfare sector.
The algorithm is claimed to have an accuracy of just under 70 per cent,
similar to that of PRM, and again this may be impressive in the abstract, but
the question arises as to how useful this is to decision-makers in the real world.
As with risk assessment in child protection mentioned above, the consequences
of a wrong decision could be catastrophic. Human decision makers do make
mistakes which lead to harm in 30 per cent of cases.
The other interesting point that the researchers mention again contradicts
some of the rhetoric that is used to promote big data approaches in social
welfare. They state that the algorithm can achieve almost the same accuracy
with only three, rather than 11, factors about potential recidivists. As with
PRM, introducing other factors, such as demographic and other behavioural
information, did not lead to any significant or real improvement in its
performance. Using only three factors leads to only small trees, which may
be preferable as they are easier for professionals to understand and explain to
others how predictions and decisions have been made.
‘The consequences
of a wrong decision
could be
catastrophic’
The Broward County Experience
In another instance where a big data approach has been developed but has yet
to be applied, Schwartz et al. (2017) demonstrate how both prescriptive and
predictive approaches to machine learning may improve the efficiency of a
child protection system. They describe how Broward County in Fort
Lauderdale, Florida has experienced increases in the numbers of children being
notified, investigated, substantiated and, in particular, returning to the child
protection system leading to overload, a problem shared by many other
jurisdictions across the English-speaking world. They used supervised
machine learning and propensity score matching to develop: predictive models
for determining the likelihood of a child being notified subsequently; and
prescriptive models for determining the type and level of service that were
most likely to prevent the child entering the child protection system again.
Schwartz et al. (2017) acknowledge that using data about subsequent
notifications of maltreatment and substantiation is not ideal as they found that
decision-making between practitioners differed even when based on the same
factors. Put simply, inconsistency in the data makes it harder for an algorithm
to identify patterns and correlations within a dataset. The other problem that they
identified within the dataset was that contact between families and supportive
services did not reduce the likelihood of a child re-entering the child protection
system. This phenomenon has also been found in other quantitative studies of
child protection data (Jenkins et al., 2017) and, as Schwartz et al. (2017) suggest,
it is an area that requires research to determine why this should be so.
Schwartz et al. (2017) conclude that if their system was implemented, it has
the potential to improve outcomes for children and families by up to 30 per cent,
© 2019 John Wiley & Sons, Ltd.
‘Inconsistency in the
data makes it harder
for an algorithm to
identify patterns and
correlations within a
dataset’
Child Abuse Rev. (2019)
DOI: 10.1002/car
Gillingham
with outcomes defined as being referred to the appropriate level of service (from
supportive services to investigative and court interventions). Another point of
interest is that they mention Russell's (2015) and Coohey et al.'s (2013) four
standards for assessing a predictive model: validity, equity, reliability and
usefulness. These standards would provide a useful and clear framework for
empirical research that aimed to evaluate the use of DSS. Though Schwartz
et al. (2017) do not state explicitly, it appears that the approach they have used
means that the algorithm, in making a recommendation to a practitioner, would
also be able to show why it had made that recommendation, that is, show which
factors about a child or his or her family it had considered. As debates about the
need for algorithmic accountability have shown (Binns, 2018), it is not only
practitioners who need to understand how an algorithm works but also those
affected by that decision. This is especially important for service users who
want to question or contest a decision made by a practitioner.
A DSS in Amsterdam
Amrit et al. (2017) developed a decision support system to support
paediatricians and other healthcare workers in Amsterdam to identify children
under their care who are at risk of maltreatment. They used healthcare records,
which they claim were complete, and both structured data (entered into a
specific field, like age, height and weight) and, as they claim uniquely,
unstructured data (the notes made by healthcare workers in free-text fields).
Amrit et al. (2017) used a complex set of procedures to test different types
of algorithms and found that by combining their approaches to both structured
and unstructured data, they produced a tool that was just over 90 per cent
accurate. It is planned that the final tool will be implemented in the healthcare
system towards the end of 2018. The health authority is, however, open to
researchers evaluating its impact on decision-making by health professionals
(personal communication).
Discussion
‘Data scientists and
social workers need
to collaborate to
ensure that both
parties are clear
about what the data
actually represent’
© 2019 John Wiley & Sons, Ltd.
From the examples in the previous section, a few points arise which need to be
considered in the future development of DSS. Data scientists and social
workers need to collaborate to ensure that both parties are clear about what
the data actually represent in order to avoid mislabelling, a point which is more
fully explored below. In terms of accuracy, 70 per cent may be acceptable for
clinical decision-making, according to Zhang et al. (2014), but this value is
too low in child and family social work given the consequences of a bad
decision. Adding more independent variables does not necessarily improve
accuracy. Many variables may have little or no actual correlation with what
an algorithm is trained to predict and there may be only two or three that are
highly correlated. In the examples of PRM and predicting recidivism in
domestic violence, the main independent variables used by the algorithm are
those which have been identified long ago by researchers and paediatricians.
Given that, it is questionable why DSS would be helpful as they are not likely
to ever perform any better than trained practitioners. Lastly, if an algorithm is
updated (and trained) with new data on a regular basis, there is the potential for
Child Abuse Rev. (2019)
DOI: 10.1002/car
Decision Support Systems and Child and Family Social Work
sample selection bias which then undermines the accuracy of DSS. How the
performance of DSS can be improved emerges as a key challenge.
What does improve performance, as Schwartz et al. (2017) demonstrate, is
adding extra statistical procedures like pair matching. The main area for
improvement, though, as hinted at by Schwartz et al. (2017) (see also Amrit
et al., 2017), is the quality, rather than the quantity, of the data used to train
algorithms. Each of the projects mentioned above has used what Salganik
(2018) refers to as ‘readymade’ datasets. He warns that practical (and, of
course, ethical) problems can arise when using data for a purpose other than
that for which they were collected. This may be the usual way that big data
approaches are applied and may be part of their appeal, but there are particular
problems with how data are generated in the child protection field which
undermine the quality. These problems stand in stark contrast to the healthcare
data (both structured and unstructured) used by Amrit et al. (2017).
Child protection data are, obviously, created by humans and, unlike data
from health databases, they constitute a particular way of representing (or
reconstructing, see D'Cruz, 2004) a version of events, decisions and opinions.
These representations are subject to the inaccuracies, biases, prejudices and
inconsistencies that the literature describes about decision-making in child
protection. As Buckley (2003) found, they are part of the ‘official’ version of
practice and might be quite different to the ‘unofficial’ version or what actually
happens in a case. As information is entered into an information system, it has
to be categorised according to the fields built into the information system,
which may or may not fit the circumstances that the practitioner has observed.
This can happen in many ways (see Gillingham, 2015a) but an obvious
example is the level of detail required by the information system. For example,
a common question in risk assessment tools concerns illicit drug use by
caregivers. Ticking a yes/no box in response to such a question is not only
overly simplistic but confounding. In terms of data, caregivers who
occasionally smoke marijuana after the children have gone to sleep are put in
the same category as caregivers who inject heroin two or three times a day
and spend much of their time finding the means to do so. Dick (2017) calls this
the ‘flattening effect’ of categorising data. Clearly, there are different levels of
risk of harm or neglect posed by each scenario.
Case file data, both structured and unstructured, may also be incomplete
(Gillingham, 2009) as busy practitioners are compelled to move on to another
case, and even if they are complete, the version of practice that they convey, as
alluded to above, is only partial (Skotte, 2018). Even when practitioners go to
some lengths to explain the rationales for decisions in case notes, it is not
possible to convey the subtleties and nuances of decision-making that research
about it has uncovered (see, for example, Keddell, 2013). As has been pointed
out, the presence of various factors in a child's life is not the only focus of risk
assessment; it is the interaction between these factors that is important (Reder
et al., 1993). For example, the use of illicit drugs, at any level, may exacerbate
mental illness and vice versa. The presence of a particular factor may also be
positive or negative or somewhere in between and vary from day to day. The
presence of grandparents in a child's life might be a protective factor but it
depends on the grandparents and requires further assessment.
The above is merely an overview of the problems with the data held in child
protection information systems, but it does demonstrate that for DSS to
© 2019 John Wiley & Sons, Ltd.
‘Problems can arise
when using data for a
purpose other than
that for which they
were collected’
Child Abuse Rev. (2019)
DOI: 10.1002/car
Gillingham
‘We may need to
move on from using
ready-made datasets
to ‘custom-made’
datasets’
‘Collaboration needs
to go far beyond a
supposed
participatory design
process that mainly
seeks ‘buy in’ from
end users’
become more accurate and therefore potentially more useful, we may need to
move on from using ready-made datasets to ‘custom-made’ datasets
(Salganik, 2018) to train algorithms. As the authors mentioned in the previous
section have demonstrated, promising methods to develop DSS exist, but the
resulting products are only going to be as good as the data used to train them.
As to what needs to be included in a custom-made dataset is a matter that
requires more research and debate. Going back to Wijenayake et al. (2018)
and PRM in New Zealand, it may be that adding extra factors about a service
user or his or her circumstances may not improve the accuracy of DSS, but
in both cases the factors that could be added were not chosen because there
was reason to believe they might be predictive. They were added because they
were available. Using the extant literature about the antecedents of child
maltreatment and the process of decision-making, it might be possible to
deduce more precisely which factors might improve the accuracy of DSS if
they were included in a dataset. For example, a Delphi study by Powell
(2003) on early indicators of child abuse and neglect found five physical, 13
behavioural/developmental and 16 parenting indicators that may occur
separately or cluster together.
Developing DSS using custom-made datasets would have to involve, as
mentioned above, close collaboration between practitioners, data scientists
and child protection and social work researchers. This collaboration needs to
go far beyond a supposed participatory design process that mainly seeks
‘buy in’ from end users (Gillingham, 2015b). It requires the skills and
knowledge of all parties to develop, as suggested previously, a ‘minimum
dataset’ (Gillingham, 2016b) which involves the collection and entry of what
are considered to be highly relevant factors or predictors of the various forms
of child maltreatment. It would require changes in the way that practitioners
collect and record data. Indeed, using information systems as a tool for
research and the development of DSS would require a very different approach
to how information systems are currently used in child and family social work,
which is mostly to ensure compliance with procedures and standards
(Gillingham and Graham, 2016). Practitioners already complain about having
to spend inordinate amounts of their work time entering data into information
systems (Gillingham, 2016c) and adding additional data entry requirements
may be unsuccessful. A focus on the entry of very specific data, for example,
the indicators of maltreatment found by Powell (2003), with a renewed
emphasis on accuracy and completeness may have to replace some of the more
compliance-driven requirements for data entry. Data must also be retrievable in
a way that best suits the development of the most promising designs of
algorithms and the process will be experimental as different factors, algorithms
and statistical procedures are tested to achieve the highest level of accuracy
possible. Starting again in terms of collecting data means that future
development of DSS will not only be a costly exercise but also a long-term
endeavour.
Conclusion
To address the question in the title of this article, it is possible that DSSs can be
developed to assist decision-making in social work with children and families.
© 2019 John Wiley & Sons, Ltd.
Child Abuse Rev. (2019)
DOI: 10.1002/car
Decision Support Systems and Child and Family Social Work
However, as the examples described above show, current attempts using purely
administrative data, or ready-made datasets, have not led to DSS that are
sufficiently useful in practice. However, the learnings from such attempts have
led to more sophisticated approaches that incorporate extra statistical
procedures and unstructured data. It has been proposed that the most promising
route to developing more accurate and reliable DSS may be achieved by
developing a custom-made dataset, based on research about the factors
associated with child maltreatment.
None of the above discussion addresses the question of whether DSS should
be developed to assist decision-making in social work with children and
families or any other field of social work. How decision-makers in social work
(both do and will) react to DSS based on algorithms and big data has yet to be
evaluated. A critic might argue that DSS, as with actuarial risk assessment
before them, only identify: (a) which potential service users most need a
service, such as children who are most likely to be renotified, re-investigated
and those whose cases are most likely to be re-substantiated; and (b) those at
highest risk of significant harm. It could be argued that professionals are
already good enough at assessing who these service users might be, with or
without actuarial risk assessment tools. Further, the area where social workers
actually need more help and resources is in the development and delivery of
interventions that improve outcomes for service users. Clearly, there needs to
be debate within the profession about whether it is desirable for DSS to be
developed for social work. This debate needs to be informed by the key
contribution of this article, namely, the conclusion that DSS can be developed
to assist decision-making in social work with children and families, but they
will be a long-term and expensive process.
Funding
‘There needs to be
debate within the
profession about
whether it is
desirable for DSS to
be developed for
social work’
Australian Research Council Future Fellowship (FT170100080).
Acknowledgements
This research was supported by the Australian Research Council
(FT170100080).
References
Allegheny County. 2017. Developing Predictive Risk Models to Support Child Maltreatment
Hotline Screening Decisions. Allegheny County Department of Human Services:
Pittsburgh, PA. Available: https://www.alleghenycountyanalytics.us/index.php/2017/04/17/
developing-predictive-risk-models-support-child-maltreatment-hotline-screening-decisions/
[22 November 2018].
Alston M, Bowles W. 2012. Research for Social Workers: An Introduction to Methods, Third
edn. Allen and Unwin: Crows Nest, NSW, Australia.
Amrit C, Paauw T, Aly R, Lavric M. 2017. Identifying child abuse through text mining and
machine learning. Expert Systems with Applications 88: 402–418. https://doi.org/10.1016/j.
eswa.2017.06.035.
Berk RA. 1983. An introduction to sample selection bias in sociological data. American
Sociological Review 48(3): 386–398.
© 2019 John Wiley & Sons, Ltd.
Child Abuse Rev. (2019)
DOI: 10.1002/car
Gillingham
Binns R. 2018. Algorithmic Accountability and Public Reason. Philosophy and Technology.
31(4): 543–556. https://doi.org/10.1007/s13347-017-0263-5.
Buckley H. 2003. Child Protection Work: Beyond the Rhetoric. Jessica Kingsley: London.
CARE. 2012. Vulnerable Children: Can Administrative Data Be Used to Identify Children
at Risk of Adverse Outcomes? Centre for Applied Research in Economics, University
of Auckland: Auckland, New Zealand. Available: https://www.msd.govt.nz/documents/
about-msd-and-our-work/publications-resources/research/vulnerable-children/auckland-uni
versity-can-administrative-data-be-used-to-identify-children-at-risk-of-adverse-outcome.pdf
[22 November 2018].
Coohey C, Johnson K, Renner LM, Easton SD. 2013. Actuarial risk assessment in child
protective services: Construction methodology and performance criteria. Children and Youth
Services Review 35: 151–161.
CRC. 2018. The SDM Model in Child Protection. Children's Research Center, National Council
on Crime and Delinquency: Madison, WI. Available: https://www.nccdglobal.org/assessment/
sdm-structured-decision-making-systems/child-welfare [22 November 2018].
Dare T. 2013. Predictive risk modelling and Child Maltreatment: an ethical review. University
of Auckland: Auckland, New Zealand. Available: https://www.msd.govt.nz/documents/
about-msd-and-our-work/publications-resources/research/predictive-modelling/00-predicitverisk-modelling-and-child-maltreatment-an-ethical-review.pdf [22 November 2018].
D'Cruz H. 2004. Constructing meanings and identities in child protection practice. Tertiary
Press: Melbourne, Australia.
Devlieghere J, Roose R. 2018. A policy, management and practitioners' perspective on social
work's rational turn: there are cracks at every level. European Journal of Social Work.
https://doi.org/10.1080/13691457.2018.1460324.
Dick S. 2017. Algorithmic Accountability. Presented at the Internet Policy Research Initiative,
Massachusetts Institute of Technology, 23 March 2017. Available: https://internetpolicy.mit.
edu/event-angwin-algo-account-2017/ [23 November 2018].
Fitch D. 2006. Examination of the Child Protective Services Decision-Making Context with
Implications for Decision Support System Design. Journal of Social Service Research
32(4): 117–134.
Fitch D. 2007. Structural equation modeling the use of a risk assessment instrument in child
protective services. Decision Support Systems 42: 2137–2152.
Foster KA, Stiffman AR. 2009. Child Welfare Workers' Adoption of Decision Support
Technology. Journal of Technology in Human Services 27(2): 106–126.
Gillingham P. 2009. The use of assessment tools in child protection: An ethnomethodological
study (PhD thesis). University of Melbourne, Australia. Available: http://repository.unimelb.
edu.au/10187/4337 [27 November 2018].
Gillingham P. 2015a. Implementing electronic information systems in human service organizations:
The challenge of categorization. Practice: Social Work in Action 27(3): 163–175.
Gillingham P. 2015b. Electronic information systems and human services organisations: Avoiding
the pitfalls of participatory design. The British Journal of Social Work 45(2): 651–666.
Gillingham P. 2016a. Predictive risk modelling to prevent child maltreatment and other adverse
outcomes for service users: Inside the “black box” of machine learning. The British Journal of
Social Work 46(4): 1044–1058.
Gillingham P. 2016b. Electronic information systems and human service organisations: The
needs of managers. Human Service Organizations: Management, Leadership & Governance
40(1): 51–61.
Gillingham P. 2016c. The use of electronic information systems to guide practice in social
welfare agencies: The perspectives of practitioners as end users. Practice: Social Work in
Action 28(5): 357–372.
Gillingham P. 2017. Predictive risk modelling to prevent child maltreatment: Insights and
implications from Aoteaora/New Zealand. Journal of Public Child Welfare 11(2): 150–165.
Gillingham P. 2018. The evaluation of practice frameworks for social work with children and
families: Exploring the challenges. Journal of Public Child Welfare 12(2): 190–203.
Gillingham P, Graham T. 2016. Designing electronic information systems for the future: Facing
the challenge of New Public Management. Critical Social Policy 36(2): 187–204.
Gloor L. 2017. Data mining program designed to predict child abuse proves unreliable, DCFS
says. Chicago Herald Tribune, 6 December 2017. Available: http://www.chicagotribune.
com/news/watchdog/ct-dcfs-eckerd-met-20171206-story.html [22 November 2018].
© 2019 John Wiley & Sons, Ltd.
Child Abuse Rev. (2019)
DOI: 10.1002/car
Decision Support Systems and Child and Family Social Work
Jenkins BQ, Tilbury C, Mazerolle P, Hayes H. 2017. The complexity of child protection
recurrence: The case for a systems approach. Child Abuse & Neglect 63: 162–171.
Kang H, Poertner J. 2006. Inter-rater reliability of the Illinois Structured Decision Support
Protocol. Child Abuse & Neglect 30: 679–689.
Keddell E. 2013. Beyond care versus control: Decision making discourses and their functions in
child protection social work (PhD thesis). University of Otago, Dunedin, New Zealand. http://
hdl.handle.net/10523/3886 [23 November 2018].
Keddell E. 2015. The ethics of predictive risk modelling in the Aotearoa/New Zealand child
welfare context: Child abuse prevention or neo-liberal tool? Critical Social Policy 35(1):
69–88.
Macchione N, Wooten W, Yphantides N, Howell JR. 2013. Integrated health and human services
information systems to enhance population-based and person-centered service. American
Journal of Preventative Medicine 45(3): 373–374.
MacFadden RJ, Schoech D. 2010. Neuroscience, the Unconscious and Professional Decision
Making: Implications for ICT. Journal of Technology in Human Services 28(4): 282–294.
Ministry of Social Development. 2014a. Final report on feasibility of using predictive risk
modelling. Ministry of Social Development: Wellington, New Zealand. Available: https://
www.msd.govt.nz/documents/about-msd-and-our-work/publications-resources/research/predi
ctive-modelling/00-feasibility-study-report.pdf [23 November 2018].
Ministry of Social Development. 2014b. The feasibility of using predictive risk modelling to
identify new-born children who are high priority for preventive services—companion technical
report. Ministry of Social Development: Wellington, New Zealand. Available: https://www.
msd.govt.nz/documents/about-msd-and-our-work/publications-resources/research/predictivemodelling/00-feasibility-study-report-technical-companion.pdf [23 November 2018].
Powell C. 2003. Early indicators of child abuse and neglect: a multi-professional Delphi study.
Child Abuse Review. 12(1): 25–40 https://doi.org/10.1002/car.778.
Reder P, Duncan S. 2004. Making the most of the Victoria Climbié Inquiry Report. Child Abuse
Review 13(2): 95–114. https://doi.org/10.1002/car.834.
Reder P, Duncan S, Gray M. 1993. Beyond Blame: Child Abuse Tragedies Revisited. Routledge:
London.
Russell J. 2015. Predictive analytics and child protection: Constraints and opportunities. Child
Abuse & Neglect 46: 182–189.
Salganik MJ. 2018. Bit by Bit: Social Research in the Digital Age. Princeton University Press:
Princeton, NJ.
Schoech D, Jennings H, Schkade LL, Hooper-Russell C. 1985. Expert Systems: Artificial
intelligence for professional decisions. Computers in Human Services 1(1): 81–115.
Schwalbe C. 2004. Re-visioning risk assessment for human service decision making. Children
and Youth Services Review 26: 561–576.
Schwartz E, Nowakowski-Sims E, Ramos-Hernandez A, York P. 2017. Predictive and
prescriptive analytics, machine learning and child welfare risk assessment: The Broward
County experience. Children and Youth Services Review 81: 309–320. https://doi.org/
10.1016/j.childyouth.2017.08.020.
Skotte PS. 2018. On caseworkers' writing in child welfare
when less is more. European
Journal of Social Work. https://doi.org/10.1080/13691457.2018.1469474
Stanley T. 2005. Making decisions: Social work processes and the construction of risk(s) in child
protection work (PhD thesis). University of Canterbury, New Zealand.
Vaithianathan R, Maloney T, Putnam-Hornstein E, Jiang N. 2013. Children in the public benefit
system at risk of maltreatment: Identification via predictive modelling. American Journal of
Preventative Medicine 45(3): 354–359.
Wijenayake S, Graham T, Christen P. 2018. A decision tree approach to predicting recidivism in
domestic violence presented at the Big Data Analytics for Social Computing (BDASC)
workshop held at the Pacific-Asia Conference on Knowledge Discovery and Data Mining
(PAKDD'18), Melbourne, Australia, June 2018. Available: https://arxiv.org/abs/1803.09862
[23 November 2018].
Zhang SX, Roberts REL, Farabee D. 2014. An Analysis of Prisoner Reentry and Parole Risk
Using COMPAS and Traditional Criminal History Measures. Crime and Delinquency 60(2):
167–192. https://doi.org/10.1177/0011128711426544.
© 2019 John Wiley & Sons, Ltd.
Child Abuse Rev. (2019)
DOI: 10.1002/car
Copyright of Child Abuse Review is the property of John Wiley & Sons, Inc. and its content
may not be copied or emailed to multiple sites or posted to a listserv without the copyright
holder's express written permission. However, users may print, download, or email articles for
individual use.
Download