References
[1] J Khadijah Abdurahman. 2021. Calculating the Souls of Black Folk: Predictive Analytics in the New York City Administration for Children鈥檚 Services. In Colum. J. Race & L. Forum, Vol. 11. HeinOnline, 75.
[2] J. Khadijah Abdurahman. 2022. Birthing Predictions of Premature Death. (2022). https://logicmag.io/home/birthing-predictions-of-premature-death/
[3] Social Security Administration. 2023. Supplemental Security Income (SSI) in Pennsylvania. https://www.ssa.gov/pubs/EN-05-11150.pdf
[4] Patricia Auspos. 2017. Using Integrated Data Systems to Improve Case Management and Develop Predictive Modeling Tools. Case Study 4. Annie E. Casey Foundation (2017).
[5] Cora Bartelink, TA Van Yperen, IJ Ten Berge, Leontien De Kwaadsteniet, and CLM Witteman. 2014. Agreement on child maltreatment decisions: A nonrandomized study on the effects of structured decision-making. In Child & Youth Care Forum, Vol. 43. Springer, 639鈥654.
[6] Cora Bartelink, Tom A Van Yperen, and J Ingrid. 2015. Deciding on child maltreatment: A literature review on methods that improve decision-making. Child Abuse & Neglect 49 (2015), 142鈥153.
[7] Elinor Benami, Reid Whitaker, Vincent La, Hongjin Lin, Brandon R Anderson, and Daniel E Ho. 2021. The distributive effects of risk prediction in environmental compliance: Algorithmic design, environmental justice, and public policy. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency. 90鈥105.
[8] Emily M Bender and Batya Friedman. 2018. Data statements for natural language processing: Toward mitigating system bias and enabling better science. Transactions of the Association for Computational Linguistics 6 (2018), 587鈥604.
[9] Umang Bhatt, Alice Xiang, Shubham Sharma, Adrian Weller, Ankur Taly, Yunhan Jia, Joydeep Ghosh, Ruchir Puri, Jos茅 MF Moura, and Peter Eckersley. 2020. Explainable machine learning in deployment. In Proceedings of the 2020 conference on fairness, accountability, and transparency. 648鈥657.
[10] Emily Black, Manish Raghavan, and Solon Barocas. 2022. Model Multiplicity: Opportunities, Concerns, and Solutions. In 2022 ACM Conference on Fairness, Accountability, and Transparency (Seoul, Republic of Korea) (FAccT 鈥22). Association for Computing Machinery, New York, NY, USA, 850鈥863. https://doi.org/10.1145/3531146.3533149
[11] Sebastian Bordt, Mich猫le Finck, Eric Raidl, and Ulrike von Luxburg. 2022. Post-hoc explanations fail to achieve their purpose in adversarial contexts. arXiv preprint arXiv:2201.10295 (2022).
[12] Anna Brown, Alexandra Chouldechova, Emily Putnam-Hornstein, Andrew Tobin, and Rhema Vaithianathan. 2019. Toward algorithmic accountability in public services: A qualitative study of affected community perspectives on algorithmic decision-making in child welfare services. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1鈥12.
[13] Carrie J Cai, Samantha Winter, David Steiner, Lauren Wilcox, and Michael Terry. 2019. 鈥淗ello AI鈥: uncovering the onboarding needs of medical practitioners for human-AI collaborative decision-making. Proceedings of the ACM on Human-computer Interaction 3, CSCW (2019), 1鈥24.
[14] Girish Chandrashekar and Ferat Sahin. 2014. A survey on feature selection methods. Computers & Electrical Engineering 40, 1 (2014), 16鈥28.
[15] Hao-Fei Cheng, Logan Stapleton, Anna Kawakami, Venkatesh Sivaraman, Yanghuidi Cheng, Diana Qing, Adam Perer, Kenneth Holstein, Zhiwei Steven Wu, and Haiyi Zhu. 2022. How child welfare workers reduce racial disparities in algorithmic decisions. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems. 1鈥22.
[16] Alexandra Chouldechova, Diana Benavides-Prado, Oleksandr Fialko, and Rhema Vaithianathan. 2018. A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. In Conference on Fairness, Accountability and Transparency. PMLR, 134鈥148.
[17] Sasha Costanza-Chock. 2020. Design justice: Community-led practices to build the worlds we need. The MIT Press.
[18] Allegheny County. [n. d.]. The DHS Data Warehouse. ([n. d.]). https://www.alleghenycounty.us/human-services/news-events/accomplishments/dhsdata-warehouse.aspx
[19] Ying Cui, Fu Chen, Ali Shiri, and Yaqin Fan. 2019. Predictive analytic models of student success in higher education: A review of methodology. Information and Learning Sciences (2019).
[20] Tim Dare and Eileen Gambrill. 2017. Ethical analysis: Predictive risk models at call screening for Allegheny County. Unpublished report (2017). https:// www.alleghenycountyanalytics.us/wp-content/uploads/2019/05/Ethical-Analysis-16-ACDHS-26_PredictiveRisk_Package_050119_FINAL-2.pdf
[21] Maria De-Arteaga, Riccardo Fogliato, and Alexandra Chouldechova. 2020. A case for humans-in-the-loop: Decisions in the presence of erroneous algorithmic scores. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1鈥12.
[22] Jessica M Eaglin. 2017. Constructing recidivism risk. Emory LJ 67 (2017), 59.
[23] Virginia Eubanks. 2018. Automating inequality: How high-tech tools profile, police, and punish the poor. St. Martin鈥檚 Press.
[24] Mary Flanagan, Daniel C. Howe, and Helen Nissenbaum. 2008. Embodying Values in Technology: Theory and Practice. Cambridge University Press, 322鈥353. https://doi.org/10.1017/CBO9780511498725.017
[25] Batya Friedman. 1996. Value-sensitive design. Interactions 3, 6 (1996), 16鈥23.
[26] Philip Gillingham. 2019. Can predictive algorithms assist decision-making in social work with children and families? Child abuse review 28, 2 (2019), 114鈥126.
[27] Stephanie K Glaberson. 2019. Coding over the cracks: predictive analytics and child protection. Fordham Urb. LJ 46 (2019), 307.
[28] James P Gleeson. 1987. Implementing structured decision-making procedures at child welfare intake. Child Welfare (1987), 101鈥112.
[29] Lauryn P Gouldin. 2016. Disentangling flight risk from dangerousness. BYU L. Rev. (2016), 837.
[30] Crystal Grant. 2022. 老澳门开奖结果 White Paper: AI in Healthcare May Worsen Medical Racism. /legal-document/aclu-white-paperai-health-care-may-worsen-medical-racism
[31] Ben Green. 2021. Data science as political action: Grounding data science in a politics of justice. Journal of Social Computing 2, 3 (2021), 249鈥265.
[32] Ben Green. 2022. Escaping the impossibility of fairness: From formal to substantive algorithmic fairness. Philosophy & Technology 35, 4 (2022), 1鈥32.
[33] Sam Harper, Nicholas B King, Stephen C Meersman, Marsha E Reichman, Nancy Breen, and John Lynch. 2010. Implicit value judgments in the measurement of health inequalities. The Milbank Quarterly 88, 1 (2010), 4鈥29.
[34] Sally Ho and Garance Burke. 2022. An algorithm that screens for child neglect in Allegheny County raises concerns. (2022). https://apnews.com/ article/child-welfare-algorithm-investigation-9497ee937e0053ad4144a86c68241ef1
[35] Benefits Tech Action Hub. 2022. Colorado Medicaid, SNAP, CHIP, and TANF Wrongful Denials. https://www.btah.org/case-study/colorado-medicaidsnap-chip-and-tanf-wrongful-denials.html
[36] Hornby Zeller Associates Inc. 2018. Allegheny County Predictive Risk Modeling Tool Implementation: Process Evaluation. (2018). https://www.alleghenycountyanalytics.us/wp-content/uploads/2019/05/Process-Evaluation-from-16-ACDHS-26_PredictiveRisk_Package_ 050119_FINAL-4.pdf
[37] Abigail Z Jacobs. 2021. Measurement as governance in and for responsible AI. arXiv preprint arXiv:2109.05658 (2021).
[38] Abigail Z Jacobs and Hanna Wallach. 2021. Measurement and fairness. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency. 375鈥385.
[39] Will Johnson. 2004. Effectiveness of California鈥檚 child welfare structured decision making (SDM) model: a prospective study of the validity of the California Family Risk Assessment. Madison (Wisconsin, USA): Children鈥檚 Research Center (2004).
[40] Nathan Kallus and Angela Zhou. 2019. The fairness of risk scores beyond classification: Bipartite ranking and the xauc metric. Advances in neural information processing systems 32 (2019).
[41] Amir-Hossein Karimi, Gilles Barthe, Bernhard Sch枚lkopf, and Isabel Valera. 2022. A Survey of Algorithmic Recourse:Contrastive Explanations and Consequential Recommendations. ACM Comput. Surv. (2022). https://doi.org/10.1145/3527848
[42] Anna Kawakami, Venkatesh Sivaraman, Hao-Fei Cheng, Logan Stapleton, Yanghuidi Cheng, Diana Qing, Adam Perer, Zhiwei Steven Wu, Haiyi Zhu, and Kenneth Holstein. 2022. Improving Human-AI Partnerships in Child Welfare: Understanding Worker Practices, Challenges, and Desires for Algorithmic Decision Support. In CHI Conference on Human Factors in Computing Systems. 1鈥18.
[43] Anna Kawakami, Venkatesh Sivaraman, Logan Stapleton, Hao-Fei Cheng, Adam Perer, Zhiwei Steven Wu, Haiyi Zhu, and Kenneth Holstein. 2022. 鈥淲hy Do I Care What鈥檚 Similar?鈥 Probing Challenges in AI-Assisted Child Welfare Decision-Making through Worker-AI Interface Design Concepts. In Designing Interactive Systems Conference. 454鈥470.
[44] Emily Keddell. 2015. The ethics of predictive risk modelling in the Aotearoa/New Zealand child welfare context: Child abuse prevention or neo-liberal tool? Critical Social Policy 35, 1 (2015), 69鈥88.
[45] Emily Keddell. 2019. Algorithmic justice in child protection: Statistical fairness, social justice and the implications for practice. Social Sciences 8, 10 (2019), 281.
[46] Kenji Kira and Larry A. Rendell. 1992. A Practical Approach to Feature Selection. In Machine Learning Proceedings 1992. San Francisco (CA), 249鈥256. https://doi.org/10.1016/B978-1-55860-247-2.50037-1
[47] Vipin Kumar and Sonajharia Minz. 2014. Feature Selection: A literature Review. Smart Comput. Rev. 4 (2014), 211鈥229.
[48] Himabindu Lakkaraju. 2021. Towards Reliable and Practicable Algorithmic Recourse. In Proceedings of the 30th ACM International Conference on Information amp; Knowledge Management (Virtual Event, Queensland, Australia) (CIKM 鈥21). Association for Computing Machinery, New York, NY, USA, 4. https://doi.org/10.1145/3459637.3482497
[49] Algorithmic Justice League. [n. d.]. The Algorithmic Justice League鈥檚 101 Overview. ([n. d.]). https://www.ajl.org/learn-more
[50] Karen Levy, Kyla E Chasalow, and Sarah Riley. 2021. Algorithms and decision-making in the public sector. Annual Review of Law and Social Science 17 (2021), 309鈥334.
[51] Gabriel Lima, Nina Grgi膰-Hla膷a, Jin Keun Jeong, and Meeyoung Cha. 2022. The Conflict Between Explainable and Accountable Decision-Making Algorithms. arXiv preprint arXiv:2205.05306 (2022).
[52] Jorge M. Lobo, Alberto Jim茅nez-Valverde, and Raimundo Real. 2008. AUC: a misleading measure of the performance of predictive distribution models. Global Ecology and Biogeography 17, 2 (2008), 145鈥151. https://doi.org/10.1111/j.1466-8238.2007.00358.x arXiv:https://onlinelibrary.wiley.com/doi/pdf/10.1111/j.1466-8238.2007.00358.x
[53] Wayne A Logan and Andrew Guthrie Ferguson. 2016. Policing criminal justice data. Minn. L. Rev. 101 (2016), 541.
[54] No毛mi Manders-Huits. 2011. What values in design? The challenge of incorporating moral values into design. Science and engineering ethics 17, 2 (2011), 271鈥287.
[55] Sandra G Mayson. 2017. Dangerous defendants. Yale LJ 127 (2017), 490.
[56] Mikaela Meyer, Aaron Horowitz, Erica Marshall, and Kristian Lum. 2022. Flipping the Script on Criminal Justice Risk Assessment: An actuarial model for assessing the risk the federal sentencing system poses to defendants. In 2022 ACM Conference on Fairness, Accountability, and Transparency. 366鈥378.
[57] Margaret Mitchell, Simone Wu, Andrew Zaldivar, Parker Barnes, Lucy Vasserman, Ben Hutchinson, Elena Spitzer, Inioluwa Deborah Raji, and Timnit Gebru. 2019. Model cards for model reporting. In Proceedings of the conference on fairness, accountability, and transparency. 220鈥229.
[58] Shira Mitchell, Eric Potash, Solon Barocas, Alexander D鈥橝mour, and Kristian Lum. 2021. Algorithmic fairness: Choices, assumptions, and definitions. Annual Review of Statistics and Its Application 8 (2021), 141鈥163.
[59] Shakir Mohamed, Marie-Therese Png, and William Isaac. 2020. Decolonial AI: Decolonial theory as sociotechnical foresight in artificial intelligence. Philosophy & Technology 33, 4 (2020), 659鈥684.
[60] Michael Muller, Ingrid Lange, Dakuo Wang, David Piorkowski, Jason Tsay, Q Vera Liao, Casey Dugan, and Thomas Erickson. 2019. How data science workers work with data: Discovery, capture, curation, design, creation. In Proceedings of the 2019 CHI conference on human factors in computing systems. 1鈥15.
[61] Michael Muller and Angelika Strohmayer. 2022. Forgetting Practices in the Data Sciences. In CHI Conference on Human Factors in Computing Systems. 1鈥19.
[62] Deirdre K Mulligan and Kenneth A Bamberger. 2018. Saving governance-by-design. California Law Review 106, 3 (2018), 697鈥784.
[63] Deirdre K Mulligan and Kenneth A Bamberger. 2019. Procurement as policy: Administrative process for machine learning. Berkeley Tech. LJ 34 (2019), 773.
[64] Debbie Nathan. 2020. The Long, Dark History of Family Separations: How politicians used the drug war and the welfare state to break up black and Native American families. (2020). https://reason.com/2020/10/13/the-long-dark-history-of-family-separations/
[65] Hina Naveed. 2022. 鈥淚f I Wasn鈥檛 Poor, I Wouldn鈥檛 Be Unfit" The Family Separation Crisis in the US Child Welfare System. (2022). https: //www.aclu.org/report/if-i-wasnt-poor-i-wouldnt-be-unfit-family-separation-crisis-us-child-welfare-system
[66] Children鈥檚 Data Network. 2022. Los Angeles County Risk Stratification Model: Methodology Implementation Report. (2022). https://dcfs.lacounty. gov/wp-content/uploads/2022/08/Risk-Stratification-Methodology-Report_8.29.22.pdf
[67] Allegheny County Department of Human Services. 2017. Ethical Analysis: Predictive Risk Models at Call Screening for Allegheny County Response by the Allegheny County Department of Human Services. Alleghany County Analytics (2017).
[68] Allegheny County Department of Human Services. 2019. Developing Predictive Risk Models to Support Child Maltreatment Hotline Screening Decisions: Predictive Risk Package. https://www.alleghenycountyanalytics.us/wp-content/uploads/2019/05/16-ACDHS-26_PredictiveRisk_ Package_050119_FINAL-2.pdf
[69] Mark A Paige and Audrey Amrein-Beardsley. 2020. 鈥淗ouston, We Have a Lawsuit鈥: A Cautionary Tale for the Implementation of Value-Added Models for High-Stakes Employment Decisions. Educational Researcher 49, 5 (2020), 350鈥359.
[70] Amandalynne Paullada, Inioluwa Deborah Raji, Emily M Bender, Emily Denton, and Alex Hanna. 2021. Data and its (dis) contents: A survey of dataset development and use in machine learning research. Patterns 2, 11 (2021), 100336.
[71] Martin Pawelczyk, Teresa Datta, Johannes van-den Heuvel, Gjergji Kasneci, and Himabindu Lakkaraju. 2022. Probabilistically Robust Recourse: Navigating the Trade-offs between Costs and Robustness in Algorithmic Recourse. https://doi.org/10.48550/ARXIV.2203.06768
[72] Tawana Petty, Mariella Saba, Tamika Lewis, Seeta Pena Gangadharan, and Virginia Eubanks. 2018. Reclaiming our data. (2018). https://www. odbproject.org/wp-content/uploads/2016/12/ODB.InterimReport.FINAL_.7.16.2018.pdf
[73] Eleanor Pratt and Heather Hahn. 2021. What Happens When People Face Unfair Treatment or Judgment When Applying for Public Assistance or Social Services? Washington, DC: Urban Institute (2021).
[74] Kaivalya Rawal, Ece Kamar, and Himabindu Lakkaraju. 2020. Can I Still Trust You?: Understanding the Impact of Distribution Shifts on Algorithmic Recourses. CoRR abs/2012.11788 (2020). arXiv:2012.11788 https://arxiv.org/abs/2012.11788
[75] Rashida Richardson, Jason M Schultz, and Kate Crawford. 2019. Dirty data, bad predictions: How civil rights violations impact police data, predictive policing systems, and justice. NYUL Rev. Online 94 (2019), 15.
[76] Putnam-Hornstein Emily Vaithianathan Rhema Rittenhouse, Katherine. 2022. Algorithms, Humans, and Racial Disparities in Child Protective Services: Evidence from the Allegheny Family Screening Tool. (2022).
[77] Dorothy Roberts. 2009. Shattered bonds: The color of child welfare. Hachette UK.
[78] Dorothy Roberts. 2022. Torn Apart: How the Child Welfare System Destroys Black Families鈥揳nd How Abolition Can Build a Safer World. Basic Books.
[79] Samantha Robertson, Tonya Nguyen, and Niloufar Salehi. 2021. Modeling Assumptions Clash with the Real World: Transparency, Equity, and Community Challenges for Student Assignment Algorithms. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI 鈥21). Association for Computing Machinery, New York, NY, USA, Article 589, 14 pages. https://doi.org/10.1145/3411764. 3445748
[80] David G. Robinson. 2022. The Kidney Transplant Algorithm鈥檚 Surprising Lessons for Ethical A.I. (2022). https://slate.com/technology/2022/08/kidneyallocation-algorithm-ai-ethics.html
[81] David G Robinson. 2022. Voices in the Code: A Story about People, Their Values, and the Algorithm They Made. Russell Sage Foundation.
[82] Anjana Samant, Aaron Horowitz, Kath Xu, and Sophie Beiers. 2022. Family Surveillance by Algorithm: The Rapidly Spreading Tools Few Have Heard Of. (2022). /fact-sheet/family-surveillance-algorithm
[83] Devansh Saxena, Karla Badillo-Urquiola, Pamela J Wisniewski, and Shion Guha. 2020. A human-centered review of algorithms used within the US child welfare system. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1鈥15.
[84] Andrew D Selbst, Danah Boyd, Sorelle A Friedler, Suresh Venkatasubramanian, and Janet Vertesi. 2019. Fairness and abstraction in sociotechnical systems. In Proceedings of the conference on fairness, accountability, and transparency. 59鈥68.
[85] Noah Simon, Jerome Friedman, Trevor Hastie, and Rob Tibshirani. 2011. Regularization Paths for Cox鈥檚 Proportional Hazards Model via Coordinate Descent. Journal of Statistical Software 39, 5 (2011), 1鈥13. https://www.jstatsoft.org/v39/i05/
[86] Logan Stapleton, Min Hun Lee, Diana Qing, Marya Wright, Alexandra Chouldechova, Ken Holstein, Zhiwei Steven Wu, and Haiyi Zhu. 2022. Imagining New Futures beyond Predictive Systems in Child Welfare: A Qualitative Study with Impacted Stakeholders. In 2022 ACM Conference on Fairness, Accountability, and Transparency (Seoul, Republic of Korea) (FAccT 鈥22). Association for Computing Machinery, New York, NY, USA, 1162鈥1177. https://doi.org/10.1145/3531146.3533177
[87] Harini Suresh and John Guttag. 2021. Understanding Potential Sources of Harm throughout the Machine Learning Life Cycle. (2021).
[88] upEND Movement. [n. d.]. Family Policing System Definition. ([n. d.]). https://upendmovement.org/family-policing-definition/
[89] Rhema Vaithianathan, Haley Dinh, Allon Kalisher, Chamari Kithulgoda, Emily Kulick, Megh Mayur, Athena Ning, Diana Benavides Prado, and Emily Putnam-Hornstein. 2019. Implementing a Child Welfare Decision Aide in Douglas County: Methodology Report. https://csda.aut.ac.nz/__ data/assets/pdf_file/0009/347715/Douglas-County-Methodology_Final_3_02_2020.pdf
[90] Rhema Vaithianathan, Emily Kulick, Emily Putnam-Hornstein, and D Benavides-Prado. 2019. Allegheny family screening tool: Methodology, version 2. Center for Social Data Analytics (2019), 1鈥22.
[91] Rhema Vaithianathan, Emily Putnam-Hornstein, Nan Jiang, Parma Nand, and Tim Maloney. 2017. Developing predictive models to support child maltreatment hotline screening decisions: Allegheny County methodology and implementation. Center for Social data Analytics (2017).
[92] Suresh Venkatasubramanian and Mark Alfano. 2020. The Philosophical Basis of Algorithmic Recourse. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (Barcelona, Spain) (FAT* 鈥20). Association for Computing Machinery, New York, NY, USA, 284鈥293. https://doi.org/10.1145/3351095.3372876
[93] Emma Williams. 2020. 鈥楩amily Regulation,鈥 Not 鈥楥hild Welfare鈥: Abolition Starts with Changing our Language. (2020). https://imprintnews.org/ opinion/family-regulation-not-child-welfare-abolition-starts-changing-language
[94] Qian Yang, Aaron Steinfeld, and John Zimmerman. 2019. Unremarkable ai: Fitting intelligent decision support into critical, clinical decision-making processes. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1鈥11.
[95] Allegheny County. [n. d.]. Office of Behavioral Health. https://www.alleghenycounty.us/Human-Services/老澳门开奖结果/Offices/Behavioral-Health.aspx