Case Study: Tackling Robodebt

Contents

Case Study: Tackling Robodebt#

This illustration shows a big machine that says 'you owe us'. It has a few people automating debt call, which is impacting the low-income community - there is person on wheelchair, a tired looking mother with an infant and a toddler and a distressed man standing next to her.

Fig. 158 Illustration of automated decision-making disproportionately affecting low-income communities. The Turing Way project illustration by Scriberia. Used under a CC-BY 4.0 licence. DOI: 10.5281/zenodo.3332807.#

The Australian Online Compliance Initiative (OCI - dubbed the ‘robodebt’ debacle by media) involved a large-scale data matching project between the Federal Tax Office and the Department of Human Services. The automated system sending out debt notices based on this data matching exercise was set up at scale quite quickly resulting in disproportionate impact on those with low socioeconomic status.

Some of the attempts at resisting these automated debt calls, very cleverly imitated the automated architecture of the Online Compliance Initiative. Activist organizations collaborated with legal aid groups to set up formal appeals processes. GetUp! created an automated appeal template through a campaign called FraudStop where someone on welfare benefit can challenge the debt claim raised against them by entering a few details, explaining why they want to appeal, and hitting send. The system compiles a formal request for review, prints it off, and sends it by post to Centrelink, the frontline social services delivery agency. It also:

  • Asks their local MP for assistance — as this could really speed up the process.

  • Sends a complaint to the Human Services Minister — responsible for this automated debt letter debacle.

  • Submits a Freedom of Information request for the individual’s Centrelink file — so that they have all the necessary information to fight back.

  • The campaign is archived here: link and here link

Further Reading#

A September 2020 report by Michele Gilman at the Data and Society Research Institute draws on existing public law frameworks to provide guidance to lawyers on how to challenge automated decision making that disproportionately harms low income communities.

POVERTYLAWGORITHMS: A Poverty Lawyer’s Guide to Fighting Automated Decision-Making Harms on Low-Income Communities Refers to rights from Consumer Law, Family Law, Housing, Public Benefits, Schools and Education, Workers’ Rights, Immigration Surveillance link