for-phone-onlyfor-tablet-portrait-upfor-tablet-landscape-upfor-desktop-upfor-wide-desktop-up
Breakingviews

Breakingviews - UK exam U-turn exposes algorithms’ deep flaws

An A-level student holds a placard as she protests about the exam results at the constituency offices of Education Secretary Gavin Williamson, amid the spread of the coronavirus disease (COVID-19), in South Staffordshire, Britain, August 17, 2020.

NEW YORK (Reuters Breakingviews) - Government U-turns don’t come much bigger. Popular fury forced the abandonment of hypothetical calculations of likely grades for the UK’s national A-level exams this week. The decision is a reminder that even well-intentioned algorithms make many harmful mistakes.

The exams, which are critical to university admissions, were canceled thanks to the coronavirus lockdown. The government used an algorithm to estimate what would have happened. The furor erupted after almost 40% of students had their scores downgraded from the predictions of their teachers, with pupils from disadvantaged schools disproportionately affected.

It’s an unusually stark and public example of a far bigger problem. For instance, models that spit out credit scores in the United States and elsewhere can make it hard for less affluent folks to borrow at reasonable interest rates, in part because a decent credit score depends on already having a record of servicing debt.

Algorithms are also used in many U.S. states to help decide how likely a criminal is to offend again. Judges adjust sentencing decisions accordingly. These recidivism models are supposed to neutralize the prejudices of the judges. But their inputs, for example whether friends of the convict have been arrested, can introduce other biases, as former hedge-fund quant Cathy O’Neil describes in her book “Weapons of Math Destruction.”

O’Neil talks about so-called proxies, meaning inputs that substitute for knowledge of a person’s actual behavior. In the UK’s exam-grade algorithm, one proxy was the past record of a student’s school. That proved a sure way to discriminate against star students from less wealthy backgrounds.

As so often, the British model’s goal was rational: to adjust forecast grades to bring them, overall, closer to the distribution of results in prior years. The government of Prime Minister Boris Johnson and its exam regulator, Ofqual, had months to consider possible unintended consequences of their algorithm. Like careless teenagers, they wasted the time.

Stories of striving kids’ heartbreak make great television, and the protests of parents, teachers and a generation of soon-to-be-voters made predictably short work of the government’s initial refusal to budge. Teachers’ estimated grades will now stand. If only it were so relatively easy for, say, prisoners handed unfairly long sentences to make themselves heard.

Breakingviews

Reuters Breakingviews is the world's leading source of agenda-setting financial insight. As the Reuters brand for financial commentary, we dissect the big business and economic stories as they break around the world every day. A global team of about 30 correspondents in New York, London, Hong Kong and other major cities provides expert analysis in real time.


Sign up for a free trial of our full service at https://www.breakingviews.com/trial and follow us on Twitter @Breakingviews and at www.breakingviews.com. All opinions expressed are those of the authors.

for-phone-onlyfor-tablet-portrait-upfor-tablet-landscape-upfor-desktop-upfor-wide-desktop-up