Proponents of algorithms say computer programs can fix the inefficiency and bias of humans. Critics point to the opacity of algorithms and argue that they are not necessarily neutral.
“Technology changes things, but perhaps not always as much as we think,” says Angelé Christin, assistant professor of communication at Stanford University . “Social context matters a lot in shaping the actual effects of the technological tools. … So, it’s important to understand that connection between humans and machines.”
After all, advanced algorithms shape and guide our every step in the online world. They are also increasingly penetrating everyday life as more sectors of society, from finance and health care to human resources and criminal justice, incorporate them into daily decision-making.
In a recent research paper in Big Data & Society, Christin found a gap between the intended and actual uses of algorithms in the fields of web journalism and criminal justice. In addition, workers in both areas developed similar strategies to minimize the impact of algorithms on their daily work, she says.
“Whereas managers and executives frequently emphasize how ‘data-driven,’ modern, and rational their organization is, the actual uses of algorithmic techniques in web newsrooms and criminal courts reveal a more complicated picture,” Christin writes in the paper.
Journalists who opt out
Christin observed five newsrooms, three in Paris and two in New York, between 2011 and 2015. Since 2015, she has also been gathering information on four criminal courts, three in the United States and one in France. She interviewed reporters and editors as well as judges, defense attorneys, probation officers, and others.
In spite of the many differences between the fields of web journalism and criminal justice, Christin found surprising similarities between newsrooms and courtrooms during her fieldwork.
In web journalism, real-time analytics software programs, such as Chartbeat, are used to measure the number of visitors to each article, the average time spent by readers on each piece, and the number of likes and shares it gets on social media platforms.
These programs purport to revolutionize journalists’ content strategies by providing this detailed data about their audience. Editorial managers order the programs installed on every employee’s computer and encourage staff writers to track traffic on their stories.
But in some of the newsrooms she studied, staff writers and employees did not use these programs and some actively avoided them on principle, Christin says.
“I have access to something that shows all those stats,” one writer told Christin. “I don’t go there and obsessively look at the stats. I find it kind of stressful and I try not to get too wrapped up in that.”
Risk of reoffending
In criminal justice, a similar discrepancy exists. A lot of courts now use various predictive risk-assessment tools that rely on defendants’ criminal history, socio-demographics, and other variables to estimate an offender’s risk of reoffending or failure to appear in court when on bail.
“The problem of transparency is a real one, and I think there is no clear way to make algorithms completely transparent for now.”
These algorithmic programs were often introduced as a result of a growing awareness that racial bias exists at every step of the criminal justice process in the United States, Christin says. The supporters of these tools believe they can reduce overcrowding in jails by reliably identifying low-risk offenders who could be released.
But many judges and prosecutors do not use those analytics, Christin says. Because some of these tools are the creations of for-profit companies, many employees in criminal justice question their reliability.
“I’d prefer to look at actual behaviors,” one former prosecutor told Christin. “With these tools the output is only as good as the input. And the input is controversial.”
The transparency problem
The question of algorithms’ reliability hits at the core of today’s debate about them. Algorithms are inherently opaque; the code they’re written in is often proprietary and difficult to understand.
Some experts propose algorithmic audits that would test the system periodically. Others say a governmental agency should be created with some oversight power, Christin says.
These solutions may address some of the issues relating to algorithmic opacity, but not all, she says.
“The problem of transparency is a real one, and I think there is no clear way to make algorithms completely transparent for now,” Christin says.
But further qualitative research can help people understand how algorithms fit in the social system and how they can make sense of them, Christin says. In addition to conducting research on the uses and practices associated with algorithmic systems, humanists and social science scholars should encourage those who make algorithms to critically analyze them.
“People who design these tools do not always follow closely how they are being used in specific organizations,” Christin says. “Sometimes people use technology in ways you want them, but sometimes they use it differently.”
Source: Stanford University