Search

AI researchers say scientific publishers help perpetuate racist algorithms - MIT Technology Review

jemputjembut.blogspot.com

The news: An open letter from a growing coalition of AI researchers is calling out scientific publisher Springer Nature for a conference paper it originally planned to include in its forthcoming book Transactions on Computational Science & Computational Intelligence. The paper, titled “A Deep Neural Network Model to Predict Criminality Using Image Processing,” presents a face recognition system purportedly capable of predicting whether someone is a criminal, according to the original press release. It was developed by researchers at Harrisburg University and was due to be presented at a forthcoming conference.

The demands: Citing the work of leading Black AI scholars, the letter debunks the scientific basis of the paper and asserts the racism of crime-prediction technologies. It also lists three demands: (1) for Springer Nature to rescind its offer to publish the study; (2) for it to issue a statement condemning the use of statistics, such as machine learning, to predict criminality and acknowledging its role in incentivizing such research; and (3) for all scientific publishers to commit to not publishing similar papers in the future. The letter, which was sent to Springer Nature on Monday, was originally written by five researchers at MIT, Rensselaer Polytechnic Institute, McGill University, and AI Now Institute. In a matter of days, it gained more than 600 signatures and counting across the AI ethics and academic communities, including from leading figures like Meredith Whittaker, co-founder of the AI Now Institute, and Ethan Zuckerman, former director of the Center for Civic Media at the MIT Media Lab.

Why it matters: While the letter highlights a specific paper, the authors’ goal is to demonstrate a systematic issue with the way scientific publishing incentivizes researchers to perpetuate unethical norms. “This is why we keep seeing race science emerging time and again,” said Chelsea Barabas, a research scientist at MIT and one of the letter’s co-authors. “It’s because publishers publish it.” “The real significance of this Springer piece is that it’s not unique whatsoever,” echoed Theodora Dryer, a postdoctoral researcher at AI Now and another co-author. “It’s emblematic of a problem and a critique that has gone on for so, so long.” 

Springer’s response: In response to the letter, Springer said that it would no longer be publishing the paper. “The paper you are referring to was submitted to a forthcoming conference for which Springer had planned to publish the proceedings,” it said. “After a thorough peer review process the paper was rejected.” Harrisburg University also took down its press release, stating “the faculty are updating the paper to address concerns raised.” The letter’s signatories said they will continue to push for a fulfillment of their second and third demands. Springer did not respond to MIT Technology Review’s independent request for comment. Harrisburg University and the co-author of the paper also denied a request for comment as well as a request for a copy of the original paper.

The bigger picture: Since George Floyd’s death sparked an international movement for racial justice, the AI field and tech industry at large has faced a reckoning in the role it has played in reinforcing structural racism. During the week of June 8th, for example, IBM, Microsoft, and Amazon all announced the end or a partial suspension of their face recognition products. The move was a culmination of two years of advocacy from researchers and activists to demonstrate a link between these technologies and the over-policing of minority communities. The open letter is the latest development in this growing movement to shift the community’s norms to a greater level of ethical accountability.

“We really wanted to contribute to this growing movement,” said Sonja Solomun, the research director of the Centre for Media, Technology and Democracy at McGill University. “Particularly when we look outside our windows and see what’s going on right now in the US and globally, the stakes are just so high.”

Let's block ads! (Why?)



"help" - Google News
June 23, 2020 at 05:41PM
https://ift.tt/316v3kO

AI researchers say scientific publishers help perpetuate racist algorithms - MIT Technology Review
"help" - Google News
https://ift.tt/2SmRddm


Bagikan Berita Ini

0 Response to "AI researchers say scientific publishers help perpetuate racist algorithms - MIT Technology Review"

Post a Comment

Powered by Blogger.