Instagram is Inspecting its Algorithm and Policies for Bias Against Black Users
With the Black Lives Matter movement in focus, a lot of tech companies are reviewing their policies to make sure black voices don’t get drowned out. Yesterday, Instagram‘s CEO, Adam Mosseri, said the company is inspecting its recommendation algorithm and policy to determine if it treats everyone equally. Mosseri talked about ‘shadowbans,’ harassment, and content takedown disagreements of black creators. He said the company will focus on four areas to improve its services: harassment, account verification, content distribution, and algorithmic bias. 2/ Addressing the feedback we get has always been an integral part of how we work, and has helped us build a better Instagram for everyone. We’re going to focus on four areas: • Harassment• Account verification• Content distribution• Algorithmic bias — Adam Mosseri 😷 (@mosseri) June 15, 2020 Out of these four, algorithmic bias and content distribution go hand in hand, as people spend most of their time on their feed or the explore tab. Mosseri said that the company has heard a lot about ‘shadowbanning’ — a practice where the platform limits the reach of the certain posts — and it’ll soon release information on what kind of content they downrank. Credit: FacebookInstagram CEO Adam Mosseri Instagram will also examine if its algorithms have some subconscious bias, but didn’t provide details on what it aims to do to find and correct it. Plus, it will look into safety issues black people face on the platform. The photo-sharing platform’s CEO said that Instagram inalso looking for ways to better serve other underrepresented groups on the app: These efforts won’t stop with the disparities people may experience solely on the basis of race; we’re also going to look at how we can better serve other underrepresented groups that use our product. In the last year alone the feedback we’ve received from communities like LGBTQ groups, body positivity activists, and artists has helped us build a more inclusive product. Ideally, Instagram should ask outside academics to study their algorithms and policies to find the bias. Plus, it needs to include marginalized communities in the discussion to solve these problems. Bias against underrepresented groups is not something that can be solved in a single day. It’ll be a long process. Read next: Tesla Model S can finally drive 400 miles between charges — but only in North America Celebrate Pride 2020 with us this month! Why is queer representation so important? What’s it like being trans in tech? How do I participate virtually? You can find all our Pride 2020 coverage here. » Read More
Like to keep reading?
This article first appeared on thenextweb.com. If you'd like to keep reading, follow the white rabbit.