Skip to main content

IBM will no longer develop or research facial recognition tech

IBM CEO Arvind Krishna says the company will no longer develop or offer general-purpose facial recognition or analysis software. In a June 8 letter addressed to Congress and written in support of the Justice in Policing Act of 2020, Krishna advocates for new reforms that support the responsible use of technology — and combat systematic racial injustice and police misconduct.

Recommended Videos

“IBM firmly opposes and will not condone uses of any technology, including facial recognition technology offered by other vendors, for mass surveillance, racial profiling, violations of basic human rights and freedoms, or any purpose which is not consistent with our values and Principles of Trust and Transparency,” wrote Krishna in the letter.

Krishna, who took over the chief executive role in April, added that it’s time for Congress to begin a national dialogue on the implications of facial recognition technology and how it “should be employed by domestic law enforcement agencies.”

The CEO also voiced his concerns regarding racial bias that is often found in artificial intelligence systems today. Krishna further called the need for more oversight to audit artificial intelligence tools, especially when they’re used in law enforcement and national policies that “bring greater transparency and accountability to policing, such as body cameras and modern data analytics techniques.”

People familiar with the matter told CNBC that the death of George Floyd, a Black man, while in the custody of Minneapolis police and the attendant focus on police reform and racial inequity convinced IBM to shut down its facial recognition products.

Over the last few years, facial recognition systems have dramatically advanced thanks to developments in fields such as machine learning. However, without any official oversight in place, they’ve been largely allowed to run unregulated and violate user privacy. Most notably, facial recognition tech was brought to the forefront of the national conversation by a startup called Clearview AI that was able to build a database of more than 3 billion images primarily scraping social media sites. Clearview has since faced a backlash from companies such as Twitter and is currently dealing with a myriad of privacy lawsuits.

Clearview AI is also reportedly being employed by law enforcement agencies in the ongoing Black Lives Matter protests across the U.S. Experts have argued that these systems can misidentify people, as they’re largely trained using white male faces.

Krishna didn’t say whether the company would reconsider its decision if and when Congress introduces new laws to bring more scrutiny to technology such as facial recognition. We’ve reached out to IBM and will update this story when we hear back.

Shubham Agarwal
Former Digital Trends Contributor
Shubham Agarwal is a freelance technology journalist from Ahmedabad, India. His work has previously appeared in Firstpost…
Google missed big chance with ChatGPT-like tech, report claims
Google Logo

Google missed a golden opportunity to lead the way with its own ChatGPT-like chatbot technology tool two years ago, but an overly cautious attitude from those at the top prevented the company from releasing it, according to a Wall Street Journal report on Tuesday.

The two Google researchers who created the powerful conversational AI technology reportedly told colleagues at the time that their creation could revolutionize how people searched on the internet and worked with computers.

Read more
Microsoft quits its creepy, emotion-reading A.I.
blonde woman with an expressionless face looks at camera while laser lights scan her features

Microsoft announced it will stop the development and distribution of controversial emotion-reading software as big tech companies pivot toward privacy and security. The company also says it will heavily restrict its own facial recognition platform.

Microsoft’s shift away from emotional recognition software is another sign of big tech’s growing prioritization of privacy. The company also admits there is little scientific evidence behind the technology.

Read more
Zoom’s A.I. tech to detect emotion during calls upsets critics
coronavirus crisis not ready for an online first world analysis zoom conference lifestyle image

Zoom has begun to develop A.I. technology which can reportedly scan the faces and speech of users in order to determine their emotions, which was first reported by Protocol.

While this technology appears to still be in its early phases of development and implementation, several human rights groups project that it could be used for more discriminatory purposes down the line, and are urging Zoom to turn away from the practice.

Read more