IBM used database of images to build new facial recognition software, reportedly without consent
Marie Donlon | March 13, 2019A photo from IBM's Diversity in Faces dataset. Source: IBMAccording to reports, IBM trained its facial recognition systems on millions of online images unbeknownst to the people appearing in those images.
The images, which were not originally collected by the tech giant but instead by former Flickr owner Yahoo, were taken with the intention of building an image repository for research purposes. Called YFCC100M, the database is home to roughly 99 million photos from Flickr containing images of people who had consented to have their pictures taken but who did not know their images would eventually be used to train facial recognition algorithms.
Made available under Creative Commons licenses and approved by IBM’s legal team, the company is reportedly using the images to help build an unbiased facial recognition dataset called Diversity in Faces amid recent reports that the technology only accurately works on the faces of white people. IBM hopes that using the diverse images from the database will battle the bias that can make artificial intelligence algorithms unfair.
While it will be a challenge for people to determine if their photos were used to train the facial recognition technology algorithms as IBM keeps dataset details private, the company does assure that only academic researchers have access to the images.
"We take the privacy of individuals very seriously and have taken great care to comply with privacy principles, including limiting the Diversity in Faces dataset to publicly available image annotations and limiting the access of the dataset to verified researchers. Individuals can opt-out of this dataset," spokesman Saswato Das said in a statement. "IBM has been committed to building responsible, fair and trusted technologies for more than a century and believes it is critical to strive for fairness and accuracy in facial recognition."
At the heart of most controversies surrounding the use of facial recognition technology is that the technology could potentially be used to unfairly target specific individuals including immigrants, people of color and religious minorities. To prove that point, the American Civil Liberties Union (ACLU) compared images of U.S. Congress members against a database of public mug shots using Amazon’s facial recognition tool, Rekognition. The result, according to the ACLU, was that 28 members of U.S. Congress were falsely identified as criminal suspects, 40% of whom were African-American. Such reports prompted a coalition of advocacy groups and activists to write open letters to tech giants Google, Amazon and Microsoft, imploring them to not sell their respective facial recognition technologies to government authorities.
'only accurately works on the faces of white people'. Well this is a new one for all those racist folks. Obviously the whites are more brain dead and don't take their privacy very seriously. Lesson to be learned here from the other folks, Good for them.
In stating that IBM takes the privacy of people seriously, why are they issuing the information to other companies and hoping they will keep it private. This is a gross invasion of people privacy for the assumed 'greater good' of God knows what. But yet so typical of the now pathetic 1st world people and the companies who deem themselves to rule our lives. Far too much liberties are allowed by companies and governments these days. Far too much.
Just to add, '28 members of U.S. Congress were falsely identified as criminal suspects'. Maybe it did work, given the controversies that are taking place these days in the US government.

It goes to show what I've always said - if there is artificial intelligence, I suggest there must also be artificial stupidity.
IBM states availability of the facial dataset to "verified researchers" only. Does such access "verification" include background checks of all academic researchers, including international students and international "visiting" or tenure-seeking academic professionals ? It would be very helpful if IBM would clearly specify details of the investigative process used to determine "verified researchers".