Daniel Howley, Yahoo! Tech, June 29, 2015
The image recognition software built into Google Photos is impressive. Show it a photo of the Eiffel Tower, and it will know that picture was taken in Paris. Snap an image of a dog or a tree, and it will automatically put them in a group with all your other pictures of dogs or trees.
But the software is far from foolproof. And when it fails, it does so in a spectacular way–as when it recently processed a photo of two black friends and labeled them “Gorillas.”
Jacky Alcine, a 21-year old programmer who lives in Brooklyn, N.Y., was checking out his Google Photos account last night when he saw that the service had automatically generated a folder titled “Gorillas.” It contained nothing but pictures of him and a friend that he had taken in 2013.
When alerted to the error, Google provided a solution to the problem within hours and issued an immediate mea culpa.
“We’re appalled and genuinely sorry that this happened,” a Google representative told Yahoo Tech.
“We are taking immediate action to prevent this type of result from appearing. There is still clearly a lot of work to do with automatic image labeling, and we’re looking at how we can prevent these types of mistakes from happening in the future,” the spokesperson added.
Alcine said that as of Monday evening, the issue had largely been addressed, though he noted, “there’s still complications with the hands obscuring the face causing it to still match to the gorilla tag. Chimp gives results as well (but not chimpanzee).”
Update: Flickr had a similar issue with its automatic tagging algorithm. In May, several users complained that tags with the words “animal” and “ape” were being automatically added to images of both black and white people.