Metaโs Oversight Board said the company should improve its automated image takedown system. The system mistakenly removed posts featuring a cartoon image showing police violence in Colombia.
The image was added to Metaโs Media Matching Service database. Once an image is added to the database, it is automatically taken down when it is posted.
According to the Oversight Board, 215 people appealed the removal of the image, and 98% of them were successful. Despite the successful appeals, Meta did not remove the image from the database.
โBy using automated systems to remove content, Media Matching Service banks can amplify the impact of incorrect decisions by individual human reviewers,โ the board said. It added that the appeals should have triggered a review of the image and its removal from the database, because one bad decision could result in an image being indefinitely prohibited, despite different conclusions by human reviewers.
The board recommended more methods of oversight.
โThe board is particularly concerned that Meta does not measure the accuracy of Media Matching Service banks for specific content policies,โ it notes. โWithout this data, which is crucial for improving how these banks work, the company cannot tell whether this technology works more effectively for some community standards than others.โ
The Oversight Board also asked Meta to publish the error rates of content that has been mistakenly included in the Media Matching Service database.
The board also considered a case surrounding Facebookโs policy on extremism, where it appeared to not make the difference between reporting on extremist groups and supporting them.
The board concluded that Meta made a mistake when it took down a post in Urdu reporting on the Taliban re-opening schools for girls and women. Facebook removed the post for violating its rule that ban โpraiseโ of extremist groups like the Taliban. The removal of the post was appealed, but was never reviewed because Facebook has less than 50 reviewers that speak Urdu.
The board said the case โmay indicate a wider problemโ with the rules on dangerous organizations. The rules are not clear to moderators and users, and the punishments are also โunclearโ and severe. The board recommended that Meta should make the rules and definition of โpraisingโ dangerous individuals clearer and add more reviewers.