Facebook doesn’t just have a fake news problem. It also has a fake ads problem.
Last month the company disclosed that it found about 3,000 ads that ran during the 2016 US election that were run by fake accounts linked to Russia. While the company hasn’t shown the ads publicly, there’s evidence that they were designed to influence voters by promoting polarizing topics. It was a wild abuse of Facebook’s automated ad platform.
Facebook’s answer: Throw bodies at the problem.
Facebook announced on Monday that it would hire 1,000 people in the coming months to monitor automated ads on the social network and remove those that don’t meet its guidelines. If that sounds familiar, it’s because it’s almost the same response Facebook had in May after a string of incidents where users live-streamed suicides and killings. Facebook said then that it would hire 3,000 new content moderators on top of the 4,500 it already had.
In addition to the 1,000 new hires announced Monday, Facebook said it’s going to beef up its algorithms using machine learning to automatically detect abusive ads.
Ignoring the obvious fact that these moves come far too late, there’s still a massive lack of transparency about what Facebook says will make its advertising more transparent.
Here’s what we still don’t know:
-
Are the new employees who will be monitoring content full time staffers or contractors?
-
Where are they based?
-
What kind of training will they go through and what specifically are they looking for?
-
What kind of machine learning improvements will Facebook make and what kind of content will the algorithm start flagging that it isn’t flagging today?
-
Is Facebook working with the FEC and other governments to come up with ad monitoring standards, or is it developing these standards on its own?