Facebook pledges tough security efforts ahead of US elections

Internal memo from outgoing security chief critical of practices at the social network

Facebook has faced fierce criticism over how it handles political propaganda and misinformation since the 2016 US election.
Facebook has faced fierce criticism over how it handles political propaganda and misinformation since the 2016 US election.

Facebook is using a range of techniques including artificial intelligence to counter Russian operatives or others who use deceptive tactics and false information to manipulate public opinion.

Officials at the company told reporters in a telephone briefing they expected to find such efforts on the social network ahead of the US mid-term elections in November, but declined to disclose whether they have already uncovered any such operations.

Facebook has faced fierce criticism over how it handles political propaganda and misinformation since the 2016 US election, which US intelligence agencies say was influenced by the Russian government, in part through social media.

The controversy has not abated despite Facebook initiatives including a new tool that shows all political advertising that is running on the network and new fact-checking efforts to inform users about obvious falsehoods.

READ MORE

But the company reiterated on Tuesday that it will not take down postings simply because they are false. Chief executive Mark Zuckerberg last week drew fire for citing Holocaust denials as an example of false statements that would not be removed if they were sincerely voiced.

Tuesday's briefing, which included Nathaniel Gleicher, head of cybersecurity policy, and Tessa Lyons, manager of Facebook's core "news feed," came just before the publication of an internal staff message from Facebook's outgoing chief security officer that was sharply critical of many company practices.

The note by Alex Stamos, written in March after he said he was going to leave the company, urged colleagues to heed feedback about "creepy" features, collect less data and "deprioritise short-term growth and revenue" to restore trust. He also urged the company's leaders to "pick sides when there are clear moral or humanitarian issues".

Mr Stamos posted the note on an internal Facebook site but Reuters confirmed its authenticity. It was first disclosed by Buzzfeed News.

Mr Stamos said the company needed to be more open in how it manages content on its network, which has become a major medium for political activity in many countries around the world. Tuesday’s media briefing was part of the company’s efforts in that direction.

Ms Lyons said the company was making progress in smoothing its process for fact-checkers assigned to label false information. Once an article is labeled false, users are warned before they share it and subsequent distribution drops 80 per cent, Ms Lyons said.

Posts from sites that often distribute false information are ranked lower in the calculations that determine what each user sees but are not entirely removed from view.

Mr Gleicher said those seeking to deliberately promote misinformation often use fake accounts to amplify their content or run afoul of community standards, both of which are grounds for removing posts or entire pages.

He said the company would use a type of artificial intelligence known as machine learning as part of its efforts to root out abuses. – Reuters