Facebook is leveraging its vast resources to help protect the 2020 U.S. election against the kind of massive manipulation and disinformation efforts that the platform failed to act on in 2016.
The moves come with Facebook taking intense criticism from the Cambridge Analytica scandal in which the British consultancy hijacked personal data from tens of millions of Americans to help Donald Trump's campaign, and for enabling foreign influence campaigns including from Russia four years ago.
Here are some of the measures Facebook has taken to avoid a repeat of the 2016 fiasco and to deliver authoritative information:
After being accused of enabling "voter suppression" efforts in 2016, Facebook has set a goal of helping register at least 4 million voters this year and on Monday said it exceeded that with 4.4 million registrations.
The social network used by more than 200 million Americans is delivering reminders to people's feeds and offering links on how and where to vote.
It is also removing posts that seek to discourage going to the polls or spreading false information such as that voting centers are risky for coronavirus infections.
Facebook also pledged to remove posts that seek to intimidate voters and took down a video from Donald Trump Jr. calling for an "army" of volunteers to protect polling stations.
Facebook has heavily invested in rooting out false information that could have an impact on voting.
It has fact-checking partnerships with dozens of media organizations, including Agence France-Presse (AFP), which can result in labeling of the post or a limit on how much it can be shared.
Facebook has also decided to place labels on media groups that are state-controlled, both financially and editorially, and to block these organizations from posting ads targeting Americans.
Facebook has been for months battling what it calls "inauthentic coordinated behavior," or those using fake or disguised accounts to spread political misinformation.
Many of these influence campaigns have been from overseas, including from Russia and Iran, but a domestic-based group using fake accounts to praise Trump was recently banned as part of the platform's latest crackdown on orchestrated deception.
Facebook and other platforms are on the watch for potential "hack and leak" operations which could be used to spread false information about a candidate. To that end, it has barred new political ads one week before the Nov. 3 election.
And it has taken steps to improve the security of the accounts of political candidates to guard against a takeover by bad actors.
Facebook has also developed contingency plans in case of a disputed election result, and the potential for unrest.
It has banned political ads after polls close on Election Day.
Any posts prematurely declaring a winner or contesting the count will be labeled with reliable information from news outlets and election officials.
"If a candidate or party declares premature victory before a race is called by major media outlets, we will add more specific information in the notifications that counting is still in progress and no winner has been determined," said Vice President of Integrity Guy Rosen.
Nick Clegg, a former deputy British prime minister who is Facebook's head of global affairs, said the top social media platform could take exceptional steps to "restrict the circulation of content" in the case of turmoil.
"We have acted aggressively in other parts of the world where we think that there is real civic instability, and we obviously have the tools to do that," Clegg said in an interview published Tuesday in the Financial Times, in comments confirmed by Agence France-Presse (AFP).
The comments are in line with prior reports that Facebook could deploy a "kill switch" to thwart the spread of misinformation in case of a dispute on U.S. election results.