SAN FRANCISCO: When a doctored video of House Speaker Nancy Pelosi — one altered to show the Democratic leader slurring her words — began making the rounds on Facebook last week, the social network didn’t take it down. Instead, it “downranked” the video, a behind-the-scenes move intended to limit its spread.
That outraged some people who believe Facebook should do more to clamp down on misinformation. Pelosi derided Facebook Wednesday for not taking down the video even though it knows it is false.
But the company and some civil libertarians warn that Facebook could evolve into an unaccountable censor if it’s forced to make judgment calls on the veracity of text, photos or videos.
Facebook has long resisted making declarations about the truthfulness of posts that could open it up to charges of censorship or political bias. It manages to get itself in enough trouble simply trying to enforce more basic rules in difficult cases, such as the time a straightforward application of its ban on nudity led it to remove an iconic Vietnam War photo of a naked girl fleeing a napalm attack. (It backed down after criticism from the prime minister of Norway, among others.)
But staying out of the line of fire is harder than it used to be, given Facebook’s size, reach and impact on global society. The social network can’t help but run into controversy given its 2.4 billion users and the sorts of decisions it must make daily— everything from which posts and links it highlights in your news feed to deciding what counts as hate speech to banning controversial figures or leaving them be.
Facebook has another incentive to keep its head down. The deeper it gets into editorial decisions, the more it looks like a publisher, which could tempt legislators to limit the liability shield it currently enjoys under federal law. In addition, making judgments about truth and falsity could quickly become one of the world’s biggest headaches.
For instance, Republican politicians and other conservatives, from President Donald Trump to Fox News personalities, have been trumpeting the charge that Facebook is biased against conservatives. That’s a “false narrative,” said Siva Vaidhyanathan, director of the Center for Media and Citizenship at the University of Virginia. But as a result, he said, “any effort to clean up Facebook now would spark tremendous fury.”
Twitter hasn’t removed the doctored Pelosi video, either, and declined comment on its handling of it. But YouTube yanked it down, pointing to community guidelines that prohibit spam, deceptive practices and scams. Facebook has a similar policy that prohibits the use of “misleading and inaccurate” information to gain likes, followers or shares, although it apparently decided not to apply it in this case.
None of these companies explicitly prohibit false news, although Facebook notes that it “significantly” reduces the distribution of such posts by pushing them lower in user news feeds.
The problem is that such downranking doesn’t quite work, Vaidhyanathan said. As of Wednesday, the video shared on Facebook by the group Politics Watchdog had been viewed nearly 3 million times and shared more than 48,000 times. By contrast, other videos posted by this group in the past haven’t had more than a few thousand views apiece.
Further complicating matters is the fact that Facebook is starting to de-emphasize the news feed itself. CEO Mark Zuckerberg has outlined a broad strategy that will emphasize private messaging over public sharing on Facebook. And Facebook groups, many of which are private, aren’t subject to downranking, Vaidhyanathan said.
Facebook didn’t respond to emailed questions about its policies and whether it is considering changes that would allow it to remove similar videos in the future. In an interview last week with CNN’s Anderson Cooper, Facebook’s head of global policy, Monika Bickert, defended the company’s decision , noting that users are “being told” that the video is false when they view or share it.
That might be a stretch. When an Associated Press reporter attempted to share the video as a test, a Facebook pop-up noted the existence of “additional reporting” on the video with links to fact-check articles, but didn’t directly describe the video as false or misleading.
Alex Stamos, Facebook’s former security chief, tweeted Sunday that few critics of the social network’s handling of the Pelosi video could articulate realistic enforcement standards beyond “take down stuff I don’t like.” Mass censorship of misleading speech on Facebook, he wrote, would be “a huge and dangerous increase in FB’s editorial power.”
Last year, Zuckerberg wrote on Facebook that the company focuses on downranking so-called “borderline content,” stuff that doesn’t violate its rules but is provocative, sensationalist, “click-bait or misinformation.”
While it’s true that Facebook could just change its rules around what is allowed — moving the line on acceptable material — Zuckerberg said this doesn’t address the underlying problem of incentive. If the line of what is allowed moves, those creating material would just push closer to that new line.
Facebook continuously grapples with the right way to deal with new forms of misinformation, Nathaniel Gleicher, the company’s head of cybersecurity policy, said in a February interview with the AP. The problem is far more complex than carefully manipulated “deepfake” videos that show people doing things they never did, or even crudely doctored videos such as the Pelosi clip.
Any consistent policy, Gleicher said, would have to account for edited images, ones presented out of context (such as a decade-old photo presented as current), doctored audio and more. He said it’s a huge challenge to accurately identify such items and decide what type of disclosure to require when they’re edited.
Facebook isn’t deleting the fake Pelosi video. Should it?
Facebook isn’t deleting the fake Pelosi video. Should it?
- Pelosi derided Facebook Wednesday for not taking down the video even though it knows it is false
- Facebook has long resisted making declarations about the truthfulness of posts that could open it up to charges of censorship or political bias
DCO and Arab News partner to combat digital misinformation, explore AI’s impact on media
KUWAIT CITY: The Digital Cooperation Organization (DCO) and the international Saudi newspaper Arab News have signed a Letter of Engagement aimed at strengthening knowledge and expertise exchange on the impact of artificial intelligence in the media sector, as well as leveraging expert insights to develop best practices to combat online misinformation amid accelerating technological advancements.
DCO said this step aligned with its efforts to strengthen collaboration with international media institutions to support responsible dialogue around digital transformation and contribute to building a more reliable, inclusive, and sustainable digital media environment.
Commenting on the agreement, Deemah AlYahya, Secretary-General of the Digital Cooperation Organization, said: “At a moment when AI is reshaping how truth is produced, distributed, and trusted, partnership with credible media institutions is essential.”
She added that “working with Arab News allows us to bridge technology and journalism in a way that protects integrity, strengthens public trust, and elevates responsible innovation. This collaboration is about equipping media ecosystems with the tools, insight, and ethical grounding needed to navigate AI’s impact, while ensuring digital transformation serves people and their prosperity.”
Faisal J. Abbas, Editor-in-Chief of Arab News, emphasized that the partnership enhances media institutions’ ability to keep pace with technological shifts, noting that engagement with representatives of DCO Member States enables deeper understanding of emerging technologies and regulatory developments in the digital space.
He added: “DCO’s commitment to initiatives addressing online content integrity reflects a clear dedication to supporting a responsible digital environment that serves societies and strengthens trust in the digital ecosystem.”
The Letter of agreement was signed on the sidelines of the Fifth DCO General Assembly held in Kuwait City under the theme “Inclusive Prosperity in the Age of AI”, alongside the second edition of the International Digital Cooperation Forum, held from 4–5 February, which brought together ministers, policymakers, business leaders, entrepreneurs, and civil society representatives from more than 60 countries to strengthen international cooperation toward a human-centric, inclusive, and sustainable digital economy.










