Step Two: How To Check Out A Link: Looking Deeper
Step Two: How To Check Out A Link: Looking Deeper
Once you’ve decided to look more deeply at a story online, it’s time to figure out where and
when it comes from. Internet news can work like a game of telephone: every time somebody
reposts or rewrites something, there’s a chance that important details will get lost.
The first step in that process is finding the date of the original story — which is one of the
most helpful pieces of information you can get. If the story’s being shared in a Facebook post
or a tweet, click on the post and find its date, otherwise known as the timestamp. You should
also look for the source of the relevant information. Sometimes a news story will explicitly
cite its sources, whether that’s by making clear that the author performed firsthand research
and interviews, or by linking to a press release or another news outlet. If it’s the latter, just
click through to see where the information is coming from, and make sure to check the
timestamp on that as well.
Sometimes, though, it’s unclear where news originated — a story might print an
inflammatory quote without saying where or when it’s from, or a Twitter account might
share a photo with a description that might be wrong. In those cases, do a quick search for
more coverage and original sourcing, generally using a search engine like Bing,
DuckDuckGo, or Google.
For more specific search tips, here are some of the strategies I use.
Looking Deeper
Check the verification
Look for names and keywords
Find survey and infographic sources
Search for quotes
Identify photos and videos
Consider how time-sensitive the story is
See if an old story is still accurate
This can go horribly awry. In January 2019, a local TV station said law enforcement was looking for a
human trafficking suspect around Waco, Texas. A radio station employee summarized the story with a
more urgent headline — “Suspected Human Trafficker, Child Predator May Be in Our Area” — and
posted it on Facebook.
The writer wanted to raise local awareness about a criminal on the loose. Instead, as Slate author Will
Oremus explains, his story got out of control. It was shared hundreds of thousands of times across the
country, likely by users who thought “our area” referred to their town instead of Texas. The suspect was
apprehended soon after, and the article was updated. But people kept sharing the original post for weeks,
because it sounded scary and urgent — apparently, too urgent to check and see if the danger was gone.
The key here is looking for gaps in a story, or mismatches between a story’s claims and its
actual source material. These might be honest mistakes — like accounts sharing satirical
news without realizing it. Or they might be a deliberate attempt to fool people.
There’s no step-by-step guide for understanding a story’s full context. But there are a few
principles you can keep in mind.
Looking Deeper
Is it satire?
Who’s providing the information?
What’s the scale of the story?
If there’s an “outrage,” are people actually upset?
How do different news outlets present the story?
CROWDFUNDING CREDIBILITY
Lots of news outlets cover cool crowdfunded products on Kickstarter and Indiegogo, or mention that a
story’s subject is raising money on GoFundMe. Before you give money to these campaigns, though, you
should make sure they’re not unrealistic or scammy.
For product-based campaigns — like a board game, an indie movie, or a gadget — does the creator have
relevant past experience? Does the funding goal seem way too low to create the product they’re
describing? If they’ve raised money in a previous campaign, were the backers satisfied?
With personal campaigns, look for a connection between the campaign and the person who’s supposed to
get the money — like a link in a news story or from the person’s known social media accounts.
GoFundMe also offers more specific guidelines on its site.
In general, be cautious of crowdfunding projects that seem far more ambitious than mainstream products
and services. If nobody — including the US government — has been able to build a giant border wall
between the US and Mexico, there might be unforeseen difficulties that help explain that. And if big
computing companies aren’t selling an ultra-thin, super-cheap laptop-tablet-phone hybrid, they may have
realized it’s just a bad idea.
You obviously don’t want to believe everything you see or read. But
uncritically disbelieving everything is just as bad. Some news sources really are more
consistently accurate than others. Some expert opinions are more trustworthy than your own
amateur research. If you only believe things that you’ve checked with your own eyes, you’ll
have an incredibly blinkered view of the world.
So the goal here isn’t to identify why a story is wrong. It’s to identify how the story works —
which parts are complicated and subjective, which parts are probably accurate, and how
much it should change your opinions or behavior.
Looking Deeper
Are important facts getting left out or distorted?
What’s the larger narrative?
What happens if you’re wrong?
Why share this story?
More commonly, sources can lie, documents can be faked, and reporters can mishear quotes. Breaking
news stories can be unreliable because nobody — including government officials and other authorities
— knows what’s going on. Radio station WNYC published an excellent “Breaking News Consumer’s
Handbook” for just this reason.
If you share stories on social media, there’s a good chance you’ll eventually post something that’s
inaccurate or misleading, even if you’re diligently doing research.
That doesn’t mean that nothing is true or that every site is equally fake. You might see a bad story from
an outlet that carefully outlines its sources, explains the context of an event, and corrects mistakes when it
finds them. You’re much more likely to see a bad story from an outlet that posts context-free rumors and
doesn’t explain where it’s getting information. If you read a site regularly over time, you’ll get a better
sense of how much to trust it.
By the same token, you might occasionally believe something false if you’re careful. But if you don’t care
about getting things right, it’ll happen much more often.
CONCLUSION
Solving misinformation and disinformation isn’t as simple as following a checklist. Getting
too invested in the checklist can even backfire. Researcher danah boyd has described a dark
side of media literacy education in schools — where asking students to think critically can
cement a blanket assumption that news outlets are lying. And I don’t want to put all the
responsibility for solving misinformation on individuals.
But here’s the thing: I think all this stuff is fun. Tracing the path of information online is one
of my favorite activities, like solving a puzzle or directing an archaeological dig. I want to
share that process with other people — and to make a case for why getting things right is
more interesting and valuable than just confirming your beliefs or scoring points online.
And above all, I want to argue for treating investigation like a shovel, not a knife. Critical
thinking shouldn’t just be a synonym for doubting or debunking something, and the point of
research isn’t simply to poke holes in a story. It’s to understand the story better, or — if
somebody is telling that story maliciously or incompetently — to get deep enough to find the
truth.
EXECUTIVE SUMMARY
Darrell M. West
Vice President and Director - Governance Studies
Founding Director - Center for Technology Innovation
@DarrWest
The news media landscape has changed dramatically over the past decades. Through digital
sources, there has been a tremendous increase in the reach of journalism, social media, and
public engagement. Checking for news online—whether through Google, Twitter, Facebook,
major newspapers, or local media websites—has become ubiquitous, and smartphone alerts
and mobile applications bring the latest developments to people instantaneously around the
world. As of 2017, 93 percent of Americans say they receive news online. [1] When asked
where they got online news in the last two hours, 36 percent named a news organization
website or app; 35 percent said social media (which typically means a post from a news
organization, but can be a friend’s commentary); 20 percent recalled a search engine; 15
percent indicated a news organization email, text, or alert; 9 percent said it was another
source; and 7 percent named a family member email or text (see Figure 1). [2]
In general, young people are most likely to get their news through online sources, relying
heavily on mobile devices for their communications. According to the Pew Research Center,
55 percent of smartphone users receive news alerts on their devices. And about 47 percent of
those receiving alerts click through to read the story. [3] Increasingly, people can customize
information delivery to their personal preferences. For example, it is possible to sign up for
news alerts from many organizations so that people hear news relevant to their particular
interests.
There have been changes overtime in sources of news overall. Figure 2 shows the results for
2012 to 2017. It demonstrates that the biggest gain has been in reliance upon social media. In
2012-2013, 27 percent relied upon social media sites, compared to 51 percent who did so in
2017.[4] In contrast, the percentage of Americans relying upon print news has dropped from
38 to 22 percent.
Figure 2: Change in overall news sources, 2012-2017
TVPrintOnline (incl.
social)Social2013201420152016Year201220170%20%40%60%80%Percent0%80%
Source: Nic Newman, “Digital News Sources,” Reuters Institute for the Study of Journalism, 2017.
In addition, the Reuters Institute for the Study of Journalism has demonstrated important
trends in news consumption. It has shown major gains in reliance upon mobile news
notifications. The percentage of people in the United States making use of this source has
risen by 8 percentage points, while there have been gains of 7 percentage points in South
Korea and 4 percentage points in Australia. There also have been increases in the use of
news aggregators, digital news sources, and voice-activated digital assistants. [6]
DECLINING TRUST IN THE NEWS MEDIA
In the United States, there is a declining public trust in traditional journalism. The Gallup
Poll asked a number of Americans over the past two decades how much trust and confidence
they have in mass media reporting the news fully, accurately, and fairly. As shown in Figure
4, the percentage saying they had a great deal or fair amount of trust dropped from 53 percent
in 1997 to 32 percent in 2016.[7]
Figure 4: Public trust in traditional news media, 1997-2016
200020052010Year199720160%10%20%30%40%50%60%Percent0%60%
Source: Gallup, “Americans’ Trust in Mass Media Sinks to New Low,” September 14, 2016.
Between news coverage they don’t like and fake news that is manipulative in nature, many
Americans question the accuracy of their news. A recent Gallup poll found that only 37
percent believe “news organizations generally get the facts straight.” This is down from
about half of the country who felt that way in 1998. There is also a startling partisan divide in
public assessments. Only 14 percent of Republicans believe the media report the news
accurately, compared to 62 percent for Democrats. Even more disturbingly, “a solid majority
of the country believes major news organizations routinely produce false information.” [8]
This decline in public trust in media is dangerous for democracies. With the current political
situation in a state of great flux in the U.S. and around the world, there are questions
concerning the quality of the information available to the general public and the impact of
marginal media organizations on voter assessments. These developments have complicated
the manner in which people hold leaders accountable and the way in which our political
system operates.
CHALLENGES FACING THE DIGITAL MEDIA LANDSCAPE
As the overall media landscape has changed, there have been several ominous developments.
Rather than using digital tools to inform people and elevate civic discussion, some
individuals have taken advantage of social and digital platforms to deceive, mislead, or harm
others through creating or disseminating fake news and disinformation.
Fake news is generated by outlets that masquerade as actual media sites but promulgate false
or misleading accounts designed to deceive the public. When these activities move from
sporadic and haphazard to organized and systematic efforts, they become disinformation
campaigns with the potential to disrupt campaigns and governance in entire countries. [9]
As an illustration, the United States saw apparently organized efforts to disseminate false
material in the 2016 presidential election. A Buzzfeed analysis found that the most widely
shared fake news stories in 2016 were about “Pope Francis endorsing Donald Trump, Hillary
Clinton selling weapons to ISIS, Hillary Clinton being disqualified from holding federal
office, and the FBI director receiving millions from the Clinton Foundation.” [10] Using a
social media assessment, it claimed that the 20 largest fake stories generated 8.7 million
shares, reactions, and comments, compared to 7.4 million generated by the top 20 stories
from 19 major news sites.
When [fake news] activities move from sporadic and
haphazard to organized and systematic efforts, they
become disinformation campaigns with the potential to
disrupt campaigns and governance in entire countries.
Fake content was widespread during the presidential campaign. Facebook has estimated that
126 million of its platform users saw articles and posts promulgated by Russian sources.
Twitter has found 2,752 accounts established by Russian groups that tweeted 1.4 million
times in 2016.[11] The widespread nature of these disinformation efforts led Columbia Law
School Professor Tim Wu to ask: “Did Twitter kill the First Amendment?” [12]
A specific example of disinformation was the so-called “Pizzagate” conspiracy, which
started on Twitter. The story falsely alleged that sexually abused children were hidden at
Comet Ping Pong, a Washington, D.C. pizza parlor, and that Hillary Clinton knew about the
sex ring. It seemed so realistic to some that a North Carolina man named Edgar Welch drove
to the capital city with an assault weapon to personally search for the abused kids. After
being arrested by the police, Welch said “that he had read online that the Comet restaurant
was harboring child sex slaves and that he wanted to see for himself if they were there.
[Welch] stated that he was armed.”[13]
A post-election survey of 3,015 American adults suggested that it is difficult for news
consumers to distinguish fake from real news. Chris Jackson of Ipsos Public Affairs
undertook a survey that found “fake news headlines fool American adults about 75 percent of
the time” and “‘fake news’ was remembered by a significant portion of the electorate and
those stories were seen as credible.”[14] Another online survey of 1,200 individuals after the
election by Hunt Allcott and Matthew Gentzkow found that half of those who saw these fake
stories believed their content.[15]
False news stories are not just a problem in the United States, but afflict other countries
around the world. For example, India has been plagued by fake news concerning cyclones,
public health, and child abuse. When intertwined with religious or caste issues, the
combination can be explosive and lead to violence. People have been killed when false
rumors have spread through digital media about child abductions. [16]
Sometimes, fake news stories are amplified and disseminated quickly through false accounts,
or automated “bots.” Most bots are benign in nature, and some major sites like Facebook ban
bots and seek to remove them, but there are social bots that are “malicious entities designed
specifically with the purpose to harm. These bots mislead, exploit, and manipulate social
media discourse with rumors, spam, malware, misinformation, slander, or even just noise.” [17]
This information can distort election campaigns, affect public perceptions, or shape human
emotions. Recent research has found that “elusive bots could easily infiltrate a population of
unaware humans and manipulate them to affect their perception of reality, with unpredictable
results.”[18] In some cases, they can “engage in more complex types of interactions, such as
entertaining conversations with other people, commenting on their posts, and answering their
questions.” Through designated keywords and interactions with influential posters, they can
magnify their influence and affect national or global conversations, especially resonating
with like-minded clusters of people.[19]
An analysis after the 2016 election found that automated bots played a major role in
disseminating false information on Twitter. According to Jonathan Albright, an assistant
professor of media analytics at Elon University, “what bots are doing is really getting this
thing trending on Twitter. These bots are providing the online crowds that are providing
legitimacy.”[20] With digital content, the more posts that are shared or liked, the more traffic
they generate. Through these means, it becomes relatively easy to spread fake information
over the internet. For example, as graphic content spreads, often with inflammatory
comments attached, it can go viral and be seen as credible information by people far from the
original post.
OTHER APPROACHES
There are several alternatives to deal with falsehoods and disinformation that can be
undertaken by various organizations. Many of these ideas represent solutions that combat
fake news and disinformation without endangering freedom of expression and investigative
journalism.
Government responsibilities
1) One of the most important thing governments around the world can do is to encourage
independent, professional journalism. The general public needs reporters who help them
make sense of complicated developments and deal with the ever-changing nature of social,
economic, and political events. Many areas are going through transformation that I elsewhere
have called “megachanges,” and these shifts have created enormous anger, anxiety, and
confusion.[32] In a time of considerable turmoil, it is vital to have a healthy Fourth Estate that
is independent of public authorities.
2) Governments should avoid crackdowns on the news media’s ability to cover the news.
Those activities limit freedom of expression and hamper the ability of journalists to cover
political developments. The United States should set a good example with other countries. If
American leaders censor or restrict the news media, it encourages other countries to do the
same thing.
3) Governments should avoid censoring content and making online platforms liable for
misinformation. This could curb free expression, making people hesitant to share their
political opinions for fear it could be censored as fake news. Such overly restrictive
regulation could set a dangerous precedent and inadvertently encourage authoritarian regimes
to weaken freedom of expression.
News industry actions
1) The news industry should continue to focus on high-quality journalism that builds trust
and attracts greater audiences. An encouraging development is that many news organizations
have experienced major gains in readership and viewership over the last couple of years, and
this helps to put major news outlets on a better financial footing. But there have been
precipitous drops in public confidence in the news media in recent years, and this has
damaged the ability of journalists to report the news and hold leaders accountable. During a
time of considerable chaos and disorder, the world needs a strong and viable news media that
informs citizens about current events and long-term trends.
2) It is important for news organizations to call out fake news and disinformation without
legitimizing them. They can do this by relying upon their in-house professionals and well-
respected fact-checkers. In order to educate users about news sites that are created to
mislead, nonprofit organizations such as Politifact, Factcheck.org, and Snopes judge the
accuracy of leader claims and write stories detailing the truth or lack thereof of particular
developments. These sources have become a visible part of election campaigns and candidate
assessment in the United States and elsewhere. Research by Dartmouth College Professor
Brendan Nyhan has found that labeling a Facebook post as “disputed” reduces the percentage
of readers believing the false news by 10 percentage points.[33] In addition, Melissa Zimdars,
a communication and media professor at Merrimack College, has created a list of 140
websites that use “distorted headlines and decontextualized or dubious information.” [34] This
helps people track promulgators of false news.
It is important for news organizations to call out fake
news and disinformation without legitimizing them.
Similar efforts are underway in other countries. In Ukraine, an organization known as
StopFake relies upon “peer-to-peer counter propaganda” to dispel false stories. Its
researchers assess “news stories for signs of falsified evidence, such as manipulated or
misrepresented images and quotes” as well as looking for evidence of systematic
misinformation campaigns. Over the past few years, it has found Russian social media posts
alleging that Ukrainian military forces were engaging in atrocities against Russian
nationalists living in eastern Ukraine or that they had swastikas painted on their vehicles.
[35]
     In a related vein, the French news outlet Le Monde has a “database of more than 600
news sites that have been identified and tagged as ‘satire,’ ‘real,’ [or] ‘fake.’” [36]
Crowdsourcing draws on the expertise of large numbers of readers or viewers to discern
possible problems in news coverage, and it can be an effective way to deal with fake news.
One example is The Guardian’s effort to draw on the wisdom of the crowd to assess 450,000
documents about Parliament member expenses in the United Kingdom. It received the
documents but lacked the personnel quickly to analyze their newsworthiness. To deal with
this situation, the newspaper created a public website that allowed ordinary people to read
each document and designate it into one of four news categories: 1) “not interesting,” 2)
“interesting but known,” 3) “interesting,” or 4) “investigate this.” [37] Digital platforms allow
news organizations to engage large numbers of readers this way. The Guardian, for example,
was able “to attract 20,000 readers to review 170,000 documents in the first 80
hours.”[38] These individuals helped the newspaper to assess which documents were most
problematic and therefore worthy of further investigation and ultimately news coverage.
Technology company responsibilities
1) Technology firms should invest in technology to find fake news and identify it for
users through algorithms and crowdsourcing. There are innovations in fake news and hoax
detection that are useful to media platforms. For example, fake news detection can be
automated, and social media companies should invest in their ability to do so. Former FCC
Commissioner Tom Wheeler argues that “public interest algorithms” can aid in identifying
and publicizing fake news posts and therefore be a valuable tool to protect consumers. [39]
In this vein, computer scientist William Yang Wang, relying upon PolitiFact.com, created a
public database of 12,836 statements labeled for accuracy and developed an algorithm that
compared “surface-level linguistic patterns” from false assertions to wording contained in
digital news stories. This allowed him to integrate text and analysis, and identify stories that
rely on false information. His conclusion is that “when combining meta-data with text,
significant improvements can be achieved for fine-grained fake news detection.” [40] In a
similar approach, Eugenio Tacchini and colleagues say it is possible to identify hoaxes with a
high degree of accuracy. Testing this proposition with a database of 15,500 Facebook posts
and over 909,000 users, they find an accuracy rate of over 99 percent and say outside
organizations can use their automatic tool to pinpoint sites engaging in fake news. [41] They
use this result to advocate the development of automatic hoax detection systems.
Algorithms are powerful vehicles in the digital era and help shape people’s quest for
information and how they find online material. They can also help with automatic hoax
detection, and there are ways to identify fake news to educate readers without censoring it.
According to Kelly Born of the William and Flora Hewlett Foundation, digital platforms
should down rank or flag dubious stories, and find a way to better identify and rank authentic
content to improve information-gathering and presentation. [42] As an example, several media
platforms have instituted “disputed news” tags that warn readers and viewers about
contentious content. This could be anything from information that is outright false to
material where major parties disagree about its factualness. It is a way to warn readers about
possible inaccuracies in online information. Wikipedia is another platform that does this.
Since it publishes crowdsourced material, it is subject to competing claims regarding factual
accuracy. It deals with this problem by adding tags to material identifying it as “disputed
news.”
Yet this cannot be relied on by itself. A survey of 7,500 individuals undertaken by David
Rand and Gordon Pennycook of Yale University argue that alerting readers about inaccurate
information doesn’t help much. They explored the impact of independent fact-checkers and
claim that “the existence of ‘disputed’ tags made participants just 3.7 percentage points more
likely to correctly judge headlines as false.”[43] The authors worry that the outpouring of false
news overwhelms fact-checkers and makes it impossible to evaluate disinformation.
Algorithms are powerful vehicles in the digital era, and
they can help establish automatic hoax detection systems.
2) These companies shouldn’t make money from fake news manufacturers and should make
it hard to monetize hoaxes. It is important to weaken financial incentives for bad content,
especially false news and disinformation, as the manufacturing of fake news is often
financially motivated. Like all clickbait, false information can be profitable due to ad
revenues or general brand-building. Indeed, during the 2016 presidential campaign, trolls in
countries such as Macedonia reported making a lot of money through their dissemination of
erroneous material. While social media platforms like Facebook have made it harder for
users to profit from fake news,[44] ad networks can do much more to stop the monetization of
fake news, and publishers can stop carrying the ad networks that refuse to do so.
3) Strengthen online accountability through stronger real-name policies and enforcement
against fake accounts. Firms can do this through “real-name registration,” which is the
requirement that internet users have to provide the hosting platform with their true identity.
This makes it easier to hold individuals accountable for what they post or disseminate online
and also stops people from hiding behind fake names when they make offensive comments
or engage in prohibited activities.[45] This is relevant to fake news and misinformation
because of the likelihood that people will engage in worse behavior if they believe their
actions are anonymous and not likely to be made public. As famed Justice Louis Brandeis
long ago observed, “sunshine is said to be the best of disinfectants.” [46] It helps to keep
people honest and accountable for their public activities.
Educational institutions
1) Funding efforts to enhance news literacy should be a high priority for governments. This
is especially the case with people who are going online for the first time. For those
individuals, it is hard to distinguish false from real news, and they need to learn how to
evaluate news sources, not accept at face value everything they see on social media or digital
news sites. Helping people become better consumers of online information is crucial as the
world moves towards digital immersion. There should be money to support partnerships
between journalists, businesses, educational institutions, and nonprofit organizations to
encourage news literacy.
2) Education is especially important for young people. Research by Joseph Kahne and
Benjamin Bowyer found that third-party assessments matter to young readers. However,
their effects are limited. Those statements judged to be inaccurate reduced reader persuasion,
although to a lower extent than alignment with the individual’s prior policy beliefs. [47] If the
person already agreed with the statement, it was more difficult for fact-checking to sway
them against the information.
How the public can protect itself
1) Individuals can protect themselves from false news and disinformation by following
a diversity of people and perspectives. Relying upon a small number of like-minded news
sources limits the range of material available to people and increases the odds they may fall
victim to hoaxes or false rumors. This method is not entirely fool-proof, but it increases the
odds of hearing well-balanced and diverse viewpoints.
2) In the online world, readers and viewers should be skeptical about news sources. In the
rush to encourage clicks, many online outlets resort to misleading or sensationalized
headlines. They emphasize the provocative or the attention-grabbing, even if that news hook
is deceptive. News consumers have to keep their guard up and understand that not everything
they read is accurate and many digital sites specialize in false news. Learning how to judge
news sites and protect oneself from inaccurate information is a high priority in the digital
age.
CONCLUSION
From this analysis, it is clear there are a number of ways to promote timely, accurate, and
civil discourse in the face of false news and disinformation. [48] In today’s world, there is
considerable experimentation taking place with online news platforms. News organizations
are testing products and services that help them identify hate speech and language that incites
violence. There is a major flowering of new models and approaches that bodes well for the
future of online journalism and media consumption.
At the same time, everyone has a responsibility to combat the scourge of fake news and
disinformation. This ranges from the promotion of strong norms on professional journalism,
supporting investigative journalism, reducing financial incentives for fake news, and
improving digital literacy among the general public. Taken together, these steps would
further quality discourse and weaken the environment that has propelled disinformation
around the globe.
Note: I wish to thank Hillary Schaub and Quinn Bornstein for their valuable research
assistance. They were very helpful in finding useful materials for this project.
The Brookings Institution is a nonprofit organization devoted to independent research and
policy solutions. Its mission is to conduct high-quality, independent research and, based on
that research, to provide innovative, practical recommendations for policymakers and the
public. The conclusions and recommendations of any Brookings publication are solely those
of its author(s), and do not reflect the views of the Institution, its management, or its other
scholars.
    Support for this publication was generously provided by Facebook. Brookings recognizes
    that the value it provides is in its absolute commitment to quality, independence, and impact.
    Activities supported by its donors reflect this commitment.
    e know people want to see accurate information on Facebook – and so do we.
    False news is harmful to our community, it makes the world less informed, and it
    erodes trust. It's not a new phenomenon, and all of us — tech companies, media
    companies, newsrooms, teachers — have a responsibility to do our part in
    addressing it. At Facebook, we're working to fight the spread of false news in three
    key areas:
         Better identifying false news through our community and third-party fact-
    checking organizations so that we can limit its spread, which, in turn, makes it
    uneconomical.
          Making it as difficult as possible for people posting false news to buy ads on
    our platform through strict enforcement of our policies.
           Working with Partners: We believe providing more context can help people
    decide for themselves what to trust and what to share. We've started a program to
    work with independent third-party fact-checking organizations. We'll use the reports
    from our community, along with other signals, to send stories to these organizations.
    If the fact-checking organizations identify a story as false, it will get flagged as
    disputed and there will be a link to a corresponding article explaining why. Stories
    that have been disputed also appear lower in News Feed.