Download Fake news spread through social media platforms was

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Internet relationship wikipedia , lookup

Social commerce wikipedia , lookup

Transcript
• Newspaper companies are losing advertisers, readers,
market value, and, in some cases, their sense of
mission. (Brown, Losing the News, p. 7).
• Independent, publicly traded American newspapers
have lost forty-two per cent of their market value in the
past three years.
• Only nineteen per cent of Americans between the ages
of eighteen and thirty-four claim to look at a daily
newspaper. The average age of the American
newspaper reader is fifty-five and rising. Since 2004,
newspapers were least preferred source for news
amongst young people. (Brown, Losing the News).
Changes in media ecosystem
• Online news sites are growing rapidly and
some low quality sites have far more readers
than traditional “respected” news sites.
• Some have argued that demand for “clickbait”
created by Facebook causes some online
reporters to become more sensationalist and
less careful about the facts.
• Raises issue of politics of algorithms and “filter
bubbles” (Pariser)
Changes in media ecosystem
• Many traditional gatekeepers, fact checkers
and filters disappear.
• Is this more democratic?
• The collapse of gatekeepers, agreement on
facts, and the emergence of fake news, with
Facebook playing a key distribution role.
• What was your Facebook page like in the
months preceding the election?
• Mainstream news, in an economic death spiral due to
the new social media landscape, edges closer to
entertainment and farther from public service? (How
many questions about climate change in the
presidential debates?)
• “The contraction in the reporting corps, combined with
the success of disinformation this year, is making for
some sleepless nights for those in Washington who will
have to govern in this bifurcated, real-news-fake-news
environment.” Kristof
• ‘It’s the biggest crisis facing our democracy - the
failing business model of real journalism.” Senator
Claire McCaskill
“Post Truth”?
The First “Facebook Election”?
• “In the last couple of weeks, Facebook, Twitter and other social media
outlets have exposed millions of Americans to false stories asserting that:
the Clinton campaign’s pollster, Joel Benenson, wrote a secret memo
detailing plans to “salvage” Hillary Clinton’s candidacy by launching a
radiological attack to halt voting (merrily shared on Twitter by Roger
Stone, an informal adviser to the Trump campaign); the Clinton campaign
senior strategist John Podesta practiced an occult ritual involving various
bodily fluids; Mrs. Clinton is paying public pollsters to skew results (shared
on Twitter by Donald Trump Jr.); there is a trail of supposedly suspicious
deaths of myriad Clinton foes.”
• “As Mike Cernovich, a Twitter star, alt-right news provocateur and
promoter of Clinton health conspiracies, boasted in last week’s New
Yorker, “Someone like me is perceived as the new Fourth Estate.” His
content can live alongside that of The Times or The Boston Globe or The
Washington Post on the Facebook newsfeed and be just as well read, if not
more so. On Saturday he called on a President Trump to disband the
White House press corps.
• Some journalists and scholars are claiming sites like
Facebook may have had a significant impact on the
election via “fake news”(sour grapes?)
• When faced with claims that news distributed via
Facebook may have impacted the election, Zuckerberg
at first dismissed this as “a pretty crazy idea.”
• Yet almost 50% of Americans get most of their news
via Facebook.
• Facebook’s algorithms actively curate the material we
see (they are no longer displayed chronologically). It is,
in some real sense, a “publisher,” not just a platform.
• Facebook took credit for its role in the Arab Spring, i.e. that
it played a vital role in political struggle overseas.
• In Myanmar misinformation on Facebook reportedly
helped fuel ethnic cleansing, creating a refugee crisis.
• Antonio Garcia-Martinez helped lead Facebook's efforts to
target ads. He says his ad team used to joke: “It would be
so easy to throw an election, just by showing vote
reminders to select counties.”
• "There's an entire political team and a massive [Facebook]
office in D.C. that tries to convince political advertisers that
Facebook can convince users to vote one way or the
other. Then Zuck gets up and says, 'Oh, by the way,
Facebook content couldn't possibly influence the election.'
It's contradictory on the face of it."
• In 2010 Facebook conducted an experiment on 61 million users before
the midterm elections. One group was shown a “go vote” message as a
plain box, while another group saw the same message with a tiny
addition: thumbnail pictures of their Facebook friends who had clicked on
“I voted.” Using public voter rolls to compare the groups after the election,
the researchers concluded that the second post had turned out hundreds
of thousands of voters.
• In 2012, Facebook researchers again secretly tweaked the newsfeed for
an experiment: Some people were shown slightly more positive posts,
while others were shown slightly more negative posts. Those shown more
upbeat posts in turn posted significantly more of their own upbeat posts;
those shown more downbeat posts responded in kind. Decades of other
research concurs that people are influenced by their peers and social
networks. All of this renders preposterous Mr. Zuckerberg’s claim that
Facebook, a major conduit for information in our society, has “no
influence.” (Zeynep Tufekci)
• On November 15, 2016, Zuckerberg vowed to “weed out fake
news.”
• “Some Facebook employees are worried about the spread of racist
and so-called alt-right memes across the network, according to
interviews with 10 current and former Facebook employees. Others
are asking whether they contributed to a “filter bubble” among
users who largely interact with people who share the same beliefs.
Even more are reassessing Facebook’s role as a media company and
wondering how to stop the distribution of false information. Some
employees have been galvanized to send suggestions to product
managers on how to improve Facebook’s powerful news feed: the
streams of status updates, articles, photos and videos that users
typically spend the most time interacting with.
• Facebook became “a sewer of misinformation.”
(Harvard’s Nieman Lab).
• “I’ve come to think that the rise of fake news —
and of the cheap-to-run, ideologically driven
aggregator sites that are only a few steps up from
fake — has weaponized those filter bubbles.
There were just too many people voting in this
election because they were infuriated by madeup things they read online.” Nieman Lab
• One example: I’m from a small town in south
Louisiana. The day before the election, I looked at
the Facebook page of the current mayor. Among
the items he posted there in the final 48 hours of
the campaign: Hillary Clinton Calling for Civil War
If Trump Is Elected. Pope Francis Shocks World,
Endorses Donald Trump for President. Barack
Obama Admits He Was Born in Kenya. FBI Agent
Who Was Suspected Of Leaking Hillary’s
Corruption Is Dead. (Nieman Lab)
•
http://www.niemanlab.org/2016/11/the-forces-that-drove-this-elections-media-failure-are-likely-to-get-worse/
• “A fake story claiming Pope Francis — actually
a refugee advocate — endorsed Mr. Trump
was shared almost a million times, likely
visible to tens of millions,” Zeynep Tufekci, an
associate professor at the University of North
Carolina who studies the social impact of
technology, said... “Its correction was barely
heard. Of course Facebook had a significant
influence in this last election’s outcome.”
Fake News & Digital Literacy
1. Fake news spread through
social media platforms
was far bigger and more
influential than previously
thought. One study
suggests “top fake election
news stories generated
more total engagement on
Facebook than top
election stories from 19
major news outlets
combined.”
• Automated AI Twitter “swarms” or “commentbots” whenever topics like climate change,
gun control, or other hot-button issues come
up? Some evidence for this emerging.
• In a way these echo fabricated letters to the
editor, astroturf groups, and other strategies
from the past. However, at present, they are
harder to identify and track.
Fake News & Digital Literacy
2. It is becoming an international phenomenon.
• The degree of the problem is suggested by the
fact that some major political figures could
not tell fake news from real news. Two
prominent examples were president Trump
and General Flynn, the (until recently)
national security advisor.
•
A day after a black activist was kicked and punched by voters at a Donald Trump rally in
Alabama, Trump retweeted an image packed with racially loaded and incorrect murder
statistics originally published by a white supremacist.
https://www.nytimes.com/interactive/2017/business/media/trump-fake-news.html?_r=0
Why it Matters
• Reading Fake News, Pakistani Minister Directs
Nuclear Threat at Israel.
Why?
• “Why is this all happening now? Clay Shirky, a professor at New
York University who has studied the effects of social networks,
suggested a few reasons.
• One is the ubiquity of Facebook, which has reached a truly epic
scale. Last month the company reported that about 1.8 billion
people now log on to the service every month. Because social
networks feed off the various permutations of interactions among
people, they become strikingly more powerful as they grow. With
about a quarter of the world’s population now on Facebook, the
possibilities are staggering. “When the technology gets boring,
that’s when the crazy social effects get interesting,” Mr. Shirky said.
• One of those social effects is what Mr. Shirky calls the “shifting of
the Overton Window,” a term coined by the researcher Joseph P.
Overton to describe the range of subjects that the mainstream
media deems publicly acceptable to discuss.
• From about the early 1980s until the very recent past, it was usually
considered unwise for politicians to court views deemed by most of
society to be out of the mainstream, things like overt calls to racial bias.
• “White ethnonationalism was kept at bay because of pluralistic
ignorance,” Mr. Shirky said. “Every person who was sitting in their
basement yelling at the TV about immigrants or was willing to say white
Christians were more American than other kinds of Americans — they
didn’t know how many others shared their views.”
• Thanks to the internet, now each person with once-maligned views can
see that he’s not alone. And when these people find one another, they can
do things — create memes, publications and entire online worlds that
bolster their worldview, and then break into the mainstream. The groups
also become ready targets for political figures like Mr. Trump, who
recognize their energy and enthusiasm and tap into it for real-world
victories.
• Mr. Shirky notes that the Overton Window isn’t
just shifting on the right. We see it happening on
the left, too. Mr. Sanders campaigned on an antiWall Street platform that would have been
unthinkable for a Democrat just a decade ago.
• The upshot is further unforeseen events. “We’re
absolutely going to get more of these insurgent
candidates, and more crazy social effects,” Mr.
Shirky said.
• Mr. Trump is just the tip of the iceberg. Prepare
for interesting times.
From an era of “Truthiness” to a “Posttruth” age?
• Oxford Dictionaries selected “post-truth” as
2016's international word of the year. The
dictionary defines “post-truth” as “relating to
or denoting circumstances in which objective
facts are less influential in shaping public
opinion than appeals to emotion and personal
belief.”
• In a 2008 book, I argued that the internet would
usher in a “post-fact” age. Eight years later, in the
death throes of an election that features a
candidate who once led the campaign to lie
about President Obama’s birth, there is more
reason to despair about truth in the online age.
Why? Because if you study the dynamics of how
information moves online today, pretty much
everything conspires against truth.
Farhad Manjoo, “How the Internet Is Loosening
Our Grip on the Truth”
• “The internet is distorting our collective grasp
on the truth. Polls show that many of us have
burrowed into our own echo chambers of
information. In a recent Pew Research Center
survey, 81 percent of respondents said that
partisans not only differed about policies, but
also about “basic facts.” Manjoo.
Fake News: Institutionalized and
“Weaponized”
• “Where hoaxes before were shared by your great-aunt who
didn’t understand the internet, the misinformation that
circulates online is now being reinforced by political
campaigns, by political candidates or by amorphous
groups of tweeters working around the campaigns,” said
Caitlin Dewey, a reporter at The Washington Post who once
wrote a column called “What Was Fake on the Internet This
Week.”
Ms. Dewey’s column began in 2014, but by the end of last
year, she decided to hang up her fact-checking hat because
she had doubts that she was convincing anyone.
What is the top story on your face
book feed today – is it this?
• What solutions are being proposed?
• Critical digital literacy is thus far more
important than we had assumed.
• So what is to be done?
• [One] option might be to empower users to
improve their online communities through peerto-peer sanctioning. To test this hypothesis,
I used Twitter accounts I controlled (“bots”) to
send messages designed to remind harassers of
the humanity of their victims and to reconsider
the norms of online behavior.
• I sent every harasser the same message:
– @[subject] Hey man, just remember that there are
real people who are hurt when you harass them with
that kind of language
• I used a racial slur as the search term because
I thought of it as the strongest evidence that a
tweet might contain racist harassment. I
restricted the sample to users who had a
history of using offensive language, and I only
included subjects who appeared to be a white
man or who were anonymous.
• I created two types of bots: white men and black men. To
manipulate the race, I used the same cartoon avatar for the
bots’ profile picture and simply changed the skin color. Using a
method that has been frequently employed to measure
discrimination in hiring, I also gave the bots characteristically
white or characteristically black first and last names.
• Here’s an example of “Greg,” a white bot:
• “Overall, I had four types of bots: High Follower/White;
Low Follower/White; High Follower/Black; and Low
Follower/Black.
• Results? Only one of the four types of bots caused a
significant reduction in the subjects’ rate of tweeting
slurs: the white bots with 500 followers.
• Overall, I found that it is possible to cause people to
use less harassing language. This change seems to be
most likely when both individuals share a social
identity. Unsurprisingly, high status people are also
more likely to cause a change.”
• Many people are already engaged in sanctioning bad
behavior online, but they sometimes do so in a way
that can backfire. If people call out bad behavior in a
way that emphasizes the social distance between
themselves and the person they’re calling out, my
research suggests that the sanctioning is less likely to
be effective.
• Physical distance, anonymity and partisan bubbles
online can lead to extremely nasty behavior, but if we
remember that there’s a real person behind every
online encounter and emphasize what we have in
common rather than what divides us, we might be
able to make the Internet a better place.
• But action can be risky – a professor and her
students created a list of fake sites and criteria
to evaluate the credibility of sites was
targeted, threatened, and doxxed, and later
took the material down.
• See the handout (“How to spot fake news”).
This is a compilation of the advice published
to help people deal with the fake news
epidemic.
• Does anything seem useful?
• Can you think of additional solutions,
strategies or tools?
• What should be done?