Connect with us


What Social Media are doing to counter misinformation in US election



Intelligence officials warned in 2018 that Russia was at it again, along with other state actors. Now in this election cycle, cybersecurity experts have also raised alarm over the increasing threat of domestic actors sowing misinformation online.

The task of policing content while avoiding the appearance of bias has been a tripwire for many of these social media giants, who have faced attacks from both sides of the political aisle for decisions to remove certain content, including allegations of censorship.

ABC News has compiled this explainer to provide readers with a guide to comparing and contrasting policy measures from some of the most-used social media platforms in the U.S. including Facebook (and Facebook-owned Instagram), Twitter, Reddit, TikTok and YouTube.


Facebook, the most-used social media platform in the U.S., took the most heat for the 2016 controversy.

In the years since 2016, Facebook’s core efforts to maintain election integrity have fallen into three major categories: Taking down inauthentic accounts and networks, tightening policies on content moderation, and unveiling an ad database with the goal of increased transparency.

Facebook also launched an Elections Operations Center in 2018, a team that it says will monitor potential democratic process abuses on the network in real-time. The company said that so far it has removed more than 120,000 pieces of content from Facebook and Instagram in the U.S. for violating voter-interference policies it has set, and displayed warnings on more than 150 pieces of content. Moreover, the company said it removed 30 networks engaged in coordinated inauthentic behavior targeting the U.S.

In August of 2020, the company unveiled a campaign to encourage people to vote and pledged to remove any content that encourages people not to vote, such as posts which state that voting requires a passport or driving license.

In the weeks ahead of the 2020 vote, the company also announced a series of last-minute changes, including banning all new political advertisements a week before the election, removing new posts with militarized language, such as “army” or “battle,” that aims to suppress voters and temporarily pausing all political ads on the site for an undisclosed period of time after the polls close on Nov. 3.

Facebook also said it will label content that seeks to delegitimize the outcome of the election, and label content from candidates or campaigns that try to declare victory before results are in — instead directing users to official results from Reuters and the National Election Pool.

Moreover, Facebook said it would start labeling some content that it doesn’t remove because it is deemed newsworthy, such as speeches from politicians.

“We’ll allow people to share this content to condemn it, just like we do with other problematic content, because this is an important part of how we discuss what’s acceptable in our society — but we’ll add a prompt to tell people that the content they’re sharing may violate our policies,” Zuckerberg said in a Facebook post at the time.

In addition, Facebook said it would remove all accounts representing the group QAnon, a baseless conspiracy theory which purports, without evidence, that Donald Trump is working in secret against a global Satanic pedophile ring. The unfounded theory was invented online shortly after the 2016 election and has made its way into the political discourse.

While these are major changes at the company compared to 2016 when many, including CEO Mark Zuckerbreg, say it was caught flat-footed, some advocates have criticized what that say is how narrow their actions surrounding political ads are.

“The policies Facebook has taken are extremely reactive,” Ben Decker, the founder of Memetica, a digital investigations consultancy firm, told ABC News. “I don’t think the measures they have taken to curb political ads are going to be particularly effective, because they have these exact stipulations.”

Dipayan Ghosh, the co-director of the Harvard Kennedy School’s digital platforms and democracy project, told ABC News that the ban on new political advertisements one week ahead of the election, “is a ban on new submittals, not on political advertising entirely” and questioned the impact of the ban when record numbers of people are voting early.

Banning political ads after polls close is also “not necessarily going to have a result on the on the election itself,” Ghosh added.

Facebook said this move was aimed to “to reduce opportunities for confusion or abuse.”

“I think many advocates would have liked to see is Facebook to extend a full ban on political advertising for a lengthy period ahead of the election, say, a month or even longer than that,” Ghosh said. “What many of us wanted to see from Facebook is a full ban, a commitment to put the democratic process over revenues.”

Facebook made nearly $70 billion in advertising revenue alone last year, according to financial disclosures.

Ghosh also expressed concerns over the way misinformation spreads on private Facebook groups, which in many cases remain largely unregulated unless they contain active calls for violence — and even then Facebook has been accused of reacting too late.

“I myself have joined groups which have amassed a big following on different sorts of issues, mostly sports related, which then all of a sudden change one day in theme from something about the New York Giants, to ‘Justice for Justice Kavanaugh,'” Ghosh said. “And you can clearly see that what’s happening here is that organizers are trying to get people into these groups, and then all of a sudden, turn a switch to try to influence the members of that group toward these kinds of political themes.”

Also concerning to many, new research from the German Marshall Digital Fund think tank published earlier this week found that more people now are engaging with outlets on Facebook that repeatedly publish verifiably false content than in the lead-up to the 2016 election.


Twitter banned all political ads worldwide in October 2019, a move that put it in stark contrast to Facebook, which at the time had recently ruled out banning political ads. Jack Dorsey, Twitter’s CEO, tweeted “while internet advertising is incredibly powerful and very effective for commercial advertisers, that power brings significant risks to politics.”

Ghosh noted that Twitter originally did not make “a lot of money off of political advertising, which likely made it an easier decision for Jack Dorsey than it would be for Mark Zuckerberg.” Twitter reported making nearly $3 billion in ad revenue in Fiscal Year 2019, according to financial disclosures.

Political ads on Twitter did not exist on the same scale as they do on Facebook, but the company has also taken a number of additional measures in recent months to show they are taking action ahead of the 2020 election. Most recently it launched what it calls a “2020 U.S. election hub” which will include a curated list of news articles, as well as live streams of debates.

As part of a suite of measures to combat misinformation, Twitter also introduced a new labelling system in May 2020 which allowed the platform to flag tweets with what it determined to be misleading content.

In the last few months, the social media platform found itself embroiled in controversy after it labeled a number of Donald Trump’s tweets, including those containing claims about mail-in-voting, as potentially misleading, It has also put labels on Trump’s tweets for violating its policies for abusive behavior as well as those regarding manipulated media. In these cases, the tweets are hidden from view but users can easily click in to see the content. Trump has accused Twitter of trying to silence conservative voices.

Critics have questioned the efficacy of the labels Twitter (and Facebook) use in actually stopping misinformation or false claims from spreading or being amplified on the platforms.

Decker noted that more research needs to be done here, but said “it’s unclear often how many of those who read the disinformation are actually reading the fact check, or the intervention response.”

Ghosh said he thinks that “these kinds of labels have a very limited, marginal impact on influencing the opinion of the people who consume that content.”

“I can’t say that these labels really resolve the core issue, which is that you’ve got a person in certain cases with a massive following, who is pushing misinformation intentionally and pushing disinformation, and trying to do so for his own political gain,” Ghosh said. “Having this sort of label does not really change the mind of anyone who’s consuming it.”

A study published in March by researchers at the Massachusetts Institute of Technology suggested that selective labeling of false news can actually have detrimental effect, dubbed the “implied-truth effect,” where unmarked and unchecked, yet still demonstrably false, content appears more legitimate.

The strongest weapon Twitter has to prevent the spread of political misinformation is the removal of tweets and the restriction of accounts, but the platform utilizes these sparingly, likely to avoid being accused of censorship. The most high profile example of this was when it restricted Donald Trump Jr.’s account in late July after he shared a video featuring doctors making false claims about the coronavirus, including that masks are unnecessary. Trump Jr.’s account was suspended for 12 hours, meaning he was unable to tweet, and it removed the video from public view.

Last week, Twitter also unveiled a slew of new updates aimed specifically at curbing the spread of misinformation on the platform ahead of the election, including efforts to stop tweets with misleading information from going viral and a policy that will not allow any person, including candidates for office, to claim an election win before it was authoritatively called.

Significantly, users will not be able to retweet or reply to tweets “with a misleading information label from U.S. political figures (including candidates and campaign accounts), U.S.-based accounts with more than 100,000 followers, or that obtain significant engagement.” Users will, however, be able to quote-tweet the messages, although they will have to click through a warning in order to see these labeled tweets in the first place.

When users attempt to retweet, they will be prompted to Quote Tweet (add their own commentary) instead.

“Though this adds some extra friction for those who simply want to Retweet, we hope it will encourage everyone to not only consider why they are amplifying a Tweet, but also increase the likelihood that people add their own thoughts, reactions and perspectives to the conversation,” the company said in a blogpost.


The video-sharing giant announced earlier this year some updates to how it was preparing for the election, saying it would remove election-related content that violated its Community Guidelines.

“These policies prohibit hate speech, harassment, and deceptive practices, including content that aims to mislead people about voting or videos that are technically manipulated or doctored in a way that misleads users (beyond clips taken out of context) and may pose a serious risk of egregious harm,” the company said.

The company also said it would remove content that contains hacked information, stating, “For example, videos that contain hacked information about a political candidate shared with the intent to interfere in an election.”

Similar to other platforms, YouTube also pledged to remove content encouraging users to interfere with the democratic process, citing an example as content “telling viewers to create long voting lines with the purpose of making it harder for others to vote.”

Some have expressed concern that YouTube (similar to Reddit) has not yet published a clear policy on how it will handle candidates claiming victory before the election is officially called.

Decker called YouTube’s policies “extremely reactive” overall.

“Oftentimes, they will apply key word filters to prevent content from being found in search, YouTube’s biggest claim is they incorporate Wikipedia pages into knowledge panels, so if it’s a video about COVID-19, regardless of where it’s from, there would also be this knowledge panel or factcheck above the description that points you toward accurate sources,” Decker said. The Wikipedia articles, while volunteer-edited, provide at least some context to content that would otherwise not have any.

“While the problem on YouTube is still bad, it’s now much less worse,” Decker said.

He also noted that their three strikes policy has been effective in booting a number of content creators off the platform, but an unintended consequence is that this has “led to the rise of fringe platforms.”

“It’s tricky because in one sense it does clean up the stream in the short term, so on the one hand it creates healthier conversations, but it moves them to another area of the internet, which is even more unregulated but there are even less dissenting views, so it’s a space where people can be radicalized,” he said.

Notably, YouTube announced in a company blogpost earlier this week it was taking new steps to curb hate by “removing more conspiracy content used to justify real-world violence.”

Specifically, YouTube cited QAnon as an example of entity “that targets an individual or group with conspiracy theories that have been used to justify real-world violence.”

“As always, context matters, so news coverage on these issues or content discussing them without targeting individuals or protected groups may stay up,” the blogpost added. “We will begin enforcing this updated policy today, and will ramp up in the weeks to come.”


In April 2020, Reddit announced that it was launching a subreddit dedicated to political transparency, which would list all political ad campaigns running on Reddit dating back to January of 2019. The company said this subreddit would give information on the individual advertiser, their targeting, impressions, and spend on a per-campaign basis. As an additional transparency measure, Reddit said it would require political advertisers to leave comments “on” for the first 24 hours of a campaign to enable them to “engage directly with users in the comments.”

While the political transparency subreddit contains significant details about political ads, it has a limited reach, with around 3,000 members since it was launched five months ago. It is also worth noting that Reddit doesn’t allow political ads in other countries.

In June 2020, Reddit also announced that it was banning a number of subreddits which it said violated company policies on hate speech. Included in these was r/TheDonald, a pro-Trump subreddit which was popular in the run-up to the 2016 election but which had been largely inactive for months despite its nearly 800,000 members. Members of this subreddit had already migrated to another platform the year before, in response to stricter content rules and increased moderation. The banning of this subreddit and others was indicative of the problems facing social media platforms, where measures to combat hate speech or misinformation do not keep pace with the dissemination of such material.

Reddit, however, noticeably has no stated policy on candidates claiming victory in the election before it is authoritatively called.


While the Chinese-owned video sharing app avoided the level of misinformation scrutiny leveled at platforms like Facebook and Twitter, it has taken a number of actions in recent months to show that it is taking a stand before the election.

Along with banning political ads, in August TikTok also announced a suite of news measures to combat misinformation ahead of the 2020 presidential election. Crucially, it banned manipulated media which it said “misleads users by distorting the truth of events in a way that could cause harm.” This included deepfakes, synthetic media produced by artificial intelligence which has the appearance of being real.

Despite these measures, political content on TikTok, like all social media platforms, is extremely popular. Videos containing the hashtag #Trump2020 have been viewed 10.3 billion times by September 2020, according to data on the app. A report from the Wall Street Journal late last year claimed that the Trump campaign had reached out to TikTok accounts with large supportive followings, including some with the Trump 2020 flag in their videos.

Source:- ABC News

Source link


Kyrie skips media availability, releases statement – theScore



As the NBA prepares for its 2020-21 season to begin on Dec. 22, the enigmatic Kyrie Irving did not provide access to reporters on Friday as part of the Brooklyn Nets‘ media week.

Instead, the 28-year-old guard released a written statement “to ensure that (his) message is conveyed properly,” per ESPN’s Malika Andrews.

Here’s Irving’s statement in full:

COVID-19 has impacted us all in many ways, so I pray for the safety and health of our communities domestically and abroad. I am truly excited for the season to start and I am also praying that everyone remains safe and healthy throughout this journey.

Instead of speaking to the media today, I am issuing this statement to ensure that my message is conveyed properly.

I am committed to show up to work everyday, ready to have fun, compete, perform, and win championships alongside my teammates and colleagues in the Nets organization. My goal this season is to let my work on and off the court speak for itself.

Life hit differently this year and it requires us, it requires me, to move differently. So, this is the beginning of that change.

The league’s collective bargaining agreement requires veteran players to “participate in photograph and media sessions” as early as the 22nd day prior to the first game of the regular season.

Additionally, under a section pertaining to promotional activities, the CBA states: “Upon request, the player shall consent to and make himself available for interviews by representatives of the media conducted at reasonable times.”

It’s unclear whether Irving’s statement constitutes participation in the Nets’ training camp media session or whether certain allowances have been made in regard to media availability within the amended agreement due to COVID-19.

In an injury-shortened debut season with the Nets in 2019-20, Irving averaged 27.4 points, 6.4 assists, 5.2 rebounds, and 1.4 steals per game across 20 appearances.

Let’s block ads! (Why?)

Source link

Continue Reading


Conversations That Matter: The state of the media – Vancouver Sun



Article content

Jack Webster was known as the king of the airways in B.C> for close to 40 years. Since his retirement the foundation in his name has been recognizing and celebrating excellence in journalism.

This year, on Dec. 8, the awards dinner is being netcast, opening it up, for the first time, to the public. 

If it was on the public agenda, Webster was there. When prisoners at the B.C. Penitentiary rioted and took hostages in 1963, they asked Webster to resolve the standoff. They asked for him because he was trusted at a time when mainstream media was believed to be fulfilling the responsibilities for the fourth estate – to step in, stand up, advocate, call out, and record the people and events of our lives.

Since Webster’s retirement in 1988, the media landscape has changed dramatically. In Vancouver, for example, the major powerhouses in print, radio and TV have all seen their constituencies dwindle. Shrinking audiences meant less ad revenue, which, in turn, led to cuts in newsrooms and that leads to further reductions in audiences.

Let’s block ads! (Why?)

Source link

Continue Reading


Kyrie Irving Imposes ‘Media Blackout,’ Won’t Speak To Reporters This Season – Forbes



Kyrie Irving was tentatively scheduled to do a Zoom interview with reporters on Friday, just as Kevin Durant and his other Nets teammates have been doing in recent days.

Now it appears Irving won’t speak to the media at all this season. ESPN’s Brian Windhorst reported on “The Jump” that Irving will conduct a “media blackout” this season, and Irving issued this statement.

Irving, the former St. Patrick (N.J.) High School and Duke star, has had some missteps with the media in the past.

In 2017, when he was with the Cleveland Cavaliers, he claimed that the Earth was flat.

It became a major story and even NBA Commissioner Adam Silver — who, like Irving, went to Duke — was asked to offer his opinion.

“Kyrie and I went to the same college,” Silver said then. “He may have taken some different courses.”

Irving later ended up apologizing for his comments.

“To all the science teachers, everybody coming up to me like, `You know I’ve got to reteach my whole curriculum?’ I’m sorry,” Irving said. “I apologize. I apologize.”

In October, Irving made some controversial comments about new Nets coach Steve Nash.

“I don’t really see us having a head coach,” Irving said on the podcast, “The ETCs With Kevin Durant.” Referring to Durant, he said, “K.D. could be a head coach. I could be a head coach.”

Speaking this week on a Zoom with reporters, Nash tried to diffuse the situation.

“I read what he said, and I think it was one phrase at the end of a bunch of things he said about being excited — about having me in this position and coaching — and then maybe taken to another level that seemed incredible in headlines, which is fine,” Nash said. “I’m in a fortunate position where I get to coach Kevin Durant and Kyrie Irving. What we’ve dealt here in a short period of time with how we’re going to prep and play is exciting. I’m thrilled to get to coach those guys.

“One statement that I don’t think necessarily was completely — wasn’t meant the way it was taken by the press, that doesn’t bother me at all.”

Without using the words “load management,” Nash also said he would be surprised if Durant or Irving played all 72 games this season. Durant is coming off Achilles surgery, and Irving off shoulder surgery.

Through it all, Irving has continued to financially support his alma mater, now called The Patrick School. In 2018, he funded the renovation of the gym, locker room, weight room and lounge at the school.

He has also mentored younger players, like Seton Hall guard Bryce Aiken and Jonathan Kuminga, the former Patrick School star now with the G League Ignite team, and pledged $1.5 million to WNBA players forgoing last season due to coronavirus or social justice concerns.

Irving, Durant and the Nets open the preseason Dec. 13 against the Wizards and the regular season Dec. 22 at home against the Warriors.

Let’s block ads! (Why?)

Source link

Continue Reading