Connect with us


What social media giants are doing to counter misinformation this election – ABC News



Intelligence officials warned in 2018 that Russia was at it again, along with other state actors. Now in this election cycle, cybersecurity experts have also raised alarm over the increasing threat of domestic actors sowing misinformation online.

The task of policing content while avoiding the appearance of bias has been a tripwire for many of these social media giants, who have faced attacks from both sides of the political aisle for decisions to remove certain content, including allegations of censorship.

ABC News has compiled this explainer to provide readers with a guide to comparing and contrasting policy measures from some of the most-used social media platforms in the U.S. including Facebook (and Facebook-owned Instagram), Twitter, Reddit, TikTok and YouTube.


Facebook, the most-used social media platform in the U.S., took the most heat for the 2016 controversy.

In the years since 2016, Facebook’s core efforts to maintain election integrity have fallen into three major categories: Taking down inauthentic accounts and networks, tightening policies on content moderation, and unveiling an ad database with the goal of increased transparency.

Facebook also launched an Elections Operations Center in 2018, a team that it says will monitor potential democratic process abuses on the network in real-time. The company said that so far it has removed more than 120,000 pieces of content from Facebook and Instagram in the U.S. for violating voter-interference policies it has set, and displayed warnings on more than 150 pieces of content. Moreover, the company said it removed 30 networks engaged in coordinated inauthentic behavior targeting the U.S.

In August of 2020, the company unveiled a campaign to encourage people to vote and pledged to remove any content that encourages people not to vote, such as posts which state that voting requires a passport or driving license.

In the weeks ahead of the 2020 vote, the company also announced a series of last-minute changes, including banning all new political advertisements a week before the election, removing new posts with militarized language, such as “army” or “battle,” that aims to suppress voters and temporarily pausing all political ads on the site for an undisclosed period of time after the polls close on Nov. 3.

Facebook also said it will label content that seeks to delegitimize the outcome of the election, and label content from candidates or campaigns that try to declare victory before results are in — instead directing users to official results from Reuters and the National Election Pool.

Moreover, Facebook said it would start labeling some content that it doesn’t remove because it is deemed newsworthy, such as speeches from politicians.

“We’ll allow people to share this content to condemn it, just like we do with other problematic content, because this is an important part of how we discuss what’s acceptable in our society — but we’ll add a prompt to tell people that the content they’re sharing may violate our policies,” Zuckerberg said in a Facebook post at the time.

In addition, Facebook said it would remove all accounts representing the group QAnon, a baseless conspiracy theory which purports, without evidence, that Donald Trump is working in secret against a global Satanic pedophile ring. The unfounded theory was invented online shortly after the 2016 election and has made its way into the political discourse.

While these are major changes at the company compared to 2016 when many, including CEO Mark Zuckerbreg, say it was caught flat-footed, some advocates have criticized what that say is how narrow their actions surrounding political ads are.

“The policies Facebook has taken are extremely reactive,” Ben Decker, the founder of Memetica, a digital investigations consultancy firm, told ABC News. “I don’t think the measures they have taken to curb political ads are going to be particularly effective, because they have these exact stipulations.”

Dipayan Ghosh, the co-director of the Harvard Kennedy School’s digital platforms and democracy project, told ABC News that the ban on new political advertisements one week ahead of the election, “is a ban on new submittals, not on political advertising entirely” and questioned the impact of the ban when record numbers of people are voting early.

Banning political ads after polls close is also “not necessarily going to have a result on the on the election itself,” Ghosh added.

Facebook said this move was aimed to “to reduce opportunities for confusion or abuse.”

“I think many advocates would have liked to see is Facebook to extend a full ban on political advertising for a lengthy period ahead of the election, say, a month or even longer than that,” Ghosh said. “What many of us wanted to see from Facebook is a full ban, a commitment to put the democratic process over revenues.”

Facebook made nearly $70 billion in advertising revenue alone last year, according to financial disclosures.

Ghosh also expressed concerns over the way misinformation spreads on private Facebook groups, which in many cases remain largely unregulated unless they contain active calls for violence — and even then Facebook has been accused of reacting too late.

“I myself have joined groups which have amassed a big following on different sorts of issues, mostly sports related, which then all of a sudden change one day in theme from something about the New York Giants, to ‘Justice for Justice Kavanaugh,'” Ghosh said. “And you can clearly see that what’s happening here is that organizers are trying to get people into these groups, and then all of a sudden, turn a switch to try to influence the members of that group toward these kinds of political themes.”

Also concerning to many, new research from the German Marshall Digital Fund think tank published earlier this week found that more people now are engaging with outlets on Facebook that repeatedly publish verifiably false content than in the lead-up to the 2016 election.


Twitter banned all political ads worldwide in October 2019, a move that put it in stark contrast to Facebook, which at the time had recently ruled out banning political ads. Jack Dorsey, Twitter’s CEO, tweeted “while internet advertising is incredibly powerful and very effective for commercial advertisers, that power brings significant risks to politics.”

Ghosh noted that Twitter originally did not make “a lot of money off of political advertising, which likely made it an easier decision for Jack Dorsey than it would be for Mark Zuckerberg.” Twitter reported making nearly $3 billion in ad revenue in Fiscal Year 2019, according to financial disclosures.

Political ads on Twitter did not exist on the same scale as they do on Facebook, but the company has also taken a number of additional measures in recent months to show they are taking action ahead of the 2020 election. Most recently it launched what it calls a “2020 U.S. election hub” which will include a curated list of news articles, as well as live streams of debates.

As part of a suite of measures to combat misinformation, Twitter also introduced a new labelling system in May 2020 which allowed the platform to flag tweets with what it determined to be misleading content.

In the last few months, the social media platform found itself embroiled in controversy after it labeled a number of Donald Trump’s tweets, including those containing claims about mail-in-voting, as potentially misleading, It has also put labels on Trump’s tweets for violating its policies for abusive behavior as well as those regarding manipulated media. In these cases, the tweets are hidden from view but users can easily click in to see the content. Trump has accused Twitter of trying to silence conservative voices.

Critics have questioned the efficacy of the labels Twitter (and Facebook) use in actually stopping misinformation or false claims from spreading or being amplified on the platforms.

Decker noted that more research needs to be done here, but said “it’s unclear often how many of those who read the disinformation are actually reading the fact check, or the intervention response.”

Ghosh said he thinks that “these kinds of labels have a very limited, marginal impact on influencing the opinion of the people who consume that content.”

“I can’t say that these labels really resolve the core issue, which is that you’ve got a person in certain cases with a massive following, who is pushing misinformation intentionally and pushing disinformation, and trying to do so for his own political gain,” Ghosh said. “Having this sort of label does not really change the mind of anyone who’s consuming it.”

A study published in March by researchers at the Massachusetts Institute of Technology suggested that selective labeling of false news can actually have detrimental effect, dubbed the “implied-truth effect,” where unmarked and unchecked, yet still demonstrably false, content appears more legitimate.

The strongest weapon Twitter has to prevent the spread of political misinformation is the removal of tweets and the restriction of accounts, but the platform utilizes these sparingly, likely to avoid being accused of censorship. The most high profile example of this was when it restricted Donald Trump Jr.’s account in late July after he shared a video featuring doctors making false claims about the coronavirus, including that masks are unnecessary. Trump Jr.’s account was suspended for 12 hours, meaning he was unable to tweet, and it removed the video from public view.

Last week, Twitter also unveiled a slew of new updates aimed specifically at curbing the spread of misinformation on the platform ahead of the election, including efforts to stop tweets with misleading information from going viral and a policy that will not allow any person, including candidates for office, to claim an election win before it was authoritatively called.

Significantly, users will not be able to retweet or reply to tweets “with a misleading information label from U.S. political figures (including candidates and campaign accounts), U.S.-based accounts with more than 100,000 followers, or that obtain significant engagement.” Users will, however, be able to quote-tweet the messages, although they will have to click through a warning in order to see these labeled tweets in the first place.

When users attempt to retweet, they will be prompted to Quote Tweet (add their own commentary) instead.

“Though this adds some extra friction for those who simply want to Retweet, we hope it will encourage everyone to not only consider why they are amplifying a Tweet, but also increase the likelihood that people add their own thoughts, reactions and perspectives to the conversation,” the company said in a blogpost.


The video-sharing giant announced earlier this year some updates to how it was preparing for the election, saying it would remove election-related content that violated its Community Guidelines.

“These policies prohibit hate speech, harassment, and deceptive practices, including content that aims to mislead people about voting or videos that are technically manipulated or doctored in a way that misleads users (beyond clips taken out of context) and may pose a serious risk of egregious harm,” the company said.

The company also said it would remove content that contains hacked information, stating, “For example, videos that contain hacked information about a political candidate shared with the intent to interfere in an election.”

Similar to other platforms, YouTube also pledged to remove content encouraging users to interfere with the democratic process, citing an example as content “telling viewers to create long voting lines with the purpose of making it harder for others to vote.”

Some have expressed concern that YouTube (similar to Reddit) has not yet published a clear policy on how it will handle candidates claiming victory before the election is officially called.

Decker called YouTube’s policies “extremely reactive” overall.

“Oftentimes, they will apply key word filters to prevent content from being found in search, YouTube’s biggest claim is they incorporate Wikipedia pages into knowledge panels, so if it’s a video about COVID-19, regardless of where it’s from, there would also be this knowledge panel or factcheck above the description that points you toward accurate sources,” Decker said. The Wikipedia articles, while volunteer-edited, provide at least some context to content that would otherwise not have any.

“While the problem on YouTube is still bad, it’s now much less worse,” Decker said.

He also noted that their three strikes policy has been effective in booting a number of content creators off the platform, but an unintended consequence is that this has “led to the rise of fringe platforms.”

“It’s tricky because in one sense it does clean up the stream in the short term, so on the one hand it creates healthier conversations, but it moves them to another area of the internet, which is even more unregulated but there are even less dissenting views, so it’s a space where people can be radicalized,” he said.

Notably, YouTube announced in a company blogpost earlier this week it was taking new steps to curb hate by “removing more conspiracy content used to justify real-world violence.”

Specifically, YouTube cited QAnon as an example of entity “that targets an individual or group with conspiracy theories that have been used to justify real-world violence.”

“As always, context matters, so news coverage on these issues or content discussing them without targeting individuals or protected groups may stay up,” the blogpost added. “We will begin enforcing this updated policy today, and will ramp up in the weeks to come.”


In April 2020, Reddit announced that it was launching a subreddit dedicated to political transparency, which would list all political ad campaigns running on Reddit dating back to January of 2019. The company said this subreddit would give information on the individual advertiser, their targeting, impressions, and spend on a per-campaign basis. As an additional transparency measure, Reddit said it would require political advertisers to leave comments “on” for the first 24 hours of a campaign to enable them to “engage directly with users in the comments.”

While the political transparency subreddit contains significant details about political ads, it has a limited reach, with around 3,000 members since it was launched five months ago. It is also worth noting that Reddit doesn’t allow political ads in other countries.

In June 2020, Reddit also announced that it was banning a number of subreddits which it said violated company policies on hate speech. Included in these was r/TheDonald, a pro-Trump subreddit which was popular in the run-up to the 2016 election but which had been largely inactive for months despite its nearly 800,000 members. Members of this subreddit had already migrated to another platform the year before, in response to stricter content rules and increased moderation. The banning of this subreddit and others was indicative of the problems facing social media platforms, where measures to combat hate speech or misinformation do not keep pace with the dissemination of such material.

Reddit, however, noticeably has no stated policy on candidates claiming victory in the election before it is authoritatively called.


While the Chinese-owned video sharing app avoided the level of misinformation scrutiny leveled at platforms like Facebook and Twitter, it has taken a number of actions in recent months to show that it is taking a stand before the election.

Along with banning political ads, in August TikTok also announced a suite of news measures to combat misinformation ahead of the 2020 presidential election. Crucially, it banned manipulated media which it said “misleads users by distorting the truth of events in a way that could cause harm.” This included deepfakes, synthetic media produced by artificial intelligence which has the appearance of being real.

Despite these measures, political content on TikTok, like all social media platforms, is extremely popular. Videos containing the hashtag #Trump2020 have been viewed 10.3 billion times by September 2020, according to data on the app. A report from the Wall Street Journal late last year claimed that the Trump campaign had reached out to TikTok accounts with large supportive followings, including some with the Trump 2020 flag in their videos.

Let’s block ads! (Why?)

Source link


Media Beat: October 19, 2020 | FYIMusicNews – FYI Music News



From Ad Contrarian Bob Hoffman’s latest newsletter

We haven’t heard much from the “TV Is Dead” imbeciles recently, so I thought a little update would be appropriate. First some background.

The concept that the TV-Is-Deadheads never quite seemed to understand is that there’s a difference between consumer behaviour and industry shit fights. The fact that broadcast tv and cable tv and satellite tv and internet-delivered tv were fighting over share meant less than nothing to consumers. Consumers like to sit on their asses and watch television. That’s all there is to know about the subject.

But the Deadheads saw the eroding share of broadcast tv and knee-jerked that into “TV Is Dying.” Whether the signal gets to peoples’ tv sets by electromagnetic waves, underground wires, satellite pulses, web streaming, carrier pigeons or rowboats is of no interest to them. As long as it’s simple to use, entertaining, and cheap, they’ll watch (Although advertising has become so horrible lately people are willing to pay way more than imagined to avoid it).

Even in an environment in which streaming is gobbling up share, over the air broadcast is still dominating. According to Nielsen…

   – The average adult spent 4 hours and 30 minutes a day watching traditional tv in Q2 2020.
   – The average adult spent 1 hour and 6 minutes on streaming. The growth of streaming has been quite substantial and impressive, but it still constitutes only 25% of video viewing.
   – Remarkably, streaming services (Amazon Prime, Apple TV, Disney+, HBO, Hulu, and Sling) bought over a quarter billion dollars of ad time on traditional tv in the past 12 months to promote their products. That’s gotta tell you something.

Regardless of who wins the internecine battle of delivery systems, one thing is clear. People love tv and reports of its death have been greatly exaggerated.  h/t Lara Bracken

Doug Ford gives thumbs up to media in a question about covid being a hoax

[embedded content]

And now, Jonathan Pie tells it like it is from the UK

[embedded content]

Postmedia financials look bleak

A $21M federal wage subsidy program helped pump Postmedia’s balance sheet to a $13.5M profit in Q4, but its annual report saw its deficit increase to $16.2M as of Aug. 31, compared with $6.3M in fiscal 2019. Revenue, mostly from advertising, print circulation and digital services, dropped nearly 18 percent for the full year to $508.4M. – The Canadian Press

Former prostitute and Edmonton radio host makes film festival debut

Illusion: The Fear is the story of Valécia Pepin’s harrowing struggle to free herself from the grip of her pimp, who attempts to reassert control. It had its debut at the Edmonton Short Film Festival on the weekend. – Andrea Huncar, CBC Edmonton

Let’s block ads! (Why?)

Source link

Continue Reading


Thai police order media probes over protest coverage – Reuters Canada



BANGKOK (Reuters) – Thai police said on Monday they had ordered an investigation of four news outlets under emergency measures imposed last week to try to stop three months of protests against the government and monarchy.

FILE PHOTO: Pro-democracy protesters shine their mobile phone lights during an anti-government protest, in Bangkok, Thailand October 18, 2020. REUTERS/Athit Perawongmetha

The announcement prompted anger from media groups and accusations of an attack on press freedom by the government of Prime Minister Prayuth Chan-ocha, the former junta leader the protesters are seeking to drive from office.

According to a police document dated Oct. 16, investigations have been ordered into content from four media outlets as well as the Facebook page of a protest group.

“We received information from intelligence units concerned that parts of the content and distorted information have been used and disseminated to cause confusion and instigate causing unrest to society,” police spokesman Kissana Phathanacharoen told a news conference.

He said it was for Thailand’s broadcast regulator and digital ministry to investigate and take appropriate action, adding that there was no plan to curb press freedom.

Putchapong Nodthaisong, a spokesman for the digital ministry, said it had requested court orders to take down content by the four media outlets and a protest page, among more than 300,000 pieces of content it said violated Thai laws.

Prachatai, an independent outlet among those being investigated, described it as a censorship order.

“Honored to report accurate info about human rights and political development in Thailand, we’ll try our best in continuing to do so,” Prachathai English said on Twitter.

The Manushya Foundation, an independent group which campaigns for online freedom, called the measures an attempt to silence free media.

“Since the ban on protests did not work, the military-backed government hopes to create fear of telling the truth,” its director Emilie Palamy Pradichit said.

“We urge free media to resist.”

The government ordered a ban on news and online information that could affect national security last Thursday as it also banned political gatherings of more than five people in the face of the growing challenge.

Protests have taken place every day since then, the latest drawing tens of thousands of people in Bangkok and across the country. Police gave a figure of 20,000 protesters in the capital.

“We will prosecute everyone,” deputy Bangkok police chief Piya Tawichai said, adding that 74 protesters had been arrested since Oct. 13.

Protesters seek the removal of Prime Minister Prayuth, accusing him of engineering last year’s election to keep hold of power he first seized in a 2014 coup. He says the election was fair.

The protesters have also grown more vocal in demanding reforms to the monarchy to reduce the powers of King Maha Vajiralongkorn. The Royal Palace has made no comment on the protests or protesters’ demands.

Protesters have said they will demonstrate every day, but had yet to set out a plan for protests on Monday.

Prayuth has said he will not quit. Speaking at Government House on Monday, Prayuth said he supported a proposal for a special parliament session to discuss the situation. His supporters have a majority in parliament.

“We are just asking people not to do wrong and destroy the government and people’s property,” he said. “What the government needs to do is to protect the monarchy.”

Writing by Matthew Tostevin; Editing by Michael Perry

Let’s block ads! (Why?)

Source link

Continue Reading


Thai police seek probe of media over protests: document – The Guardian



BANGKOK (Reuters) – Thai police are seeking an investigation of four media outlets and a protest group over their coverage of protests on suspicion of breaking emergency measures restricting information, according to a document published by local media on Monday.

“It appears that there was content that affected state security, peace and order, or the good morals of the people,” the Oct. 16 document quoted police as saying.

Police would request Thailand’s broadcast regulator and digital ministry to investigate the outlets “and suspend broadcasting or remove computer information,” the statement said.

Police said there would be a briefing on the subject, but made no immediate comment.

(Reporting by Patpicha Tanakasempipat; Writing by Matthew Tostevin)

Let’s block ads! (Why?)

Source link

Continue Reading