Remove election-related content that violates our policies
Our Community Guidelines provide clear guidance on content that is not allowed on YouTube. One common question we get is how these policies apply to election-related content. Here are some examples of where our established deceptive practices policies apply:
- Content that has been technically manipulated or doctored in a way that misleads users (beyond clips taken out of context) and may pose a serious risk of egregious harm; for example, a video that has been technically manipulated to make it appear that a government official is dead.
- Content that aims to mislead people about voting or the census processes, like telling viewers an incorrect voting date.
- Content that advances false claims related to the technical eligibility requirements for current political candidates and sitting elected government officials to serve in office, such as claims that a candidate is not eligible to hold office based on false information about citizenship status requirements to hold office in that country.
Additionally, we terminate channels that:
- Attempt to impersonate another person or channel, misrepresent their country of origin, or conceal their association with a government actor.
- Artificially increase the number of views, likes, comments, or other metric either through the use of automatic systems or by serving up videos to unsuspecting viewers.
As always, we enforce our policies consistently, without regard to a video’s political viewpoint.
The best way to quickly remove content is to stay ahead of new technologies and tactics that could be used by malicious actors, including technically-manipulated content. We also heavily invest in research and development. In 2018, we formed an Intelligence Desk to detect new trends surrounding inappropriate content and problematic behaviors, and to make sure our teams are prepared to address them before they become a larger issue. For example, in 2018, as a result of the Intelligence Desk’s work to detect the evolving online tactics and impending statements of terrorist organizations, we shared 100,000 digital fingerprints (also known as hashes) of terror content to the Global Internet Forum to Counter Terrorism’s hash-sharing database.
To combat foreign and domestic coordinated influence operations looking to interfere in electoral processes, we coordinate closely with Google’s Threat Analysis Group (TAG) to identify bad actors and terminate their channels and accounts. Through TAG, we work with other technology companies to share intelligence and best practices, and share threat information with law enforcement.
Raise up authoritative election news
Political news and events can be subject to misinformation, so the availability of quality information sources is crucial. That’s why we raise up authoritative voices, including news sources, for news and information in search results and “watch next” panels. Millions of search queries and recommendations are getting this authoritative ranking treatment today, and we’re continuing to improve and expand our systems.
Additionally, we introduced Top News and Breaking News shelves to highlight quality journalism, as well as information panels that indicate funding sources below videos from publishers that receive public or government funding. During breaking news events, we provide short previews of text-based news articles in search results, along with a reminder that developing news can rapidly change. Because of all these efforts, last year we saw consumption of content from authoritative news partners’ channels grow by 60%.
For the 2018 U.S. midterm and the 2019 EU Parliamentary elections, when users searched for a candidate on YouTube, we surfaced an information panel with additional information about that candidate—for example, party affiliation and district—above search results. We also highlighted the official YouTube channels of candidates when available. We will have a similar feature for candidates in the 2020 U.S. elections in the coming months.
Reduce the spread of election misinformation
Content that comes close to violating our Community Guidelines is a fraction of 1% of what’s watched on YouTube in the U.S. To reduce this even further, in January 2019, we launched changes to our recommendations systems to limit the spread of harmful misinformation and borderline content. The result is a 70% average drop in watch time of this content coming from non-subscribed recommendations in the U.S. These changes are now implemented in the U.S. and other English-language markets, and we’ve begun expanding this effort to non-English-language markets, helping us reduce recommendations of borderline content, including election-related misinformation, around the world.
Recognize and reward campaigns, candidates, and political creators
Politicians, commentators and news publishers can access a suite of YouTube features and resources, including support from our partnerships team. These specialists work with news organizations, political creators and candidates on both sides of the aisle to optimize their presence on YouTube, helping them more effectively reach viewers, engage their community and keep their accounts secure.
In addition to our work at YouTube to support elections, other teams at Google are also working hard ahead of Election Day—including by expanding our political advertising policies to provide users with even more visibility into who is buying election ads on YouTube, Google, and partner properties. YouTube remains committed to maintaining the balance of openness and responsibility, before, during and after the 2020 U.S. election. We’ll have even more to share on this work in the coming months.
Leslie Miller, VP of Government Affairs & Public Policy, YouTube