The deployment of disinformation has long been part of Soviet and Russian military strategy. During the Cold War, the Soviet Union saw the West’s open media environment as an ideal space to undermine confidence in Western government institutions and the West’s governing model.16 Ivan Agayants, a general for the KGB—the former Russian secret police and intelligence agency—who political scientist Thomas Rid describes as “the KGB’s grandmaster of dezinformatsiya,” commented in 1965 that “if they [the West] did not have press freedom, we would have to invent it for them.”17
Soviet intelligence sought to amplify failings within the United States, such as in the area of civil rights, and pushed fake stories and conspiracy theories to discredit the West, including the claim that the CIA was responsible for inventing AIDs.18 According to Rid, by the 1960s, “disinformation—or active measures—were well-resourced and nearly on par with [intelligence] collection in the KGB … The Cold War saw more than 10,000 individual Soviet bloc disinformation operations.”19 While Soviet dezinformatsiya efforts were vast, the relative concentration of the news media and the slower pace of the news cycle limited its effectiveness. The information and communication revolution of the past two decades, however, has transformed the information domain and created new access points for the Kremlin to conduct disinformation operations.
Putin was slow to recognize the disruptive power of the internet, focusing his energy after coming to power in 2000 on exerting control over Russian television and newspapers. However, the role social media played in helping organize protests in Moscow and St. Petersburg in 2011 and 2012 led to a change in the Kremlin’s approach.20 As the Kremlin woke to the disruptive power of the internet and social media, it sought to exert control and worked to develop its capacity to disrupt the ability of opponents of the Kremlin to organize and communicate online. Time reported that after being re-elected in 2012, “Putin dispatched his newly installed head of military intelligence, Igor Sergun, to begin repurposing cyberweapons previously used for psychological operations in war zones for use in electioneering.”21 A report from the Institute of Modern Russia explains:
If at the advent of the Internet age, online activity was seen as essentially politically liberating, a censorship-busting tool that would undermine authoritarian regimes, it is quickly turning into a weapon for postmodern dictatorships like the Kremlin’s, which rely more on manipulating societies from inside than on direct oppression. The underlying mindset … [is] the idea that “truth” is a lost cause and that reality is essentially malleable.22
As the Kremlin disrupted its domestic opposition online, it also learned tactics, techniques, and procedures that it would use to inform its disinformation operations against Europe and the United States.
In recent years, information operations have become an ever more important part of Russian military strategy. In 2013, the chief of the general staff of the Russian army, Valery Gerasimov, stated as much when he said that the use of nonmilitary tools such as disinformation had become as important if not more important than more traditional military means.23 Following the Maidan Revolution in Ukraine, Russia followed a hybrid warfare doctrine, in which information operations played a critical role. Russia used “little green men”—Russian forces without insignia—to occupy Crimea and instigate an uprising in eastern Ukraine and used information operations both to deny its involvement and to advance so-called alternative, or fake, stories to attack Ukraine and mask its military intervention.24 When Malaysia Airlines Flight MH17 was shot down in Ukraine with advanced Russian anti-aircraft weaponry provided by the Kremlin to poorly trained Russian-backed rebels, Russia immediately went on a disinformation offensive, blaming the CIA or Ukrainian forces.25
A declassified U.S. intelligence community (IC) report on Russia’s activities during the 2016 U.S. election concluded that “Moscow’s campaign aimed at the US election reflected years of investment in its capabilities, which Moscow has honed in the former Soviet states.”26 This was underscored in February when Russian Defense Minister Sergey Shoigu acknowledged in an address to the Russian Duma that the Russian military operates a “cyber army” of 1,000 operatives at a cost of $300 million annually.27 Commenting on Shoigu’s address, Gen. Yuri Baluyevsky, the former Russian military commander-in-chief, said that success in information warfare “can be much more important than victory in a classical military conflict, because it is bloodless, yet the impact is overwhelming and can paralyze all of the enemy state’s power structures.”28
While information operations have become a key part of Russian military strategy, the tools that the Kremlin uses to conduct these operations extend well beyond its military. Russian information operations are often integrated whole-of-Kremlin efforts, using Russian intelligence and espionage capabilities, criminal networks of cyberhackers, official Russian media networks, and social media users or trolls paid by Kremlin-linked oligarchs. According to an IC assessment, Russian influence campaigns are “designed to be deniable because they use a mix of agents of influence, cutouts, front organizations, and false-flag operations.”29 As such, these operations utilize tactics that blend covert and overt operations, creating a very complex and sophisticated disinformation system that works on multiple, mutually reinforcing levels.
Russian information operations can be broken down into three lines of effort: covert; semi-covert; and overt.30 While distinct, each of these efforts feeds into and strengthens the others.
Covert: Spy, hack, steal, and launder
What makes Russian disinformation operations incredibly effective is that Russia uses its immense espionage capabilities in the service of its information operations.
The Russian intelligence services employ highly advanced information gathering tools—tools also used by U.S. intelligence agencies.31 Just as the U.S. National Security Agency has the ability to monitor and capture electronic communications, so does Russia. Additionally, Russia’s intelligence services—the SVR, or foreign intelligence, and the GRU, or military intelligence—are also very adept at espionage and collecting human intelligence, just like the United States’ CIA or the United Kingdom’s MI6. While Russia’s intelligence prowess is well-known, it is the Kremlin’s willingness to use information gained through intelligence means for information operations that makes these operations unique and so effective. The Russians are willing to deploy information gained through espionage in a way from which other governments have largely shied away.
Russia’s interference in the U.S. election and its willingness to use espionage tools for information operations was foreshadowed by the Kremlin’s 2014 release of a recording of a phone call between former U.S. Assistant Secretary of State for European and Eurasian Affairs Victoria Nuland and former U.S. Ambassador to Ukraine Geoffrey Pyatt. The call was uploaded to YouTube. On the call, Nuland and Pyatt discussed Ukraine’s leadership transition, and Nuland, now infamously, says “f— the EU.”32 Russia used its intelligence tools to capture the communications and then release the recording to advance its narrative that the United States was meddling in Ukraine, as well as to attempt to sow discord between the European Union and the United States. What is notable, however, is that most intelligence services would have shied away from deploying information gained from sensitive intelligence tools in such an operational way. The objective of most intelligence services is to gain intelligence to inform policymakers. By making that call public, the Russians alerted Nuland and Pyatt to the fact that their calls were being monitored and thereby burned a potentially valuable information source. Moreover, this incident also prompted other U.S. government officials to increase their vigilance about their own communications, possibly burning other Russian sources of information. In other words, releasing the call had a clear intelligence cost to Russia, yet the Kremlin released the call anyway. Moscow placed greater priority on advancing a narrative than on maintaining its access to intelligence. This is a calculation that Western intelligence agencies would almost never make.
Russia has invested in developing its cyberhacking capabilities. For instance, the Russian military’s Main Intelligence Directorate, or GRU, has developed units of cyberhackers, which have been responsible for a number of high-profile hacking efforts. The GRU hacking units have been identified variously as Fancy Bear, APT28, STRONTIUM, and Operation Pawn Storm.33 The U.S. IC found “that the GRU operations resulted in the compromise of the personal e-mail accounts of Democratic Party officials and political figures.”34 This same GRU group was also responsible for the 2015 hack of the German Bundestag, or Parliament. Hans-Georg Maassen, the head of Germany’s domestic security agency, stated, “We recognize this as a campaign being directed from Russia … for disinformation or for influencing operations … Whether they do it or not is a political decision … that I assume will be made in the Kremlin.”35 Furthermore, private cybersecurity companies have identified a clear pattern. Ed Cabrera with Trend Micro, a security software firm, said, “the victimology—when they are attacking, how they are attacking, and who they are attacking”—all align with Kremlin interests. He said these GRU-tied units push well-crafted phishing campaigns to gain access to accounts. These groups may also use a technique known as tabnabbing—essentially, spoofing a fake webmail page that says the user’s log-in information is expired and to re-enter the user’s credentials.36
One unique aspect of Russian efforts, according to a senior British intelligence official in an interview with the Financial Times, is that Moscow has also “fostered a network of ‘modern privateers’,” effectively emulating state-sanctioned piracy of centuries ago where monarchs green lighted pirates to plunder foreign ships. Another U.S. intelligence official told the Financial Times that “the links between Russia’s state agencies and criminal networks when it comes to aggressive cyber activities are deep and developing.”37 These criminal cyber actors are essentially allowed to operate unimpeded if they confine their efforts outside Russia and serve the needs of the Kremlin when called upon. As a result, Russia serves in effect as a safe haven for cybercrime against the West. Moreover, these private hackers also give the Kremlin additional cyber firepower and the ability to surge its efforts when needed.
The Russian hacking and subsequent release of emails from the Democratic National Committee (DNC) and from Clinton campaign chairman John Podesta represented, according to the IC, an “unprecedented” intervention in the U.S. election process.38 The private cybersecurity firm CrowdStrike identified Russian hacking units Fancy Bear and Cozy Bear as the culprits behind the DNC and Podesta hacks, which was later corroborated by U.S. intelligence.39 Similarly, in the French elections, a massive document dump of stolen files from the campaign of Emmanuel Macron, the eventual winner, was released in the final few days of the election. It is not uncommon for foreign intelligence agencies to penetrate political campaigns in an effort to gain insight into future policy approaches. However, Russia was not seeking intelligence; it was seeking to influence the U.S. and French elections and the democratic process in both countries.
Russia also does not seem particularly concerned about covering its tracks and may even want the victims to know that it was behind the hackings. Following the DNC hack, Michael Buratowski of Fidelis Cybersecurity commented that it was clear Russia was behind the hacking, as it used “Russian internet addresses, Russian language keyboards, and the time codes corresponding to business hours in Russia.”40 Similar “digital ‘fingerprints’” pointing to Russia were found following the hacking of the Macron campaign.41
After hacking into and collecting emails from the DNC and Podesta, the GRU, posing as the persona of Guccifer 2.0, laundered the content through WikiLeaks. It remains unclear whether WikiLeaks serves as a witting or an unwitting accomplice used by a foreign intelligence agency. Regardless, WikiLeaks served as the vehicle for Russia to publicize the stolen information with a veneer of deniability. While WikiLeaks released and promoted thousands of stolen emails, it appears to have done so selectively and strategically, with the goal of influencing the 2016 campaign and election. For instance, many emails were never released.42 That being the case, the information being released and publicized is not “leaks” from a concerned actor with legitimate access to sensitive information but rather stolen information laundered by a foreign intelligence service.
The timing of the email releases was also designed for political effect. The DNC emails were released right before the Democratic National Convention in an effort to drive a wedge within the Democratic Party. Podesta’s emails were released just 29 minutes after the release of Donald Trump’s infamous “Access Hollywood” tape, and Macron’s campaign emails were released in the final days of the French election.43
A critical factor in the success of these operations is the complicity of the press. While WikiLeaks served as a vehicle to launder the stolen information during the American campaign, the American press, as noted by Rid during his testimony to the Senate Intelligence Committee, also effectively served as “unwitting agents” of Russia in its eagerness to report on the stolen content.44 Here the contrast to the reaction of the French press is striking, including the fact that the main French newspaper Le Monde refused to write about the stolen material until after the election, while The New York Times and other key U.S. media gave extensive coverage to Podesta’s emails and the hacked DNC material.45
What makes the new online information landscape so troubling is that Russia, as well as other foreign actors, has been able to greatly expand its espionage efforts against the United States with little consequence. Before the internet and social media, cultivating intelligence assets in the United States largely had to be done in person and was therefore more difficult and incredibly risky. This forced Soviet and Russian intelligence to be highly selective with their efforts. But now Russian intelligence can target Americans en masse and can do so with impunity from thousands of miles away. For instance, citing information from the cybersecurity firm SecureWorks, Rid found that in a period of 14 months, the GRU sent “19,300 malicious links, targeting around 6,730 individuals.”46 This yielded for Russia troves of information that it can deploy to influence events, attack its enemies, extract financial or business data, shape public opinion, and potentially blackmail and recruit foreign agents.
Semi-covert: Troll, forge, disrupt, and amplify
Where Russia has innovated is in its incorporation of semi-covert cyber operators into its information operations. Russia uses paid social media users, or trolls, to form an army of online cyber operators and propagandists. These operators serve as online foot soldiers engaging in keyboard-to-keyboard combat on the front lines of Russia’s information war, working to elevate, disseminate, and amplify information that advances Kremlin objectives. Individually, their impact may be minimal, but collectively and operating in concert with other efforts, these operators have a strategic effect.
To build up these forces, the Kremlin has set up front organizations such as the Internet Research Agency, or the Agency, based in St. Petersburg. Funded by a Russian oligarch with ties to the Kremlin, the Agency was estimated to employ around 400 people with a budget of about $400,000 per month, with a typical employee working a 12-hour shift for approximately $700 per month.47 In 2014, BuzzFeed reported on leaked documents regarding the Agency, finding, “On an average working day, the Russians are to post on news articles 50 times. Each blogger is to maintain six Facebook accounts publishing at least three posts a day and discussing the news in groups at least twice a day. By the end of the first month, they are expected to have won 500 subscribers and get at least five posts on each item a day. On Twitter, the bloggers are expected to manage 10 accounts with up to 2,000 followers and tweet 50 times a day.”48
These groups, known as troll farms, operate like a campaign operation. They have certain messages or themes that they are pushing or advancing for that day or a week. This action can be as basic as defending the Kremlin or pushing pro-Russian content, but it can also involve advancing conspiracy theories that cast doubt on Western governments or pushing attacks on globalism. During the 2016 campaign, a major focus was spreading messages that attacked Hillary Clinton or that cast doubt on the credibility of U.S. institutions or on the election itself—in this case, claiming the election is “rigged.”49 The American intelligence community even noted that, “Pro-Kremlin bloggers had prepared a Twitter campaign, #DemocracyRIP, on election night in anticipation of Secretary Clinton’s victory, judging from their social media activity.”50
But the online efforts often go much further than amplifying content or spreading vitriol. For instance, Finnish journalist Jessikka Aro, who sought to highlight Russian disinformation, was harassed viciously online and falsely accused of being a drug dealer on alternative news websites, which posted personal photos of her online. Saara Jantunen, a researcher at the Finnish Defense Forces, explained such tactics, saying, “They fill the information space with so much abuse and conspiracy talk that even sane people start to lose their minds.”51 Aro made herself a target for dezinformatsiya and of Russia’s online troll army. The result was a multidimensional attack: vitriolic online harassment; the production and posting of fake or slanderous stories; even the organization of protests outside the offices of her employer. Russia effectively made an example of Aro with the designed aim of creating a chilling effect on other journalists covering Russia.52
While disinformation campaigns can have specific targets, they are also simply trying to create noise online and complicate or pollute anti-Kremlin narratives, meaning that the effectiveness of a singular post matters much less than the volume of posts. As Ben Nimmo, now of the Atlantic Council, explained in 2015, Russian online propaganda “largely relies on four tactics: dismiss the critic, distort the facts, distract from the main issue and dismay the audience.”53 The Kremlin therefore accepts failure as part of the process—some disinformation ploys will not catch on—and understands that if even a few of these efforts are visibly successful, the broader effort has an impact.54 Unlike Western public diplomacy efforts, Russian information efforts enjoy broad funding support, and since there is no focus on accuracy or avoiding unwanted international or domestic blowback, there is no detailed bureaucratic organizational chart through which these operators must work. These troll farms can thus operate rapidly and with agility.
These information operations exploit the fact that social media platforms promote or elevate content that is being talked about, shared, or trending, regardless of what that content is and who and what is causing the content to be shared. A Russian operative at a troll farm can operate a number of Twitter and Facebook accounts simultaneously by programing them to operate automatically. On Twitter, a Russian operative can use automated accounts, or bots, that can rapidly amplify content. Operators can create a network of fake bot accounts known as botnets to create a reinforcing ecosystem, creating the illusion of an online community and audience. While just a single person may be operating this ecosystem, it has the appearance of credibility due to the way that social media elevates content that is shared widely. A Russian operative can therefore amplify content and have their own bots expand the popularity of a specific tweet or post, such that it gains the attention of actual users and, potentially, the press. This amplification matters because it makes it more likely for this content to spread and spill over into Facebook feeds or Google news lists. Data scientists and digital media specialists Jonathon Morgan and Kris Shaffer found that, “In addition to Trump support and anti-Semitic, nativist rhetoric, both bots and sockpuppet accounts were more likely to discuss Russia than normal users. The bot accounts in particular mentioned Russia four times as often as other users, and were especially interested in countering the narrative that Russia was involved in swaying the election.”55
The use of botnets also helps embolden fringe and extremist groups by creating a faux sense of their community being larger than what it is. This makes individuals with extremist views more willing to vocalize them and encourage others. This creates a degree of normalization—suddenly, for example, posting racist statements brings followers and praise instead of scorn. Samuel Woolley of Oxford University, co-author of a report on “pro-Trump bots,” explained the objective of using of bots: “The goal here is not to hack computational systems but to hack free speech and to hack public opinion.”56
Often, these trolls, bots, and botnets push fake news stories, which are routinely produced and laundered through fake news websites. For instance, in an effort to dissuade Sweden from joining NATO, stories emerged that NATO would secretly store nuclear weapons in Sweden, that NATO could attack Russia from Sweden without Swedish approval, and that NATO soldiers could rape Swedish women with impunity. More recently, the Russians have spread rumors about German soldiers stationed in Lithuania, accusing them of rape.57
While pro-Russia troll farms can fabricate their own fake news, they can also manipulate the online marketplace to incentivize others to create fake news for them. Many of the fake news items seen during the election were linked back to Eastern Europe—specifically, to a town in Veles, Macedonia, where an industry emerged to write fake news stories and profit off the traffic-driven advertising revenue. While these fake news creators seem to be motivated by profit, what also seems likely is that they owe their profits in part to Russian amplification through botnets, which may have driven traffic to fake news stories, helping drive up their revenue from traffic-based advertising. Reporting has indicated that fake news on the right was more popular than fake news on the left; while progressives may like to believe that they are simply harder to fool, there were also fewer bots and other artificial traffic drivers during the election seeking to elevate fake progressive content.58
After receiving significant media attention in 2015, the Internet Research Agency claimed to have closed, rebranding itself instead as a pro-Kremlin media company. But Russian news organization RBC recently revealed that it was still operational. The Moscow Times explained, “Despite its new status as a network of legitimate—if heavily biased news outlets—the ‘troll agency’ hasn’t quite abandoned its old ways, RBC’s report suggests.” The report found that a popular pro-Trump, anti-Clinton Facebook group called Secure Borders,59 which boasts 140,000 subscribers, was actually “managed from the St. Petersburg troll factory. … One of its posts published at the height of the election campaign and heavily advertised on Facebook, reached 4 million people on Facebook, was ‘liked’ more than 300 thousand times and shared more than 80 thousand times.” Reporting on the RBC report, The Moscow Times says it discovered that a right-wing Twitter account called Tea Party News was managed from the same location: “All in all, RBC’s sources say that at the zenith of the U.S. election campaign, the troll factory’s accounts across different social media platforms would churn out as many as 50 million posts a month, with anti-Clinton messages getting the most attention.”60
It is important to note, however, that Russia did not create the caustic online environment. The methods they have utilized are also widely used by others online, including by those on the right and the left. It is therefore hard to pinpoint the precise impact of Russia’s online efforts, both on the election and on public discourse—which, of course, is largely by design. What is clear is that social media manipulation played a key role during the 2016 election. For example, an Oxford University professor, Philip Howard, described how pro-Trump bot networks began to use pro-Clinton hashtags to inject negative memes, links, and political messages into pro-Clinton circles.61 Like a virus, they essentially co-opted the opponent’s messaging and infiltrated her supporters. Using pro-Clinton hashtags such as #ImWithHer and #uniteblue, memes describing Clinton as corrupt ricocheted across both blue and red feeds. In one joint academic study on botnets, researchers found that “pro-Trump hashtags were inserted into more and more combinations of neutral and pro-Clinton hashtags, such that by the time of the election fully 81.9 percent of the highly automated content involved some pro-Trump messaging.”62
It is difficult to determine precisely how much of this traffic was instigated or amplified by Russian operatives or was simply a campaign tactic by pro-Trump or far-right Americans. This degree of ambiguity has prompted defenders of the Kremlin to claim there is no proof. While it is difficult to disaggregate, the impact of Russian trolling efforts in influencing the election and polluting public discourse was no doubt significant. Russia, as a state actor, has resources and capabilities at its disposal that are much greater than those of individuals, groups, or political campaigns. Furthermore, the U.S. intelligence community report on Russian interference highlighted the Internet Research Agency in its unclassified report and concluded that “Russia used trolls … as part of its influence efforts to denigrate Secretary Clinton.”63 Sen. Mark Warner (D-VA), ranking member of the Senate Select Committee on Intelligence, noted that, “there were upwards of a 1,000 paid internet trolls, working out of a facility in Russia … they can generate news down to specific areas … in Wisconsin, Michigan, Pennsylvania.” This enabled them to push anti-Clinton messages.64 Additionally, Time reported that U.S. intelligence officials found that “Moscow’s agents bought ads on Facebook to target specific populations with propaganda.” According to a senior intelligence official interviewed for the article, “‘They buy the ads, where it says sponsored by—they do that just as much as anybody else does.’”65
Russian influence efforts were also heavily enmeshed with U.S. alternative media. A number of academic experts in social media analysis have documented the role of Russian trolls, bots, and botnets in amplifying content and in Russia’s growing links to the alt-right.66 In 2015, Nimmo explained that “the Kremlin media use Western commentators to amplify and validate Moscow’s messages.”67 Russia’s messaging and posturing has also demonstrated an intimacy with alt-right content, as shown by Russia’s tweeting of a racist meme used by white supremacists.68 Kate Starbird of the University of Washington found that “the structure of the alternative media ecosystem and the content that is hosted and spread there suggest the use of intentional disinformation tactics—meant to create ‘muddled thinking’ and a general mistrust in information.”69 These links are not surprising given Russia’s well-documented backing of far-right political parties and extremist groups.70
However, the links between the U.S. alt-right media ecosystem and Russia may have involved more than just amplification. According to the American publishing company McClatchy, the FBI, as part of its investigation into Russian interference in the election, is examining whether far-right websites such as Breitbart and Infowars knowingly coordinated with Russian cyber operators. McClatchy reported that “operatives for Russia appear to have strategically timed the computer commands, known as ‘bots,’ to blitz social media with links to the pro-Trump stories … Investigators examining the bot attacks are exploring whether the far-right news operations took any actions to assist Russia’s operatives.”71
Overt: Propaganda pushers and fake news launderers
Russian media outlets such as RT and Sputnik serve to advance the Kremlin’s agenda domestically and internationally and act as key players in dezinformatsiya operations. These outlets effectively serve as propaganda tools and so-called information underers, playing a role similar to that of money launderers but for information.
While news organizations compete for viewers, RT’s low ratings have little impact on its funding. According to the U.S. intelligence community report, RT alone spends more than $190 million per year on the “distribution and dissemination” of its programming, though other reports have indicated that its budget has been more than $300 million per year.72 This is a dramatic increase over its initial budget of $30 million per year in 2005. While this budget is incredibly large for a poorly rated news network, compared with a weapon system, it is a relative bargain. The lavish funding by the Russian state enables RT and Sputnik to attract mainstream talent and to create a slick, modern news platform. For instance, RT paid former national security adviser Lt. Gen. Michael Flynn to attend its 10th anniversary gala, where he sat at a table with President Putin and gave a talk.73 RT has also hired prominent television personalities such as former CNN host Larry King, adding to its image as a mainstream network. However, RT was dramatically unmasked as the mouthpiece of the Kremlin when RT anchors Sara Firth and Liz Wahl publicly resigned out of disgust at being part of the Russian propaganda machinery.74
As part of their goal to advance the interests of the Kremlin, these so-called news organizations work to sow doubt and discredit the American and European democratic systems. RT, for instance, uses the tagline “Question More” to justify pushing conspiracy theories and sowing doubt in Western state institutions. Because RT does not have a domestic partisan agenda, it eagerly highlights voices on both ends of the political spectrum—as long as they are critical of Western governments. For instance, RT has hired prominent progressives such as Ed Schultz, who now works for RT, and had Green Party candidate Jill Stein seated at the same table at the RT gala as Flynn.75 RT will give extensive coverage to events that portray America in a negative light, such as the protests in Ferguson, Missouri, as this coverage highlights America’s continuing problems with racism and police brutality. RT will also give considerable coverage to anti-fracking stories, as it is in Russia’s interest for America not to develop its natural gas industry, which could rival Russia’s. RT and Sputnik’s willingness to selectively highlight critical voices on the right and the left also adds to the credibility of these organizations on both sides of the political spectrum, which enhances the ability of RT and Sputnik to push disinformation.
The Institute of Modern Russia explained Russia’s approach as combining “Soviet-era ‘whataboutism’ and Chekist ‘active measures’ with a wised-up, post-modern smirk that says everything is a sham. Where the Soviets once co-opted and repurposed concepts such as ‘democracy,’ ‘human rights’ and ‘sovereignty’ to mask their opposites, the Putinists use them playfully to suggest that not even the West really believes in them. Gitmo, Iraq, Ferguson, BP, Jobbik, Schröder—all liberalism is cant, and anyone can be bought.”76
Key to RT and Sputnik’s ability to launder information is creating a veneer of credibility. Much of the content published or broadcast may be legitimate news of the day. This makes it harder for viewers and readers to weed out stories that are either completely fabricated or pure propaganda. These outlets will take fake information, often originating online, and give it the veneer of credibility by reproducing it in RT- or Sputnik-produced stories. These stories are then reinjected into social media and advanced and promoted by troll farms and botnets.
Occasionally, the volume of attention given to a fake story will prompt legitimate mainstream media to report on it as well. One telling example of this chain of information laundering came in August 2016 with the spread of a false story about Turkish forces surrounding the U.S. airbase in Incirlik. According to cybersecurity experts Clint Watts and Andrew Weisburd, the false story started on Twitter, then migrated to RT’s and Sputnik’s Twitter accounts, and was then picked up and promoted in an “hours-long storm of activity from a small, vocal circle of users,” many of whom were pro-Trump and pro-Russia.77 A couple weeks later, during an interview with CNN, Paul Manafort—then campaign chair for Donald Trump—tried to call out the media for not covering the fabricated attack on the NATO base in Turkey, referring to the Incirlik base, which houses NATO troops. In trying to criticize the media’s lack of coverage on an attack that never occurred, Trump and Manafort demonstrated a willingness to espouse false reporting from Russian state media.78
While RT will highlight stories on the left, during the 2016 election, stories on RT quickly made their way to alt-right U.S. media such as Breitbart. In one recent example, an RT commentator floated a conspiracy theory that President Barack Obama asked the British intelligence of the Government Communications Headquarters (GCHQ) to surveil the Trump campaign. Fox News commentator Judge Andrew Napolitano allegedly saw this content and repeated it on Fox, prompting Sean Spicer, White House press secretary, to repeat the allegation at a White House press briefing. In other words, disinformation spread from a Russian propaganda network to Fox News to the White House.79
An integrated disinformation campaign
These different disinformation tools in the covert, semi-covert, and overt space can also all work together in synchronized fashion to reinforce and amplify each other’s efforts. One vivid example of this came on September 11, 2014, when the Office of Homeland Security and Emergency Preparedness for St. Mary Parish, Louisiana, received reports that there had been a chemical plant explosion in Centerville, Louisiana. News of the alleged explosion spread across Twitter, with hundreds of users documenting what appeared to be eyewitness accounts and videos of the explosion and one user even posting a screenshot of CNN’s homepage reporting on the story. According to one YouTube video, the Islamic State took credit for the attack. In the end, however, the entire incident proved to be an extremely well-coordinated hoax by the Internet Research Agency, which involved not only the use of dozens of fake Twitter accounts but also the creation of clone news sites, a Wikipedia page documenting the explosion, and a fake YouTube video.80 These complex efforts are designed to sow public distrust of the U.S. media and U.S. government institutions.
The post-Cold War strategy of seeking to integrate Russia into the liberal global order is no longer operable. Russia under Vladimir Putin will continue to position itself as a geopolitical adversary of the United States. The United States needs to reset its posture toward Russia to recognize that it now confronts a global ideological competition not seen since the Cold War. The ideological challenge that communism posed to liberal democracy has been replaced by authoritarian nationalism.
While disinformation operations are a key Russian tool in this ideological contest, they are ultimately an asymmetric tool used by a weaker power to level the playing field. Russia lacks the economic and military capacity to challenge the United States and its allies directly, and strongman-authoritarian governance lacks the ideological pull of liberal democracy. A key facet of responding to an adversary using an asymmetric tool is not to respond in a way that undercuts one’s own strengths. Countering disinformation is not about fighting fire with fire. The United States should not conduct disinformation campaigns, as doing so would only play into the hands of the Kremlin by adding to the cognitive dissonance that undercuts democracy and would be relatively ineffective against relatively closed societies.
Instead, the United States must seek to improve efforts to deter, detect, and counter disinformation. The United States must also recognize that countering disinformation and defending the truth means playing defense. Therefore, the United States needs to look to other tools through which it can go on the offensive, impose costs, and deter future attacks. Countering Russian disinformation requires playing offense and defense and should be embedded within a broader strategy toward challenging Russia in this new ideological contest, especially by standing up strongly in defense of democracy and human rights. It should be noted that effectively countering Russian disinformation will require vastly more commitment on the part of the U.S. government, especially as it is currently headed by a president who, along with his surrogates, has actively pushed disinformation.
To effectively respond to the Russian threat to democracy, the United States should:
Immediately impose meaningful costs on Russia to deter future attacks against the United States and its allies
While an outgoing administration could only do so much, evicting diplomats and intelligence agents, as the Obama administration did on December 29, was not enough. Russia benefits economically from its access to open liberal societies, while at the same time it uses its access to undermine them. Congress should impose robust sanctions that further isolate Russia’s economy and starve its oligarchs of their easy access to launder money into the West.90 The United States must also send a message to other countries that if they intervene in the U.S. democratic process or that of its allies, the costs will be severe. Unfortunately, while there seems to be bipartisan support for stronger sanctions in Congress, the Republican leadership, including Chairman of the Senate Foreign Relations Committee Sen. Bob Corker (R-TN), are blocking action.91 As a result, the administration and the congressional majority are leaving the United States and its allies in Europe dangerously exposed. Putting in place stronger economic sanctions is a necessary first step in establishing a modicum of deterrence, since more in-depth policy responses will take time to put in place.
Establish an independent commission to conduct a top-to-bottom assessment of the U.S. government’s tools and capabilities to counter the Russian threat
The U.S. government is simply not positioned to respond effectively to foreign information operations. As former FBI Special Agent and cyber and homeland security analyst Clint Watts noted in his testimony to the Senate Intelligence Committee, the U.S. national security bureaucracy has become heavily tilted toward counterterrorism efforts over the past 15 years.92 While these efforts are vital, it is clear that more attention and resources need to be allocated to countering the threat from foreign information operations. Following the attacks of September 11, an independent 9/11 Commission was set up and proved crucial not just to getting to the bottom of what happened but also to providing recommendations for how the U.S. national security bureaucracy should reform itself to counter the terrorist threat. The United States needs an independent commission modeled on the 9/11 Commission to conduct a top-to-bottom examination of the U.S. national security bureaucracy to identify new approaches, reforms, and actions necessary to confront the threat of foreign information operations. This should involve a thorough review of U.S. cyber, intelligence, homeland security, law enforcement, diplomatic, and public diplomacy tools and provide recommendations to better posture the United States to tackle this threat.
- The U.S. Department of Homeland Security and state governments need to take urgent steps to protect the integrity of our voting systems. While the Department of Homeland Security and the intelligence community determined that Russia did not alter vote tallies in the 2016 election, they did conclude that Russia had gained access into the systems of some state voter registration systems.93 This represents a clear and present danger to the integrity of elections. While Russia may not have manipulated voter tallies this time, there is no guarantee that it—or another actor—will not do so next time. As former Assistant Attorney General for National Security John Carlin argues, voting systems should be seen as “critical infrastructure,” and the United States should have a “nonpartisan warning system” going into the next election if there are indications of hacking efforts.94
- Give greater priority and resources to combating cyberhacking and intrusions, especially those related to elections. The Obama administration significantly expanded the government’s attention to cyberhacking. However, the lackadaisical FBI response to the Russian hacking of the Democratic National Committee is unacceptable and demonstrates that greater vigilance is needed, especially relating to elections and political campaigns. The government should be ready and able to respond rapidly when there are potential state-sponsored hacking attempts on political parties, candidates, and election-related offices.
Political campaigns must also take stronger steps to protect and defend themselves. Beyond simply taking cybersecurity seriously and adopting smart and thorough cybersecurity procedures, campaigns should look to some of the steps taken by the Macron campaign in France. Well aware that it was a target for Russian espionage, the Marcon campaign went on a counteroffensive against the hackers to sow confusion and waste their time. The campaign claimed to have planted fake documents, signed on to phishing emails, and flooded the hackers with fake passwords and logins.95 This enabled the campaign to highlight that some of the files that were dumped online were fake, which complicates the media’s ability to report on the information.
- U.S. Cyber Command and the intelligence agencies need to assess their cyber tools in light of the Russian cyber operations. What seems evident from the Obama administration’s policy response is that U.S. cyber tools, while likely the most powerful in the world, are not nearly as nimble, creative, or flexible as believed. In light of the 2016 attack, the United States needs to assess whether it is prioritizing the right operational objectives and is appropriately postured to counter disinformation campaigns in the cyber realm. Actions should include going after Russian hackers or exposing Russian censorship tools, as suggested by retired Admiral James Stavridis, former NATO supreme allied commander.96 There may not be as much to be done in this space as it seems, as countering Russian actions in cyberspace with cyber tools may not be effective. The possible costs of the development of international cyber norms and the potential escalatory nature of an overt cyber response may outweigh the potential gains of such a response. But given the threat posed by Russian information operations, it is necessary to assess whether U.S. cyber resources are being deployed effectively and whether the United States is making the investments needed to develop the right capabilities in light of Russian efforts. For instance, Hans-Georg Maassen, the head of Germany’s domestic security agency, expressed the need for Germany to develop offensive cyber capabilities: “We believe it is necessary that we are in a position to be able to wipe out these servers if the providers and the owners of the servers are not ready to ensure that they are not used to carry out attacks.”97 The United States should similarly explore using its cyber tools to destroy stolen information before it can be deployed for disinformation campaigns.
- Structure intelligence disclosure policies and practices to give greater priority to countering disinformation and advancing public diplomacy goals. The United States could potentially use its intelligence tools to counter disinformation in a more effective manner and in so doing advance public diplomacy goals. In general, the United States is extremely reluctant to make intelligence information public to advance public diplomacy policy goals. This reluctance is sensible, but it can also severely limit the United States’ ability to make its case to a foreign government or public or to push back on incorrect information. For instance, U.S. diplomats are often put in the position of pressing countries to take action on nonproliferation cases—such as providing warnings to a country to stop and search a cargo ship in its port carrying illicit nuclear materials—but cannot fully disclose the information in order to protect intelligence sources and methods. This often leaves the United States in a “trust us” mode—a posture that can hinder the ability to convince countries to take action. In the case of information operations, this means that even if the United States has information that could advance U.S. public diplomacy objectives vis-à-vis Russia or could be effective in countering Russian disinformation or in embarrassing the Kremlin—information akin to Russian kompromat, such as knowledge that a senior leader has a mistress or evidence of corruption—that information is rarely injected into the public domain. Admittedly, this is a difficult balance to strike, but in weighing intelligence disclosure decisions, it is likely that greater weight should be given to the policy objectives of countering disinformation and advancing U.S. public diplomacy. After all, the intelligence community should serve broader U.S. policy objectives.
- Expand counterintelligence efforts. The United States needs to up its intelligence and counterintelligence efforts concerning Russia. Since 9/11, counterterrorism has appropriately been the priority. However, given the escalation of Russian influence and espionage efforts, more resources and personnel need to be devoted to countering Russian espionage within the United States.
- Push at a high level to strengthen international cyber norms. Developing the rules of the road in cyberspace is immensely challenging. Given the difficulties of assessing attribution, the prevalence of nonstate actors, and the difficulty in determining the line between what constitutes a cybercrime or traditional espionage and what constitutes a cyberattack, which is an act of war, is extremely unclear. These issues complicate efforts to emulate nuclear arms control efforts. Nevertheless, the fact that cyberspace represents a unique challenge does not mean that efforts should not be pursued to establish and reinforce norms, develop codes of conduct, or reach an arms-control style treaty. Rep. Jim Himes (D-CT) has similarly pointed to the need to negotiate a treaty to lay out international standards of behavior in cyberspace, just as the Geneva Conventions did for conventional warfare.98
Progress has been made. In 2015, following bilateral discussions between the United States and China, the G-20 affirmed in a statement that no country should “conduct or support cyber-enabled theft of intellectual property” with the goal of gaining a competitive economic advantage.99 In an effort to advance international cyber norms, the U.S. Department of State has sought to affirm the applicability of existing international law to the cyber domain, forge norms relating to state behavior during peacetime, and build confidence to foster cooperation among nations. But for ongoing U.S. efforts to gain momentum, high-level effort is needed. In 2010, for instance, President Obama held a nuclear security summit, which put a spotlight on loose nuclear material and helped prompt international action.100 A similar high-level effort is needed to elevate this issue on the international agenda and prompt international action.
- Establish boundaries and deterrence in cyberspace through the clear messaging of U.S. cyber redlines and by loudly calling out cyber intrusions. Developing clear messages and redlines about what the United States would deem to be a cyberattack under the law of war could decrease ambiguity and help deter such attacks against America. Indeed, President Obama privately confronted President Putin at the G-20 summit in China and warned him that hacking the voting systems would cross the line and merit a strong retaliatory response.101 The United States should more clearly articulate sectors that it believes should be off limits to a cyberattack and warn that if these sectors are deemed to be under attack—such as interference in an election or an attack on critical infrastructure—the United States will respond forcefully.
Develop tools to shine a spotlight on disinformation to build public resilience
Ultimately, efforts to expose disinformation are about playing defense. These efforts will inevitably manifest as a reaction to events, rather than shaping those events. As a result, a response is likely to be slower and more bureaucratically cumbersome than the more agile Russian disinformation bureaucracy, which is not worried about accuracy. Nonetheless, improving U.S. defenses is critical, and when it comes to disinformation, sunshine can be a great disinfectant. U.S. efforts should not focus on debunking every falsehood online but instead on raising public awareness and resilience. The Department of Homeland Security’s U.S. Computer Emergency Readiness Team (US-CERT), which primarily focuses on cybersecurity and serves as an important liaison with the private sector, should have its mission expanded. It should be given the resources and capabilities to publicly expose foreign operatives online and to debunk fake news relating to U.S. government activities. US-CERT should also work closely with the State Department’s Global Engagement Center, which has a similar mission abroad. This expanded mission for US-CERT could involve efforts to:
- Detect and monitor. Identifying and monitoring foreign information operations will require additional resources and the hiring additional of personnel. As Clint Watts explained in his testimony to the Senate Select Committee on Intelligence, the intelligence agencies have not prioritized unclassified open source information.102 The United States needs to expand its efforts to monitor and track foreign actors operating in cyberspace.
- Notify and highlight. Once the United States is better able to detect information operations, it should do more to notify the public and news media of the current environment. Many Nordic and Baltic states are much more resilient to Russian disinformation efforts, as their publics have become familiar with Russian actions and tactics. Russian disinformation efforts have been identified and highlighted by the political leadership of these states; as a result, their press members have become more skilled at detecting these campaigns. Following suit, U.S. government officials should do more to notify and brief the public and press on foreign information operations.
- Debunk misinformation. The United States could also emulate the European Union’s efforts to counter Russian disinformation by publicly identifying and highlighting fake information. In 2015, the European Union established a disinformation task force dubbed East StratCom to address Russian disinformation campaigns. The task force scours the internet for fake news and disinformation, highlighting disinformation efforts on its Twitter account and sending it out in biweekly newsletters called the Disinformation Review.103 For instance, following the April terrorist attack in Stockholm, the Disinformation Review highlighted how pro-Kremlin outlets spread conspiracy theories and manipulated photos after the tragedy, noting that this fit a familiar pattern.104 As of November 2016, the newsletter had 20,000 readers each week,105 with a major audience for the material being journalists who can use it as a resource to avoid being duped.
- Troll the trolls. Just as the United States monitors social media for counterterrorism purposes, it could develop its ability to detect state-sponsored operators. The United States could alert American users who interact with these operatives, or their controlled bots, that the accounts they are interacting with are either suspected agents of a foreign government or are bots controlled by suspected agents. A U.S. government account, operated by US-CERT, could effectively troll the trolls by pointing them out and shining a spotlight on their efforts. This sort of interaction would be similar to Google’s warning to Gmail users that a state-sponsored entity is seeking to hack someone’s email.106 Establishing this capacity would require additional resources and manpower but could likely rely on the same type of automation used by Kremlin-controlled bots.
- Name and shame. The United States should more aggressively name and shame countries that violate norms of cyberspace. The Obama administration’s more aggressive approach to China in the cyber domain, for instance, prompted China to come to the table.107 While it is unlikely that naming and shaming would have a significant impact on Russia, it would help highlight its behavior for the public and would make Russian information operations a major bilateral irritant, forcing it on to the bilateral agenda. It would also prompt the rest of the international community to take notice and can serve to isolate.
Significantly expand public diplomacy efforts
Efforts to counter disinformation are defensive and reactive in nature. But the United States can go on the offensive in the information space by significantly expanding its public diplomacy efforts. In testimony to the Senate in January, James Clapper, then-director of national intelligence, said that America needs a “USIA [U.S. Information Agency] on steroids,” as we need “to fight this information war a lot more aggressively.”108 While an independent commission should assess recreating the USIA, as a first step Congress should increase funding for the existing, though severely underfunded and neglected, public diplomacy tools.
As the United States cut funding for public diplomacy efforts after the Cold War, Russia—as well as countries such as Iran and China—significantly expanded funding of state-supported media.109 Meanwhile, many Western news networks decided it was not profitable enough to invest in foreign language media in small markets in places such as the Balkans. Russia has sought to fill this gap in the news media marketplace through the expansion of Russian-state funded media. Many of the local media environments are increasingly Russian-dominated—where anti-U.S., anti-NATO, anti-EU, and anti-democratic messages carry the day.110
Yet despite the urgent need for increased U.S. efforts, the Trump administration is proposing massively cutting funding for the State Department and, as a consequence, cutting efforts to counter Russian propaganda. Specifically, President Trump’s fiscal year 2018 budget calls for a 31 percent cut in funds to the State Department and the U.S. Agency for International Development, which, if enacted, would reduce spending as a percentage of U.S. gross domestic product to its lowest levels since World War II.111 While Trump’s budget was light on details, these cuts would seriously affect critical diplomacy programs aimed at countering attacks on the United States and its allies. This is exactly the wrong approach.
The United States needs to support and expand efforts to provide an independent alternative to Russian disinformation. Doing so requires significant expansion in funding efforts for U.S.-sponsored outlets such as Radio Free Europe/Radio Liberty and Voice of America, which are funded by the United States but governed by the Broadcasting Board of Governors; therefore, the U.S. government has no operational or editorial input. Like the Public Broadcasting Service (PBS) and National Public Radio (NPR) here in the United States, these outlets serve as a source of independent news and as the surrogate free press where the press is stifled, producing content in more than 25 languages. Radio Free Europe/Radio Liberty and Voice of America also launched “Current Time” earlier this year, a fact-based Russian language 24-hour news channel designed to provide a fact-based alternative for Russian speakers.112 These efforts, however, remain woefully underfunded and fall short of what is needed to challenge Russian-backed media, which has become entrenched in many countries.
As part of this investment, the State Department should also revamp its approach to public diplomacy. The National Defense Authorization Act for Fiscal Year 2017 (NDAA), expanded the mandate of the Global Engagement Center (GEC) beyond countering the Islamic State messaging to countering disinformation from state actors. The NDAA also authorized a significant increase in the GEC budget from $5 million to up to $80 million.113 Secretary of State Rex Tillerson should implement the changes authorized in the NDAA and prioritize the expansion of the GEC, including expanding its collaboration with NATO and the European Union. Additionally, the State Department’s Bureau of Public Affairs Rapid Response Unit, which monitors foreign news and reports trends, should feed its efforts into the GEC.114
In addition, the State Department should prioritize public diplomacy tours and trainings for its foreign service officers, and Congress should support education and cultural exchanges that are supported by the State Department’s Bureau of Educational and Cultural Affairs. The State Department should also take steps to end a climate of caution prevalent among public diplomacy officers, who fear being punished if their engagement with the public ever goes awry. For example, Senior Public Affairs Officer Larry Schwartz, in Egypt, was recalled from Cairo back to Washington because of a tweet directed at an Egyptian audience with which some on the U.S. far right took issue.115 His recall to Washington sent a chilling effect throughout the State Department—engaging the public could put your career at risk. The State Department and the U.S. government need to develop a thicker skin if public diplomacy efforts are to be effective.
Social media companies should take steps to avoid being unwitting agents
The erosion of trust in online content and information could have far-reaching economic impacts. The emergence of a new “sharing” economy is built on public trust of online tools. It took years for the public to gain trust in online venues, as the public was dubious of online interactions in the early years of the internet. If consumers no longer trust the information they are receiving on social media sites and fear sharing financial information or conducting online transactions due to threats from hacking, this will have a far-reaching impact on Silicon Valley. Social media companies need to do more to address the fact that their platforms are being used by foreign governments as vehicles to conduct information operations.
Since the election, Facebook and Google have taken some steps to tackle this issue. In fact, Facebook recently released a report acknowledging that bad actors, including governments, have exploited the platform to “manipulate civic discourse and deceive people.” In the report, Facebook claims that it has developed “new analytical techniques” specifically “to uncover and disrupt” such abuse on its platform, which in the case of France, recently enabled Facebook to take action against 30,000 fake accounts.116 Facebook has also started to implement a third-party fact-checking tool to combat fake news that warns users when content is disputed.117 Additionally, Facebook has taken steps to change the algorithm for its trending section, which will now try to promote topics that are not only popular but also have multiple related articles to try to prevent viral false stories from being listed.118 Meanwhile, Google has announced the expansion of its use of fact-checking tags, whereby news search results are tagged with such phrases as “mostly true” or “false” if stories have been checked. Google has paired up with more than 100 news and fact-checking organizations whose conclusions appear in search results if they have met certain criteria.119 Both Facebook and Google have also taken actions against fake news sites directly, with Google banning websites that spread misinformation from using its online advertising service and Facebook clarifying its ad placement policies to include not displaying ads for sites that include fake news.120 Google reported that as a result of this change, it had banned 200 publishers from its advertising network.121
In a Daily Beast article, Clint Watts and Andrew Weisburd recommend that social media companies fund the creation of an independent news rating system akin to Consumer Reports.122 This rating agency would create a system for rating the accuracy of the source of the information, as opposed to assessing the content of the information. Watts compares it to providing nutrition labels on food products, which gives consumers the information to make informed judgments about what they eat but does not stop consumers from eating fatty food. Similarly, a rating system on news would provide social media users with information about the news they see but would not block this news from being shared or violate anyone’s freedom to consume the fake news. As Watts and Weisburd note, “If social media users choose to read junk news and be mind fat, they have no one to blame but themselves.”
Notably absent from the list of giant tech companies trying to tackle this issue is Twitter, which is problematic given that a recent study found that up to 15 percent of Twitter accounts are run by bots, not people, which translates to nearly 48 million bot accounts on the platform.123
However, some outside groups have developed tools for users to track disinformation themselves. For instance, Indiana University and the Center for Complex Networks and Systems Research have created a new tool called Hoaxy that allows users to conduct searches to track the origins of claims and how they spread on Twitter and to analyze the information available to determine for themselves whether a claim or rebuttal is true.124 Researchers at Indiana University previously developed a tool that helps identify whether a Twitter account is being operated by a bot.125
There is, however, still considerable doubt that these steps are sufficient to address the problem of the spread of disinformation on these platforms. Social media companies may be reluctant to crack down on bots and other artificial users because these accounts and the traffic they generate increase companies’ user statistics, making their platforms seem more widely used—and therefore more valuable, which increases share price. This is manipulation, and advertisers on Twitter are beginning to notice.126
News media must avoid being unwitting agents
The news media is critical to building social resilience to disinformation. The press needs to assess its journalistic standards when it comes to its approach to stolen information. How material is accessed and received should come into play when journalists and news organizations—measuring against ethical standards—decide how to treat certain information. The race to publish emails stolen from an individual’s account and laundered through WikiLeaks creates a dangerous precedent. The press needs to treat stolen and laundered information with a high degree of caution and skepticism, similar to how it has treated the publication of the Christopher Steele dossier on Trump’s Russia ties—refusing to report on some elements and prefacing any mention of the dossier as being “unverified.”127 For instance, ahead of its recent presidential elections and following the leaking of documents allegedly stolen from the Macron campaign, France’s electoral commission called on the media not to cover the content of the leaked materials, warning that “the dissemination of such data, which has been fraudulently obtained and in all likelihood may have been mingled with false information, is liable to be classified as a criminal offense.” As a result, the French media largely complied with the commission’s request, including such major outlets as Le Monde.128
In a climate of fake news, there is a great need to cultivate news media that is independent of outside funding sources abroad. The same is true within the United States. The United States should increase investments into PBS and NPR.
The United States must stand up for democracy and human rights
Combating disinformation is also about standing up in support of freedom and democracy around the world at the highest levels of the U.S. government. Unfortunately, the Trump administration has not only failed to speak up for democracy and human rights, but it has also actively sought to undermine these values and what America stands for by, for instance, hosting Russian Foreign Minister Sergey Lavrov at the White House days after Russian interference in the French election—as well as ignoring human rights abuses and inviting President of the Philippines Rodrigo Duterte to the White House.129 The Trump administration has bent over backward to accommodate and offer support to authoritarian governments and strongman rulers. Moreover, Trump’s attack on the press, Secretary of State Tillerson’s unwillingness to allow the press to travel with him, the abandonment of the State Department’s daily press briefing, and rhetoric about locking up political opponents cause tremendous damage to the moral authority of the United States and serve to weaken America’s ability to lead.130 When the leader of the free world no longer values democracy, these principles are weakened worldwide.