Virtually every month it seems, new controversies emerge swirling around Facebook, Google, Amazon, Twitter and other Silicon Valley companies. The latest controversy, which involved allies of Donald Trump swiping personal data from as many as 87 million Facebook users in the US presidential election, is yet another window into the nature of these companies. Before that there was the scourge of fake news and Russian propaganda bots that had spread across the Facebook platform like a metastasized cancer. And before that Google being fined €2.4 billion by the European Commission for manipulating its search results in favor of preferred companies.
In the case of Facebook, we are left with the daunting question of whether Mr. Zuckerberg and his computer geniuses really understand their own creation. Facebook’s AI has been built (or more accurately, cobbled together) over several years by hundreds of different developers and programmers. Professor Zeynep Tufekci from Harvard University describes the Facebook algorithm as “giant matrices, maybe millions of rows and columns, and not even the programmers understand anymore how exactly it is operating.” There are so many variables that go into its complex and proprietary sorting that Facebook cannot say with authority why something will or will not appear in a user’s news feed, or how and why suddenly Russian trolls and their bots were able to manipulate the algorithms to reach nearly half of all US voters with targeted fake news. (This included such whoppers as the Pope had endorsed Donald Trump for president, which received nearly 2m Facebook “engagements” (total number of shares, likes and comments) in the three months leading up to the U.S. election).
Nevertheless, a number of experts have been closely observing this company and have figured out a few of its behavioural patterns. Combined with recent revelations from a whistleblower, here’s what we have learned about how Facebook and its algorithms actually work. It is even more alarming than anyone thought.
AI is being aimed at users to affect not only advertising but also news and elections. Given Facebook’s widespread use and influence – with 2 billion global users, it is fast replacing television as the most dominant news, entertainment and commercial medium in the world — this manipulation strikes at the very heart of our democratic societies.
First, Facebook’s “engagement algorithms” use technological surveillance of our online behavior to capture our personal data in a way that would have made the Stasi drool with envy. The goal is to generate increasingly accurate, automated predictions of what advertisements we are most influenced by.
Recently Wired reported on a leaked Facebook memo which revealed that the company had offered advertisers the opportunity to target 6.4m younger users, some only 14 years old, during moments of psychological vulnerability. Facebook monitored posts, photos, and interactions in real time to track emotional lows, and the 23-page document actually highlighted Facebook’s ability to micro-target ads down to “moments when young people need a confidence boost.” As Dr Tufekci explains: “Humans are a social species…We are particularly susceptible to glimmers of novelty, messages of affirmation and belonging, and messages of outrage toward perceived enemies. These kinds of messages are to human community what salt, sugar, and fat are to the human appetite.”
So Facebook offers these to its users in a grand gluttony of feeding, creating what the company’s first president, Sean Parker, recently called “a social-validation feedback loop.” The platform is specifically designed to keep users clicking, tapping, and scrolling down a bottomless feed, and in the process deliver us to various advertisers.
But that’s not all. Based on our individual profiles, the Facebook engagement algorithms are also designed to feed us sensationalist news (both fake and real) selected to provoke powerful emotions of anger and fear. By reacting to, clicking on and sharing these stories, users are herded by the Facebook “persuasion architecture” into hyper-partisan information ghettos of opinion and alternative facts, referred to as “cognitive bubbles”.
One outrageous example was the false conspiracy theory blasted around Facebook during the presidential campaign that Hillary Clinton and her former campaign chair ran a child sex ring in the basement of a pizzeria in Washington DC. Besides the fact that the restaurant, Comet Ping Pong, doesn’t even have a basement, the restaurant’s staff and its owner were hit with a fusillade of abuse and death threats on social media. Matters went from alarming to dangerous when a man walked into Comet Ping Pong with an assault rifle and began shooting (fortunately no one was injured). That was just one of dozens of fake news stories, all of them with absurd storylines. Other stories claimed that Hillary Clinton sold weapons to Isis, and that an FBI agent linked to Clinton’s email leaks had been mysteriously found dead.
A BuzzFeed News analysis found that 17 of the 20 top-performing false election stories were overtly pro-Donald Trump or anti-Hillary Clinton. In the last three months of the presidential campaign, top fake election news stories generated nearly 9m Facebook engagements, which was 20 percent greater than the number received by election stories from 19 major news outlets combined.
We need your support
Social Europe is an independent publisher and we believe in freely available content. For this model to be sustainable, however, we depend on the solidarity of our readers. Become a Social Europe member for less than 5 Euro per month and help us produce more articles, podcasts and videos. Thank you very much for your support!
Choosing the target
But the impacts go even deeper than fake news. Whistleblower Christopher Wylie revealed to the New York Times and UK’s Observer how his former company, Cambridge Analytica, used personal information swiped from 87 million Facebook users to build a system that created psychological-political profiles of most US voters via online questionnaires. Cambridge Analytica was headed by Trump’s key adviser Steve Bannon, from the alt-right media outlet Breitbart. For the 2016 US presidential election, the Guardian reports that Cambridge Analytica deployed a set of techniques that were adopted from the US Department of Defense and UK Ministry of Defence, in particular their “psychological operations,” or psyops. Those methods were focused on changing people’s minds, not through persuasion but through “informational dominance” that relies on disinformation, fake news, rumor and “psychographic messaging.”As the New York Times pointed out, a voter found to be neurotic might be shown a gun-rights commercial featuring burglars breaking into a home, rather than a dry legal defense of the Second Amendment; voters troubled by anxiety would be targeted with ads warning of the dangers posed by the Islamic State, but such ads would be presumed ineffective with those identified as ‘optimistic.’
Cambridge Analytica’s precursor, called SCL Elections, claims that it has used a similar suite of tools in more than 200 elections around the world, including in Italy, Ukraine, Romania, South Africa, Nigeria, Kenya, India, Indonesia, Thailand and many other democracies. When the Facebook data breach first occurred back in 2014, it included a third of active North American users, and nearly a quarter of potential US voters. Wylie told the newspapers; “We exploited Facebook…and built models to exploit what we knew about [their users] and target their inner demons.” Paul-Olivier Dehaye, a data expert and academic based in Switzerland, who published some of the first research into Cambridge Analytica’s processes, says it’s become increasingly apparent that Facebook is “abusive by design.”
Even if the Cambridge Analytica techniques were not wholly influential in swaying voters, as some pundits have claimed, that misses an important point. Left to their own devices, these companies are writing the rules of our collective digital future. And we are just at the beginning stage of what these technologies eventually will be capable of. The linkage of powerful emotive responses that have been digitally stimulated to orchestrate a kind of partisan groupthink is chillingly reminiscent of George Orwell’s Two Minutes of Hate in his novel 1984. Dr Tufekci says that “the core business model underlying the Big Tech platforms—harvesting attention with a massive surveillance infrastructure to allow for targeted and mostly automated advertising at a very large scale—is far too compatible with authoritarianism, propaganda, misinformation, and polarization.”
Beyond social networks
With its 2bn users worldwide, Facebook has grown from a pet project started in Zuckerberg’s Harvard dorm to become far more than a social networking platform. It has morphed into a huge news, entertainment and advertising platform that is viewed by more people than any US or European television network, any newspaper or magazine and any online news outlet. It also reaches hundreds of millions of users in the developing world, where the company has tailored its app for low-bandwidth connections and less expensive Android phones. Many grey- and black market entrepreneurs in the developing world use a Facebook app as their eBay-like commercial gateway for buying and selling.
Google also has mastered these kinds of architectures of engagement, and Amazon is well on its way. In so doing, these companies have become three of the most valued companies in the world. Together Google and Facebook now account for a whopping 73% of all global online advertising revenue (84% outside China) and 25% of all ad sales (online or off). Amazon is now slowly playing catch up in the advertising game. But this growing duopoly is crowding out other media outlets, forcing even Twitter and Snapchat to fight for the advertising scraps needed to survive.
In other words, we the public are the guinea pigs for these platform monopolies’ algorithmic experiments. The power of algorithms is being put to work for very questionable ends, and has a propensity to result in unforeseen consequences. Viktor Mayer-Schönberger, Professor at Oxford University and co-author of Reinventing Capitalism in the Age of Big Data, says; “The algorithms and datasets behind them will become black boxes that offer us no accountability, traceability, or confidence.” In the past, he says, most computer code could be opened and inspected, making it effectively transparent. But with AI and its enormous datasets enhanced by “machine learning,” the human ability to monitor these technological puzzles is declining.
Europe has long operated by the “precautionary principle,” which is like the Hippocratic oath in medicine that says, “First, do no harm.” The products and services being created by these Silicon Valley companies have enjoyed widespread access to European markets and consumers. Europeans have great faith in an open Internet, but the pitfalls of this over-optimism are becoming more and more apparent. Now that Facebook has grown into a large global monopoly, a major media platform for the entire planet, it has become less benign and is raising more alarm bells.
A European model platform?
Many people have long lamented “Where is the European Facebook and Google?” But the emperor no longer has any clothes; the digital leadership of Silicon Valley platform companies has been shown to be sinisterly irresponsible and teetering on being dangerous. Now is the time for some European start-ups to come together, with seed money from the EU and/or individual governments, and create a new version of Facebook (and Google, and Amazon) that incorporates European values. Call it Facebook 2.0. China has accomplished this – why can’t Europe?
At this point, a kind of renationalization of the Internet seems natural and almost inevitable. Nations and regional blocs like the EU and China must each re-configure the Net in ways that work for their populations, their values and their future needs. Currently these platform companies seem to exist everywhere and nowhere, growing into global behemoths. One idea that has been discussed is that of turning these kinds of services into public utilities. Another is that of breaking them up as overly big monopolies. Yet another option is that of requiring digital licenses that map out the rules and regulations of operation for Internet-based platforms, much in the way that traditional brick-and-mortar companies must be granted business permits and licenses.
The evolution of the Digital Age is proceeding rapidly, and just as in bygone eras when oil, phone and Microsoft monopolies eventually needed to be yoked, it is time to figure out the right digital harness for these platform companies. The alternative is to leave the standards and norms that will rule the future to be defined by these Frankenstein companies from Silicon Valley.
Steven Hill is the former policy director at the Center for Humane Technology and author of seven books, including Raw Deal: How the Uber Economy and Runaway Capitalism Are Screwing American Workers and The Startup Illusion: How the Internet Economy Threatens Our Welfare.