How Facebook’s Data Has Been Turned Against Its Users—and Democracy
What state should you move to based on your personality? What character on Downton Abbey would you be? What breed of dog is best for you? Some enormous percentage of must have seen Facebook friends sharing results of various online quizzes. They are sometimes annoying, senseless, and a total waste of time. But . Besides, you’re only sharing the results with your family and friends. There’s nothing more innocent, right? Wrong. Facebook is in the . The company is worth billions of dollars because it to advertisers. Users are encouraged to like, share, and comment their lives away in the name of staying connected to family and friends. However, as , I know that there is a lot more to the story. The .
You are being psychographically profiled
Most people have heard of demographics—the term used by advertisers to slice up a market by age, gender, ethnicity, and other variables to help them understand customers. In contrast, psychographics measure people’s personality, values, opinions, attitudes, interests, and lifestyles. They help advertisers understand the way you act and who you are. Historically, psychographic data were much harder to than demographics. Today, Facebook is the world’s largest treasure trove of this data. Every day billions of people give the company huge amounts of information about their lives and dreams. This isn’t a problem when the data are used ethically—like when a company shows you an ad for a pair of sunglasses you recently searched for. However, it matters a lot when the data are used maliciously—segmenting society into disconnected echo chambers, and custom-crafting misleading messages to manipulate individuals’ opinions and actions. That’s exactly what Facebook allowed to happen.
Quizzes: reading your mind and predicting your politics
Recent reports have revealed how Cambridge Analytica, a U.K.-based company owned by an and led at the time by candidate Donald Trump’s key adviser Steve Bannon, used psychographic data from Facebook to profile American voters in the months before the 2016 presidential election. Why? To target them with and . A whistleblower from Cambridge Analytica, Christopher Wylie, described in detail how the company exploited Facebook users by harvesting their data and building models to “.” How did Facebook let this happen? The company does more than just sell your data. Since the early 2000s, Facebook has provided access to . Many psychologists and social scientists have made their careers by asking simple questions. These questions, like the ones used in social media quizzes, do not appear to have obvious connections to politics. Even a decision like which web browser you are using to read this article is . In 2015, Facebook gave permission to academic researcher . Like other quizzes, his was able to capture , including name, profile picture, age, gender, and birthday; everything you’ve ever posted on your timeline; your entire friends list; all of your photos and the photos you’re tagged in; education history; hometown and current city; everything you’ve ever liked; and information about the device you’re using including your web browser and preferred language. Kogan shared the data he collected with Cambridge Analytica, which was against Facebook policy—but apparently the company .
Going shopping for impressionable users
Analyzing these data, Cambridge Analytica determined topics that would intrigue users, what kind of political messaging users were susceptible to, how to frame the messages, the content and tone that would motivate users, and how to get them to share it with others. It compiled a that could be predicted about voters. Then the company was able to and encourage them to spread the word. : “they see it … they click it … they go down the rabbit hole.” This is how American voters were targeted with fake news, and intended to influence how they voted—or . This is how Facebook users’ relationships with family and friends are being exploited for monetary profit, and for .
Knowingly putting users at risk
Facebook could have done more to protect users. The company encouraged developers to build apps for its platform. In return, the apps had access to vast amounts of user data—supposedly subject to those . But Facebook collected 30 percent of payments made through the apps, so its business interest made it want more apps, doing more things. People who didn’t fill out quizzes were vulnerable, too. Facebook allowed companies like Cambridge Analytica to collect personal data of friends of quiz takers, without their knowledge or consent. of people’s data were harvested—and many more Facebook users could have been affected by other apps.
Changing culture and politics
In a video interview with the Observer, Wylie explained that “.” That’s exactly what Facebook enabled Cambridge Analytica to do. In 2017, the company’s CEO boasted publicly that it was “able to use data to identify … very large quantities of persuadable voters … that .” To exert that influence, Cambridge Analytica—which claims to have — to . This included what became known as “fake news.” In an undercover investigation, Britain’s Channel 4 expressing their willingness to disseminate misinformation, with its CEO saying, “These are things that don’t necessarily need to be true, .” U.S. society was unprepared: , and many people who see fake news stories . So Cambridge Analytica’s tactics worked: on Facebook a total of 30 million times. In fact, the most popular fake news stories were . For this , the Trump campaign .
A healthy dose of skepticism
U.S. history is filled with stories of people sharing their thoughts in the public square. If interested, a passerby could come and listen, sharing in the experience of the narrative. With combined psychographic profiling, analysis of big data, and , public discourse in the U.S. has entered a new era. What used to be a public exchange of information and democratic dialogue is now a customized whisper campaign: Groups both ethical and malicious can divide Americans, whispering into the ear of each and every user, nudging them based on their fears and encouraging them to whisper to others who share those fears. : “There are two fundamental human drivers … hopes and fears … and many of those are unspoken and even unconscious. You didn’t know that was a fear until you saw something that evoked that reaction from you. Our job is … to understand those really deep-seated underlying fears, concerns. It’s no good fighting an election campaign on the facts because actually it’s all about emotion.” The information that you shared on Facebook exposed your hopes and fears. That innocent-looking Facebook quiz isn’t so innocent. The problem isn’t that this psychographic data were exploited at a massive scale. It’s that platforms like Facebook enable people’s data to be used in ways that take power away from voters and give it to data-analyzing campaigners. In my view, this kills democracy. Even Facebook can see that, saying in January that at its worst, social media “ y.” My advice: Use Facebook with a healthy dose of skepticism. This article was originally published by . It has been edited for èßäÉçÇø! Magazine.