The Perfect Technocracy: Facebook's Attempt to Create Good Government for 900 Million People

Facebook's desire for efficiency means democracy is out and technocratic, developer-king rule is in.

crowns-615.jpg

Reuters/FakeTV

Let's stipulate that Facebook is not a country, that real governments fulfill many more functions, and that people are not citizens of their social networks.

Nonetheless, 900 million human beings do something like live in the blue-and-white virtual space of the world's largest structured web of people. And those people get into disputes that they expect to be adjudicated. They have this expectation in part because Facebook has long said it wants to create a safe environment for connecting with other people. (How else can you get people to be "more open and connected"?) But people also want someone to be in charge, they want an authority to whom they can appeal if some other person is being a jerk.

Except in this case, the someone really is a corporate person. So when you report something or someone reports something of yours, it is Facebook that makes the decision about what's been posted, even if we know that somewhere down the line, some human being has to embody the corporate we, if only for long enough to click a button.

Any individual decision made by Facebook's team -- like taking down this photo of a gay couple kissing -- is easy to question. Ars Technica's Ken Fisher detailed a whole bunch of one-off problems that people have encountered with Facebook's reporting system. In each, there is an aggrieved party, but we're only hearing one side of the conflict when these problems bubble up. Across many single events, you have two people (or entities like businesses) with conflicting desires. This is a classic case where you need some sort of government.

It's not hard to imagine making one or 20 or even 200 decisions about photographs or status updates in a week, but it's mindboggling to consider that Facebook has to process 2 million reports per week, and that's not including simple "mark as spam" messages.

How do you design a system to deal with that workload? I spoke with James Mitchell, who helms what Facebook calls "site integrity" within its user-operations department, and Jud Hoffman, the company's global policy manager about the reporting process. They are the architects of Facebook's technocracy.

"The amount of thought and debate that goes into the process of creating and managing these rules is not that different from a legislative and judicial process all rolled up into one," Hoffman, a lawyer, told me. "And James has the executive/judicial element. I don't think it is a stretch to think about this in a governance context, but it's a different form and we take it really, really seriously."

The key step, Mitchell told me, was to put some structure into the reporting process. Back when he started in 2006, there wasn't any form to complaints from users. That meant there was a massive queue of undifferentiated problems. So, he and his team started to think about what kinds of problems they received and created categories of problems, which they refined over time.

That allows the reports to be channeled through a complex set of processes and teams so that they arrive in front of human beings or computers that know what to do with them.

Facebook has revealed this infrastructure for the first time today. It's the product of more than five years of work by several teams within Facebook, who have worked to make the process of handling this flood of user inquiries as efficient as possible. (Click the graphic to enlarge it.)

FB_Reporting_Guide_1.6-1.jpg

At the end of many of these reporting lines, there's a person who has to make a decision about the user's message. Some of these decisions are binary -- Does this photograph contain nudity? -- and those are generally outsourced to teams that can apply simple and rigorous formulas such as asking, "Is this person naked?" Other decisions are complex in ways that make machines very good at dealing with them. (For example, there are more than 50 signals that Facebook's algorithms look at to determine whether a profile is spam, and the automated responses are more accurate than human ones would be.)

But the bulk of the reports are fielded by a faceless team of several hundred Facebook employees in Mountain View, Austin, Dublin, and Hyderabad. These people and the tools they've built have become the de facto legislators, bureaucrats, police, and judges of the quasi-nation of Facebook. Some decisions they make impact hundreds of millions of people in some small way; other decisions will change some small number of people's lives in a big way.

Presented by

How to Cook Spaghetti Squash (and Why)

Cooking for yourself is one of the surest ways to eat well. Bestselling author Mark Bittman teaches James Hamblin the recipe that everyone is Googling.

Join the Discussion

After you comment, click Post. If you’re not already logged in you will be asked to log in or register.

blog comments powered by Disqus

Video

How to Cook Spaghetti Squash (and Why)

Cooking for yourself is one of the surest ways to eat well.

Video

Before Tinder, a Tree

Looking for your soulmate? Write a letter to the "Bridegroom's Oak" in Germany.

Video

The Health Benefits of Going Outside

People spend too much time indoors. One solution: ecotherapy.

Video

Where High Tech Meets the 1950s

Why did Green Bank, West Virginia, ban wireless signals? For science.

Video

Yes, Quidditch Is Real

How J.K. Rowling's magical sport spread from Hogwarts to college campuses

Video

Would You Live in a Treehouse?

A treehouse can be an ideal office space, vacation rental, and way of reconnecting with your youth.

More in Technology

Just In