The New Politics of Knowledge
Speech delivered at the Jefferson Society, University of Virginia, Charlottesville, Virginia, November 9, 2007, and at the Institute of European Affairs, Dublin, Ireland, September 28, 2007, as the inaugural talk for the IEA's "Our Digital Futures" program.
I want to begin by asking a question that might strike you as perhaps a little absurd. The question is, "Why haven't governments tried to regulate online communities more?" To be sure, there have been instances where governments have stepped in. For instance, in January of last year in Germany, the father of a deceased computer hacker used the German court system to try to have an article about his son removed from the German Wikipedia. As a result, wikipedia.de actually went offline for a brief period. It's come back online, of course, and in fact the article in question is still up.
Here's another example. In May of last year, attorneys general from eight U.S. states demanded that MySpace turn over the names of registered sex offenders lurking on the website, which as you probably know is heavily frequented by teenagers. The website deleted pages of some 7,000 registered sex offenders. And the following July, they said that in fact some 29,000 registered sex offenders had accounts, which were subsequently deleted.
Those are just a few examples. But we can make some generalizations. The Internet is famously full of outrageously false, defamatory, and offensive information, and is said to be a haven for criminal activity. This leads back to the question I asked earlier: why haven't governments tried to regulate online communities even more than they have?
We might well find this question a little absurd, especially if we champion the liberal ideals that form the foundation of Western civil society. Indeed, no doubt one reason is our widespread commitment to freedom of speech. But consider another possible reason—one that, I think, is very interesting.
Governments, and everyone else, implicitly recognize that social groups, however new and different, have their own interests and are usually capable of regulating themselves. It is a truly striking thing that people come together from across the globe and, out of their freely donated labor and strings of electrons, form a powerful new corporate body. When they do so—as I have repeatedly observed—they develop a sense of themselves as a group, in which they invest some time and can take some pride, and which they govern by rules.
In fact, these groups are a new kind of political entity, the birth of which our generation has been privileged to witness. Such groups are not like supra-national organizations, like the United Nations; nor are they like international aid organizations, like Doctors Without Borders; nor are they quite like international scientific groups, like the Intergovernmental Panel on Climate Change. The existence and primary activity of these online communities is all online. Their membership is self-selecting, international, and connected online in real time. This makes it possible for enormous numbers and varieties of groups to arise, of arbitrary size and arbitrary nationality, to achieve arbitrary purposes. They essentially make up a new kind of political community, a cyber-polity if you will, and so there is a presumption that they can regulate themselves. Government steps in, as in the case of MySpace, only when they cannot regulate themselves responsibly.
The idea that online communities are a kind of polity is, I think, very suggestive and fruitful. I want to talk in particular about how online communities, considered as polities, are engaged in a certain new kind of politics—a politics of knowledge. Let me explain what I mean by this.
Speaking of a "politics of knowledge," I assume that what passes for knowledge, or what we in some sense take ourselves to know as a society, is determined by those who have authority or power of a sort. You don't of course have to like this situation, and you might disagree with the authorities, or scoff at their authority in some cases. Nevertheless, when for example professors at the University of Virginia say that something is well known and not seriously doubted by anyone who knows about the subject, those professors are in effect establishing what "we all know," or what we as a society take ourselves to know. Since those professors, and many others, speak from a position of authority about knowledge—a powerful force in society—surely it makes some sense to speak of a politics of knowledge. I just hope you won't understand me to be saying that what really is known, in fact, is determined by whoever happens to be in authority. I'm no relativist, and I think the authorities can be, and frequently are, wrong.
If we talk about a politics of knowledge, and we take the analogy with politics seriously, then we assume that there is a sort of hierarchy of authority, with authority in matters of knowledge emanating from some agency that is "sovereign." In short, if we put stock in the notion of the politics of knowledge, then we're saying that, when it comes to knowing stuff, some people are at the top of the heap.
Our new online communities—our cyber-polities—are increasingly influential forces, when it comes to the politics of knowledge. When Wikipedia speaks, like it or not, people listen. So in this talk I want to discuss in particular something I call the new politics of knowledge. Any talk of a new politics of knowledge raises questions about what agency is sovereign. Well, it is often said that in the brave new world of online communities, everyone is in charge. Time Magazine's "Man of the Year" is, by practice, usually some influential political figure. When its "Person of the Year" last year was "You," Time didn't break its practice. Time was rightly claiming that, through Internet communities we are all newly empowered. In the new politics of knowledge, we can all, through blogs, wikis, and many other venues, compete with real experts for epistemic authority—for power over what is considered to be known.
If this sounds like a political revolution, that's because it is. It is frequently described as a democratic revolution. So what I'm going to do in the rest of this talk is examine exactly what sense in which the new cyber-polities, like Wikipedia, do indeed represent a sort of democratic revolution. This discussion will have the interesting result that we should be more concerned than we might already be about the internal governance of Internet communities—because that internal governance has real-world effects. And I will conclude by making some recommendations for how cyber-polities should be internally governed.
As a philosopher, I find myself impelled to ask: what exactly is democratic about the so-called Internet revolution?
Democracy in one very basic sense means that sovereignty rests ultimately with the people, that is, with all of us. Bearing that in mind, the new Internet revolution might be democratic, I think, both in a narrow sense and in a broad sense. The narrow sense concerns project governance: the new content production systems are themselves governed ultimately by the participants, and for that reason can be called democratic. In the broad sense, the Internet revolution gives everyone "a voice" which formerly many did not have, a stake in determining "what is known" not just for a narrow website or Internet practice, but for society as a whole. To draw the distinction by analogy, we might say that each online community has a domestic policy, about its own internal affairs, and a foreign policy, through which it manages its influence on the world at large.
Now, I'd like to point something out that you might not immediately notice. It is that the broad sense depends in a certain way on the narrow sense. The contributors are ultimately sovereign in various Internet projects, and that is precisely why they are able to have their newfound broader influence over society. Let's take Digg.com as an example. This is a website that allows people to post any link, and then others vote, a simple up or down, on whether they "digg" the link. It's one person, one vote. Of course, no one checks anybody's credentials on Digg. The highest-voted links are placed most prominently on the website. So the importance of a Web article, and presumably whatever the article has to say, is determined democratically, at least as far as the Digg community goes. But Digg's influence goes beyond its own community. A relatively obscure story can become important by being highly rated on Digg. In this way, all those people voting on Digg—and these can be as expert as you hope, or as uneducated, ignorant, biased, immature, and foolish as you fear—they can wield a power to highlight different news stories, a power hitherto usually reserved only to professional journalists.
Similarly, Wikipedia articles are now well-known for being the #1 Google search result for many popular searches. Any website with that much reach is, like it or not, very influential. That is, in effect, practical epistemic authority. That is real authority, given to anyone who has the time and patience to work on Wikipedia and do the hand-to-hand battle necessary to get your edits to "stick" in Wikipedia articles. That power, to define what is known about a general topic, was formerly reserved only to the professional intellectuals who wrote and edited encyclopedias, and more broadly to experts generally speaking. And again, of course, no one checks anybody's credentials before they get on Wikipedia. So amateurs are to some extent displacing experts, in the new politics of knowledge.
So that's why we call the Internet revolution democratic. But this needs some qualification. There is one fundamental reason that we describe as "democratic" such websites as Digg, Wikipedia, MySpace, YouTube, and all the rest, and that is that anyone can, virtually without restriction, go to the website and get involved. This, however, is only to say that they have a certain benchmark level of "user empowerment," which we might call the "right to contribute." But frequently, a large variety of governance structures are superimposed upon this basic "right to contribute." While the content is generally determined by largely self-governing contributors, some policies and decisions are left in the hands of the website owners, like Slashdot and YouTube, who are officially answerable to no one else within the project. Granted, if these privileged persons anger their contributors, the contributors can vote with their feet—and this has happened on numerous occasions. And in some cases, such as Wikipedia, the community is almost completely self-governing. Still, we probably should qualify claims about the democratic nature of cyber-polities: just because there is a basic right to contribute, it does not follow that there will also be an equal right to determine the project's internal governance.
So, as I said before, the Internet revolution is democratic in the broad sense because it is democratic, however qualifiedly, in the narrow sense. In other words, internal Web project governance bears directly on real-world political influence. But how closely connected are Web community politics and real-world influence?
Consider Wikipedia again—and I think this is particularly interesting. If you've followed the news about Wikipedia at all in the last few years, you have might noticed that when they make larger changes to their policy, it is no longer of interest just to their contributors. It is of interest to the rest of the world, too. It gets reported on. Two recent news items illustrate this very well.
First item. A few months ago, a student posted a website, called the WikiScanner, that allows people to look up government agencies and corporations to see just who has been editing which Wikipedia articles. This was fairly big news—all around the world. I was asked to comment about the story by reporters in Canada and Australia. Journalists think it's absolutely fascinating that someone from a politician's office made a certain edit to an article about that politician, or that a corporation's computers were used to remove criticisms about the corporation. At the same time, reporters and others observe that Wikipedia's anonymity has allowed people to engage in such PR fiddling with impunity. And that is the interesting internal policy point: anyone can contribute to Wikipedia without identifying him- or herself. You can even mask your IP address, which those political aids and corporation employees should have done; all they had to do was make up some random username, which one can still do without giving Wikipedia an e-mail address, and then the WikiScanner couldn't track the IP address. Nobody who was signed in was caught by the WikiScanner. Anyway, it was an internal policy that has had some very interesting external ramifications.
Second item. It was reported recently by the London Times that the German Wikipedia would be changing its editing system. In the future, all edits by unregistered and newer contributors will have to be approved by the older contributors before they can appear on the website. In fact, this was old news—the system described has been under development for well over a year, and it still hasn't been put into use. Nevertheless, it has been touted as a very big concession on the part of Wikipedia. It's said now that Wikipedia has a role for "trusted editors" on the website, but this is incorrect; it has a role only for people who have been in the system for a while, and these can be very untrustworthy indeed. However unlikely this is to have any significant effect, it was still touted as important news. And again, what was touted as big news was a change in internal policy, the policy about how the wiki can be edited by newer and anonymous contributors. This is supposed to be important, because it might help make Wikipedia a more responsible global citizen.
In general, it is becoming increasingly clear that the "domestic policy," so to speak, of cyber-polities is closely connected with their real-world impact. Wikipedia isn't the only example I might give. Here's another—although in this case, the effect is economic, not epistemic. There is an amazingly huge website, called craigslist, which lists, they say, over 12 million new classified ads every month. This website has proven to be a real thorn in the side of local newspapers, which depend on revenue from ads. Increasingly, people are posting their classified ads in craigslist instead of in their local newspapers. This is the effect of a policy, an internal policy, that anyone can post an ad for free, except for employment ads in certain markets. What might have originally seemed to be an optional feature of a small Web community has turned out, in fact, to cost jobs at newspapers.
But let's get back to the politics of knowledge. In the intellectual sphere, I think the full power of collaboration and aggregation has yet to be demonstrated. Try to imagine Wikipedia done right—not just enormous, but credible and well-written. If this sounds impossible to believe, consider that just a few years ago, Wikipedia itself, a reasonably useful general encyclopedia with over two millions articles in English, would have sounded equally impossible to believe. I can tell you that, when Wikipedia was first starting out, there were many people who sneered that we didn't have a chance.
Let me describe briefly my new project, which is relevant here. It is called the Citizendium, or the Citizens' Compendium. It is a non-profit, free wiki encyclopedia that invites contributions from the general public—and to that extent it's like Wikipedia. There are three very important differences, however. First, we require the use of real names and do not allow anonymous contribution; we also require contributors to submit at least a brief biography. So we all know who we're actually working with. Second, we distinguish between rank-and-file authors, which do not require any special qualifications, and editors, who must demonstrate expertise in a field; our editors may approve articles, and they may make decisions about content in their areas of expertise. Still, they work side-by-side with authors on the wiki. Nobody assigns anybody any work; it's still very much a bottom-up process. Third, we are a rather more mature community. All contributors must sign onto a sort of social contract, which states the rules of the community; we expect people to behave professionally; and we have people called "constables" who are actually willing to enforce our rules by kicking out troublemakers.
So how is the project going? We started a pilot project just over a year ago, and in that time we created 3,500 articles, and we have over 2,000 authors and well over 200 expert editors on board. We also have more words than Wikipedia did after its first year—our average article is six times as long as the average Wikipedia article after its first year. Our pace of article production has accelerated—it has doubled in the past 100 days or so and tripled since last January. And we are pretty much free of vandalism, and I think our articles are pretty high-quality for such a wide-open project. The project is doing rather well, and I think that we are probably, with continued development, poised to replicate Wikipedia's sort of growth. We too could have a million articles in under ten years.
Well, imagine that the Citizendium had a million articles, together with loads of ancillary reference material such as images, tables, tutorials, and so forth—all free, credible, and managed by experts. The sort of influence that such a website would wield would, I think, far outweigh Wikipedia's. The one thing that really holds Wikipedia back, from the end user's perspective, is its reliability. So suppose there were a similar website that solved that problem.
If you ask me, this is somewhat of a frightening prospect. After all, already, far too many students and even members of the general public treat Wikipedia as if it were reliable. Already, for far too many students, Wikipedia is their only source of reference information. If humanity were to produce a similarly giant encyclopedia that were really reliable, you can just imagine how it would probably be received by the general public. It would become, essentially, the world's textbook and omnipresent reference library. There would be a general presumption that what it says is correct, and that if anyone asserts something in contradiction to it, they would have to explain in as much detail as they would have to do if they contradicted the Encyclopedia Britannica today. Sure, a good encyclopedia can be wrong; but it usually isn't. Unlike Wikipedia, it's innocent until proven guilty.
This is frightening, I say, precisely because of how powerful such a resource would be. Imagine the article about, for example, the Iraq War, after it had been written and rewritten, and checked and rechecked, by hundreds of real experts. It would no doubt be a thing of beauty, as I think the Citizendium's best articles are. But it would also be taken as the starting-point for serious conversation. What claims it makes could have real-world political ramifications, as much as, if not more than, any U.N. report. So you can easily imagine the attention given to major changes of policy, or to internal rulings on controversial cases in the project. Again: the internal policymaking for a truly successful collaborative reference project would have major external consequences.
We don't want governments to take over or closely regulate collaborative projects, but if they continue to act as irresponsibly as Wikipedia has, I fear that they might attempt to do so. That is, for me, a disturbing scenario, because in a civilized, modern, liberal society—one that deeply values the freedom of speech—the authority to say what we know is one power that should not be in the hands of the government. Every government regulation of online collaborative communities is a direct threat to the sovereignty of that community, and an implicit threat to the free speech of its members.
It is, therefore, extremely important that online projects, ones with any influence, be well-governed. We want to remove every excuse governments might have for exerting their own political authority. At this point I might argue that Wikipedia's governance has failed in various ways, but the root problem is that Wikipedia is absolutely committed to anonymous contribution; this ultimately makes it impossible to enforce many rules effectively. However much oppressive bureaucracy Wikipedia layers on, it will always be possible for people to sidestep rules, simply by creating a new identity. The unreliability of Wikipedia's enforcement of its own rules, in turn, provides a deep explanation of the unreliability of its information. The pretentious mediocrities and ideologues, as well as the powerful vested interests—generally, anyone with a strong motive to make Wikipedia articles read their way—can always create new accounts if they are ousted. Wikipedia's content will remain unreliable, and it will continue to have various public scandals, because its governance is unreliable. And this, I'm afraid, opens Wikipedia up to the threat of government regulation. I wouldn't wish that on them, of course, and I don't mean to give anyone ideas.
After all, if the Citizendium's more sensible system succeeds, it will have the power to do far more damage than Wikipedia can. To get an idea of the damage Wikipedia can do, consider another example. In late 2005, John Seigenthaler, Sr., long-time editor of the American newspaper The Tennessean, was accused in a Wikipedia article of being complicit in the assassination of John F. Kennedy. Well, it was rather easy for him to protect his reputation by pointing out publicly how unreliable Wikipedia is. He simply shamed Wikipedia, and he came off looking quite good.
But imagine that Seigenthaler were accused by some better, more reliable source. Then he couldn't have gotten relief in this way; he no doubt would have had to sue. I hate the thought, but I have to concede that it is barely possible that the Citizendium could be sued for defamation. After all, the effect of defamation by a more credible source would be much more serious. Then the government might be called in, and this worries me.
As I said, my horror scenario is that the Citizendium grows up to be as influential as its potential implies, only to be overregulated by zealous governments with a weak notion of free speech. As I said at the beginning of this talk, I think cyber-polities can generally regulate themselves. But communities with poor internal governance may well incur some necessary correction by governments, if they violate copyright on a massive scale or if they permit, irresponsibly, a pattern of libel. Why should this be disturbing to me? Government intervention is perhaps all right when we are talking about child molesters on MySpace; but when we are talking about projects to sum up what is known, that is when more serious issues of free speech enter in.
You can think of government intervention in something like Wikipedia or the Citizendium as akin to government intervention in the content of academic lectures and the governance of universities. When this happens, what should be an unimpeded search for the truth risks becoming politicized and politically controlled.
But you can imagine, perhaps, a series of enormous scandals on Wikipedia that has government leaders calling for the project to be taken over by the Department of Education, or by some private entity that is nevertheless implicitly answerable to the government. Wikipedia is far from being in such a position now, but it is conceivable. The argument would go as follows:
End of argument. Nationalization might seem hard to conceive, but it has happened quite a bit in the last century. Why couldn't it happen to something that is already a free, public trust?
As both an academic (or former academic, anyway) and as an online project organizer, the thought of this scenario bothers me greatly, and in fact I must admit that I have given it no small amount of thought in the last few years. Fear of government intrusions on what should be a fully independent enterprise is one reason that I have spent so much time in the last year working on a sensible governance framework for the Citizendium. In short, the best protection against undue government interference in open content projects is good internal governance. So let me describe the Citizendium's current governance and its future plans.
The Citizendium works now under an explicit Statement of Fundamental Policies, which calls for the adoption of a Charter, not unlike a constitution, within the next few months. The Charter will no doubt solidify the governance system we are developing right now. This system involves an Editorial Council which is responsible for content policy; a Constabulary which gets new people on board and encourages good behavior; and a Judicial Board which will handle conflict resolution and appeals. While editors will make up the bulk of our Editorial Council, both authors and editors may participate in each of these bodies. Each of these bodies will have mutually exclusive membership, to help ensure a separation of powers, and there will be some other checks and balances. In addition, I as Editor-in-Chief am head of an Executive Committee. But to set a positive precedent, before even launching the Citizendium I have committed to stepping down within two to three years, so that we have an appropriate and regular succession of leadership.
Another perhaps interesting point concerns the Editorial Council. It has actually adopted a digitized version of Robert's Rules of Order, and we have passed five resolutions using e‑mail and the wiki exclusively. Recall that contributors must agree to uphold this system, as a condition of their participation. They must also be identified by their real-world identity if they wish to participate—although we will make exceptions in truly extraordinary cases.
I think you can recognize what we are trying to build: a traditional constitutional republic, but moved online. Only time will tell, but my hope is that this nascent governance structure will help us to avoid some of the problems that have beset not just Wikipedia, but a wide variety of Web communities.
I have covered a pretty wide variety of topics in my talk. I hope you have been able to follow the thread, at least a little; I doubt I have spent all the time I would need to make everything perfectly clear. But let me sum up my main argument anyway. Online communities, I say, are political entities. As such, they can govern their own "domestic" affairs, as well have various "foreign" or external effects. And so they can be democratic insofar as their members have authority internally or externally. I've discussed mainly one kind of authority, namely epistemic authority, or the authority over what society takes to be knowledge.
Then I pointed out that the external authority a project has depends on its internal governance—and so, the more externally influential, the more important it is that we get the internal governance right. I pointed to Wikipedia as an example of a cyber-polity that is not particularly well-governed. I worried a fair bit about the fallout, in terms of government regulation, that this might incur. In part to help avoid such fallout, I have briefly sketched a governance system that the Citizendium uses, which is a traditional constitutional, representative republic—mapped online.