nettime's roving reporter on Fri, 16 Nov 2001 22:08:34 +0100 (CET) |
[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]
<nettime> The Internet Under Siege |
[via: Digital Bearer Settlement List <dbs@philodox.com>] http://www.foreignpolicy.com/issue_novdec_2001/lessig.html The Internet Under Siege Who owns the Internet? Until recently, nobody. That's because, although the Internet was "Made in the U.S.A.," its unique design transformed it into a resource for innovation that anyone in the world could use. Today, however, courts and corporations are attempting to wall off portions of cyberspace. In so doing, they are destroying the Internet's potential to foster democracy and economic growth worldwide. By Lawrence Lessig The Internet revolution has ended just as surprisingly as it began. None expected the explosion of creativity that the network produced; few expected that explosion to collapse as quickly and profoundly as it has. The phenomenon has the feel of a shooting star, flaring unannounced across the night sky, then disappearing just as unexpectedly. Under the guise of protecting private property, a series of new laws and regulations are dismantling the very architecture that made the Internet a framework for global innovation. Neither the appearance nor disappearance of this revolution is difficult to understand. The difficulty is in accepting the lessons of the Internet's evolution. The Internet was born in the United States, but its success grew out of notions that seem far from the modern American ideals of property and the market. Americans are captivated by the idea, as explained by Yale Law School professor Carol Rose, that the world is best managed "when divided among private owners" and when the market perfectly regulates those divided resources. But the Internet took off precisely because core resources were not "divided among private owners." Instead, the core resources of the Internet were left in a "commons." It was this commons that engendered the extraordinary innovation that the Internet has seen. It is the enclosure of this commons that will bring about the Internet's demise. This commons was built into the very architecture of the original network. Its design secured a right of decentralized innovation. It was this "innovation commons" that produced the diversity of creativity that the network has seen within the United States and, even more dramatically, abroad. Many of the Internet innovations we now take for granted (not the least of which is the World Wide Web) were the creations of "outsiders"-foreign inventors who freely roamed the commons. Policymakers need to understand the importance of this architectural design to the innovation and creativity of the original network. The potential of the Internet has just begun to be realized, especially in the developing world, where many "real space" alternatives for commerce and innovation are neither free nor open. Yet old ways of thinking are reasserting themselves within the United States to modify this design. Changes to the Internet's original core will in turn threaten the network's potential everywhere-staunching the opportunity for innovation and creativity. Thus, at the moment this transformation could have a meaningful effect, a counterrevolution is succeeding in undermining the potential of this network. The motivation for this counterrevolution is as old as revolutions themselves. As Niccolò Machiavelli described long before the Internet, "Innovation makes enemies of all those who prospered under the old regime, and only lukewarm support is forthcoming from those who would prosper under the new." And so it is today with us. Those who prospered under the old regime are threatened by the Internet. Those who would prosper under the new regime have not risen to defend it against the old; whether they will is still a question. So far, it appears they will not. The Neutral Zone A "commons" is a resource to which everyone within a relevant community has equal access. It is a resource that is not, in an important sense, "controlled." Private or state-owned property is a controlled resource; only as the owner specifies may that property be used. But a commons is not subject to this sort of control. Neutral or equal restrictions may apply to it (an entrance fee to a park, for example) but not the restrictions of an owner. A commons, in this sense, leaves its resources "free." Commons are features of all cultures. They have been especially important to cultures outside the United States-from communal tenure systems in Switzerland and Japan to irrigation communities within the Philippines. But within American intellectual culture, commons are treated as imperfect resources. They are the object of "tragedy," as ecologist Garrett Hardin famously described. Wherever a commons exists, the aim is to enclose it. In the American psyche, commons are unnecessary vestiges from times past and best removed, if possible. For most resources, for most of the time, the bias against commons makes good sense. When resources are left in common, individuals may be driven to overconsume, and therefore deplete, them. But for some resources, the bias against commons is blinding. Some resources are not subject to the "tragedy of the commons" because some resources cannot be "depleted." (No matter how much we use Einstein's theories of relativity or copy Robert Frost's poem "New Hampshire," those resources will survive.) For these resources, the challenge is to induce provision, not to avoid depletion. The problems of provision are very different from the problems of depletion-confusing the two only leads to misguided policies. This confusion is particularly acute when considering the Internet. At the core of the Internet is a design (chosen without a clear sense of its consequences) that was new among large-scale computer and communications networks. Named the "end-to-end argument" by network theorists Jerome Saltzer, David Clark, and David Reed in 1984, this design influences where "intelligence" in the network is placed. Traditional computer-communications systems located intelligence, and hence control, within the network itself. Networks were "smart"; they were designed by people who believed they knew exactly what the network would be used for. But the Internet was born at a time when a different philosophy was taking shape within computer science. This philosophy ranked humility above omniscience and anticipated that network designers would have no clear idea about all the ways the network could be used. It therefore counseled a design that built little into the network itself, leaving the network free to develop as the ends (the applications) wanted. The motivation for this new design was flexibility. The consequence was innovation. Because innovators needed no permission from the network owner before different applications or content got served across the network, innovators were freer to develop new modes of connection. Technically, the network achieved this design simply by focusing on the delivery of packets of data, oblivious to either the contents of the packets or their owners. Nor does the network concern itself that all the packets make their way to the other side. The network is "best efforts"; anything more is provided by the applications at both ends. Like an efficient post office (imagine!), the system simply forwards the data along. Since the network was not optimized for any single application or service, the Internet remained open to new innovation. The World Wide Web is perhaps the best example. The Web was the creation of computer scientist Tim Berners-Lee at the European Organization for Nuclear Research (CERN) laboratory in Geneva in late 1990. Berners-Lee wanted to enable users on a network to have easy access to documents located elsewhere on the network. He therefore developed a set of protocols to enable hypertext links among documents located across the network. Because of end-to-end, these protocols could be layered on top of the initial protocols of the Internet. This meant the Internet could grow to embrace the Web. Had the network compromised its commitment to end-to-end-had its design been optimized to favor telephony, for example, as many in the 1980s wanted-then the Web would not have been possible. This end-to-end design is the "core" of the Internet. If we can think of the network as built in layers, then the end-to-end design was created by a set of protocols implemented at the middle layer-what we might call the logical, or code layer, of the Internet. Below the code layer is a physical layer (computers and the wires that link them). Above the code layer is a content layer (material that gets served across the network). Not all these layers were organized as commons. The computers at the physical layer are private property, not "free" in the sense of a commons. Much of the content served across the network is protected by copyright. It, too, is not "free." At the code layer, however, the Internet is a commons. By design, no one controls the resources for innovation that get served across this layer. Individuals control the physical layer, deciding whether a machine or network gets connected to the Internet. But once connected, at least under the Internet's original design, the innovation resources for the network remained free. No other large scale network left the code layer free in this way. For most of the history of telephone monopolies worldwide, permission to innovate on the telephone platform was vigorously controlled. In the United States in 1956, AT&T successfully persuaded the U.S. Federal Communications Commission to block the use of a plastic cup on a telephone receiver, designed to block noise from the telephone microphone, on the theory that AT&T alone had the right to innovation on the telephone network. The Internet might have remained an obscure tool of government-backed researchers if the telephone company had maintained this control. The Internet would never have taken off if ordinary individuals had been unable to connect to the network by way of Internet service providers (ISPs) through already existing telephone lines. Yet this right to connect was not preordained. It is here that an accident in regulatory history played an important role. Just at the moment the Internet was emerging, the telephone monopoly was being moved to a different regulatory paradigm. Previously, the telephone monopoly was essentially free to control its wires as it wished. Beginning in the late 1960s, and then more vigorously throughout the 1980s, the government began to require that the telephone industry behave neutrally-first by insisting that telephone companies permit customer premises equipment (such as modems) to be connected to the network, and then by requiring that telephone companies allow others to have access to their wires. This kind of regulation was rare among telecommunications monopolies worldwide. In Europe and throughout the world, telecommunications monopolies were permitted to control the uses of their networks. No requirement of access operated to enable competition. Thus no system of competition grew up around these other monopolies. But when the United States broke up AT&T in 1984, the resulting companies no longer had the freedom to discriminate against other uses of their lines. And when ISPs sought access to the local Bell lines to enable customers to connect to the Internet, the local Bells were required to grant access equally. This enabled a vigorous competition in Internet access, and this competition meant that the network could not behave strategically against this new technology. In effect, through a competitive market, an end-to-end design was created at the physical layer of the telephone network, which meant that an end-to-end design could be layered on top of that. This innovation commons was thus layered onto a physical infrastructure that, through regulation, had important commons-like features. Common-carrier regulation of the telephone system assured that the system could not discriminate against an emerging competitor, the Internet. And the Internet itself was created, through its end-to-end design, to assure that no particular application or use could discriminate against any other innovations. Neutrality existed at the physical and code layer of the Internet. An important neutrality also existed at the content layer of the Internet. This layer includes all the content streamed across the network-Web pages, MP3s, e-mail, streaming video-as well as application programs that run on, or feed, the network. These programs are distinct from the protocols at the code layer, collectively referred to as TCP/IP (including the protocols of the World Wide Web). TCP/IP is dedicated to the public domain. Recent FP Articles on the Internet But the code above these protocols is not in the public domain. It is, instead, of two sorts: proprietary and nonproprietary. The proprietary includes the familiar Microsoft operating systems and Web servers, as well as programs from other software companies. The nonproprietary includes open source and free software, especially the Linux (or GNU/Linux) operating system, the Apache server, as well as a host of other plumbing-oriented code that makes the Net run. Nonproprietary code creates a commons at the content layer. The commons here is not just the resource that a particular program might provide-for example, the functionality of an operating system or Web server. The commons also includes the source code of software that can be drawn upon and modified by others. Open source and free software ("open code" for short) must be distributed with the source code. The source code must be free for others to take and modify. This commons at the content layer means that others can take and build upon open source and free software. It also means that open code can't be captured and tilted against any particular competitor. Open code can always be modified by subsequent adopters. It, therefore, is licensed to remain neutral among subsequent uses. There is no "owner" of an open code project. In this way, and again, parallel to the end-to-end principle at the code layer, open code decentralizes innovation. It keeps a platform neutral. This neutrality in turn inspires innovators to build for that platform because they need not fear the platform will turn against them. Open code builds a commons for innovation at the content layer. Like the commons at the code layer, open code preserves the opportunity for innovation and protects innovation against the strategic behavior of competitors. Free resources induce innovation. An Engine of Innovation The original Internet, as it was extended to society generally, mixed controlled and free resources at each layer of the network. At the core code layer, the network was free. The end-to-end design assured that no network owner could exercise control over the network. At the physical layer, the resources were essentially controlled, but even here, important aspects were free. One had the right to connect a machine to the network or not, but telephone companies didn't have the right to discriminate against this particular use of their network. And finally, at the content layer, many of the resources served across the Internet were controlled. But a crucial range of software building essential services on the Internet remained free. Whether through an open source or free software license, these resources could not be controlled. This balance of control and freedom produced an unprecedented explosion in innovation. The power, and hence the right, to innovate was essentially decentralized. The Internet might have been an American invention, but creators from around the world could build upon this network platform. Significantly, some of the most important innovations for the Internet came from these "outsiders." As noted, the most important technology for accessing and browsing the Internet (the World Wide Web) was not invented by companies specializing in network access. It wasn't America Online (AOL) or Compuserve. The Web was developed by a researcher in a Swiss laboratory who first saw its potential and then fought to bring it to fruition. Likewise, it wasn't existing e-mail providers who came up with the idea of Web-based e-mail. That was co-created by an immigrant to the United States from India, Sabeer Bhatia, and it gave birth to one of the fastest growing communities in history-Hotmail. And it wasn't traditional network providers or telephone companies that invented the applications that enabled online chatting to take off. The original community-based chatting service (ICQ) was the invention of an Israeli, far from the trenches of network design. His service could explode (and then be purchased by AOL for $400 million) only because the network was left open for this type of innovation. Similarly, the revolution in bookselling initiated by Amazon.com (through the use of technologies that "match preferences" of customers) was invented far from the traditional organs of publishers. By gathering a broad range of data about purchases by customers, Amazon-drawing upon technology first developed at MIT and the University of Minnesota to filter Usenet news-can predict what a customer is likely to want. These recommendations drive sales, but without the high cost of advertising or promotion. Consequently, booksellers such as Amazon can outcompete traditional marketers of books, which may account for the rapid expansion of Amazon into Asia and Europe. These innovations are at the level of Internet services. Far more profound have been innovations at the level of content. The Internet has not only inspired invention, it has also inspired publication in a way that would never have been produced by the world of existing publishers. The creation of online archives of lyrics and chord sequences and of collaborative databases collecting information about compact discs and movies demonstrates the kind of creativity that was possible because the right to create was not controlled. Again, the innovations have not been limited to the United States. OpenDemocracy.org, for example, is a London-based, Web-centered forum for debate and exchange about democracy and governance throughout the world. Such a forum is possible only because no coordination among international actors is needed. And it thrives because it can engender debate at a low cost. This history should be a lesson. Every significant innovation on the Internet has emerged outside of traditional providers. The new grows away from the old. This trend teaches the value of leaving the platform open for innovation. Unfortunately, that platform is now under siege. Every technological disruption creates winners and losers. The losers have an interest in avoiding that disruption if they can. This was the lesson Machiavelli taught, and it is the experience with every important technological change over time. It is also what we are now seeing with the Internet. The innovation commons of the Internet threatens important and powerful pre-Internet interests. During the past five years, those interests have mobilized to launch a counterrevolution that is now having a global impact. This movement is fueled by pressure at both the physical and content layers of the network. These changes, in turn, put pressure on the freedom of the code layer. These changes will have an effect on the opportunity for growth and innovation that the Internet presents. Policymakers keen to protect that growth should be skeptical of changes that will threaten it. Broad-based innovation may threaten the profits of some existing interests, but the social gains from this unpredictable growth will far outstrip the private losses, especially in nations just beginning to connect. Fencing Off the Commons The Internet took off on telephone lines. Narrowband service across acoustic modems enabled millions of computers to connect through thousands of ISPs. Local telephone service providers had to provide ISPs with access to local wires; they were not permitted to discriminate against Internet service. Thus the physical platform on which the Internet was born was regulated to remain neutral. This regulation had an important effect. A nascent industry could be born on the telephone wires, regardless of the desires of telephone companies. But as the Internet moves from narrowband to broadband, the regulatory environment is changing. The dominant broadband technology in the United States is currently cable. Cable lives under a different regulatory regime. Cable providers in general have no obligation to grant access to their facilities. And cable has asserted the right to discriminate in the Internet service it provides. Consequently, cable has begun to push for a different set of principles at the code layer of the network. Cable companies have deployed technologies to enable them to engage in a form of discrimination in the service they provide. Cisco, for example, developed "policy-based routers" that enable cable companies to choose which content flows quickly and which flows slowly. With these, and other technologies, cable companies will be in a position to exercise power over the content and applications that operate on their networks. This control has already begun in the United States. ISPs running cable services have exercised their power to ban certain kinds of applications (specifically, those that enable peer-to-peer service). They have blocked particular content (advertising from competitors, for example) when that content was not consistent with their business model. The model for these providers is the model of cable television generally-controlling access and content to the cable providers' end. The environment of innovation on the original network will change according to the extent that cable becomes the primary mode of access to the Internet. Rather than a network that vests intelligence in the ends, the cable-dominated network will vest an increasing degree of intelligence within the network itself. And to the extent it does this, the network will increase the opportunity for strategic behavior in favor of some technologies and against others. An essential feature of neutrality at the code layer will have been compromised, reducing the opportunity for innovation worldwide. Far more dramatic, however, has been the pressure from the content layer on the code layer. This pressure has come in two forms. First, and most directly related to the content described above, there has been an explosion of patent regulation in the context of software. Second, copyright holders have exercised increasing control over new technologies for distribution. The changes in patent regulation are more difficult to explain, though the consequence is not hard to track. Two decades ago, the U.S. Patent Office began granting patents for software-like inventions. In the late 1990s, the court overseeing these patents finally approved the practice and approved their extension to "business methods." The European Union (EU), meanwhile, initially adopted a more skeptical attitude toward software patents. But pressure from the United States will eventually bring the EU into alignment with American policy. In principle, these patents are designed to spur innovation. But with sequential and complementary innovation, little evidence exists that suggests such patents will do any good, and there is increasing evidence that they will do harm. Like any regulation, patents tax the innovative process generally. As with any tax, some firms-large rather than small, U.S. rather than foreign-are better able to bear that tax than others. Open code projects, in particular, are threatened by this trend, as they are least able to negotiate appropriate patent licenses. The most dramatic restrictions on innovation, however, have come at the hands of copyright holders. Copyright is designed to ensure that artists control their "writings" for a limited time. The aim is to secure to copyright holders a sufficient interest to produce new work. But copyright laws were crafted in an era long before the Internet. And their effect on the Internet has been to transfer control over innovation in distribution from many innovators to a concentrated few. The clearest example of this effect is online music. Before the Internet, the production and distribution of music had become extraordinarily concentrated. In 2000, for example, five companies controlled 84 percent of music distribution in the world. The reasons for this concentration are many-including the high costs of promotion-but the effect of concentration on artist development is profound. Very few artists make any money from their work, and the few that do are able to do so because of mass marketing from record labels. The Internet had the potential to change this reality. Both because the costs of distribution were so low, and because the network also had the potential to significantly lower the costs of promotion, the cost of music could fall, and revenues to artists could rise. Five years ago, this market took off. A large number of online music providers began competing for new ways to distribute music. Some distributed MP3s for money (eMusic.com). Some built technology for giving owners of music easier access to their music (mp3.com). And some made it much easier for ordinary users to "share" their music with other users (Napster). But as quickly as these companies took off, lawyers representing old media succeeded in shutting them down. These lawyers argued that copyright law gave the holders (some say hoarders) of these copyrights the exclusive right to control how they get used. American courts agreed. To keep this dispute in context, we should think about the last example of a technological change that facilitated a much different model for distributing content: cable TV, which has been accurately hailed as the first great Napster. Owners of cable television systems essentially set up antenna and "stole" over-the-air broadcasts and then sold that "stolen property" to their customers. But when U.S. courts were asked to stop this "theft," they refused. Twice the U.S. Supreme Court held that this use of someone else's copyrighted material was not inconsistent with copyright law. When the U.S. Congress finally got around to changing the law, it struck an importantly illustrative balance. Congress granted copyright owners the right to compensation from the use of their material on cable broadcasts, but cable companies were given the right to broadcast the copyrighted material. The reason for this balance is not hard to see. Copyright owners certainly are entitled to compensation for their work. But the right to compensation shouldn't translate into the power to control innovation. Rather than giving copyright holders the right to veto a particular new use of their work (in this case, because it would compete with over-the-air broadcasting), Congress assured copyright owners would get paid without having the power to control-compensation without control. The same deal could have been struck by Congress in the context of online music. But this time, the courts did not hesitate to extend control to the copyright holders. So the concentrated holders of these copyrights were able to stop the deployment of competing distributors. And Congress was not motivated to respond by granting an equivalent compulsory right. The aim of the recording company's strategy was plain enough: shut down these new and competing models of distribution and replace them with a model for distributing music online more consistent with the traditional model. This trend has been supported by the actions of Congress. In 1998, Congress passed the Digital Millennium Copyright Act (DMCA), which (in)famously banned technologies designed to circumvent copyright protection technologies and also created strong incentives for ISPs to remove from their sites any material claimed to be a violation of copyright. On the surface both changes seem sensible enough. Copyright protection technologies are analogous to locks. What right does anyone have to pick a lock? And ISPs are in the best position to assure that copyright violations don't occur on their Web sites. Why not create incentives for them to remove infringing copyrighted material? But intuitions here mislead. A copyright protection technology is just code that controls access to copyrighted material. But that code can restrict access more effectively (and certainly less subtly) than copyright law does. Often the desire to crack protection systems is nothing more than a desire to exercise what is sometimes called a fair-use right over the copyrighted material. Yet the DMCA bans that technology, regardless of its ultimate effect. More troubling, however, is that the DMCA effectively bans this technology on a worldwide basis. Russian programmer Dimitry Sklyarov, for example, wrote code to crack Adobe's eBook technology in order to enable users to move eBooks from one machine to another and to give blind consumers the ability to "read" out loud the books they purchased. The code Sklyarov wrote was legal where it was written, but when it was sold by his company in the United States, it became illegal. When he came to the United States in July 2001 to talk about that code, the FBI arrested him. Today Sklyarov faces a sentence of 25 years for writing code that could be used for fair-use purposes, as well as to violate copyright laws. Similar trouble has arisen with the provision that gives ISPs the incentive to take down infringing copyrighted material. When an ISP is notified that material on its site violates copyright, it can avoid liability if it removes the material. As it doesn't have any incentive to expose itself to liability, the ordinary result of such notification is for the ISP to remove the material. Increasingly, companies trying to protect themselves from criticism have used this provision to silence critics. In August 2001, for example, a British pharmaceutical company invoked the DMCA in order to force an ISP to shut down an animal rights site that criticized the British company. Said the ISP, "It's very clear [the British company] just wants to shut them up," but ISPs have no incentive to resist the claims. In all these cases, there is a common pattern. In the push to give copyright owners control over their content, copyright holders also receive the ability to protect themselves against innovations that might threaten existing business models. The law becomes a tool to assure that new innovations don't displace old ones-when instead, the aim of copyright and patent law should be, as the U.S. Constitution requires, to "promote the progress of science and useful arts." These regulations will not only affect Americans. The expanding jurisdiction that American courts claim, combined with the push by the World Intellectual Property Organization to enact similar legislation elsewhere, means that the impact of this sort of control will be felt worldwide. There is no "local" when it comes to corruption of the Internet's basic principles. As these changes weaken the open source and free software movements, countries with the most to gain from a free and open platform lose. Those affected will include nations in the developing world and nations that do not want to cede control to a single private corporation. And as content becomes more controlled, nations that could otherwise benefit from vigorous competition in the delivery and production of content will also lose. An explosion of innovation to deliver MP3s would directly translate into innovation to deliver telephone calls and video content. Lowering the cost of this medium would dramatically benefit nations that still suffer from weak technical infrastructures. Policymakers around the world must recognize that the interests most strongly protected by the Internet counterrevolution are not their own. They should be skeptical of legal mechanisms that enable those most threatened by the innovation commons to resist it. The Internet promised the world-particularly the weakest in the world-the fastest and most dramatic change to existing barriers to growth. That promise depends on the network remaining open to innovation. That openness depends upon policy that better understands the Internet's past. Lawrence Lessig is professor of law at Stanford University. He is author of The Future of Ideas: The Fate of the Commons in the Connected World (New York: Random House, 2001) and Code and Other Laws of Cyberspace (New York: Basic Books, 1999). # distributed via <nettime>: no commercial use without permission # <nettime> is a moderated mailing list for net criticism, # collaborative text filtering and cultural politics of the nets # more info: majordomo@bbs.thing.net and "info nettime-l" in the msg body # archive: http://www.nettime.org contact: nettime@bbs.thing.net