nettime's roving reporter on Fri, 16 Nov 2001 22:08:34 +0100 (CET)


[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]

<nettime> The Internet Under Siege


[via: Digital Bearer Settlement List <dbs@philodox.com>]


http://www.foreignpolicy.com/issue_novdec_2001/lessig.html


The Internet Under Siege

Who owns the Internet? Until recently, nobody. That's because, although
the Internet was "Made in the U.S.A.," its unique design transformed it
into a resource for innovation that anyone in the world could use. Today,
however, courts and corporations are attempting to wall off portions of
cyberspace. In so doing, they are destroying the Internet's potential to
foster democracy and economic growth worldwide.

By Lawrence Lessig

The Internet revolution has ended just as surprisingly as it began. None
expected the explosion of creativity that the network produced; few
expected that explosion to collapse as quickly and profoundly as it has.
The phenomenon has the feel of a shooting star, flaring unannounced across
the night sky, then disappearing just as unexpectedly. Under the guise of
protecting private property, a series of new laws and regulations are
dismantling the very architecture that made the Internet a framework for
global innovation.

Neither the appearance nor disappearance of this revolution is difficult
to understand. The difficulty is in accepting the lessons of the
Internet's evolution. The Internet was born in the United States, but its
success grew out of notions that seem far from the modern American ideals
of property and the market. Americans are captivated by the idea, as
explained by Yale Law School professor Carol Rose, that the world is best
managed "when divided among private owners" and when the market perfectly
regulates those divided resources. But the Internet took off precisely
because core resources were not "divided among private owners." Instead,
the core resources of the Internet were left in a "commons." It was this
commons that engendered the extraordinary innovation that the Internet has
seen. It is the enclosure of this commons that will bring about the
Internet's demise.

This commons was built into the very architecture of the original network.
Its design secured a right of decentralized innovation. It was this
"innovation commons" that produced the diversity of creativity that the
network has seen within the United States and, even more dramatically,
abroad. Many of the Internet innovations we now take for granted (not the
least of which is the World Wide Web) were the creations of
"outsiders"-foreign inventors who freely roamed the commons. Policymakers
need to understand the importance of this architectural design to the
innovation and creativity of the original network. The potential of the
Internet has just begun to be realized, especially in the developing
world, where many "real space" alternatives for commerce and innovation
are neither free nor open.

Yet old ways of thinking are reasserting themselves within the United
States to modify this design. Changes to the Internet's original core will
in turn threaten the network's potential everywhere-staunching the
opportunity for innovation and creativity. Thus, at the moment this
transformation could have a meaningful effect, a counterrevolution is
succeeding in undermining the potential of this network.

The motivation for this counterrevolution is as old as revolutions
themselves. As Niccolò Machiavelli described long before the Internet,
"Innovation makes enemies of all those who prospered under the old regime,
and only lukewarm support is forthcoming from those who would prosper
under the new." And so it is today with us. Those who prospered under the
old regime are threatened by the Internet. Those who would prosper under
the new regime have not risen to defend it against the old; whether they
will is still a question. So far, it appears they will not.

The Neutral Zone

A "commons" is a resource to which everyone within a relevant community
has equal access. It is a resource that is not, in an important sense,
"controlled." Private or state-owned property is a controlled resource;
only as the owner specifies may that property be used. But a commons is
not subject to this sort of control. Neutral or equal restrictions may
apply to it (an entrance fee to a park, for example) but not the
restrictions of an owner. A commons, in this sense, leaves its resources
"free."

Commons are features of all cultures. They have been especially important
to cultures outside the United States-from communal tenure systems in
Switzerland and Japan to irrigation communities within the Philippines.
But within American intellectual culture, commons are treated as imperfect
resources. They are the object of "tragedy," as ecologist Garrett Hardin
famously described. Wherever a commons exists, the aim is to enclose it.
In the American psyche, commons are unnecessary vestiges from times past
and best removed, if possible.

For most resources, for most of the time, the bias against commons makes
good sense. When resources are left in common, individuals may be driven
to overconsume, and therefore deplete, them. But for some resources, the
bias against commons is blinding. Some resources are not subject to the
"tragedy of the commons" because some resources cannot be "depleted." (No
matter how much we use Einstein's theories of relativity or copy Robert
Frost's poem "New Hampshire," those resources will survive.) For these
resources, the challenge is to induce provision, not to avoid depletion.
The problems of provision are very different from the problems of
depletion-confusing the two only leads to misguided policies.

This confusion is particularly acute when considering the Internet. At the
core of the Internet is a design (chosen without a clear sense of its
consequences) that was new among large-scale computer and communications
networks. Named the "end-to-end argument" by network theorists Jerome
Saltzer, David Clark, and David Reed in 1984, this design influences where
"intelligence" in the network is placed. Traditional
computer-communications systems located intelligence, and hence control,
within the network itself. Networks were "smart"; they were designed by
people who believed they knew exactly what the network would be used for.

But the Internet was born at a time when a different philosophy was taking
shape within computer science. This philosophy ranked humility above
omniscience and anticipated that network designers would have no clear
idea about all the ways the network could be used. It therefore counseled
a design that built little into the network itself, leaving the network
free to develop as the ends (the applications) wanted.

The motivation for this new design was flexibility. The consequence was
innovation. Because innovators needed no permission from the network owner
before different applications or content got served across the network,
innovators were freer to develop new modes of connection. Technically, the
network achieved this design simply by focusing on the delivery of packets
of data, oblivious to either the contents of the packets or their owners.
Nor does the network concern itself that all the packets make their way to
the other side. The network is "best efforts"; anything more is provided
by the applications at both ends. Like an efficient post office
(imagine!), the system simply forwards the data along.

Since the network was not optimized for any single application or service,
the Internet remained open to new innovation. The World Wide Web is
perhaps the best example. The Web was the creation of computer scientist
Tim Berners-Lee at the European Organization for Nuclear Research (CERN)
laboratory in Geneva in late 1990. Berners-Lee wanted to enable users on a
network to have easy access to documents located elsewhere on the network.
He therefore developed a set of protocols to enable hypertext links among
documents located across the network. Because of end-to-end, these
protocols could be layered on top of the initial protocols of the
Internet. This meant the Internet could grow to embrace the Web. Had the
network compromised its commitment to end-to-end-had its design been
optimized to favor telephony, for example, as many in the 1980s
wanted-then the Web would not have been possible.

This end-to-end design is the "core" of the Internet. If we can think of
the network as built in layers, then the end-to-end design was created by
a set of protocols implemented at the middle layer-what we might call the
logical, or code layer, of the Internet. Below the code layer is a
physical layer (computers and the wires that link them). Above the code
layer is a content layer (material that gets served across the network).
Not all these layers were organized as commons. The computers at the
physical layer are private property, not "free" in the sense of a commons.
Much of the content served across the network is protected by copyright.
It, too, is not "free."

At the code layer, however, the Internet is a commons. By design, no one
controls the resources for innovation that get served across this layer.
Individuals control the physical layer, deciding whether a machine or
network gets connected to the Internet. But once connected, at least under
the Internet's original design, the innovation resources for the network
remained free.

No other large scale network left the code layer free in this way. For
most of the history of telephone monopolies worldwide, permission to
innovate on the telephone platform was vigorously controlled. In the
United States in 1956, AT&T successfully persuaded the U.S. Federal
Communications Commission to block the use of a plastic cup on a telephone
receiver, designed to block noise from the telephone microphone, on the
theory that AT&T alone had the right to innovation on the telephone
network.

The Internet might have remained an obscure tool of government-backed
researchers if the telephone company had maintained this control. The
Internet would never have taken off if ordinary individuals had been
unable to connect to the network by way of Internet service providers
(ISPs) through already existing telephone lines. Yet this right to connect
was not preordained. It is here that an accident in regulatory history
played an important role. Just at the moment the Internet was emerging,
the telephone monopoly was being moved to a different regulatory paradigm.
Previously, the telephone monopoly was essentially free to control its
wires as it wished. Beginning in the late 1960s, and then more vigorously
throughout the 1980s, the government began to require that the telephone
industry behave neutrally-first by insisting that telephone companies
permit customer premises equipment (such as modems) to be connected to the
network, and then by requiring that telephone companies allow others to
have access to their wires.

This kind of regulation was rare among telecommunications monopolies
worldwide. In Europe and throughout the world, telecommunications
monopolies were permitted to control the uses of their networks. No
requirement of access operated to enable competition. Thus no system of
competition grew up around these other monopolies. But when the United
States broke up AT&T in 1984, the resulting companies no longer had the
freedom to discriminate against other uses of their lines. And when ISPs
sought access to the local Bell lines to enable customers to connect to
the Internet, the local Bells were required to grant access equally. This
enabled a vigorous competition in Internet access, and this competition
meant that the network could not behave strategically against this new
technology. In effect, through a competitive market, an end-to-end design
was created at the physical layer of the telephone network, which meant
that an end-to-end design could be layered on top of that.

This innovation commons was thus layered onto a physical infrastructure
that, through regulation, had important commons-like features.
Common-carrier regulation of the telephone system assured that the system
could not discriminate against an emerging competitor, the Internet. And
the Internet itself was created, through its end-to-end design, to assure
that no particular application or use could discriminate against any other
innovations. Neutrality existed at the physical and code layer of the
Internet.

An important neutrality also existed at the content layer of the Internet.
This layer includes all the content streamed across the network-Web pages,
MP3s, e-mail, streaming video-as well as application programs that run on,
or feed, the network. These programs are distinct from the protocols at
the code layer, collectively referred to as TCP/IP (including the
protocols of the World Wide Web). TCP/IP is dedicated to the public
domain. Recent FP Articles on the Internet


But the code above these protocols is not in the public domain. It is,
instead, of two sorts: proprietary and nonproprietary. The proprietary
includes the familiar Microsoft operating systems and Web servers, as well
as programs from other software companies. The nonproprietary includes
open source and free software, especially the Linux (or GNU/Linux)
operating system, the Apache server, as well as a host of other
plumbing-oriented code that makes the Net run.

Nonproprietary code creates a commons at the content layer. The commons
here is not just the resource that a particular program might provide-for
example, the functionality of an operating system or Web server. The
commons also includes the source code of software that can be drawn upon
and modified by others. Open source and free software ("open code" for
short) must be distributed with the source code. The source code must be
free for others to take and modify. This commons at the content layer
means that others can take and build upon open source and free software.
It also means that open code can't be captured and tilted against any
particular competitor. Open code can always be modified by subsequent
adopters. It, therefore, is licensed to remain neutral among subsequent
uses. There is no "owner" of an open code project.

In this way, and again, parallel to the end-to-end principle at the code
layer, open code decentralizes innovation. It keeps a platform neutral.
This neutrality in turn inspires innovators to build for that platform
because they need not fear the platform will turn against them. Open code
builds a commons for innovation at the content layer. Like the commons at
the code layer, open code preserves the opportunity for innovation and
protects innovation against the strategic behavior of competitors. Free
resources induce innovation.

An Engine of Innovation

The original Internet, as it was extended to society generally, mixed
controlled and free resources at each layer of the network. At the core
code layer, the network was free. The end-to-end design assured that no
network owner could exercise control over the network. At the physical
layer, the resources were essentially controlled, but even here, important
aspects were free. One had the right to connect a machine to the network
or not, but telephone companies didn't have the right to discriminate
against this particular use of their network. And finally, at the content
layer, many of the resources served across the Internet were controlled.
But a crucial range of software building essential services on the
Internet remained free. Whether through an open source or free software
license, these resources could not be controlled.

This balance of control and freedom produced an unprecedented explosion in
innovation. The power, and hence the right, to innovate was essentially
decentralized. The Internet might have been an American invention, but
creators from around the world could build upon this network platform.
Significantly, some of the most important innovations for the Internet
came from these "outsiders."

As noted, the most important technology for accessing and browsing the
Internet (the World Wide Web) was not invented by companies specializing
in network access. It wasn't America Online (AOL) or Compuserve. The Web
was developed by a researcher in a Swiss laboratory who first saw its
potential and then fought to bring it to fruition. Likewise, it wasn't
existing e-mail providers who came up with the idea of Web-based e-mail.
That was co-created by an immigrant to the United States from India,
Sabeer Bhatia, and it gave birth to one of the fastest growing communities
in history-Hotmail.

And it wasn't traditional network providers or telephone companies that
invented the applications that enabled online chatting to take off. The
original community-based chatting service (ICQ) was the invention of an
Israeli, far from the trenches of network design. His service could
explode (and then be purchased by AOL for $400 million) only because the
network was left open for this type of innovation.

Similarly, the revolution in bookselling initiated by Amazon.com (through
the use of technologies that "match preferences" of customers) was
invented far from the traditional organs of publishers. By gathering a
broad range of data about purchases by customers, Amazon-drawing upon
technology first developed at MIT and the University of Minnesota to
filter Usenet news-can predict what a customer is likely to want. These
recommendations drive sales, but without the high cost of advertising or
promotion. Consequently, booksellers such as Amazon can outcompete
traditional marketers of books, which may account for the rapid expansion
of Amazon into Asia and Europe.

These innovations are at the level of Internet services. Far more profound
have been innovations at the level of content. The Internet has not only
inspired invention, it has also inspired publication in a way that would
never have been produced by the world of existing publishers. The creation
of online archives of lyrics and chord sequences and of collaborative
databases collecting information about compact discs and movies
demonstrates the kind of creativity that was possible because the right to
create was not controlled.

Again, the innovations have not been limited to the United States.
OpenDemocracy.org, for example, is a London-based, Web-centered forum for
debate and exchange about democracy and governance throughout the world.
Such a forum is possible only because no coordination among international
actors is needed. And it thrives because it can engender debate at a low
cost.

This history should be a lesson. Every significant innovation on the
Internet has emerged outside of traditional providers. The new grows away
from the old. This trend teaches the value of leaving the platform open
for innovation. Unfortunately, that platform is now under siege. Every
technological disruption creates winners and losers. The losers have an
interest in avoiding that disruption if they can. This was the lesson
Machiavelli taught, and it is the experience with every important
technological change over time. It is also what we are now seeing with the
Internet. The innovation commons of the Internet threatens important and
powerful pre-Internet interests. During the past five years, those
interests have mobilized to launch a counterrevolution that is now having
a global impact.

This movement is fueled by pressure at both the physical and content
layers of the network. These changes, in turn, put pressure on the freedom
of the code layer. These changes will have an effect on the opportunity
for growth and innovation that the Internet presents. Policymakers keen to
protect that growth should be skeptical of changes that will threaten it.
Broad-based innovation may threaten the profits of some existing
interests, but the social gains from this unpredictable growth will far
outstrip the private losses, especially in nations just beginning to
connect.

Fencing Off the Commons

The Internet took off on telephone lines. Narrowband service across
acoustic modems enabled millions of computers to connect through thousands
of ISPs. Local telephone service providers had to provide ISPs with access
to local wires; they were not permitted to discriminate against Internet
service. Thus the physical platform on which the Internet was born was
regulated to remain neutral. This regulation had an important effect. A
nascent industry could be born on the telephone wires, regardless of the
desires of telephone companies.

But as the Internet moves from narrowband to broadband, the regulatory
environment is changing. The dominant broadband technology in the United
States is currently cable. Cable lives under a different regulatory
regime. Cable providers in general have no obligation to grant access to
their facilities. And cable has asserted the right to discriminate in the
Internet service it provides.

Consequently, cable has begun to push for a different set of principles at
the code layer of the network. Cable companies have deployed technologies
to enable them to engage in a form of discrimination in the service they
provide. Cisco, for example, developed "policy-based routers" that enable
cable companies to choose which content flows quickly and which flows
slowly. With these, and other technologies, cable companies will be in a
position to exercise power over the content and applications that operate
on their networks.

This control has already begun in the United States. ISPs running cable
services have exercised their power to ban certain kinds of applications
(specifically, those that enable peer-to-peer service). They have blocked
particular content (advertising from competitors, for example) when that
content was not consistent with their business model. The model for these
providers is the model of cable television generally-controlling access
and content to the cable providers' end.

The environment of innovation on the original network will change
according to the extent that cable becomes the primary mode of access to
the Internet. Rather than a network that vests intelligence in the ends,
the cable-dominated network will vest an increasing degree of intelligence
within the network itself. And to the extent it does this, the network
will increase the opportunity for strategic behavior in favor of some
technologies and against others. An essential feature of neutrality at the
code layer will have been compromised, reducing the opportunity for
innovation worldwide.

Far more dramatic, however, has been the pressure from the content layer
on the code layer. This pressure has come in two forms. First, and most
directly related to the content described above, there has been an
explosion of patent regulation in the context of software. Second,
copyright holders have exercised increasing control over new technologies
for distribution.

The changes in patent regulation are more difficult to explain, though the
consequence is not hard to track. Two decades ago, the U.S. Patent Office
began granting patents for software-like inventions. In the late 1990s,
the court overseeing these patents finally approved the practice and
approved their extension to "business methods." The European Union (EU),
meanwhile, initially adopted a more skeptical attitude toward software
patents. But pressure from the United States will eventually bring the EU
into alignment with American policy.

In principle, these patents are designed to spur innovation. But with
sequential and complementary innovation, little evidence exists that
suggests such patents will do any good, and there is increasing evidence
that they will do harm. Like any regulation, patents tax the innovative
process generally. As with any tax, some firms-large rather than small,
U.S. rather than foreign-are better able to bear that tax than others.
Open code projects, in particular, are threatened by this trend, as they
are least able to negotiate appropriate patent licenses.

The most dramatic restrictions on innovation, however, have come at the
hands of copyright holders. Copyright is designed to ensure that artists
control their "writings" for a limited time. The aim is to secure to
copyright holders a sufficient interest to produce new work. But copyright
laws were crafted in an era long before the Internet. And their effect on
the Internet has been to transfer control over innovation in distribution
from many innovators to a concentrated few.

The clearest example of this effect is online music. Before the Internet,
the production and distribution of music had become extraordinarily
concentrated. In 2000, for example, five companies controlled 84 percent
of music distribution in the world. The reasons for this concentration are
many-including the high costs of promotion-but the effect of concentration
on artist development is profound. Very few artists make any money from
their work, and the few that do are able to do so because of mass
marketing from record labels. The Internet had the potential to change
this reality. Both because the costs of distribution were so low, and
because the network also had the potential to significantly lower the
costs of promotion, the cost of music could fall, and revenues to artists
could rise.

Five years ago, this market took off. A large number of online music
providers began competing for new ways to distribute music. Some
distributed MP3s for money (eMusic.com). Some built technology for giving
owners of music easier access to their music (mp3.com). And some made it
much easier for ordinary users to "share" their music with other users
(Napster). But as quickly as these companies took off, lawyers
representing old media succeeded in shutting them down. These lawyers
argued that copyright law gave the holders (some say hoarders) of these
copyrights the exclusive right to control how they get used. American
courts agreed.

To keep this dispute in context, we should think about the last example of
a technological change that facilitated a much different model for
distributing content: cable TV, which has been accurately hailed as the
first great Napster. Owners of cable television systems essentially set up
antenna and "stole" over-the-air broadcasts and then sold that "stolen
property" to their customers. But when U.S. courts were asked to stop this
"theft," they refused. Twice the U.S. Supreme Court held that this use of
someone else's copyrighted material was not inconsistent with copyright
law.

When the U.S. Congress finally got around to changing the law, it struck
an importantly illustrative balance. Congress granted copyright owners the
right to compensation from the use of their material on cable broadcasts,
but cable companies were given the right to broadcast the copyrighted
material. The reason for this balance is not hard to see. Copyright owners
certainly are entitled to compensation for their work. But the right to
compensation shouldn't translate into the power to control innovation.
Rather than giving copyright holders the right to veto a particular new
use of their work (in this case, because it would compete with
over-the-air broadcasting), Congress assured copyright owners would get
paid without having the power to control-compensation without control.

The same deal could have been struck by Congress in the context of online
music. But this time, the courts did not hesitate to extend control to the
copyright holders. So the concentrated holders of these copyrights were
able to stop the deployment of competing distributors. And Congress was
not motivated to respond by granting an equivalent compulsory right. The
aim of the recording company's strategy was plain enough: shut down these
new and competing models of distribution and replace them with a model for
distributing music online more consistent with the traditional model.

This trend has been supported by the actions of Congress. In 1998,
Congress passed the Digital Millennium Copyright Act (DMCA), which
(in)famously banned technologies designed to circumvent copyright
protection technologies and also created strong incentives for ISPs to
remove from their sites any material claimed to be a violation of
copyright.

On the surface both changes seem sensible enough. Copyright protection
technologies are analogous to locks. What right does anyone have to pick a
lock? And ISPs are in the best position to assure that copyright
violations don't occur on their Web sites. Why not create incentives for
them to remove infringing copyrighted material?

But intuitions here mislead. A copyright protection technology is just
code that controls access to copyrighted material. But that code can
restrict access more effectively (and certainly less subtly) than
copyright law does. Often the desire to crack protection systems is
nothing more than a desire to exercise what is sometimes called a fair-use
right over the copyrighted material. Yet the DMCA bans that technology,
regardless of its ultimate effect.

More troubling, however, is that the DMCA effectively bans this technology
on a worldwide basis. Russian programmer Dimitry Sklyarov, for example,
wrote code to crack Adobe's eBook technology in order to enable users to
move eBooks from one machine to another and to give blind consumers the
ability to "read" out loud the books they purchased. The code Sklyarov
wrote was legal where it was written, but when it was sold by his company
in the United States, it became illegal. When he came to the United States
in July 2001 to talk about that code, the FBI arrested him. Today Sklyarov
faces a sentence of 25 years for writing code that could be used for
fair-use purposes, as well as to violate copyright laws.

Similar trouble has arisen with the provision that gives ISPs the
incentive to take down infringing copyrighted material. When an ISP is
notified that material on its site violates copyright, it can avoid
liability if it removes the material. As it doesn't have any incentive to
expose itself to liability, the ordinary result of such notification is
for the ISP to remove the material. Increasingly, companies trying to
protect themselves from criticism have used this provision to silence
critics. In August 2001, for example, a British pharmaceutical company
invoked the DMCA in order to force an ISP to shut down an animal rights
site that criticized the British company. Said the ISP, "It's very clear
[the British company] just wants to shut them up," but ISPs have no
incentive to resist the claims.

In all these cases, there is a common pattern. In the push to give
copyright owners control over their content, copyright holders also
receive the ability to protect themselves against innovations that might
threaten existing business models. The law becomes a tool to assure that
new innovations don't displace old ones-when instead, the aim of copyright
and patent law should be, as the U.S. Constitution requires, to "promote
the progress of science and useful arts."

These regulations will not only affect Americans. The expanding
jurisdiction that American courts claim, combined with the push by the
World Intellectual Property Organization to enact similar legislation
elsewhere, means that the impact of this sort of control will be felt
worldwide. There is no "local" when it comes to corruption of the
Internet's basic principles. As these changes weaken the open source and
free software movements, countries with the most to gain from a free and
open platform lose. Those affected will include nations in the developing
world and nations that do not want to cede control to a single private
corporation. And as content becomes more controlled, nations that could
otherwise benefit from vigorous competition in the delivery and production
of content will also lose. An explosion of innovation to deliver MP3s
would directly translate into innovation to deliver telephone calls and
video content. Lowering the cost of this medium would dramatically benefit
nations that still suffer from weak technical infrastructures.

Policymakers around the world must recognize that the interests most
strongly protected by the Internet counterrevolution are not their own.
They should be skeptical of legal mechanisms that enable those most
threatened by the innovation commons to resist it. The Internet promised
the world-particularly the weakest in the world-the fastest and most
dramatic change to existing barriers to growth. That promise depends on
the network remaining open to innovation. That openness depends upon
policy that better understands the Internet's past.


Lawrence Lessig is professor of law at Stanford University. He is author
of The Future of Ideas: The Fate of the Commons in the Connected World
(New York: Random House, 2001) and Code and Other Laws of Cyberspace (New
York: Basic Books, 1999).





#  distributed via <nettime>: no commercial use without permission
#  <nettime> is a moderated mailing list for net criticism,
#  collaborative text filtering and cultural politics of the nets
#  more info: majordomo@bbs.thing.net and "info nettime-l" in the msg body
#  archive: http://www.nettime.org contact: nettime@bbs.thing.net