September 15, 2008
Copyright's Default Rules and the Internet
This article studies the rise of copy-reliant technologies - technologies such as Internet search engines and plagiarism detection software that, although they do not read, understand or enjoy copyrighted works, necessarily copy them in large quantities. This article provides a unifying theoretical framework for the legal analysis of topics that tend to be viewed discretely. Search engines, plagiarism detection software, reverse engineering and Google's nascent library cataloging effort, are each part of a broader phenomenon brought about by digitization, that of copy-reliant technologies. These technologies raise two novel, yet central, questions of copyright law. First, whether a non-expressive use that nonetheless requires copying the entirety of a copyright work should be found to infringe the exclusive rights of the copyright owner. Second, whether the transaction costs associated with copy-reliant technologies justify switching copyright's default rule that no copying may take place without permission to one in which copyright owners must affirmatively opt-out of specific uses of their works.
September 15, 2008 | Permalink
TrackBack URL for this entry:
Listed below are links to weblogs that reference Copyright's Default Rules and the Internet: