07.10.2013 Views

A companion to digital humanities (2001)(649s).pdf - FTP Directory ...

A companion to digital humanities (2001)(649s).pdf - FTP Directory ...

A companion to digital humanities (2001)(649s).pdf - FTP Directory ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Notes on Contribu<strong>to</strong>rs<br />

Howard Besser is Direc<strong>to</strong>r of the Moving Image Archive and Preservation Program at<br />

New York University's Tisch School of the Arts. He has been extensively involved in the<br />

movement <strong>to</strong> construct <strong>digital</strong> libraries and museums, and has taught, researched, and<br />

published extensively in the areas of technological change, and the social and cultural<br />

impact of new information environments.<br />

John Bradley was the original author and designer of TACT from 1985 until 1992. He<br />

now works within the Centre for Computing in the Humanities at King's College London,<br />

where he is involved in both teaching and research. His research interests focus on<br />

alternative approaches <strong>to</strong> computer-assisted textual analysis, and on issues that arise from<br />

the modeling of his<strong>to</strong>rical data in computing systems. He has played an important role in<br />

the design and implementation of systems behind the Prosopographies of the Byzantine<br />

Empire/World and of Anglo-Saxon England, of the Clergy of the Church of England, of<br />

the Corpus Vitrearum Medii Aevi and the Corpus of Romanesque Sculpture in Britain<br />

and Ireland, of the Stellenbibliographie zum "Parzival" Wolframs von Eschen-bach, of<br />

the Modern Poetry in Translation website, and a number of other research projects.<br />

John Burrows is Emeritus Professor of English at the University of Newcastle, NSW,<br />

Australia, and was the Foundation Direc<strong>to</strong>r of the Centre for Literary and Linguistic<br />

Computing in that university. His many publications in the field of computational<br />

stylistics include Computation in<strong>to</strong> Criticism (1987). In <strong>2001</strong> he became the second<br />

recipient of the Rober<strong>to</strong> Busa Award for Humanities Computing. The lecture he gave on<br />

that occasion appeared in Computers and the Humanities (February 2003).<br />

Rober<strong>to</strong> A. Busa entered the Jesuit order in 1933, and was ordained priest on May 20,<br />

1940. He is Professor of Philosophy at Aloisianum's Department of Philosophy in<br />

Gallarate, at Gregorian Pontifical University in Rome, and at the Catholic University in<br />

Milan. He is internationally recognized as the pioneer of computational linguistics.<br />

Hugh Craig is Direc<strong>to</strong>r of the Centre for Literary and Linguistic Computing at the<br />

University of Newcastle, NSW, Australia. He also teaches English in the School of<br />

Language and Media where he is currently Head of School. His research in recent years<br />

has been in computational stylistics, applying evidence from the frequencies of very<br />

common words <strong>to</strong> authorial and stylistic problems in Early Modern English literature.<br />

Greg Crane is Winnick Family Chair of Technology and Entrepreneurship, Professor of<br />

Classics and Direc<strong>to</strong>r of the Perseus Project at Tufts University. Originally trained as a<br />

classicist, his current interests focus more generally on the application of information<br />

technology <strong>to</strong> the <strong>humanities</strong>.<br />

Marilyn Deegan has a PhD in medieval studies: her specialism is Anglo-Saxon medical<br />

texts and herbals and she has published and lectured widely in medieval studies, <strong>digital</strong><br />

library research, and <strong>humanities</strong> computing. She is Direc<strong>to</strong>r of Research Development,<br />

Centre for Computing in the Humanities, King's College London, and was formerly


Direc<strong>to</strong>r of Forced Migration Online at the Refugee Studies Centre at Oxford University,<br />

a major <strong>digital</strong> library and portal for materials concerned with all aspects of refugee<br />

studies. She is Edi<strong>to</strong>r-in-Chief of Literary and Linguistic Computing, the Journal of the<br />

Association for Literary and Linguistic Computing, and Direc<strong>to</strong>r of Publications for the<br />

Office for Humanities Communication based at King's College London. Dr. Deegan has<br />

recently published a book, Digital Futures: Strategies for the Information Age, with<br />

Simon Tanner.<br />

Johanna Drucker is currently the Robertson Professor of Media Studies at the<br />

University of Virginia, where she is Professor in the Department of English and Direc<strong>to</strong>r<br />

of Media Studies. She helped establish the Speculative Computing Labora<strong>to</strong>ry in 2000 <strong>to</strong><br />

explore experimental projects in <strong>humanities</strong> computing. She is well known for her work<br />

in the his<strong>to</strong>ry of written forms, typography, design, and visual poetics. Her scholarly<br />

books include: Theorizing Modernism (1994), The Visible Word: Experimental<br />

Typography and Modern Art (1994); The Alphabetic Labyrinth (1995), and The Century<br />

of Artists' Books (1995). Her most recent collection, Figuring the Word, was published in<br />

November 1998.<br />

Harrison Eiteljorg, II, is a classical archaeologist who specializes in the architecture of<br />

classical Greece. He has worked <strong>to</strong> use computer technology in his own work and <strong>to</strong><br />

explore how that technology can best be applied <strong>to</strong> the work of archaeologists generally.<br />

Charles Ess is Distinguished Research Professor, Interdisciplinary Studies, Drury<br />

University (Springfield, Missouri) and Professor II in the Applied Ethics Programme,<br />

Norwegian University of Science and Technology (Trondheim). He researches, lectures,<br />

and publishes on computer-mediated communication (CMC) and Information Ethics,<br />

especially from cross-cultural perspectives; with Fay Sudweeks, organizes conferences<br />

and edits publications on cultural attitudes <strong>to</strong>ward technology and CMC.<br />

Ichiro Fujinaga is an Assistant Professor at the Faculty of Music at McGill University<br />

and the Chair of the Music Technology Area. He has degrees in music/percussion and<br />

mathematics from the University of Alberta, and a Master's degree in music theory, and a<br />

PhD in music technology from McGill University.<br />

Michael Greenhalgh has been interested in computing applications in the <strong>humanities</strong><br />

since playing with a Commodore PET-<strong>2001</strong> in 1977 at Leicester University, England.<br />

Graduating <strong>to</strong> large Cyber machines, PDF 11s and then VAXes, and writing books on<br />

them, he started teaching <strong>humanities</strong> computing courses in 1980. Elected <strong>to</strong> the Sir<br />

William Dobell Foundation Chair in Art His<strong>to</strong>ry at the Australian National University in<br />

1987, he introduced similar courses there and, using powerful graphics machines, was<br />

ready with over 300 digitized images <strong>to</strong> start his web-server "ArtServe"<br />

(http://rubens.anu.edu.au) in January 1994. This server now offers over 190,000 images,<br />

and receives about 1.3 million hits per week. Since 1997, his lecturing and seminar work<br />

has been done exclusively over the network from servers, via a video projec<strong>to</strong>r.


Jan Hajič is an Associate Professor of Computational Linguistics at Charles University,<br />

Prague, Czech Republic. His interests range from morphology of inflective languages <strong>to</strong><br />

syntax and treebanking <strong>to</strong> machine translation, using extensively statistical methods in<br />

natural language processing. He has previously worked at Johns Hopkins University<br />

(Maryland), and at IBM Research (New York).<br />

Susan Hockey was Professor of Library and Information Studies and Direc<strong>to</strong>r of the<br />

School of Library, Archive, and Information Studies at University College London until<br />

July 31, 2004. She was Chair of the Association for Literary and Linguistic Computing<br />

(1984–97) and a member, twice Chair, of the Steering Committee for the Text Encoding<br />

Initiative. Her current interests are markup technologies for the <strong>humanities</strong> and the<br />

his<strong>to</strong>ry of <strong>humanities</strong> computing.<br />

Nancy Ide is Professor and Chair of Computer Science at Vassar College in<br />

Poughkeepsie, New York. She has been involved in <strong>humanities</strong> computing and<br />

computational linguistics for 20 years. She was president of the Association for<br />

Computers and the Humanities from 1985 <strong>to</strong> 1995, and is currently co-edi<strong>to</strong>r of the<br />

journal Computers and the Humanities. She is also co-edi<strong>to</strong>r of the Kluwer book series<br />

Text, Speech, and Language Technology, and has co-directed the past four EUROLAN<br />

summer schools on various <strong>to</strong>pics in computational linguistics. In 1987 she spearheaded<br />

the Text Encoding Initiative and served on its steering committee until 1997. She has<br />

published numerous papers on the application of statistical methods <strong>to</strong> language analysis,<br />

including computational lexicography, word sense disambiguation, and discourse<br />

analysis. Most recently she has been involved in developing standards for the<br />

representation of linguistically annotated resources, the creation of the American<br />

National Corpus, the development of an "intelligently searchable" corpus for his<strong>to</strong>rical<br />

research comprising materials from the Franklin D. Roosevelt Library, and adapting<br />

language processing practices <strong>to</strong> the Semantic Web.<br />

Michael Jensen was recently appointed Direc<strong>to</strong>r of Web Communications for the<br />

National Academies: the National Academy of Sciences, the National Research Council,<br />

the Institute of Medicine, and the National Academy of Engineering. He remains Direc<strong>to</strong>r<br />

of Publishing Technologies at the National Academies Press, which makes more than<br />

2,800 books (more than 500,000 pages) fully searchable and browsable online for free. In<br />

the mid-1990s, he helped establish Project Muse, the pioneering online journals project of<br />

Johns Hopkins University Press. For the University of Nebraska Press, he produced the<br />

first searchable publisher's catalogue available on the Internet, via Telnet, in 1990.<br />

Matthew G. Kirschenbaum is Assistant Professor of English and Digital Studies at the<br />

University of Maryland, College Park. He does both theoretical and applied work in the<br />

<strong>digital</strong> <strong>humanities</strong>, participating in projects from the William Blake Archive <strong>to</strong> the<br />

Electronic Literature Organization. He has long-standing interests in images, interface,<br />

and visualization, and is co-developer of a software <strong>to</strong>ol called the Virtual Lightbox. His<br />

book-in-progress, entitled Mechanisms: New Media and the New Textuality, is<br />

forthcoming from the MIT Press.


Robert Kolker is the author of a number of books in cinema studies, including the third<br />

edition of A Cinema of Loneliness: Penn, S<strong>to</strong>ne, Kubrick, Scorsese, and Altman (2000).<br />

His textbook, Film, Form, and Culture (2nd edn., 2002) is an introduction <strong>to</strong> film with an<br />

accompanying interactive CD-ROM, containing moving image clips. Kolker was one of<br />

the first film scholars <strong>to</strong> put moving images on the Web, in a Postmodern Culture essay,<br />

"The Moving Image Reclaimed." He will soon publish a Casebook of essays on<br />

Hitchcock's Psycho and a critical work, Dreadful Landscapes in the Spaces of Modernity;<br />

Welles, Kubrick and the Imagination of the Visible. He has been Professor of English at<br />

the University of Maryland, and Chair of the School of Literature, Communication, and<br />

Culture at Georgia Tech.<br />

Ian Lancashire, Professor of English at the University of Toron<strong>to</strong>, founded the Center<br />

for Computing in the Humanities there in 1986 and co-developed TACT (Text Analysis<br />

Computing Tools), with which he does research on Chaucer and Shakespeare. He now<br />

edits Representative Poetry Online, teaches a fully online course in reading poetry,<br />

develops the Early Modern English Dictionaries (LEME) database, and is a member of<br />

the Inter-PARES2 project. From 1992 <strong>to</strong> 2003 he presided over a Canadian learned<br />

society, the Consortium for Computers in the Humanities/Consortium pour ordinateurs en<br />

sciences humaines (COCH/COSH).<br />

Andrea Laue is a PhD candidate at the University of Virginia. She is a technical edi<strong>to</strong>r<br />

for the William Blake Archive, a researcher in the Speculative Computing Lab, and a<br />

graduate instruc<strong>to</strong>r in the media studies program. Her research interests include narra<strong>to</strong>logy,<br />

cognitive poetics, textual studies, <strong>digital</strong> media, and information visualization. In her<br />

dissertation, she investigates narrative structuring as manifested in the emergent text, the<br />

nexus produced by the interactions of an interpreter and a literary artifact.<br />

Andrew Mactavish is an Assistant Professor of Multimedia in the School of the Arts at<br />

McMaster University. He has published and presented papers in the areas of computer<br />

games, <strong>humanities</strong> computing, and multimedia. He currently holds a research grant from<br />

the Social Sciences and Humanities Research Council of Canada (SSHRC) <strong>to</strong> study the<br />

cultural politics of computer game play. He is also a member of collaborative research<br />

projects, including Globalization and Au<strong>to</strong>nomy (SSHRC-MCRI) and two projects<br />

funded by the Canada Foundation for Innovation: TAPoR (Text Analysis Portal for<br />

Research) and IRIS (Infrastructure for Research on Internet Streaming).<br />

Willard McCarty is Senior Lecturer in Humanities Computing, King's College London,<br />

and edi<strong>to</strong>r of Humanist. Recently his research has centered on modeling and more<br />

broadly explored the intellectual integrity of <strong>humanities</strong> computing by probing and<br />

interrelating its disciplinary kinships. He is currently finishing a book on the field, for<br />

which the chapter included here is a preview. His primary research on modeling has<br />

drawn heavily on his Analytical Onomasticon <strong>to</strong> the Metamorphoses of Ovid,<br />

forthcoming when a sufficiently adventurous publisher can be found. Details at<br />

.


Jerome McGann is the John Stewart Bryan University Professor, University of Virginia<br />

and Adjunct Professor at Royal Holloway, University of London. His recent book<br />

Radiant Textuality: Literature after the World Wide Web, was awarded the Modern<br />

Language Association's James Russell Lowell Award (2002). He is currently developing<br />

<strong>digital</strong> resources for the interpretation of literary works. These include the collaborative<br />

environment IVANHOE and the 'Patacritical Demon.<br />

Bethany Nowviskie is a doc<strong>to</strong>ral candidate at the University of Virginia. She serves as<br />

design edi<strong>to</strong>r for the Rossetti Archive, and is the lead designer and manager of the<br />

Temporal Modeling Project. Her other SpecLab projects include Biblioludica, the<br />

Ivanhoe Game, and the 'Patacritical Demon. Nowviskie's dissertation theorizes and<br />

describes the production of <strong>digital</strong> environments that both promote humanistic<br />

interpretation and emerge from the interpretative acts of their users.<br />

Carole L. Palmer is an Associate Professor at the Graduate School of Library and<br />

Information Science at the University of Illinois at Urbana-Champaign. Her research<br />

explores how information systems and services can best support the work of researchers.<br />

She is engaged in projects <strong>to</strong> develop information technologies that support<br />

interdisciplinary inquiry, discovery, and collaboration in the <strong>humanities</strong> and the sciences.<br />

Her recent publications include Scholarly Work in the Humanities and the Evolving<br />

Information Environment and Work at the Boundaries of Science: Information and the<br />

Interdisciplinary Research Process.<br />

Daniel V. Pitti is project direc<strong>to</strong>r at the Institute for Advanced Technology in the<br />

Humanities (IATH) at the University of Virginia. As project direc<strong>to</strong>r, he is responsible<br />

for project design in general, and Extensible Markup Language (XML) and objectrelational<br />

databases design and development in particular. Before coming <strong>to</strong> IATH in<br />

1997, he was Librarian for Advanced Technologies Projects at the University of<br />

California at Berkeley Library.<br />

Stephen Ramsay worked as a programmer and software engineer for the Institute for<br />

Advanced Technology in the Humanities at the University of Virginia before becoming<br />

an Assistant Professor of English at the University of Georgia. He edits the online version<br />

of the journal TEXT Technology, and has lectured widely on subjects related <strong>to</strong><br />

<strong>humanities</strong> computing, public policy, and software design.<br />

Allen H. Renear is an Associate Professor in the Graduate School of Library and<br />

Information Science (GSLIS) at the University of Illinois, Urbana-Champaign, where he<br />

teaches courses in knowledge representation and document modeling, and, as head of the<br />

Electronic Publishing Research Group, leads research on XML semantics and document<br />

on<strong>to</strong>logy. He has been involved in <strong>humanities</strong>-oriented electronic publishing, standards<br />

development, and research for over twenty years. Currently Chair of the Open eBook<br />

Publication Structure Working Group, he has served as President of the Association for<br />

Computers and the Humanities, on the Advisory Board of the Text Encoding Initiative,<br />

and prior <strong>to</strong> joining GSLIS, was Direc<strong>to</strong>r of the Scholarly Technology Group at Brown<br />

University.


Geoffrey Rockwell is an Associate Professor of Humanities Computing and Multimedia<br />

in the School of the Arts at McMaster University. He received a BA in philosophy from<br />

Haverford College, an MA and PhD in philosophy from the University of Toron<strong>to</strong> and<br />

worked at the University of Toron<strong>to</strong> as a Senior Instructional Technology Specialist. He<br />

has published and presented papers in the area of textual visualization and analysis,<br />

<strong>humanities</strong> computing, instructional technology, computer games, and multimedia. With<br />

colleagues at McMaster University he set up an honors Multimedia program. He is<br />

currently the project leader for the CFI (Canada Foundation for Innovation) funded<br />

project TAPoR, a Text Analysis Portal for Research, which is developing a text <strong>to</strong>ol<br />

portal for researchers who work with electronic texts. He recently published a book,<br />

Defining Dialogue: From Socrates <strong>to</strong> the Internet (2003).<br />

Thomas Rommel is Professor of English at International University Bremen (IUB). He<br />

is a member of the executive committee of the Association for Literary and Linguistic<br />

Computing (ALLC). His publications include a book on Byron's poetry (1995), Anglistik<br />

im Internet (1996), a study of Adam Smith (1999), and an edition of essays on literary<br />

hypertexts (forthcoming). He is co-edi<strong>to</strong>r of Prolepsis, The Heidelberg Review of English<br />

Studies. His research interests include theories of electronic text and the methodological<br />

implications of computer-assisted studies of literature.<br />

Marie-Laure Ryan is a native of Geneva, Switzerland, and is currently an independent<br />

scholar based in Colorado. She is the author of Possible Worlds, Artificial Intelligence<br />

and Narrative Theory (1991), and of Narrative as Virtual Reality: Immersion and<br />

Interactivity in Literature and Electronic Media (<strong>2001</strong>), which received the Jeanne and<br />

Aldo Scaglione Prize for Comparative Literature from the Modern Language<br />

Association. She is also the edi<strong>to</strong>r of two collections of essays, Cyberspace Textuality<br />

(1999) and Narrative Across Media (2004).<br />

David Z. Saltz is Associate Professor of Drama at the University of Georgia. He is<br />

Principal Investiga<strong>to</strong>r of Virtual Vaudeville: A Live Performance Simulation System,<br />

funded by the NSF, and has published essays about performance theory and interactive<br />

media in scholarly books and journals including Theatre Research International,<br />

Performance Research, and the Journal of Aesthetics and Art Criticism. He is also a<br />

practicing direc<strong>to</strong>r and installation artist whose work focuses on the interaction between<br />

<strong>digital</strong> media and live performance.<br />

Susan Schreibman is General Edi<strong>to</strong>r and Project Manager of the MacGreevy Archive,<br />

and edi<strong>to</strong>r of Collected Poems of Thomas MacGreevy: An Annotated Edition (1991). She<br />

is the founder and edi<strong>to</strong>r of the web-based Irish Resources in the Humanities. She is<br />

currently serving a two-year term on the TEI Council. Dr. Schreibman is Assistant<br />

Direc<strong>to</strong>r of the Maryland Institute for Technology in the Humanities (MITH), and an<br />

Affiliate Faculty member in the Department of English. Previously she was an Assistant<br />

Professor of Professional and Technical Communication at New Jersey Institute of<br />

Technology (2000–1), and the Semester in Irish Studies Newman Fellow at University<br />

College Dublin (1997–2000).


Ray Siemens is Canada Research Chair in Humanities Computing and Associate<br />

Professor of English at the University of Vic<strong>to</strong>ria; formerly, he was Professor of English<br />

at Malaspina University-College (1999–2004). He is President (English) of the<br />

Consortium for Computers in the Humanities/Consortium pour ordinateurs en sciences<br />

humaines and, in 2003, was Visiting Senior Research Fellow at the Centre for Computing<br />

in the Humanities at King's College London. Direc<strong>to</strong>r of the Digital<br />

Humanities/Humanities Computing Summer Institute, founder of Malaspina University-<br />

College's Center for Digital Humanities Innovation, and founding edi<strong>to</strong>r of the electronic<br />

scholarly journal Early Modern Literary Studies, he is also author of works chiefly<br />

focusing on areas where literary studies and computational methods intersect, is edi<strong>to</strong>r of<br />

several Renaissance texts, and is co-edi<strong>to</strong>r of several book collections on <strong>humanities</strong><br />

computing <strong>to</strong>pics.<br />

Abby Smith is Direc<strong>to</strong>r of Programs at the Council on Library and Information<br />

Resources, where her work focuses on the development and preservation of research<br />

collections in all formats and genres. She worked as program specialist at the Library of<br />

Congress, and taught intellectual and Russian his<strong>to</strong>ry at Harvard and Johns Hopkins. Her<br />

recent publications include: New-Model Scholarship: How Will It Survive? (2003); The<br />

Evidence in Hand: Report of the Task Force on the Artifact in Library Collections<br />

(<strong>2001</strong>); Strategies for Building Digitized Collections (<strong>2001</strong>); Building and Sustaining<br />

Digital Collections: Models for Libraries and Museums (<strong>2001</strong>); and Collections, Content,<br />

and the Web (2000).<br />

Martha Nell Smith is Professor of English and Direc<strong>to</strong>r of the Maryland Institute for<br />

Technology in the Humanities (MITH) at the University of Maryland. Her numerous<br />

publications include three award-winning books – Open Me Carefully: Emily Dickinson's<br />

Intimate Letters <strong>to</strong> Susan Dickinson, co-authored with Ellen Louise Hart (1998), Comic<br />

Power in Emily Dickinson, co-authored with Cristanne Miller and Suzanne Juhasz<br />

(1993), Rowing in Eden: Rereading Emily Dickinson (1992) – and more than thirty<br />

articles in such journals as American Literature, Studies in the Literary Imagination,<br />

South Atlantic Quarterly, Women's Studies Quarterly, Profils Americains, San Jose<br />

Studies, and The Emily Dickinson Journal. With Mary Loeffelholz, she is editing the<br />

Blackwell Companion <strong>to</strong> Emily Dickinson (forthcoming in 2005). The recipient of<br />

numerous awards from the National Endowment for the Humanities (NEH), the<br />

American Council of Learned Societies (ACLS), and the Fund for the Improvement of<br />

Postsecondary Education (FIPSE) for her work on Dickinson and in new media, Smith is<br />

also Coordina<strong>to</strong>r and General Edi<strong>to</strong>r of the Dickinson Electronic Archives projects at the<br />

Institute for Advanced Technology in the Humanities (IATH) at the University of<br />

Virginia. With Lara Vetter, Smith is a general edi<strong>to</strong>r of Emily Dickinson's<br />

Correspondences, forthcoming from the Mellon-sponsored University of Virginia Press<br />

Electronic Imprint.<br />

C. M. Sperberg-McQueen is a member of the technical staff of the World Wide Web<br />

Consortium, an international membership organization responsible for developing Web<br />

standards. He co-edited the XML 1.0 specification and the Guidelines of the Text<br />

Encoding Initiative.


Simon Tanner is the Direc<strong>to</strong>r of Digital Consultancy Services at King's College London<br />

(KDCS). He has an international reputation as a consultant and has consulted for<br />

prestigious digitization projects in Europe and America. Tanner has a library and<br />

information science background and prior <strong>to</strong> KDCS was Senior Consultant with HEDS<br />

Digitization Services and held librarian and systems posts at Loughborough University,<br />

Rolls Royce, and IBM. He recently authored a book, Digital Futures: Strategies for the<br />

Information Age, with Marilyn Deegan; he co-edits the Digital Futures series of books<br />

from Facet Publishing and is a guest edi<strong>to</strong>r for the Journal of Digital Information.<br />

William G. Thomas, III, is an Assistant Professor of His<strong>to</strong>ry and Direc<strong>to</strong>r of the<br />

Virginia Center for Digital His<strong>to</strong>ry at the University of Virginia. He is the co-author with<br />

Edward L. Ayers of "The Difference Slavery Made: A Close Analysis of Two American<br />

Communities", a fully electronic journal article for the American His<strong>to</strong>rical Review. He is<br />

the author of Lawyeringfor the Railroad: Business Law and Power in the New South<br />

(1999) and co-author of the Emmy nominee documentary film Massive Resistance. He is<br />

currently researching and producing a <strong>digital</strong> project on the environmental and social<br />

his<strong>to</strong>ry of the Chesapeake Bay.<br />

John Unsworth served from 1993 <strong>to</strong> 2003 as the first Direc<strong>to</strong>r of the Institute for<br />

Advanced Technology in the Humanities and as faculty in the Department of English at<br />

the University of Virginia. As the Institute's Direc<strong>to</strong>r, he oversaw research projects across<br />

the disciplines in the <strong>humanities</strong> and published widely on electronic scholarship,<br />

<strong>humanities</strong> computing, and other <strong>to</strong>pics. In 2003, he was appointed Dean of the Graduate<br />

School of Library and Information Science (GSLIS) at the University of Illinois, Urbana-<br />

Champaign, with appointments as Professor in GSLIS, in the Department of English, and<br />

on the Library faculty.<br />

Claire Warwick is a lecturer at the School of Library Archive and Information Studies,<br />

University College London, where she is programme direc<strong>to</strong>r of the MA in Electronic<br />

Communication and Publishing. She is also a research supervisor for Cambridge<br />

University's MSt in International Relations. She has previously worked at Sheffield<br />

University, Department of Information Studies, and Oxford University English Faculty<br />

and Humanities Computing Unit, and at Chadwyck-Healey Ltd. Her research interests<br />

center on the study of the use and impact of computing on <strong>humanities</strong> scholarship, and on<br />

the societal effects of electronic publishing. She is a member of the advisory panel for the<br />

Portsmouth His<strong>to</strong>rical Records Series, e-press and the Digital Egypt for Universities<br />

project and is a visiting lecturer at City College Thessaloniki, Greece.<br />

Susan Forscher Weiss holds a joint appointment in the Departments of Musicology and<br />

Romance Languages and Literature at the Johns Hopkins University. Her numerous<br />

publications include Bologna Q 18: l-Bologna, Civico Museo Bibliografico Musicale,<br />

Ms.BolC Q 18 (olim 143), Introduction and Facsimile Edition (1998), chapters, articles,<br />

reviews, and entries in numerous scholarly publications. She has collaborated in webbased<br />

programming, exhibits, CD-ROMs, and audio <strong>to</strong>urs. She has also been the recipient<br />

of numerous awards and fellowships for teaching and research including grants from the


National Endowment for the Humanities, the American Council of Learned Societies,<br />

Harvard University, the Johns Hopkins University, and the Folger Shakespeare Library.<br />

Perry Willett was appointed the Head of the Digital Library Production Service at the<br />

University of Michigan in 2004. He served as the Assistant Direc<strong>to</strong>r of the Digital<br />

Library Program at Indiana University from <strong>2001</strong>, and was a bibliographer at the Main<br />

Library of Indiana University from 1992 <strong>to</strong> <strong>2001</strong>. He is the general edi<strong>to</strong>r of the Wright<br />

American Fiction project and the Vic<strong>to</strong>rian Women Writers Project, and is a member of<br />

the Text Encoding Initiative Consortium Council. He has written on electronic text and<br />

<strong>digital</strong> libraries.<br />

William Winder is Assistant Professor of French at the University of British Columbia's<br />

French, Hispanic, and Italian Studies Department. He is on the board of direc<strong>to</strong>rs of the<br />

Consortium for Computers in the Humanities and the edi<strong>to</strong>rial board of TEXT<br />

Technology, and he co-edits Computing in the Humanities Working Papers. His interests<br />

lie in computational and formalist approaches <strong>to</strong> the semantics of language and literature.<br />

See his website (http://www.fhis.ubc.ca/winder) for recent publications and research.<br />

Russon Wooldridge is Professor in the Department of French, University of Toron<strong>to</strong>. He<br />

is a teacher/researcher in French language, French lexicography, translation, and corpus<br />

linguistics of the Web. Research details and content at:<br />

http://www.chass.u<strong>to</strong>ron<strong>to</strong>.ca/~wulfric.<br />

Foreword: Perspectives on the Digital Humanities<br />

Rober<strong>to</strong> A. Busa<br />

During World War II, between 1941 and 1946, I began <strong>to</strong> look for machines for the<br />

au<strong>to</strong>mation of the linguistic analysis of written texts. I found them, in 1949, at IBM in<br />

New York City. Today, as an aged patriarch (born in 1913) I am full of amazement at the<br />

developments since then; they are enormously greater and better than I could then<br />

imagine. Digitus Dei est hic! The finger of God is here!<br />

I consider it a great honor <strong>to</strong> be asked <strong>to</strong> write the Foreword <strong>to</strong> this fine book. It<br />

continues, underlines, completes, and recalls the previous Survey of An<strong>to</strong>nio Zampolli<br />

(produced in Pisa, 1997), who died before his time on August 22, 2003. In fact, the book<br />

gives a panoramic vision of the status artis. It is just like a satellite map of the points <strong>to</strong><br />

which the wind of the ingenuity of the sons of God moves and develops the contents of<br />

computational linguistics, i.e., the computer in the <strong>humanities</strong>.<br />

Humanities computing is precisely the au<strong>to</strong>mation of every possible analysis of human<br />

expression (therefore, it is exquisitely a "humanistic" activity), in the widest sense of the<br />

word, from music <strong>to</strong> the theater, from design and painting <strong>to</strong> phonetics, but whose<br />

nucleus remains the discourse of written texts.


In the course of the past sixty years I have added <strong>to</strong> the teaching of scholastic philosophy,<br />

the processing of more than 22 million words in 23 languages and 9 alphabets, registering<br />

and classifying them with my teams of assistants. Half of those words, the main work, are<br />

in Latin. I will summarize the three different perspectives that I have seen and<br />

experienced in these sixty years.<br />

Technological "Miniaturization"<br />

According <strong>to</strong> the perspective of technological miniaturization, the first perspective I will<br />

treat, the Index Thomisticus went through three phases. The first one lasted less than 10<br />

years. I began, in 1949, with only electro-countable machines with punched cards. My<br />

goal was <strong>to</strong> have a file of 13 million of these cards, one for each word, with a context of<br />

12 lines stamped on the back. The file would have been 90 meters long, 1.20 m in height,<br />

1 m in depth, and would have weighed 500 <strong>to</strong>nnes.<br />

In His mercy, around 1955, God led men <strong>to</strong> invent magnetic tapes. The first were the<br />

steel ones by Reming<strong>to</strong>n, closely followed by the plastic ones of IBM. Until 1980, I was<br />

working on 1,800 tapes, each one 2,400 feet long, and their combined length was 1,500<br />

km, the distance from Paris <strong>to</strong> Lisbon, or from Milan <strong>to</strong> Palermo. I used all the<br />

generations of the dinosaur computers of IBM at that time. I finished in 1980 (before<br />

personal computers came in) with 20 final and conclusive tapes, and with these and the<br />

au<strong>to</strong>matic pho<strong>to</strong>composi<strong>to</strong>r of IBM, I prepared for offset the 20 million lines which filled<br />

the 65,000 pages of the 56 volumes in encyclopedia format which make up the Index<br />

Thomisticus on paper.<br />

The third phase began in 1987 with the preparations <strong>to</strong> transfer the data on<strong>to</strong> CD-ROM.<br />

The first edition came out in 1992, and now we are on the threshold of the third. The<br />

work now consists of 1.36 GB of data, compressed with the Huffman method, on one<br />

single disk.<br />

Textual Informatics<br />

The second perspective is textual informatics, and it has branched in<strong>to</strong> three different<br />

currents. Today the two greater and richer ones must be clearly distinguished from the<br />

third, the smallest and poorest. I must say that many people still do not realize this.<br />

I call the first current "documentaristic" or "documentary", in memory of the American<br />

Documentation Society, and of the Deutsche Gesellschaft für Dokumentation in the<br />

1950s. It includes databanks, the Internet, and the World Wide Web, which <strong>to</strong>day are the<br />

infrastructures of telecommunications and are in continuous ferment. The second current<br />

I call "edi<strong>to</strong>rial." This is represented by CDs and their successors, including the<br />

multimedia ones, a new form of reproduction of a book, with audio-visual additions. Both<br />

these, albeit in different ways, provide for the multiplication, distribution, and swift<br />

traceability of both information and of a text. Both are recognizable by the fact that they<br />

substantially transfer and present, on an electronic support system, words and punctuation


plus some operative commands. Because they provide a service and so have a quick<br />

return on investment, both have grown abundantly – the first, despite significant<br />

obstacles, more solidly and with fewer disappointments than the second.<br />

I call the third current "hermeneutic" or interpretative, that informatics most associated<br />

with linguistic analysis and which I would describe as follows. In the electronic Index<br />

Thomisticus each of the 11 million words is encapsulated in a record of 152 bytes. Some<br />

22 are reserved for the word, and 130 contain 300 alternating "internal hypertexts", which<br />

specify the values within the levels of the morphology.<br />

At the moment, I am trying <strong>to</strong> get another project under way, which will obviously be<br />

posthumous, the first steps of which will consist in adding <strong>to</strong> the morphological encoding<br />

of each single separate word of the Thomistic lexicon (in all there are 150,000, including<br />

all the particles, such as et, non, etc.), the codes that express its syntax (i.e., its direct<br />

elementary syntactic correlations) within each single phrase in which it occurs. This<br />

project is called Lessico Tomistico Biculturale (LTB). Only a computer census of the<br />

syntactic correlations can document what concepts the author wanted <strong>to</strong> express with that<br />

word. Of a list of syntactic correlations, the "conceptual" translation can thus be given in<br />

modern languages. I have already published, mainly in the series of the Lessico<br />

Intellectuale Europeo (directed by T. Gregory of the University of Rome), the results of<br />

such syntactical analysis of a dozen words in their more than 500,000 context lines. To<br />

give one example, in the mind of St Thomas ratio seminalis meant then what <strong>to</strong>day we<br />

call genetic programme. Obviously, St Thomas did not know of either DNA or genes,<br />

because at the time microscopes did not exist, but he had well unders<strong>to</strong>od that something<br />

had <strong>to</strong> perform their functions.<br />

Hermeneutic Informatics<br />

This third sort of informatics was the first <strong>to</strong> come in<strong>to</strong> being, with the Index Thomisticus<br />

project, in 1949. It brought the following facts <strong>to</strong> my attention. First, everyone knows<br />

how <strong>to</strong> use his own mother <strong>to</strong>ngue, but no one can know "how", i.e., no one can explain<br />

the rules and no one can list all the words of the lexicon that he uses (the active lexicon)<br />

nor of that which he understands but never uses (the passive lexicon).<br />

What scholar could answer the following questions? How many verbs does he know at<br />

least passively? Which and how many of them are always and only transitive? Which and<br />

how many of them are always and only intransitive? Which and how many of them are<br />

sometimes the one, and sometimes the other, and what is the percentage of each? Lastly,<br />

which contextual situations characteristically mark the transitive or intransitive use of the<br />

latter?<br />

Second, there is still no scientific grammar of any language that gives, in a systematized<br />

form, all the information necessary <strong>to</strong> program a computer for operations of artificial<br />

intelligence that may be currently used on vast quantities of natural texts, at least, e.g., for<br />

indexing the key words under which <strong>to</strong> archive or summarize these texts <strong>to</strong> achieve<br />

"au<strong>to</strong>matic indexing – au<strong>to</strong>matic abstracting."


Third, it is thus necessary for the use of informatics <strong>to</strong> reformulate the traditional<br />

morphology, syntax, and lexicon of every language. In fact all grammars have been<br />

formed over the centuries by nothing more than sampling. They are not <strong>to</strong> be<br />

revolutionized, abandoned, or destroyed, but subjected <strong>to</strong> a re-elaboration that is<br />

progressive in extent and depth.<br />

Schematically, this implies that, with integral censuses of a great mass of natural texts in<br />

every language, in synchrony with the discovered data, methods of observation used in<br />

the natural sciences should be applied <strong>to</strong>gether with the apparatus of the exact and<br />

statistical sciences, so as <strong>to</strong> extract categories and types and, thus, <strong>to</strong> organize texts in a<br />

general lexicological system, each and all with their probability index, whether great or<br />

small.<br />

Hermeneutic informatics hinges on the Alpac Report (Washing<strong>to</strong>n, DC, 1966) and, now,<br />

this perspective is perhaps awaiting its own globalization. I have already said that<br />

hermeneutic informatics was the first <strong>to</strong> come in<strong>to</strong> existence. Shortly afterwards, in the<br />

early 1950s, if I am correct, the move <strong>to</strong>ward au<strong>to</strong>matic translation started. The magazine<br />

MT – Mechanical Translation was started at MIT, launched, I think, by Professor Billy<br />

Locke and others. The Pentagon financed various centers. I was involved in this. I<br />

connected the Anglo-Russian project of Leon Dostert and Peter Toma (of George<strong>to</strong>wn<br />

University, Washing<strong>to</strong>n, DC) with the Computing Center of the Eura<strong>to</strong>m of Ispra, which<br />

is on Lake Maggiore in Lombardy. My contributions were on an exchange basis. I<br />

supplied them, from my labora<strong>to</strong>ry at Gallarate, with Russian abstracts of biochemistry<br />

and biophysics in Cyrillic script on punched cards, a million words. These were<br />

translated with the George<strong>to</strong>wn programs. The translation was sufficient for an expert on<br />

the subject <strong>to</strong> be able <strong>to</strong> evaluate the merits of a more accurate translation done by hand,<br />

i.e., by a person's brain.<br />

Unfortunately, in 1966, as a result of the Alpac Report, the Pentagon cut off all funding.<br />

This was not because computers at that time did not have sufficient memory capability or<br />

speed of access, but precisely because the information on the categories and their<br />

linguistic correspondences furnished by the various branches of philology were not<br />

sufficient for the purpose. The "machine" required greater depth and more complex<br />

information about our ways of thinking and modes of expression!<br />

Future Perspectives<br />

In certain respects this was a boon. In fact, as this volume, <strong>to</strong>o, documents, the number<br />

and devotion of those few volunteers, who in almost every part of the world have a<br />

passion for computational informatics, increased. They are not an organized army, but<br />

hunters who range freely, and this produces some obvious side effects.<br />

It also provokes a valuable consideration. Namely, it makes us realize that no single<br />

research center ever seems <strong>to</strong> have been able <strong>to</strong> answer, alone and completely, the<br />

linguistic challenge of globalized telematics. It seems that the answer <strong>to</strong> globalization, at<br />

least in principle, should be global as well, i.e., collective, or rather undertaken by one or


more supranational organizations, this for the obvious reason of the commitments<br />

required. Speaking is an interface between infinity and the cosmos, between time and<br />

eternity, and is evidence of the thirst for knowledge, understanding, possession, and<br />

manipulation of everything, according <strong>to</strong> one's own personal freedom, but on track with a<br />

common ballast of logic and beauty. Speaking must thus be taken seriously; it is sacred,<br />

as is every human person. We are far from having exhausted the precept inscribed on<br />

Apollo's temple at Delphi, "Know thyself." It seems, therefore, that the problem must be<br />

attacked: in its <strong>to</strong>tality – with comprehensive, i.e., global, research; collectively – by<br />

exploiting informatics with its enormous intrinsic possibilities, and not by rushing, just <strong>to</strong><br />

save a few hours, in<strong>to</strong> doing the same things which had been done before, more or less in<br />

the same way as they were done before.<br />

It seems that the attack on the Twin Towers of New York City on September 11, <strong>2001</strong>,<br />

has brought in an unforeseen season of lean kine. In Italy, as everywhere else, this has<br />

meant reductions in public funds for research. This period will pass, as surely as all the<br />

others we have experienced. Such reductions were also in evidence at the Explora<strong>to</strong>ry<br />

Workshop on Computer Texts, which the European Science Foundation of the European<br />

Union held at Strasbourg on June 14 and 15, 2002. On the one hand, these cutbacks in<br />

finance are certainly worrying for the many opera<strong>to</strong>rs of computational linguistics, which<br />

<strong>to</strong>day is fragmented, but on the other hand, it could also lead <strong>to</strong> the according of priority<br />

<strong>to</strong> a definitive solution of the linguistic problem, like that which could facilitate the<br />

fulfillment of the globalization of economic exchange.<br />

A Proposal<br />

I should like <strong>to</strong> summarize the formula of a global solution <strong>to</strong> the linguistic challenge that<br />

I presented at the above-mentioned conference at Strasburg, much as if it were my<br />

spiritual testament, although I am uncertain whether <strong>to</strong> call it prophecy or U<strong>to</strong>pia.<br />

I suggest that – care of, for example, the European Union – for every principal language<br />

A, B, C, D, etc., from each of the principal university textbooks in present use in the<br />

various disciplines (for these represent the present state of what is knowable) there should<br />

be extracted its integral "lexicological system" (with the help of the instruments tested in<br />

the Index Thomisticus). In it, two "hemispheres" of each lexicon should be distinguished,<br />

the few words of very high frequency present in every argument which express the logic,<br />

and which are sometimes called "grammatical", and the very many words of relatively<br />

minor frequency that specify messages and arguments. The systems of each text of each<br />

discipline in each language should be integrally compared with each other by isolating<br />

and specifying both the coincidences and divergences with their quantities and<br />

percentages, including those of hapax, extracting from them, i.e., mixing them, in one<br />

single system of the same language, a system which <strong>to</strong>tals statistically and with<br />

percentages both how much they have in common and their co-respective divergences.<br />

Then these statistical summaries of the various languages A, B, C, D, etc., should be<br />

compared with each other, with the same method of reporting their convergences and<br />

divergences, with quantity and percentage in a single system. (One has only <strong>to</strong> think of


how much data could be published a latere as valid and useful documents, although <strong>to</strong> be<br />

constantly updated, for example for contrastive grammars, etc.)<br />

Thus there would be on the computer a common interlingual system consisting solely of<br />

strings of bits and bytes with correspondence links both between convergences and<br />

divergences in themselves and between each other. It would be a sort of universal<br />

language, in binary alphabet, "antiBabel", still in virtual reality. From this, going in the<br />

reverse direction, there could be extracted in the respective alphabets of the individual<br />

languages A, B, C, D, etc., the number of words and expressions, situations of<br />

morphology and syntax, of each language which have found correspondence in other<br />

languages, and the number of those which have not. The number of such correspondences<br />

thus extracted (lexicon and grammar) would be a set of "disciplined" basic languages, <strong>to</strong><br />

be adopted for the telematic use of the computer, <strong>to</strong> be also printed and then updated<br />

according <strong>to</strong> experience.<br />

In input, therefore, everybody could use their own native disciplined language and have<br />

the desired translations in output. The addressee could even receive the message both in<br />

their own language, in that of the sender, and in others.<br />

In addition, the problems of keys and privacy will have <strong>to</strong> be solved, as (one step at a<br />

time!) will those of the phonetic version of both input and output.<br />

These thoughts have formed gradually in my mind over the years, starting from the<br />

realization that my programs for Latin, which I always wanted broken up for<br />

monofunctional use, could be applied with the same operative philosophy <strong>to</strong> more than<br />

twenty other languages (all in a phonetic script), even those that do not descend from<br />

Latin, such as Arabic and Hebrew, which are written from right <strong>to</strong> left. I had only <strong>to</strong><br />

transfer elements from one table <strong>to</strong> another, changing the length of fields, or adding a<br />

field. (However, I cannot say anything about languages written in ideograms or<br />

pic<strong>to</strong>grams.)<br />

Conclusion<br />

In conclusion, I will therefore summarize the third perspective, that of textual<br />

hermeneutic informatics, as follows. The first period began with my Index Thomisticus<br />

and ended, though not for me, with the Alpac Report. The second, after the Alpac Report,<br />

is the parcelization in progress of free research. The third would begin if and when<br />

comparative global informatics begins in the principal languages, more or less in the<br />

sense I have tried <strong>to</strong> sketch here. Those who live long enough will see whether these<br />

roses (and thorns), which <strong>to</strong>day are merely thoughts, will come <strong>to</strong> flower, and so will be<br />

able <strong>to</strong> tell whether they were prophecy or dream.<br />

The Digital Humanities and Humanities Computing:<br />

An Introduction


Susan Schreibman, Ray Siemens, and John Unsworth<br />

His<strong>to</strong>ry<br />

This collection marks a turning point in the field of <strong>digital</strong> <strong>humanities</strong>: for the first time, a<br />

wide range of theorists and practitioners, those who have been active in the field for<br />

decades, and those recently involved, disciplinary experts, computer scientists, and<br />

library and information studies specialists, have been brought <strong>to</strong>gether <strong>to</strong> consider <strong>digital</strong><br />

<strong>humanities</strong> as a discipline in its own right, as well as <strong>to</strong> reflect on how it relates <strong>to</strong> areas<br />

of traditional <strong>humanities</strong> scholarship.<br />

This collection has its origins in the research carried out over the past half a century in<br />

textually focused computing in the <strong>humanities</strong>, as Susan Hockey notes in this volume,<br />

found in Father Rober<strong>to</strong> Busa's adaptation of early computing <strong>to</strong> textual location,<br />

comparison, and counting as part of his work in creating the Index Thomisticus, a<br />

concordance <strong>to</strong> the works of St Thomas Aquinas. Yet even a cursory glance at this<br />

Companion's table of contents reveals how broadly the field now defines itself. It remains<br />

deeply interested in text, but as advances in technology have made it first possible, then<br />

trivial <strong>to</strong> capture, manipulate, and process other media, the field has redefined itself <strong>to</strong><br />

embrace the full range of multimedia. Especially since the 1990s, with the advent of the<br />

World Wide Web, <strong>digital</strong> <strong>humanities</strong> has broadened its reach, yet it has remained in <strong>to</strong>uch<br />

with the goals that have animated it from the outset: using information technology <strong>to</strong><br />

illuminate the human record, and bringing an understanding of the human record <strong>to</strong> bear<br />

on the development and use of information technology.<br />

The first section of the Companion addresses the field of <strong>digital</strong> <strong>humanities</strong> from<br />

disciplinary perspectives. Although the breadth of fields covered is wide, what is revealed<br />

is how computing has cut across disciplines <strong>to</strong> provide not only <strong>to</strong>ols, but methodological<br />

focal points. There is, for example, a shared focus on preserving physical artifacts<br />

(written, painted, carved, or otherwise created), that which is left <strong>to</strong> us by chance (ruin,<br />

and other debris of human activity), or that which has been near-impossible <strong>to</strong> capture in<br />

its intended form (music, performance, and event). Yet many disciplines have gone<br />

beyond simply wishing <strong>to</strong> preserve these artifacts, what we might now call early forms of<br />

data management, <strong>to</strong> re-represent and manipulate them <strong>to</strong> reveal properties and traits not<br />

evident when the artifact was in its native form. Moreover, <strong>digital</strong> <strong>humanities</strong> now also<br />

concerns itself with the creation of new artifacts which are born <strong>digital</strong> and require<br />

rigorous study and understanding in their own right.<br />

Eiteljorg notes that archaeologists, like most other early adopters in the arts and<br />

<strong>humanities</strong>, first used the computer for record making and record keeping, in the<br />

knowledge that data in this form would see more flexible utilization, particularly in<br />

computer-assisted statistical analysis. More recent applications derive from the<br />

introduction of global data-recording standards, allowing large corpora of archaeological<br />

data <strong>to</strong> be navigated, as well as the integration of global information systems (GlS) –<br />

derived data <strong>to</strong> represent standard locational information across these corpora. Art<br />

his<strong>to</strong>rians, as described by Greenhalgh, use computers <strong>to</strong> image, order, sort, interrogate,


and analyze data about artworks, and increasingly use the Internet as the carrier for<br />

multimedia research or teaching/learning projects. Classical studies has always been a<br />

data-intensive enterprise, Crane demonstrates, and has seen the development of lexica,<br />

encyclopedias, commentaries, critical editions, and other elements of scholarly<br />

infrastructure that are well suited <strong>to</strong> an electronic environment which, ultimately, reflects<br />

a natural impulse <strong>to</strong>ward systematic knowledge management and engineering within the<br />

field. So, <strong>to</strong>o, is this impulse essential <strong>to</strong> the understanding of computing's role in literary<br />

studies (as noted by Rommel), linguistics (as discussed by Hajic), and lexicography (as<br />

charted by Wooldridge). In discussing the discipline of musicology, Fujinaga and Weiss<br />

note that the Internet has revolutionized not only the distribution potential for the artifacts<br />

that lie at the heart of their consideration but, also, other more analytical applications<br />

pertinent <strong>to</strong> the future of the field. Thomas documents the intense methodological debates<br />

sparked by the introduction of computing in his<strong>to</strong>ry, debates which computing ultimately<br />

lost (in the United States, at least), after which it <strong>to</strong>ok a generation for his<strong>to</strong>rians <strong>to</strong><br />

reconsider the usefulness of the computer <strong>to</strong> their discipline. The rhe<strong>to</strong>ric of revolution<br />

proved more predictive in other disciplines, though – for example, in philosophy and<br />

religion. Today, one hears less and less of it, perhaps because (as Ess notes) the<br />

revolution has succeeded: in almost all disciplines, the power of computers, and even<br />

their potential, no longer seem revolutionary at all. While this may be true of a number of<br />

disciplines, in fields such as the performing arts, as discussed by Saltz, and new media<br />

studies, by Rockwell and Mactavish, there is an inherent kinship between the everevolving<br />

developments in computing and their performative and analytical potentials.<br />

Principles, Applications, and Dissemination<br />

The <strong>digital</strong> <strong>humanities</strong>, then, and their interdisciplinary core found in the field of<br />

<strong>humanities</strong> computing, have a long and dynamic his<strong>to</strong>ry best illustrated by examination<br />

of the locations at which specific disciplinary practices intersect with computation. Even<br />

so, just as the various fields that make up the <strong>humanities</strong> share a focus on the examination<br />

of artifactual evidence of that which makes us human, so, <strong>to</strong>o, do these fields share a<br />

number of commonly held assumptions about the way in which such examination is<br />

carried out, both with and without the assistance of the computer. Widely spread through<br />

the <strong>digital</strong> <strong>humanities</strong> community is the notion that there is a clear and direct relationship<br />

between the interpretative strategies that humanists employ and the <strong>to</strong>ols that facilitate<br />

exploration of original artifacts based on those interpretative strategies; or, more simply<br />

put, those working in the <strong>digital</strong> <strong>humanities</strong> have long held the view that application is as<br />

important as theory. Thus, exemplary tasks traditionally associated with <strong>humanities</strong><br />

computing hold the <strong>digital</strong> representation of archival materials on a par with analysis or<br />

critical inquiry, as well as theories of analysis or critical inquiry originating in the study<br />

of those materials. The field also places great importance on the means of disseminating<br />

the results of these activities and, as Pitti discusses, recognizes that project conception<br />

and management can be as important pragmatic concerns as others which are more<br />

traditionally associated with disciplinary pursuits.<br />

The representation of archival material involves the use of computer-assisted means <strong>to</strong><br />

describe and express print-, visual-, and audio-based material in tagged and searchable


electronic form, as discussed by Deegan and Tanner. This representation is a critical and<br />

self-conscious activity, from the choice of what <strong>to</strong> represent <strong>to</strong> the reproduction of<br />

primary materials, for example, in the preparation of an electronic edition or <strong>digital</strong><br />

facsimile (as discussed by Smith) and the exploration of the relationship between <strong>digital</strong><br />

surrogates <strong>to</strong> legacy data (Warwick). Related <strong>to</strong> the process of representation is the<br />

necessity of understanding the <strong>to</strong>ols being used, as Laue outlines, and the implications of<br />

decisions that we make in the use of those <strong>to</strong>ols and the impact they have on analytical<br />

processes (McGann). The growing field of knowledge representation, which draws on the<br />

field of artificial intelligence and seeks <strong>to</strong> "produce models of human understanding that<br />

are tractable <strong>to</strong> computation" (Unsworth <strong>2001</strong>), provides a lens through which we might<br />

understand such implications. This is especially true in issues related <strong>to</strong> archival<br />

representation and textual editing, high-level interpretative theory and criticism, and<br />

pro<strong>to</strong>cols of knowledge transfer – as modeled with computational techniques (discussed<br />

by McCarty), and captured via encoding and classification systems (Renear; Sperberg-<br />

McQueen) and represented in data structures (Ramsay), some of which have great impact<br />

on the ways in which we associate human information (Ryan) and interpret the ways in<br />

which it has influence upon us (Drucker).<br />

In the <strong>digital</strong> <strong>humanities</strong>, critical inquiry involves the application of algorithmically<br />

facilitated search, retrieval, and critical processes that, originating in <strong>humanities</strong>-based<br />

work, have been demonstrated <strong>to</strong> have application far beyond. Associated with critical<br />

theory, this area is typified by interpretative studies that assist in our intellectual and<br />

aesthetic understanding of humanistic works. It also involves the application (and<br />

applicability) of critical and interpretative <strong>to</strong>ols and analytic algorithms, as discussed by<br />

Bradley, on those artifacts produced through processes associated with archival<br />

representation made available via resources associated with processes of publishing and<br />

the communication of results. Manifested in the analysis techniques that Burrows and Ide<br />

each discuss – and seeing utility in a wide-ranging array of applications, from authorship<br />

attribution (Craig) <strong>to</strong> cognitive stylistics (Lancashire) – the basis of such analysis is the<br />

encoded and <strong>digital</strong>ly s<strong>to</strong>red corpora governed by strategies of knowledge representation<br />

that, themselves, are capable of possessing what we might term a "poetics" (Winder). So,<br />

<strong>to</strong>o, with <strong>digital</strong> media such as film (Kolker), and with issues of interface and usability<br />

that are, as Kirschenbaum discusses, integral <strong>to</strong> all materials in electronic form and our<br />

interaction with them. Further, efforts <strong>to</strong>ward dissemination have their roots, ultimately,<br />

in issues related <strong>to</strong> re-presentation but are themselves manifested in concerns pertinent <strong>to</strong><br />

the nature of computer-facilitated communities (Willett): preservation in the electronic<br />

medium (discussed by Abby Smith), professional electronic publication (treated by<br />

Jensen's chapter, and addressed further by Palmer), and the unique array of challenges<br />

and opportunities that arise with the emergence of <strong>digital</strong> libraries, as outlined by Besser.<br />

Conclusion<br />

The edi<strong>to</strong>rs intended this collection <strong>to</strong> serve as a his<strong>to</strong>rical record of the field, capturing a<br />

sense of the <strong>digital</strong> <strong>humanities</strong> as they have evolved over the past half century, and as<br />

they exist at the moment. Yet, if one looks at the issues that lie at the heart of nearly all<br />

contributions <strong>to</strong> this volume, one will see that these contributions reflect a relatively clear


view of the future of the <strong>digital</strong> <strong>humanities</strong>. In addition <strong>to</strong> charting areas in which past<br />

advances have been made, and in which innovation is currently taking place, this volume<br />

reveals that <strong>digital</strong> <strong>humanities</strong> is addressing many of the most basic research paradigms<br />

and methods in the disciplines, <strong>to</strong> focus our attention on important questions <strong>to</strong> be asked<br />

and answered, in addition <strong>to</strong> important new ways of asking and answering that are<br />

enabled by our interaction with the computer.<br />

What this collection also reveals is that there are central concerns among <strong>digital</strong><br />

humanists which cross disciplinary boundaries. This is nowhere more evident than in the<br />

representation of knowledge-bearing artifacts. The process of such representation –<br />

especially so when done with the attention <strong>to</strong> detail and the consistency demanded by the<br />

computing environment – requires humanists <strong>to</strong> make explicit what they know about their<br />

material and <strong>to</strong> understand the ways in which that material exceeds or escapes<br />

representation. Ultimately, in computer-assisted analysis of large amounts of material that<br />

has been encoded and processed according <strong>to</strong> a rigorous, well thought-out system of<br />

knowledge representation, one is afforded opportunities for perceiving and analyzing<br />

patterns, conjunctions, connections, and absences that a human being, unaided by the<br />

computer, would not be likely <strong>to</strong> find.<br />

The process that one goes through in order <strong>to</strong> develop, apply, and compute these<br />

knowledge representations is unlike anything that <strong>humanities</strong> scholars, outside of<br />

philosophy, have ever been required <strong>to</strong> do. This method, or perhaps we should call it a<br />

heuristic, discovers a new horizon for <strong>humanities</strong> scholarship, a paradigm as powerful as<br />

any that has arisen in any <strong>humanities</strong> discipline in the past – and, indeed, maybe more<br />

powerful, because the rigor it requires will bring <strong>to</strong> our attention undocumented features<br />

of our own ideation. Coupled with enormous s<strong>to</strong>rage capacity and computational power,<br />

this heuristic presents us with patterns and connections in the human record that we<br />

would never otherwise have found or examined.<br />

Acknowledgments<br />

The edi<strong>to</strong>rs would like <strong>to</strong> thank Emma Bennett, Andrew McNeillie, and Karen Wilson at<br />

Blackwell for their assistance, encouragement, and support. Ray Siemens would also like<br />

<strong>to</strong> acknowledge the assistance of Karin Armstrong and Barbara Bond with some<br />

materials present in this volume, and the Malaspina Research and Scholarly Activity<br />

Committee, for their support.<br />

McCarty, Willard. What is Humanities Computing? Toward a Definition of the Field.<br />

URL: http://www.cch.kcl.ac.uk/legacy/staff/wlm/essays/what/<br />

Schreibman, Susan (2002). Computer-mediated Discourse: Reception Theory and<br />

Versioning. Computers and the Humanities 36, 3: 283–93.<br />

Siemens, R. G. (2002). A New Computer-assisted Literary Criticism? Introduction <strong>to</strong> A<br />

New Computer-assisted Literary Criticism?, ed. R. G. Siemens. [A special issue of]<br />

Computers and the Humanities 363: 259–67.


Unsworth, John (<strong>2001</strong>). Knowledge Representation in Humanities Computing. Inaugural<br />

E-<strong>humanities</strong> Lecture at the National Endowment for the Humanities (April 3). URL:<br />

http://www.iath.virginia.edu/~jmu2m/KR/.<br />

1.<br />

The His<strong>to</strong>ry of Humanities Computing<br />

Susan Hockey<br />

Introduction<br />

Tracing the his<strong>to</strong>ry of any interdisciplinary academic area of activity raises a number of<br />

basic questions. What should be the scope of the area? Is there overlap with related areas,<br />

which has impacted on the development of the activity? What has been the impact on<br />

other, perhaps more traditional, disciplines? Does a straightforward chronological<br />

account do justice <strong>to</strong> the development of the activity? Might there be digressions from<br />

this, which could lead us in<strong>to</strong> hither<strong>to</strong> unexplored avenues? Each of these questions could<br />

form the basis of an essay in itself but within the space and context available here, the<br />

approach taken is <strong>to</strong> present a chronological account which traces the development of<br />

<strong>humanities</strong> computing. Within this, the emphasis is on highlighting landmarks where<br />

significant intellectual progress has been made or where work done within <strong>humanities</strong><br />

computing has been adopted, developed or drawn on substantially within other<br />

disciplines.<br />

It is not the place of this essay <strong>to</strong> define what is meant by <strong>humanities</strong> computing. The<br />

range of <strong>to</strong>pics within this Companion indeed sends plenty of signals about this. Suffice it<br />

<strong>to</strong> say that we are concerned with the applications of computing <strong>to</strong> research and teaching<br />

within subjects that are loosely defined as "the <strong>humanities</strong>", or in British English "the<br />

arts." Applications involving textual sources have taken center stage within the<br />

development of <strong>humanities</strong> computing as defined by its major publications and thus it is<br />

inevitable that this essay concentrates on this area. Nor is it the place here <strong>to</strong> attempt <strong>to</strong><br />

define "interdisciplinarity", but by its very nature, <strong>humanities</strong> computing has had <strong>to</strong><br />

embrace "the two cultures", <strong>to</strong> bring the rigor and systematic unambiguous procedural<br />

methodologies characteristic of the sciences <strong>to</strong> address problems within the <strong>humanities</strong><br />

that had hither<strong>to</strong> been most often treated in a serendipi<strong>to</strong>us fashion.<br />

Beginnings: 1949 <strong>to</strong> early 1970s<br />

Unlike many other interdisciplinary experiments, <strong>humanities</strong> computing has a very wellknown<br />

beginning. In 1949, an Italian Jesuit priest, Father Rober<strong>to</strong> Busa, began what even<br />

<strong>to</strong> this day is a monumental task: <strong>to</strong> make an index verborum of all the words in the<br />

works of St Thomas Aquinas and related authors, <strong>to</strong>taling some 11 million words of<br />

medieval Latin. Father Busa imagined that a machine might be able <strong>to</strong> help him, and,


having heard of computers, went <strong>to</strong> visit Thomas J. Watson at IBM in the United States<br />

in search of support (Busa 1980). Some assistance was forthcoming and Busa began his<br />

work. The entire texts were gradually transferred <strong>to</strong> punched cards and a concordance<br />

program written for the project. The intention was <strong>to</strong> produce printed volumes, of which<br />

the first was published in 1974 (Busa 1974).<br />

A purely mechanical concordance program, where words are alphabetized according <strong>to</strong><br />

their graphic forms (sequences of letters), could have produced a result in much less time,<br />

but Busa would not be satisfied with this. He wanted <strong>to</strong> produce a "lemmatized"<br />

concordance where words are listed under their dictionary headings, not under their<br />

simple forms. His team attempted <strong>to</strong> write some computer software <strong>to</strong> deal with this and,<br />

eventually, the lemmatization of all 11 million words was completed in a semiau<strong>to</strong>matic<br />

way with human beings dealing with word forms that the program could not handle. Busa<br />

set very high standards for his work. His volumes are elegantly typeset and he would not<br />

compromise on any levels of scholarship in order <strong>to</strong> get the work done faster. He has<br />

continued <strong>to</strong> have a profound influence on <strong>humanities</strong> computing, with a vision and<br />

imagination that reach beyond the horizons of many of the current generation of<br />

practitioners who have been brought up with the Internet. A CD-ROM of the Aquinas<br />

material appeared in 1992 that incorporated some hypertextual features ("cum<br />

hypertextibus") (Busa 1992) and was accompanied by a user guide in Latin, English, and<br />

Italian. Father Busa himself was the first recipient of the Busa award in recognition of<br />

outstanding achievements in the application of information technology <strong>to</strong> humanistic<br />

research, and in his award lecture in Debrecen, Hungary, in 1998 he reflected on the<br />

potential of the World Wide Web <strong>to</strong> deliver multimedia scholarly material accompanied<br />

by sophisticated analysis <strong>to</strong>ols (Busa 1999).<br />

By the 1960s, other researchers had begun <strong>to</strong> see the benefits of working with<br />

concordances. A series of four articles by Dolores Bur<strong>to</strong>n in the journal Computers and<br />

the Humanities in 1981–2 attempted <strong>to</strong> bring these <strong>to</strong>gether, beginning with a discussion<br />

of the 1950s (Bur<strong>to</strong>n 1981a, 1981b, 1981c, 1982). Some of these researchers were<br />

individual scholars whose interests concentrated on one set of texts or authors. In the UK,<br />

Roy Wisbey produced a series of indexes <strong>to</strong> Early Middle High German texts (Wisbey<br />

1963). In the USA Stephen Parrish's concordances <strong>to</strong> the poems of Matthew Arnold and<br />

W B. Yeats introduced the series of concordances published by Cornell University Press<br />

(Parrish 1962). This period also saw the establishment of computing facilities in some<br />

major language academies in Europe, principally <strong>to</strong> assist with the compilation of<br />

dictionaries. Examples include the Trésor de la Langue Française (Gorcy 1983), which<br />

was established in Nancy <strong>to</strong> build up an archive of French literary material, and the<br />

Institute of Dutch Lexicology in Leiden (De Tollenaere 1973).<br />

Although much activity at this time was concentrated on the production of concordances<br />

as ends in themselves, one application of these <strong>to</strong>ols began <strong>to</strong> take on a life of its own.<br />

The use of quantitative approaches <strong>to</strong> style and authorship studies predates computing.<br />

For example, Augustus de Morgan in a letter written in 1851 proposed a quantitative<br />

study of vocabulary as a means of investigating the authorship of the Pauline Epistles<br />

(Lord 1958) and T. C. Mendenhall, writing at the end of the nineteenth century, described


his counting machine, whereby two ladies computed the number of words of two letters,<br />

three, and so on in Shakespeare, Marlowe, Bacon, and many other authors in an attempt<br />

<strong>to</strong> determine who wrote Shakespeare (Mendenhall 1901). But the advent of computers<br />

made it possible <strong>to</strong> record word frequencies in much greater numbers and much more<br />

accurately than any human being can. In 1963, a Scottish clergyman, Andrew Mor<strong>to</strong>n,<br />

published an article in a British newspaper claiming that, according <strong>to</strong> the computer, St<br />

Paul only wrote four of his epistles. Mor<strong>to</strong>n based his claim on word counts of common<br />

words in the Greek text, plus some elementary statistics. He continued <strong>to</strong> examine a<br />

variety of Greek texts producing more papers and books concentrating on an examination<br />

of the frequencies of common words (usually particles) and also on sentence lengths,<br />

although it can be argued that the punctuation which identifies sentences was added <strong>to</strong> the<br />

Greek texts by modern edi<strong>to</strong>rs (Mor<strong>to</strong>n 1965; Mor<strong>to</strong>n and Winspear 1971).<br />

It is believed that the first use of computers in a disputed authorship study was carried out<br />

on the Junius Letters by Alvar Ellegard. Published in 1962, this study did not use a<br />

computer <strong>to</strong> make the word counts, but did use machine calculations which helped<br />

Ellegard get an overall picture of the vocabulary from hand counts (Ellegard 1962). What<br />

is probably the most influential computer-based authorship investigation was also carried<br />

out in the early 1960s. This was the study by Mosteller and Wallace of the Federalist<br />

Papers in an attempt <strong>to</strong> identify the authorship of the twelve disputed papers (Mosteller<br />

and Wallace 1964). With so much material by both authorship candidates on the same<br />

subject matter as the disputed papers, this study presented an ideal situation for<br />

comparative work. Mosteller and Wallace were primarily interested in the statistical<br />

methods they employed, but they were able <strong>to</strong> show that Madison was very likely <strong>to</strong> have<br />

been the author of the disputed papers. Their conclusions generally have been accepted,<br />

<strong>to</strong> the extent that the Federalist Papers have been used as a test for new methods of<br />

authorship discrimination (Holmes and Forsyth 1995; Tweedie et al. 1996).<br />

At this time much attention was paid <strong>to</strong> the limitations of the technology. Data <strong>to</strong> be<br />

analyzed were either texts or numbers. They were input laboriously by hand either on<br />

punched cards, with each card holding up <strong>to</strong> eighty characters or one line of text<br />

(uppercase letters only), or on paper tape, where lower-case letters were perhaps possible<br />

but which could not be read in any way at all by a human being. Father Busa has s<strong>to</strong>ries<br />

of truckloads of punched cards being transported from one center <strong>to</strong> another in Italy. All<br />

computing was carried out as batch processing, where the user could not see the results at<br />

all until prin<strong>to</strong>ut appeared when the job had run. Character-set representation was soon<br />

recognized as a substantial problem and one that has only just begun <strong>to</strong> be solved now<br />

with the advent of Unicode, although not for every kind of <strong>humanities</strong> material. Various<br />

methods were devised <strong>to</strong> represent upper- and lower-case letters on punched cards, most<br />

often by inserting an asterisk or similar character before a true upper-case letter. Accents<br />

and other non-standard characters had <strong>to</strong> be treated in a similar way and non-Roman<br />

alphabets were represented entirely in transliteration.<br />

Most large-scale datasets were s<strong>to</strong>red on magnetic tape, which can only be processed<br />

serially. It <strong>to</strong>ok about four minutes for a full-size tape <strong>to</strong> wind from one end <strong>to</strong> the other<br />

and so software was designed <strong>to</strong> minimize the amount of tape movement. Random access


<strong>to</strong> data such as happens on a disk was not possible. Data had therefore <strong>to</strong> be s<strong>to</strong>red in a<br />

serial fashion. This was not so problematic for textual data, but for his<strong>to</strong>rical material it<br />

could mean the simplification of data, which represented several aspects of one object<br />

(forming several tables in relational database technology), in<strong>to</strong> a single linear stream.<br />

This in itself was enough <strong>to</strong> deter his<strong>to</strong>rians from embarking on computer-based projects.<br />

Representation problems extended far beyond specific characters. For concordance and<br />

retrieval programs there was a need <strong>to</strong> identify citations by their location within the text.<br />

The methods used by conventional document retrieval systems were inadequate because<br />

they tended <strong>to</strong> assume document structures similar <strong>to</strong> those of journal articles and were<br />

unable <strong>to</strong> cope with the structures found in poetry or drama, or in manuscript sources<br />

where the lineation is important. Various methods of defining document structures were<br />

proposed, but the most sophisticated one developed at this time was that used by the<br />

COCOA concordance program (Russell 1967). Modeled on a format developed by Paul<br />

Bratley for an Archive of Older Scottish texts (Hamil<strong>to</strong>n-Smith 1971), COCOA enables<br />

the user <strong>to</strong> define a specification for the document structure which matches the particular<br />

set of documents. It also enables the markup of overlapping structures, making it<br />

possible, for example, <strong>to</strong> encode a citation system for a printed version in parallel with<br />

that for the manuscript source of the material. COCOA is also economical of file space,<br />

but is perhaps less readable for the human.<br />

The other widely used citation scheme was more dependent on punched card format. In<br />

this scheme, often called "fixed format", every line began with a coded sequence of<br />

characters giving citation information. Each unit within the citation was positioned in<br />

specific columns across the line, for example the title in columns 1–3, verse number in<br />

columns 5–6, and line number in columns 7–9. The entry of this information was speeded<br />

up by functions on the punched card machine, but the information also occupied more<br />

space within the computer file.<br />

The legacy of these citation schemes can still be found in electronic texts created some<br />

time ago. COCOA, particularly, was very influential and other schemes were derived<br />

from it. COCOA cannot easily handle the markup of small features within the content<br />

such as names, dates, and abbreviations, but its ability <strong>to</strong> deal with overlapping structures<br />

outstrips that of almost all modern markup schemes.<br />

This period also saw the first opportunities for those interested in <strong>humanities</strong> computing<br />

<strong>to</strong> get <strong>to</strong>gether <strong>to</strong> share ideas and problems. In 1964, IBM organized a conference at<br />

York<strong>to</strong>wn Heights. The subsequent publication, Literary Data Processing Conference<br />

Proceedings, edited by Jess Bessinger and Stephen Parrish (1965), almost reads like<br />

something from twenty or so years later, except for the reliance on punched cards for<br />

input. Papers discuss complex questions in encoding manuscript material and also in<br />

au<strong>to</strong>mated sorting for concordances where both variant spellings and the lack of<br />

lemmatization are noted as serious impediments.<br />

As far as can be ascertained, the York<strong>to</strong>wn Heights conference was a one-off event. The<br />

first of a regular series of conferences on literary and linguistic computing and the


precursor of what became the Association for Literary and Linguistic<br />

Computing/Association for Computers and the Humanities (ALLC/ACH) conferences<br />

was organized by Roy Wisbey and Michael Farringdon at the University of Cambridge in<br />

March, 1970. This was a truly international event with good representation from both<br />

sides of the Atlantic as well as from Australia. The proceedings, meticulously edited by<br />

Wisbey (1971), set the standard for subsequent publications. A glance through them<br />

indicates the emphasis of interest on input, output, and programming as well as<br />

lexicography, textual editing, language teaching, and stylistics. Even at this time the need<br />

for a methodology for archiving and maintaining electronic texts was fully recognized.<br />

Another indication of an embryonic subject area is the founding of a new journal.<br />

Computers and the Humanities began publication in 1966 under the edi<strong>to</strong>rship of Joseph<br />

Raben. With characteristic energy, Raben nurtured the new journal and during its first<br />

years, at least until the regular series of conferences and associations that developed from<br />

them got going, it became the main vehicle for dissemination of information about<br />

<strong>humanities</strong> computing. Raben recognized the need just <strong>to</strong> know what is going on and the<br />

journal's Direc<strong>to</strong>ry of Scholars Active was the first point of call for people who were<br />

thinking about starting a project. Other informal newsletters also served specific<br />

communities, notably Calculi for computers and classics, edited by Stephen Waite.<br />

The 1960s also saw the establishment of some centers dedicated <strong>to</strong> the use of computers<br />

in the <strong>humanities</strong>. Wisbey founded the Centre for Literary and Linguistic Computing in<br />

Cambridge in 1963 as support for his work with Early Middle High German Texts. In<br />

Tübingen, Wilhelm Ott established a group which began <strong>to</strong> develop the suite of programs<br />

for text analysis, particularly for the production of critical editions. The TuStep software<br />

modules are in use <strong>to</strong> this day and set very high standards of scholarship in dealing with<br />

all phases from data entry and collation <strong>to</strong> the production of complex print volumes.<br />

Work in this early period is often characterized as being hampered by technology, where<br />

technology is taken <strong>to</strong> mean character sets, input/output devices and the slow turnaround<br />

of batch processing systems. However, researchers did find ways of dealing with some of<br />

these problems, albeit in a cumbersome way. What is more characteristic is that key<br />

problems which they identified are still with us, notably the need <strong>to</strong> look at "words"<br />

beyond the level of the graphic string, and <strong>to</strong> deal effectively with variant spellings,<br />

multiple manuscripts, and lemmatization.<br />

Consolidation: 1970s <strong>to</strong> mid-1980s<br />

If any single-word term can be used <strong>to</strong> describe this period, it would almost certainly be<br />

"consolidation." More people were using methodologies developed during the early<br />

period. More electronic texts were being created and more projects using the same<br />

applications were started. Knowledge of what is possible had gradually spread through<br />

normal scholarly channels of communication, and more and more people had come<br />

across computers in their everyday life and had begun <strong>to</strong> think about what computers<br />

might do for their research and teaching.


The diffusion of knowledge was helped not only by Computers and the Humanities but<br />

also by a regular series of conferences. The 1970 symposium in Cambridge was the start<br />

of a biennial series of conferences in the UK, which became a major focal point for<br />

computing in the <strong>humanities</strong>. Meetings in Edinburgh (1972), Cardiff (1974), Oxford<br />

(1976), Birmingham (1978), and Cambridge (1980) all produced high-quality papers. The<br />

Association for Literary and Linguistic Computing was founded at a meeting in King's<br />

College London in 1973. Initially it produced its own Bulletin three times per year. It also<br />

began <strong>to</strong> organize an annual meeting with some invited presentations and by 1986 had a<br />

journal, Literary and Linguistic Computing. By the mid-1970s, another series of<br />

conferences began in North America, called the International Conference on Computing<br />

in the Humanities (ICCH), and were held in odd-numbered years <strong>to</strong> alternate with the<br />

British meetings. The British conference and the ALLC annual meetings gradually began<br />

<strong>to</strong> coalesce. They continued <strong>to</strong> concentrate on literary and linguistic computing with<br />

some emphasis on "linguistic", where they offered a forum for the growing number of<br />

European researchers in what became known as corpus linguistics. ICCH attracted a<br />

broader range of papers, for example on the use of computers in teaching writing, and on<br />

music, art, and archaeology. The Association for Computers and the Humanities (ACH)<br />

grew out of this conference and was founded in 1978.<br />

The requirements of <strong>humanities</strong> computing also began <strong>to</strong> be recognized within academic<br />

computing centers. Still in the days of mainframe computing, it was necessary <strong>to</strong> register<br />

<strong>to</strong> use any computing facilities and that registration provided an opportunity for academic<br />

computing staff <strong>to</strong> find out what users wanted and <strong>to</strong> consider providing some standard<br />

software that could be used by many different people. The second version of the COCOA<br />

concordance program in Britain was designed <strong>to</strong> be run on different mainframe<br />

computers for exactly this purpose (Berry-Rogghe and Crawford 1973). It was distributed<br />

<strong>to</strong> different computing centers in the mid-1970s and many of these centers designated one<br />

person <strong>to</strong> act as support. Dissatisfaction with its user interface coupled with the<br />

termination of support by the Atlas Labora<strong>to</strong>ry, where it was written, led the British<br />

funding bodies <strong>to</strong> sponsor the development of a new program at Oxford University.<br />

Called the Oxford Concordance Program (OCP), this software was ready for distribution<br />

in 1982 and attracted interest around the world with users in many different countries<br />

(Hockey and Marriott 1979a, 1979b, 1979c, 1980). Other packaged or generic software<br />

also appeared at this time and significantly reduced the cost of a project in terms of<br />

programming support.<br />

The need <strong>to</strong> avoid duplication of effort also led <strong>to</strong> consolidation in the area of text<br />

archiving and maintenance. With the advent of packaged software and the removal of the<br />

need for much programming, preparing the electronic text began <strong>to</strong> take up a large<br />

proportion of time in any project. The key driver behind the establishment of the Oxford<br />

Text Archive (OTA) in 1976 was the need simply <strong>to</strong> ensure that a text that a researcher<br />

had finished with was not lost. The OTA under<strong>to</strong>ok <strong>to</strong> maintain electronic texts and,<br />

subject <strong>to</strong> the permission of the deposi<strong>to</strong>r and with appropriate copyright permissions, <strong>to</strong><br />

make these texts available <strong>to</strong> anyone else who wanted <strong>to</strong> use them for academic purposes.<br />

It was the beginnings of a <strong>digital</strong> library, although nobody called it this initially, and its<br />

staff had <strong>to</strong> devise their own method of describing and documenting the material (Proud


1989). The amount of undocumented material highlighted the need for recognized<br />

procedures for describing electronic texts.<br />

The OTA's approach was <strong>to</strong> offer a service for maintenance of anything that was<br />

deposited. It managed <strong>to</strong> do this for some considerable time on very little budget, but was<br />

not able <strong>to</strong> promote the creation of specific texts. Groups of scholars in some discipline<br />

areas made more concerted attempts <strong>to</strong> create an archive of texts <strong>to</strong> be used as a source<br />

for research. Notable among these was the Thesaurus Linguae Graecae (TLG) begun at<br />

the University of California Irvine and directed for many years by Theodore Brunner.<br />

Brunner raised millions of dollars <strong>to</strong> support the creation of a "databank" of Ancient<br />

Greek texts, covering all authors from Homer <strong>to</strong> about ad 600, some 70 million words<br />

(Brunner 1993). A complementary collection of Classical Latin was later produced by the<br />

Packard Humanities Institute, and <strong>to</strong>gether with the TLG gave scholars in classical<br />

studies a research resource that was unrivaled in other disciplines for many years. Only<br />

Old English scholars had access <strong>to</strong> a similar comprehensive, but smaller corpus with the<br />

completion of the Old English Corpus for the Dictionary of Old English (Healey 1989).<br />

More centers for <strong>humanities</strong> computing were also established during this period. Some,<br />

for example the Norwegian Computing Center for the Humanities (now HIT) at Bergen,<br />

with substantial government support, incorporated a wide range of applications and<br />

projects. Others such as the Center for Computer Analysis of Texts (CCAT) at the<br />

University of Pennsylvania were more narrowly focused on the interests of the academics<br />

who had initially promoted them. Pockets of interest had become established around the<br />

world and scholars in those institutions on the whole enjoyed a good deal of support.<br />

This period also saw the introduction of courses on various aspects of <strong>humanities</strong><br />

computing. Some courses were given by staff within academic computing centers and<br />

concentrated mostly on the mechanics of using specific software programs. Others looked<br />

more broadly at application areas. Those given by academics tended <strong>to</strong> concentrate on<br />

their own interests giving rise <strong>to</strong> student projects in the same application areas. A debate<br />

about whether or not students should learn computer programming was ongoing. Some<br />

felt that it replaced Latin as a "mental discipline" (Hockey 1986). Others thought that it<br />

was <strong>to</strong>o difficult and <strong>to</strong>ok <strong>to</strong>o much time away from the core work in the <strong>humanities</strong>. The<br />

string handling language SNOBOL was in vogue for some time as it was easier for<br />

<strong>humanities</strong> students than other computer languages, of which the major one was still<br />

Fortran.<br />

There were some developments in processing <strong>to</strong>ols, mostly through the shift from tape <strong>to</strong><br />

disk s<strong>to</strong>rage. Files no longer had <strong>to</strong> be searched sequentially. For a time there were<br />

various technologies for organizing material in databases, some of which were very<br />

effective for <strong>humanities</strong> material (Burnard 1987b), but gradually the relational model<br />

prevailed. In mainframe implementations this presented a better structure within which<br />

his<strong>to</strong>rians and others working with material drawn from sources (rather than the sources<br />

themselves) could work. However, relational technologies still presented some problems<br />

for the representation of information that needed <strong>to</strong> be fitted in<strong>to</strong> tables. At least two<br />

hardware devices were invented in the 1970s for assisting searching. One was


implemented in David Packard's Ibycus computer, which was built <strong>to</strong> work with the TLG<br />

and some other classics material (Lancashire 1991: 204–5). The other was the Content<br />

Addressing File S<strong>to</strong>re (CAFS), which worked on the British ICL computers (Burnard<br />

1987a). The idea of transferring processing in<strong>to</strong> the hardware was very attractive <strong>to</strong><br />

<strong>humanities</strong> researchers who had <strong>to</strong> deal with large amounts of material, but it did not<br />

catch on in a big way, possibly because it was overtaken by advances in the speed of<br />

conventional hardware.<br />

A glance through the various publications of this period shows a preponderance of papers<br />

based on vocabulary studies generated initially by concordance programs. The results<br />

were of interest either for some kinds of stylistic analyses or for linguistic applications.<br />

Increasingly complex mathematics were brought <strong>to</strong> bear on vocabulary counts, leaving<br />

some more <strong>humanities</strong>-oriented conference participants out in the cold. Apart from these,<br />

there was little really new or exciting in terms of methodology and there was perhaps less<br />

critical appraisal of methodologies than might be desirable. The important developments<br />

during this period lay more in support systems generated by the presence of more outlets<br />

for dissemination (conferences and journals) and the recognition of the need for standard<br />

software and for archiving and maintaining texts. Dissemination was concentrated in<br />

outlets for <strong>humanities</strong> computing and much less in mainstream <strong>humanities</strong> publications. It<br />

seems that we were still at a stage where academic respectability for computer-based<br />

work in the <strong>humanities</strong> was questionable and scholars preferred <strong>to</strong> publish in outlets<br />

where they were more likely <strong>to</strong> be accepted.<br />

New Developments: Mid-1980s <strong>to</strong> Early 1990s<br />

This period saw some significant developments in <strong>humanities</strong> computing. Some of these<br />

can be attributed <strong>to</strong> two new technologies, the personal computer and electronic mail.<br />

Others happened simply because of the increase of usage and the need <strong>to</strong> reduce<br />

duplication of effort.<br />

At first there were several different and competing brands of personal computers. Some<br />

were developed for games, some were standalone word processors and could not be used<br />

for anything else, and others were specifically aimed at the educational market rather than<br />

for general use. Gradually IBM PCs and models based on the IBM architecture began <strong>to</strong><br />

dominate, with Apple Macin<strong>to</strong>shes also attracting plenty of use, especially for graphics.<br />

The personal computer is now a necessity of scholarly life, but in its early days it was<br />

considerably more expensive in relation <strong>to</strong> now and early purchasers were enthusiasts and<br />

those in the know about computing. The initial impact in <strong>humanities</strong> computing was that<br />

it was no longer necessary <strong>to</strong> register at the computer center in order <strong>to</strong> use a computer.<br />

Users of personal computers could do whatever they wanted and did not necessarily<br />

benefit from expertise that already existed. This encouraged duplication of effort, but it<br />

also fostered innovation where users were not conditioned by what was already available.<br />

By the end of the 1980s, there were three DOS-based text analysis programs: Word-<br />

Cruncher, TACT, and MicroOCP, all of which had very good functionality. Owners of


personal computers would work with these at home and, in the case of WordCruncher<br />

and TACT, obtain instantaneous results from searches. MicroOCP was developed from<br />

the mainframe program using a batch concordance technique rather than interactive<br />

searching. However, the main application of personal computers was that shared with all<br />

other disciplines, namely word processing. This attracted many more users who knew<br />

very little about other applications and tended <strong>to</strong> assume that the functions within word<br />

processing programs might be all that computers could do for them.<br />

The Apple Macin<strong>to</strong>sh was attractive for <strong>humanities</strong> users for two reasons. Firstly, it had a<br />

graphical user interface long before Windows on PCs. This meant that it was much better<br />

at displaying non-standard characters. At last it was possible <strong>to</strong> see Old English<br />

characters, Greek, Cyrillic, and almost any other alphabet, on the screen and <strong>to</strong><br />

manipulate text containing these characters easily. Secondly, the Macin<strong>to</strong>sh also came<br />

with a program that made it possible <strong>to</strong> build some primitive hypertexts easily.<br />

HyperCard provided a model of file cards with ways of linking between them. It also<br />

incorporated a simple programming <strong>to</strong>ol making it possible for the first time for<br />

<strong>humanities</strong> scholars <strong>to</strong> write computer programs easily. The benefits of hypertext for<br />

teaching were soon recognized and various examples soon appeared. A good example of<br />

these was the Beowulf Workstation created by Patrick Conner (Conner 1991). This<br />

presents a text <strong>to</strong> the user with links <strong>to</strong> a modern English version and linguistic and<br />

contextual annotations of various kinds. The first version of the Perseus Project was also<br />

delivered <strong>to</strong> the end user in HyperCard.<br />

Networking, at least for electronic mail, was previously confined <strong>to</strong> groups of computer<br />

scientists and research institutes. By the mid-1980s, facilities for sending and receiving<br />

electronic mail across international boundaries were provided by most academic<br />

computing services. At the 1985 ALLC conference in Nice, electronic mail addresses<br />

were exchanged avidly and a new era of immediate communication began. Soon e-mail<br />

was being sent <strong>to</strong> groups of users and the ListServ software for electronic discussion lists<br />

was established. Ansaxnet, the oldest electronic discussion list for the <strong>humanities</strong>, was<br />

founded by Patrick Conner in 1986 (Conner 1992).<br />

At the ICCH conference in Columbia, South Carolina, in spring 1987 a group of people<br />

mostly working in support roles in <strong>humanities</strong> computing got <strong>to</strong>gether and agreed that<br />

they needed <strong>to</strong> find a way of keeping in <strong>to</strong>uch on a regular basis. Willard McCarty, who<br />

was then at the University of Toron<strong>to</strong>, agreed <strong>to</strong> look in<strong>to</strong> how they might do this. On his<br />

return from the conference he discovered the existence of ListServ, and Humanist was<br />

born (McCarty 1992). The first message was sent out on May 7, 1987. McCarty launched<br />

himself in<strong>to</strong> the role of editing what he prefers <strong>to</strong> call an "electronic seminar" and, except<br />

for a hiatus in the early 1990s when Humanist was edited from Brown University, has<br />

continued in this role ever since.<br />

Humanist has become something of a model for electronic discussion lists. McCarty has<br />

maintained excellent standards of editing and the level of discussion is generally high.<br />

For those of us in Europe the regular early morning diet of three <strong>to</strong> six Humanist digests<br />

is a welcome start <strong>to</strong> the day. Humanist has become central <strong>to</strong> the maintenance and


development of a community and it has made a significant contribution <strong>to</strong> the definition<br />

of <strong>humanities</strong> computing. Its archives going back <strong>to</strong> 1987 are a vast source of information<br />

on developments and concerns during this period and it was taken as an exemplar by the<br />

founders of the Linguist List, the key electronic forum for linguistics.<br />

This period also saw the publication in print form of the only large-scale attempt <strong>to</strong><br />

produce a bibliography of projects, software, and publications. Two volumes of the<br />

Humanities Computing Yearbook (HCY) were published. The first, edited by Ian<br />

Lancashire and Willard McCarty appeared in 1988 with some 400 pages. The second<br />

volume, for 1989–90, has almost 700 pages with a much better index. For several years,<br />

until it began <strong>to</strong> get out of date, the HCY was an extremely valuable resource, fulfilling<br />

the role originally taken by the Computers and the Humanities Direc<strong>to</strong>ry of Scholars<br />

Active, which had ceased <strong>to</strong> appear by the early 1970s. Preparing the HCY was a truly<br />

enormous undertaking and no further volumes appeared. By the early 1990s, the general<br />

consensus was that in future an online database would be a more effective resource.<br />

Although there have been various attempts <strong>to</strong> start something similar, nothing on a<br />

serious scale has emerged, and the picture of overall activity in terms of projects and<br />

publications is once again incomplete.<br />

In terms of intellectual development, one activity stands out over all others during this<br />

period. In November 1987 Nancy Ide, assisted by colleagues in ACH, organized an<br />

invitational meeting at Vassar College, Poughkeepsie, <strong>to</strong> examine the possibility of<br />

creating a standard encoding scheme for <strong>humanities</strong> electronic texts (Burnard 1988).<br />

There had been various previous attempts <strong>to</strong> address the problem of many different and<br />

conflicting encoding schemes, a situation that was described as "chaos" by one of the<br />

participants at the Vassar meeting. Now, the time was ripe <strong>to</strong> proceed. Scholars were<br />

increasingly tired of wasting time reformatting texts <strong>to</strong> suit particular software and had<br />

become more frustrated with the inadequacies of existing schemes. In 1986, a new<br />

encoding method had appeared on the scene. The Standard Generalized Markup<br />

Language (SGML), published by ISO, offered a mechanism for defining a markup<br />

scheme that could handle many different types of text, could deal with metadata as well<br />

as data, and could represent complex scholarly interpretation as well as the basic<br />

structural features of documents.<br />

Participants at the meeting agreed on a set of principles ("the Poughkeepsie Principles")<br />

as a basis for building a new encoding scheme and entrusted the management of the<br />

project <strong>to</strong> a Steering Committee with representatives from ACH, ALLC, and the<br />

Association for Computational Linguistics (Text Encoding Initiative <strong>2001</strong>).<br />

Subsequently, this group raised over a million dollars in North America and oversaw the<br />

development of the Text Encoding Initiative (TEI) Guidelines for Electronic Text<br />

Encoding and Interchange. The work was initially organized in<strong>to</strong> four areas, each served<br />

by a committee. Output from the committees was put <strong>to</strong>gether by two edi<strong>to</strong>rs in<strong>to</strong> a first<br />

draft version, which was distributed for public comment in 1990. A further cycle of work<br />

involved a number of work groups that looked at specific application areas in detail. The<br />

first full version of the TEI Guidelines was published in May 1994 and distributed in<br />

print form and electronically.


The size, scope, and influence of the TEI far exceeded what anyone at the Vassar meeting<br />

envisaged. It was the first systematic attempt <strong>to</strong> categorize and define all the features<br />

within <strong>humanities</strong> texts that might interest scholars. In all, some 400 encoding tags were<br />

specified in a structure that was easily extensible for new application areas. The<br />

specification of the tags within the Guidelines illustrates some of the issues involved, but<br />

many deeper intellectual challenges emerged as the work progressed. Work in the TEI led<br />

<strong>to</strong> an interest in markup theory and the representation of <strong>humanities</strong> knowledge as a <strong>to</strong>pic<br />

in itself. The publication of the TEI Guidelines coincided with full-text <strong>digital</strong> library<br />

developments and it was natural for <strong>digital</strong> library projects, which had not previously<br />

come in<strong>to</strong> contact with <strong>humanities</strong> computing, <strong>to</strong> base their work on the TEI rather than<br />

inventing a markup scheme from scratch.<br />

Much of the TEI work was done by e-mail using private and public discussion lists,<br />

<strong>to</strong>gether with a fileserver where drafts of documents were posted. From the outset anyone<br />

who served on a TEI group was required <strong>to</strong> use e-mail regularly and the project became<br />

an interesting example of this method of working. However, participants soon realized<br />

that it is not easy <strong>to</strong> reach closure in an e-mail discussion and it was fortunate that<br />

funding was available for a regular series of face-<strong>to</strong>-face technical meetings <strong>to</strong> ensure that<br />

decisions were made and that the markup proposals from the different working groups<br />

were rationalized effectively.<br />

Apart from major developments in personal computing, networking, and the TEI, the<br />

kind of <strong>humanities</strong> computing activities which were ongoing in the 1970s continued <strong>to</strong><br />

develop, with more users and more projects. Gradually, certain application areas spun off<br />

from <strong>humanities</strong> computing and developed their own culture and dissemination routes.<br />

"Computers and writing" was one <strong>to</strong>pic that disappeared fairly rapidly. More important<br />

for <strong>humanities</strong> computing was the loss of some aspects of linguistic computing,<br />

particularly corpus linguistics, <strong>to</strong> conferences and meetings of its own. Computational<br />

linguistics had always developed independently of <strong>humanities</strong> computing and, despite the<br />

efforts of Don Walker on the TEI Steering Committee, continued <strong>to</strong> be a separate<br />

discipline. Walker and An<strong>to</strong>nio Zampolli of the Institute for Computational Linguistics in<br />

Pisa worked hard <strong>to</strong> bring the two communities of <strong>humanities</strong> computing and<br />

computational linguistics <strong>to</strong>gether but with perhaps only limited success. Just at the time<br />

when <strong>humanities</strong> computing scholars were beginning seriously <strong>to</strong> need the kinds of <strong>to</strong>ols<br />

developed in computational linguistics (morphological analysis, syntactic analysis, and<br />

lexical databases), there was an expansion of work in computational and corpus<br />

linguistics <strong>to</strong> meet the needs of the defense and speech analysis community. In spite of a<br />

landmark paper on the convergence between computational linguistics and literary and<br />

linguistic computing given by Zampolli and his colleague Nicoletta Calzolari at the first<br />

joint ACH/ALLC conference in Toron<strong>to</strong> in June 1989 (Calzolari and Zampolli 1991),<br />

there was little communication between these communities, and <strong>humanities</strong> computing<br />

did not benefit as it could have done from computational linguistics techniques.<br />

The Era of the Internet: Early 1990s <strong>to</strong> the Present


One development far outstripped the impact of any other during the 1990s. This was the<br />

arrival of the Internet, but more especially the World Wide Web. The first graphical<br />

browser, Mosaic, appeared on the scene in 1993. Now the use of the Internet is a vital<br />

part of any academic activity. A generation of students has grown up with it and naturally<br />

looks <strong>to</strong> it as the first source of any information.<br />

Initially, some long-term <strong>humanities</strong> computing practitioners had problems in grasping<br />

the likely impact of the Web in much the same way as Microsoft did. Those involved<br />

with the TEI felt very much that HyperText Markup Language (HTML) was a weak<br />

markup system that perpetuated all the problems with word processors and appearancebased<br />

markup. The Web was viewed with curiosity but this tended <strong>to</strong> be rather from the<br />

outside. It was a means of finding some kinds of information but not really as a serious<br />

<strong>to</strong>ol for <strong>humanities</strong> research. This presented an opportunity for those institutions and<br />

organizations that were contemplating getting in<strong>to</strong> <strong>humanities</strong> computing for the first<br />

time. They saw that the Web was a superb means of publication, not only for the results<br />

of their scholarly work, but also for promoting their activities among a much larger<br />

community of users. A new group of users had emerged.<br />

Anyone can be a publisher on the Web and within a rather short time the focus of a<br />

broader base of interest in <strong>humanities</strong> computing became the delivery of scholarly<br />

material over the Internet. The advantages of this are enormous from the producer's point<br />

of view. The format is no longer constrained by that of a printed book. Theoretically<br />

there is almost no limit on size, and hypertext links provide a useful way of dealing with<br />

annotations, etc. The publication can be built up incrementally as and when bits of it are<br />

ready for publication. It can be made available <strong>to</strong> its audience immediately and it can<br />

easily be amended and updated.<br />

In the early <strong>to</strong> mid-1990s, many new projects were announced, some of which actually<br />

succeeded in raising money and getting started. Particularly in the area of electronic<br />

scholarly editions, there were several meetings and publications devoted <strong>to</strong> discussion<br />

about what an electronic edition might look like (Finneran 1996; Bornstein and Tinkle<br />

1998). This was just at the time when edi<strong>to</strong>rial theorists were focusing on the text as a<br />

physical object, which they could represent by <strong>digital</strong> images. With the notable exception<br />

of work carried out by Peter Robinson (Robinson 1996, 1997, 1999) and possibly one or<br />

two others, few of these publications saw the light of day except as pro<strong>to</strong>types or small<br />

samples, and by the second half of the decade interest in this had waned somewhat. A<br />

good many imaginative ideas had been put forward, but once these reached the stage<br />

where theory had <strong>to</strong> be put in<strong>to</strong> practice and projects were faced with the laborious work<br />

of entering and marking up text and developing software, attention began <strong>to</strong> turn<br />

elsewhere.<br />

Debates were held on what <strong>to</strong> call these collections of electronic resources. The term<br />

"archive" was favored by many, notably the Blake Archive and other projects based in<br />

the Institute for Advanced Technology in the Humanities at the University of Virginia.<br />

"Archive" meant a collection of material where the user would normally have <strong>to</strong> choose a<br />

navigation route. "Edition" implies a good deal of scholarly added value, reflecting the


views of one or more edi<strong>to</strong>rs, which could be implemented by privileging specific<br />

navigation routes. SGML (Standard Generalized Markup Language), mostly in<br />

applications based on the TEI, was accepted as a way of providing the hooks on which<br />

navigation routes could be built, but significant challenges remained in designing and<br />

building an effective user interface. The emphasis was, however, very much on<br />

navigation rather than on the analysis <strong>to</strong>ols and techniques that had formed the major<br />

application areas within <strong>humanities</strong> computing in the past. In the early days of the Web,<br />

the technology for delivery of SGML-encoded texts was clunky and in many ways<br />

presented a less satisfying user interface than what can be delivered with raw HTML.<br />

Nevertheless, because of the easy way of viewing them, the impact of many of these<br />

publishing projects was substantial. Many more people became familiar with the idea of<br />

technology in the <strong>humanities</strong>, but in a more limited sense of putting material on<strong>to</strong> the<br />

Web.<br />

Although at first most of these publishing projects had been started by groups of<br />

academics, it was not long before libraries began <strong>to</strong> consider putting the content of their<br />

collections on the Internet. Several institutions in the United States set up electronic text<br />

or <strong>digital</strong> library collections for <strong>humanities</strong> primary source material, most usually using<br />

the OpenText SGML search engine (Price-Wilkin 1994). While this provides good and<br />

fast facilities for searching for words (strings), it really provides little more than a<br />

reference <strong>to</strong>ol <strong>to</strong> look up words. Other projects used the DynaText SGML electronic book<br />

system for the delivery of their material. This offered a more structured search but with<br />

an interface that is not particularly intuitive.<br />

A completely new idea for an electronic publication was developed by the Orlando<br />

Project, which is creating a His<strong>to</strong>ry of British Women's Writing at the Universities of<br />

Alberta and Guelph. With substantial research funding, new material in the form of short<br />

biographies of authors, his<strong>to</strong>ries of their writing, and general world events was created as<br />

a set of SGML documents (Brown et al. 1997). It was then possible <strong>to</strong> consider extracting<br />

portions of these documents and reconstituting them in<strong>to</strong> new material, for example <strong>to</strong><br />

generate chronologies for specific periods or <strong>to</strong>pics. This project introduced the idea of a<br />

completely new form of scholarly writing and one that is fundamentally different from<br />

anything that has been done in the past. It remains <strong>to</strong> be seen whether it will really be<br />

usable on a large scale.<br />

The Internet also made it possible <strong>to</strong> carry out collaborative projects in a way that was<br />

never possible before. The simple ability for people in different places <strong>to</strong> contribute <strong>to</strong> the<br />

same document collections was a great advance on earlier methods of working. In the<br />

Orlando Project, researchers at both institutions add <strong>to</strong> a document archive developed as a<br />

web-based document management system, which makes use of some of the SGML<br />

markup for administrative purposes. Ideas have also been floated about collaborative<br />

editing of manuscript sources where people in different locations could add layers of<br />

annotation, for example for the Peirce Project (Neuman et al. 1992) and the Codex<br />

Leningradensis (Leningrad Codex Markup Project 2000). The technical aspects of this<br />

are fairly clear. Perhaps less clear is the management of the project, who controls or vets<br />

the annotations, and how it might all be maintained for the future.


The TEI's adoption as a model in <strong>digital</strong> library projects raised some interesting issues<br />

about the whole philosophy of the TEI, which had been designed mostly by scholars who<br />

wanted <strong>to</strong> be as flexible as possible. Any TEI tag can be redefined and tags can be added<br />

where appropriate. A rather different philosophy prevails in library and information<br />

science where standards are defined and then followed closely – this <strong>to</strong> ensure that<br />

readers can find books easily. It was a pity that there was not more input from library and<br />

information science at the time that the TEI was being created, but the TEI project was<br />

started long before the term "<strong>digital</strong> library" came in<strong>to</strong> use. A few people made good<br />

contributions, but in the library community there was not the widespread range of many<br />

years' experience of working with electronic texts as in the scholarly community. The<br />

TEI was, however, used as a model by the developers of the Encoded Archival<br />

Description (EAD), which has had a very wide impact as a standard for finding aids in<br />

archives and special collections.<br />

An additional dimension was added <strong>to</strong> <strong>humanities</strong> electronic resources in the early 1990s,<br />

when it became possible <strong>to</strong> provide multimedia information in the form of images, audio,<br />

and video. In the early days of <strong>digital</strong> imaging there was much discussion about file<br />

formats, pixel depth, and other technical aspects of the imaging process and much less<br />

about what people can actually do with these images other than view them. There are of<br />

course many advantages in having access <strong>to</strong> images of source material over the Web, but<br />

<strong>humanities</strong> computing practitioners, having grown used <strong>to</strong> the flexibility offered by<br />

searchable text, again tended <strong>to</strong> regard imaging projects as not really their thing, unless,<br />

like the Beowulf Project (Kiernan 1991), the images could be manipulated and enhanced<br />

in some way. Interesting research has been carried out on linking images <strong>to</strong> text, down <strong>to</strong><br />

the level of the word (Zweig 1998). When most of this can be done au<strong>to</strong>matically we will<br />

be in a position <strong>to</strong> reconceptualize some aspects of manuscript studies. The potential of<br />

other forms of multimedia is now well recognized, but the use of this is only really<br />

feasible with high-speed access and the future may well lie in a gradual convergence with<br />

television.<br />

The expansion of access <strong>to</strong> electronic resources fostered by the Web led <strong>to</strong> other areas of<br />

theoretical interest in <strong>humanities</strong> computing. Electronic resources became objects of<br />

study in themselves and were subjected <strong>to</strong> analysis by a new group of scholars, some of<br />

whom had little experience of the technical aspects of the resources. Hypertext in<br />

particular attracted a good many theorists. This helped <strong>to</strong> broaden the range of interest in,<br />

and discussion about, <strong>humanities</strong> computing but it also perhaps contributed <strong>to</strong><br />

misapprehensions about what is actually involved in building and using such a resource.<br />

Problems with the two cultures emerged again, with one that was actually doing it and<br />

another that preferred talking about doing it.<br />

The introduction of academic programs is another indication of the acceptance of a<br />

subject area by the larger academic community. For <strong>humanities</strong> computing this began <strong>to</strong><br />

happen by the later 1990s although it is perhaps interesting <strong>to</strong> note that very few of these<br />

include the words "Humanities Computing" in the program title. King's College London<br />

offers a BA Minor in Applied Computing with a number of <strong>humanities</strong> disciplines, and<br />

its new MA, based in the Centre for Humanities Computing, is also called MA in Applied


Computing. McMaster University in Canada offers a BA in Multimedia. The MA that the<br />

University of Virginia is soon <strong>to</strong> start is called Digital Humanities and is under the<br />

auspices of the Media Studies Program. The University of Alberta is, as far as I am<br />

aware, the first <strong>to</strong> start a program with Humanities Computing in its title, although the<br />

University of Glasgow has had an MPhil in His<strong>to</strong>ry and Computing for many years.<br />

As the Internet fostered the more widespread use of computers for <strong>humanities</strong><br />

applications, other organizations began <strong>to</strong> get involved. This led <strong>to</strong> some further attempts<br />

<strong>to</strong> define the field or at least <strong>to</strong> define a research agenda for it. The then Getty Art His<strong>to</strong>ry<br />

Information Program published what is in my view a very interesting Research Agenda<br />

for Networked Cultural Heritage in 1996 (Bearman 1996). It contains eight papers<br />

tackling specific areas that cover <strong>to</strong>pics which really bridge across <strong>digital</strong> libraries, and<br />

<strong>humanities</strong> research and teaching. Each of these areas could form a research program in<br />

its own right, but the initiative was not taken further. Meanwhile the ALLC and ACH<br />

continued <strong>to</strong> organize a conference every year with a predominance of papers on markup<br />

and other technical issues. An attempt <strong>to</strong> produce a roadmap and new directions for<br />

<strong>humanities</strong> computing for the 2002 conference in Germany produced a useful survey<br />

(Robey 2002), but little new, and would perhaps have benefited from more input from a<br />

broader community. But how <strong>to</strong> involve other communities was becoming more of a<br />

problem in an era when many more electronic resources for the <strong>humanities</strong> were being<br />

developed outside the <strong>humanities</strong> computing community.<br />

Conclusion<br />

If one <strong>humanities</strong> computing activity is <strong>to</strong> be highlighted above all others, in my view it<br />

must be the TEI. It represents the most significant intellectual advances that have been<br />

made in our area, and has influenced the markup community as a whole. The TEI<br />

attracted the attention of leading practitioners in the SGML community at the time when<br />

XML (Extensible Markup Language) was being developed and Michael Sperberg-<br />

McQueen, one of the TEI edi<strong>to</strong>rs, was invited <strong>to</strong> be co-edi<strong>to</strong>r of the new XML markup<br />

standard. The work done on hyperlinking within the TEI formed the basis of the linking<br />

mechanisms within XML. In many ways the TEI was ahead of its time, as only with the<br />

rapid adoption of XML in the last two <strong>to</strong> three years has the need for descriptive markup<br />

been recognized by a wider community. Meanwhile, the community of markup theorists<br />

that has developed from the TEI continues <strong>to</strong> ask challenging questions on the<br />

representation of knowledge.<br />

There are still other areas <strong>to</strong> be researched in depth. Humanities computing can contribute<br />

substantially <strong>to</strong> the growing interest in putting the cultural heritage on the Internet, not<br />

only for academic users, but also for lifelong learners and the general public. Tools and<br />

techniques developed in <strong>humanities</strong> computing will facilitate the study of this material<br />

and, as the Perseus Project is showing (Rydberg-Cox 2000), the incorporation of<br />

computational linguistics techniques can add a new dimension. Our <strong>to</strong>ols and techniques<br />

can also assist research in facilitating the digitization and encoding processes, where we<br />

need <strong>to</strong> find ways of reducing the costs of data creation without loss of scholarly value or<br />

of functionality. Through the Internet, <strong>humanities</strong> computing is reaching a much wider


audience, and students graduating from the new programs being offered will be in a<br />

position <strong>to</strong> work not only in academia, but also in electronic publishing, educational<br />

technologies, and multimedia development. Throughout its his<strong>to</strong>ry, <strong>humanities</strong><br />

computing has shown a healthy appetite for imagination and innovation while continuing<br />

<strong>to</strong> maintain high scholarly standards. Now that the Internet is such a dominant feature of<br />

everyday life, the opportunity exists for <strong>humanities</strong> computing <strong>to</strong> reach out much further<br />

than has hither<strong>to</strong> been possible.<br />

References for Further Reading<br />

Bearman, D., (ed.) (1996). Research Agenda for Networked Cultural Heritage. Santa<br />

Monica, CA: Getty Art His<strong>to</strong>ry Information Program.<br />

Berry-Rogghe, G. L. M. and T. D. Crawford (1973). Developing a Machine-independent<br />

Concordance Program for a Variety of Languages. In A. J. Aitken, R. W. Bailey, and N.<br />

Hamil<strong>to</strong>n-Smith (eds.), The Computer and Literary Studies (pp. 309–16). Edinburgh:<br />

Edinburgh University Press.<br />

Bessinger, J. B. and S. M. Parrish (1965). Literary Data Processing Conference<br />

Proceedings. White Plains, NY: IBM.<br />

Bornstein, G. and T. Tinkle (1998). The Iconic Page in Manuscript, Print, and Digital<br />

Culture. Ann Arbor: University of Michigan Press.<br />

Brown, S., S. Fisher, P. Clements, K. Binhammer, T. Butler, K. Carter, I. Grundy, and S.<br />

Hockey (1997). SGML and the Orlando Project: Descriptive Markup for an Electronic<br />

His<strong>to</strong>ry of Women's Writing. Computers and the Humanities 31: 271–84.<br />

Brunner, T. F. (1993). Classics and the Computer: The His<strong>to</strong>ry of a Relationship. In J.<br />

Solomon (ed.), Accessing Antiquity: The Computerization of Classical Studies (pp. 10–<br />

33). Tucson: University of Arizona Press.<br />

Burnard, L. (1987a). CAFS: A New Solution <strong>to</strong> an Old Problem. Literary and Linguistic<br />

Computing 2: 7–12.<br />

Burnard, L. (1987b). Principles of Database Design. In S. Rahtz (ed.), Information<br />

Technology in the Humanities (pp. 54–68). Chichester: Ellis Horwood.<br />

Burnard, L. (1988). Report of Workshop on Text Encoding Guidelines. Literary and<br />

Linguistic Computing 3: 131–3.<br />

Bur<strong>to</strong>n, D. M. (1981a). Au<strong>to</strong>mated Concordances and Word Indexes: The Fifties.<br />

Computers and the Humanities 15: 1–14.<br />

Bur<strong>to</strong>n, D. M. (1981b). Au<strong>to</strong>mated Concordances and Word Indexes: The Early Sixties<br />

and the Early Centers. Computers and the Humanities 15: 83–100.


Bur<strong>to</strong>n, D. M. (1981c). Au<strong>to</strong>mated Concordances and Word Indexes: The Process, the<br />

Programs, and the Products. Computers and the Humanities 15: 139–54.<br />

Bur<strong>to</strong>n, D. M. (1982). Au<strong>to</strong>mated Concordances and Word Indexes: Machine Decisions<br />

and Edi<strong>to</strong>rial Revisions. Computers and the Humanities 16: 195–218.<br />

Busa, R. (1974-). Index Thomisticus. Stuttgart: Frommann-Holzboog.<br />

Busa, R. (1980). The Annals of Humanities Computing: The Index Thomisticus.<br />

Computers and the Humanities 14: 83–90.<br />

Busa, R., (ed.) (1992). Thomae Aquinatis Opera Omnia Cum Hypertextibus in CD-ROM.<br />

Milano: Edi<strong>to</strong>ria Elettronica Editel.<br />

Busa, R. (1999). Picture a Man.… Busa Award Lecture, Debrecen, Hungary, July 6,<br />

1998. Literary and Linguistic Computing 14: 5–9.<br />

Calzolari, N. and A. Zampolli (1991). Lexical Databases and Textual Corpora: A Trend<br />

of Convergence between Computational Linguistics and Literary and Linguistic<br />

Computing. In S. Hockey, N. Ide, and I. Lancashire (eds.), Research in Humanities<br />

Computing 1: Selected Papers from the ALLC/ACH Conference, Toron<strong>to</strong>, June 1989 (pp.<br />

272–307). Oxford: Clarendon Press.<br />

Conner, P. W (1991). The Beowulf Workstation: One Model of Computer-assisted<br />

Literary Pedagogy. Literary and Linguistic Computing 6: 50–8.<br />

Conner, P. W (1992). Networking in the Humanities: Lessons from Ansaxnet. Computers<br />

and the Humanities 26: 195–204.<br />

De Tollenaere, F. (1973). The Problem of the Context in Computer-aided Lexicography.<br />

In A. J. Aitken, R. W. Bailey, and N. Hamil<strong>to</strong>n-Smith (eds.), The Computer and Literary<br />

Studies (pp. 25–35). Edinburgh: Edinburgh University Press.<br />

Ellegård, A. (1962). A Statistical Method for Determining Authorship: The Junius Letters<br />

1769–1772. Gothenburg: Gothenburg Studies in English.<br />

Finneran, R. J. (1996). The Literary Text in the Digital Age. Ann Arbor: University of<br />

Michigan Press.<br />

Gorcy, G. (1983). L'informatique et la mise en oeuvre du trésor de la langue française<br />

(TLF), dictionnaire de la langue du 19 e et du 20 e siècle (1789–1960). In A. Cappelli and<br />

A. Zampolli (eds.), The Possibilities and Limits of the Computer in Producing and<br />

Publishing Dictionaries: Proceedings of the European Science Foundation Workshop,<br />

Pisa 1981. Linguistica Computazionale III (pp. 119–44). Pisa: Giardini.


Hamil<strong>to</strong>n-Smith, N. (1971). A Versatile Concordance Program for a Textual Archive. In<br />

R. A. Wisbey (ed.), The Computer in Literary and Linguistic Research (pp. 235–44).<br />

Cambridge: Cambridge University Press.<br />

Healey, A. (1989). The Corpus of the Dictionary of Old English: Its Delimitation,<br />

Compilation and Application. Paper presented at the Fifth Annual Conference of the UW<br />

Centre for the New Oxford English Dictionary. Oxford, September, 1989.<br />

Hockey, S. (1986). Workshop on Teaching Computers and the Humanities Courses.<br />

Literary and Linguistic Computing 1: 228–9.<br />

Hockey, S. and I. Marriott (1979a). The Oxford Concordance Project (OCP) – Part 1.<br />

ALLC Bulletin 7: 35–43.<br />

Hockey, S. and I. Marriott (1979b). The Oxford Concordance Project (OCP) – Part 2.<br />

ALLC Bulletin 7: 155–64.<br />

Hockey, S. and I. Marriott (1979c). The Oxford Concordance Project (OCP) – Part 3.<br />

ALLC Bulletin 7: 268–75.<br />

Hockey, S. and I. Marriott (1980). The Oxford Concordance Project (OCP) – Part 4.<br />

ALLC Bulletin 8: 28–35.<br />

Holmes, D. I. and R. S. Forsyth (1995). The Federalist Revisited: New Directions in<br />

Authorship Attribution. Literary and Linguistic Computing 10: 111–27.<br />

Kiernan, K. S. (1991). Digital Image Processing and the Beowulf Manuscript. Literary<br />

and Linguistic Computing 6: 20–7.<br />

Lancashire, I., (ed.) (1991). The Humanities Computing Yearbook 1989–90: A<br />

Comprehensive Guide <strong>to</strong> Software and Other Resources. Oxford: Clarendon Press.<br />

Lancashire, I. and W. McCarty, (eds.) (1988). The Humanities Computing Yearbook<br />

1988. Oxford: Clarendon Press.<br />

Leningrad Codex Markup Project (2000). Project "EL": The XML Leningrad Codex.<br />

Available at: , accessed May 15, 2003.<br />

Lord, R. D. (1958). Studies in the His<strong>to</strong>ry of Probability and Statistics: viii. de Morgan<br />

and the Statistical Study of Literary Style. Biometrika 45: 282.<br />

McCarty, W. (1992). Humanist: Lessons from a Global Electronic Seminar. Computers<br />

and the Humanities 26: 205–22.<br />

Mendenhall, T. C. (1901). A Mechanical Solution of a Literary Problem. The Popular<br />

Science Monthly 60: 97–105.


Mor<strong>to</strong>n, A. Q. (1965). The Authorship of the Pauline Epistles: A Scientific Solution.<br />

Saska<strong>to</strong>on: University of Saskatchewan.<br />

Mor<strong>to</strong>n, A. Q. and Winspear, A. D. (1971). It's Greek <strong>to</strong> the Computer. Montreal:<br />

Harvest House.<br />

Mosteller, F. and D. L. Wallace (1964). Inference and Disputed Authorship: The<br />

Federalist. Reading, MA: Addison-Wesley.<br />

Neuman, M., M. Keeler, C. Kloesel, J. Ransdell, and A. Renear (1992). The Pilot Project<br />

of the Electronic Peirce Consortium (abstract). ALLC-ACH92 Conference Abstracts and<br />

Program (pp. 25–7). Oxford.<br />

Parrish, S. M. (1962). Problems in the Making of Computer Concordances. Studies in<br />

Bibliography 15: 1–14.<br />

Price-Wilkin, J. (1994). Using the World Wide Web <strong>to</strong> Deliver Complex Electronic<br />

Documents: Implications for Libraries. The Public-Access Computer Systems Review 5:<br />

5–21. , accessed July 21, 2004.<br />

Proud, J. K. (1989). The Oxford Text Archive. London: British Library Research and<br />

Development Report.<br />

Robey, D. (2002). New Directions in Humanities Computing, ,<br />

accessed May 15, 2003.<br />

Robinson, P., (ed.) (1996). Geoffrey Chaucer: The Wife of Bath's Prologue on CD-ROM.<br />

Cambridge: Cambridge University Press.<br />

Robinson, P. M. W. (1997). New Directions in Critical Editing. In K. Sutherland (ed.),<br />

Electronic Text: Investigations in Method and Theory (pp. 145–71). Oxford: Clarendon<br />

Press.<br />

Robinson, P. M. W. (1999). New Methods of Editing, Exploring and Reading The<br />

Canterbury Tales. http://www.cta.dmu.ac.uk/projects/ctp/desc2.html, accessed May 14,<br />

2003.<br />

Russell, D. B. (1967). COCOA - A Word Count and Concordance Genera<strong>to</strong>r for Atlas.<br />

Chil<strong>to</strong>n: Atlas Computer Labora<strong>to</strong>ry.<br />

Rydberg-Cox, J. A. (2000). Co-occurrence Patterns and Lexical Acquisition in Ancient<br />

Greek Texts. Literary and Linguistic Computing 15: 121–30.<br />

Text Encoding Initiative (<strong>2001</strong>). Text Encoding Initiative, http://www.tei-c.org, accessed<br />

May 15, 2003.


Tweedie, F. J., S. Singh, and D. I. Holmes (1996). Neural Network Applications in<br />

Stylometry: The Federalist Papers. Computers and the Humanities 30: 1–10.<br />

Wisbey, R. (1963). The Analysis of Middle High German Texts by Computer: Some<br />

Lexicographical Aspects. Transactions of the Philological Society, 28–48.<br />

Wisbey, R. A., (ed.) (1971). The Computer in Literary and Linguistic Research.<br />

Cambridge: Cambridge University Press.<br />

Zweig, R. W. (1998). Lessons from the Palestine Post Project. Literary and Linguistic<br />

Computing 13: 89–97.<br />

2.<br />

Computing for Archaeologists<br />

Harrison Eiteljorg, II<br />

Archaeology 1 is the study of his<strong>to</strong>ry as written not with ink on paper but with the debris<br />

of human activity found where it fell. Whereas his<strong>to</strong>rians read the written records of our<br />

ances<strong>to</strong>rs, archaeologists read the material record, either <strong>to</strong> augment the his<strong>to</strong>ric one or <strong>to</strong><br />

reconstruct prehis<strong>to</strong>ry when there is no written record. As with the his<strong>to</strong>ric record, the<br />

archaeological record is often prejudiced by accidental survival, biased sources, and<br />

skewed representations of certain materials. Of course, both his<strong>to</strong>ric and material records<br />

from the past may carry inadvertent meaning. When Shelley's Ozymandias cried out,<br />

"My name is Ozymandias, king of kings, / Look on my works, ye Mighty, and despair!"<br />

he, as have many figures from the past, sent an ironic and unintended message.<br />

The archaeologist's work is, first, <strong>to</strong> find the material remains of our ances<strong>to</strong>rs, second, <strong>to</strong><br />

unearth those remains in ways that maximize the information they can convey, and,<br />

finally, <strong>to</strong> interpret the evidence. Finding those remains may be done by excavation or<br />

surface survey, but both processes require the destruction of the very evidence that is the<br />

first fruit of the work. In some cases, the destruction is literal and complete, as when an<br />

archaeologist must dig through one level of a site <strong>to</strong> find another; often the destruction is<br />

not so complete but involves the removal of objects from their physical contexts. Since<br />

contexts provide crucial clues <strong>to</strong> both physical and temporal relationships, evidence is<br />

lost even by removing objects. The destructive nature of the work demands<br />

extraordinarily careful record keeping <strong>to</strong> avoid accidental information loss. Indeed, it can<br />

be argued that record keeping is the real occupation of the field archaeologist.<br />

The interpretative process requires careful examination of the records of excavation or<br />

survey; the contexts of the objects are as important <strong>to</strong> a full understanding as the objects<br />

themselves. For instance, a figurine found in the remains of a <strong>to</strong>mb later covered by a<br />

house must have been deposited before the construction of the house and before the<br />

deposit of anything in that house. On the other hand, a potsherd found in the bot<strong>to</strong>m of a


trash pit must have been deposited at roughly the time the pit was dug, which may have<br />

been long after the deposit of flint <strong>to</strong>ols located near the potsherd but not in the trash pit.<br />

The importance of context <strong>to</strong> the archaeologist highlights the importance of good records.<br />

Records of both the archaeological contexts and the artifacts, not <strong>to</strong> mention a great many<br />

other aspects of an excavation or survey, provide the keys <strong>to</strong> analyses. Those records are<br />

truly crucial, and the potential utility of the computer for that record keeping is obvious<br />

<strong>to</strong>day. Less obvious is the fact that, if computers are used <strong>to</strong> record the basic information<br />

on which archaeologists build their understanding of the past, all practitioners must<br />

ultimately be able <strong>to</strong> use computers <strong>to</strong> retrieve that information. That is, computers and<br />

computer skills will be needed by all archaeologists if a substantial portion of the<br />

archaeological record is maintained in computer form.<br />

It is now obvious that computers are ideal for record keeping, but when archaeologists<br />

first started using computers, in the late 1950s, computers were arcane and foreign. Their<br />

record-keeping potential could not then be utilized by academics because of costs and<br />

access limits; affordable microcomputers lay well in the future. Data entry required<br />

punchcards or tape, and results were only available on paper. As a natural result,<br />

computers were used for tasks that required considerable processing power, not routine<br />

data s<strong>to</strong>rage. Early uses therefore tended <strong>to</strong> be for statistical processing, and, though<br />

statistics had been used in archaeology for decades, more statistical procedures and more<br />

mathematically complex statistical procedures could be performed with computers.<br />

During these early years of computer usage in archaeology, archaeologists had <strong>to</strong> learn<br />

computer languages in order <strong>to</strong> prepare data for processing and then <strong>to</strong> carry out the<br />

statistical processes. That made the use of computers less likely <strong>to</strong> penetrate deeply in<strong>to</strong><br />

the field. Nevertheless, some saw even in the late 1960s the enormous potential for<br />

computers <strong>to</strong> s<strong>to</strong>re large quantities of information for retrieval from what were then<br />

called databanks.<br />

By the early 1970s, there had already been some conferences on archaeological<br />

computing, and the microcomputer revolution began later in that decade, though the first<br />

microcomputers did not have the advantage of the IBM name. One of the most important<br />

things the early microcomputers did have was the database management system, dBase.<br />

That program and its many offspring have had an enormous impact on archaeology<br />

because of their record-keeping potential. Database management programs make it<br />

possible <strong>to</strong> manage large datasets without first learning <strong>to</strong> write long and complex<br />

computer programs (though the need <strong>to</strong> write routines for specific operations remains).<br />

The database software of the mid-1970s and after brought the promise of efficient record<br />

keeping, and the new capabilities were desperately needed as excava<strong>to</strong>rs were then<br />

expanding the quantity of material collected. For instance, studies of plant and animal<br />

remains in the archaeological record (<strong>to</strong> understand food sources and the surrounding<br />

ecosystem) required sifting through large quantities of earth <strong>to</strong> find seeds and bones that<br />

could only be interpreted with statistical analyses; such work cried out for sophisticated<br />

data-handling techniques. Similarly, more careful and conscientious attention <strong>to</strong> small


finds and fragmentary evidence could only become common with the advent of better<br />

recording techniques. It goes without saying that the recording of all these data would<br />

have been of little use had the programs not also made the retrieval of information – in an<br />

incredibly wide variety of forms – more efficient and flexible.<br />

Microcomputers – with relatively easy-<strong>to</strong>-use software – arrived just as those needs for<br />

more sophisticated data handling were being felt. As a result, database technology<br />

seemed <strong>to</strong> many a godsend. That is not <strong>to</strong> say that all archaeologists immediately began<br />

using computers and databases, or even that all have adopted them <strong>to</strong>day, but the need for<br />

assistance with ever-more-voluminous records and the availability of sophisticated<br />

computer programs fed on one another. Archaeologists who were eager <strong>to</strong> deal with<br />

seeds, shells, bones, and small, fragmentary finds could do so, and the fact that they could<br />

record so much information with the aid of computers encouraged them <strong>to</strong> treat that level<br />

of record keeping as standard. That pushed others <strong>to</strong> attend <strong>to</strong> similar levels of detail and,<br />

of necessity, <strong>to</strong> turn <strong>to</strong> computers <strong>to</strong> help. The process is continuing, with new<br />

technologies being regularly added <strong>to</strong> the <strong>to</strong>ol kit.<br />

The acceleration of computer use is ongoing, and the extent <strong>to</strong> which the quantity of<br />

recorded data may overwhelm the scholar's ability <strong>to</strong> synthesize is a matter of some<br />

debate. Until archaeologists can dispassionately evaluate the utility of different<br />

approaches <strong>to</strong> data collection, there is a natural tendency <strong>to</strong> record more, whether or not<br />

all the information is useful.<br />

The growth of computer use spawned the organization called Computer Applications and<br />

Quantitative Methods in Archaeology (CAA). It began with a small meeting at the<br />

University of Birmingham (England) in 1973 and has grown <strong>to</strong> an international<br />

organization with annual meetings in various cities in Europe. In December 1984 the<br />

Archaeological Computing Newsletter was launched <strong>to</strong> report on information about<br />

archaeological computing, and by the mid-1980s established professional archaeology<br />

organizations featured regular sessions about one or another aspect of computing at their<br />

annual meetings.<br />

The early and lasting interest in databases stemmed not only from the need <strong>to</strong> control<br />

huge quantities of excavation data but also from the hope that data s<strong>to</strong>rehouses could be<br />

used by scholars <strong>to</strong> retrieve and analyze information from related excavations, thus<br />

permitting broader syntheses. Indeed, many archaeologists still hope for such<br />

aggregations of data. While that giant data warehouse has been seen by many as the pot<br />

at the end of the rainbow, those most intimately familiar with the technology have, from<br />

the beginning, seen aggregated databases as a far more distant goal at best. Even<br />

archaeologists working in the same cultural and geographic areas do not – and cannot –<br />

excavate, survey, record their results, or use terms in precisely the same ways. As a<br />

result, combining data from multiple projects remains an illusive goal. Efforts <strong>to</strong> impose<br />

standardization have met with little success, and even such a carefully crafted and<br />

unthreatening potential aid as the Getty Art and Architecture Thesaurus has not been<br />

noticeably helpful in bringing the scholars in the field in<strong>to</strong> terminological uniformity for<br />

a single language, much less across languages.


The difficulties with common terms and data structures have been exacerbated by the<br />

divergence between the needs of the archaeological community and those of museum<br />

professionals. Archaeologists and the museum cura<strong>to</strong>rs who ultimately receive the<br />

excavated artifacts record their information about those artifacts in strikingly different<br />

arrangements, the one beginning with excavation context and the other with either a<br />

cultural classification or an object-based classification system. As a result, the databases<br />

of the two groups are organized differently, making large-scale cooperation problematic.<br />

Large-scale constructed datasets such as the collection of information about<br />

archaeological sites in Turkey (The Archaeological Settlements of Turkey website online<br />

at ) are less ambitious than the enormous, au<strong>to</strong>matic aggregations<br />

once anticipated, but they are now becoming more common and often prove <strong>to</strong> be<br />

remarkably useful. Another good example of such constructed databases is the National<br />

Archeological Database of US public archaeological sites (online at<br />

) developed by the National Park Service and<br />

the Center for Advanced Spatial Technologies at the University of Arkansas (CAST,<br />

). Websites that serve as gateways <strong>to</strong><br />

information, for instance, ARGE, the Archaeological Guide <strong>to</strong> Europe<br />

(), have also proved <strong>to</strong> be very valuable, and a recent project,<br />

called OASIS (), which makes possible timely<br />

access <strong>to</strong> information about archaeological projects throughout Britain, is an excellent<br />

example of real benefits for scholars by providing access <strong>to</strong> disparate data through a<br />

common central point but without imposing unrealistic new standards.<br />

Combining disparate datasets may not be a realistic near-term goal, but preserving<br />

datasets for future access is a necessity now. There are active <strong>digital</strong> reposi<strong>to</strong>ries now<br />

available <strong>to</strong> archaeologists for the long-term preservation of their data files, and<br />

preservation of <strong>digital</strong> data is a major responsibility. The recovered artifacts and the data<br />

about them are the only surviving evidence of fieldwork. Neither is fully meaningful<br />

without the other. It may also be argued that economic development will result in fewer<br />

possibilities for excavation and field survey, making excavations in museum basements<br />

and old records the archaeology of the future.<br />

There is a serious problem with the expansion of data quantity and the consequently<br />

increasing use of databases. When archaeologists use computers <strong>to</strong> view data rather than<br />

spending time with the objects themselves, they risk losing the familiarity with the<br />

objects that can only come from sustained, intimate, physical contact.<br />

Databases were recognized as valuable <strong>to</strong>ols quickly; so were graphical applications.<br />

Early graphics programs could only generate printed output – and slowly, very slowly.<br />

Nevertheless, maps and drawings have been so integral <strong>to</strong> the record keeping of the<br />

discipline that the development of graphics aids was an obvious necessity. The earliest<br />

graphics products were maps, but there were also programs developed before the end of<br />

the 1970s for drawing plans and sections. The early programs, often written by scholars,<br />

replicated many of the hand-drawing processes, and the drawings – the physical products


– were the desired results; the underlying computer data simply represented a means <strong>to</strong> an<br />

end.<br />

Programs were written by scholars <strong>to</strong> create maps, and even <strong>digital</strong> terrain models<br />

(draped grids illustrating the undulations of terrain) could be created from survey data<br />

before the advent of the personal computer. Maps were simply drawings, but powerful<br />

mapping software providing other sophisticated features was developed for other<br />

disciplines; that software, called geographic information system (GIS) programs, has<br />

been eagerly used by archaeologists, starting in the mid-1980s. For these programs the<br />

underlying data are far more important than any particular drawing.<br />

GIS programs combine maps and data about maps in ways that bring significant benefits<br />

<strong>to</strong> archaeology. The data about the maps are of two kinds, standard relational data tables<br />

(with information about artifacts, flora, fauna, etc.) linked <strong>to</strong> areas or points on maps, and<br />

information derived from map data, such as the steepness of the grade in a given area<br />

(from con<strong>to</strong>ur lines or point-source elevation data). The crucial benefit of GIS is the<br />

connection between bounded portions or individual points on a map and data about them<br />

– and the ability <strong>to</strong> analyze the data according <strong>to</strong> any of the available criteria. The<br />

resulting ability <strong>to</strong> analyze material remains in concert with the physical environment is<br />

extremely powerful.<br />

Vec<strong>to</strong>r-based maps rely upon s<strong>to</strong>red data points and are only scaled when output <strong>to</strong> a<br />

screen or paper; as a result the precision of coordinates available from a vec<strong>to</strong>r-based<br />

map is limited only by the survey technology that produced the data points. On the other<br />

hand, raster-based maps are scaled drawings, and a given map may be at virtually any<br />

scale. Issues of scale and precision may thus become very complex, and using <strong>to</strong>gether<br />

maps with differing scales, precision, and levels of generalization (e.g., how many survey<br />

points define a river course or a coastline?) can yield very misleading results. These<br />

issues of scale, precision, and generalization with GIS make it necessary that users of GIS<br />

data be sophisticated and self-conscious in their use of maps and the data connected <strong>to</strong><br />

them.<br />

The use of GIS has been aided in archaeology, as in other disciplines, by the fact that GIS<br />

data – maps and data files – can be created for one use/user and re-used by many others.<br />

Archaeologists may use GIS maps created by and for others, e.g., <strong>digital</strong> elevation maps<br />

made for the military, in addition <strong>to</strong> maps and data tables of their own. This is not an<br />

unmixed blessing, however, since available data – whether maps or data tables – may<br />

determine the questions asked or, more worrisome, the ones not asked.<br />

GIS software is the graphics road taken for mapping. The other graphics road, taken for<br />

plans and other record drawings, is computer-assisted design (CAD) software. The<br />

advent of CAD programs for personal computers brought significant and almost<br />

immediate change <strong>to</strong> record keeping for many archaeologists. Although CAD programs<br />

were created <strong>to</strong> assist with design processes, it was a short step <strong>to</strong> see their utility for<br />

modeling the existing world and existing structures; archaeologists began <strong>to</strong> use them in<br />

the mid-1980s.


Many archaeologists initially used CAD programs as drafting aids only, treating the<br />

computer-generated drawing as the final product. Plans and elevations can be generated<br />

at various scales and with differing emphases without time-consuming redrafting 2 .<br />

However, dimensional information can be retrieved from the CAD file at original<br />

measurement precision, unaffected by drawing scale, and the computer data can be<br />

segmented so that specific phases or other selections from the entire <strong>digital</strong> file can be<br />

included or excluded in any specific drawing. As a result, CAD data eventually came <strong>to</strong><br />

be recognized as more full and complex than any individual drawing.<br />

The larger promise of CAD lay in its ability <strong>to</strong> record three-dimensionally and <strong>to</strong> generate<br />

drawings based on the full 3-D geometry of a site or structure. As database technology<br />

made it possible <strong>to</strong> record the added information that archaeologists were unearthing in<br />

the 1960s and 1970s, CAD made it possible for archaeologists <strong>to</strong> record 3-D information<br />

easily and with the precision of the original field measurements. Fortunately, advanced<br />

surveying equipment, particularly the <strong>to</strong>tal station and, <strong>to</strong> a lesser extent, desk<strong>to</strong>p<br />

pho<strong>to</strong>grammetry, made the gathering of 3-D data far easier. The 3-D capabilities of CAD<br />

and advanced surveying instruments have led <strong>to</strong> very complex 3-D models as<br />

archaeological records. The models record the full geometry, and all dimensions can be<br />

retrieved at surveyed precision. The 3-D models can also be segmented <strong>to</strong> permit any onscreen<br />

view or paper drawing <strong>to</strong> include or exclude parts of the whole according <strong>to</strong> the<br />

needs of the moment.<br />

Scholars conceive of CAD models of structures in two quite different ways. For some,<br />

the CAD model is the kind of record that a drawing once was – containing all the<br />

dimensional and geometric information known from surveying. For others the model is<br />

the starting point for reconstructing missing parts or phases, for making realistic images,<br />

or for envisioning larger ensembles, cityscapes, or landscapes. Those intending <strong>to</strong><br />

reconstruct, illustrate, or envision the past may not need an exact record of existing<br />

conditions. As a result, two approaches <strong>to</strong> CAD modeling – one based on record-keeping<br />

practices and the other less concerned with precise dimensions – have been common. In<br />

general, archaeologists have been more likely <strong>to</strong> use CAD as a record-keeping<br />

technology, since their approach <strong>to</strong> data gathering emphasizes such recording.<br />

Archaeologists and architectural his<strong>to</strong>rians dealing with older and less complete<br />

structures have often used CAD for precise records as well.<br />

Archaeologists using CAD models principally for reconstructing, illustrating, or<br />

envisioning the past have worried less about precise dimensions and more about<br />

appearances. Buildings or larger ensembles imagined with the aid of the computer and<br />

with simpler, idealized dimensions are much easier and less expensive <strong>to</strong> make and can<br />

be presented in ways that are extremely compelling. However, using CAD as an aid <strong>to</strong><br />

illustrate reconstructed realities nearly always involves subsidiary technologies, usually<br />

rendering programs or virtual reality programs that use CAD models as their starting<br />

points.<br />

Reconstructions based on CAD data – even idealized and simplified CAD data – have<br />

one very important benefit over hand-drawn reconstructions: they must be based on


fitting geometric shapes <strong>to</strong> one another. Therefore, CAD models are limited by the rules<br />

of geometry. A CAD program will only generate a reconstructed view based on the rules<br />

of geometry, not the hopes or dreams of a scholar.<br />

Computer reconstructions can be truly pho<strong>to</strong>realistic, providing believably real images.<br />

That has drawbacks as well as obvious benefits. The pho<strong>to</strong>realistic views generally omit<br />

the inevitable irregularities of real structures, dirt and grime, marks of age and<br />

deterioration, nearby vegetation, and so on. The surrounding structures must often be<br />

omitted as well – or be included as if they were fully known when they are not – or be<br />

shown only as featureless blocks. Standing alone or in a generalized context, an ancient<br />

structure looks unnatural; placed in a hypothetical context it provides a sense of reality<br />

that exceeds our knowledge. The root problem here is not with the computer but with the<br />

necessarily partial state of our knowledge.<br />

Virtual reality (VR) systems based on CAD models promise realistic visions of computer<br />

worlds through which users may navigate. However, they face similar problems of<br />

inadequate data. A new hybrid system provides reconstructions that are visible in glasses<br />

through which the actual remains can be seen at the same time. This promises a better<br />

marriage of the real and the reconstructed, but the technology has yet <strong>to</strong> be proved in<br />

actual use.<br />

VR and generalized models permit serious examination of larger settings. Both VR<br />

worlds and very generalized models can provide information about landscapes, urban<br />

scale, open spaces, communication arteries, and other aspects of both the natural and the<br />

constructed world.<br />

Archaeologists have used CAD, GIS, and databases primarily for data recording. In fact,<br />

that is the most important benefit the computer has brought <strong>to</strong> archaeology – the ability <strong>to</strong><br />

manage more data more effectively. Of course, managing the data effectively implies<br />

accessing the data effectively as well. Databases, CAD models, and GIS programs all<br />

provide many ways <strong>to</strong> retrieve information, <strong>to</strong> analyze, <strong>to</strong> ask questions of the data, and<br />

<strong>to</strong> understand better what has been found and the relationships between and among those<br />

finds. That retrieval, of course, requires some computer skill.<br />

There are concerns unique <strong>to</strong> archaeology that complicate the use of all these<br />

technologies. For example, terminology issues greatly complicate the potential <strong>to</strong> share<br />

data with all these technologies, as already mentioned regarding databases. In addition,<br />

the ways archaeologists apply CAD and GIS technologies <strong>to</strong> data recording are quite<br />

different from the ways the software developers expected. Standards designed by the<br />

developers do not generally suffice for scholarly users, and the development of scholarly<br />

standards has been slow and ill-focused. The Archaeology Data Service in England has<br />

been especially helpful in this area, producing guides <strong>to</strong> good practice for scholars.<br />

Much early archaeological computing required scholars <strong>to</strong> write their own programs, but<br />

that is no longer necessary. Most now use commercial software, although special<br />

archaeological problems such as certain statistical procedures and seriation routines


continue <strong>to</strong> inspire programming by archaeologists. Another area in which programming<br />

has remained a necessity is that of simulation. Using computers <strong>to</strong> simulate development<br />

in relatively simple societies was considered a very promising technique as early as the<br />

1960s, but, as the popularity of the "new archaeology" waned, so did the enthusiasm for<br />

simulation. Those interested in simulation <strong>to</strong>day are more likely <strong>to</strong> use GIS, CAD, or<br />

statistics and <strong>to</strong> compare various predefined scenarios <strong>to</strong> test possible explanations of<br />

development, than <strong>to</strong> use simulation algorithms or artificial intelligence algorithms,<br />

which have seen little use in archaeology.<br />

The use of commercial software has made it easier for scholars <strong>to</strong> work cooperatively on<br />

data, since many can access the same data files – whether data tables, CAD models, or<br />

GIS datasets – for data entry, examination, or analysis. The potential <strong>to</strong> share <strong>digital</strong> files<br />

over the Internet, of course, has made collaboration even easier and more common. In<br />

some cases, excava<strong>to</strong>rs have entered data in<strong>to</strong> on-site database systems connected <strong>to</strong> the<br />

Internet so that the data can be accessed almost immediately after being entered.<br />

The value of the Internet for communications is easily missed in a discussion of<br />

technology; it seems <strong>to</strong>o obvious and pervasive <strong>to</strong> need mentioning. In archaeology,<br />

however, colleagues are routinely working <strong>to</strong>gether on projects while living on different<br />

continents. Simple e-mail can be crucial in such cases. For communicating <strong>to</strong> a wider<br />

audience, of course, use of the Web has already changed the extent <strong>to</strong> which scholars can<br />

and will offer their information <strong>to</strong> the public.<br />

Computers have also aided archaeologists' use of pho<strong>to</strong>graphs. Pho<strong>to</strong>graphs of buildings,<br />

sites, features, and objects are central <strong>to</strong> the practice of archaeology, and <strong>digital</strong> imagery<br />

has not changed that. However, it has made possible the use of color for a much greater<br />

portion of the pho<strong>to</strong>graphs – not in printed publications but over the Internet where the<br />

cost of a color pho<strong>to</strong> is not noticeably different from the cost of a black-and-white pho<strong>to</strong>.<br />

(In printed publications the problem remains. Small print runs make color prohibitively<br />

expensive.) The value of color is greater than one might imagine; scholars need <strong>to</strong> see<br />

colors accurately if they are fully <strong>to</strong> grasp the appearance of a structure or object. In<br />

particular, ceramic analysis demands careful attention <strong>to</strong> colors, but it is much <strong>to</strong>o<br />

expensive <strong>to</strong> print color pho<strong>to</strong>s of the sherds that make up such a large portion of the<br />

recovered material from an archaeological project.<br />

Digital imagery has also brought better, less time-consuming, and less expensive<br />

enhancements of pho<strong>to</strong>graphs. Such enhancements aid in interpreting satellite images or<br />

the stratigraphy represented in a trench wall by subtly changing colors. In at least one<br />

instance, pho<strong>to</strong>graphs of faded and damaged fragments of frescoes were <strong>digital</strong>ly<br />

enhanced <strong>to</strong> aid the res<strong>to</strong>ration work, and satellite pho<strong>to</strong>graphs of the Near East have<br />

been used <strong>to</strong> locate ancient roadways that seemed <strong>to</strong> have left no evidence on the<br />

landscape.<br />

Only recently coming in<strong>to</strong> use has been pho<strong>to</strong>graphy/digitizing that includes a 3-D<br />

modeling component so that pho<strong>to</strong>graphs and 3-D digitizers can be used <strong>to</strong> provide 3-D<br />

geometry of objects as well as color and <strong>to</strong>ne. Objects can be modeled with these


techniques so that scholars can see – and rotate and measure – fully 3-D representations<br />

of objects on a computer. While still <strong>to</strong>o expensive for general use, these processes<br />

provide unprecedented access <strong>to</strong> objects without risk of damage, although they also<br />

remove the scholars from the objects themselves.<br />

Sharing <strong>digital</strong> data over the Internet may be important for archaeologists, but scholarly<br />

electronic publication – over the Internet or on CDs – has not met the predictions of its<br />

early supporters. Some web-based monographs have appeared, as have some publications<br />

on CDs, and Internet Archaeology has been extraordinarily innovative in putting complex<br />

<strong>digital</strong> data on the Web as journal "articles." Sadly, very few of the electronic<br />

monographs seem <strong>to</strong> offer real permanence. CDs have a lifespan that can probably be<br />

measured in years, not decades, and <strong>to</strong>o many websites are ephemeral. Thus, <strong>to</strong>o few<br />

electronic publications provide the permanence required. Probably more important,<br />

electronic publication still does not normally carry the value of paper publication when<br />

advancement or promotion is considered.<br />

Although electronic monographs may not become common, web-delivered journals such<br />

as Internet Archaeology and supplements <strong>to</strong> printed articles, as pioneered by the<br />

American Journal of Archaeology, promise <strong>to</strong> bring more <strong>digital</strong> information <strong>to</strong> more<br />

archaeologists. That is not sufficient, however, because archaeology desperately needs<br />

the potential offered by electronic publication. The volume of material collected in the<br />

course of a project simply cannot be presented in a book or even a multi-volume series.<br />

Indeed, the use of databases, CAD models, and GIS programs creates a concomitant need<br />

<strong>to</strong> "publish" databases, CAD models, and GIS datasets – which can only be done<br />

<strong>digital</strong>ly. Since electronic publication in the sense of a single, unified item that can be<br />

called the publication of a project now seems unlikely, and since there must be ways for<br />

archaeologists <strong>to</strong> obtain the electronic data along with the analyses, syntheses, and other<br />

exposi<strong>to</strong>ry texts that combine <strong>to</strong> make the record of a project, the future seems <strong>to</strong> lie in<br />

hybrid publications involving either printed or web-served text coupled with permanent<br />

access <strong>to</strong> the <strong>digital</strong> files that are the basic record of the project. Archival reposi<strong>to</strong>ries for<br />

those files are required as part of this approach, and they are necessary anyway, as<br />

mentioned previously, <strong>to</strong> preserve the original data. There have been some successful<br />

reposi<strong>to</strong>ries, most notably the Archaeology Data Service in the United Kingdom. At the<br />

same time, however, there is not a universally unders<strong>to</strong>od responsibility on the part of all<br />

archaeologists <strong>to</strong> prepare <strong>digital</strong> material for such reposi<strong>to</strong>ries – not a trivial task – and<br />

then <strong>to</strong> make the actual deposit; nor have funding agencies generally recognized the<br />

importance of the long-term preservation of <strong>digital</strong> data 3 . Therefore, hybrid publications<br />

– text plus access <strong>to</strong> <strong>digital</strong> files – remain uncommon.<br />

Should the day come when hybrid publications are common, a major obstacle remains.<br />

Despite the importance of computers and computing <strong>to</strong> archaeology, <strong>to</strong>o few scholars are<br />

being trained in the use of computers. Scholars have bemoaned the problem for more<br />

than 15 years (see Richards 1985; Eiteljorg <strong>2001</strong>), but there remains a divide between<br />

those who understand the computer technologies required for the discipline and those<br />

who do not. The divide is not a matter of age but of education; young scholars are not<br />

required <strong>to</strong> learn computer skills in graduate programs. In addition, young scholars who


seem computer-savvy often lack precisely the skills most needed for archaeological<br />

computing – skills with database design, CAD programs, and GIS. Although it is now a<br />

given that any archaeology project will involve the use of computers, it is not a given that<br />

the project direc<strong>to</strong>rs will know how <strong>to</strong> use them well or have the requisite skills <strong>to</strong> find<br />

helpers who do. Nor is it a given that the archaeologists of the future will be able <strong>to</strong> use<br />

the <strong>digital</strong> data created in the field <strong>to</strong>day. Unfortunately, those who are adept with the<br />

technologies must often be self-taught, although these technologies are better learned<br />

from experts who understand the problems and pitfalls that are likely <strong>to</strong> be encountered.<br />

This problem of untrained or self-taught users of computer technology is not widely<br />

recognized or acknowledged in the field of archaeology at large, at least in part because<br />

archaeologists have not realized that all archaeologists need at least <strong>to</strong> be able <strong>to</strong> retrieve<br />

<strong>digital</strong> data from computer databases, CAD models, or GIS datasets. The absence of<br />

formal training represents a serious impediment both <strong>to</strong> effective application of computer<br />

technology and <strong>to</strong> the reuse of the <strong>digital</strong> data already gathered in computer form.<br />

Despite the many changes computers have brought <strong>to</strong> archaeology, the transformation<br />

from paper-based <strong>to</strong> <strong>digital</strong> recording is still incomplete. The discipline at large has not<br />

fully absorbed the need <strong>to</strong> preserve access <strong>to</strong> <strong>digital</strong> data for future scholars. It has not yet<br />

found an effective and relatively standard way <strong>to</strong> present <strong>digital</strong> data as part of a final<br />

publication. Its educational institutions have not accepted the need <strong>to</strong> prepare all<br />

archaeologists in the use of those computer technologies necessary in the field and the<br />

labora<strong>to</strong>ry. In part, these problems reflect the nature of the discipline, a uniquely<br />

fragmented one consisting of practitioners who may have begun as his<strong>to</strong>rians, art<br />

his<strong>to</strong>rians, students of ancient languages, or anthropologists – but not as scientists<br />

dependent upon a tradition of prompt and full data sharing. The problems also reflect the<br />

unique independence of archaeologists, who must have a strong entrepreneurial spirit in<br />

order <strong>to</strong> fund and operate complex projects. To the extent that these problems will be<br />

solved, as they surely will, it is likely that centralized initiatives such as the Archaeology<br />

Data Service in England will be crucial <strong>to</strong> the process. Other countries are pursuing<br />

similar avenues, but the more decentralized United States may find a voice in the process<br />

only through individual institutions seeking practical standards as a matter of sheer<br />

necessity.<br />

See also chapter 15: Databases.<br />

Notes<br />

1 Archaeology as a discipline may be seen as monolithic, and archaeologists may be<br />

expected <strong>to</strong> know about the archaeology of any cultural or geographic area. In fact,<br />

however, archaeologists specialize very early in their educational careers, and areas of<br />

specialization can be remarkably narrow. Consequently, readers should be forewarned<br />

that the author may not be aware of developments far from his archaeological ken.<br />

2 Other programs can produce drawings at various scales and may seem <strong>to</strong> have the<br />

necessary virtues of CAD, but the use of 3-D Cartesian grid systems as the core for data<br />

s<strong>to</strong>rage makes CAD more than a drafting aid.


3 Despite the obvious need for archiving archaeological information, archival<br />

preservation of paper records has not been routine. Thus, the situation with <strong>digital</strong> data<br />

can be seen as a continuation of sadly normal archaeological practices.<br />

References for Further Reading<br />

Allen, Kathleen M. S., W. Green Stand<strong>to</strong>n, and Ezra B. W. Zubrow, (eds.) (1990).<br />

Interpreting Space: CIS and Archaeology. London, New York, Philadelphia: Taylor and<br />

Francis.<br />

Badler, Norman and Virginia R. Badler (1977). SITE: A Color Computer Graphics<br />

System for the Display of Archaeological Sites and Artifacts. Philadelphia: University of<br />

Pennsylvania, Moore School of Electrical Engineering, Department of Computer and<br />

Information Science.<br />

Binford, Sally R. and Binford, Lewis R., (eds.) (1968). New Perspectives in Archeology.<br />

Chicago: Aldine.<br />

Duncan, J. M. and P. L. Main (1977). The Drawing of Archaeological Sections and Plans<br />

by Computer. Science and Archaeology 20: 17–26.<br />

Eiteljorg, H., II (<strong>2001</strong>). Computing in the Archaeological Curriculum. CSA Newsletter<br />

14, 2. Available at: http://csan.org/newsletter/fall01/nlf0104.html.<br />

Richards, J. D. (1985). Training Archaeologists <strong>to</strong> Use Computers. Archaeological<br />

Computing Newsletter 2: 2–5.<br />

Richards, J. D. and N. S. Ryan, (eds.) (1985). Data Processing in Archaeology (from the<br />

series Cambridge Manuals in Archaeology). Cambridge: Cambridge University Press.<br />

Sabloff, Jeremy A., (ed.) (1981). Simulations in Archaeology. Albuquerque: University of<br />

New Mexico Press.<br />

Upham, Steadman, (ed.) (1979). Computer Graphics in Archaeology: Statistical<br />

Car<strong>to</strong>graphic Applications <strong>to</strong> Spatial Analysis in Archaeological Contexts<br />

(Anthropological Research Papers no. 15). Tempe: Arizona State University.<br />

The following contain numerous articles concerning archaeological computing and<br />

should be consulted broadly:<br />

Annual proceedings of the meetings on Computer Applications and Quantitative Methods<br />

in Archaeology beginning in 1973 (published for that one year only as Science in<br />

Archaeology, 9).<br />

Archaeological Computing Newsletter (tables of contents from number 45 (Spring 1996)<br />

onward may be found at . The


Archaeological Computing Newsletter will be published by Archeologia & Calcola<strong>to</strong>ne<br />

from the end of 2004.<br />

CSA Newsletter (All issues beginning with volume 8 are available on the Web,<br />

; volumes 13 and following are only available on the<br />

Web.)<br />

Selected Websites<br />

– the website for the excavations at<br />

Çatalhöyük, Turkey, where modern technology has not only been extensively used but<br />

discussed and debated.<br />

– information about the site Teotihuacan, a site that<br />

has benefited from extensive use of computer technology and at which computers have<br />

been used for a very long time.<br />

– should be used in concert with the<br />

publication by James Packer (1997), The Forum of Trajan in Rome (Berkeley: University<br />

of California Press), for an example of an architectural his<strong>to</strong>rian using both traditional<br />

and computer methods.<br />

http://ads.ahds.ac.uk/> – the Archaeology Data Service in England (and<br />

for the ADS Guides <strong>to</strong> Good<br />

Practice and for a different version of<br />

the CAD Guide <strong>to</strong> Good Practice).<br />

– online journal, Internet Archaeology.<br />

- CSA Propylaea Project, still in early stages, <strong>to</strong> construct a s<strong>to</strong>neby-s<strong>to</strong>ne<br />

CAD model of the Propylaea. Discussions of methods, but only the CAD model<br />

of the predecessor is available.<br />

– excavations at Lahav, Israel, with<br />

discussion of data collection methods and computer use in the field.<br />

– Amiens cathedral project at<br />

Columbia University, with extensive imagery showing the illustrative power of CAD and<br />

rendering <strong>to</strong>ols.<br />

– Savannah urban design/his<strong>to</strong>ry/development project of the<br />

Savannah College of Art and Design, still in early stages. Only Windows users need<br />

apply, alas.<br />

– a study of hill forts in France with<br />

effective use of GIS technology.


– the Corinth Computer Project at the<br />

University of Pennsylvania.<br />

– the Digital Archive Network for Anthropology at<br />

North Dakota State University, a project involving the use of 3-D models of<br />

archaeological objects that may be examined on the Web.<br />

3.<br />

Art His<strong>to</strong>ry<br />

Michael Greenhalgh<br />

Introduction<br />

The discipline itself in this context should be unders<strong>to</strong>od <strong>to</strong> include not only lecturers and<br />

students who impart and acquire knowledge, and cura<strong>to</strong>rs who look after resources such<br />

as slides, but also the staff of museums and galleries, with the oversight and care of large<br />

quantities of works, usually many more than can be displayed. And we should remember<br />

that other his<strong>to</strong>rians need and use images from the past, 1 a fact reflected in the broad<br />

church of participants included in conferences. 2 In large collections especially, computers<br />

were early recognized as an essential <strong>to</strong>ol in regimenting what was often incomplete<br />

paper documentation for the use of cura<strong>to</strong>rs, conserva<strong>to</strong>rs, registrars, and, eventually, the<br />

general public.<br />

Art his<strong>to</strong>rians use computers <strong>to</strong> order, sort, interrogate, and analyze data about artworks,<br />

preferably with images; increasingly they wish <strong>to</strong> use the Web as the carrier for<br />

multimedia research or teaching/learning projects. The fact that primary data for art<br />

his<strong>to</strong>rians are visual – images, still and perhaps moving – did not prevent an early interest<br />

in the ordering and interrogation of text data arranged in<strong>to</strong> databases (following the<br />

development of au<strong>to</strong>mation in libraries) or studying how <strong>to</strong> deal with the quantitative data<br />

churned out by a counting machine (Floud 1979).<br />

This chapter charts such developing use, from the development of text databases through<br />

the introduction of <strong>digital</strong> imaging and then <strong>to</strong> the use of the Web in teaching and<br />

learning. It does not consider the use of computers in art, 3 nor yet text bibliographies of<br />

art his<strong>to</strong>rical material, such as the Bibliographic d'His<strong>to</strong>ire de I'Art, which are librarybased<br />

activities. It illustrates how disciplinary innovation in this area is inevitably<br />

dependent upon and triggered by technical innovation, and how the introduction and<br />

popularization of the Web with its consistent interface has brought new possibilities for<br />

computer use in learning. Each section introduces and explains the various problems, and<br />

then charts with relevant links how various people and organizations tried <strong>to</strong> solve them.<br />

Far from hymning <strong>to</strong>day's computing world as a nirvana, the theme throughout is of<br />

difficulties with expertise, communication, and finance – that is, with people rather than<br />

technology. These are problems that intensify as the technologies and the potential


applications become more sophisticated (cf. the Virtual Museum of Computing: see<br />

http://vmoc.museophile.org/).<br />

Text Databases<br />

Early computers were unsuitable for <strong>humanities</strong> use: large, slow, expensive, and with<br />

software designed by and therefore for scientists, they were jealously guarded. Using one<br />

<strong>to</strong> write a book (with the attraction of supposed "au<strong>to</strong>mated indexing") was seen as nearblasphemy<br />

and, in any case, was difficult given the restricted software and the way such<br />

machines dealt with text. The construction of databases was no easier, because interface<br />

programs, which softened and simplified data entry and retrieval, were rare. And the<br />

discipline had two problems, which did not confront run-of-the-mill author-title-date<br />

databases: the part-whole problem (i.e., how <strong>to</strong> catalogue a multipart altarpiece, let alone<br />

a Gothic portal or a complete cathedral), and the element of imprecision, which is often<br />

part of his<strong>to</strong>rical data.<br />

What is more, the interchangeability and communication we take for granted (and<br />

demand) <strong>to</strong>day did not exist: operating system did not necessarily talk un<strong>to</strong> operating<br />

system; input was usually via a text terminal if not via punchcards or paper tape<br />

(certainly not a VDU); interfaces hovered between the obscure and the feral; output was<br />

<strong>to</strong> line printer or daisy-wheel or, with some difficulty, <strong>to</strong> pho<strong>to</strong>typesetter; and intercomputer<br />

communication was tricky because of the myriad of physical and logical<br />

formats in use. Nevertheless, some institutions (such as the Detroit Institute of Arts: cf.<br />

<strong>to</strong>day's Visual Resources Art Image Database) used computers <strong>to</strong> develop catalogues of<br />

their collection, perforce using specially commissioned software, or adapting mainframe<br />

programs (such as IBM's STAIRS) <strong>to</strong> their needs. By the early 1980s, art his<strong>to</strong>rians<br />

began <strong>to</strong> develop database projects that grappled with the problem of describing images<br />

(see Corti 1984). As in other areas (such as theater his<strong>to</strong>ry: cf. The London Stage, see<br />

Donahue, website) the road was often rocky and long, with memory and s<strong>to</strong>rage<br />

restrictions of early systems demanding laconic descrip<strong>to</strong>rs (Villard 1984); and funding<br />

for staff, machinery, and programming difficult <strong>to</strong> find. The Universit` degli Studi in<br />

Siena was early in the field, with a how-<strong>to</strong>-do-it manual (Bisogni and Corti 1980)<br />

followed by one on methodology describing projects (Bisogni 1980), two giving<br />

exemplary applications (di Bari et al. 1981) and yet another tackling the problem of<br />

dealing with images (Bisogni 1981).<br />

The first problem in establishing art his<strong>to</strong>rical databases was terminology. Even well after<br />

the penetration of the microcomputer, the gap between the attitudes of visual resources<br />

(VR) cura<strong>to</strong>rs and academics was clear: few of the latter would take an interest in<br />

computerization until machine and network could demonstrably help with teaching or<br />

research; and all who had done battle with different slide rooms were quite used <strong>to</strong> a<br />

flexible (read: vague) use of terminology. In any case, most departments had "legacy"<br />

collections of slides, and few intended <strong>to</strong> relabel them all <strong>to</strong> fit some grand scheme. The<br />

VR cura<strong>to</strong>rs were more precise, reasoning that chaos would result unless the same<br />

terminology was employed. Several conferences concentrated on this important matter<br />

(see Roberts 1990). Standards are fine, but is it the researchers who want them or the


librarians? When somebody at such a conference opined that For every simple problem,<br />

there is a complicated solution, perhaps she was thinking of the Art His<strong>to</strong>ry Information<br />

Program 4 at the Getty, and its Art and Architecture Thesaurus: begun pre-computers, it<br />

offers an excellent set of standards, but these are expensive <strong>to</strong> implement and far from<br />

universally applicable 5 . The fears and expectations of museum professionals are well<br />

reflected in the journals of the period 6 .<br />

In any case, how <strong>to</strong> catalogue an artwork? One possibility was by meaning; and a fruitful<br />

project which began without any connection with a computer is ICONCLASS, the origins<br />

of which go back <strong>to</strong> 1954, and which provides an iconographical classification system,<br />

and now a CD-ROM, and web browser software.<br />

In these days before effective and cheap graphics, the discipline's associations and<br />

journals naturally concentrated on database matters. Thus, Computers and the His<strong>to</strong>ry of<br />

Art publishes since 1991 a journal (CHArt, a development of a Newsletter begun in<br />

1985), holds annual conferences, usually themed, and now publishes them on the Web.<br />

The Visual Resources Association, initially cura<strong>to</strong>rs meeting annually since 1968 at<br />

College Art Association conferences, was founded formally in 1982: it now has some 600<br />

members and a quarterly Bulletin, and correctly titles itself "The international<br />

organization of image media professionals" with members working in "educational and<br />

cultural heritage environments." The range of the membership indicates just how broadly<br />

"Art His<strong>to</strong>ry" images are used and the problems and possibilities they entail are<br />

recognized: they include information specialists; <strong>digital</strong> image specialists; art,<br />

architecture, film, and video librarians; museum cura<strong>to</strong>rs; slide, pho<strong>to</strong>graph, microfilm,<br />

and <strong>digital</strong> archivists; architectural firms; galleries; publishers; image system vendors;<br />

rights and reproductions officials; pho<strong>to</strong>graphers; art his<strong>to</strong>rians; artists; and scientists.<br />

The Lure of "Intelligent" Software<br />

But surely computers are "thinking" machines, or so it was <strong>to</strong>o often assumed. Much was<br />

made in the 1980s of computers as "intelligent", and able <strong>to</strong> learn and therefore respond<br />

just like a teacher. It should be difficult <strong>to</strong>day <strong>to</strong> conceive of the computer as any more<br />

than a speedy idiot obeying a set of precise instructions; but many in the <strong>humanities</strong> do<br />

indeed continue <strong>to</strong> regard it as a black and magical (or black-magical) box – throw the<br />

data at it, and somehow the computer will make sense of them.<br />

Such a continuing substitute for idol-worship has had many adverse consequences, which<br />

continue <strong>to</strong> this day. The first was in the early 1980s, when the heralded "Fifth<br />

Generation", <strong>to</strong> be developed by the Japanese (the very mention of MITI struck<br />

apprehension in<strong>to</strong> the Western heart), was <strong>to</strong> produce new levels of software. Second, this<br />

was the age, not coincidentally, when Artificial Intelligence programs – AI – would<br />

supposedly take over medical diagnosis, mineral exploration, and, of course, education.<br />

Intelligent systems were also proposed for art his<strong>to</strong>rians, because computer databases<br />

were seen as able <strong>to</strong> provide intelligent answers <strong>to</strong> researchers' questions, such as Cerri's<br />

expert system for archaeologists and art his<strong>to</strong>rians. Characteristically, Cerri was from the<br />

Computer Department at the Scuola Normale Superiore (SNS) at Pisa. The SNS provided


conferences, examples, and publications <strong>to</strong> drive along the application of computer<br />

processing <strong>to</strong> art his<strong>to</strong>ry 7 .<br />

Graphics<br />

The breakthrough for art his<strong>to</strong>rians and other humanists came not with the introduction of<br />

microcomputers in the later 1970s (usually tricky <strong>to</strong> program and <strong>to</strong> use, with strange<br />

s<strong>to</strong>rage possibilities and nil communication, and intended for aficionados) but with the<br />

IBM PC in 1981 and then the Macin<strong>to</strong>sh. I borrowed one from a computer labora<strong>to</strong>ry at a<br />

large American University in late 1981: perhaps not recognizing the cuckoo in their<br />

superior mainframe nest, they hadn't played with it and didn't want <strong>to</strong> do so (just as many<br />

IBM engineers considered it a <strong>to</strong>y). As is well known, that machine with its open (i.e.,<br />

easily and freely imitable) architecture led <strong>to</strong> an explosion of cheap machines with<br />

increasingly user-friendly and sometimes useful software, a cheap s<strong>to</strong>rage medium (the<br />

floppy disk) and, eventually, a hard disk and then CD-ROM. In 1984 arrived the Apple<br />

Macin<strong>to</strong>sh, the first PC with a graphical user interface. The cheapness and apparent<br />

utility of micros set many people pondering how they might organize, access, and<br />

disseminate the mountains of text and graphics data used in art his<strong>to</strong>rical research and<br />

teaching (see Heusser 1987).<br />

Pictures on a computer were now theoretically possible for every art his<strong>to</strong>rian, even if<br />

only in greyscale. Given the puny power of such machines, however, they were not yet<br />

feasible – but that would change in the following decade. If the first programs that<br />

ensured the popularity of micros were spreadsheets, the second were database packages,<br />

for the first time enabling art his<strong>to</strong>rians and museum personnel <strong>to</strong> construct at minimal<br />

cost databases of artifacts. Until the late 1980s, these were text-only, and sometimes<br />

relational. Only with the more powerful hard-disk machines of the late 1980s did image<br />

databases come along, and then the users faced a perennial dilemma: should they<br />

purchase a commercial product, paying for the privilege of testing it as version follows<br />

version, and lock themselves in<strong>to</strong> a proprietary setup? Or should they go for open rollyour-own<br />

software? Luckily, with the development of the freeware/shareware ethos,<br />

good open programs could be had for little money; while the development of avowedly<br />

non-proprietary operating systems such as Linux not only made people look at the PC<br />

with greater fondness, but also encouraged the development of freely distributable<br />

software in all application areas.<br />

So was it immediately "goodbye 35 mm slides, hello <strong>digital</strong> images"? If all successful<br />

computers have been <strong>digital</strong> machines from the start, the last twenty years have<br />

highlighted the problem of how <strong>to</strong> deal with analogue materials in an increasingly <strong>digital</strong><br />

age. For the art his<strong>to</strong>rians, this means principally slides and pho<strong>to</strong>graphs. The 1960s had<br />

seen the move from 6 × 6 lantern slides (usually greyscale) <strong>to</strong> 35 mm slides (usually<br />

color), and most university departments built up large collections which had <strong>to</strong> be bought,<br />

labeled, catalogued, perhaps mounted, and sometimes repaired. With strong light<br />

projected through them, slides fade: a prize exhibit in one collection was six slides of the<br />

selfsame Monet Haystack, all with different coloration, and none that of the actual<br />

painting. Every collection had much the same basic set of slides, curated by expensive


personnel: might money not be saved by collecting such images on<strong>to</strong> a more cheaply<br />

reproducible medium?<br />

The late 1980s also saw the introduction of powerful workstations with sufficient<br />

memory and screen resolution <strong>to</strong> manipulate graphics. These were still <strong>to</strong>o expensive for<br />

most people, but pioneering projects such as Marilyn Aronberg Lavin's Piero Project and<br />

the English VASARI Project were funded, and demonstrated the possibilities of highpower<br />

computer graphics (including VRML models [Lavin 1992]: see below) <strong>to</strong> explore<br />

Piero della Francesca's monumental chapel in Arezzo.<br />

Examples<br />

1 Prince<strong>to</strong>n Index of Christian Art, founded 1917. By 1990 it had some 800,000 file cards<br />

documenting over 25,000 different subjects, plus a cross-referenced pho<strong>to</strong> file with<br />

nearly a quarter of a million images. Copies also in the Vatican, UCLA, Utrecht and<br />

Dumbar<strong>to</strong>n Oaks. In 1990, with Mellon Foundation and Prince<strong>to</strong>n funding, a<br />

computerized database was begun.<br />

2 The Marburger Index Inven<strong>to</strong>ry of Art in Germany is another long-lived venture 8 ,<br />

which as Fo<strong>to</strong> Marburg goes back <strong>to</strong> 1913, and collected pho<strong>to</strong>graphs of art in Germany.<br />

Placed on microfiche in the 1980s, the Web version 9 currently offers access <strong>to</strong> some<br />

220,000 images. It offers a variety of iconographic catalogues ordered by ICONCLASS,<br />

as well as orderings according <strong>to</strong> function, form, material, techniques, and so on.<br />

3 The Witt Computer Index at the Courtauld Institute in London: controlled vocabulary,<br />

authority files, ICONCLASS for subject matter.<br />

4 The Census of Antique Works of Art Known <strong>to</strong> the Renaissance: beginning in 1946 in<br />

the pho<strong>to</strong>graphic collection of the Warburg Institute, funding from the Getty Art His<strong>to</strong>ry<br />

Information Program allowed a start <strong>to</strong> be made on its computerization; from 1981<br />

housed at both the Warburg and the Biblioteca Hertziana, Rome, a demonstra tion<br />

videodisk was pressed in 1988. Its home from 1995 has been the Kunstgeschichtlichen<br />

Seminar of the Humboldt University in Berlin. Cf. the conference Inaugurating a<br />

Scholarly Database: the Census of Antique Works of Art and Architecture Known <strong>to</strong> the<br />

Renaissance, Warburg Institute, March 19–21, 1992.<br />

Laserdisks<br />

One viable technology, which lost the race for watch-at-home films <strong>to</strong> the VCR, was the<br />

laserdisk or videodisk, which could hold about 40,000 still frames, was cheap <strong>to</strong> copy,<br />

and played <strong>to</strong> an ordinary TV moni<strong>to</strong>r. What is more, it was a random-access device<br />

(simply punch in the desired frame number), and the images thereon did not fade. The<br />

images could be enregistered professionally, or a recorder (e.g., by Sony) bought and<br />

used. Although it could be operated standalone, the laserdisk player was easily linked <strong>to</strong> a<br />

computer, with the latter holding the text data controlling the display of images.


The 1980s therefore saw the production of several very ambitious laserdisk projects.<br />

Some of these were by Art His<strong>to</strong>ry departments who laserdisked their holdings; little<br />

information is now available about these copyright-precarious ventures, and the results<br />

certainly did not get much circulated, so that economies of scale were unavailable.<br />

However, several museums and galleries also produced laserdisks of some of their<br />

holdings. High-quality projects included the Vatican Library videodisks, on sale from<br />

1993: the first three disks offered a <strong>to</strong>tal of 75,000 images, and it was estimated that the<br />

Latin MSS alone would occupy 400,000 images 10 . But the relatively high cost of the<br />

playing equipment, and their specialized market, meant that few such ventures were a<br />

financial (as opposed <strong>to</strong> a scholarly) success. Where laserdisks were useful was in<br />

supplementing the speed and s<strong>to</strong>rage deficiencies of computers; thus the first edition of<br />

Perseus had the software and text on a CD-ROM, with the images displayed on the linked<br />

laserdisk player.<br />

However, although laid down in <strong>digital</strong> format on the platter, the laserdisk is an analogue<br />

technology, and is really video, so that only one dimension of image is possible (NTSC<br />

and PAL being slightly different but neither offering more in modern <strong>digital</strong> parlance<br />

than a 300,000-pixel image). By comparing such images with what is possible <strong>to</strong>day (up<br />

<strong>to</strong> about 5,000,000 pixels, or over 16 times that resolution), it is easy <strong>to</strong> see why the<br />

laserdisk has died. But it poses the dilemma: if laserdisks had a ten-year life, and even<br />

assuming <strong>to</strong>day's support media will remain readable, will <strong>to</strong>day's <strong>digital</strong> images soon<br />

seem puny, and need redigitizing?<br />

The decade 1984–94 therefore saw much activity as text databases began <strong>to</strong> be populated<br />

with images, at first by scanner or video. This was <strong>to</strong> be a long and difficult journey<br />

because of the huge number of records involved, the cost of digitizing, and worries about<br />

image resolution and longevity of the support (tapes, hard disks).<br />

3-D graphics<br />

Interest in graphics was not circumscribed by still images, because it seemed that the new<br />

power of computers and speedy networks could deliver much more interesting scenarios<br />

in the form of virtual reality. The construction in a computer of a graphical representation<br />

of the real world or an imaginary one was seen by some visionaries as offering much<br />

more attractive and educationally rich presentations. Why display St Peter's Square or the<br />

inside of the Louvre in flat images? Why not build them in the machine, and allow the<br />

user <strong>to</strong> zoom in, out, and around, accessing other resources (text, sound, video) with the<br />

click of a mouse? Was this not an effective way of allowing students <strong>to</strong> experience real<br />

environments it was impractical <strong>to</strong> visit? Package the presentations for the Web, using the<br />

Virtual Reality Modeling Language (VRML), and use them for teaching – although the<br />

flaw in the argument is that technologies developed for gaming do not necessarily meet<br />

exacting standards either of detail or of accuracy – of "realism."<br />

VRML models are difficult <strong>to</strong> prepare and expensive <strong>to</strong> construct; so that even if one<br />

convincing model were <strong>to</strong> be constructed, this would not help much in art his<strong>to</strong>ry courses


dealing with tens of monuments. Nor are there any signs that more "intelligent" software<br />

will rescue what looks increasingly like a dead end (but not before many art his<strong>to</strong>rians,<br />

architects, and archaeologists had spent time on ambitious plans for a computer-generated<br />

version of the world). Of course, a continuing problem in anything <strong>to</strong> do with computers<br />

is hype versus reality; so that while we can all agree that three dimensions should be an<br />

improvement on two (cf. human sight), computer reconstructions of the real world (I<br />

exclude drawn reconstructions of archaeological sites – cf. titles such as Rediscovering<br />

Ancient Sites through New Technology – where such technologies probably do have a<br />

future) remain the triumph of hope over experience. Even if for the devotees of virtual<br />

reality we are "virtually there" (Lanier <strong>2001</strong>), it is sensible <strong>to</strong> recognize the scale of the<br />

problem from data acquisition <strong>to</strong> processing and display 11 . Nevertheless, computer<br />

graphics, both raster and bitmapped, have a very large part <strong>to</strong> play in the study of art and<br />

architecture (cf. Stenvert 1991) 12 . We may even hope for the further development of<br />

au<strong>to</strong>stereoscopic 3-D displays, which will allow the user <strong>to</strong> view the displayed scene<br />

without using stereo glasses 13 .<br />

CD-ROM and <strong>digital</strong> camera<br />

In the late 1980s and early 1990s, the generation of <strong>digital</strong> images for manipulation on<br />

standalone computers was done by the end user with a video camera, or by Kodak putting<br />

images on<strong>to</strong> their multi-resolution Pho<strong>to</strong>-CD format. Two technologies that between<br />

1993 and 1998 moved from the high-price, low-volume specialist <strong>to</strong> the low-price, highvolume<br />

consumer category were the CD-ROM burner and the <strong>digital</strong> camera. CD-ROMs<br />

could be professionally pressed in the 1980s, but at high cost; and the first expensive and<br />

tricky "end user" burners appeared in 1994. For art his<strong>to</strong>ry the burner and the camera go<br />

<strong>to</strong>gether (with their cousin the scanner), because the output of the latter is conveniently<br />

s<strong>to</strong>red on the former; as images get larger, so more CD-ROM blanks are needed – hence<br />

the recent introduction of its child, the DVD burner, one version of which holds 4.2 GB,<br />

rather than the 700 MB of the conventional CD-ROM.<br />

By 1999, prices were reasonable for both items (about a tenfold drop in six years), and<br />

light and cheap portable cameras and burners helped scholars on research trips. Only now<br />

did the old argument of the art-his<strong>to</strong>rical computer-deniers that <strong>digital</strong> images were<br />

inferior <strong>to</strong> 35 mm slides come <strong>to</strong> grief. At last it became feasible and cheap <strong>to</strong> digitize<br />

sets of images for class use, although difficulties in the provision of such material <strong>to</strong> the<br />

students remained, as we shall see below.<br />

Examples<br />

1 To encourage involvement, the CIHA Conference in 2000 organized a competition for<br />

"the best <strong>digital</strong> productions in Art His<strong>to</strong>ry."<br />

2 The Deutsche Forschungsgemeinschaft supports an initiative of the Art His<strong>to</strong>ry<br />

Department of the Humboldt University intended <strong>to</strong> build up an international archive of<br />

Virtual Art (see Grau website).


The Internet, Web, and Communication<br />

In the mid-1980s, micros communicated little better than mainframes. For museums and<br />

galleries with data, how <strong>to</strong> interchange them? By tape, certainly; but would the tape from<br />

machine A fit machine B and, if so, could the data thereon be read with or without the<br />

program with which they were written? Were two sets of data compatible <strong>to</strong> the extent<br />

that they could be merged, and thereby fulfill a main reason for communication? In<br />

general, dog-in-the-manger attitudes have kept data discrete; and it is ironic that the<br />

development of the Web has meant that anyone's data can be read in a web browser,<br />

thereby negating any urgency in actually sharing them. Idealists, who expected that the<br />

Internet and then the Web would mean access <strong>to</strong> artifacts beyond gallery boundaries,<br />

after the manner of the librarians' union catalogue, 14 have been largely disappointed.<br />

If the PC and graphical user interface held the promise of the easy manipulation of<br />

images and concomitant text by art his<strong>to</strong>rians, their communication across the Internet<br />

(Naugh<strong>to</strong>n 1999) in universally readable formats was lacking. The enthusiasm and effort<br />

in <strong>humanities</strong> computing may be measured by the 701 pages of the 1989–90 Humanities<br />

Computing Yearbook: A Comprehensive Guide <strong>to</strong> Software and Other Resources 15 .<br />

Although digitizing images was in hand by 1990 (see Durran 1992–3, for examples), the<br />

Web and its pro<strong>to</strong>cols and software, a European invention from CERN (Naugh<strong>to</strong>n 1999:<br />

230–9), was the breakthrough (in 1993/4), offering a consistent way of viewing data in<br />

various formats, and hyperlinking between them. The data could be served from the local<br />

machine or, more normally, pulled down from a server Out There somewhere on the<br />

Web. Text and still images would write <strong>to</strong> the browser quickly; sound and video would<br />

continue <strong>to</strong> be problematical on slower networks. For the first time, communication<br />

looked easy, in that any connected machine that could run the browser could get the data<br />

– no need <strong>to</strong> bother about incompatible operating systems any more. This cheap, easy,<br />

and attractively multimedia pro<strong>to</strong>col saw the designers of earlier teaching and learning<br />

packages (most of which began with the penetration of the micro) eager <strong>to</strong> upgrade their<br />

work and reach a larger audience.<br />

Education<br />

Suddenly, with the Web, computing in an education context also looked practicable at a<br />

reasonable price. Some departments offered course images on the Web using Mosaic as<br />

early as 1994. Although useful websites grew only slowly, and access around the world<br />

also (spreading from the G8 outwards), educa<strong>to</strong>rs could now contemplate the use of these<br />

learning materials, with teacher and students being in different locations, and the data<br />

being accessed asynchronously, even if they stipulated the survival of the campus-based<br />

university.<br />

The minutiae and problems of communication and cooperation bypassed many<br />

administra<strong>to</strong>rs, in government as well as in education, who since its introduction had seen<br />

the microcomputer as a cost-effective way of teaching without teachers. Typical of this<br />

period (and our own!) was an increasing demand for education and a reluctance <strong>to</strong> fund


personnel: machines are cheaper, do not get sick or sue, and can be written down for tax<br />

purposes. So computers came in<strong>to</strong> education, sometimes as a requirement based on faith<br />

rather than need, but more often as cost-cutters, not as education-enhancers or facilita<strong>to</strong>rs.<br />

The consequences are with us <strong>to</strong>day – usually as much hardware as anyone can use, but<br />

an inadequate level of personnel, and no understanding of the need for programmers <strong>to</strong><br />

write <strong>humanities</strong>-specific programs. The current hope is <strong>to</strong> have computers search visual<br />

databases by subject, and have the software recognize and retrieve – all images<br />

containing a tree and a river. Such content-based retrieval might be realized: it is logical<br />

<strong>to</strong> wish for software be able <strong>to</strong> "recognize" real-world objects, and it would be interesting<br />

<strong>to</strong> see such a process demonstrated on a wide range of art his<strong>to</strong>rical data 16 .<br />

Examples<br />

1 Journals promoted education via computer. 17<br />

2 In the museum world, the Museum Computer Network played an important role, with<br />

its archives online since March 1999 (eSpectra website).<br />

3 The Computers in Teaching Initiative (CTI) project was begun in 1984 by the British<br />

Computer Board for Universities and Research Councils <strong>to</strong> help staff use computers in<br />

support of teaching. In 1989, 24 specialist centers were established, including a CTI<br />

Center for His<strong>to</strong>ry, Archaeology, and Art His<strong>to</strong>ry. This offered workshops, and published<br />

a Newsletter welcoming "reports on the use of computers for teaching or research …<br />

announcements of new courses, reports on running courses, news about projects,<br />

software, datasets and conferences, and letters." The mission of the CTI<br />

(), according <strong>to</strong> its publicity, was "<strong>to</strong> maintain and enhance the<br />

quality of learning and increase the effectiveness of teaching through the applica tion of<br />

appropriate learning technologies." The aim was <strong>to</strong> "enable change in higher education,<br />

because with computers teachers and learners can determine the place, time and pace for<br />

learning." Their refereed journal, Active Learning, emphasizes "learning outcomes rather<br />

than enabling technologies." In 1991, 314 packages for his<strong>to</strong>rians were listed (Spaeth<br />

1991).<br />

Art His<strong>to</strong>ry and Computing Today<br />

The discipline continues <strong>to</strong> be a minnow in the <strong>humanities</strong> pond, and a relatively<br />

expensive one (all those slides and pho<strong>to</strong>graphs, and glossy books), but books which in<br />

years past targeted a very broad audience now focus on art his<strong>to</strong>rians. 18 There are many<br />

projects and ventures in the area (cf. CNI and Düsseldorf), and several ventures (<strong>to</strong> add <strong>to</strong><br />

the pioneering Marburg Index – see above) which offer an overview of national holdings,<br />

such as the Joconde database of works in 75 French museums, which joins a series of<br />

other French initiatives (see, for example, the Ministère de la culture website). These are<br />

important undertakings: the architecture database called Mérimée, for instance, has over<br />

150,000 records. It also derives from a base prepared in the 1980s: as early as 1985, the<br />

Ministère de la culture committed in hardware alone FF18.2 million <strong>to</strong> archaeology,<br />

FF10.8 million for labora<strong>to</strong>ries, and FF10.3 million for the digitization of images and


sound (Ministère de la culture 1985). Why are not the governments or cultural<br />

organizations of other first-world, computer-rich countries doing likewise?<br />

But as in many other disciplines, moves <strong>to</strong>ward the use of computers and the Web in<br />

teaching and learning have been left <strong>to</strong> the initiative of individual academics; some of<br />

these have been <strong>to</strong>ld sententiously that academics should leave such technical matters <strong>to</strong><br />

programmers; others have seen their enterprise met by lack of interest, help, or adequate<br />

financing from the administration. On the other hand, institutions in Germany and<br />

Switzerland promote computers in research and conferences (see Hamburg University<br />

website), and in courses on computing in the discipline (see Universitat Trier website). A<br />

collaborative venture – Scbule des Sebens - offers web-based seminars, using Web-CT.<br />

And although the connectivity of computers should also mean the enthusiastic<br />

cooperation of academics across institutions, and a willingness <strong>to</strong> share data, in no<br />

country has any such organizational initiative been taken nationally, and international<br />

initiatives tend <strong>to</strong> address narrow research issues rather than <strong>to</strong> aim <strong>to</strong> facilitate teaching.<br />

"No man is an island", wrote John Donne; but then he had no experience of our<br />

supposedly connected world, and of the fears throughout the system. If I develop a webbased<br />

course, who owns it, the institution or me? If it works, am I still needed? If ArtAl<br />

exists as such a course, how many young lecturers will be needed, except <strong>to</strong> keep it<br />

updated?<br />

So although the technologies (including Web software) are clever and advanced, it is the<br />

human element that restricts obvious developments in the discipline. These (some of<br />

which are valid across other disciplines) include: acknowledgment that computer literacy<br />

is essential for all staff and students; rethinking of traditional modes of course delivery<br />

(which doesn't mean their complete replacement by the Web); gradual replacement of<br />

analogue slide collections by <strong>digital</strong> ones retrieved from some international database-inthe-sky;<br />

insistence that web-based learning should improve the quality of education,<br />

rather than lower the unit of resource for the administra<strong>to</strong>rs; planning for classrooms<br />

which encompass existing web-based technologies, including videoconferencing.<br />

Conclusion: The Future<br />

Although computing for art his<strong>to</strong>rians is considerably simpler now than it was ten or<br />

twenty years ago, many problems remain, and they still revolve around communication,<br />

expertise, and finance.<br />

Ease of communication, and a relative freedom from hackers as from the 90 percent or<br />

more of idiocy which overloads the net, requires the development of private or restricted<br />

networks which will allow the easy transmission and viewing of visual material of<br />

spectacularly high resolution. In many ways it is a pity that the Internet and the Web,<br />

both developed in universities or university-like institutions on public money, did not tie<br />

all the technologies up with patents (against the ideology, I know), and then recoup funds<br />

by making commerce pay hugely for their use. This would certainly have softened the<br />

funding crisis in higher education.


A growing desire of art his<strong>to</strong>rians <strong>to</strong> teach using the Web requires either the upgrading of<br />

useful but general-purpose course-construction programs such as Blackboard or WebCT,<br />

or the development of suitable programs usable by newly computer-savvy staff.<br />

Yet financial problems in a small discipline are certain <strong>to</strong> remain. If there are more than<br />

1,999 museums online worldwide (MCN website), funding prevents many teaching<br />

institutions from presenting their materials on the Web. This will happen gradually, as 35<br />

mm slides degrade and as <strong>digital</strong> cameras take over from analogue ones. Then the day<br />

will soon come when VR cura<strong>to</strong>rs must be as computer-literate as staff in university<br />

libraries (now usually part of something like a Division of Scholarly Information), as the<br />

Visual Resources Association (VRA) (see above) has recognized for years. In other<br />

words, such personnel must capture the high ground of indispensability just as librarians<br />

have so successfully presented themselves as the natural high priests for the curating of<br />

<strong>digital</strong> as well as of paper-based information. So how will such a small discipline keep its<br />

head above water if it goes thoroughly <strong>digital</strong>? We already have a system where numbers<br />

count instead of quality (5 students taking Poussin: bad; 200 students taking Desire and<br />

the Body. good). In the nineteenth century academics were sometimes paid according <strong>to</strong><br />

the number of students they attracted (plus ça change …); so can we envisage those<br />

academics who develop expensive web-based courses recouping the outlay by some kind<br />

of pay-per-use micro-money, since computers are good at such calculations? Since<br />

<strong>to</strong>urism is a crucial element in the economy, and since universities are increasingly<br />

commercial organizations, should art his<strong>to</strong>rians relocate themselves (as many museums<br />

have done) as part of the heritage industry (see<br />

), and produce cultural <strong>to</strong>urism<br />

courses for paying cus<strong>to</strong>mers as spin-offs from their student-directed university work?<br />

The wind of commerce hit museums harder than many universities. In some instances<br />

necessity occasioned the creation of good websites as the public face of the institution<br />

(interest logged via hits per week); in others it provoked a rethink of the mission of<br />

museums and a dumbing down <strong>to</strong> a lower common denomina<strong>to</strong>r, in displays, special<br />

exhibitions, and websites. Museums are in a state of flux because of computers and the<br />

Web: can/should technology be applied <strong>to</strong> these dusty mausolea of out-of-context<br />

artifacts <strong>to</strong> provide a learned, multimedia, immersive context – the nearest thing now<br />

possible <strong>to</strong> reintegrating the works with the context from which they were removed? Or is<br />

the museum of the future <strong>to</strong> be a theme park, lacking any truly educational mission, with<br />

Web displays as part of the fun? Believers in culture should hope for the former; but, as<br />

the inscription on Wren's <strong>to</strong>mb nearly has it, Si monumentum requiris, cinumclicke.<br />

Indeed, should we extrapolate from the existence of the Web <strong>to</strong> the decline of campusbased<br />

universities and physical museums and galleries? If galleries can only display a<br />

small proportion of their collection at any one time, should not funding go in<strong>to</strong> webbased<br />

image databases? And if seminars and tu<strong>to</strong>rials should remain because everyone<br />

recognizes the human and pedagogic value of face-<strong>to</strong>-face sessions, the fixed-term, fixedplace<br />

lecture might disappear in the face of the twin pressures of web-based profusely<br />

illustrated "lectures", and the increasing inability of students working <strong>to</strong> pay for their<br />

education <strong>to</strong> attend them. The challenge is <strong>to</strong> ensure that web-based "lectures" are of a


quality equivalent <strong>to</strong> or higher than traditional ones. We might hope that the <strong>digital</strong><br />

image, available well-catalogued and in profusion in a discipline for the development of<br />

which the pho<strong>to</strong>graph was partly responsible, will enhance art his<strong>to</strong>ry's popularity and<br />

underline the worldwide importance of culture in an age where networks reach around<br />

the world.<br />

Notes<br />

1 Thus the Institut de Recherche et d'His<strong>to</strong>ire des Textes published Le Médiéviste et<br />

I'Ordinateur, issues 26/7 of which (Autumn 1992-Spring 1993) were devoted <strong>to</strong> digitized<br />

images, "source essentielle pour l'his<strong>to</strong>rien et la plupart du temps difficile à consulter."<br />

2 Cf. Peter Denley and Deian Hopkin's His<strong>to</strong>ry and Computing (1987): the result of a<br />

conference in 1986 <strong>to</strong> inaugurate an association for his<strong>to</strong>rians who used or wished <strong>to</strong> use<br />

computers in their work, whether that work was research or teaching; or Jean-Philippe<br />

Genet's L!0rdinateur et le Métier d'His<strong>to</strong>rien (1990), from the conference at Talence,<br />

September 1989.<br />

3 See for a <strong>to</strong>ur d'horizon.<br />

4 Earlier, .<br />

5 Online at . Other Getty projects have<br />

included the Union List of Artist Names; Guide for the Description of Architectural<br />

Drawings and Archives; Anthology of Museum Au<strong>to</strong>mation; Bibliography of the His<strong>to</strong>ry<br />

of Art (BHA); The Provenance Index; The Avery Index <strong>to</strong> Architectural Periodicals; Witt<br />

Computer Index; Census of Antique Art and Architecture Known <strong>to</strong> the Renaissance.<br />

6 The International Journal of Museum Management and Cura<strong>to</strong>rship 8,1 (March 1989),<br />

is largely devoted <strong>to</strong> computerization; the False Gods of the Edi<strong>to</strong>rial's title refers <strong>to</strong> the<br />

process of intellectual cheapening, not computers.<br />

7 The SNS Centre di Elaborazione Au<strong>to</strong>matica di Dati e Documenti S<strong>to</strong>rico-Artistici<br />

produced a Bollettino d'Informazioni. Vol. 2 for 1981, for example, offered papers on<br />

computerization of a lexicon of goldsmiths' work; a report on a seminar <strong>to</strong> discuss<br />

digitization at the Bargello and Museo Stibbert in Florence, a catalogue of reused antique<br />

sarcophagi, and a report on the computerization of a Corpus dell'Arte Senese. For the<br />

later his<strong>to</strong>ry of this organization, cf. .<br />

8 See .<br />

9 See .<br />

10 Details in Baryla (1992–3: 23–4). NB: these were accompanied by a database.


11 See Debevec's notes for SIGGRAPH 99 Course no. 28, and cf.<br />

; a course at SIGGRAPH 2002 was devoted <strong>to</strong><br />

Recreating the Past: ;<br />

also Polleyes et al., Reconstruction Techniques with Applications in Archeology, at<br />

.<br />

12 Cf. Stenvert (1991), with its excellent bibliography.<br />

13 For example, or the list at<br />

.<br />

14 Cf. virtual library catalogues of art his<strong>to</strong>ry, such as that at<br />

with the interface<br />

available in English, French, German, and Italian.<br />

15 See Lancashire (1991: 9–17). These pages are devoted <strong>to</strong> art his<strong>to</strong>ry. This gives an<br />

indication of the relatively slow start of this image-based discipline vis-a-vis text-based<br />

ones.<br />

16 Romer, who had previously published a paper entitled A Keyword is Worth 1,000<br />

Images, correctly concluded that "image and multimedia databases are heavily dependent<br />

on the quality of their s<strong>to</strong>red descriptions" (1996: 55).<br />

17 Such as the Journal of Educational Multimedia and Hypermedia, or the CTI's Active<br />

Learning, no. 2, for July 1995, having the theme Using the Internet for Teaching.<br />

18 See also Andrews and Greenhalgh (1987) and Richard and Tiedemann's Internet für<br />

Kunsthis<strong>to</strong>riker: Eine praxisorientierte Einführung (1999).<br />

References for Further Reading<br />

AHWA-AWHA:CIHA London 2000. Section 23: Digital Art His<strong>to</strong>ry Time. Accessed:<br />

September 28, 2003. At http://www.unites.uqam.ca/AHWA/Meetings/2000.CIHA/.<br />

Andrews, Derek and Michael Greenhalgh (1987). Computing for Non-Scientific<br />

Applications. Leicester: Leicester University Press.<br />

Au<strong>to</strong>stereoscopic display project plan. Accessed September 28, 2003. At<br />

http://www.mrl.nyu.edu/~perlin/courses/fall98/projects.au<strong>to</strong>stereo.html.<br />

Baryla, C. (1992–93). Les vidéodisques de la Bibliothèque Vaticane [The videodisks of<br />

the Vatican Library]. Le Médiéviste et I'Ordinateur 26–7: 23–4.<br />

Bisogni, Fabio, (ed.) (1980). Me<strong>to</strong>dologie di Analisi e di Catalogazione dei bent<br />

Culturali[Methodologies for the analysis and cataloguing of cultural property]. In the


series Quaderni di Informatica e Beni Culturali, 2. Siena: Università degli Studi di<br />

Siena/Regione Toscana.<br />

Bisogni, Fabio (ed.), (ed.) (1981). Sistemi di Trattamen<strong>to</strong> di Dati e Immagini [The<br />

processing of data and images]. In the series Quaderni di Informatica e Beni Culturali, 4.<br />

Siena: Università degli Studi di Siena/Regione Toscana.<br />

Bisogni, Fabio and Laura Corti (1980). Manuals Tecnico per I'Elaborazione Au<strong>to</strong>matica<br />

di Dati S<strong>to</strong>rico-Artistici [Technical manual for the au<strong>to</strong>matic processing of art-his<strong>to</strong>rical<br />

data]. In the series Quaderni di Informatica e Beni Culturali, 1. Siena: Universita degli<br />

Studi di Siena/Regione Toscana.<br />

Bowen, Jonathan (2003). Virtual Museum of Computing. The Virtual Library Museums<br />

Pages. Accessed September 28, 2003. Available at: http://vmoc.museophile.com/.<br />

CENSUS: Census of Antique Works of Art and Architecture Known in the Renaissance.<br />

Accessed September 28, 2003. At http://bak-information.ub.tuberlin.de/fachinfo/DBANKEN/CENSUS.html.<br />

Cerri, Stefano A. (1983). Sistema esper<strong>to</strong> di elaborazione elettronica sui beni<br />

archeologici di Roma: Come un calcula<strong>to</strong>re puo usare competenze archeologica per<br />

rispondere a domande ingenue e complesse [Expert system for the processing of Roman<br />

archaeological artifacts: how a computer can exploit archaeological information <strong>to</strong><br />

respond <strong>to</strong> subtle and complex questions]. In the Conference Roma: Archeologia e<br />

Projet<strong>to</strong>, Campidoglio, Rome, pp. 23–28, 20.<br />

CHArt: Computers and the His<strong>to</strong>ry of Art. Accessed September 28, 2003. At<br />

http://www.chart.ac.uk/.<br />

Coalition for Networked Information (CNI). Humanities and Arts Report: Appendices.<br />

Last updated July 3, 2002. Accessed September 28, 2003. At<br />

http://www.cni.org/projects.humartiway/humartiway-rpt.append.html.<br />

Corti, Laura (1984). Census: Computerization in the His<strong>to</strong>ry of Art, I, Pisa and Los<br />

Angeles: Scuola Normale Superiore Pisa / J. Paul Getty Trust Los Angeles.<br />

Courtauld Institute of Art. University of London. Accessed September 28, 2003. At<br />

http://www.courtauld.ac.uk/.<br />

Crane, Gregory, (ed.) The Perseus Digital Library. Accessed September 28, 2003. At:<br />

http://www.perseus.tufts.edu/.<br />

Cribecu Online. Centre Ricerche Informatiche per i Beni Culturali. Accessed September<br />

28, 2003. At http://www.cribecu.sns.it/info/s<strong>to</strong>ria.html.


Daniels, Dieter Medien Künst Netz [Strategies of Interactivity], tr. Tom Morrison. At<br />

http://www.mediaartnet.org/Texteud.html.<br />

Debevec, Paul. Modelling and Rendering Architecture from Pho<strong>to</strong>graphs. Notes for<br />

SIGGRAPH 99 Course no. 28. Accessed September 28, 2003. At<br />

http://www.debevec.org/Research/.<br />

Denley, Peter, and Deian Hopkin, (eds.) (1987). His<strong>to</strong>ry and Computing. Manchester:<br />

Manchester University Press.<br />

Detroit Institute of Arts. DIA Visual Resource Home Page. Accessed September 28,<br />

2003. At http://www.diamondial.org/.<br />

di Bari, Vit<strong>to</strong>ria Carla et al. (1981). Elaborazione Au<strong>to</strong>matica dei Dati relativi alle<br />

Monete Romans dello Scavo di Campo all'Oro (Siena) [Computer processing of Roman<br />

coin data from the excavation at Campo all'Oro (Siena)]. In the series Quaderni di<br />

Informatica e Beni Culturali, 3. Siena: Universita degli Studi di Siena/Regione Toscana.<br />

Direction des Musées de France base Joconde. Accessed September 28, 2003. At<br />

http://www.culture.fr/documentation/joconde/pres.htm.<br />

Donahue, Joseph. The London Stage 1800–1900: A Documentary Record and Calendar<br />

of Performances. Accessed September 28, 2003. At http://wwwunix.oit.umass.edu/~aOfsOOO/lsp.html.<br />

Dota, Michele and Franco Niccolucci (1981). Elaborazione Au<strong>to</strong>matica dei Dati relativi<br />

ai Conventi del Centra S<strong>to</strong>rico di Firenze, Siena. [Computer processing of data<br />

concerning the convents of the his<strong>to</strong>rical centers of Florence and Siena]. In the series<br />

Quaderni di Informatica e Beni Culturali, 5. Siena: Università degli Studi di<br />

Siena/Regione Toscana.<br />

Durran, Jennifer. Developments in Electronic Image Databases for Art His<strong>to</strong>ry [1992–3].<br />

Last updated April 14, 2000. First published in: Drinking From A Fire Hose: Managing<br />

Networked Information Conference Preprints (pp. 86–96). Vic<strong>to</strong>rian Association for<br />

Library Au<strong>to</strong>mation 7th Biennial Conference and Exhibition, November 9–11, 1993.<br />

Accessed September 28, 2003. At<br />

http://www.arts.monash.edu.au/visual_culture/about/staff/jdurran/vala_report.html.<br />

Durran, Jennifer (1994). EVA: Electronic Imaging and the Visual Arts Conference,<br />

Tu<strong>to</strong>rials and Exhibition: A Review. Last updated April 14, 2000. First published:<br />

ARLIS/ANZ News 39, November 1994: 71–5. Accessed September 28, 2003. At<br />

http://www.arts.monash.edu.au/visual_culture/about/staff.jdurran/eva93.html.<br />

Düsseldorf. Die Diisseldorfer Virtuelle Bibliothek: Kunstwissenschaft. Last updated<br />

September 18, 2003. Accessed September 28, 2003. At http://www.rz.uniduesseldorf.de/WWW/ulb.kun.html.


eSpectra Features: March 1999 Accessed September 28, 2003. At<br />

http://www.mcn.edu/eSpectra/03_1999.html.<br />

Floud, Roderick (1979). An Introduction <strong>to</strong> Quantitative Methods for His<strong>to</strong>rians, 2nd<br />

edn. London: Methuen.<br />

Genet, Jean-Philippe (1990). L'Ordinateur et le métier d'his<strong>to</strong>rien. From the conference<br />

at Talence, September 1989. Bordeaux: Maison des Pays Ibériques.<br />

Getty Vocabulary Program. The J. Paul Getty Trust. Accessed September 28, 2003. At<br />

http://www.getty.edu/research/<strong>to</strong>ols/vocabulary/.<br />

Grau, Oliver. Ances<strong>to</strong>rs of the Virtual, His<strong>to</strong>rical Aspects of Virtual Reality and its<br />

Contemporary Impact. CIHA London, 2000. Section 23: Digital Art His<strong>to</strong>ry Time.<br />

Accessed September 28, 2003. At<br />

http://www.unites.uqam.ca/AHWA/Meetings/2000.CIHA/Grau.html.<br />

Hamber, Anthony (1991). The VASARI Project. Computers and the His<strong>to</strong>ry of Art 1, 2:<br />

3–19.<br />

Hamburg University. Schule des Sehens - Neue Medien der Kunstgeschichte. Accessed<br />

September 28, 2003. At http://www.unihamburg.de/Wiss/FB/09/KunstgeS/schulesehen.html.<br />

Heusser, Hans-Jörg, (ed.) (1987). Computers and the Future of Research: Visions,<br />

Problems, Projects. AICARC: Bulletin of the Archives and Documentation Centers for<br />

Modern and Contemporary Art 14,2/15,1.<br />

ICONCLASS. Accessed September 28, 2003. At http://www.iconclass.nl/.<br />

INIST-CNRS. Bibliographic de I'His<strong>to</strong>ire de I'Art [Bibliography of the His<strong>to</strong>ry of Art].<br />

Accessed September 28, 2003. At http://www.inist.fr/BHA/.<br />

Institut de Recherche et d'His<strong>to</strong>ire des Textes (1992–3). Le Médiéviste et I'Ordinateur<br />

26/7. International Journal of Museum Management and Cura<strong>to</strong>rship (1989). 8, 1.<br />

Lancashire, Ian, (ed.) (1991). The Humanities Computing Yearbook, 1989–90. Oxford:<br />

Clarendon Press.<br />

Lanier, Jaron (<strong>2001</strong>). Virtually There. Scientific American. Available at<br />

http://www.scientificamerican.com/<strong>2001</strong>/0401issue/9401lanier.html.<br />

Lavin, Marilyn Aronberg (1992). Researching Visual Images with Computer Graphics.<br />

Computers and the His<strong>to</strong>ry of Art 2, 2: 1–5.


Lavin, Marilyn Aronberg and Kirk Alexander. Piero Project: ECIT/Electronic<br />

Compendium of Images and Text. Accessed September 28, 2003. At<br />

http://etc.prince<strong>to</strong>n.edu/art430/art430.html.<br />

MCN (Museum Computer Network). Museum Sites Online. Accessed September 28,<br />

2003. At http://www.mcn.edu/resources/sitesonline/htm.<br />

Ministère de la culture. Documentation: Bases de données patrimoniales.<br />

[Documentation: national heritage databases]. Accessed September 28, 2003. At<br />

http://www.culture.fr/documentation/docum.htm.<br />

Ministère de la culture (1985). Mission de la Recherche [brochure]. Paris: Ministère de<br />

la culture.<br />

Naugh<strong>to</strong>n, John (1999). A Brief His<strong>to</strong>ry of the Future: The Origins of the Internet.<br />

London: Weidenfeld and Nicolson.<br />

Otey, Astrid. Visual Resources Association. Last updated September 9, 2003. Accessed<br />

September 28, 2003. At http://www.vraweb.org/.<br />

Polleyes, M., et al. Reconstruction Techniques with Applications in Archeology.<br />

Accessed September 28, 2003. At http://www.esat.kuleuven/ac.be/sagalassos/3-<br />

Dreconstruction/3-D.html.<br />

Richard, Birgit and Tiedemann, Paul (1999). Internet für Künsthis<strong>to</strong>riker: Eine<br />

praxisorientierte Einführung. Darmstadt.<br />

Roberts, D. Andrew, (ed.) (1990). Terminology for Museums. Proceedings of the<br />

International Conference, Cambridge, England, September 21–4, 1988.<br />

Romer, Donna M. (1996). Image and Multimedia Retrieval. In David Bearman et al.,<br />

Research Agenda for Networked Cultural Heritage (pp. 49–56). Santa Monica, CA:<br />

Getty AHIP.<br />

SIGGRAPH (2002). Course 27: Recreating the Past. Accessed September 28, 2003. At<br />

http://www.siggraph.org/s2002/conference/courses/crs27.html.<br />

Spaeth, Donald (1991). A Guide <strong>to</strong> Software for His<strong>to</strong>rians. Glasgow: CTI.<br />

Stenvert, Ronald (1991). Constructing the Past: Computer-assisted Architectural-<br />

His<strong>to</strong>rical Research: The Application of Image-processing using the Computer and<br />

Computer-Aided Design for the Study of the Urban Environment, Illustrated by the Use<br />

of Treatises in Seventeenth-century Architecture. The Hague: self-published in<br />

association with the Dutch Government Service for the Preservation of Monuments and<br />

His<strong>to</strong>ric Buildings.


Universität Trier. D. W. Dörrbecker: Veranstaltungsangebote im Wintersemester 2002–<br />

2003. Accessed September 28, 2003. At http://www.unitrier.de/uni/fb3/kunstgeschichte/2002/WS/dwd.html.<br />

Villard, Laurence (1984). Systèms descriptif des antiquités classiques. Paris: Réunion des<br />

Musées Nationaux.<br />

VRND: A Real-Time Virtual Reconstruction, Notre Dame Cathedral. Accessed<br />

September 28, 2003. At http://www.vrndproject.com/.<br />

4.<br />

Classics and the Computer: An End of the His<strong>to</strong>ry<br />

Greg Crane<br />

Many non-classicists from academia and beyond still express surprise that classicists<br />

have been aggressively integrating computerized <strong>to</strong>ols in<strong>to</strong> their field for a generation.<br />

The study of Greco-Roman antiquity is, however, a data-intensive enterprise. Classicists<br />

have for thousands of years been developing lexica, encyclopedias, commentaries,<br />

critical editions, and other elements of scholarly infrastructure that are best suited <strong>to</strong> an<br />

electronic environment. Classicists have placed great emphasis on systematic knowledge<br />

management and engineering. The adoption of electronic methods thus reflects a very old<br />

impulse within the field of classics. The paper knowledge base on Greco-Roman<br />

antiquity is immense and well organized; classicists, for example, established standard,<br />

persistent citations schemes for most major authors, thus allowing us <strong>to</strong> convert<br />

nineteenth-century reference <strong>to</strong>ols in<strong>to</strong> useful electronic databases. Classicists are thus<br />

well prepared <strong>to</strong> exploit emerging <strong>digital</strong> systems. For many classicists, electronic media<br />

are interesting not (only) because they are new and exciting but because they allow us <strong>to</strong><br />

pursue more effectively intellectual avenues than had been feasible with paper. While<br />

many of us compare the impact of print and of new electronic media, classicists can see<br />

the impact of both revolutions upon the 2,500-year his<strong>to</strong>ry of their field. Electronic media<br />

thus allow us <strong>to</strong> pursue both our deepest and most firmly established scholarly values and<br />

challenge us <strong>to</strong> rethink every aspect of our field.<br />

The title of this chapter alludes <strong>to</strong> a 1993 article, "Classics and the Computer: the<br />

His<strong>to</strong>ry", in which Theodore Brunner, the founder of the Thesaurus Linguae Graecae<br />

(TLG), described the development of computing in the classics up <strong>to</strong> the end of the<br />

1980s. Those wishing an account of the first generation of computer-based work in<br />

classics will find the s<strong>to</strong>ry well documented in Brunner's work. This chapter takes a<br />

broader, more schematic, and more tendentious approach. To some extent, this approach<br />

is reactive: the rise of the personal computer and, more recently, of the World Wide Web<br />

has diffused computation throughout the daily life of scholars, and especially of students.<br />

A Foucauldian scholar might compare the shift from a few very specialized projects and<br />

extremely expensive hardware <strong>to</strong> the ubiqui<strong>to</strong>us e-mail viewers, web browsers, word


processing systems, bibliographic database systems, etc., <strong>to</strong> the shift from the spectacle of<br />

the sovereign <strong>to</strong> the diffused microphysics of modern power. New computer-based<br />

projects continue <strong>to</strong> emerge; the December 2002 Ancient Studies – New Technology<br />

conference at Rutgers hosted presentations on roughly thirty computer-based projects<br />

about the ancient world. Perhaps more significant, even superficially traditional<br />

publications distributed even now in traditional format (whether paper-only publication<br />

or as electronic files that mimic paper presentation) depend upon an infrastructure that is<br />

almost entirely electronic. Digital systems have quietly become the norm. A follow-on <strong>to</strong><br />

Brunner's his<strong>to</strong>ry would lead far beyond the growing but much more tractable body of<br />

work that Brunner confronted more than a decade ago.<br />

The title of this chapter reflects an argument as well as a defensive strategy. There should<br />

not be a his<strong>to</strong>ry of classics and the computer, for the needs of classicists are simply not so<br />

distinctive as <strong>to</strong> warrant a separate "classical informatics." Disciplinary specialists<br />

learning the strengths and weaknesses have, in the author's experience, a strong tendency<br />

<strong>to</strong> exaggerate the extent <strong>to</strong> which their problems are unique and <strong>to</strong> call for a specialized,<br />

domain-specific infrastructure and approach. Our colleagues in the biological sciences<br />

have been able <strong>to</strong> establish bioinformatics as a vigorous new field – but the biologists can<br />

bring <strong>to</strong> bear thousands of times more resources than can classicists. A tiny field such as<br />

classics, operating at the margins of the <strong>humanities</strong>, cannot afford a distinctive and<br />

au<strong>to</strong>nomous his<strong>to</strong>ry of its own. For classicists <strong>to</strong> make successful use of information<br />

technology, they must insinuate themselves within larger groups, making allies of other<br />

disciplines and sharing infrastructure. For classicists, the question is not whether they can<br />

create a classical informatics but whether such broad rubrics as computational<br />

<strong>humanities</strong>, computing in the <strong>humanities</strong>, cultural informatics, and so on, are sufficient or<br />

whether they should more aggressively strive <strong>to</strong> situate themselves within an informatics<br />

that covers the academy as a whole. Even academic technology, strictly defined, may be<br />

<strong>to</strong>o limiting, since the most revolutionary information technology for twenty-first-century<br />

philologists may well emerge from programs such as TIDES (Translingual Information<br />

Detection Extraction Summarization), supported by the US military.<br />

But much as some of us may struggle <strong>to</strong> minimize the future his<strong>to</strong>ry of classical<br />

computing as a separate movement, classicists have had <strong>to</strong> forge their own his<strong>to</strong>ry. Even<br />

in the best of futures, where classicists cus<strong>to</strong>mize general <strong>to</strong>ols and share a rich<br />

infrastructure with larger disciplines, classicists will have <strong>to</strong> struggle mightily for their<br />

voices <strong>to</strong> be heard so that emerging systems and standards meet their needs. This chapter<br />

seeks <strong>to</strong> explain why classicists needed <strong>to</strong> create this separate his<strong>to</strong>ry in the past and <strong>to</strong><br />

ground current trends in a larger his<strong>to</strong>rical context. It is relatively easy <strong>to</strong> pass judgments<br />

on the past, but in using the past <strong>to</strong> argue for trends in the future this chapter justifies any<br />

unfairness in its own perfect hindsight by exposing its own predictions for the future.<br />

The following broad movements – which overlap with one another and all continue <strong>to</strong> the<br />

present – provide one way of looking at the his<strong>to</strong>ry of classical computing. First,<br />

beginning with Father Rober<strong>to</strong> Busa's concordance of Aquinas's Latin writing in the late<br />

1940s, standard machines were used <strong>to</strong> produce particular projects (especially<br />

concordances such as Packard's Livy Concordance). Second, beginning at roughly 1970,


large, fairly centralized efforts <strong>to</strong>ok shape that sought <strong>to</strong> address major issues of classical<br />

computing infrastructure. These projects included David Packard's Ibycus system, the<br />

TLG, the Database of Classical Bibliography, the Bryn Mawr Classical Review, the Duke<br />

Databank of Documentary Papyri, and the Perseus Project. Third, beginning in the mid-<br />

1980s, the rise of the personal computer distributed computing throughout classics and,<br />

indeed, the <strong>humanities</strong>. Fourth, beginning in the mid-1990s, the rise of the Web spurred a<br />

vast outpouring of smaller projects in classics as in other fields.<br />

The following section takes one particular point in time twenty years ago (1982–3). The<br />

selection reflects the situation that existed when the author of this piece began his own<br />

work, but focusing on this point in time provides a brief case study with the particular<br />

illustrating more general conditions. The subsequent section provides a schematic view of<br />

classical computing. The conclusion briefly suggests how trends rooted in the past may<br />

carry us forward in<strong>to</strong> the future.<br />

Cincinnati 1983: Retrospectives and Prospectives<br />

In the American Philological Association convention in 1983, held at Cincinnati, a<br />

crowded session considered the merits of two very different approaches <strong>to</strong> classical<br />

computing. I was part of a group describing a Unix-based approach that we had begun<br />

developing at the Harvard Classics Department in the summer of 1982. David Packard,<br />

who was then, and remains <strong>to</strong>day, arguably the most significant figure in classical<br />

computing, was prepared <strong>to</strong> explain the rationale for his own Ibycus system, which he<br />

had spent years developing from the ground up <strong>to</strong> serve the needs of classicists. Then a<br />

graduate student and very conscious of my junior position, I was very nervous – if there<br />

was a tradition of young entrepreneurial leadership in technology, I had certainly not<br />

encountered it as a classicist. The underlying issues remain interesting <strong>to</strong>pics of debate:<br />

how far do we follow generic standards, and at what point do the benefits of<br />

specialization justify departures from broader practice?<br />

Looking back after twenty years, we could argue that the positions we espoused had<br />

prevailed. The Ibycus minicomputer system is long gone and its successor, the Ibycus<br />

Scholarly Computer (SC) PC system, is no longer under development (although a few<br />

venerable Ibycus SCs continue <strong>to</strong> serve dedicated users <strong>to</strong> this day). The benefits which<br />

my colleagues and I placed on standards have, <strong>to</strong> some extent, proven themselves: the<br />

10,000 lines of source code, written in the C programming language under Unix, which<br />

provided an efficient searching environment for Greek and other languages, still compiles<br />

and can run on any Unix system (including Linux and OS X) – it would not be possible <strong>to</strong><br />

buy <strong>to</strong>day a computer that was less powerful than the fastest systems <strong>to</strong> which we had<br />

access twenty years ago. The Thesaurus Linguae Graecae itself now uses a Unix server<br />

<strong>to</strong> provide the core string-searching operations on which both Packard and my group<br />

were working twenty years ago.<br />

Such a triumphalist retrospective upon the past would, however, be unjust and inaccurate.<br />

First, systems are only a means <strong>to</strong> an end. David Packard opened his presentation in<br />

December 1983 with an observation that put our discussions in<strong>to</strong> perspective and that has


shaped my own decision making ever since. He observed that software and systems were<br />

ephemeral but that primary sources such as well structured, cleanly entered source texts<br />

were objects of enduring value. In fact, the TLG had released a core set of Greek authors<br />

in machine-readable form and our work concentrated above all on providing textsearching<br />

facilities for this collection. Word processing and typesetting were major<br />

advantages of the new technology and the decrease in per page typesetting costs helped<br />

justify the considerable expense of any systems – Unix, Ibycus, or other – at the time.<br />

The TLG had begun work more than a decade before in 1972. The TLG began creating its<br />

<strong>digital</strong> library of classical Greek source texts by using the standard <strong>to</strong>ols available from<br />

the UC Irvine computer center, but the problems of entering, formatting, and verifying<br />

Greek texts were very different from those of the number-crunching experiments and<br />

administrative databases for which those <strong>to</strong>ols were developed. Packard's Ibycus system<br />

provided an environment far more suited <strong>to</strong> their needs than anything else available.<br />

Packard had gone so far as <strong>to</strong> modify the microcode of the Hewlett Packard<br />

minicomputer <strong>to</strong> enhance the speed of text searching. He created an entire environment,<br />

from the operating system up through a high-level programming language, aimed initially<br />

at serving the needs of classicists. Decades later, the boldness and achievement of<br />

creating this system seems only greater <strong>to</strong> this observer. It would have been immensely<br />

more difficult for the TLG - and many other smaller projects (such as the Duke Databank<br />

of Documentary Papyri and Robert Kraft's Septuagint Project) – <strong>to</strong> have been as<br />

successful. Without the Ibycus environment, few departments would have been able <strong>to</strong><br />

justify the use of computers in the 1970s or early 1980s. Nor might the field have been<br />

able <strong>to</strong> provide the National Endowment <strong>to</strong> the Humanities funders with the same level of<br />

support for the TLG.<br />

The early 1980s represented a tipping point, for at that time new systems were emerging<br />

that would provide inexpensive and, even more important, relatively stable platforms for<br />

long-term development. The Unix operating system, C programming language, and<br />

similar resources provided <strong>to</strong>ols that were independent of any one commercial vendor<br />

and that have continued <strong>to</strong> evolve ever since. MS DOS and the IBM PC appeared in 1981<br />

– before most of our current undergraduates were born. The Macin<strong>to</strong>sh (which has now<br />

built its current operating system on a Berkeley Unix base) appeared in 1984 – at the<br />

same time as many students who entered American universities as freshmen in 2002.<br />

New languages such as Java and Perl have emerged, and web browsers have provoked a<br />

substantially new model of cross-platform interaction, but <strong>to</strong>ols developed under Unix<br />

twenty years ago can still run. They may be abandoned, but only if better <strong>to</strong>ols have<br />

emerged – not because the systems in which they were created are gone and they need <strong>to</strong><br />

be, at the least, reconstituted.<br />

But even if the early 1980s represented the beginning of a new phase in the evolution of<br />

<strong>digital</strong> technology, the gap between our needs as classicists and the infrastructure at hand<br />

remained immense. The labor needed <strong>to</strong> adapt existing infrastructure <strong>to</strong> our needs was not<br />

so daunting as that which faced David Packard in building the Ibycus, but it was<br />

substantial.


Two classes of problems faced us, and they continue <strong>to</strong> face us even now. The first, <strong>to</strong><br />

which I will return, consisted of tangible problems that we needed <strong>to</strong> solve. The second,<br />

however, arose from the gap in understanding between those of us who were new <strong>to</strong> the<br />

technology and our technological colleagues who were innocent of classics. Because we<br />

were unable <strong>to</strong> communicate our real needs, we made serious initial errors, investing<br />

heavily in <strong>to</strong>ols that seemed suitable but proved, on closer examination, <strong>to</strong> be<br />

fundamentally unable <strong>to</strong> do what we needed. While many of the problems that we faced<br />

then have resolved themselves, the general problem remains: the biggest government<br />

funders of academic technology are the National Institutes of Health and the National<br />

Science Foundation whose aggregate funding ($20 billion and $5 billion respectively)<br />

exceeds that of the National Endowment for the Humanities ($135 million requested for<br />

2003) by a fac<strong>to</strong>r of 185. The academic technology specialists in higher education surely<br />

devote at least 1 percent of their time <strong>to</strong> the <strong>humanities</strong>, but the staggering disparity in<br />

support – governmental and private – for science, technology, and medicine means that<br />

the <strong>humanities</strong> are trapped at the margins of decision making. If our needs require<br />

substantial added investment beyond those of our colleagues outside the <strong>humanities</strong> (not<br />

<strong>to</strong> mention classics in particular), we will have great difficulties. We must be proactive<br />

and influence the shape of information technology as early as possible, tirelessly<br />

exploring common ground with larger disciplines and taking responsibility for pointing<br />

out where our challenges do, in fact, overlap with those of our colleagues from beyond<br />

the <strong>humanities</strong>.<br />

Our lack of sophistication, which became clear when Harvard began its own computing<br />

project in the summer of 1982, had one advantage. If we had been more knowledgeable,<br />

the department probably would not have moved forward but would have waited for<br />

technology <strong>to</strong> advance further. Instead, the department invested so many resources that<br />

we could not easily pull back. In the end, our work on the specialized problems of<br />

classical typesetting helped lower the publication costs of Harvard Studies in Classical<br />

Philology and the Loeb Classical Library, thus providing a justification for the initial<br />

investment. But if the results were ultimately satisfac<strong>to</strong>ry, the process was deeply flawed,<br />

and we were profoundly lucky <strong>to</strong> enjoy as much success as we did. Many other projects<br />

in classics and throughout the academy have faced similar problems and not been so<br />

fortunate.<br />

Several months of intensive work gave us a clearer idea of the challenges that we faced in<br />

1982.1 offer the following as a representative survey <strong>to</strong> document the state of the art at<br />

the time.<br />

Computer power and s<strong>to</strong>rage<br />

In 1965, Gordon E. Moore observed that the number of transis<strong>to</strong>rs that could be s<strong>to</strong>red<br />

per unit area had been doubling since the transis<strong>to</strong>r was invented and he argued that the<br />

trend would continue for the foreseeable future. The pace has slowed a bit – density has<br />

doubled every year and a half- but the exponential change has continued. Thus, twenty<br />

years ago we felt ourselves able <strong>to</strong> control staggering computational resources and looked<br />

upon the previous years with satisfaction. Now, of course, the subsequent twenty years


have made our initial systems appear primitive. The Psychology Department's Digital<br />

Equipment Corporation PDB 11/44 on which we did our work had far less computational<br />

power than the smallest desk<strong>to</strong>p machine now available, but served dozens of users<br />

typing in manuscripts or conducting experiments. For us as classicists, disk s<strong>to</strong>rage was a<br />

crucial issue. Modern disks allowed us <strong>to</strong> imagine keeping vast libraries – dozens of<br />

megabytes – of text online for casual searching and analysis. The Psychology lab at the<br />

time had two 80-megabyte hard drives, each the size of a small washing machine. The<br />

Harvard Classics Department needed more s<strong>to</strong>rage <strong>to</strong> mount the TLG and purchased the<br />

largest machine then generally available. The Control Data Corporation disk held 660<br />

megabytes – four times the s<strong>to</strong>rage of both disks already installed. It cost $34,000<br />

(including our educational discount) and had a service contract of $4,000 per year. The<br />

disk arrived in a crate that we had <strong>to</strong> pry open. We needed a special-purpose disk<br />

controller ($2,000). Worst of all, we had <strong>to</strong> write a software driver <strong>to</strong> mount this disk,<br />

hacking the source code <strong>to</strong> the Berkeley Unix system. It <strong>to</strong>ok months before we could<br />

exploit more than a tiny fraction of the disk s<strong>to</strong>rage on this heavy, loud, expensive device.<br />

Our colleagues in the Ibycus world chose different hardware (at the time, the largest<br />

Ibycus systems had, if I recall correctly, 400-megabyte drives) but the basic parameters<br />

were the same for all of us. Disk s<strong>to</strong>rage was cumbersome and expensive. As I write this,<br />

the smallest disk that I can find contains 20 gigabytes (30 times as much as our CDC<br />

behemoth) and costs $200 (150 times less). The price/performance ratio has thus<br />

increased by a fac<strong>to</strong>r of c.45,000. This does not even consider the fact that this 20gigabyte<br />

drive plugs directly in<strong>to</strong> a computer without modification and that it fits in a<br />

notebook.<br />

Machines have grown so fast and inexpensive that it is perhaps difficult for most of us <strong>to</strong><br />

imagine the extent <strong>to</strong> which hardware constrained the way we designed systems and thus<br />

the questions that we could pursue. The extraordinarily high cost of disk s<strong>to</strong>rage meant<br />

that Packard chose not <strong>to</strong> create indices for the TLG. All searches read through the texts<br />

from start <strong>to</strong> finish. Packard modified the microcode of the HP minicomputer <strong>to</strong> increase<br />

the search speed, thus providing another reason <strong>to</strong> build an entirely new operating<br />

system. Limitations on s<strong>to</strong>rage meant that <strong>digital</strong> images of any kind were impractical.<br />

The first TLG texts that we received at Harvard were in a compressed format that reduced<br />

s<strong>to</strong>rage by c.2Q percent. We had <strong>to</strong> write a program <strong>to</strong> decompress the files – the<br />

program was short but we had <strong>to</strong> write one ourselves as none existed, and this simply<br />

added <strong>to</strong> the overhead of working with the TLG documents.<br />

Greek display<br />

The graphical displays which we now take for granted were not in standard circulation.<br />

Displays were monochrome terminals that could cope with letters but not drawings, much<br />

less color. Even displaying textual data posed major technical barriers. The TLG had<br />

already developed an excellent ASCII encoding scheme for classical Greek and we had<br />

no scientific reason <strong>to</strong> use anything other than BETA Code, but our colleagues insisted<br />

that they needed <strong>to</strong> see fully accented classical Greek. We thus had <strong>to</strong> devote substantial<br />

energy <strong>to</strong> the font problem – some of us who worked on Greek fonts in the period still<br />

view "font" as the one most disturbing four-letter word in English.


To display Greek, we needed <strong>to</strong> use special terminals that could display cus<strong>to</strong>mized<br />

character sets. We designed Greek fonts on graph paper, converted the dot patterns in<strong>to</strong><br />

hexadecimal codes, programmed the data on <strong>to</strong> chips and then physically inserted these<br />

chips in<strong>to</strong> the displays. The moni<strong>to</strong>rs that we used cost $1,700 each and provided users<br />

with shared access <strong>to</strong> the overtaxed minicomputers in the psychology department.<br />

Networks<br />

The Internet was still tiny, connecting a few key research institutions over a relatively<br />

slow network. When we first began work in 1982, we had access <strong>to</strong> no machine-<strong>to</strong>machine<br />

networking other than log-ins via dial-up modems. To provide our colleagues<br />

access <strong>to</strong> the machine in the William James building, we purchased massive spools with<br />

thousands of feet of twisted pair cable and then ran this cable through a network of steam<br />

tunnels and conduits <strong>to</strong> Widener Library and other buildings. We did have e-mail within<br />

our single machine but we only gained access <strong>to</strong> inter-machine e-mail when we joined a<br />

network called UUCP. At the time, the machine would effectively grind <strong>to</strong> a halt every<br />

time it processed mail. Nor was mail a very effective <strong>to</strong>ol, since we knew very few<br />

people outside of our own group who had e-mail accounts. Simple file transfers were<br />

beyond us and the current Internet would have seemed outlandish: experience near the<br />

bleeding of technology tends <strong>to</strong> generate a schizophrenic attitude, alternating between the<br />

visionary and the cynical.<br />

Multilingual text editing<br />

David Packard developed an edi<strong>to</strong>r for the Ibycus that could manage Greek and English.<br />

When we first considered working with Unix our best consultant suggested that it would<br />

take twenty minutes <strong>to</strong> modify the source code for the standard Unix text edi<strong>to</strong>r (Vi) <strong>to</strong><br />

handle Greek. In fact, the Unix text edi<strong>to</strong>r assumed an ASCII character set and would<br />

have required a complete rewrite <strong>to</strong> manage any other character sets. We spent a good<br />

deal of time trying <strong>to</strong> develop a multilingual text edi<strong>to</strong>r. We had made good progress<br />

when the Macin<strong>to</strong>sh arrived. The Macin<strong>to</strong>sh knew nothing about languages at the time,<br />

but it unders<strong>to</strong>od fonts, and multiple fonts were enough <strong>to</strong> serve the basic needs of<br />

classicists. We abandoned our edi<strong>to</strong>r and resolved never <strong>to</strong> address a general problem that<br />

the marketplace would solve for us.<br />

Text retrieval<br />

Pioneers such as Gerald Sal<strong>to</strong>n had laid the foundation for the science of information<br />

retrieval in the 1950s and 1960s. Techniques already existed <strong>to</strong> provide efficient<br />

searching of textual databases. The <strong>to</strong>ols at our disposal were also powerful and flexible.<br />

Unix provided a superb scripting language and development environment within which <strong>to</strong><br />

reuse existing programs. Unix provided several text-searching programs (the infamously<br />

non-mnemonic grep, egrep and fgrep). We had the source code for everything and could<br />

thus modify any existing Unix program. Searching the TLG and other scholarly textual<br />

databases should have been easy.


In fact, it proved quite difficult <strong>to</strong> build services that our colleagues would actually use on<br />

the generic <strong>to</strong>ols of Unix. Three issues confronted us. First, we needed a reasonable<br />

interface. Few classicists even <strong>to</strong>day have proven willing <strong>to</strong> learn how <strong>to</strong> use a Unix<br />

environment directly. Since we were working more than a decade before web browsers<br />

radically reduced the labor needed <strong>to</strong> create simple interfaces, this task required both<br />

programming and design work.<br />

Second, we needed <strong>to</strong> generate standard citations for the search results. The Unix search<br />

utilities returned lines that had a particular pattern. The TLG files had a complex scheme<br />

for encoding changes in book, chapter, section, or line numbers. The search program had<br />

<strong>to</strong> examine each line in a file and update the various registers, as well as deal with the<br />

exceptions (e.g., line 178a, line 38 appearing before line 34, etc.) that have found their<br />

way in<strong>to</strong> our texts. Adding such routines <strong>to</strong> the fine-tuned text-scanning modules of the<br />

Unix search routines proved non-trivial and would have required virtual rewrites.<br />

Third, speed was an issue. The systems <strong>to</strong> which we had access were <strong>to</strong>o slow. We<br />

learned quickly why David Packard had modified the microcode of this HP computer <strong>to</strong><br />

increase linear search speeds. Ultimately, we were able <strong>to</strong> match the linear search speeds<br />

on the Ibycus on DEC VAX computers by rewriting the core search loop in VAX<br />

assembly language so that we could utilize a special pattern-matching language. Of<br />

course, the VAX computers were more expensive (and more powerful) than the HP<br />

computers. Also, while classics departments owned Ibycus computers, we shared all of<br />

our machines with many other users – and others did not appreciate our clogging the<br />

system with searches that slowed down the disks and the CPU alike.<br />

Fourth, searching classical Greek raises two problems. First, classical Greek has a<br />

complex system of accentuation, with the accent shifting around the word as inflections<br />

vary. Thus, searches need <strong>to</strong> be able <strong>to</strong> ignore accents and scan for underlying stems: e.g.,<br />

searching for forms of the verb π µπω ("<strong>to</strong> send"), we need <strong>to</strong> match "' πεµπoν" and "π<br />

µπεις" We can write regular expressions <strong>to</strong> accommodate this or simply search for "πεµπ"<br />

or "π µπ" but such permutations can become complex and require knowledge of vowel<br />

length and other features of the language. More significantly, Greek is a highly inflected<br />

language. The search <strong>to</strong>ols developed under Unix implicitly assumed English – with its<br />

minimal system of inflections – as its model. The <strong>to</strong>ols at our disposal simply were not<br />

designed for a language in which a single verb can have a thousand different forms.<br />

Ultimately, we developed a multilingual full text retrieval system from scratch. The<br />

system used a set of inverted indices, added 50 percent <strong>to</strong> the s<strong>to</strong>rage needs of the TLG (a<br />

significant fac<strong>to</strong>r then), but provided almost instantaneous lookups. The system<br />

comprised more than 15,000 lines of code when completed and provided a reasonable<br />

TLG solution for a decade, finally yielding <strong>to</strong> personal computer-based programs <strong>to</strong><br />

search the subsequent TLG CDs.<br />

From 1983 <strong>to</strong> 2003: Past Trends and Prospects


The details listed in the section above provide an insight in<strong>to</strong> one particular point in time.<br />

The problems described above warrant documentation in part precisely because it is hard<br />

<strong>to</strong> remember now what barriers they posed. Looking back over the past twenty years, the<br />

following broad themes stand out.<br />

Increasingly powerful hardware<br />

Moore's law continues <strong>to</strong> hold. Even if technology were <strong>to</strong> freeze at 2003 levels, we<br />

would still need a generation <strong>to</strong> digest its implications. The cost of s<strong>to</strong>ring textual<br />

databases such as the TLG is almost zero. We can s<strong>to</strong>re hundreds of thousands of images,<br />

vast geographic datasets and anything that was published in print.<br />

Visualizations<br />

These include not only virtual reality displays and geographic information systems but<br />

also au<strong>to</strong>matically generated timelines and other data visualization techniques. The result<br />

will be possibilities for more holistic analysis of the Greco-Roman world, with<br />

philologists making much more effective use of art and archaeological materials than<br />

before.<br />

Language technologies<br />

I single out the broad class of "language technologies", a rubric that includes machine<br />

translation, cross-lingual information retrieval (e.g., type in "guest friend" and locate<br />

passages in Greek, Latin, Arabic, Sanskrit, etc.), summarization, clustering, syntactic<br />

analysis (and the possibility of generating large syntactic databases for Greco-Roman<br />

source texts), etc. The US defense establishment is investing heavily in rapidly<br />

deployable <strong>to</strong>ols for "low-density languages" (e.g., languages for which few if any<br />

computational resources exist) – intelligence analysts have found themselves compelled<br />

<strong>to</strong> develop capabilities in languages such as Albanian and Pashtun. The underlying<br />

resources for these techniques are bilingual text corpora, morphological analyzers, online<br />

lexica, grammars and other knowledge sources. Classicists already have these resources<br />

online, and other languages promise <strong>to</strong> follow suit. The next twenty years promise <strong>to</strong><br />

introduce a golden age of philology, in which classicists not only explore new questions<br />

about Greek and Latin but also explore corpora in many languages which they will never<br />

have the opportunity <strong>to</strong> master.<br />

Annotation managers<br />

Classicists have a long tradition of standalone commentaries and small notes on<br />

individual words and passages. All major literary classical Greek source texts are<br />

available from the TLG and many from Perseus. Authors can already publish annotations<br />

directly linked <strong>to</strong> the passages that readers see (rather than buried in separate<br />

publications). The hypertex-tual nature of web reading is stimulating new <strong>to</strong>ols and new<br />

opportunities with classicists, who can bring online a rich tradition of annotations.


Rise of library reposi<strong>to</strong>ries<br />

The World Wide Web spurred a generation of pseudo-publication: documents more<br />

broadly available than any print publication in his<strong>to</strong>ry could at any given time reach<br />

millions of machines. The same documents often ran, however, under individual<br />

accounts, with many URLs being changed or pointing <strong>to</strong> documents that were no longer<br />

online or, arguably worse, that had been substantively changed since the original link had<br />

been added. A variety of library reposi<strong>to</strong>ries are now coming in<strong>to</strong> use 1 . Their features<br />

differ but all are dedicated <strong>to</strong> providing long-term s<strong>to</strong>rage of core documents and <strong>to</strong><br />

separating authors from preservation. In the world of publication, alienation is a virtue,<br />

because in alienating publications, the author can entrust them <strong>to</strong> libraries that are<br />

designed <strong>to</strong> provide stable access beyond the lifespan of any one individual. Unless we<br />

transfer stewardship – and control – of our work at some point, then our work will not<br />

outlive us.<br />

Convergence of needs<br />

The examples listed above reflect a common theme: as our computing infrastructure<br />

grows in power, the generality of the <strong>to</strong>ols developed increases and the degree <strong>to</strong> which<br />

classicists (and other humanists) need <strong>to</strong> cus<strong>to</strong>mize general <strong>to</strong>ols becomes more defined.<br />

Where Packard had <strong>to</strong> create a whole operating system, we were, twenty years ago, able<br />

<strong>to</strong> build on Unix. Where we needed <strong>to</strong> work on our own multilingual text edi<strong>to</strong>r, Unicode<br />

provides multilingual support now at the system level. The set of problems particular <strong>to</strong><br />

classicists is shrinking. We are better able now than ever before <strong>to</strong> share infrastructure<br />

with our colleagues not only in the <strong>humanities</strong> but in the rest of the academy as well. The<br />

rising NSF-sponsored National Science Digital Library (NSDL) will, if it is successful,<br />

probably establish a foundation for the integration of academic resources across the<br />

curriculum. Nevertheless, classicists need <strong>to</strong> reach out <strong>to</strong> their colleagues and <strong>to</strong> begin<br />

influencing projects such as the NSDL if these science-based efforts are <strong>to</strong> serve our<br />

needs in the future. Otherwise, we may find that simple steps that could radically improve<br />

our ability <strong>to</strong> work in the future will have been overlooked at crucial points in the coming<br />

years. Our his<strong>to</strong>ry now lies with the larger s<strong>to</strong>ry of computing and academia in the<br />

twenty-first century.<br />

Notes<br />

1 Among the best known are D-Space (at ) and FEDORA (at<br />

).<br />

References for Further Reading<br />

Ancient Studies – New Technology, conference held at Rutgers University, December<br />

2002. At http://tabula.rutgers.edu/conferences/ancient_studies2002/. Accessed April 5,<br />

2004.


Brunner, T. F. (1993). Classics and the Computer: The His<strong>to</strong>ry. In J. Solomon (ed.),<br />

Accessing Antiquity: The Computerization of Classical Databases (pp. 10–33). Tucson:<br />

University of Arizona Press.<br />

The Bryn Mawr Classical Review. Accessed April 5, 2004. At<br />

http://ccat.sas.upenn.edu/bmcr/.<br />

Busa, R. (1949). La terminologia <strong>to</strong>mistica dell'interiorita; saggi di me<strong>to</strong>do per<br />

un'interpretazione della metafisica della presenza (p. 279). Milano: Fratelli Bocca.<br />

The Database of Classical Bibliography. Accessed April 5, 2004. At<br />

http://www.library.usyd.edu.au/databases/deb.html.<br />

The Duke Databank of Documentary Papyri. Accessed April 5, 2004. At<br />

http://scrip<strong>to</strong>rium.lib.duke.edu/papyrus/texts/DDBDP.html.<br />

Packard, D. W. (1968). A Concordance <strong>to</strong> Livy. Cambridge, MA: Harvard University<br />

Press.<br />

The Perseus Project. Accessed April 5, 2004. At http://www.perseus.tufts.edu.<br />

Thesaurus Linguae Graecae Project. Accessed April 5, 2004. At http://www.tlg.uci.edu/.<br />

TIDES (Translingual Information Detection Extraction Summarization). Accessed April<br />

5, 2004. At http://tides.nist.gov/.<br />

5.<br />

Computing and the His<strong>to</strong>rical Imagination<br />

William G.Thomas, II<br />

In the late 1960s and early 1970s his<strong>to</strong>rians seemed <strong>to</strong> think that their profession, the craft<br />

and art of his<strong>to</strong>ry itself, was on the brink of change. Everywhere one looked the signs<br />

were unmistakable. A kind of culture war broke out in the profession and a flurry of tense<br />

conference panels, public arguments, and roundtables <strong>to</strong>ok place with subtitles, such as<br />

"The Muse and Her Doc<strong>to</strong>rs" and "The New and the Old His<strong>to</strong>ry." This culture war pitted<br />

the "new" his<strong>to</strong>ry, largely influenced by social science theory and methodology, against<br />

the more traditional practices of narrative his<strong>to</strong>rians. The "new" his<strong>to</strong>rians used<br />

computers <strong>to</strong> make calculations and connections never before undertaken, and their<br />

results were, at times, breathtaking. Giddy with success, perhaps simply enthusiastic <strong>to</strong><br />

the point of overconfidence, these his<strong>to</strong>rians saw little purpose in anyone offering<br />

resistance <strong>to</strong> their findings or their techniques. When challenged at a conference, more<br />

than one his<strong>to</strong>rian responded with nothing more than a mathematical equation as the<br />

answer. Computers and the techniques they made possible have over the years altered


how many his<strong>to</strong>rians have unders<strong>to</strong>od their craft. To some they have opened the his<strong>to</strong>rical<br />

imagination <strong>to</strong> new questions and forms of presentation, while <strong>to</strong> others they have instead<br />

shuttered the his<strong>to</strong>rical imagination, at best limiting and channeling his<strong>to</strong>rical thinking<br />

and at worst confining it <strong>to</strong> procedural, binary steps. This chapter traces where the<br />

his<strong>to</strong>rical profession has come in the years since these professional debates and tries <strong>to</strong><br />

assess how computing technologies have affected the discipline and how they will shape<br />

its future scholarship.<br />

Not all his<strong>to</strong>rians in the 1960s and 1970s considered the computer the future of the<br />

profession. One railed against worshiping "at the shrine of that bitch goddess<br />

QUANTIFICATION." Another, Arthur Schlesinger, Jr., one of America's foremost<br />

his<strong>to</strong>rians, wrote in reply <strong>to</strong> the rising confidence in cliometrics and quantitative his<strong>to</strong>ry,<br />

"Almost all important questions are important precisely because they are not susceptible<br />

<strong>to</strong> quantitative answers" (Swierenga 1970: 33). Other critics of quantitative methods <strong>to</strong>ok<br />

aim not at the methods these his<strong>to</strong>rians used but instead questioned whether the "armies"<br />

of graduate students needed <strong>to</strong> develop large-scale projects and the technicians <strong>to</strong><br />

maintain them were worth the cost. One prominent British social his<strong>to</strong>rian, Lawrence<br />

S<strong>to</strong>ne, disparaged the contributions of the new his<strong>to</strong>ry and the costs it entailed: "It is just<br />

those projects that have been the most lavishly funded, the most ambitious in the<br />

assembly of vast quantities of data by armies of paid researchers, the most scientifically<br />

processed by the very latest in computer technology, the most mathematically<br />

sophisticated in presentation, which have so far turned out <strong>to</strong> be the most disappointing."<br />

These sentiments are still widely held in the discipline and at the heart of them lay<br />

fundamental disagreements over the practice and method of his<strong>to</strong>ry.<br />

In the United States the greatest fight in the contest over computational methods in<br />

his<strong>to</strong>ry <strong>to</strong>ok place in 1974 with the publication of Robert Fogel and Stanley Engerman's<br />

Time on the Cross: The Economics of American Negro Slavery. The book was a<br />

magisterial work of quantitative methodology and his<strong>to</strong>rical analysis, and it came in two<br />

volumes, one of which was dedicated entirely <strong>to</strong> quantitative methods and data. Fogel<br />

and Engerman admitted in their prologue that the book "will be a disturbing one <strong>to</strong> read."<br />

They also admitted that cliometrics, or quantitative his<strong>to</strong>ry, had limitations and that<br />

"there is no such thing as errorless data" (Fogel and Engerman 1974: 8, 10). Fogel and<br />

Engerman concentrated their study on some very hotly contested issues: the economic<br />

profitability of slavery, the economic success of the South in the years before the Civil<br />

War, and the relative productivity of slave and free agriculture.<br />

Time on the Cross received extensive criticism for two main reasons. First, the book<br />

rested so much of its key findings and interpretations on purely quantitative analysis and<br />

addressed purely economic questions, largely ignoring the textual and qualitative analysis<br />

of other his<strong>to</strong>rians as well as the social and political context of slavery. Second, the book<br />

addressed one of the most explosive and challenging subjects in American his<strong>to</strong>ry –<br />

slavery. It seemed everyone was interested in the subject. Fogel and Engerman appeared<br />

on national television and in Time magazine, and were reviewed in nearly every<br />

newspaper and magazine. The publication of the methods and evidence volume alone<br />

was enough <strong>to</strong> intimidate many scholars.


C. Vann Woodward, a distinguished his<strong>to</strong>rian of the American South, in an early review<br />

of the book was understanding of the authors but clearly concerned about where their<br />

computers had led (or misled) them. He noted that "the rattle of electronic equipment is<br />

heard off stage, and the reader is coerced by references <strong>to</strong> Vast research effort involving<br />

thousands of man and computer hours' and inconceivable mountains of statistical data."<br />

Woodward let it be known what the stakes were: "The object of the attack is the entire<br />

'traditional' interpretation of the slave economy." Woodward could hardly believe the line<br />

of reasoning that led Fogel and Engerman <strong>to</strong> assert that on average only "'2 percent of the<br />

value of income produced by slaves was expropriated by their masters,' and that this falls<br />

well within modern rates of expropriation." It was one of many disturbingly cold and<br />

skeptically received findings that the cliometricians put forward. Still, Woodward<br />

concluded that "it would be a great pity" if the controversy enflamed by Fogel and<br />

Engerman's conclusions were <strong>to</strong> "discredit their approach and obscure the genuine merits<br />

of their contribution" (Woodward 1974).<br />

Other reviewers were unwilling <strong>to</strong> concede any ground <strong>to</strong> the cliometricians. Withering<br />

criticism began shortly after the book was published, and it was led by Herbert Gutman,<br />

whose Slavery and the Numbers Game (1975) put forward the most devastating attack.<br />

As if <strong>to</strong> mock the cliometricians' elaborate reliance on numerical analysis and technical<br />

jargon, Gutman consistently referred <strong>to</strong> the book as "T/C" and <strong>to</strong> the authors as "F + E."<br />

Gutman's blood boiled when Fogel and Engerman icily concluded about the frequency of<br />

slave sales that "most slave sales were either of whole families or of individuals who<br />

were at an age when it would have been normal for them <strong>to</strong> have left the family." Gutman<br />

focused not on the statistical accuracy or the techniques of the cliometricians, but on the<br />

implications of their assumptions, the quality of their evidence, and how they used<br />

evidence. "Computers are helpful", Gutman pointed out, but not necessary for<br />

understanding the assumptions behind the statistical evidence (Gutman 1975: 3). Much of<br />

his critique faulted Fogel and Engerman's model of analysis for making almost no room<br />

for enslaved persons' agency, and instead making consistent and deeply embedded<br />

assumptions that everything in the enslaved person's life was directly subject <strong>to</strong> control<br />

and action by the slave holder. Gutman suggested that Fogel and Engerman's<br />

methodological, statistical, and interpretative errors all consistently aligned <strong>to</strong> produce a<br />

deeply flawed book, one that depicted the plantation South and slavery in a benign, even<br />

successful, light. In the end, the Fogel and Engerman finding that the typical enslaved<br />

person received less than one whipping per year (0.7, <strong>to</strong> be exact) helped drive home<br />

Gutman's critique. The number was presented as relatively minimal, as if such a thing<br />

could be counted and its effect established through quantification. Gutman pointed out<br />

that the really important measure was <strong>to</strong> account for the whip as an instrument of "social<br />

and economic discipline." The whipping data for Time on the Cross came from one<br />

plantation, and when Gutman re-examined the data he found that "a slave – 'on average' –<br />

was whipped every 4.56 days" (1975: 19).<br />

After Time on the Cross, advocates of numerical analysis and computing technology<br />

found themselves on the defensive. But this was not always the case. At the end of World<br />

War II in 1945 Vannevar Bush, the Direc<strong>to</strong>r of the Office of Scientific Research and<br />

Development and one of the United States' leading scientists, tried <strong>to</strong> envision how the


advances in science during the war might be most usefully directed. In an Atlantic<br />

Monthly essay, titled "As We May Think", Bush turned <strong>to</strong> examples from his<strong>to</strong>rical<br />

inquiry <strong>to</strong> make his key point – technology might be turned <strong>to</strong> the possibilities for<br />

handling the growing mass of scientific and humanistic data. The problem Bush<br />

described seems only more pressing now: "The investiga<strong>to</strong>r is staggered by the findings<br />

and conclusions of thousands of other workers – conclusions which he cannot find the<br />

time <strong>to</strong> grasp, much less remember, as they appear. Yet specialization becomes<br />

increasingly necessary for progress, and the effort <strong>to</strong> bridge between disciplines is<br />

correspondingly superficial" (Bush 1945).<br />

Bush's vision was for a machine he called "the memex", a strikingly prescient description<br />

of a networked desk<strong>to</strong>p computer. The machine would enable a scholar <strong>to</strong> map what<br />

Bush called a "trail" through the massive and growing scholarly record of evidence, data,<br />

interpretation, and narrative. He wanted machines <strong>to</strong> make the same kinds of connections,<br />

indeed <strong>to</strong> emulate the human mind with its fantastic power of association and linkage.<br />

Bush's principal examples for the memex's applications were spun out of his<strong>to</strong>ry – a<br />

research investigating the origins of the Turkish longbow – and he considered his<strong>to</strong>rians<br />

strong candidates <strong>to</strong> become "a new profession of trail blazers, those who find delight in<br />

the task of establishing useful trails through the enormous mass of the common record."<br />

As his<strong>to</strong>rians went about their noble work, Bush thought, they would leave nothing<br />

hidden from view, instead producing scholarship that was intricately connected in ways<br />

that could be accessed, replicated, and extended: "The inheritance from the master<br />

becomes, not only his additions <strong>to</strong> the world's record, but for his disciples the entire<br />

scaffolding by which they were erected."<br />

Some American his<strong>to</strong>rians were already working in the ways Bush described. They were<br />

using Hollerith (IBM) punchcards and undertaking large-scale analysis of a wide array of<br />

records. Frank Owsley and Harriet Owsley in their landmark 1949 study of the Old South<br />

used statistical methods <strong>to</strong> make linkages across disparate federal census records and <strong>to</strong><br />

create new quantitative datasets <strong>to</strong> analyze the class structure of the region. Frank Owsley<br />

and a large number of his graduate students worked on thousands of manuscript census<br />

returns for several counties from Alabama, Mississippi, Georgia, and Tennessee, and<br />

linked record-by-record the individuals and households in the population, slave owners,<br />

and agricultural schedules. Owsley's work challenged the idea that the planter elite<br />

dominated the South and instead suggested that a plain folk democracy characterized the<br />

region. The questions Owsley asked demanded methods of computational linking and<br />

quantitative analysis. Other his<strong>to</strong>rians in the 1940s and 1950s also worked on large<br />

statistical projects, including Merl Curtis analysis of the American frontier and William<br />

O. Aydelotte's study of the British Parliament in the 1840s.<br />

In retrospect, we can see three distinct phases in the ways his<strong>to</strong>rians have used computing<br />

technologies. Owsley, Curti, Aydelotte, and a few others in the 1940s were part of the<br />

first phase of quantitative his<strong>to</strong>ry. These his<strong>to</strong>rians used mathematical techniques and<br />

built large datasets. A second phase began in the early 1960s and was associated with an<br />

emerging field of social science his<strong>to</strong>ry. The "new" social, economic, and political his<strong>to</strong>ry<br />

concentrated on mobility, political affiliation, urbanization, patterns of assimilation, and


legislative behavior. It included his<strong>to</strong>rians using a range of statistical techniques, such as<br />

Lee Benson and Allan Bogue of the so-called "Iowa school" of quantification, as well as<br />

Olivier Zunz, Michael F. Holt, Steven Thernstrom, J. Morgan Kousser, and many others.<br />

Many "new" social, political, and economic his<strong>to</strong>rians drew on the massive data collected<br />

in the Inter-University Consortium for Political and Social Research (ICPSR) which was<br />

founded at the University of Michigan in 1962 with support from the American His<strong>to</strong>rical<br />

Association and the American Political Science Association <strong>to</strong> collect and make available<br />

his<strong>to</strong>rical data on county elections, census data, congressional roll call votes, and other<br />

miscellaneous political files. The computer made possible, or at least more practical and<br />

compelling, the study of his<strong>to</strong>ry "from the bot<strong>to</strong>m up." Political his<strong>to</strong>rians examined the<br />

influences at play in voting, not just the rhe<strong>to</strong>ric of a few leaders; social his<strong>to</strong>rians found<br />

patterns <strong>to</strong> describe the world of average people; and economic his<strong>to</strong>rians developed<br />

models <strong>to</strong> account for multiple variables of causation.<br />

The technology was alluring, but his<strong>to</strong>rians worried about the costs of the research and<br />

the time and training it <strong>to</strong>ok. Robert P. Swierenga, a his<strong>to</strong>rian with considerable<br />

experience in quantitative methods, confessed in a 1974 article that "<strong>to</strong> compute the<br />

Yules Q coefficient statistic for congressional roll call analysis" might require five hours<br />

of mainframe computer time <strong>to</strong> run the analysis, not <strong>to</strong> mention the time and cost of<br />

coding the 100,000 data cards needed for the computation. Swierenga was skeptical that<br />

"desk consoles" would solve the problem because his<strong>to</strong>rians' data were simply <strong>to</strong>o large<br />

for the machines of the foreseeable future (Swierenga 1974: 1068).<br />

The PC revolution and the rise of the Internet, along with the Moore's law exponential<br />

increase in speed and disk capacity of computing technology, led <strong>to</strong> the third phase. In<br />

this current phase, the networking capacity of the Internet offers the greatest<br />

opportunities and the most challenges for his<strong>to</strong>rians. At the same time as the Internet has<br />

created a vast network of systems and data, personal computers and software have<br />

advanced so far that nearly every his<strong>to</strong>rian uses information technologies in their daily<br />

work.<br />

Some significant methodological issues have emerged with the rise of PCs and "off the<br />

shelf" software. Manfred Thaller, an experienced quantitative his<strong>to</strong>rian, argued in 1987<br />

that database programs suited for businesses simply did not work for his<strong>to</strong>rians, who<br />

continually face "fuzzy" or incomplete data. The most obvious limitation that his<strong>to</strong>rians<br />

encounter with commercial database programs is the date function, since most programs<br />

cannot interpret or query nineteenth-century or earlier dates and a nest of problems in<br />

date functionality accompany any attempt <strong>to</strong> use his<strong>to</strong>rical dating in these programs. The<br />

nature of his<strong>to</strong>rical sources and the complexity of his<strong>to</strong>rical relationships led Thaller and<br />

others <strong>to</strong> challenge the easy comfort his<strong>to</strong>rians have with these programs. Thaller has<br />

been a voice crying in the wilderness, as his<strong>to</strong>rians in the 1980s and 1990s snapped up<br />

commercial database and spreadsheet programs with little hesitation and bent them <strong>to</strong><br />

their needs. Meanwhile, Thaller and other his<strong>to</strong>rians at Gottingen developed a software<br />

system, KLEIO, for the needs of his<strong>to</strong>rical researchers and scholars. KLEIO's data<br />

processing system differs from standard packages because of its data structure. The<br />

software allows for three forms of related information: documents, groups, and elements.


KLEIO handles a wide range of "fuzzy" data, his<strong>to</strong>rical dating systems, and types of<br />

documents (see Denley and Hopkin 1987: 149; Harvey and Press 1996: 190–3).<br />

Throughout the 1980s his<strong>to</strong>rians, especially in Europe, refined the use of databases for<br />

his<strong>to</strong>rical computing. In the process they questioned how these <strong>to</strong>ols affected their<br />

interpretations and whether the relational database models simply could not account for<br />

non-tabular data such as long texts, sound, images, and maps. Manfred Thaller's system<br />

tried <strong>to</strong> separate what he called the "knowledge environment" from the system<br />

environment or software, so that ideally the meaning of the records in a database<br />

remained independent of its source information. Database design for his<strong>to</strong>rical computing<br />

led Thaller and other quantitative his<strong>to</strong>rians <strong>to</strong> make distinctions between designs that<br />

build around the method of analysis or model from those which build around the source.<br />

In the former, his<strong>to</strong>rians extract information out of sources from the data's original<br />

context in<strong>to</strong> a set of well-defined tables arranged <strong>to</strong> allow for predetermined queries. In<br />

the latter, his<strong>to</strong>rians concentrate on capturing the original text or data and its entire source<br />

context, only later making decisions about analysis and organization of the data. Thaller's<br />

purist position on the use of databases in his<strong>to</strong>rical computing has gained attention in the<br />

field and sharpened the debate, but it has not slowed down the widespread use of<br />

commercial software for his<strong>to</strong>rical analysis. Most his<strong>to</strong>rians have taken a far more<br />

pragmatic approach.<br />

Despite the fundamental changes in information technologies and in his<strong>to</strong>rians' use of<br />

them, the recovery from the "battle royale" over Time on the Cross has been slow for<br />

those American his<strong>to</strong>rians working with computers. Manfred Thaller and the debates over<br />

database methodology and theory remained well out of the mainstream of American<br />

his<strong>to</strong>ry in the 1980s. Twenty years after Time on the Cross, when Daniel Greenstein<br />

published A His<strong>to</strong>rian's Guide <strong>to</strong> Computing (1994), he felt compelled <strong>to</strong> convince his<br />

readers that computers should not be "tarred with the same brush as the social science<br />

his<strong>to</strong>rians" (1994: 1). His treatment of regression analysis amounted <strong>to</strong> less than a page,<br />

while e-mail, bibliographic software, and note-capturing programs <strong>to</strong>ok up more than a<br />

chapter. Similarly, Evan Mawdsley and Thomas Munck, in their Computing for<br />

His<strong>to</strong>rians: An Introduc<strong>to</strong>ry Guide (1993), cautioned that they were "not champions of<br />

'cliometrics', 'quantification', the 'new' his<strong>to</strong>ry, 'scientific his<strong>to</strong>ry', or even what is called<br />

'social science his<strong>to</strong>ry'" The idea that some his<strong>to</strong>rians pushed the profession <strong>to</strong> use<br />

computers and quantification techniques struck these authors as "the tail wagging the<br />

dog" (Mawdsley and Munck 1993: xiii). J. Morgan Kousser, in his review of social<br />

science his<strong>to</strong>ry in the 1980s, admitted that most his<strong>to</strong>rians were moving abruptly away<br />

from quantitative methods. Time and distance tempered some critics, and by 1998 Joyce<br />

Appleby in her presidential address <strong>to</strong> the American His<strong>to</strong>rical Association remarked that<br />

the quantifying social his<strong>to</strong>rians had "immediate, substantive, conceptual, and ideological<br />

effects" on the profession. Appleby considered them responsible for an important shift in<br />

the practice of his<strong>to</strong>ry from unexplained <strong>to</strong> "explicit" assumptions about research<br />

methodology and from descriptive <strong>to</strong> analytical narrative (Appleby 1998: 4).<br />

British his<strong>to</strong>rians, however, have established a long and comparatively polite respect for<br />

his<strong>to</strong>rical computing. The Association for His<strong>to</strong>ry and Computing (AHC) was founded at


Westfield College, University of London, in 1986, and has sponsored several large<br />

conferences and published proceedings. At the 1988 meeting, Charles Harvey speculated<br />

on the reasons for the strong interest in AHC among British his<strong>to</strong>rians. Chief among<br />

them, according <strong>to</strong> Harvey, is the "strength of support for empirical, scientific,<br />

developmental his<strong>to</strong>ry in Britain" (Mawdsley et al. 1990: 206). Harvey described the<br />

British his<strong>to</strong>rical profession as focused on process not outcomes, rooted in scientific<br />

inquiry and only marginally concerned with theoretical debates. British his<strong>to</strong>rical<br />

scholarship was characterized by a focus on the importance of primary sources, the<br />

application of scientific notions in his<strong>to</strong>rical analysis, an emphasis on the specialized<br />

skills of the profession, the wide gulf between his<strong>to</strong>rians and social scientists, and the<br />

emphasis on applied skills and research methods in teaching students. Few of these<br />

characteristics describe the American his<strong>to</strong>rical profession and this may explain the<br />

thriving interest in "his<strong>to</strong>ry and computing" in England.<br />

Despite the success of and interest in AHC, some British his<strong>to</strong>rians have viewed<br />

computing technology variously as the handmaiden of postmodernism, as a witless<br />

accomplice in the collapse of narrative, and as the silent killer of his<strong>to</strong>ry's obligation <strong>to</strong><br />

truth and objectivity. One recent British his<strong>to</strong>rian argued: "The declining importance of<br />

the so-called grand narratives of national and class his<strong>to</strong>ries, and the fragmentation and<br />

loss of cultural authority of scholarly his<strong>to</strong>ry in the face of increasingly diffuse popular<br />

and political uses of 'his<strong>to</strong>ry,' cannot be separated from the impact of the new<br />

technologies" (Morris 1995: 503). Another suggested that when a his<strong>to</strong>rian "embarks on a<br />

statistical analysis he crosses a kind of personal Rubicon" (Swierenga 1974: 1062).<br />

Across that divide, in this view, the his<strong>to</strong>rian finds restrictions imposed by the software<br />

that defy the his<strong>to</strong>rian's allegiance <strong>to</strong> basic craft and adherence <strong>to</strong> evidence.<br />

If Time on the Cross soured many American his<strong>to</strong>rians on computational methods, the<br />

Philadelphia Social His<strong>to</strong>ry Project helped sustain them for a time. The project began in<br />

the 1970s under the leadership of Theodore Hershberg and <strong>to</strong>ok over a decade <strong>to</strong><br />

complete. At first, skeptical critics considered it a painful example of the tail wagging the<br />

dog, long on money, time, and detail and short on results. Hershberg and his colleagues at<br />

the University of Pennsylvania used the computer <strong>to</strong> make linkages between a vast array<br />

of his<strong>to</strong>rical data and set out <strong>to</strong> establish guidelines for large-scale his<strong>to</strong>rical relational<br />

databases. The project <strong>to</strong>ok in nearly $2 million in grants and, in the process, offered<br />

his<strong>to</strong>rians a model for large-scale research and interdisciplinary activity. Numerous<br />

dissertations and theses drew on its data collection, and great expectations for the project<br />

were widely claimed. Still, the project ended with what many considered a thud – the<br />

publication in 1981 of a book of essays by sixteen his<strong>to</strong>rians working with the<br />

Philadelphia Social His<strong>to</strong>ry Project that offered no larger synthesis for urban his<strong>to</strong>ry.<br />

Critics faulted the project and its authors for over-quantification, for paying <strong>to</strong>o much<br />

attention <strong>to</strong> cultivating an interdisciplinary "research culture", and for allowing federal<br />

dollars <strong>to</strong> push a research agenda about "public policy" matters removed from his<strong>to</strong>ry.<br />

While his<strong>to</strong>rians continued <strong>to</strong> ponder the pros and cons of quantitative methods and while<br />

the profession increasingly turned <strong>to</strong> cultural studies, or <strong>to</strong>ok the "linguistic turn", as<br />

some have called the move <strong>to</strong>ward the textual and French theory, computer scientists


were hammering out a common language for shared files over the Internet. The World<br />

Wide Web's opening in 1993 with the creation of HTML and browser technologies<br />

offered his<strong>to</strong>rians a new medium in which <strong>to</strong> present their work. One of the first<br />

his<strong>to</strong>rians <strong>to</strong> understand the implications of the Web for scholarship, teaching, and<br />

his<strong>to</strong>rical study generally was Edward L. Ayers at the University of Virginia. He was<br />

already embarking on a computer-aided analysis of two communities in the American<br />

Civil War when he first saw the Mosaic (TM) browser and the Web in operation at the<br />

Institute for Advanced Technology in the Humanities at the University of Virginia. Ayers<br />

immediately discarded his initial plans <strong>to</strong> distribute his research project on tape <strong>to</strong><br />

libraries and instead shifted the direction entirely <strong>to</strong> the World Wide Web. As a result,<br />

Ayers's Valley of the Shadow Project was one of the earliest sites on the World Wide<br />

Web and perhaps the first work of his<strong>to</strong>rical scholarship on it.<br />

From its inception the Valley of the Shadow Project was more than met the eye. The<br />

general public unders<strong>to</strong>od it as a set of Civil War letters, records, and other accounts.<br />

Students and teachers praised it for opening up the past <strong>to</strong> them and allowing everyone<br />

"<strong>to</strong> be their own his<strong>to</strong>rian." Scholars s<strong>to</strong>od initially aloof, wondering what could possibly<br />

be so exciting about an electronic archive. Gradually his<strong>to</strong>rians began <strong>to</strong> pay attention <strong>to</strong><br />

the Web and <strong>to</strong> the Valley of the Shadow Project in particular. Some his<strong>to</strong>rians saw the<br />

Project as perhaps a potentially troublesome upstart that threatened <strong>to</strong> change the<br />

narrative guideposts laid down in other media. James M. McPherson, one of the leading<br />

his<strong>to</strong>rians of the Civil War period whose work had done so much <strong>to</strong> influence Ken<br />

Burns's The Civil War, considered the Project potentially <strong>to</strong>o narrow and argued in a<br />

major conference panel on the Valley Project that the communities Ayers had chosen<br />

were not representative of the North and South. McPherson, and other critics as well,<br />

were beginning <strong>to</strong> recognize that the <strong>digital</strong> medium allowed Ayers <strong>to</strong> create a thoroughly<br />

captivating, technically savvy, and wholly unexpected comparative approach <strong>to</strong> the Civil<br />

War, one so complex and interconnected that such a thing seemed impossible in more<br />

linear media such as film and books.<br />

While Ayers was getting started on the Valley of the Shadow Project, another group of<br />

his<strong>to</strong>rians was already producing an electronic textbook for the American his<strong>to</strong>ry survey<br />

course. The American Social His<strong>to</strong>ry Project, based at City University of New York,<br />

included Roy Rosenzweig, Steve Brier, and Joshua Brown. Their Who Built America?, a<br />

CD-ROM of film, text, audio, images, and maps, aggressively and successfully placed<br />

social his<strong>to</strong>ry, especially labor his<strong>to</strong>ry, at the center of the national s<strong>to</strong>ry. The CD-ROM<br />

won a major prize from the American His<strong>to</strong>rical Association in 1994 and one reviewer<br />

called it a "massive <strong>to</strong>ur-de-force, setting the standard for his<strong>to</strong>rians who aim <strong>to</strong> make<br />

their work accessible <strong>to</strong> broad audiences via multimedia" (Darien 1998). Other reviewers,<br />

for N-Net, speculated that Who Built America? was part of a larger trend <strong>to</strong>ward CD-<br />

ROMs and multimedia his<strong>to</strong>ry, an "experiment" that would undoubtedly inspire others<br />

and lead <strong>to</strong> new forms of scholarship and teaching (Frost and Saillant 1994). These<br />

reviewers, though, also listed the daunting system requirements <strong>to</strong> run the CD-ROM:<br />

"Macin<strong>to</strong>sh computer running system 6.0.7 or higher; 4 megabytes of installed RAM in<br />

System 6 or 5 megabytes in System 7, with a minimum of 3–5 megabytes <strong>to</strong> allocate <strong>to</strong>


HyperCard; hard disk with 7 megabytes of free space, or 8 if QuickTime and HyperCard<br />

must be installed; 13-inch color moni<strong>to</strong>r; QuickTime-ready CD-ROM drive."<br />

It turns out that Edward Ayers's early decision <strong>to</strong> produce the Valley of the Shadow<br />

Project for the World Wide Web was one of the keys <strong>to</strong> that Project's long-term success.<br />

While the team at the University of Virginia working with Ayers has produced CD-<br />

ROMs and Ayers himself is publishing a narrative book out of the electronic archive, it is<br />

the website that has reached millions of users and <strong>to</strong> which all of the other scholarly<br />

objects point. The CD-ROMs of the Valley of the Shadow and Who Built America?<br />

contain remarkable materials, but their self-contained systems and off-the-network<br />

approach hobbled them, and by the late 1990s the CD-ROM seemed a relic in the fastmoving<br />

technology marketplace. The World Wide Web offered connectivity and<br />

hypertext on a scale that the public demanded and that scholars were beginning <strong>to</strong> see as<br />

immensely advantageous. "His<strong>to</strong>rians might begin <strong>to</strong> take advantage of the new media",<br />

Ayers wrote, "by trying <strong>to</strong> imagine forms of narrative on paper that convey the<br />

complexity we see in the <strong>digital</strong> archives." In his call for a "hypertext his<strong>to</strong>ry", Ayers<br />

admitted that while the technology offers grand possibilities, even with the crude <strong>to</strong>ols<br />

presently in use, there are significant barriers for his<strong>to</strong>rians. Ayers called hypertext<br />

his<strong>to</strong>ry potentially a "culmination of a long-held desire <strong>to</strong> present a more<br />

multidimensional his<strong>to</strong>ry and a threat <strong>to</strong> standard practice" (Ayers 1999).<br />

All of the connectivity and digitization has opened up his<strong>to</strong>ry and his<strong>to</strong>rical sources in<br />

unprecedented ways, yet the technology has not come without tensions, costs, and<br />

unexpected sets of alliances and demands for his<strong>to</strong>rians, educa<strong>to</strong>rs, administra<strong>to</strong>rs, and<br />

the public. The opportunities of <strong>digital</strong> technology for his<strong>to</strong>ry notwithstanding, Roy<br />

Rosenzweig, one of the leading scholars of the Who Built America? CD-ROM, and<br />

Michael O'Malley questioned whether professional his<strong>to</strong>rians can "compete with<br />

commercial operations" (Rosenzweig and O'Malley 1997: 152). Permission <strong>to</strong> pay for<br />

copyright, the costs of design and graphical layout, maintaining programming<br />

technologies and software all conspire <strong>to</strong> favor commercial publishing companies rather<br />

than professional his<strong>to</strong>rians. More recently, Rosenzweig has cautioned his<strong>to</strong>rians about<br />

the prospects of writing his<strong>to</strong>ry in a world of information overload. "His<strong>to</strong>rians, in fact,<br />

may be facing a fundamental paradigm shift from a culture of scarcity <strong>to</strong> a culture of<br />

abundance", Rosenzweig observed, while at the same time archivists and librarians warn<br />

that vast electronic records are being lost every day in government, business, and<br />

academic institutions, not <strong>to</strong> mention in homes, churches, schools, and nonprofits.<br />

Problems of "authenticity", Rosenzweig pointed out, plague <strong>digital</strong> preservation, and<br />

libraries and archives face skyrocketing costs and difficult choices. But the his<strong>to</strong>rians<br />

face equally demanding problems. "The his<strong>to</strong>rical narratives that future his<strong>to</strong>rians write<br />

may not actually look much different from those that are crafted <strong>to</strong>day", according <strong>to</strong><br />

Rosenzweig, "but the methodologies they use may need <strong>to</strong> change radically"<br />

(Rosenzweig 2003). The vast size of born <strong>digital</strong> electronic data collections and the<br />

interrelationships among these data present his<strong>to</strong>rians with a fundamental methodological<br />

issue, according <strong>to</strong> Rosenzweig. They will need <strong>to</strong>ols and methods, perhaps borrowed<br />

from the tradition of social science his<strong>to</strong>ry, <strong>to</strong> make sense of these records.


If his<strong>to</strong>rians face an unprecedented scale of information, they also encounter in the <strong>digital</strong><br />

medium an unexpectedly versatile mode of presenting their work. Many his<strong>to</strong>rians are<br />

beginning <strong>to</strong> ask what can we expect his<strong>to</strong>rical scholarship <strong>to</strong> look like in the networked<br />

electronic medium of the Internet and what forms of his<strong>to</strong>rical narrative might be<br />

enhanced or enabled. Robert Darn<strong>to</strong>n, past president of the American His<strong>to</strong>rical<br />

Association, National Book Award finalist, and innova<strong>to</strong>r in his<strong>to</strong>rical forms of<br />

scholarship, sketched out his ideas for the future of electronic publishing in a 1999 New<br />

York Review of Books essay, titled "The New Age of the Book." He was concerned in<br />

large part with the future of the book, especially the academic monograph, and the<br />

university presses that produce and publish them. Darn<strong>to</strong>n considered his<strong>to</strong>ry <strong>to</strong> be "a<br />

discipline where the crisis in scholarly publishing is particularly acute" (Darn<strong>to</strong>n 1999).<br />

Books won prizes and sold fewer than 200 copies; academic presses struggled <strong>to</strong> stay out<br />

of the red, and authors, especially young scholars before tenure, frantically tried <strong>to</strong><br />

publish their work. Darn<strong>to</strong>n used Middle East Studies as his example of "an endangered<br />

field of scholarship" about which the public cared and thought little. Only a few years<br />

after Darn<strong>to</strong>n's review article, the importance of Middle East Studies could hardly be<br />

argued, as the events of September 11, <strong>2001</strong>, brought the field <strong>to</strong> the immediate attention<br />

of the American public. Darn<strong>to</strong>n observed that the unpredictability of the market and the<br />

pressures on presses, tenure committees, and scholars seemed <strong>to</strong> conspire against the<br />

future of the academic book.<br />

Darn<strong>to</strong>n asked whether electronic publishing "could provide a solution." He outlined<br />

significant advantages for the field of his<strong>to</strong>ry where connections among evidence were so<br />

eye-opening during research in the archive and often so difficult <strong>to</strong> reproduce in<br />

narrative. Darn<strong>to</strong>n wished for a new form for his<strong>to</strong>rical scholarship: "If only I could show<br />

how themes crisscross outside my narrative and extend far beyond the boundaries of my<br />

book … instead of using an argument <strong>to</strong> close a case, they could open up new ways of<br />

making sense of the evidence, new possibilities of making available the raw material<br />

embedded in the s<strong>to</strong>ry, a new consciousness of the complexities involved in construing<br />

the past" (Darn<strong>to</strong>n 1999). Darn<strong>to</strong>n cautioned against "bloating" the book and piling on<br />

appendages <strong>to</strong> narrative. Instead, he called for a trim pyramid of layers: summary at the<br />

<strong>to</strong>p and documentation, evidence, commentary, and other materials below.<br />

It is not yet clear how <strong>digital</strong> technology will affect the practice of his<strong>to</strong>ry or whether<br />

his<strong>to</strong>rians will heed Darn<strong>to</strong>n's call <strong>to</strong> consider the advantages of electronic publication. In<br />

a show of leadership Darn<strong>to</strong>n offered his<strong>to</strong>rians an example of what new electronic<br />

scholarship might look like, publishing a dynamic essay in the American His<strong>to</strong>rical<br />

Review (Darn<strong>to</strong>n 2000). In recent years several his<strong>to</strong>rians are working <strong>to</strong>ward common<br />

ends. Philip Ething<strong>to</strong>n has written an electronic essay and presentation for the Web on the<br />

urban his<strong>to</strong>ry of Los Angeles. Ething<strong>to</strong>n "explores the hypothesis that the key concept in<br />

the search for his<strong>to</strong>rical certainty should be 'mapping' in a literal, not a metaphoric, sense"<br />

(Ething<strong>to</strong>n 2000). His work includes a wide range of media and sources <strong>to</strong> create, or<br />

rather recreate, the "panorama" of the city. Ething<strong>to</strong>n suggests that the website can be<br />

read like a newspaper, inviting readers <strong>to</strong> wander through it, skipping from section <strong>to</strong><br />

section, and focusing on what strikes their interest. Motivated "simultaneously by two<br />

ongoing debates: one among his<strong>to</strong>rians about 'objective knowledge,' and another among


urbanists about the depthless postmodern condition", Ething<strong>to</strong>n's electronic scholarship<br />

grasps the "archetype of 'hyperspace'" <strong>to</strong> address these concerns.<br />

Finally, Ayers and this author are working on an American His<strong>to</strong>rical Review electronic<br />

article, titled "The Differences Slavery Made: A Close Analysis of Two American<br />

Communities." This piece of electronic scholarship operates on several levels <strong>to</strong> connect<br />

form and analysis. First, it allows one <strong>to</strong> reconstruct the process by which our argument<br />

was arrived at, <strong>to</strong> "follow the logic" of our thinking, in effect <strong>to</strong> reconstruct the kind of<br />

"trails" that Vannevar Bush expected the technology <strong>to</strong> allow his<strong>to</strong>rians. This electronic<br />

scholarship also uses spatial analysis and spatial presentation <strong>to</strong> locate its subjects and its<br />

readers within the context of the his<strong>to</strong>rical evidence and interpretation. And it presents<br />

itself in a form that allows for unforeseen connections with future scholarship.<br />

His<strong>to</strong>rians will increasingly use and rely on "born <strong>digital</strong>" objects for evidence, analysis,<br />

and reference, as libraries and other government agencies increasingly digitize, catalogue,<br />

and make accessible his<strong>to</strong>rical materials (see Rosenzweig 2003). Some of these materials<br />

are hypertextual maps, others are annotated letters, edited video, oral his<strong>to</strong>ries, or<br />

relational databases. These <strong>digital</strong> objects vary widely according <strong>to</strong> their origin, format,<br />

and purpose. A born <strong>digital</strong> object is one created expressly for and in the <strong>digital</strong> medium,<br />

and therefore is more than a <strong>digital</strong> replication of an analogue object. For these objects,<br />

such as a reply <strong>to</strong> an email message, there is no complete analogue surrogate and as a<br />

consequence his<strong>to</strong>rians will need <strong>to</strong> understand not only what these objects explain but<br />

also how they were created. The electronic text, for example, marked up in Text<br />

Encoding Initiative (TEI) language becomes transformed in the process of both<br />

digitization and markup. Its unique markup scheme, as well as the software and hardware<br />

at both the server and client ends, affect how the text behaves and how its readers<br />

encounter it. Literary scholars, such as Espen Aarseth for example, have widely discussed<br />

the nature of narrative in cyberspace, and Aarseth calls cybertexts "ergotic" <strong>to</strong> distinguish<br />

them as non-linear, dynamic, explorative, configurative narratives (Aarseth 1997: 62).<br />

For his<strong>to</strong>rians the first stage in such textual developments for narrative have already been<br />

expressed in a wide range of <strong>digital</strong> archives. While these archives might appear for<br />

many users as undifferentiated collections of evidence, they represent something much<br />

more interpreted. Digital archives are often themselves an interpretative model open for<br />

reading and inquiry, and the objects within them, whether marked-up texts or hypermedia<br />

maps, derive from a complex series of authored stages. What Jerome McGann called<br />

"radiant textuality", the dynamic multi-layered expressions that <strong>digital</strong> technologies<br />

enable, applies <strong>to</strong> huge edited <strong>digital</strong> texts as well as <strong>to</strong> discrete objects within larger<br />

electronic archives (see McGann <strong>2001</strong>: for example 151–2, 206–7).<br />

The step from these archives <strong>to</strong> second-order his<strong>to</strong>rical interpretation necessarily involves<br />

the incorporation and explication of born <strong>digital</strong> objects, and "The Differences Slavery<br />

Made" offers an example of scholarship that emerges from and in relationship <strong>to</strong> born<br />

<strong>digital</strong> scholarly objects. The work fuses interpretative argument with the <strong>digital</strong><br />

resources of the Valley of the Shadow Project, and it is designed for a future of<br />

networked scholarship in which interpretation, narrative, evidence, commentary, and<br />

other scholarly activities will interconnect. The resulting piece is intended <strong>to</strong> fuse form


and argument in the <strong>digital</strong> medium. The authors propose a prismatic model as an<br />

alternative <strong>to</strong> Darn<strong>to</strong>n's pyramid structure, one that allows readers <strong>to</strong> explore angles of<br />

interpretation on the same evidentiary and his<strong>to</strong>riographical background. The prismatic<br />

functionality of the article offers <strong>to</strong> open the process of his<strong>to</strong>rical interpretation <strong>to</strong> the<br />

reader, providing sequential and interrelated nodes of analysis, evidence, and their<br />

relationship <strong>to</strong> previous scholarship. As important, the article tries <strong>to</strong> use its form as an<br />

explication of its argument and subject. Slavery, in other words, must be unders<strong>to</strong>od as<br />

having no single determinative value, no one experience or effect; instead, its refractive<br />

powers <strong>to</strong>uched every aspect of society. The article's form – its modules of refracted<br />

analysis, evidence, and his<strong>to</strong>riography – is meant <strong>to</strong> instruct and carry forward the<br />

argument.<br />

Given the sweeping changes within the field of his<strong>to</strong>ry and computing, we might ask<br />

what <strong>digital</strong> his<strong>to</strong>ry scholarship might tackle in the future. A range of opportunities<br />

present themselves. The most anticipation and attention currently surround what is<br />

loosely called "his<strong>to</strong>rical GIS." His<strong>to</strong>rical GIS (geographic information systems) refers <strong>to</strong><br />

a methodology and an emergent interdisciplinary field in which computer-aided spatial<br />

analysis is applied <strong>to</strong> archaeology, his<strong>to</strong>ry, law, demography, geography, environmental<br />

science, and other areas (see Knowles 2002). His<strong>to</strong>rians are building large-scale systems<br />

for rendering his<strong>to</strong>rical data in geographic form for places across the world from Great<br />

Britain <strong>to</strong> Tibet, and they are finding new answers <strong>to</strong> old questions from Salem,<br />

Massachusetts, <strong>to</strong> Tokyo, Japan. Legal scholars have begun <strong>to</strong> examine "legal<br />

geographies" and <strong>to</strong> theorize about the implications of spatial understandings and<br />

approaches <strong>to</strong> legal questions (see Blomley 1994). These scholars seem only a step away<br />

from adopting his<strong>to</strong>rical GIS approaches <strong>to</strong> their studies of segregation, slavery, race<br />

relations, labor relations, and worker safety. Environmental his<strong>to</strong>rians and scientists, <strong>to</strong>o,<br />

have developed new approaches <strong>to</strong> human and ecological change, examining subjects<br />

ranging from salt marsh economies and cultures in North America <strong>to</strong> the character and<br />

culture of Native American Indian societies in the river basins of the Chesapeake. Taken<br />

<strong>to</strong>gether these efforts represent remarkable possibilities for an integrated and networked<br />

spatial and temporal collection.<br />

But his<strong>to</strong>rians might dream up even more highly interpretive and imaginative <strong>digital</strong><br />

creations. Extending his<strong>to</strong>rical GIS, they might attempt <strong>to</strong> recreate "lost landscapes" in<br />

ways that fully allow readers <strong>to</strong> move and navigate through them. These fourdimensional<br />

models might res<strong>to</strong>re buildings, roads, and dwellings <strong>to</strong> his<strong>to</strong>ric landscapes<br />

as well as the legal, economic, social, and religious geographies within them. Networks<br />

of information, finance, trade, and culture might also find expression in these models.<br />

Readers might do more than query these datasets; they might interact within them <strong>to</strong>o,<br />

taking on roles and following paths they could not predict but cannot ignore. Readers of<br />

these interpretations will have some of the same questions that the critics of earlier<br />

computer-aided his<strong>to</strong>ry had. The goal for his<strong>to</strong>rians working in the new <strong>digital</strong> medium<br />

needs <strong>to</strong> be <strong>to</strong> make the computer technology transparent and <strong>to</strong> allow the reader <strong>to</strong> focus<br />

his or her whole attention on the "world" that the his<strong>to</strong>rian has opened up for<br />

investigation, interpretation, inquiry, and analysis. Creating these worlds, developing the


sequences of evidence and interpretation and balancing the demands and opportunities of<br />

the technology will take imagination and perseverance.<br />

References for Further Reading<br />

Aarseth, Espen (1997). Cybertext: Perspectives on Ergotic Literature. Baltimore and<br />

London: Johns Hopkins University Press.<br />

Appleby, Joyce (1998). The Power of His<strong>to</strong>ry, American His<strong>to</strong>rical Review 103, 1<br />

(February): 1–14.<br />

Ayers, Edward L. (1999). His<strong>to</strong>ry in Hypertext. Accessed April 5, 2004. At<br />

http://www.vcdh.virginia.edu/Ayers.OAH.html.<br />

Barzun, J. (1972). His<strong>to</strong>ry: The Muse and Her Doc<strong>to</strong>rs. American His<strong>to</strong>rical Review 77:<br />

36–64.<br />

Barzun, J. (1974). Clio and the Doc<strong>to</strong>rs: Psycho-His<strong>to</strong>ry, Quanta-His<strong>to</strong>ry and His<strong>to</strong>ry.<br />

Chicago: University of Chicago Press.<br />

Blomley, Nicholas K. (1994). Law, Space, and the Geographies of Power. New York:<br />

Guilford Press.<br />

Bogue, A. G. (1983). Clio and the Bitch Goddess: Quantification in American Political<br />

His<strong>to</strong>ry. Beverly Hills, CA: Sage Publications.<br />

Bogue, A. G. (1987). Great Expectations and Secular Depreciation: The First 10 Years<br />

of the Social Science His<strong>to</strong>ry Association. Social Science His<strong>to</strong>ry 11.<br />

Bur<strong>to</strong>n, Orville, (ed.) (2002). Computing in the Social Sciences and Humanities. Urbana<br />

and Chicago: University of Illinois Press.<br />

Bush, Vannevar (1945). As We May Think. Atlantic Monthly (July).<br />

Clubb, J. M. and H. Allen (1967). Computers and His<strong>to</strong>rical Studies, Journal of<br />

American His<strong>to</strong>ry 54: 599–607.<br />

Darien, Andrew (1998). Review of Who Built America? From the Centennial Celebration<br />

of 1876 <strong>to</strong> the Great War of 1914. Journal of Multimedia His<strong>to</strong>ry 1, 1 (Fall). At<br />

http://www.albany.edu/jmmh.<br />

Darn<strong>to</strong>n, Robert (1999). The New Age of the Book. New York Review of Books (March):<br />

18.


Darn<strong>to</strong>n, Robert (2000). An Early Information Society: News and the Media in<br />

Eighteenth-century Paris. American His<strong>to</strong>rical Review 105, 1 (February). Accessed April<br />

5, 2004. At http://www.his<strong>to</strong>rycooperative.org/journals/ahr/105.1/ah000001.html.<br />

David, Paul et al. (1976). Reckoning with Slavery: A Critical Study in the Quantitative<br />

His<strong>to</strong>ry of American Negro Slavery. New York: Oxford University Press.<br />

Degler, Carl (1980). Remaking American His<strong>to</strong>ry. Journal of American His<strong>to</strong>ry 67, 1: 7–<br />

25.<br />

Denley, Peter and Deian Hopkin (1987). His<strong>to</strong>ry and Computing. Manchester:<br />

Manchester University.<br />

Press, Denley, Peter, Stefan Fogelvik, and Charles Harvey (1989). His<strong>to</strong>ry and<br />

Computing II. Manchester: Manchester University Press.<br />

Erikson, C. (1975). Quantitative His<strong>to</strong>ry. American His<strong>to</strong>rical Review 80: 351–65.<br />

Ething<strong>to</strong>n, Philip J. (2000). Los Angeles and the Problem of Urban His<strong>to</strong>rical<br />

Knowledge. American His<strong>to</strong>rical Review (December). At<br />

http://www.usc.edu/dept/LAS/his<strong>to</strong>ry/his<strong>to</strong>rylab/LAPUHK/.<br />

Fitch, N. (1984). Statistical Fantasies and His<strong>to</strong>rical Facts: His<strong>to</strong>ry in Crisis and its<br />

Methodological Implications. His<strong>to</strong>rical Methods 17: 239–54.<br />

Fogel, R. W. and G. R. El<strong>to</strong>n (1983). Which Road <strong>to</strong> the Past? Two Views of His<strong>to</strong>ry.<br />

New Haven, CT: Yale University Press.<br />

Fogel, R. W. and Stanley Engerman (1974). Time on the Cross: The Economics of<br />

American Negro Slavery. London: Little, Brown.<br />

Frost, Carol J. and John Saillant (1994). Review of Who Built America? From the<br />

Centennial Celebration of 1876 <strong>to</strong> the Great War of 1914. H-Net Reviews. Accessed<br />

April 5, 2004. At http://www.h-net.org/mmreviews/showrev.cgi?path=230.<br />

Greenstein, Daniel I. (1994). A His<strong>to</strong>rian's Guide <strong>to</strong> Computing. Oxford: Oxford<br />

University Press.<br />

Gutman, Herbert (1975). Slavery and the Numbers Game: A Critique of Time on the<br />

Cross. Urbana: University of Illinois Press.<br />

Harvey, Charles and Jon Press (1996). Databases in His<strong>to</strong>rical Research: Theory,<br />

Methods, and Applications. London: Macmillan Press.<br />

Himmelfarb, G. (1987). The New His<strong>to</strong>ry and the Old. Cambridge, MA: Harvard<br />

University Press.


Jenson, Richard (1974). Quantitative American Studies: The State of the Art. American<br />

Quarterly 26, 3: 225–220.<br />

Knowles, Anne Kelly, (ed.) (2002). Past Time, Past Place: GIS for His<strong>to</strong>ry. Redlands,<br />

CA: ESRI.<br />

Kousser, J. Morgan (1989). The State of Social Science His<strong>to</strong>ry in the late 1980s.<br />

His<strong>to</strong>rical Methods 22: 13–20.<br />

McGann, Jerome (<strong>2001</strong>). Radiant Textuality: Literature after the World Wide Web. New<br />

York: Palgrave.<br />

Mawdsley, Evan and Thomas Munck (1993). Computing for His<strong>to</strong>rians: An Introduc<strong>to</strong>ry<br />

Guide. Manchester: Manchester University Press.<br />

Mawdsley, Evan, Nicholas Morgan, Lesley Richmond, and Richard Trainor (1990).<br />

His<strong>to</strong>ry and Computing III: His<strong>to</strong>rians, Computers, and Data, Applications in Research<br />

and Teaching. Manchester and New York: Manchester University Press.<br />

Middle<strong>to</strong>n, Roger and Peter Wardley (1990). Information Technology in Economic and<br />

Social His<strong>to</strong>ry: The Computer as Philosopher's S<strong>to</strong>ne or Pandora's Box? Economic<br />

His<strong>to</strong>ry Review (n.s.) 43, 4: 667–96.<br />

Morris, R. J. (1995). Computers and the Subversion of British His<strong>to</strong>ry, journal of British<br />

Studies 34: 503–28.<br />

Rosenzweig, Roy (1994). "So What's Next for Clio?" CD-ROM and His<strong>to</strong>rians. Journal<br />

of American His<strong>to</strong>rians 81, 4 (March): 1621–40.<br />

Rosenzweig, Roy (2003). Scarcity or Abundance? Preserving the Past in a Digital Era.<br />

American His<strong>to</strong>rical Review 108, 3 (June): 735–62.<br />

Rosenzweig, Roy and Michael O'Malley (1997). Brave New World or Blind Alley?<br />

American His<strong>to</strong>ry on the World Wide Web. Journal of American His<strong>to</strong>ry 84, 3 (June):<br />

132–55.<br />

Rosenzweig, Roy and Michael O'Malley (<strong>2001</strong>). The Road <strong>to</strong> Xanadu: Public and<br />

Private Pathways on the His<strong>to</strong>ry Web. Journal of American His<strong>to</strong>ry 88, 2 (September):<br />

548–79.<br />

Shorter, Edward (1971). The His<strong>to</strong>rian and the Computer: A Practical Guide. Englewood<br />

Cliffs, NJ: Prentice Hall.<br />

Swierenga, Robert P., (ed.) (1970). Quantification in American His<strong>to</strong>ry. New York:<br />

Atheneum.


Swierenga, Robert P. (1974). Computers and American His<strong>to</strong>ry: The Impact of the "New"<br />

Generation. journal of American His<strong>to</strong>ry 60, 4: 1045–220.<br />

Woodward, C. Vann (1974). The Jolly Institution. New York Review of Books, May 2.<br />

6.<br />

Lexicography<br />

Russon Wooldridge<br />

Lexicography is here considered in the literal and concrete sense of the word: the writing<br />

of the lexicon, the ordered description of the lexicon of a language in the form of a<br />

reference work usually called a dictionary.<br />

The following is intended as a typological and partial account of the use of computers in<br />

lexicography, dealing with the essential applications and the main examples, for English<br />

and French, of computer-assisted lexicographical products. The dictionaries considered<br />

are those intended for the general public; scant mention will be made of those created for<br />

specialists.<br />

Nature of the Dictionary Text<br />

A dictionary has fundamentally the same structure as a telephone direc<strong>to</strong>ry, a hospital's or<br />

general practitioner's medical files, or a library catalogue. Each unit of these collections is<br />

a record containing a number of fields, potentially the same for each record (some fields<br />

are blank) and placed in the same order, the essential characteristic of this relational<br />

database being its recursiveness.<br />

• telephone direc<strong>to</strong>ry entry: name, address, telephone number<br />

• medical record: name, personal coordinates, medical his<strong>to</strong>ry, progress notes,<br />

consultations, lab reports, etc.<br />

• library catalogue record: title, author, place and date of publication, subject, material,<br />

ISBN number, holdings, etc.<br />

• dictionary entry: headword, pronunciation, part of speech, definition, examples,<br />

etymology, etc.<br />

Example of two dictionary entries (source: Dictionnaire universel francophone):<br />

dictionnaire n. m. Ouvrage qui recense et décrit, dans un certain ordre, un ensemble<br />

particulier d'éléments du lexique (sens 4). Dictionnaire médical, étymologique. –<br />

Dictionnaire de la langue ou dictionnaire de langue, qui décrit le sens, les valeurs, les


emplois, etc. des mots d'une langue. Le dictionnaire de I'Académiefrançaise. –<br />

Dictionnaire bilingue, qui donne les équivalents des mots et expressions d'une langue<br />

dans une autre langue. Un dictionnaire français – vietnamien. – Dictionnaire<br />

encyclopédique, qui, outre les descriptions de mots, fournit des développements<br />

encyclopédiques consacrés aux objets désignés par les mots. Syn. Fam. dico.<br />

informatique n. f. et adj. Technique du traitement au<strong>to</strong>matique de l'information au<br />

moyen des calculateurs et des ordinateurs. Informatique de gestion. adj. Relatif à cette<br />

technique. Traitement par des moyens informatiques.<br />

Encycl. L'informatique est apparue avec le développement des calculateurs électroniques<br />

à grande capacité, les ordinateurs (le mot informatique date de 1962). La rapidité d'accès<br />

et de traitement de l'information, l'au<strong>to</strong>matisme du fonctionnement des ordinateurs et la<br />

systématique des résolutions ont ouvert un très vaste champ d'application à<br />

l'informatique: recherche scientifique (ex.: contrôle de la trajec<strong>to</strong>ire d'un satellite);<br />

Industrie (conception assistée par ordinateur, contrôle et commande des machines, des<br />

processus); gestion des entreprises (opérations administratives, simulation, recherche<br />

opérationnelle); enseignement programmé; documentation, banques d'informations;<br />

informatique individuelle. La liaison de plusieurs ordinateurs accroit la puissance de leur<br />

traitement, la télématique assurant la transmission (V. télématique, ordinateur, réseau).<br />

The recursiveness of the informational fields of the above two dictionary entries is<br />

indicated by typography, position, and abbreviation: (1) headword in bolded large letters;<br />

(2) part of speech conventionally abbreviated; (3) definition; (4) examples of usage in<br />

italics. Fields 1, 3, and 4 are given in full because of the idiosyncratic nature of lexical<br />

units; field 2 is given in abbreviated form since its values belong <strong>to</strong> a small finite class.<br />

Typography, position, abbreviation, and ellipsis (none of the four fields is explicitly<br />

named) are the features of dictionary recursiveness and economy (the dictionary is also a<br />

commercial product). Occasional fields tend <strong>to</strong> be named: "Syn." for synonyms;<br />

"Encycl." for encyclopedic information (normal, systematic, unlabeled information is<br />

linguistic); "V." for cross-reference <strong>to</strong> related terms.<br />

Even the simplest dictionary entries, such as the ones quoted above, tend <strong>to</strong> be<br />

structurally complex. Besides the main binary equations – (a) dictionnaire = masculine<br />

noun; (b) dictionnaire [means] "Ouvrage qui recense et décrit, dans un certain ordre, un<br />

ensemble particulier d'éléments du lexique"; (c) [the word] dictionnaire [typically occurs<br />

in expressions such as] Dictionnaire médical (domain of experience), [dictionnaire]<br />

étymologique (domain of language) – there are also ternary ones: dictionnaire ><br />

[exemplified in] Dictionnaire bilingue > [which means] "[dictionnaire] qui donne les<br />

équivalents des mots et expressions d'une langue dans une autre langue." (The implicit<br />

copulas and other terms are here made explicit and enclosed in square brackets.)<br />

Idiosyncrasy is characteristic of lexis and also of the dictionary, which in the vast<br />

majority of its realizations is composed by (fallible) human beings. Just as the treatment<br />

of lexical units will vary enormously according <strong>to</strong> part of speech, frequency of usage,<br />

monosemy or polysemy, register, and other variables, so will dictionary-writing tend <strong>to</strong><br />

vary according <strong>to</strong> time (beginning, middle, or end of the alphabet, or of the writing of the


dictionary, even day of the week) and writer (entry-writer A and entry-writer B are<br />

individual human beings and not clones or machines).<br />

Setting aside the question of idiosyncrasy and variability of lexical units and dictionarywriting<br />

(the latter nevertheless an important obstacle in computerizing the Trésor de la<br />

langue française – see below), the well-ordered dictionary requires three types of<br />

sophisticated competence on the part of the user: (1) linguistic competence obviously; (2)<br />

dictionary competence, a particular type of textual competence, enabling one, for<br />

example, <strong>to</strong> find a word beginning with m by opening the dictionary more or less in the<br />

middle, <strong>to</strong> know that adj. means adjective/adjectif (and through linguistic competence <strong>to</strong><br />

know what an adjective is), etc.; (3) pragmatic competence <strong>to</strong> make sense of references <strong>to</strong><br />

the outside world: Le dictionnaire de l'Académie française, "calculateurs électroniques",<br />

etc.<br />

The requirement of different types of user competence combined with the frequent use of<br />

ellipsis can result in cases of ambiguity that tax the analytical faculties of the dictionary<br />

reader and render powerless the analytical parser of the computer. The following<br />

examples are taken from the entry for GAGNER in Lexis (Wooldridge et al. 1992):<br />

(a) Gagner quelque chose (moyen de subsistance, récompense), l'acquérir par son travail<br />

(b) Gagner son biftek (pop.) [= gagner <strong>to</strong>ut juste sa vie]<br />

(c) Gagner le Pérou (= des sommes énormes)<br />

(d) Gagner le maquis (syn. PRENDRE)<br />

(e) C'est <strong>to</strong>ujours ça de gagné (fam.) [= c'est <strong>to</strong>ujours ça de pris]<br />

(f) Il ne gagne pas lourd (= peu)<br />

(g) Je suis sorti par ce froid, j'ai gagné un bon rhume (syn. plus usuels: ATTRAPER,<br />

PRENDRE, fam. CHIPER)<br />

Each of the seven items contains an equation of synonymy, the equation concerning<br />

either the whole or part of the first term: the object of the verb in (a) and (c), the adverbial<br />

qualifier in (f), the verb in (d) and (g), the whole expression in (b) and (e). Linguistic<br />

competence is necessary <strong>to</strong> equate quelque chose and l' (a), ne … pas lourd with peu (f),<br />

the conjugated form ai gagné with the infinitives attraper, prendre, and chiper (g). The<br />

dictionary user also has <strong>to</strong> deal with the variability of the synonymy delimiters and<br />

indica<strong>to</strong>rs (parentheses, brackets, equals sign, syn. label, upper case).<br />

In brief, the dictionary, in theory a systematic relational database, with ordered records<br />

and recurrent fields, may in human practice be as variable as the lexicon it sets out <strong>to</strong><br />

describe. Successful applications of the computer <strong>to</strong> man-made dictionaries are, then,<br />

usually modest in their ambitions. Computer-driven dictionaries (machine dictionaries)


tend <strong>to</strong> be procrustean in their treatment of language, or limit themselves <strong>to</strong> relatively<br />

simple areas of lexis such as terminology.<br />

Pre-WWW (World Wide Web)<br />

Modern lexicography did not wait for the invention of the computer, nor even that of the<br />

seventeenth-century calculating machines of Leibniz and Pascal, <strong>to</strong> apply computer<br />

methods <strong>to</strong> dictionaries. In 1539, the father of modern lexicography, Robert Estienne,<br />

King's Printer, bookseller, humanist, and lexicographer, published his Dictionaire<br />

francois-latin (1539), a "mirror-copy" of his Dictionarium latinogallicum of the previous<br />

year. Each French word and expression contained in the glosses and equivalents of the<br />

Latin-French dictionary had its own headword or sub-headword in the French-Latin; each<br />

Latin word or expression contained in the headwords and examples of the Latin-French<br />

occurred as an equivalent <strong>to</strong> the corresponding French in the French-Latin. Example of<br />

the words aboleo and abolir.:<br />

1 Dispersion:<br />

1538: "ABOLEO […] Abolir, Mettre a neant."<br />

> 1539: "Abolir, Abolere" (s.v. Abolir), "Mettre a neant […] Abolere" (s.v. Mettre).<br />

2 Collection:<br />

1538: "ABOLEO […] Abolir", "ABROGO […] Abolir", "Antique […] Abolir"<br />

"Conuellere […] abolir." "Exterminare […] abolir." "Inducere […] Abolir", "Interuertere<br />

[…] Abolir", "OBLITERO […] abolir.", "Resignare, Abolir."<br />

> 1539: "Abolir, Abolere, Abrogare, Antiquare, Conuellere, Exterminare, Inducere,<br />

Interuertere, Obliterare, Resignare."<br />

Moving forward four centuries and several decades, we find the first applications of the<br />

computer <strong>to</strong> lexicography in the 1960s and 1970s. In the 1960s, the Centre pour un<br />

Trésor de la langue française in Nancy started keyboarding representative works of<br />

literature and technical treatises <strong>to</strong> provide source material for its print dictionary, the<br />

Dictionnaire de la langue du XIX e et du XX e siècle, commonly known as the Trésor de la<br />

langue française, or TLF. In the late 1970s, there appeared in England two machinereadable<br />

dictionaries, the Oxford Advanced Learners' Dictionary and the Longman<br />

Dictionary of Contemporary English; the latter used the computer not only <strong>to</strong> print off<br />

the paper dictionary, but also <strong>to</strong> help in its writing (Meijs 1992: 143–5).<br />

The first early dictionary <strong>to</strong> be computerized was Jean Nicot's Thresor de la langue<br />

françoyse of 1606. The text was keyboarded in Nancy and Toron<strong>to</strong> between 1979 and<br />

1984, indexed at the University of Toron<strong>to</strong> with the mainframe concordance program<br />

COGS, published in microfiche concordance form in 1985, indexed as a standalone<br />

interactive database with WordCruncher in 1988, and put on the World Wide Web in


1994 (see sections on the WWW era and on technological change, below). It is not<br />

without interest <strong>to</strong> note that in the early 1980s funding agencies expected concordance<br />

projects <strong>to</strong> undertake lemmatization of text forms. An argument had <strong>to</strong> be made <strong>to</strong><br />

demonstrate the absurdity of attempting <strong>to</strong> lemmatize an already partly lemmatized text:<br />

dictionary headwords are lemmas. The Nicot project initially had the ambition of<br />

labeling information fields (Wooldridge 1982), until it quickly became obvious that such<br />

fields, though present and analyzable by the human brain, are impossible <strong>to</strong> delimit<br />

systematically in a complex early dictionary such as Nicot's Thresor, where position,<br />

typography, and abbreviation are variable, and functional polyvalence is common. The<br />

challenge is not negligible in modern dictionaries, where explicit field-labeling is the<br />

norm. Other early dictionaries have since been <strong>digital</strong>ly retro-converted, notably Samuel<br />

Johnson's Dictionary of the English Language, published on CD-ROM in 1996.<br />

The 1980s saw the emergence of large-scale computer-assisted lexicographical<br />

enterprises. The COBUILD (Collins and Birmingham University International Language<br />

Database) Project started in 1980, with the intention of creating a corpus of contemporary<br />

English for the writing of an entirely new dictionary and grammar. The young discipline<br />

of corpus linguistics and the COBUILD Project fed off each other in this innovative<br />

lexicographical environment (Sinclair 1987; Renouf 1994). The New Oxford English<br />

Dictionary Project was formed <strong>to</strong> produce the second edition of the OED with the aid of<br />

computer technology. The project was international in scope: it was conceived and<br />

directed in England, the role of the computer was defined and implemented in Canada,<br />

the text was keyboarded in the United States of America. The second edition appeared in<br />

print in 1989 and on CD-ROM in 1992.<br />

Where early electronic tagging of dictionaries was restricted <strong>to</strong> typographical codes for<br />

printing the finished product, it soon became necessary <strong>to</strong> add information tags so that<br />

not only could the text be correctly displayed on screen or paper, but it could also be<br />

searched and referenced by fields. The dictionary was just one of the types of text whose<br />

structure was analyzed by the Text Encoding Initiative (TEI) (Ide et al. 1992).<br />

The last decade of the twentieth century witnessed the proliferation of electronic<br />

dictionaries distributed on CD-ROM. For example, the 1993 edition of the Random<br />

House Unabridged Dictionary came both in print and on CD-ROM, the two being sold<br />

<strong>to</strong>gether for the price of one. As one might expect of a freebie, the functionality of the<br />

Random House CD-ROM is rudimentary. On the other hand, the CD-ROM version of the<br />

Petit Robert, published in 1996, offers many advantages over the print edition: besides<br />

basic look-up of words and their entries, the user can search for anagrams (the search<br />

term dome produces dome and mode), homophones (saint produces sain, saint, sein,<br />

seing), etymologies by language (families: African, Amerindian, Arabic, etc., or specific<br />

idioms: Bantu, Hotten<strong>to</strong>t, Somali, etc.), quotations by author, work, or character, plus<br />

full-text searches in either the complete dictionary text (full entries), or the particular<br />

fields of examples of usage or synonyms and an<strong>to</strong>nyms.<br />

It is often the case that a number of entries are made more complete through the access <strong>to</strong><br />

the full text granted by an electronic version of a dictionary. For example, <strong>to</strong> take the case


of the Petit Robert, sabotage, applied <strong>to</strong> work, organizations, or machinery, in the entry<br />

for the word, is used in an important – and common – figurative sense in a quotation<br />

concerning speaker. "Sabotage de la prononciation de notre belle langue par les speakers<br />

de la radio." Dictionaries tend <strong>to</strong> be more conservative in their treatment of a word in its<br />

own entry than elsewhere.<br />

A brief mention should be made of computer-assisted lexicographical <strong>to</strong>ols for the<br />

everyday user, the main ones being word-processor spellcheckers and thesauruses.<br />

A good broad account of the period 1960-early 1990s, that of computer-assisted and<br />

computer-driven lexicography resulting in print and CD-ROM dictionaries, is given by<br />

Meijs (1992); an in-depth one by Knowles (1990).<br />

Lexicography in the WWW Era<br />

Like many other human practices, lexicography – and particularly lexicography – has<br />

been transformed by the World Wide Web. The Web works by virtue of words; <strong>to</strong> quote<br />

from the title of a well-known book by James Murray's granddaughter (Murray 1977), the<br />

Web, like the dictionary, is a "web of words." One reads the words in a book, one looks<br />

up headwords in a dictionary, one surfs the Web by keywords. The millions of documents<br />

published on the Web constitute, through the structuring of search engine keywords, a<br />

vast dictionary, an encyclopedic dictionary of concepts and words. Conventional<br />

dictionaries, whether paper or electronic, pale by comparison, though many of the latter<br />

are caught in the online Web of words.<br />

An early demonstration of the Web as super- or meta-dictionary can be found in<br />

Wooldridge et al. (1999). A Web search of the Canadian French word enfirouaper<br />

(search term: enfirou*) collected occurrences of the verb and derivatives both used and<br />

commented on; the documents were of all types: political and personal, newspaper report<br />

and manifes<strong>to</strong>, poetry and prose, dialogue and dictionary. The occurrences of the word in<br />

use showed that dictionary and glossary treatment is narrow and out-of-date (cf. sabotage<br />

above). Applying the principles of corpus creation and analysis learned from the<br />

COBUILD Project, the WebCorp Project at the University of Liverpool uses standard<br />

Web search engines such as Google and AltaVista <strong>to</strong> collect results from the Web and<br />

format them in easily analyzable KWIC concordances (Kehoe and Renouf 2002). For<br />

example, expressions such as one Ave short of a rosary, two leeks short of a harvest<br />

supper, or two sheets short of a bog roll, encountered in the novels of Reginald Hill, are<br />

individual realizations of the commonly used pattern "one/two/three/a/an/several X short<br />

of a Y" (X being constituent parts of the whole Y), which can be expressed in Google or<br />

AltaVista by variants of the search term "one * short of a". Since WebCorp is, at least at<br />

the time of writing, freely available on the Web, corpus linguistics has become a<br />

lexicographical <strong>to</strong>ol for the general public.<br />

Meta-sites are a good source of information about online dictionaries. For French, two<br />

good ones are Robert Peckham's Leximagne – l'Empereur des pages dico and Carole<br />

Netter's ClicNet: Dictionnaires. The latter gives links for the following categories (I


translate): Multilingual dictionaries; French-language dictionaries and encyclopedias;<br />

Grammar, morphology, orthography, and linguistics; His<strong>to</strong>rical dictionaries; Dictionaries<br />

of Architecture, Visual arts, Slang, Law, Economics and Finance, Gastronomy and<br />

Dietetics, His<strong>to</strong>ry, Humor, Games, Multicultural dictionaries, Literature, Media, Music,<br />

Nature and Environment, Science, Political science, Services, Social sciences and<br />

Humanities, Sports, Techniques, Tourism, Various vocabularies; Internet glossaries;<br />

Discussion lists; Lexical columns; Other servers.<br />

Most online dictionaries are fairly modest in scope and are published as straight text, just<br />

like a print dictionary. A few, however, can be queried interactively as relational<br />

databases, and may offer other features. It is interesting then <strong>to</strong> compare those of the two<br />

main online dictionaries of English and French, the OED and the TLF.<br />

• In the late 1990s, a first online version of the second edition of the OED became<br />

available through the OED Project at the University of Waterloo; a more generally<br />

available edition, the OED Online, was launched on the main OED website in 2000. Both<br />

versions are accessible by subscription only and allow the following types of search:<br />

"lookup" (as in the print version), "entire entry" = full-text search, "etymology", "label"<br />

(= field label). Electronic network technology has also made a significant contribution <strong>to</strong><br />

the OED's reading program: in place of the parcels of paper slips from around the globe<br />

delivered by the Post Office <strong>to</strong> the Oxford Scrip<strong>to</strong>rium of James Murray's day, readers<br />

can now send in words, references, and details via the Web. The OED website gives a<br />

detailed his<strong>to</strong>ry of the dictionary, thus adding <strong>to</strong> a scholarly value rare on online<br />

dictionary sites.<br />

• The complete version of the TLFI (Trésor de la langue française informatisé),<br />

published on the Web in 2002, is both free and, somewhat ambitiously, lets the user limit<br />

queries <strong>to</strong> one or several of 29 fields, including "entrée" (the whole entry), "exemple"<br />

(with subcategories of various types of example), "auteur d'exemple" (examples by<br />

author), "date d'exemple" (by date), "code grammatical", "définition", "domaine<br />

technique", "synonyme/an<strong>to</strong>nyme." The 16-volume print TLF suffered from a high<br />

degree of writing variation (cf. the first section, above), making field tagging an<br />

extremely difficult task and forcing the Institut National de la Langue Française (INaLF)<br />

team <strong>to</strong> adopt in part a probabilistic approach in creating the electronic version (Henry<br />

1996).<br />

A characteristic of the Web is the hyperlink, which facilitates, among other things, the<br />

association between text and footnote (intratextual link), that between the bibliographical<br />

reference and the library (intertextual link), or that between a word A encountered within<br />

the dictionary entry for word B and the entry for word A (e.g. "anaptyxis: epenthesis of a<br />

vowel" > epenthesis). The Dictionnaire universel francophone en ligne (DUF), an<br />

important free language resource for speakers and learners of all varieties of French and<br />

the online equivalent of the one-volume general language print dictionary <strong>to</strong> be found in<br />

most homes, has hyperlinks for every word contained within its entries, allowing the user<br />

<strong>to</strong> refer <strong>to</strong> the entry of any given word with a single click (e.g. "sabotage n. m. 1. TECH


Action de saboter (un pieu, une traverse, etc.)" > nom, masculin, technique,<br />

technologie, technologique, action, de, saboter, un, pieu, traverse, et caetera).<br />

Apart from dictionaries of general contemporary language, there are a large number of<br />

marked or specialized ones. In the field of early dictionaries, there are several sixteenth-<br />

<strong>to</strong> early twentieth-century ones freely accessible in database form in the section<br />

Dictionnaires d'autrefois on the site of the ARTFL (American and French Research on<br />

the Treasury of the French Language) Project at the University of Chicago: Estienne,<br />

Nicot, Bayle, Académie française (also on an INaLF server in Nancy). Interactive<br />

databases of several of these and others are on a server of the University of Toron<strong>to</strong>.<br />

Terminology, once the reserve of paying specialists, is now freely available on the Web.<br />

For example, a Glossaire typographique et linguistique or a Terminology of Pediatric<br />

Mas<strong>to</strong>cy<strong>to</strong>sis;, a pediatric mas<strong>to</strong>cy<strong>to</strong>sis term such as anaphylaxis occurs on tens of<br />

thousands of Web pages (69,700 hits with Google on September 28, 2002).<br />

Web lexicography offers a number of <strong>to</strong>ols, a significant one being au<strong>to</strong>matic translation<br />

(e.g., Babelfish) intended <strong>to</strong> translate the gist of a Web document in<strong>to</strong> a language<br />

unders<strong>to</strong>od by the user. A good account of the merits and drawbacks of au<strong>to</strong>matic<br />

translation on the Web is given by Austermühl (<strong>2001</strong>).<br />

Along with the professional team dictionaries, such as the OED, the TLF, or the DUF,<br />

and the specialized lexicons accessible on the Web, there are also <strong>to</strong> be found dictionaries<br />

and glossaries compiled by amateurs and individuals. If one wants, for example, <strong>to</strong><br />

explore the Dublin slang encountered in Roddy Doyle's Barry<strong>to</strong>wn Trilogy, the most<br />

ready source of online information is the O'Byrne Files.<br />

A final word should be reserved for recreational lexicography. The word games of the<br />

parlor, radio, television, books, and the press proliferate on the Web. The OED site<br />

proposes "Word of the Day"; COBUILD has "Idiom of the Day", "The Definitions<br />

Game", and "Cobuild Competition." Many sites offer "Hangman" or "Le Jeu du pendu."<br />

There are various types of "Crossword" or "Mots croisés", "Anagrams" and<br />

"Anagrammes;" Online "Scrabble" has interactive play sites and <strong>to</strong>ol box (dictionary)<br />

sites.<br />

A Case Study in Technological Change<br />

This last section takes a single dictionary-computerization project and looks at the<br />

various technological stages it has gone through over the years. The project in question is<br />

that concerning Jean Nicot's Thresor de la langue françoyse.<br />

(a) Mechanography. When the present writer set out <strong>to</strong> analyze Nicot's Thresor in<br />

Besançon, the technology of the time used <strong>to</strong> manipulate textual data involved a BULL<br />

mechanographical computer using IBM cards and only capable of handling small, simple<br />

corpora such as the plays of Corneille or the poems of Baudelaire. The idea, which<br />

occurred in the 1960s, of putting the Thresor in<strong>to</strong> <strong>digital</strong>, full-text searchable form had <strong>to</strong><br />

wait for technological advances.


(b) Keyboarding, tape-perforation, and s<strong>to</strong>ring on magnetic tape. In 1979, the manual<br />

capture of half of the Thresor was begun at the Institut national de la langue française in<br />

Nancy, followed in 1980 by the commencement of capture of the other half at the<br />

University of Toron<strong>to</strong>. In Nancy, keyboarded data were captured on<strong>to</strong> paper tape and<br />

then transferred <strong>to</strong> magnetic tape; Toron<strong>to</strong> entry was sent directly from a keyboard via a<br />

telephone modem <strong>to</strong> an IBM mainframe computer. Data sent from Nancy <strong>to</strong> Toron<strong>to</strong> by<br />

mail on magnetic tape were made compatible with the Toron<strong>to</strong> input through a variety of<br />

routines written in various languages including Wylbur.<br />

(c) Concordancing on microfiches. In 1984, the complete unified text was run through the<br />

COGS mainframe concordance program written at the University of Toron<strong>to</strong>. Practically<br />

the entire resources of the central computing service of the University of Toron<strong>to</strong> were<br />

reserved for one night <strong>to</strong> index and concord the approximately 900,000 words of the text<br />

of the Thresor. Some of the concordance output was done with Spitbol routines. The<br />

thirty magnetic tapes were output commercially <strong>to</strong> microfiches.<br />

(d) WordCruncher on a standalone. In 1988, the data were transferred via modem and<br />

five-and-a-quarter inch floppy diskettes from the mainframe <strong>to</strong> a midi-computer and<br />

thence <strong>to</strong> an IBM AT personal computer with a 20 MB hard disk. This time it only <strong>to</strong>ok<br />

the resources of one small machine <strong>to</strong> index the full text of the Thresor and create an<br />

interactive concordance database.<br />

(e) The World Wide Web. The Thresor was first put online in 1994 as an interactive<br />

database at the ARTFL Project of the University of Chicago, the ASCII data files being<br />

converted <strong>to</strong> run under the program Philologic. In 2000, ARTFL's Dictionnaires<br />

d'autrefois were installed on a server at the INaLF in Nancy using the Stella interactive<br />

database program. In the same year, the Thresor was put up as an interactive database on<br />

a Windows server at the University of Toron<strong>to</strong> running under TACTweb, the data files<br />

first of all being indexed by TACT on an IBM-compatible. The reference fields – entry<br />

headword, page, and typeface – derive from the tags entered manually during the first<br />

stage of data capture in Nancy and Toron<strong>to</strong>.<br />

Conclusion<br />

The most radical effect that the computer has had on lexicography – from dictionaries on<br />

hard disk or CD-ROM, through <strong>to</strong> dictionaries on the World Wide Web and <strong>to</strong> the Webas-mega-dictionary<br />

– has been <strong>to</strong> supplement the limited number of paths for information<br />

retrieval determined in advance by author and publisher with the infinite number of paths<br />

chosen by the dictionary-user. It is now normal for the user <strong>to</strong> feel in charge of<br />

information retrieval, whether it be through access <strong>to</strong> the full text of a dictionary or the<br />

entire reachable resources of the Web. Headwords have been supplanted by keywords.<br />

References for Further Reading


Austermühl, Frank (<strong>2001</strong>). Electronic Tools for Transla<strong>to</strong>rs. Manchester: St Jerome<br />

Publishing.<br />

ClicNet: Dictionnaires et lexiques [Dictionaries and lexicons] (2002). Carole Netter,<br />

Swarthmore College. At http://clicnet.swarthmore.edu/dictionnaires.html.<br />

Collins COBUILD. At http://titania.cobuild.collins.co.uk/.<br />

Dictionnaire universel francophone en ligne (since 1997). [Online universal francophone<br />

dictionary]. Hachette and AUPELF-UREF. At http://www.francophonie.hachettelivre.fr/.<br />

Dictionnaires d'autrefois [Early dictionaries]. ARTFL, University of Chicago. Accessed<br />

April 5, 2004. At http://www.lib.uchicago.edu/efts/ARTFL/projects/dicos/.<br />

Estienne, Robert (1538). Dictionarium latinogallicum [Latin-French dictionary]. Paris:<br />

R. Estienne.<br />

Estienne, Robert (1539). Dictionaire francoislatin [French-Latin dictionary]. Paris: R.<br />

Estienne.<br />

Glossaire typographique et linguistique (since 1996–7) [Typographical and linguistic<br />

dictionary]. Alis Technologies Inc. Accessed April 5, 2004. At<br />

http://babel.alis.com:8080/glossaire/index.fr.html.<br />

Henry, Françoise (1996). Pour une informatisation du TLF [For a computerization of the<br />

TLF]. In D. Piotrowski (ed.), Lexicographie et informatique: au<strong>to</strong>ur de l'informatisation<br />

du Trésor de la langue française (pp. 79–139). Paris: Didier Érudition.<br />

Ide, Nancy, Jean Véronis, Susan Warwick-Armstrong, and Nicoletta Calzolari (1992).<br />

Principles for Encoding Machine Readable Dictionaries. In H. Tommola, K. Varan<strong>to</strong>la,<br />

T Salmi-Tolonen, and J. Schopp (eds.), Euralex 1992 Proceedings (pp. 239–46).<br />

Tampere: University of Tampere.<br />

Johnson, Samuel (1996). A Dictionary of the English Language on CD-ROM, ed. Anne<br />

McDermott. Cambridge and New York: Cambridge University Press.<br />

Kehoe, Andrew and An<strong>to</strong>inette Renouf (2002). WebCorp: Applying the Web <strong>to</strong><br />

Linguistics and Linguistics <strong>to</strong> the Web. In WWW2002: Eleventh International World<br />

Wide Web Conference. Accessed April 5, 2004. At<br />

http://www2002.org/CDROM/poster/67/.<br />

Knowles, Francis E. (1990). The Computer in Lexicography. In F. J. Hausmann, O.<br />

Reichmann, H. E. Wiegand, and L. Zgusta (eds.), Wörterbucher: Ein Internationales<br />

Handbuch zur Lexicographie, vol. 1 (pp. 1645–72). Berlin and New York: Walter de<br />

Gruyter.


Leximagne - l'empereur des pages dico [The emperor of dictionary pages]. TennesseeBob<br />

Peckham, University of Tennessee-Martin. Accessed April 5, 2004. At<br />

http://www.utm.edu/departments/french/dico.shtml.<br />

Meijs, Willem (1992). Computers and Dictionaries. In Chris<strong>to</strong>pher S. Butler (ed.),<br />

Computers and Written Texts (pp. 141–65). Oxford and Cambridge, MA: Blackwell.<br />

Murray, K. M. Elisabeth (1977). Caught in the Web of Words. New Haven, CT: Yale<br />

University Press.<br />

The O'Byrne Files (since 2000). Accessed April 5, 2004. At<br />

http://homepage.tinet.ie/~nobyrne/slang.html.<br />

Oxford English Dictionary. Accessed April 5, 2004. At http://www.oed.com/.<br />

Renouf, An<strong>to</strong>inette (1994). Corpora and His<strong>to</strong>rical Dictionaries. In I. Lancashire and R.<br />

Wooldridge (eds.), Early Dictionary Databases (pp. 219–35). Toron<strong>to</strong>: Centre for<br />

Computing in the Humanities.<br />

Sinclair, John M., (ed.) (1987). Looking Up. London and Glasgow: Collins.<br />

Terminology of Pediatric Mas<strong>to</strong>cy<strong>to</strong>sis. Mas<strong>to</strong>Kids.org. Accessed April 5, 2004. At<br />

http://www.mas<strong>to</strong>kids.org/index.php?x=terminology.php.<br />

Le Trésor de la langue française informatisé [Computerized treasury of the French<br />

language] (2002). Accessed April 5, 2004. At http://atilf.atilf.fr/tlf.htm.<br />

WebCorp. Accessed April 5, 2004. At http://www.webcorp.org.uk/.<br />

Wooldridge, Russon (1982). Projet de traitement informatique des dictionnaires de<br />

Robert Estienne et de Jean Nicot. [Project for the computerization of the dictionaries of<br />

Robert Estienne and Jean Nicot]. In Manfred Höfler (ed.), La Lexicographie française du<br />

XVI e au XVIII e siècle (pp. 21–32). Wolfenbüttel: Herzog August Bibliothek.<br />

Wooldridge, Russon (1985). Concordance du Thresor de la langue françoyse de Jean<br />

Nicot (1606) [Concordance of Nicot's Thresor de la langue françoyse (1606)]. Toron<strong>to</strong>:<br />

Éditions Paratexte. Wooldridge, Russon (2000). Interactive database of Dictionnaires de<br />

la Renaissance. Accessed April 5, 2004. At<br />

http://www.chass.u<strong>to</strong>ron<strong>to</strong>.ca/~wulfric/dico_tactweb/tiden.htm.<br />

Wooldridge, Russon, Astra Ikse-Vi<strong>to</strong>ls, and Terry Nadasdi (1992). Le Projet CopuLex.<br />

[The CopuLex project]. In R. Wooldridge (ed.), His<strong>to</strong>rical Dictionary Databases (pp.<br />

107–24). Toron<strong>to</strong>: Centre for Computing in the Humanities; and in CH Working Papers,<br />

B.9 (1996). Accessed April 5, 2004. At http://www.chass.u<strong>to</strong>ron<strong>to</strong>.ca/epc/chwp/copulex/.


Wooldridge, Russon, Maryam McCubben, John Planka, and Snejina Sonina (1999).<br />

Enfirouaper dans le World Wide Web [Enfirouaper on the WWW]. Accessed April 5,<br />

2004. At http://www.chass.u<strong>to</strong>ron<strong>to</strong>.ca/~wulfric/lexperimenta/enfirouaper/.<br />

7.<br />

Linguistics Meets Exact Sciences<br />

Jan Hajič<br />

Linguistics is a science that studies human language, both spoken and written. It studies<br />

the structure of language and its usage, the function of its elements, and its relation <strong>to</strong><br />

other bordering sciences (psycholinguistics, sociolinguistics, etc.), of which the most<br />

important related field <strong>to</strong>day is computer science. Diachronic linguistics studies language<br />

from the perspective of its development over time, whereas synchronic linguistics studies<br />

the function and structure of the current live language. General linguistics studies human<br />

language as a system, but particular languages (such as English, Mandarin, Tamil, or any<br />

other) are studied in detail as well. Due <strong>to</strong> its complexity, the study of language is often<br />

divided in<strong>to</strong> several areas. Phonetics and phonology are related <strong>to</strong> speech (or more<br />

precisely, <strong>to</strong> the spoken language), whereas orthography deals with the standard written<br />

form of a particular language, including capitalization and hyphenation where<br />

appropriate. Morphology studies the composition of words by morphemes (prefixes,<br />

roots, suffixes, endings, segmentation in general, etc.) and its relation <strong>to</strong> syntax, which<br />

introduces structure in<strong>to</strong> the description of language at the level of phrases, clauses, and<br />

sentences. Sentences are typically considered <strong>to</strong> be very important basic language units.<br />

Semantics and pragmatics study the relation of the lower levels <strong>to</strong> meaning and contents,<br />

respectively.<br />

A description of a (correct) behavior of a particular language is typically called a<br />

grammar. A grammar usually generalizes: it describes the language structurally and in<br />

terms of broad categories, avoiding the listing of all possible words in all possible clauses<br />

(it is believed that languages are infinite, making such a listing impossible). An English<br />

grammar, for example, states that a sentence consisting of a single clause typically<br />

contains a subject, expressed by a noun phrase, and a verb phrase, expressed by a finite<br />

verb form as a minimum. A grammar refers <strong>to</strong> a lexicon (set of lexical units) containing<br />

word-specific information such as parts of speech (noun, verb, adjective, particle, etc.) or<br />

syntactic subcategorization (e.g., that the verb "<strong>to</strong> attach" has a subject and an indirect<br />

object with the preposition "<strong>to</strong>").<br />

From the his<strong>to</strong>ric perspective, the scientific, economic, and political developments in the<br />

world before and during World War II were preparing the field of linguistics for what<br />

happened shortly thereafter. Natural language moved in<strong>to</strong> the focus of people working in<br />

several scientific fields previously quite distant from linguistics (and other <strong>humanities</strong> as<br />

well): computer science, signal processing, and information theory, supported by<br />

mathematics and statistics. Today, we can say that one of the turning points for linguistics


was Claude Shannon's work (1948). Shannon was an expert in communication theory,<br />

and his work belongs <strong>to</strong> what is known <strong>to</strong>day as information theory, a probabilistic and<br />

statistical description of information contents. However, he was interested not only in the<br />

mathematical aspects of technical communication (such as signal transfer over telegraph<br />

wires), but he and Warren Weaver also tried <strong>to</strong> generalize this approach <strong>to</strong> human<br />

language communication. Although forgotten by many, this work was the first attempt <strong>to</strong><br />

describe the use of natural language by strictly formal (mathematical and statistical, or<br />

s<strong>to</strong>chastic) methods. The recent revival of s<strong>to</strong>chastic methods in linguistics only<br />

underscores the his<strong>to</strong>rical importance of their work.<br />

Developments in theoretical linguistics that led <strong>to</strong> the use of strictly formal methods for<br />

language description were strongly influenced by Ferdinand de Saussure's work (1916).<br />

His work turned the focus of linguists <strong>to</strong> so-called synchronic linguistics (as opposed <strong>to</strong><br />

diachronic linguistics). Based on this shift in paradigm, the language system as a whole<br />

began <strong>to</strong> be studied. Later, Noam Chomsky (even though his view differs from de<br />

Saussure's in many respects) came up with the first systematic formalization of the<br />

description of the sentences of natural language (1957). It should be noted, however, that<br />

Chomsky himself has always emphasized that his motivation for introducing formal<br />

grammar has never been connected with computerization. Moreover, he has renounced<br />

probabilistic and statistical approaches. For various reasons, most notably the lack of<br />

computer power needed for probabilistic and other computationally intensive approaches,<br />

his work stayed dominant in the field of computational linguistics for more than thirty<br />

years.<br />

In line with the formal means he proposed, Chomsky also adopted the view (introduced<br />

by the descriptivist school of American linguists) that sentence structure in natural<br />

language can be represented essentially by recursive bracketing, which puts <strong>to</strong>gether<br />

smaller, immediately adjacent phrases (or so-called constituents) <strong>to</strong> form bigger and<br />

bigger units, eventually leading <strong>to</strong> a treelike sentence structure. An alternative theory<br />

known in essence from the nineteenth century (Becker 1837) but developed in its modern<br />

form by Tesniere (1959) and other European linguists states that the relation between two<br />

sentence constituents is not that of closeness but that of dependence. This theory offers<br />

some more flexibility in expressing relations among constituents that are not immediately<br />

adjacent, and it is also more adequate for a functional view of language structure.<br />

Other formal theories emerged during the 1960s, 1970s, and early 1980s. The generalized<br />

phrase structure grammars (Gazdar) are still close <strong>to</strong> the context-free grammar formalism<br />

as proposed by Chomsky. The formalism later developed in<strong>to</strong> so-called head-driven<br />

phrase structure grammars, a formalism that has characteristics of both the immediate<br />

constituent structure and dependency structure. Similarly, the independently developed<br />

lexicalized functional grammar (Kaplan and Bresnan 1983) explicitly separates the<br />

surface form, or constituent structure (so-called c-structure) from functional structure (fstructure,<br />

which is close <strong>to</strong> the predicate-argument structure) and uses lexical information<br />

heavily.


Chomsky himself proposed a number of modifications of his original theory during the<br />

same period, most notably the government and binding theory (1981), later referred <strong>to</strong> as<br />

the principles and parameters theory, and recently the minimalist theory (1993). While<br />

substantially different from the previous theories and enthusiastically followed by some,<br />

they seem <strong>to</strong> contribute little <strong>to</strong> actual computational linguistics goals, such as building<br />

wide-coverage parsers of naturally occurring sentences. Chomsky's work is thus more<br />

widely used and respected in the field of computer science itself, namely in the area of<br />

formal languages, such as syntax of programming or markup languages.<br />

During the 1980s, s<strong>to</strong>chastic methods (based largely on Shannon and Weaver's work, cf.<br />

above) re-emerged on a higher level, primarily thanks <strong>to</strong> the greatly increased power of<br />

computers and their widespread availability even for university-based research teams.<br />

First, the use of s<strong>to</strong>chastic methods has led <strong>to</strong> significant advances in the area of speech<br />

recognition (Bahl et al. 1983), then it was applied <strong>to</strong> machine translation (Brown et al.<br />

1990), and in the late 1990s, almost every field of computational linguistics was using<br />

s<strong>to</strong>chastic methods for au<strong>to</strong>matic text and speech processing. More complex, more<br />

sophisticated ways of using probability and statistics are now available, and formal nonprobabilistic<br />

means of language descriptions are being merged with the classic<br />

information-theoretic methods, even though we are still waiting for a real breakthrough in<br />

the way the different methods are combined.<br />

Computational Linguistics<br />

Computational linguistics is a field of science that deals with computational processing of<br />

a natural language. On its theoretical side, it draws from modern (formal) linguistics,<br />

mathematics (probability, statistics, information theory, algebra, formal language theory,<br />

etc.), logic, psychology and cognitive science, and theoretical computer science. On the<br />

applied side, it uses mostly the results achieved in modern computer science, user studies,<br />

artificial intelligence and knowledge representation, lexicography (see chapter 6), and<br />

language corpora (see chapter 21). Conversely, the results of computational linguistics<br />

contribute <strong>to</strong> the development of the same fields, most notably lexicography, electronic<br />

publishing (see chapter 35), and access <strong>to</strong> any kind of textual or spoken material in <strong>digital</strong><br />

libraries (see chapter 36).<br />

Computational linguistics can be divided (with many overlappings, of course) in<strong>to</strong><br />

several subfields, although there are often projects that deal with several of them at once.<br />

Theoretical computational linguistics deals with formal theories of language description<br />

at various levels, such as phonology, morphology, syntax (surface shape and underlying<br />

structure), semantics, discourse structure, and lexicology. Accordingly, these subfields<br />

are called computational phonology, computational semantics, etc. Definition of formal<br />

systems of language description also belongs here, as well as certain research directions<br />

combining linguistics with cognitive science and artificial intelligence.<br />

S<strong>to</strong>chastic methods provide the basis for application of probabilistic and s<strong>to</strong>chastic<br />

methods and machine learning <strong>to</strong> the specifics of natural language processing. They use


heavy mathematics from probability theory, statistics, optimization and numerical<br />

mathematics, algebra and even calculus – both discrete and continuous mathematical<br />

disciplines are used.<br />

Applied computational linguistics (not <strong>to</strong> be confused with commercial applications) tries<br />

<strong>to</strong> solve well-defined problems in the area of natural language processing. On the analysis<br />

side, it deals with phonetics and phonology (sounds and phonological structure of words),<br />

morphological analysis (discovering the structure of words and their function), tagging<br />

(disambiguation of part-of-speech and/or morphological function in sentential context),<br />

and parsing (discovering the structure of sentences; parsing can be purely structureoriented<br />

or deep, trying <strong>to</strong> discover the linguistic meaning of the sentence in question).<br />

Word sense disambiguation tries <strong>to</strong> solve polysemy in sentential context, and it is closely<br />

related <strong>to</strong> lexicon creation and use by other applications in computational linguistics. In<br />

text generation, correct sentence form is created from some formal description of<br />

meaning. Language modeling (probabilistic formulation of language correctness) is used<br />

primarily in systems based on s<strong>to</strong>chastic methods. Machine translation usually combines<br />

most of the above <strong>to</strong> provide translation from one natural language <strong>to</strong> another, or even<br />

from many <strong>to</strong> many (see Zarechnak 1979). In information retrieval, the goal is <strong>to</strong> retrieve<br />

complete written or spoken documents from very large collections, possibly across<br />

languages, whereas in information extraction, summarization (finding specific<br />

information in large text collections) plays a key role. Question answering is even more<br />

focused: the system must find an answer <strong>to</strong> a targeted question not just in one, but<br />

possibly in several documents in large as well as small collections of documents. Topic<br />

detection and tracking classifies documents in<strong>to</strong> areas of interest, and it follows a s<strong>to</strong>ry<br />

<strong>to</strong>pic in a document collection over time. Keyword spotting (or, flagging a document that<br />

might be of interest based on preselected keywords) has obvious uses. Finally, dialogue<br />

systems (for man-machine communication) and multi-modal systems (combining<br />

language, gestures, images, etc.) complete the spectrum.<br />

Development <strong>to</strong>ols for linguistic research and applications are needed for quick<br />

pro<strong>to</strong>typing and implementation of systems, especially in the area of applied<br />

computational linguistics as defined in the previous paragraph. Such support consists of<br />

lexicons for natural language processing, <strong>to</strong>ols for morphological processing, tagging,<br />

parsing and other specific processing steps, creation and maintenance of language<br />

corpora (see chapter 21) <strong>to</strong> be used in systems based on s<strong>to</strong>chastic methods, and<br />

annotation schemas and <strong>to</strong>ols.<br />

Despite sharing many of the problems with written language processing, speech (spoken<br />

language) processing is usually considered a separate area of research, apart even from<br />

the field of computational linguistics. Speech processing uses almost exclusively<br />

s<strong>to</strong>chastic methods. It is often referred <strong>to</strong> as au<strong>to</strong>matic speech recognition (ASR), but the<br />

field is wider: it can be subdivided in<strong>to</strong> several subfields, some of them dealing more<br />

with technology, some with applications. Acoustic modeling relates the digitized speech<br />

signal and phonemes as they occur in words, whereas language modeling for speech<br />

recognition shares many common features with language modeling of written language,<br />

but it is used essentially <strong>to</strong> predict what the speaker will utter next based on what she said


in the immediate past. Speaker identification is another obvious application, but it is<br />

typically only loosely related <strong>to</strong> the study of language, since it relies more on acoustic<br />

features. Small-vocabulary speech recognition systems are the basis for the most<br />

successful commercial applications <strong>to</strong>day. For instance, systems for recognizing digits,<br />

typically in difficult acoustic conditions such as when speaking over a telephone line, are<br />

widely used. Speaker-independent speech recognition systems with large vocabularies<br />

(sometimes called dictation) are the focus of the main research direction in au<strong>to</strong>matic<br />

speech recognition. Speaker adaptation (where the speech recognizers are adapted <strong>to</strong> a<br />

particular person's voice) is an important subproblem that is believed <strong>to</strong> move <strong>to</strong>day's<br />

systems <strong>to</strong> a new level of precision. Searching spoken material (information retrieval,<br />

<strong>to</strong>pic detection, keyword spotting, etc.) is similar <strong>to</strong> its text counterpart, but the problems<br />

are more difficult due <strong>to</strong> the lack of perfect au<strong>to</strong>matic transcription systems. A less<br />

difficult, but commercially very viable field is text-<strong>to</strong>-speech synthesis (TTS).<br />

It is clear however that any such subdivision can never be exact; often, a project or<br />

research direction draws upon several of the above fields, and sometimes it uses methods<br />

and algorithms even from unrelated fields, such as physics or biology, since more often<br />

than not their problems share common characteristics.<br />

One very important aspect of research in computational linguistics, as opposed <strong>to</strong> many<br />

other areas in <strong>humanities</strong>, is its ability <strong>to</strong> be evaluated. Usually, a gold-standard result is<br />

prepared in advance, and system results are compared (using a predefined metric) <strong>to</strong> such<br />

a gold-standard (called also a test) dataset. The evaluation metric is usually defined in<br />

terms of the number of errors that the system makes; when this is not possible, some<br />

other measure (such as test data probability) is used. The complement of error rate is<br />

accuracy. If the system produces multiple results, recall (the rate at which the system hits<br />

the correct solution by at least one of its outputs) and precision (the complement of false<br />

system results) have <strong>to</strong> be used, usually combined in<strong>to</strong> a single figure (called F-measure).<br />

Objective, au<strong>to</strong>matic system evaluation entered computational linguistics with the revival<br />

of statistical methods and is considered one of the most important changes in the field<br />

since its inception – it is believed that such evaluation was the driving force in the fast<br />

pace of advances in the recent past.<br />

The Most Recent Results and Advances<br />

Given the long his<strong>to</strong>ry of research in computational linguistics, one might wonder what<br />

the state-of-the-art systems can do for us <strong>to</strong>day, at the beginning of the twenty-first<br />

century. This section summarizes the recent results in some of the subfields of<br />

computational linguistics.<br />

The most successful results have been achieved in the speech area. Speech synthesis and<br />

recognition systems are now used even commercially (see also the next section). Speech<br />

recognizers are evaluated by using so-called word error rate, a relatively simple measure<br />

that essentially counts how many words the system missed or confused. The complement<br />

of word error rate is the accuracy, as usual. The latest speech recognizers can handle<br />

vocabularies of 100,000 or more words (some research systems already contain million-


word vocabularies). Depending on recording conditions, they have up <strong>to</strong> a 95 percent<br />

accuracy rate (with a closely mounted microphone, in a quiet room, speaker-adapted).<br />

Broadcast news is recognized at about 75–90 percent accuracy. Telephone speech (with<br />

specific <strong>to</strong>pic only, smaller vocabulary) gives 30–90 percent accuracy, depending on<br />

vocabulary size and domain definition; the smaller the domain, and therefore the more<br />

restricted the grammar, the better the results (very tiny grammars and vocabularies are<br />

usually used in successful commercial applications). The worst situation <strong>to</strong>day is in the<br />

case of multi-speaker spontaneous speech under, for example, standard video recording<br />

conditions: only 25–45 percent accuracy can be achieved.<br />

It depends very much on the application whether these accuracy rates are sufficient or<br />

not; for example, for a dictation system even 95 percent accuracy might not be enough,<br />

whereas for spoken material information retrieval an accuracy of about 30 <strong>to</strong> 40 percent<br />

is usually sufficient.<br />

In part-of-speech tagging (morphological disambiguation), the accuracy for English<br />

(Brants 2000) has reached 97–98 percent (measuring the tag error rate). However, current<br />

accuracy is substantially lower (85–90 percent) for other languages that are<br />

morphologically more complicated or for which not enough training data for s<strong>to</strong>chastic<br />

methods is available. Part-of-speech tagging is often used for experiments when new<br />

analytical methods are considered because of its simplicity and test data availability.<br />

In parsing, the number of crossing brackets (i.e., wrongly grouped phrases) is used for<br />

measuring the accuracy of parsers producing a constituent sentence structure, and a<br />

dependency error rate is used for dependency parsers. Current state-of-the-art English<br />

parsers (Collins 1999; Charniak <strong>2001</strong>) achieve 92–94 percent combined precision and<br />

recall in the crossing-brackets measure (the number would be similar if dependency<br />

accuracy is measured). Due <strong>to</strong> the lack of training data (treebanks) for other languages,<br />

there are only a few such parsers; the best-performing published result of a foreignlanguage<br />

parser (on Czech, see Collins et al. 1999) achieves 80 percent dependency<br />

accuracy.<br />

Machine translation (or any system that produces free, plain text) is much more difficult<br />

(and expensive) <strong>to</strong> evaluate. Human evaluation is subjective, and it is not even clear what<br />

exactly should be evaluated, unless a specific task and environment of the machine<br />

translation system being evaluated is known. Machine translation evaluation exercises<br />

administered by DARPA (Defense Advanced Research Projects Agency) in the 1990s led<br />

<strong>to</strong> about 60 percent subjective translation quality of both the statistical machine<br />

translation (MT) systems as well as the best commercial systems. Great demand for an<br />

au<strong>to</strong>matic evaluation of machine translation systems by both researchers and users has<br />

led <strong>to</strong> the development of several au<strong>to</strong>mated metrics (most notably, those of Papineni, see<br />

Papineni et al. <strong>2001</strong>) that try <strong>to</strong> simulate human judgment by computing a numeric match<br />

between the system's output and several reference (i.e., human) translations. These<br />

numbers cannot be interpreted directly (current systems do not go over 0.30–0.35 even<br />

for short sentences), but only in relation <strong>to</strong> a similarly computed distance between human


translations (for example, that number is about 0.60 with 4 reference translations; the<br />

higher the number, the better).<br />

Successful Commercial Applications<br />

The success of commercial applications is determined not so much by the absolute<br />

accuracy of the technology itself, but rather by the relative suitability <strong>to</strong> the task at hand.<br />

Speech recognition is currently the most successful among commercial applications of<br />

language processing, using almost exclusively s<strong>to</strong>chastic methods. But even if there are<br />

several dictation systems on the market with accuracy over 95 percent, we are for the<br />

most part still not dictating e-mails <strong>to</strong> our computers, because there are many other<br />

fac<strong>to</strong>rs that make dictation unsuitable: integration with corrections is poor, noisy<br />

environments decrease the accuracy, sometimes dramatically. On the other hand,<br />

scanning broadcast news for spotting certain <strong>to</strong>pics is routinely used by intelligence<br />

agencies. Call routing in the cus<strong>to</strong>mer service departments of major telephone companies<br />

(mostly in the USA) is another example of a successful application of speech recognition<br />

and synthesis: even if the telephone speech recognition is far from perfect, it can direct<br />

most of the cus<strong>to</strong>mer's calls correctly, saving hundreds of frontline telephone opera<strong>to</strong>rs.<br />

Direc<strong>to</strong>ry inquiries and call switching in some companies are now also handled<br />

au<strong>to</strong>matically by speech recognition; and we now have cell phones with ten-number<br />

voice dialing – a trivial, but fashionable application of speech recognition technology –<br />

not <strong>to</strong> mention mass-market <strong>to</strong>ys with similar capabilities.<br />

Successful applications of non-speech natural language processing technology are much<br />

more rare, if we do not count the ubiqui<strong>to</strong>us spelling and grammar checkers in word<br />

processing software. The most obvious example is machine translation. General machine<br />

translation systems, an application that has been being developed at many places for<br />

almost fifty years now, is still lousy in most instances, even if it sort of works for rough<br />

information gathering (on the Web for example); most translation bureaus are using<br />

translation memories that include bilingual and multilingual dictionaries and previously<br />

translated phrases or sentences as a much more effective <strong>to</strong>ol. Only one system –<br />

SYSTRAN – stands out as a general machine translation system, now being used also by<br />

the European Commission for translating among many European languages (Wheeler<br />

1984). Targeted, sharply constrained domain-oriented systems are much more successful:<br />

an excellent example is the Canadian METEO system (Grimaila and Chandioux 1992), in<br />

use since May 24, 1977, for translating weather forecasts between English and French.<br />

Research systems using s<strong>to</strong>chastic methods do surpass current general systems, but a<br />

successful commercial system using them has yet <strong>to</strong> be made.<br />

Searching on the Web is an example of a possible application of natural language<br />

processing, since most of the Web consists of natural language texts. Yet most of the<br />

search engines (with the notable exception of Askjeeves and its products<br />

, which allows queries <strong>to</strong> be posted in plain English) still use simple<br />

string matching, and even commercial search applications do not usually go beyond<br />

simple stemming. This may be sufficient for English, but with the growing proportion of


foreign-language web pages the necessity of more language-aware search techniques will<br />

soon become apparent.<br />

Future Perspectives<br />

Due <strong>to</strong> the existence of corpora, <strong>to</strong> the ascent of s<strong>to</strong>chastic methods and evaluation<br />

techniques, computational linguistics has become mostly an experimental science –<br />

something that can hardly be said about many other branches of <strong>humanities</strong> sciences.<br />

Research applications are now invariably tested against real-world data, virtually<br />

guaranteeing quick progress in all subfields of computational linguistics. However,<br />

natural language is neither a physical nor a mathematical system with deterministic<br />

(albeit unknown) behavior; it remains a social phenomenon that is very difficult <strong>to</strong> handle<br />

au<strong>to</strong>matically and explicitly (and therefore, by computers), regardless of the methods<br />

used. Probability and statistics did not solve the problem in the 1950s (weak computer<br />

power), formal computational linguistics did not solve the problem in the 1960s and<br />

1970s (language seems <strong>to</strong> be <strong>to</strong>o complex <strong>to</strong> be described by mere introspection), and<br />

s<strong>to</strong>chastic methods of the 1980s and 1990s apparently did not solve the problem either<br />

(due <strong>to</strong> the lack of data needed for current data-hungry methods). It is not unreasonable <strong>to</strong><br />

expect that we are now, at the beginning of the twenty-first century, on the verge of<br />

another shift of research paradigm in computational linguistics. Whether it will be more<br />

linguistics (a kind of return <strong>to</strong> the 1960s and 1970s, while certainly not leaving the new<br />

experimental character of the field), or more data (i.e., the advances in computation –<br />

statistics plus huge amounts of textual data which are now becoming available), or neural<br />

networks (a long-term promise and failure at the same time), a combination of all of<br />

those, or something completely different, is an open question.<br />

References for Further Reading<br />

There are excellent textbooks now for those interested in learning the latest developments<br />

in computational linguistics and natural language processing. For speech processing,<br />

Jelinek (2000) is the book of choice; for those interested in (text-oriented) computational<br />

linguistics, Charniak (1996), Manning and Schutze (1999), and Jurafsky and Martin<br />

(2000) are among the best.<br />

Bahl, L. R., F. Jelinek, and R. L. Mercer (1983). A Maximum Likelihood Approach <strong>to</strong><br />

Continuous Speech Recognition. IEEE Transactions on Pattern Analysis and Machine<br />

Intelligence 5, 2: 179–220.<br />

Becker, K. F. (1837). Ausfürliche Deutsche Grammatik als Kommentar der<br />

Shulgrammatik. Zweite Abtheilung [Detailed Grammar of German as Notes <strong>to</strong> the School<br />

Grammar]. Frankfurt am Main: G. F. Kettembeil.<br />

Brants, T. (2000). TnT - A Statistical Part-of-Speech Tagger. In S. Nirenburg,<br />

Proceedings of the 6th ANLP (pp. 224–31). Seattle, WA: ACL.


Brown, P. F., J. Cocke, S. A. Della Pietra, V. J. Della Pietra, F. Jelinek, J. D. Lafferty, R.<br />

L. Mercer, and P. S. Roossin (1990). A Statistical Approach <strong>to</strong> Machine Translation.<br />

Computational Linguistics 16, 2: 79–220.<br />

Charniak, E. (1996). Statistical Language Learning. Cambridge, MA: MIT Press.<br />

Charniak, E. (<strong>2001</strong>). Immediate-head Parsing for Language Models. In N. Reithinger and<br />

G. Satta, Proceedings of the 39th Annual Meeting of the Association for Computational<br />

Linguistics (pp. 116–23). Toulouse: ACL.<br />

Chomsky, A. N. (1957). Syntactic Structures. The Hague: Mou<strong>to</strong>n.<br />

Chomsky, A. N. (1981). Lectures on Government and Binding (The Pisa Lectures).<br />

Dordrecht and Cinnaminson, NJ: Foris.<br />

Chomsky, A. N. (1993). A Minimalist Program for Linguistic Theory. In K. Hale and S.<br />

J. Keyser (eds.), The View from Building 20: Essays in Linguistics in Honor of Sylvain<br />

(pp. 1–52). Cambridge, MA: Bromberger.<br />

Collins, M. (1999). Head-driven Statistical Models for Natural Language Parsing. PhD<br />

dissertation, University of Pennsylvania.<br />

Collins, M., J. Hajič, L. Ramshaw, and C. Tillmann (1999). A Statistical Parser for<br />

Czech. In R. Dale and K. Church, Proceedings of ACL 99 (pp. 505–12). College Park,<br />

MD: ACL.<br />

Gazdar, G. et al. (1985). Generalized Phrase Structure Grammar. Oxford: Blackwell.<br />

Grimaila, A. and J. Chandioux (1992). Made <strong>to</strong> Measure Solutions. In J. New<strong>to</strong>n (ed.),<br />

Computers in Translation, A Practical Appraisal (pp. 33–45). New York: Routledge.<br />

Jelinek, F. (2000). Statistical Methods for Speech Recognition. Cambridge, MA: MIT<br />

Press.<br />

Jurafsky, D. and J. H. Martin (2000). Speech and Language Processing. New York:<br />

Prentice Hall.<br />

Kaplan, R. M., and J. Bresnan (1983). Lexical-functional Grammar: A Formal System for<br />

Grammatical Representation. In J. Bresnan, The Mental Representation of Grammatical<br />

Relations (pp. 173–381). Cambridge, MA: MIT Press.<br />

Manning, C. D. and H. Schütze (1999). Foundations of Statistical Natural Language<br />

Processing. Cambridge, MA: MIT Press.


Papineni, K., S. Roukos, T. Ward, and Wei-Jing Zhu (<strong>2001</strong>). Bleu: A Method for<br />

Au<strong>to</strong>matic Evaluation of Machine Translation. Published as IBM Report RC22176.<br />

York<strong>to</strong>wn Heights, NY: IBM T. J. Watson Research Center.<br />

Pollard, C. and I. Sag (1992). Head-driven Phrase Structure Grammar. Chicago:<br />

University of Chicago Press.<br />

Saussure de, F. (1949). Court de linguistique générale, 4th edn. Paris: Libraire Payot.<br />

Shannon, C. (1948). The Mathematical Theory of Communication. Bell Systems<br />

Technical Journal 27: 398–403.<br />

Tesnière, L. (1959). Elements de Syntaxe Structural. Paris: Editions Klincksieck.<br />

Wheeler, P. J. (1984). Changes and Improvements <strong>to</strong> the European Comission's<br />

SYSTRAN MT System, 1976–1984. Terminologie Bulletin 45: 25–37. European<br />

Commission, Luxembourg.<br />

Zarechnak, M. (1979). The His<strong>to</strong>ry of Machine Translation. In B. Henisz-Dostert, R.<br />

Ross Macdonald, and M. Zarechnak (eds.), Machine Translation. Trends in Linguistics:<br />

Studies and Monographs, vol. 11 (pp. 20–8). The Hague: Mou<strong>to</strong>n.<br />

8.<br />

Literary Studies<br />

Thomas Rommel<br />

The systematic study and analysis of literature dates back <strong>to</strong> the beginnings of literary<br />

"text production"; even the earliest forms of oral literature were practiced in a context of<br />

descriptive and prescriptive aesthetics. With the rise of written literature emerged a canon<br />

of rules that could be applied <strong>to</strong> text in order <strong>to</strong> evaluate its adherence <strong>to</strong> poetic norms<br />

and values, and very soon quantitative and qualitative methods of text analysis were<br />

applied in textual exegesis. But the analysis of literature is traditionally seen as a<br />

subjective procedure. Objectivity, based on empirical evidence, does not seem <strong>to</strong> figure<br />

prominently in studies that elucidate meaning from literary texts. In most studies,<br />

however, some kind of exemplary textual sampling does take place, and scholars<br />

occasionally arrive at value judgments that are based on the observation of frequent<br />

occurrences or the absence of certain textual features. The exact number of occurrences<br />

and/or their distribution in long texts is difficult <strong>to</strong> establish, because literary texts, in<br />

particular novels, make a thorough analysis of every single word or sentence almost<br />

impossible. Empirical evidence that is truly representative for the whole text is extremely<br />

difficult <strong>to</strong> come by, and mainstream literary scholarship has come <strong>to</strong> accept this<br />

limitation as a given fact:


A simultaneous possession by the reader of all the words and images of Middlemarch), À<br />

la recherche du temps perdu, or Ulysses may be posited as an ideal, but such an ideal<br />

manifestly cannot be realized. It is impossible <strong>to</strong> hold so many details in the mind at<br />

once.<br />

(Miller 1968: 23)<br />

The first computer-assisted studies of literature of the 1960s and 1970s used the potential<br />

of electronic media for precisely these purposes – the identification of strings and<br />

patterns in electronic texts. Word lists and concordances, initially printed as books but<br />

later made available in electronic format, <strong>to</strong>o, helped scholars come <strong>to</strong> terms with all<br />

occurrences of observable textual features. The "many details", the complete sets of<br />

textual data of some few works of literature, suddenly became available <strong>to</strong> every scholar.<br />

It was no longer acceptable, as John Burrows pointed out, <strong>to</strong> ignore the potential of<br />

electronic media and <strong>to</strong> continue with textual criticism based on small sets of examples<br />

only, as was common usage in traditional literary criticism: "It is a truth not generally<br />

acknowledged that, in most discussions of works of English fiction, we proceed as if a<br />

third, two-fifths, a half of our material were not really there" (Burrows 1987: 1).<br />

Literary computing was seen <strong>to</strong> remedy this shortcoming, and it has provided substantial<br />

insights in<strong>to</strong> some questions of style and literary theory. Most studies on patterns and<br />

themes that were published in the last twenty years question concepts of text and method,<br />

and by investigating literature with the help of a powerful <strong>to</strong>ol these studies situate<br />

themselves in a context of meta-discourse: the question of "method" remains at the heart<br />

of most electronic analysis of literature.<br />

Seen as a mere <strong>to</strong>ol without any inherent analytical power of its own, the computer in<br />

literary studies enhances the critic's powers of memory electronically, thereby providing<br />

a complete database of findings that meet all predefined patterns or search criteria. As<br />

error-prone manual sampling becomes obsolete, textual analysis as well as the ensuing<br />

interpretation of a text as a whole can be based on a complete survey of all passages that<br />

promise results, no matter how long the text is. Comparative approaches spanning large<br />

literary corpora have become possible, and the proliferation of primary texts in electronic<br />

form has contributed significantly <strong>to</strong> the corpus of available <strong>digital</strong> texts. In order <strong>to</strong> be<br />

successful, literary computing needs <strong>to</strong> use techniques and procedures commonly<br />

associated with the natural sciences and fuse them with <strong>humanities</strong> research, thereby<br />

bringing in<strong>to</strong> contact the Two Cultures: "What we need is a principal use of technology<br />

and criticism <strong>to</strong> form a new kind of literary study absolutely comfortable with scientific<br />

methods yet completely suffused with the values of the <strong>humanities</strong>" (Potter 1989: xxix).<br />

The his<strong>to</strong>ry of literary computing, however, shows that only a limited number of textual<br />

phenomena can be analyzed profitably in the context of quantitative and qualitative<br />

computer-based analyses of style. These phenomena have <strong>to</strong> be linked <strong>to</strong> some surface<br />

features that can be identified by electronic means, usually by some form of pattern<br />

matching. Computers are exceptionally well suited for this kind of analysis, and only<br />

human intuition and insight, in combination with the raw computing power of machines


programmed <strong>to</strong> act as highly specialized electronic <strong>to</strong>ols, can make some texts or textual<br />

problems accessible <strong>to</strong> scholars. As Susan Hockey writes:<br />

In the most useful studies, researchers have used the computer <strong>to</strong> find features of interest<br />

and then examined these instances individually, discarding those that are not relevant,<br />

and perhaps refining the search terms in order <strong>to</strong> find more instances. They have also<br />

situated their project within the broader sphere of criticism on their author or text, and<br />

reflected critically on the methodology used <strong>to</strong> interpret the results.<br />

(Hockey 2000: 84)<br />

The methodological implications of such approaches <strong>to</strong> literary texts accommodate<br />

computer-based and computer-assisted studies within the theoretical framework of<br />

literary-linguistic stylistics. In this context, texts are seen as aesthetic constructs that<br />

achieve a certain effect (on the reader) by stylistic features on the surface structure of the<br />

literary text. These features are sometimes minute details that the reader does not<br />

normally recognize individually but that nevertheless influence the overall impression of<br />

the text. The presence or absence of such features can only be traced efficiently by<br />

electronic means, and while the reader may be left with a feeling of having been<br />

manipulated by the text without really knowing how, the computer can work out<br />

distribution patterns that may help understand how a particular effect is achieved.<br />

"[U]nexpectedly high or low frequencies or occurrences of a feature or some atypical<br />

tendency of co-occurrence are, in their very unexpectedness or atypicality, noteworthy",<br />

Michael Toolan maintains, but then continues that "[elaborate statistical computations are<br />

unlikely <strong>to</strong> be illuminating in these matters of subtle textual effect" (1990: 71). This view,<br />

frequently expounded by scholars who see literary computing more critically, points at<br />

one of the central shortcomings of the discipline: in order <strong>to</strong> be acknowledged by<br />

mainstream criticism, computer-based literary studies need <strong>to</strong> clarify that the computer is<br />

a <strong>to</strong>ol used for a specific result in the initial phases of literary analysis. No final result, let<br />

alone an "interpretation" of a text, can be obtained by computing power alone; human<br />

interpretation is indispensable <strong>to</strong> arrive at meaningful results. And in particular the aim of<br />

the investigation needs <strong>to</strong> be clarified; every "computation in<strong>to</strong> criticism", <strong>to</strong> use<br />

Burrows's term, has <strong>to</strong> provide results that transcend the narrow confines of stylostatistical<br />

exercises.<br />

As for studies of punctuation, sentence length, word length, vocabulary distribution<br />

curves, etc., the numbers have been crunched for about twenty years now. It is clearly<br />

established that the distribution of such features is not random, or normal in the statistical<br />

sense. The extent of such variance from the models has been measured with great<br />

precision. But since no one ever claimed that a literary text was a random phenomenon,<br />

or a statistically normal distribution, it is difficult <strong>to</strong> see the point of the exercise.<br />

(Fortier 1991: 193)<br />

Statistics, in conjunction with quantifiable data and a (supposedly) positivistic attitude<br />

<strong>to</strong>ward textual phenomena, have contributed <strong>to</strong> the image of computer-based literary


analysis as a "difficult" or "marginal" pursuit. And in the context of a shift away from<br />

close reading <strong>to</strong>ward a more theory-oriented approach <strong>to</strong> literary texts, new models of<br />

textuality seemed <strong>to</strong> suggest that literary computing was occupied with fixed meanings<br />

that could be elucidated by counting words and phrases. This image was further enhanced<br />

by references <strong>to</strong> this procedure in literature itself, as David Lodge shows in his novel<br />

Small World:<br />

"What's the use? Let's show him, Josh." And he passed the canister <strong>to</strong> the other guy, who<br />

takes out a spool of tape and fits it on <strong>to</strong> one of the machines. "Come over here", says<br />

Dempsey, and sits me down in front of a kind of typewriter with a TV screen attached.<br />

"With that tape", he said, "we can request the computer <strong>to</strong> supply us with any information<br />

we like about your ideolect." "Come again?" I said. "Your own special, distinctive,<br />

unique way of using the English language. What's your favourite word?" "My favourite<br />

word. I don't have one." "Oh yes you do!" he said. "The word you use most frequently."<br />

The simplistic view of computer-based studies as "counting words" has been a major<br />

fac<strong>to</strong>r for later studies that were seen in this light. Contrary <strong>to</strong> received opinion, studies of<br />

literature that use electronic means are mostly concerned with questions of theory and<br />

method. Especially the notion of what constitutes a "text" and how, therefore, a given<br />

theory of text influences the procedures of analysis and interpretation, form the basis of<br />

every literary analysis.<br />

A literary text, interpreted as an aesthetic construct that achieves a certain effect through<br />

the distribution of words and images, works on various levels. Without highly elaborate<br />

thematic – and therefore by definition interpretative – markup, only surface features of<br />

texts can be analyzed. These surface features are read as significant in that they influence<br />

the reader's understanding and interpretation of the text. This has a number of theoretical<br />

implications: if a literary text carries meaning that can be detected by a method of close<br />

reading, then computer-assisted studies have <strong>to</strong> be seen as a practical extension of the<br />

theories of text that assume that "a" meaning, trapped in certain words and images and<br />

only waiting <strong>to</strong> be elicited by the informed reader, exists in literature. By focusing<br />

primarily on empirical textual data, computer studies of literature tend <strong>to</strong> treat text in a<br />

way that some literary critics see as a reapplication of dated theoretical models:<br />

One might argue that the computer is simply amplifying the critic's powers of perception<br />

and recall in concert with conventional perspectives. This is true, and some applications<br />

of the concept can be viewed as a lateral extension of Formalism, New Criticism,<br />

Structuralism, and so forth.<br />

(Smith 1989: 14)<br />

Most studies, both quantitative and qualitative, published in the context of literary<br />

<strong>humanities</strong> computing after powerful desk<strong>to</strong>p computers became available, tend <strong>to</strong><br />

prioritize empirical data, either in the form of au<strong>to</strong>matically extracted stylistic features, or<br />

as encoded thematic units that are then quantified, mapped, and interpreted.


Most suitable for this kind of literary analysis are studies of repeated structures in texts.<br />

These are usually characters, syllables, words, or phrases that reappear throughout a text<br />

or a collection of texts. These repetitions are frequently recognized by readers as<br />

structural devices that help segment a text, or link passages in texts. Chapters, characters,<br />

locations, thematic units, etc., may thus be connected, parallels can be established, and a<br />

systematic study of textual properties, such as echoes, contributes substantially <strong>to</strong> the<br />

understanding of the intricate setup of (literary) texts. This type of analysis is closely<br />

linked <strong>to</strong> theoretical models of intertextuality used in non-computer-based literary<br />

studies, and here the impact of electronic procedures is felt most acutely. Repetitions and<br />

echoes can be traced throughout a text in a consistent fashion; it takes, however, a sound<br />

theoretical model that allows one, first <strong>to</strong> identify, and then <strong>to</strong> isolate common formal<br />

properties of these textual units. The criterion of reliability and verifiability of results and<br />

findings is all-important in studies of repeated structures, and maps of distribution and<br />

significant presences and absences of textual features are used as the basis for a more<br />

detailed analysis. In this area computer-assisted approaches substantially contributed <strong>to</strong><br />

the understanding of literary texts, and electronic studies of literary texts provided<br />

empirical evidence for the analysis of a broad range of intertextual features.<br />

The methodological problems connected with this kind of approach feature prominently<br />

in nearly all electronic studies of literary texts: how does traditional criticism deal with<br />

formal properties of text, and where do electronic studies deviate from and/or enhance<br />

established techniques. Most computer-assisted studies of literature published in the last<br />

twenty years examine their own theoretical position and the impact of formal(ized)<br />

procedures on literary studies very critically. Nearly all come <strong>to</strong> the conclusion that<br />

rigorous procedures of textual analysis are greatly enhanced by electronic means, and that<br />

the basis for scholarly work with literary texts in areas that can be formalized is best<br />

provided by studies that compile textual evidence on an empirical basis.<br />

The concept of rigorous testing, ideally unbiased by personal preferences or interpretation<br />

by the critic, relies on the assumption that textual properties can be identified and isolated<br />

by au<strong>to</strong>matic means. If au<strong>to</strong>matic procedures cannot be applied, stringent procedures for<br />

the preparation of texts have <strong>to</strong> be designed. It has been, and still is, one of the particular<br />

strengths of most electronic studies of literature that the criteria used in the process of<br />

analysis are situated in a theoretical model of textuality that is based on a critical<br />

examination of the role of the critic and the specific properties of the text.<br />

These textual properties often need <strong>to</strong> be set off against the rest of the text, and here<br />

markup as the most obvious form of "external intervention" plays a leading role. The<br />

importance of markup for literary studies of electronic texts cannot be overestimated,<br />

because the ambiguity of meaning in literature requires at least some interpretative<br />

process by the critic even prior <strong>to</strong> the analysis proper. Words as discrete strings of<br />

characters, sentences, lines, and paragraphs serve as "natural" but by no means value-free<br />

textual segments. Any other instance of disambiguation in the form of thematic markup is<br />

a direct result of a critic's reading of a text, which by definition influences the course of<br />

the analysis. As many computer-based studies have shown, laying open one's criteria for<br />

encoding certain textual features is of prime importance <strong>to</strong> any procedure that aspires <strong>to</strong>


produce quantifiable results. The empirical nature of the data extracted from the<br />

electronic text and then submitted <strong>to</strong> further analysis allows for a far more detailed<br />

interpretation that is indeed based on procedures of close reading, and this "new critical<br />

analyse de texte, as well as the more recent concepts of inter-textuality or Riffaterrian<br />

micro-contexts can lead <strong>to</strong> defensible interpretations only with the addition of the rigour<br />

and precision provided by computer analysis" (Fortier 1991: 194).<br />

As the majority of literary critics still seem reluctant <strong>to</strong> embrace electronic media as a<br />

means of scholarly analysis, literary computing has, right from the very beginning, never<br />

really made an impact on mainstream scholarship. Electronic scholarly editions, on the<br />

contrary, are readily accepted in the academic community and they are rightly seen as<br />

indispensable <strong>to</strong>ols for both teaching and research. But even the proliferation of<br />

electronic texts, some available with highly elaborate markup, did not lead <strong>to</strong> an<br />

increasing number of computer-based studies.<br />

This can no longer be attributed <strong>to</strong> the lack of user-friendly, sophisticated software<br />

specifically designed for the analysis of literature. If early versions of TACT, Word-<br />

Cruncher, OCP, or TuStep required considerable computing expertise, modern versions<br />

of these software <strong>to</strong>ols allow for easy-<strong>to</strong>-use routines that literary scholars without<br />

previous exposure <strong>to</strong> <strong>humanities</strong> computing can master. In addition, dedicated scholarly<br />

software has become very flexible and allows the user <strong>to</strong> dictate the terms of analysis,<br />

rather than superimpose certain routines (word lists, concordances, limited pattern<br />

matching) that would prejudice the analysis.<br />

Early computer-based studies suffered greatly from hardware and software constraints,<br />

and as a result software <strong>to</strong>ols were developed that addressed the specific requirements of<br />

scholarly computing. Although these <strong>to</strong>ols proved remarkably effective and efficient<br />

given that the hardware available for <strong>humanities</strong> computing was rather slow and basic, it<br />

still <strong>to</strong>ok considerable expertise <strong>to</strong> prepare electronic texts and convert them in<strong>to</strong><br />

machine-readable form. As no standardized form of encoding existed until the Text<br />

Encoding Initiative (TEI) was formed, most scholars adopted some system of markup<br />

that reflected their particular research needs. Initially, these systems were nonstandardized,<br />

but later the majority of studies used COCOA tags for markup, but these<br />

systems for the scholarly encoding of literary texts needed <strong>to</strong> be adjusted <strong>to</strong> the specific<br />

software requirements of the programs used for the analysis. Accessing the results of<br />

computer-assisted studies in the form of prin<strong>to</strong>uts was equally cumbersome, and any<br />

statistical evaluation that extended the range of predefined options of standard software<br />

would have <strong>to</strong> be designed specifically for every individual application. Visualization, the<br />

plotting of graphs, or the formatting of tables required considerable expertise and<br />

expensive equipment and was thus mostly unavailable <strong>to</strong> one-person projects.<br />

In the light of these technical difficulties it seemed that once hardware limitations no<br />

longer existed and the computing infrastructure was up <strong>to</strong> the demands of scholarly<br />

computing, the electronic analysis of literature would become a major field of research.<br />

Methodological problems addressed in studies that wanted <strong>to</strong> but could not, for technical<br />

reasons, attempt more demanding tasks that required large sets of data, access <strong>to</strong> a


multitude of different texts and enough computing power <strong>to</strong> scan long texts for strings,<br />

for example, seemed a direct result of technical limitations.<br />

But the three basic requirements, seen as imperative for eventually putting literary<br />

computing on the map of mainstream scholarship, have been met since the early 1960s<br />

and 1970s:<br />

• virtually unlimited access <strong>to</strong> high-quality electronic texts;<br />

• sophisticated software that lets the user define the terms of analysis rather than vice<br />

versa;<br />

• powerful computing equipment that supplies unlimited computing power and s<strong>to</strong>rage<br />

capacity.<br />

Despite impressive advances in both hardware and software development, and although<br />

electronic texts with markup based on the TEI guidelines have become available on the<br />

net, literary computing still remains a marginal pursuit. Scholarly results are presented at<br />

international conferences organized by the Association for Literary and Linguistic<br />

Computing (ALLC) and the Association for Computers and the Humanities (ACH) that<br />

are designed <strong>to</strong> inform humanists with a background in the discipline. The results are<br />

published in scholarly journals (L&LC, Literary and Linguistic Computing;, and CHum,<br />

Computers and the Humanities) but rarely make an impact on mainstream scholarship.<br />

This dilemma has been commented on repeatedly: Thomas Corns, Rosanne Potter, Mark<br />

Olsen, and Paul Fortier show that even the most sophisticated electronic studies of<br />

canonical works of literature failed <strong>to</strong> be seen as contributions <strong>to</strong> the discourse of literary<br />

theory and method. Computer-based literary criticism has not "escaped from the ghet<strong>to</strong> of<br />

specialist periodicals <strong>to</strong> the mainstream of literary periodicals", Corns writes, and<br />

continues that the "tables and graphs and scattergrams and word lists that are so<br />

characteristic of computer-based investigation are entirely absent from mainstream<br />

periodicals" (Corns 1991: 127).<br />

One reason for this, apart from a general aversion <strong>to</strong> all things electronic in traditional<br />

literary criticism, is described by Jerome McGann as the notion of relevance, because<br />

the general field of <strong>humanities</strong> education and scholarship will not take the use of <strong>digital</strong><br />

technology seriously until one demonstrates how its <strong>to</strong>ols improve the ways we explore<br />

and explain aesthetic works – until, that is, they expand our interpretational procedures.<br />

(McGann <strong>2001</strong>: xii)<br />

It is important that computer-assisted studies position themselves in the field of recent<br />

scholarship, take up the theoretical issues of text and textuality, and convey <strong>to</strong> the field of<br />

non-experts that the results merit closer inspection. Computers are not used for the sake<br />

of using new <strong>to</strong>ols, but computers can supplement the critic's work with information that<br />

would normally be unavailable <strong>to</strong> a human reader. Speed, accuracy, unlimited memory,<br />

and the instantaneous access <strong>to</strong> virtually all textual features constitute the strength of the


electronic <strong>to</strong>ol. By tapping in<strong>to</strong> the ever-growing pool of knowledge bases and by linking<br />

texts in ways that allow them <strong>to</strong> be used as huge reposi<strong>to</strong>ries of textual material <strong>to</strong> draw<br />

on, traditional literary criticism can profit substantially from the knowledge and expertise<br />

accumulated in the search for a more rigorous analysis of literature as practiced in<br />

computer-based studies.<br />

By looking at the his<strong>to</strong>ry of literary computing, however, one cannot fail <strong>to</strong> see that most<br />

contributions add significant insight in a very narrow spectrum of literary analysis – in<br />

the area of stylistic studies that focus on textual features. The input of computing in these<br />

studies is limited <strong>to</strong> the preparation and prepara<strong>to</strong>ry analysis of the material under<br />

consideration. No immediate result, of course, can be obtained by the computer, but data<br />

are collected that allow for and require further analysis and interpretation by the<br />

researcher. The results, however, are impressive. Numerous studies of individual, and<br />

collections of, texts show that empirical evidence can be used productively for literary<br />

analysis. The his<strong>to</strong>ry of literary computing shows that the field itself is changing. Stylostatistical<br />

studies of isolated textual phenomena have become more common, even if the<br />

computing aspect does not always figure prominently. More and more scholars use<br />

electronic texts and techniques designed for computing purposes, but the resulting studies<br />

are embedded in the respective areas of traditional research. The methods, <strong>to</strong>ols, and<br />

techniques have thus begun <strong>to</strong> influence literary criticism indirectly.<br />

Right from the very beginning, <strong>humanities</strong> computing has always maintained its multidimensional<br />

character as far as literary genre, socio-cultural context and his<strong>to</strong>ricgeographical<br />

provenance of literary texts is concerned. Studies have focused on poetry,<br />

drama, and narrative from antiquity <strong>to</strong> the present day. Although an emphasis on<br />

literature in English can be observed, texts in other languages have also been analyzed.<br />

The variety of approaches used <strong>to</strong> come <strong>to</strong> terms with heterogeneous textual objects, the<br />

multitude of theoretical backgrounds and models of literature brought <strong>to</strong> bear on studies<br />

that share as a common denomina<strong>to</strong>r neither one single technique nor one "school of<br />

thought", but the application of a common <strong>to</strong>ol, are the strong points of studies of<br />

literature carried out with the help of the computer. Discussions of literary theory,<br />

textuality, and the interdisciplinary nature of computer-assisted literary analysis feature<br />

prominently in modern studies. In this respect, mainstream literary criticism is most open<br />

<strong>to</strong> contributions from a field that is, by its very nature, acutely aware of its own<br />

theoretical position. In the future, the discourse of meta-criticism, however, may be fused<br />

with innovative approaches <strong>to</strong> literary texts. As Jerome McGann points out:<br />

A new level of computer-assisted textual analysis may be achieved through programs that<br />

randomly but systematically deform the texts they search and that submit those<br />

deformations <strong>to</strong> human consideration. Computers are no more able <strong>to</strong> "decode" rich<br />

imaginative texts than human beings are. What they can be made <strong>to</strong> do, however, is<br />

expose textual features that lie outside the usual purview of human readers.<br />

(McGann <strong>2001</strong>: 190–1)<br />

References for Further Reading


Ball, C. N. (1994). Au<strong>to</strong>mated Text Analysis: Cautionary Tales. Literary and Linguistic<br />

Computing 9: 293–302.<br />

Burrows, J. F. (1987). A Computation in<strong>to</strong> Criticism. A Study of Jane Austen's Novels and<br />

an Experiment in Method. Oxford: Oxford University Press.<br />

Burrows, J. F. (1992). Computers and the Study of Literature. In C. S. Butler (ed.),<br />

Computers and Written Texts (pp. 167–204). Oxford: Blackwell.<br />

Busa, R. (1992). Half a Century of Literary Computing: Towards a "New" Philology.<br />

Literary and Linguistic Computing 7: 69–73.<br />

Corns, T. N. (1991). Computers in the Humanities: Methods and Applications in the<br />

Study of English Literature. Literary and Linguistic Computing 6: 127–30.<br />

Feldmann, D., F.-W. Neumann, and T. Rommel, (eds.) (1997). Anglistik im Internet.<br />

Proceedings of the 1996 Erfurt Conference on Computing in the Humanities. Heidelberg:<br />

Carl Winter.<br />

Finneran, R. J., (ed.) (1996). The Literary Text in the Digital Age. Ann Arbor: University<br />

of Michigan Press.<br />

Fortier, P. A. (1991). Theory, Methods and Applications: Some Examples in French<br />

Literature. Literary and Linguistic Computing 6, 192–6.<br />

Fortier, P. A., (ed.) (1993–4). A New Direction for Literary Studies? Computers and the<br />

Humanities 27 (special double issue).<br />

Hockey, S. (1980). A Guide <strong>to</strong> Computer Applications in the Humanities. London:<br />

Duckworth.<br />

Hockey, S. (2000). Electronic Texts in the Humanities. Principles and Practice. Oxford:<br />

Oxford University Press.<br />

Landow, G. P. and P. Delany, (eds.) (1993). The Digital Word: Text-Based Computing in<br />

the Humanities. Cambridge, MA: MIT Press.<br />

McGann, J. (<strong>2001</strong>). Radiant Textuality: Literature After the World Wide Web. New York:<br />

Palgrave.<br />

Miall, D. S., (ed.) (1990). Humanities and the Computer: New Directions. Oxford:<br />

Oxford University Press.<br />

Miller, J. H. (1968). Three Problems of Fictional Form: First-person Narration in David<br />

Copperfield and Huckleberry Finn. In R. H. Pearce (ed.), Experience in the Novel:


Selected Papers from the English Institute (pp. 21–48). New York: Columbia University<br />

Press.<br />

Opas, L. L. and T. Rommel, (eds.) (1995). New Approaches <strong>to</strong> Computer Applications in<br />

Literary Studies. Literary and Linguistic Computing 10: 4.<br />

Ott, W (1978). Metrische Analysen zu Vergil, Bucolica. Tübingen: Niemeyer.<br />

Potter, R. G., (ed.) (1989). Literary Computing and Literary Criticism: Theoretical and<br />

Practical Essays on Theme and Rhe<strong>to</strong>ric. Philadelphia: University of Pennsylvania Press.<br />

Renear, A. (1997). Out of Praxis: Three (Meta) Theories of Textuality. In K. Sutherland<br />

(ed.), Electronic Textuality: Investigations in Method and Theory (pp. 107–26). Oxford:<br />

Oxford University Press.<br />

Robey, D. (1999). Counting Syllables in the Divine Comedy: A Computer Analysis.<br />

Modern Language Review 94: 61–86.<br />

Rommel, T. (1995). "And Trace It in This Poem Every Line." Methoden und Verfahren<br />

computerunterstützter Textanalyse am Beispiel von Lord Byrons Don Juan. Tübingen:<br />

Narr.<br />

Smedt, K. et al., (eds.) (1999). Computing in Humanities Education: A European<br />

Perspective. ACO*HUM Report. Bergen: University of Bergen HIT Center.<br />

Smith, J. B. (1989). Computer Criticism. In R. G. Potter (ed.), Literary Computing and<br />

Literary Criticism: Theoretical and Practical Essays on Theme and Rhe<strong>to</strong>ric (pp. 13–44).<br />

Philadelphia: University of Pennsylvania Press.<br />

Sutherland, K., (ed.) (1997). Electronic Textuality: Investigations in Method and Theory.<br />

Oxford: Oxford University Press.<br />

Toolan, M. (1990). The Stylistics of Fiction: A Literary-Linguistic Approach. London:<br />

Routledge.<br />

9.<br />

Music<br />

Ichiro Fujinaga and Susan Forscher Weiss<br />

Introduction<br />

Technological advances are often accompanied by growing pains. Following the advent<br />

of printing, texts were produced in an effort <strong>to</strong> reach wider audiences. Many of these, in a


wide variety of subjects, especially music, were based on the most popular source<br />

materials. As with many other disciplines, computer applications in music have been<br />

dramatically affected by the advent of the Internet. The best-known phenomenon is the<br />

proliferation of files – mostly popular music – in compressed format known as MP3.<br />

There are also peer-<strong>to</strong>-peer file sharing programs such as Napster. The availability of<br />

sample music through online music s<strong>to</strong>res such as amazon.com and barnesandnoble.com<br />

as well as in other specialized sites such as the Classical Music Archives or the web<br />

pages of professors at numerous universities around the world has changed the lives of<br />

scholars and music listeners in general. In addition, music notation software has<br />

revolutionized the music publishing industry. It allows composers <strong>to</strong> produce<br />

professional-quality scores and scholars <strong>to</strong> prepare critical editions more efficiently and<br />

economically. Optical music recognition (OMR) is still in the process of developing<br />

adequate techniques. Once accomplished, OMR and other newer technologies will<br />

redefine the way musicians are able <strong>to</strong> analyze musical scores.<br />

Perhaps the technology that had the most profound influence on music scholarship has<br />

been the availability of e-mail, discussion lists, and society listservs, providing scholars<br />

and lay people with opportunities <strong>to</strong> exchange ideas and materials without the delay of<br />

traditional modes of communication. This trend is documented as late as 1995 (Troutman<br />

1995).<br />

E-mail has provided scholars with opportunities for communication and the sharing of<br />

ideas. It also has paved the way for the formation of national and international<br />

consortiums. Some professional listservs include the AMS (American Musicological<br />

Society) and SMT (Society of Music Theory) lists and their many offshoots. Hundreds of<br />

websites provide information <strong>to</strong> scholars in various music disciplines. Online versions of<br />

important music reference materials include the New Grove Dictionary of Music and<br />

Musicians, the Reper<strong>to</strong>ire International des Sources Musicales (RISM), the Reper<strong>to</strong>ire<br />

International de Litterature Musicale (RILM), the International Index <strong>to</strong> Music<br />

Periodicals (IIMP), the Music Index, Musical America, and Thesaurus Musicarum<br />

Latinarum, among countless others.<br />

The following review of computer applications in music scholarship will focus on the last<br />

decade, examining developments in a number of areas including notation software, MIDI<br />

(Musical Instrument Data Interchange), and MP3, databases, specialized resources,<br />

computer applications for music scholars, and music information retrieval (MIR)<br />

systems. For previous reviews of the discipline, see Alphonce (1989), Davis (1988), Bent<br />

and Morehen (1978), and Hewlett and Selfridge-Field (1991).<br />

Notation Software<br />

Music notation software has been developed over the course of the last two decades of<br />

the twentieth century, now allowing composers <strong>to</strong> prepare their own music without the<br />

necessity of hiring professional copyists. Before the advent of this technology, individual<br />

parts were manually copied from the score. Large ensemble pieces that require a score for<br />

the conduc<strong>to</strong>r as well as parts for individual musicians can now be produced in record


time and at a fraction of the cost. Once the master score is entered in<strong>to</strong> the computer, the<br />

parts can be generated au<strong>to</strong>matically with a minimal amount of manual editing. This<br />

software is also widely used in the music publishing industry and allows musicologists <strong>to</strong><br />

produce scholarly editions of works much more accurately, efficiently, and economically<br />

than before.<br />

One drawback, and an explanation for the lack of widespread use, is the absence of<br />

standard music representation format, either as de fac<strong>to</strong> proprietary (equivalent <strong>to</strong> Word<br />

in the text world) or as an open standard, such as ASCII. Despite numerous efforts since<br />

the advent of computer applications in music <strong>to</strong> create such a format (Hewlett and<br />

Selfridge-Field 1991), no such standard has emerged. Some may argue that the<br />

proprietary format of the very popular music notation software Finale is becoming a<br />

standard, but others contend that another proprietary format of an increasingly popular<br />

competi<strong>to</strong>r called Sibelius may become the standard of the future. Even if one or both of<br />

these become the standard, without some open formats, equivalent <strong>to</strong> ASCII or RTF,<br />

these files will be <strong>to</strong>o cumbersome for other purposes such as searching and analysis.<br />

Optical Music Recognition<br />

One way <strong>to</strong> create a computer-readable format of music notation is <strong>to</strong> use optical music<br />

recognition software, similar <strong>to</strong> optical character recognition technology. A variety of<br />

research pro<strong>to</strong>cols are being tested and commercial software is becoming available, but<br />

the accuracy of the recognition is highly dependent on the type and quality of the original<br />

document and on the complexity of the music. Newer techniques are being applied that<br />

will enable the recognition of lute tablature, medieval music notation, and music notation<br />

systems from other cultures (Selfridge-Field 1994b).<br />

MIDI and MP3<br />

Outside of music notation, there are two open formats for music that have emerged as<br />

standards: MIDI and MP3. Musical Instrument Data Interchange (MIDI) was developed<br />

by a consortium of music synthesizer manufacturers in the early 1980s. It has become<br />

extremely widespread, becoming the de fac<strong>to</strong> standard for popular music and the world<br />

standard for exchanging information between electronic musical instruments and music<br />

software.<br />

Originally designed <strong>to</strong> capture the gestures of keyboard performers, it does have some<br />

limitations. A MIDI file contains data as <strong>to</strong> when, which, and how fast a key was<br />

depressed. It does not contain the timbre of the instrument or much notational<br />

information such as the precise pitch of a note (A♯ or B♭) or which notes are grouped<br />

<strong>to</strong>gether. Nevertheless, a large amount of music is available in MIDI format on the<br />

Internet and in commercial classical piano music recordings. (There are also acoustic<br />

pianos that can play back MIDI files.)


In general, most traditional music scholars base their research on scores. The majority of<br />

analytical methods and theories are dependent on notated music. Thus, the lack of crucial<br />

notational information in MIDI files renders this format less useful for scholarly work.<br />

The recent explosion in the proliferation of MP3 files can be explained as a convergence<br />

of various fac<strong>to</strong>rs. Among these is the standardization of audio format, a higher<br />

bandwidth access <strong>to</strong> the Internet, a dramatic increase in computing power on the desk<strong>to</strong>p,<br />

and the invention of clever compression methods.<br />

The representation of <strong>digital</strong> audio (as sound) was firmly established in the 1980s by the<br />

invention of compact discs (CDs). The standard requires that the sound is captured<br />

(digitized) 44,100 times per second and each sample is s<strong>to</strong>red as a 16-bit integer.<br />

Although there are a number of competing file formats for audio, these are not difficult <strong>to</strong><br />

interchange because they are basically a series of numbers. The s<strong>to</strong>rage required for one<br />

minute of stereo audio is about 10 MB (megabytes). Even with a relatively high-speed<br />

modem (28.8 KB), it takes a couple of hours <strong>to</strong> download a pop song of three minutes (30<br />

MB). The MP3 compression technology reduces the size of the audio files <strong>to</strong> up <strong>to</strong> onehundredth<br />

of its original size. Finally, a computer powerful enough <strong>to</strong> decompress and<br />

play back the downloaded MP3 file was needed. Home computers in the late 1980s could<br />

barely play back a CD-quality stereo audio. Thanks <strong>to</strong> an almost hundredfold increase in<br />

the computing power of desk<strong>to</strong>p computers within the last decade, <strong>to</strong>day almost any<br />

computer can download, decompress, and play back music while the users perform other<br />

tasks.<br />

Databases<br />

General bibliographical databases<br />

The searchable online bibliographical databases are among the best computer <strong>to</strong>ols for all<br />

music scholars in the last decade. Two of the most useful and important bibliographical<br />

databases for music scholars are RILM (Reper<strong>to</strong>ire International de Litterature Musicale)<br />

and Music Index. RILM Abstracts of Music Literature contains over 275,000 entries from<br />

over 500 scholarly periodicals from 60 countries with 140 languages and includes<br />

original-language titles; title translations in English; full bibliographic information; and<br />

abstracts in English. The Music Index: A Subject-Author Guide <strong>to</strong> Music Periodical<br />

Literature covers 650 publications from 1979 and covers a wider range of publications<br />

including popular music magazines. Both of these databases are subscription-based, but<br />

because of their importance they are available in most music libraries. International Index<br />

<strong>to</strong> Music Periodicals (IIMP) is a relatively new subscription service that indexes more<br />

than 370 international music periodicals with over 60 full-text titles. It also includes<br />

retrospective coverage from over 185 periodicals dating back as far as 1874.<br />

The most important music reference in English has been the Grove's Dictionaries for<br />

generations of music scholars. Now both The New Grove Dictionary of Music and<br />

Musicians and The New Grove Dictionary of Opera are available online and have greatly<br />

enhanced the usability of these valuable sources.


Also, The Concise Oxford Dictionary of Music, The Concise Oxford Dictionary of Opera,<br />

and Who's Who in Opera are now part of the Oxford Reference Online. The utility of this<br />

system is that items can be searched in other Oxford dictionaries. For example, query of<br />

"Beethoven" produces a quote by E. M. Forster: "It will be generally admitted that<br />

Beethoven's Fifth Symphony is the most sublime noise that has ever penetrated in<strong>to</strong> the<br />

ear of man" (Howards End [1910], from The Oxford Dictionary of Twentieth Century<br />

Quotations).<br />

Other major resources include RISM (Réper<strong>to</strong>ire International des Sources Musicales).<br />

"Music manuscripts after 1600" is an annotated index and guide <strong>to</strong> music manuscripts<br />

produced between 1600 and 1850 containing more than 380,000 works by over 18,000<br />

composers found in manuscripts over 595 libraries and archives in 31 countries. The<br />

music manuscript database contains over 506,000 searchable musical incipits, which can<br />

be viewed as musical scores.<br />

RIPM (Réper<strong>to</strong>ire International de la Presse Musicale) contains over 410,000 annotated<br />

records from 127 volumes of nineteenth- and twentieth-century music periodicals from<br />

over 15 countries. Another important source is Doc<strong>to</strong>ral Dissertations in Musicology<br />

(DDM) – Online, which is a database of bibliographic records for completed dissertations<br />

and new dissertation <strong>to</strong>pics in the fields of musicology, music theory, and ethnomusicology,<br />

as well as in related musical, scientific, and humanistic disciplines. Available since<br />

1996, it contains more than 12,000 records.<br />

IIPA (International Index <strong>to</strong> the Performing Arts) Full Text contains 182,550 records<br />

drawn from over 210 periodicals and 33 full-text titles. It also contains retrospective<br />

coverage with citations dating back <strong>to</strong> 1864.<br />

The Canadian Music Periodical Index, although restricted <strong>to</strong> Canadian music periodicals,<br />

should be mentioned because, unlike other databases mentioned above, it is free. A<br />

service provided by the National Library of Canada, it includes nearly 30,000 entries on<br />

articles dating from the late nineteenth century <strong>to</strong> the present day. Some 500 Canadian<br />

music journals, newsletters, and magazines are represented here, almost 200 of which are<br />

currently active and continue <strong>to</strong> be indexed. Another free service provided by the<br />

National Library of Canada is an online Encyclopedia of Music in Canada, available both<br />

in English and French.<br />

Music databases<br />

The Scribe Medieval Music Database (La Trobe University) is a collection of 6,000<br />

musical scores, color images, texts, and bibliographic information on medieval music,<br />

searchable by text or melody. This includes the complete annual cycle of liturgical chant<br />

taken from original medieval sources, and the complete works of selected composers<br />

from the twelfth <strong>to</strong> the fifteenth centuries (Stinson 1992).<br />

CANTUS database, started at Catholic University in the 1980s and now hosted at the<br />

University of Western Ontario, contains indices of Latin ecclesiastical chants in over 70


selected manuscripts and early printed sources of the liturgical Office. It is searchable by<br />

first few words, keywords, chant identification number, or liturgical occasion. Further<br />

downloadable chant resources can be found at Cantus Planus Archiv at the University of<br />

Regensburg.<br />

In development since 1989, Thesaurus Musicarum Latinarum (TML) is a full-text<br />

searchable database (with ASCII text and associated image files) containing a large<br />

corpus of Latin music theory written during the Middle Ages and the Renaissance. A<br />

similar database of texts on Italian music theory and aesthetics is Saggi Musicali Italiani<br />

(SMI) and Texts on Music in English from the Medieval and Early Modern Eras (TME).<br />

Another source for Italian treatises, from the Renaissance and early Baroque, is<br />

Thesaurus Musicarum Italicarum (TMI). One of the features of this collection, not<br />

available in others, is that the many terms are hyperlinked. The CD-ROM version of this<br />

database contains complete facsimiles and they are being made available online.<br />

Themefinder is a collaborative project of Stanford University and Ohio State University.<br />

It consists of three databases: Classical Instrumental Music (10,000 themes), European<br />

Folksongs (7,000 folksongs), and Latin Motets from the sixteenth century (18,000<br />

incipits). A web-based searching interface is provided and the matched themes are<br />

displayed in graphical notation (Kornstdädt 1998).<br />

MELDEX, a music search engine developed at the University of Waika<strong>to</strong>, New Zealand,<br />

forms part of the New Zealand Digital Library. The database consists of 9,400 folk<br />

melodies. This is one of the first search engines that allows query-by-humming, where<br />

search can be accomplished by providing a sung audio file online (Bainbridge 1998).<br />

MuseData at Stanford University is a collection of full-text databases of music for several<br />

composers, including J. S. Bach, Beethoven, Corelli, Handel, Haydn, Mozart, Telemann,<br />

and Vivaldi. It currently contains 2,461 complete movements of 634 classical pieces,<br />

including 185 Bach chorales (Selfridge-Field 1994a).<br />

For scholars working with popular music there are a variety of large databases available.<br />

For MIDI data searchable by title and artist, there are sites that claim <strong>to</strong> index over 1<br />

million MIDI files and 95,000 lyrics. Within classical music there are sites with over<br />

20,000 MIDI files.<br />

One can also find opera libret<strong>to</strong>s (RISM-US Libret<strong>to</strong> Database of over 13,000 libretti),<br />

old English ballads, or even hip-hop lyrics, where one can use "Rhymera<strong>to</strong>r" <strong>to</strong> find<br />

rhyming words online.<br />

Specialized bibliography<br />

The Online Bach Bibliography was first launched on the Web in May 1997 by Tomita<br />

Yo, and ever since it has been updated regularly. At present, it contains over 18,000<br />

records of bibliographical references that are considered useful for the scholarly<br />

discussion of Bach's life and works.


The Beethoven Bibliography Database begun in 1990 currently includes 2,700 records<br />

for books and scores and is available from the Center for Beethoven Studies at San Jose<br />

State University (Elliott 1994).<br />

Other notable databases include:<br />

• OCLC (Online Computer Library Center) Music Library, which is a subset of the<br />

OCLC Online Union Catalogue; it provides citations <strong>to</strong> nearly 800,000 musical sound<br />

recordings.<br />

• National Sound Archive at the British Library, whose catalogue includes entries for<br />

almost 2.5 million recordings, published and unpublished, in all genres from pop, jazz,<br />

classical, and world music, <strong>to</strong> oral his<strong>to</strong>ry, drama and literature, dialect, language, and<br />

wildlife sounds.<br />

• SONIC of Library of Congress. The Library of Congress Recorded Sound Collection<br />

contains some 2.5 million audio recordings including multiple copies. It currently has<br />

350,000 bibliographic records representing roughly 25 percent of the Library's entire<br />

sound recording holding of, among others, 78s, 45s, and radio broadcasts.<br />

• Ufficio Ricerca Fondi Musicali di Milano is a national catalogue of music manuscript<br />

and printed scores <strong>to</strong> 1900, containing over 300,000 entries (Gentili-Tedeschi <strong>2001</strong>).<br />

Sheet music<br />

There are many large collections of North American sheet music available on the Web.<br />

Most of these are scanned images including the cover artwork and the musical score, with<br />

varying amount of metadata; for example, some have subjects for the cover art, and some<br />

have partial lyrics. For example, see the Library of Congress 1820–1885 (62,500 pieces<br />

registered at the Library for copyright), Brown University 1850–1920 (1,305 pieces of<br />

African-American sheet music), Duke University 1850–1920 (3,042 pieces), the Lester<br />

Levy Collection of Sheet Music at the Johns Hopkins University 1780–1960 (29,000<br />

pieces), Mississippi State University (22,000 pieces), University of California at Berkeley<br />

1852–1900 (2,000 pieces published in California). The next step for these collections is<br />

<strong>to</strong> provide audio, and full-text/music searching capabilities. In order <strong>to</strong> achieve these<br />

goals, au<strong>to</strong>mated optical music recognition system would be required.<br />

Computer Application<br />

Perhaps the most intensive utilization of computers in music is by composers who use<br />

them for various functions. The machines may be used <strong>to</strong> create new sounds, create new<br />

instruments and interfaces, imitate existing instruments with fine control, generate new<br />

composition, or train computers <strong>to</strong> listen <strong>to</strong> music for interaction with human performers.<br />

Other research activities in this field include: pitch recognition, tempo/beat induction, and<br />

expressive performance, where the computer attempts <strong>to</strong> imitate musicians. For thorough<br />

introduction and his<strong>to</strong>ry in this area, see the excellent book by Roads (1996). Similar


esearch <strong>to</strong>pics are also investigated by music psychologists and music educa<strong>to</strong>rs, where<br />

computers are used <strong>to</strong> model human musical cognition and abilities. Educa<strong>to</strong>rs are also<br />

interested in using computers <strong>to</strong> aid in music teaching. Music theorists have also actively<br />

used computers for analysis of notated music. Applications include melodic and<br />

harmonic analysis, key-finder, segmentation, and development of interactive graphic<br />

analysis <strong>to</strong>ols.<br />

Music his<strong>to</strong>rians have used computers for a range of tasks. Samples of published works<br />

can be divided by subject area. Medieval and Renaissance applications include examples<br />

of computer-assisted examination and analysis of notational signs in twelfth-century<br />

manuscripts (Loos 1996), of the oral transmission of Old Roman chant (Haas 1997), and<br />

a computer program <strong>to</strong> compare a number of motets with questionable attributions<br />

(Thomas 1999). There are a number of computer analyses of motets by William Byrd and<br />

other English composers in an effort <strong>to</strong> confirm authorship (see Morehen 1992; Wulstan<br />

1992, 1995). There are also systematic studies of vertical sonorities and melodic features<br />

of works by Italian Renaissance theorists and composers (Moll 1996). One example <strong>to</strong>ok<br />

Palestrina's Masses from the Casimiri edition, entered them in<strong>to</strong> a database without text<br />

using a MIDI synthesizer, examined by cus<strong>to</strong>mized programs, and manipulated by<br />

spreadsheet software. Areas of investigation included prolations and contexts of note<br />

spans and rests, including metric and textural aspects; the distribution of pitch classes,<br />

including the effects of mode; and the distribution of individual pitches correlated with<br />

voice ranges. Conclusions revealed that Palestrina's pitch-class collection was<br />

conservative and common voice ranges were narrower than previously reported (Miller<br />

1992).<br />

Computer applications in Baroque music include a database, music processor, and coding<br />

system for analyzing the music of (Telemann Lange 1995); software <strong>to</strong> analyze modality<br />

versus <strong>to</strong>nality in Bach's four-part chorales (Rasmussen 1996); a dissertation that used<br />

statistics and computer routines <strong>to</strong> resolve conflicting attributions in works by Bach's<br />

sons (Knoll 1998); the use of spreadsheets for statistical analysis in text critical study of<br />

Bach's Well-Tempered Clavier (Tomita 1993); and a computer analysis of the variable<br />

role of dissonance and contrapuntal techniques in Corelli Trio Sonatas (Peiper 1996).<br />

Studies of Classical and Romantic era composers include analysis of melodic affinities<br />

and identical melodic elements in music by Mozart and Siissmayr <strong>to</strong> refute claims of the<br />

latter's authorship of portions of Mozart's Requiem (Leeson 1995); programs <strong>to</strong> detect<br />

melodic patterns and search text in Schubert's lieder (Nettheim 2000; Yi 1990); and an<br />

analytical comparison, based on statistics, using relational database and spreadsheet<br />

software, of the final movements of Brahms's symphonies nos. 3 and 4 and Bruckner's<br />

symphonies nos. 3 and 4, with findings that contradict a number of related previous<br />

studies (McAr<strong>to</strong>r 1995).<br />

Musicologists and theorists are applying computer programs <strong>to</strong> study <strong>to</strong>nal, a<strong>to</strong>nal, and<br />

post-<strong>to</strong>nal music of the last century. An example is a study of excerpts from Bartók's Ten<br />

Easy Pieces in which the key is ambiguous. A program was created <strong>to</strong> determine the key


(as recorded in MIDI) on the basis of the relative frequency and duration of pitch classes<br />

(Cooper 1998).<br />

In the area of musical images and iconography, work has been done with the superimposition<br />

of divergent sixteenth- and seventeenth-century prints, such as those in the<br />

William Byrd edition (Brett and Smith <strong>2001</strong>). Other studies focusing on images include<br />

one that describes some post-processing procedures for scanned images in re-establishing<br />

the content of medieval sources for sacred vocal music preserved in the Vatican Library<br />

(Planchart <strong>2001</strong>).<br />

Another, the Digital Image Archive of Medieval Music (DIAMM), provides a new<br />

resource for scholars desiring <strong>to</strong> digitize, archive, and make available images of primary<br />

sources and <strong>to</strong> develop techniques of <strong>digital</strong> image enhancement, or "virtual res<strong>to</strong>ration",<br />

<strong>to</strong> retrieve lost data or improve the legibility of materials that cannot at present be read<br />

(Wathey et al. <strong>2001</strong>).<br />

Another area of research concerns the transcription of tablatures. Software has been<br />

developed <strong>to</strong> facilitate the assignment of voices when transcribing tablature, lute and<br />

guitar, in<strong>to</strong> modern notation (Charnasse 1991; Kelly 1995; Derrien-Peden et al. 1991).<br />

Performance practice is an area in need of further investigation. There is one study that<br />

au<strong>to</strong>matically assigns proper fingering for English keyboard music (1560–1630),<br />

examining sixty surviving manuscripts, half of which contain fingerings. Ten important<br />

works with fingerings are s<strong>to</strong>red and analyzed and then applied <strong>to</strong> those without<br />

fingerings (Morehen 1994). Composer sketches is another area just beginning <strong>to</strong> benefit<br />

from the application of <strong>digital</strong> techniques. A recent study describes pho<strong>to</strong>graphic<br />

techniques used in an electronic facsimile of the sketches for Alban Berg's Wozzeck. By<br />

showing various stages of sketches for one scene from this opera, one of the most<br />

prominent works of the twentieth century, the paper suggests typical uses of an electronic<br />

facsimile for a large body of sketches. The archaeological nature of the task involves the<br />

capture of pencil drafts and paper paste-overs as well as the penetration of lacquer<br />

occlusions (Hall <strong>2001</strong>).<br />

Perhaps one of the most exciting of the recent articles in Computing in Musicology is a<br />

study of watermarks by Dexter Edge. The traditional methods used by music his<strong>to</strong>rians<br />

for imaging watermarks are problematic. Freehand tracing is inherently prone <strong>to</strong><br />

inaccuracy, the Dylux method often seems ineffective with musical manuscripts, and<br />

beta-radiography has become prohibitively expensive. Using a scanner equipped with a<br />

transparency adapter and manipulating the resulting images, Edge is able <strong>to</strong> create <strong>digital</strong><br />

images of watermarks with good results.<br />

Folksong is one category that has inspired a number of studies. A paper on the mapping<br />

of European folksong discusses issues related <strong>to</strong> encoding, displaying, and analyzing<br />

geographical information pertaining <strong>to</strong> music (Aarden and Huron <strong>2001</strong>). Other studies<br />

examine pulse in German folksong, melodic arch in Western folksongs, and con<strong>to</strong>ur<br />

analysis of Hungarian folk music (Nettheim 1993; Huron 1996; Juhasz 2000; Pintér


1999). For further information on application in musicology see Nettheim (1997) and<br />

Selfridge-Field (1992).<br />

Music Information Retrieval<br />

As more and more music is becoming available on the Internet, it has become evident<br />

that there is a need for a method of searching for the required music information.<br />

Although this branch of research is in its initial stage, it is growing rapidly. With<br />

increased interest in <strong>digital</strong> libraries, in general, preservation and retrieval of musical<br />

material both new and old is necessary. Although searching for textual information is<br />

becoming easier, as with image searching, searching for music, in audio format or in<br />

notated format, is still difficult (Byrd and Crawford <strong>2001</strong>).<br />

Topics in this field include: query-by-humming, au<strong>to</strong>matic genre and style classification,<br />

music indexing, melodic similarity, and au<strong>to</strong>matic transcription. The latter, especially<br />

transcribing polyphonic music, remains the "holy grail" of audio engineers.<br />

The copyright issue is still under intense debate (Anthony et al. <strong>2001</strong>).<br />

Although most of these issues have been the academic concerns of music scholars, there<br />

is also an ever-increasing commercial interest in knowing more about music. Record<br />

companies are eager for buyers <strong>to</strong> find music more easily and for the companies <strong>to</strong> be<br />

able <strong>to</strong> classify musical styles and genre so that they can match listeners' preferences. To<br />

accomplish this, companies are asking questions such as "What is a melody or theme?"<br />

and "Where do they occur?" Knowing the answers will aid in discovering new music that<br />

a consumer would like. Thus, with funding from the commercial sec<strong>to</strong>r, there may now<br />

be additional research opportunities for music investiga<strong>to</strong>rs.<br />

Conclusions<br />

This is a very exciting time for music. Technology is making new sounds and<br />

unprecedented access <strong>to</strong> music. As more and more musical data become readily available,<br />

musical research methodology may undergo a fundamental change. As David Huron<br />

hints, the discipline will go from a "data-poor" field <strong>to</strong> a "data-rich" field (Huron 1999).<br />

As of now, significant research has been limited <strong>to</strong> subjects such as composer<br />

attributions, image enhancements, determination of provenance based on watermarks,<br />

and some notational studies. More needs <strong>to</strong> be done. Also, with pressure from music<br />

information retrieval (MIR) areas, there may be greater need for scholars from different<br />

musical disciplines <strong>to</strong> work <strong>to</strong>gether <strong>to</strong> solve a variety of musical puzzles. Whether it is<br />

for musicologists and theorists or members of the business and lay communities, we will<br />

need <strong>to</strong> develop more sophisticated analytical technologies and programs for search and<br />

retrieval of an ever-expanding mass of information and sound.<br />

References for Further Reading


Aarden, B. and D. Huron (<strong>2001</strong>). Mapping European Folksong: Geographical<br />

Localization of Musical Features. Computing in Musicology 12: 169–83.<br />

Alphonce, B. (1989). Computer Applications in Music Research: A Retrospective.<br />

Computers in Music Research 1: 1–74.<br />

Anthony, D., C. Cronin, and E. Selfridge-Field (<strong>2001</strong>). The Electronic Dissemination of<br />

Notated Music: An Overview. Computing in Musicology 12: 135–66.<br />

Bainbridge, D. (1998). MELDEX: A Web-based Melodic Loca<strong>to</strong>r Service. Computing in<br />

Musicology 11: 223–9.<br />

Bent, I. and J. Morehen (1978). Computers in Analysis of Music. Proceedings of the<br />

Royal Society of Music 104: 30–46.<br />

Brett, P. and J. Smith (<strong>2001</strong>). Computer Collation of Divergent Early Prints in the Byrd<br />

Edition. Computing in Musicology 12: 251–60.<br />

Byrd, D. and T. Crawford (<strong>2001</strong>). Problems of Music Information Retrieval in the Real<br />

World. Information Processing and Management 38, 2: 249–220.<br />

Charnassé, H. (1991). ERATTO Software for German Lute Tablatures. Computing in<br />

Musicology 7: 60–2.<br />

Cooper, D. (1998). The Unfolding of Tonality in the Music of Béla Bartók. Music<br />

Analysis 17, 1: 21–220.<br />

Davis, D. S. (1988). Computer Applications in Music. Madison, WI: A-R Editions.<br />

Derrien-Peden, D., I. Kanellos, and J.-F. Maheas (1991). French Sixteenth-century Guitar<br />

Tablatures: Transcriptions and Analysis. Computing in Musicology 7: 62–4.<br />

Edge, D. (<strong>2001</strong>). The Digital Imaging of Watermarks. Computing in Musicology 12: 261–<br />

74.<br />

Elliott, P. (1994). Beethoven Bibliography Online. Computing and Musicology 9: 51–2.<br />

Gentili-Tedeschi, M. (<strong>2001</strong>). Il lavoro dell'Ufficio Ricerca Fondi Musicali di Milano.<br />

Libreria Musicale Italiana 18: 483–7.<br />

Haas, M. (1997). Mündliche Überlieferung und altrömischer Choral: His<strong>to</strong>rische und<br />

analytische computergestützte Untersuchungen. Bern: Lang.<br />

Hall, P. (<strong>2001</strong>). The Making of an Electronic Facsimile: Berg's Sketches for Wozzeck.<br />

Computing in Musicology 12: 275–82.


Hewlett, W B. and Selfridge-Field, E. (1991). Computing in Musicology, 1966–91.<br />

Computers and the Humanities 25, 6: 381–92.<br />

Huron, D. (1996). The Melodic Arch in Western Folksongs. Computing in Musicology 10:<br />

3–23.<br />

Huron, D. (1999). The New Empiricism: Systematic Musicology in a Postmodern Age.<br />

1999 Ernest Bloch Lecture, University of California, Berkeley.<br />

Juhász, Z. (2000). Con<strong>to</strong>ur Analysis of Hungarian Folk Music in a Multidimensional<br />

Metric-Space. Journal of New Music Research 29, 1: 71–220.<br />

Kelly, W. (1995). Calculating Fret Intervals with Spreadsheet Software. American<br />

Lutherie 43: 46–7.<br />

Knoll, M. W. (1998). Which Bach Wrote What? A Cumulative Approach <strong>to</strong> Clarification<br />

of Three Disputed Works. PhD dissertation, University of Michigan.<br />

Kornstädt, A. (1998). Themefinder: A Web-based Melodic Search Tool. Computing in<br />

Musicology 11: 231–6.<br />

Lange, C. (1995). Die Telemann Datenbank: Aspekte der Datenspeicherung und der<br />

Nutzungsmöglichkeiten [The Telemann Database: Aspects of data s<strong>to</strong>rage and possible<br />

uses]. Magdeburger Telemann Studien 13: 128–44.<br />

Leeson, D. N. (1995). Franz Xaver Süssmayr and the Mozart Requiem: A Computer<br />

Analysis of Authorship Based on Melodic Affinity. Mozart-Jahrbuch: 111–53.<br />

Loos, I. (1996). Ein Beispiel der Computeranalyse mittelalterlicher Neumen: Das<br />

Quilisma im Antiphonar U 406 (12. Jh.) [An example of computer analysis of Medieval<br />

neumes: the Quilisma in Antiphoner U 406 (12th c.)]. Musicologica Austriaca 14–15:<br />

173–81.<br />

McAr<strong>to</strong>r, M. J. (1995). Comparison of Design in the Finales of the Symphonies of<br />

Bruckner and Brahms. DMA thesis, Arizona State University.<br />

Miller, E. J. (1992). Aspects of Melodic Construction in the Masses of Palestrina: A<br />

Computer-assisted Study. PhD dissertation, Northwestern University.<br />

Moll, K. N. (1996). Vertical Sonorities in Renaissance Polyphony: A Music-analytic<br />

Application of Spreadsheet Software. Computing in Musicology 10: 59–77.<br />

Morehen, J. (1992). Byrd's Manuscript Motet: A New Perspective. In A. M. Brown and<br />

A. Martin (eds.), Byrd studies (pp). 51–62). Cambridge: Cambridge University Press.


Morehen, J. (1994). Aiding Authentic Performance: A Fingering Databank for<br />

Elizabethan Keyboard Music. Computing in Musicology 9: 81–92.<br />

Nettheim, N. (1993). The Pulse in German Folksong: A Statistical Investigation.<br />

Musikometrika 5: 69–89.<br />

Nettheim, N. (1997). A Bibliography of Statistical Applications in Musicology.<br />

Musicology Australia – Journal of the Musicological Society of Australia 20: 94–107.<br />

Nettheim, N. (2000). Melodic Pattern-detection Using MuSearch in Schubert's Die<br />

schöne Müllerin. Computing in Musicology 11: 159–64.<br />

Peiper, C. E. (1996). Dissonance and Genre in Corelli's Trio Sonatas: A LISP-based<br />

Study of Opp. 1 and 2. Computing in Musicology 10: 34–8.<br />

Pintér, I. (1999). Computer-aided Transcription of Folk Music. Studia Musicologica<br />

Academiae Scientiarum Hungaricae 40, 1–3: 189–209.<br />

Planchart, A. E. (<strong>2001</strong>). Image-enhancement Procedures for Medieval Manuscripts.<br />

Computing in Musicology 12: 241–50.<br />

Rasmussen, S. C. (1996). Modality vs. Tonality in Bach's Chorale Harmonizations.<br />

Computing in Musicology 10: 49–58.<br />

Roads, C. (1996). Computer Music Tu<strong>to</strong>rial. Cambridge, MA: MIT Press.<br />

Selfridge-Field, E. (1992). Music Analysis by Computer. In G. Haus (ed.), Music<br />

Processing (pp. 3–24). Madison, WI: A-R Editions.<br />

Selfridge-Field, E. (1994a). The MuseData Universe: A System of Musical Information.<br />

Computing in Musicology 9: 9–30.<br />

Selfridge-Field, E. (1994b). Optical Recognition of Musical Notation: A Survey of<br />

Current Work. Computing in Musicology 9: 109–5.<br />

Stinson, J. (1992). The SCRIBE Database. Computing in Musicology 8: 65.<br />

Thomas, J. S. (1999). The Sixteenth-century Motet: A Comprehensive Survey of the<br />

Reper<strong>to</strong>ry and Case Studies of the Core Texts, Composers, and Reper<strong>to</strong>ry. PhD<br />

dissertation, University of Cincinnati.<br />

Tomita, Y. (1993). The Spreadsheet in Musicology: An Efficient Working Environment<br />

for Statistical Analysis in Text Critical Study. Musicus 3: 31–7.<br />

Troutman, L. A. (1995). MLA-L: A New Mode of Communication. Fontes Artis Musicae<br />

42, 3: 271–220.


Wathey, A., M. Bent, and J. Feely-McCraig (<strong>2001</strong>). The Art of Virtual Res<strong>to</strong>ration:<br />

Creating the Digital Image Archive of Medieval Music (DIAMM). Computing in<br />

Musicology 12: 227–40.<br />

Wulstan, D. (1992). Birdus Tantum Natus Decorare Magistrum. In A. M. Brown and A.<br />

Martin (eds.), Byrd Studies (pp. 63–82). Cambridge: Cambridge University Press.<br />

Wulstan, D. (1995) Byrd, Tallis, and Ferrabosco. In J. Morehen (ed.), English Choral<br />

Practice C.1400-C.1650: A Memorial Volume <strong>to</strong> Peter Le Huray (pp. 109–2).<br />

Cambridge: Cambridge University Press.<br />

Yi, S. W. (1990). A Theory of Melodic Con<strong>to</strong>ur as Pitch – Time Interaction: The<br />

Linguistic Modeling and Statistical Analysis of Vocal Melodies in Selected "Lied". PhD<br />

dissertation, University of California, Los Angeles.<br />

10.<br />

Multimedia<br />

Geoffrey Rockwell and Andrew Mactavish<br />

How do we think through the new types of media created for the computer? Many names<br />

have emerged <strong>to</strong> describe computer-based forms, such as <strong>digital</strong> media, new media,<br />

hypermedia, or multimedia. In this chapter we will start with multimedia, one possible<br />

name that captures one of the features of the emerging genre.<br />

What is Multimedia?<br />

Thinking through a definition starts with a name. Definitions help bring in<strong>to</strong> view limits<br />

<strong>to</strong> that about which you think. Here are some definitions of "multimedia":<br />

A multimedia computer system is one that is capable of input or output of more than one<br />

medium. Typically, the term is applied <strong>to</strong> systems that support more than one physical<br />

output medium, such as a computer display, video, and audio.<br />

(Blattner and Dannenberg 1992: xxiii)<br />

Blattner and Dannenberg further make the observation that "multimedia systems strive <strong>to</strong><br />

take the best advantage of human senses in order <strong>to</strong> facilitate communication" (1992:<br />

xix). Embedded in their discussion is a view of communication where the communica<strong>to</strong>r<br />

chooses <strong>to</strong> combine the media best suited <strong>to</strong> her communicative goals; therefore,<br />

multimedia, which encompasses other media, provides the greatest breadth of<br />

communicative possibilities.


The Encyclopaedia Britannica Online defines "Interactive Multimedia" as "any<br />

computer-delivered electronic system that allows the user <strong>to</strong> control, combine, and<br />

manipulate different types of media." In this definition the emphasis is placed on<br />

interactivity and the computer control over the delivery of information in different media.<br />

This control includes the release of control <strong>to</strong> the reader or viewer so that they can<br />

participate in the development of meaning through interaction with a multimedia work.<br />

While similar, what is interesting in these definitions is what they are defining. The first<br />

defines a "multimedia system" while the second specifies "interactive multimedia." This<br />

chapter proposes a third and shorter definition that combines many of the features in the<br />

others with a focus on multimedia as a genre of communicative work.<br />

A multimedia work is a computer-based rhe<strong>to</strong>rical artifact in which multiple media are<br />

integrated in<strong>to</strong> an interactive whole.<br />

We can use the parts of this definition <strong>to</strong> analyze multimedia.<br />

Computer-based<br />

The word "multimedia" originally referred <strong>to</strong> works of art that combined multiple<br />

traditional art media, as in a multimedia art installation. By defining multimedia as<br />

"computer-based" such mixed-media works are deliberately excluded. In other words, a<br />

multimedia work is a <strong>digital</strong> work that is accessed through the computer even if parts<br />

were created in analogue form and then digitized for integration on the computer. This<br />

definition also excludes works that might have been created on a computer, like a desk<strong>to</strong>p<br />

publishing file, but are accessed by readers through an analogue medium like print.<br />

Rhe<strong>to</strong>rical artifact<br />

A multimedia work is one designed <strong>to</strong> convince, delight, or instruct in the classical sense<br />

of rhe<strong>to</strong>ric. It is not a work designed for administrative purposes or any collection of data<br />

in different media. Nor is it solely a technological artifact. This is <strong>to</strong> distinguish a<br />

multimedia work, which is a work of human expression, from those works that may<br />

combine media and reside on the computer, but are not designed by humans <strong>to</strong><br />

communicate <strong>to</strong> humans.<br />

Multiple media<br />

Central <strong>to</strong> all definitions of multimedia is the idea that multimedia combines types of<br />

information that traditionally have been considered different media and have therefore<br />

had different traditions of production and distribution. Digitization makes this possible as<br />

the computer s<strong>to</strong>res all information, whatever its original form, as binary <strong>digital</strong> data.<br />

Thus it is possible <strong>to</strong> combine media, especially media that are incompatible in other<br />

means of distribution, like synchronous or time-dependent media (audio and video) and<br />

asynchronous media (text and still images).


Integrated … artistic whole<br />

A multimedia work is not just a random collection of different media gathered<br />

somewhere on the system. By this definition the integration of media is the result of<br />

deliberate artistic imagination aimed at producing a work that has artistic unity, which is<br />

another way of saying that we treat multimedia as unified works that are intended by their<br />

crea<strong>to</strong>r <strong>to</strong> be experienced as a whole. Likewise, consumers of multimedia treat such<br />

works as integrated in their consumption. The art of multimedia consists in how you<br />

integrate media.<br />

Interactive<br />

One of the features of multimedia is the interactivity or the programming that structures<br />

the viewer's experience. Some level of interactivity is assumed in any computer-based<br />

work, but by this definition interactivity becomes a defining feature that helps weave the<br />

multiplicity in<strong>to</strong> a whole. Interactivity is thus important <strong>to</strong> the artistic integrity of<br />

multimedia. We might go further and say that interactivity, in the sense of the<br />

programming that structures the work, is the form that integrates the others.<br />

The names given for multimedia works emphasize different characteristics of these<br />

works. "New media" emphasizes the experience of these works as "new" and different<br />

from existing forms of entertainment and instruction, but new media can also refer <strong>to</strong><br />

media new <strong>to</strong> the twentieth century, including electronic (but not necessarily <strong>digital</strong>)<br />

media like television. "Hypermedia" evolved out of "hypertext" and emphasizes the way<br />

these works are multi-linear labyrinths of information that the user navigates. This name,<br />

however, suggests that all new media are organized as hypertexts with nodes and links,<br />

which is not the case for works like arcade games. While "hypermedia" is a useful term<br />

for those works that make use of hypertext features, "multimedia" emphasizes the<br />

combination of traditional media in<strong>to</strong> rhe<strong>to</strong>rical unities.<br />

Defining multimedia as a way of thinking about the new medium made possible by the<br />

computer runs the risk of fixing a moving target inappropriately. It could turn out that<br />

multimedia works are not a new form of expression, but that they are remediated forms of<br />

existing genres of expression (Bolter and Grusin 1999). These traditional forms, when<br />

represented <strong>digital</strong>ly, are transformed by the limitations and capabilities of the computer.<br />

They can be processed by the computer; they can be transmitted instantaneously over the<br />

Internet without loss of quality; they can be extended with other media annotations; they<br />

can be transcoded from one form <strong>to</strong> another (a text can be visualized or read out as<br />

synthesized audio).<br />

The ways in which traditional media are created, distributed and consumed are also<br />

transformed when represented <strong>digital</strong>ly. Multimedia books are not only bought at<br />

books<strong>to</strong>res and read in bed, they can be distributed over the Internet by an e-text library<br />

for your PDA (personal <strong>digital</strong> assistant) and consumed as concordances with text<br />

analysis <strong>to</strong>ols. In short, even if we think of multimedia as a way of <strong>digital</strong>ly re-editing<br />

(re-encoding) traditional works, there are common limitations and possibilities <strong>to</strong> the


<strong>digital</strong> form. Multimedia works, whether born <strong>digital</strong> or remediated, share common<br />

characteristics including emerging modes of electronic production, distribution, and<br />

consumption. They can be defined as multimedia for the purposes of thinking through the<br />

effects of the merging of multiple media in<strong>to</strong> interactive <strong>digital</strong> works <strong>to</strong> be accessed on<br />

the computer.<br />

What are the Types of Multimedia?<br />

Classifying is a second way of thinking through multimedia, and one that involves<br />

surveying the variety of the phenomena. It is also a common move in any discussion of<br />

multimedia <strong>to</strong> give examples of these types of multimedia, especially <strong>to</strong> make the point<br />

that these types are no longer academic experiments inaccessible <strong>to</strong> the everyday<br />

consumer. The challenge of multimedia <strong>to</strong> the <strong>humanities</strong> is thinking through the variety<br />

of multimedia artifacts and asking about the clusters of works that can be aggregated in<strong>to</strong><br />

types. Here are some examples:<br />

Web hypermedia<br />

The first multimedia works <strong>to</strong> be considered seriously in <strong>humanities</strong> computing circles<br />

were hypertexts like The Dickens Web by George P. Landow, a work created <strong>to</strong> explore<br />

the possibilities for hypertext and multimedia in education. It was an exemplary<br />

educational hypertext that illustrated and informed Landow's theoretical work around<br />

hypertext theory. With the evolution of the World Wide Web as a common means for<br />

distributing and accessing hypertextual information, we now have thousands of<br />

educational and research Web hypertexts, some of which combine multiple media and<br />

can be called hypermedia works. The early technologies of the Web, like HTML, have<br />

been extended with technologies like XML and the Macromedia Flash file format (SWF<br />

for Shockwave-Flash) that make sophisticated interactive graphics and animation<br />

possible.<br />

Computer games<br />

By far the most commercially successful multimedia works are computer games, whose<br />

short but rich his<strong>to</strong>ry is interwoven with the development of multimedia technologies.<br />

Games like Myst (Cyan) introduced consumers of all ages <strong>to</strong> the effective use of images,<br />

animations, and environmental sound <strong>to</strong> create a fictional world characterized by<br />

navigation and puzzle-solving. More recently, advancements in hardware and software<br />

technologies for graphics, audio, animation, and video, and sophisticated artificial<br />

intelligence and physics models are making game worlds look and act more convincing.<br />

Games are normally distributed on CD-ROM or DVD, but the Web is frequently used for<br />

distributing software updates and game demos.<br />

Digital art<br />

Artists have been using multimedia <strong>to</strong> create interactive installations that are controlled<br />

by computers and use multiple media. An example would be David Rokeby's Very


Nervous System (1986–90), an interactive sound installation where the user or a<br />

performer generates sound and music through body movement. These playful works are<br />

exhibited in galleries and museums as works of art that bring multimedia in<strong>to</strong> the<br />

traditions of art exhibition. Other <strong>digital</strong> artists have created Web works that are<br />

submitted <strong>to</strong> online exhibitions like those mounted by the San Francisco Museum of<br />

Modern Art in their E•SPACE, which collects and commissions Web art objects.<br />

Multimedia encyclopedia<br />

Multimedia has been used widely in education and for the presentation of research. A<br />

common form of educational and reference multimedia is the multimedia encyclopedia,<br />

like the Encyclopaedia Britannica Online and Microsoft's Encarta (on CD-ROM).<br />

Multimedia encyclopedias are the logical extension of the print genre, taking advantage<br />

of the computer's capability <strong>to</strong> play time-dependent media like audio, animation, and<br />

video <strong>to</strong> enhance the accessibility of information.<br />

These are but examples of types of multimedia. A proper <strong>to</strong>pology would be based on<br />

criteria. For example, we could classify multimedia works in terms of their perceived use,<br />

from entertainment <strong>to</strong> education. We could look at the means of distribution and the<br />

context of consumption of such works, from free websites that require a high-speed<br />

Internet connection, <strong>to</strong> expensive CD-ROM games that require the latest video cards <strong>to</strong><br />

be playable. We could classify multimedia by the media combined, from remediated<br />

works that take a musical work and add synchronized textual commentary, <strong>to</strong> virtual<br />

spaces that are navigated. Other criteria for classification could be the technologies of<br />

production, the sensory modalities engaged, the type of organization that created the<br />

work, or the type of interactivity.<br />

What is the His<strong>to</strong>ry of Multimedia?<br />

A traditional way of thinking through something that is new is <strong>to</strong> recover its his<strong>to</strong>ries.<br />

The his<strong>to</strong>ries of multimedia are still being negotiated and include the his<strong>to</strong>ries of different<br />

media, the his<strong>to</strong>ry of computing, and the his<strong>to</strong>ry of the critical theories applied <strong>to</strong><br />

multimedia. One his<strong>to</strong>ry of multimedia is the his<strong>to</strong>ry of the personal computer as it<br />

evolved from an institutional machine designed for numerical processing <strong>to</strong> a multimedia<br />

personal computer that most of us can afford. The modern computer as it emerged after<br />

World War II is a general-purpose machine that can be adapted <strong>to</strong> new purposes through<br />

programming and peripherals. The his<strong>to</strong>ry of the computer since the ENIAC (1946) can<br />

be seen as the working out of this idea in different ways, including the techniques for<br />

managing different media. While the first computers were designed solely <strong>to</strong> do scientific<br />

and applied numerical calculations, they were eventually extended <strong>to</strong> handle<br />

alphanumeric strings (text), raster and vec<strong>to</strong>r graphics, audio, moving pictures (video and<br />

animation), and finally, three-dimensional objects and space. Today's personal computer<br />

can handle all these media with the appropriate peripherals, making multimedia<br />

development and consumption available <strong>to</strong> the home user.


Numbers and text<br />

If the first computers were designed for number crunching and data processing for<br />

military, scientific, and then business applications, they soon became adapted <strong>to</strong> text<br />

editing or the manipulation of alphanumeric strings. The first commercial word processor<br />

was the IBM MT/ST (magnetic tape / Selectric typewriter), which was marketed by IBM<br />

as a "word processor" and released in 1964. It s<strong>to</strong>red text on a tape for editing and<br />

reprinting through a Selectric typewriter. A word processor, as opposed <strong>to</strong> a text edi<strong>to</strong>r,<br />

was meant for producing rhe<strong>to</strong>rical documents while text edi<strong>to</strong>rs were for programming<br />

and interacting with the system. By the late 1970s, personal computers had primitive<br />

word processing programs that allowed one <strong>to</strong> enter, edit, and print documents. MicroPro<br />

International's WordStar (1979) was one of the first commercially successful word<br />

processing programs for a personal computer, expanding the media that could be handled<br />

by a home user from numbers <strong>to</strong> text.<br />

Images<br />

The next step was access <strong>to</strong> graphics on a personal computer, a development that came<br />

with the release of the Apple Macin<strong>to</strong>sh in 1984. The Macin<strong>to</strong>sh (Mac), which made<br />

innovations from the Xerox Palo Al<strong>to</strong> Research Center accessible on a commercially<br />

successful personal computer, was designed from the start <strong>to</strong> handle graphics. It came<br />

bundled with a "paint" program, MacPaint, and a mouse for painting and interacting with<br />

the graphical user interface (GUI). While it was not the first computer with graphical<br />

capabilities, it was the first widely available computer with standard graphical<br />

capabilities built-in so that anyone could paint simple images, edit them, print them or<br />

integrate them in<strong>to</strong> other documents like word processing documents created with Mac-<br />

Write, a WYSIWIG (what-you-see-is-what-you-get) word processor also bundled with<br />

the early Macs.<br />

Desk<strong>to</strong>p publishing<br />

In 1986, the capabilities of the Macin<strong>to</strong>sh were extended with the release of the Mac<br />

Plus, Aldus PageMaker and the PostScript capable Apple LaserWriter. The combination<br />

of these three technologies made "desk<strong>to</strong>p publishing" accessible on the personal<br />

computer where before it had been limited <strong>to</strong> very expensive specialized systems. While<br />

MacPaint was a playful <strong>to</strong>ol that could not compete with commercial graphics systems, a<br />

designer outfitted with PageMaker and a LaserWriter could compete with professional<br />

designers working on dedicated typesetting systems for low-end, monochrome publishing<br />

jobs like manuals and newsletters. It was not long before a color-capable Macin<strong>to</strong>sh was<br />

released (the Mac II), which, when combined with image-editing software like Adobe<br />

Pho<strong>to</strong>Shop, helped the Mac replace dedicated systems as the industry standard for<br />

graphic design and publishing. Now, just about any publication, from newspapers <strong>to</strong><br />

glossy annual reports, is created, edited, and proofed on personal computer systems. The<br />

only components still beyond the budget of the home user are the high-resolution <strong>digital</strong><br />

cameras, scanners, and printers necessary <strong>to</strong> produce <strong>to</strong>p-quality publications. But even<br />

these components are slowly moving in<strong>to</strong> the reach of everyday computer users.


Desk<strong>to</strong>p publishing is the precursor <strong>to</strong> multimedia, even though desk<strong>to</strong>p publishing aims<br />

at rhe<strong>to</strong>rical artifacts that are not viewed on a computer. Computer-aided graphic design<br />

and desk<strong>to</strong>p publishing are arts that use computers instead of traditional technologies <strong>to</strong><br />

produce rhe<strong>to</strong>rical artifacts that combine media, such as text and images. The challenge<br />

of combining two media, each with different creative and interpretative traditions,<br />

predates desk<strong>to</strong>p publishing – designers before the computer struggled <strong>to</strong> design the word<br />

and image. What was new, however, was that the personal computer user now had the<br />

opportunity <strong>to</strong> experiment with the design and placement of content in two-dimensional<br />

space. The initial result was a proliferation of horrid, over-designed newsletters and<br />

posters that frequently exhibited unrestrained use of fonts and visual styles.<br />

Authoring environments<br />

Further, the desk<strong>to</strong>p publishing <strong>to</strong>ols were themselves multimedia environments that<br />

provided for the direct manipulation of images and text. Desk<strong>to</strong>p publishing was a<br />

precursor <strong>to</strong> multimedia; desk<strong>to</strong>p publishers typically spent most of their time viewing<br />

the for-print documents they manipulated on the interactive screen, not on paper. Graphic<br />

designers comfortable with design for print (but on a screen) were ready when the first<br />

authoring <strong>to</strong>ols became available for the design of screen-based media. They knew how<br />

<strong>to</strong> work with images and text in the two-dimensional screen space and were competent<br />

with the graphics <strong>to</strong>ols needed <strong>to</strong> lay out and create computer graphics. When Apple<br />

released HyperCard in 1987, the graphics community was positioned <strong>to</strong> take advantage of<br />

their new skills in screen-based design. HyperCard, developed by the crea<strong>to</strong>r of MacPaint<br />

(Andy Hertzfield), was an immediate success, especially since it came free with every<br />

Macin<strong>to</strong>sh and allowed multimedia authors <strong>to</strong> distribute HyperCard stacks without<br />

licensing costs <strong>to</strong> other Macin<strong>to</strong>sh users. Given the high penetration of Macs in schools, it<br />

is not surprising that within a year of the release of HyperCard there were thousands of<br />

simple educational multimedia works that combined text, images, simple animations, and<br />

simple interactivity.<br />

Authoring environments like HyperCard are important <strong>to</strong> the growth of multimedia as<br />

they were easier <strong>to</strong> learn than the programming languages needed previously <strong>to</strong> create<br />

multimedia, and they were designed specifically for the combination of media in<strong>to</strong><br />

interactive works. HyperCard, as its name suggests, was inspired by hypertext theory.<br />

The metaphor of HyperCard was that authors created a stack of cards (nodes of<br />

information), which could have text, graphics, and but<strong>to</strong>ns on them. The but<strong>to</strong>ns were the<br />

hypertext links <strong>to</strong> other cards. HyperCard had a scripting language with which one could<br />

create more complex behaviors or add extensions <strong>to</strong> control other media devices like<br />

audio CDs and videodisk players. One of the most popular computer games of its time,<br />

Myst (1993), was first developed on HyperCard. The card stack metaphor was quickly<br />

imitated by Asymetrix ToolBook, one of the more popular multimedia authoring<br />

environments for the IBM PC. ToolBook's metaphor was a book of pages with text,<br />

graphics, and but<strong>to</strong>ns and it added color capability.<br />

Today, the most popular authoring environments other than HTML edi<strong>to</strong>rs such as<br />

Dreamweaver and GoLive are <strong>to</strong>ols like Macromedia Direc<strong>to</strong>r and Macromedia Flash.


Both Direc<strong>to</strong>r and Flash use a cell and timeline metaphor that evolved out of animation<br />

environments. Flash is used extensively <strong>to</strong> add animations and interactive components <strong>to</strong><br />

websites while Direc<strong>to</strong>r is used for more complex projects that are typically delivered on<br />

a CD-ROM. The Flash file format (SWF) has been published so that other <strong>to</strong>ols can<br />

manipulate SWF.<br />

Sound<br />

The Macin<strong>to</strong>sh also incorporated sound manipulation as a standard feature. The first<br />

Macs released in the mid-1980s had built-in sound capabilities beyond a speaker for<br />

beeps. The 128K Mac had 8-bit mono sound output capability. By 1990, Apple was<br />

bundling microphones with standard Macs. HyperCard could handle audio, though it<br />

could not edit it. The standard Macin<strong>to</strong>sh thus had simple audio capabilities suitable for<br />

interactive multimedia. With the addition of Musical Instrument Digital Interface (MIDI)<br />

controllers and software, Macin<strong>to</strong>shes became popular in the electronic music community<br />

along with the now discontinued Atari ST (1985), which came with a built in MIDI port.<br />

One of the first multimedia works <strong>to</strong> make extensive use of audio was Robert Winter's<br />

interactive Beethoven's Ninth Symphony. This 1989 work came with HyperCard stacks<br />

on floppy disk, which could control a commercial audio CD of Beethoven's Ninth<br />

Symphony. The user could navigate the audio and read critical notes that were<br />

synchronized <strong>to</strong> the symphony.<br />

Digital video<br />

The latest media threshold <strong>to</strong> be overcome in affordable personal computers is <strong>digital</strong><br />

video. The challenge of multimedia is <strong>to</strong> combine not just asynchronous media like text<br />

and images, neither of which need <strong>to</strong> be played over time, but also time-dependent media<br />

like audio, animation, and video. Video puts the greatest stress on computer systems<br />

because of the demands of accessing, processing, and outputting the 29.97 frames-persecond<br />

typical of television-quality video. Only recently, with the introduction of<br />

computers with Fire Wire or IEEE-1394 ports, has it become easy <strong>to</strong> shoot video,<br />

download it <strong>to</strong> the personal computer for editing, and transfer it back <strong>to</strong> tape, CD, or<br />

DVD, or even <strong>to</strong> stream it over the Internet. Given the challenge of integrating video,<br />

there have been some interesting hybrid solutions. One of the first multimedia works, the<br />

Aspen Movie Map (1978), by Andrew Lippman (and others) from what is now called the<br />

MIT Media Lab, combined pho<strong>to</strong>graphs on a videodisk with computer control so that the<br />

user could wander through Aspen, going up and down streets in different seasons. With<br />

the release of <strong>digital</strong> video standards like MPEG (MPEG-1 in 1989, MPEG-2 in 1991)<br />

and Apple QuickTime (1991), it became possible <strong>to</strong> manage video entirely in <strong>digital</strong><br />

form. An early published work that <strong>to</strong>ok advantage of QuickTime was the Voyager CD-<br />

ROM of the Beatles' A Hard Day's Night (1993). This was built around a <strong>digital</strong> video<br />

version of the innovative Beatles' music movie. It is now common for multimedia works<br />

<strong>to</strong> include low-resolution <strong>digital</strong> video elements.


Virtual space and beyond<br />

Current multimedia systems present the user with a two-dimensional graphical user<br />

interface. While such systems can manipulate three-dimensional information (3-D), they<br />

do not typically have the 3-D input and output devices associated with virtual reality<br />

(VR) systems. Is VR the next step in the evolution of the multimedia computer and user<br />

interface? In the 1990s it seemed that cyberspace, as described by William Gibson in<br />

Neuromancer (1984), was the next frontier for multimedia computing. Gibson's vision<br />

was implemented in systems that combine head-tracking systems, data gloves, and 3-D<br />

goggles <strong>to</strong> provide an immersive experience of a virtual space. The metaphor for<br />

computing would no longer be the desk<strong>to</strong>p, but would be virtual spaces filled with<br />

avatars representing people and 3-D objects. The relationship between user and computer<br />

would go from one of direct manipulation of iconographic representations <strong>to</strong> immersion<br />

in a simulated world. Space and structure were the final frontier of multimedia.<br />

While this projected evolution of the multimedia interface is still the subject of academic<br />

research and development, it has been miniaturization and the Internet that have driven<br />

the industry instead. The desk<strong>to</strong>p multimedia systems of the 1990s are now being<br />

repackaged as portable devices that can play multiple media. The keyboard and the<br />

mouse are being replaced by input devices like pen interfaces on personal <strong>digital</strong><br />

assistants (PDAs). Rather than immersing ourselves in virtual caves, we are bringing<br />

multimedia computing out of the office or lab and weaving it in our surroundings. The<br />

challenge <strong>to</strong> multimedia design is how <strong>to</strong> scale interfaces appropriately for hand-held<br />

devices like MP3 players and mobile phones.<br />

What are the Academic Issues in the Study of<br />

Multimedia?<br />

How can we study multimedia in the academy? What are the current issues in multimedia<br />

theory and design? The following are some of the issues that the community is thinking<br />

through.<br />

Best practices in multimedia production<br />

The academic study of multimedia should be distinguished from the craft of multimedia.<br />

Learning <strong>to</strong> create multimedia works is important <strong>to</strong> the study of multimedia in applied<br />

programs, but it is possible <strong>to</strong> study <strong>digital</strong> media in theory without learning <strong>to</strong> make it.<br />

That said, a rich area of academic research is in the study of appropriate practices in<br />

multimedia design. For example, the field of Human Computer Interface (HCI) design is<br />

one area that crosses computer science, information science, psychology, and design.<br />

HCI tends <strong>to</strong> be the scientific study of interface and interactivity. In art and design<br />

schools the issue of interface tends <strong>to</strong> be taken up within the traditions of visual design<br />

and the his<strong>to</strong>ry of commercial design. An important issue for computing humanists<br />

building multimedia is digitization – what <strong>to</strong> digitize, how <strong>to</strong> <strong>digital</strong>ly represent evidence,<br />

and how <strong>to</strong> digitize evidence accurately.


Game criticism and interactivity<br />

If the practice of digitization creates the media that make up multimedia, it is the practice<br />

of combining multiple media in<strong>to</strong> rhe<strong>to</strong>rically effective works that is the play of<br />

multimedia. The possibilities of interactivity are what characterize computer-based<br />

media. In particular, interactive game designers have created complex systems for<br />

interaction with media. For this reason, the emerging field of Digital Game Criticism that<br />

attempts <strong>to</strong> study computer games seriously as popular culture and rhe<strong>to</strong>ric is important<br />

<strong>to</strong> the study of multimedia. What is a game and how can we think of games as forms of<br />

human art? What makes an effective or playable game? What are the possibilities for<br />

playful interaction through the computer? The interactive game may be the paradigmatic<br />

form of multimedia, or for that matter, the paradigmatic form of expression in the <strong>digital</strong><br />

age.<br />

Theories and his<strong>to</strong>ries of multimedia<br />

The study of multimedia as a form of expression has yet <strong>to</strong> develop a theoretical tradition<br />

of its own. Instead, critical theories from existing disciplines are being applied with<br />

increasing ingenuity from film studies <strong>to</strong> literary theory. The very issue of which existing<br />

theoretical traditions can be usefully applied <strong>to</strong> multimedia is a source of debate and<br />

discussion. This essay has taken a philosophical/his<strong>to</strong>rical approach, asking questions<br />

about how <strong>to</strong> think through multimedia. Theorists like Brenda Laurel (Computers as<br />

Theatre, 1991) look at multimedia as dramatic interactions with users. George Landow,<br />

in Hypertext: The Convergence of Contemporary Critical Theory and Technology (1992),<br />

has applied literary theory <strong>to</strong> computing. Lev Manovich, in The Language of New Media<br />

(<strong>2001</strong>), looks at the his<strong>to</strong>rical, social, and cultural continuity of film and new media. In<br />

Hamlet on the Holodeck: The future of Narrative in Cyberspace (1997), Janet H. Murray<br />

considers the new aesthetic possibilities of multimedia within the context of narrative<br />

tradition.<br />

The intersection of technology, communication, and culture has also been a <strong>to</strong>pic of wide<br />

interest. Marshall McLuhan, in Understanding Media (1964), popularized an approach <strong>to</strong><br />

thinking about the effects of technology and media on content. He and others, like Walter<br />

Ong (Orality and Literacy, 1982), draw our attention <strong>to</strong> the profound effects that changes<br />

in communications technology can have on what is communicated and how we think<br />

through communication. Influential industry magazines like Wired take it as a given that<br />

we are going through a communications revolution as significant as the development of<br />

writing or print. There is no shortage of enthusiastic evangelists, like George Gilder (Life<br />

After Television, 1992) and critics like Neil Postman (Technopoly, 1993). There are also<br />

influential popular works on personal computing and media technology – works that have<br />

introduced ideas from the research community in<strong>to</strong> popular culture, like those of Stewart<br />

Brand (The Media Lab, 1987), Howard Rheingold (Tools for Thought, 1985, and Virtual<br />

Communities, 1994), and Nicholas Negroponte (Being Digital, 1995).<br />

Conclusion


There are two ways we can think through multimedia. The first is <strong>to</strong> think about<br />

multimedia through definitions, his<strong>to</strong>ries, examples, and theoretical problems. The<br />

second way is <strong>to</strong> use multimedia <strong>to</strong> think and <strong>to</strong> communicate thought. The academic<br />

study of multimedia is a "thinking-about" that is typically communicated through<br />

academic venues like textbooks, articles, and lectures. "Thinking-with" is the craft of<br />

multimedia that has its own traditions of discourse, forms of organization, <strong>to</strong>ols, and<br />

outcomes. To think-with multimedia is <strong>to</strong> use multimedia <strong>to</strong> explore ideas and <strong>to</strong><br />

communicate them. In a field like multimedia, where what we think about is so new, it is<br />

important <strong>to</strong> think-with. Scholars of multimedia should take seriously the challenge of<br />

creating multimedia as a way of thinking about multimedia and attempt <strong>to</strong> create<br />

exemplary works of multimedia in the traditions of the <strong>humanities</strong>.<br />

References for Further Reading<br />

This bibliography is organized along the lines of the chapter <strong>to</strong> guide readers in further<br />

study.<br />

Introduction <strong>to</strong> Multimedia<br />

Ambron, Sueann and Kristina Hooper, (eds.) (1988). Interactive Multimedia: Visions of<br />

Multimedia for Developers, Educa<strong>to</strong>rs, and Information Providers. Redmond, WA:<br />

Microsoft Press.<br />

Blattner, Meera M. and Roger B. Dannenberg, (eds.) (1992). Multimedia Interface<br />

Design. New York: ACM Press.<br />

Buford, John F. Koegel, (ed.) (1994). Multimedia Systems. New York: Addison-Wesley.<br />

Cot<strong>to</strong>n, Bob, and Richard Oliver (1994). The Cyberspace Lexicon: An Illustrated<br />

Dictionary of Terms from Multimedia <strong>to</strong> Virtual Reality. London: Phaidon.<br />

Cot<strong>to</strong>n, Bob, and Richard Oliver (1997). Understanding Hypermedia 2.000: Multimedia<br />

Origins, Internet Futures. London: Phaidon.<br />

Elliot, John and Tim Worsley, (eds.) (1996). Multimedia: The Complete Guide. Toron<strong>to</strong>:<br />

Élan Press.<br />

Encyclopaedia Britannica. Interactive multimedia. Encyclopaedia Britannica Online.<br />

URL: http://www.search.eb.com/bol/<strong>to</strong>pic?eu=146l&sctn=1. Accessed Oc<strong>to</strong>ber 1999.<br />

Haykin, Randy, (ed.) (1994). Multimedia Demystified: A Guide <strong>to</strong> the World of<br />

Multimedia from Apple Computer, Inc. New York: Random House.<br />

Hofstetter, Fred T. (1995). Multimedia Literacy. New York: McGraw-Hill.


Keyes, Jessica (ed) (1994). The McGraw-Hill Multimedia Handbook. New York:<br />

McGraw-Hill.<br />

Nielsen, Jakob (1995). Multimedia and Hypertext: The Internet and Beyond. Bos<strong>to</strong>n: AP<br />

Professional.<br />

Nyce, J. M. and P. Kahn, (eds.) (1991). From Memex <strong>to</strong> Hypertext. Bos<strong>to</strong>n: Academic<br />

Press. (Includes As We May Think, by Vannevar Bush).<br />

Reisman, Sorel, (ed.) (1994). Multimedia Computing: Preparing for the 21st Century.<br />

Harrisburg and London: Idea Group Publishing.<br />

Tannenbaum, Robert S. (1998). Theoretical Foundations of Multimedia. New York:<br />

Computer Science Press.<br />

Theories and Multimedia<br />

Barrett, Edward, (ed.) (1992). Sociomedia: Multimedia, Hypermedia, and the Social<br />

Construction of Knowledge. Cambridge, MA: MIT Press.<br />

Benjamin, Walter (1968). The Work of Art in the Age of Mechanical Reproduction, tr.<br />

Harry Zohn. In. Hannah Arendt, (ed.), Illuminations (pp. 217–51). New York: Schocken<br />

Books.<br />

Bolter, Jay David and Richard Grusin (1999). Remediation. Cambridge, MA: MIT Press.<br />

Landow, George P. (1992). Hypertext: The Convergence of Contemporary Critical<br />

Theory and Technology. Baltimore: Johns Hopkins University Press.<br />

Landow, George P., (ed.) (1994). Hyper/Text/Theory. Baltimore: Johns Hopkins<br />

University Press.<br />

Laurel, Brenda (1991). Computers as Theatre. New York: Addison-Wesley.<br />

Lévy, Pierre (1997). Cyberculture; Rapport au Conseil de l'Europe. Paris: Edition Odile<br />

Jacob/Editions du Conseil de l'Europe.<br />

Liestøl, Gunnar, et al., (eds.) (2003). Digital Media Revisited: Theoretical and<br />

Conceptual Innovations in Digital Domains. Cambridge, MA: MIT Press.<br />

McLuhan, Marshall (1964). Understanding Media: The Extensions of Man. London:<br />

Routledge.<br />

Manovich, Lev (<strong>2001</strong>). The Language of New Media. Cambridge, MA: MIT Press.


Mitchell, William J. (1992). The Reconfigured Eye: Visual Truth in the Post-<br />

Pho<strong>to</strong>graphic Era. Cambridge, MA: MIT Press.<br />

Ong, Walter J. (1982). Orality and Literacy: The Technologhing of the Word. New York:<br />

Routledge.<br />

Stephens, Mitchell (1998). The Rise of the Image and the Fall of the Word. Oxford:<br />

Oxford University Press.<br />

Interactivity, Interface, and Game Criticism<br />

Aarseth, Espen J. (1997). Cybertext: Perspectives on Ergotic Literature. Baltimore, MD:<br />

Johns Hopkins University Press.<br />

Baecker, Ronald M., et al., (eds.) (1995). Human – Computer Interaction: Toward the<br />

Year 2000, 2nd edn. San Francisco: Morgan Kaufmann.<br />

Birringer, Johannes (1998). Media and Performance: Along the Border. Baltimore: Johns<br />

Hopkins University Press.<br />

Burnham, Van (<strong>2001</strong>). Supercade: A Visual His<strong>to</strong>ry of the Videogame Age 1971–1984.<br />

Cambridge, MA: MIT Press.<br />

Cohen, Scott (1984). Zap! The Rise and Fall of Atari. New York: McGraw-Hill.<br />

Crawford, C. The Art of Computer Game Design. URL: http://www.erasmataz2.com/.<br />

Accessed December 2002.<br />

Huizinga, Johan (1950). Homo Ludens: A Study of the Play-Element in Culture. Bos<strong>to</strong>n:<br />

Beacon Press.<br />

King, Geoff and Tanya Krzywinska, (eds.) (2002). Screen Play:<br />

Cinema/Videogames/Interfaces. New York: Wallflower Press.<br />

King, Lucien, (ed.) (2002). Game On: The His<strong>to</strong>ry and Culture of Video Games. New<br />

York: Universe Publishing.<br />

Laurel, Brenda, (ed.) (1990). The Art of Human–Computer Interface Design. New York:<br />

Addison-Wesley.<br />

Murray, Janet H. (1997). Hamlet on the Holodeck: The Future of Narrative in<br />

Cyberspace. Cambridge, MA: MIT Press.<br />

Norman, Donald (1988). The Psychology of Everyday Things. New York: Basic Books.


Preece, Jenny et al., (eds.) (1994). Human–Computer Interaction. New York: Addison-<br />

Wesley.<br />

Provenzo, Eugene E, Jr. (1991). Video Kids: Making Sense of Nintendo. Cambridge, MA:<br />

Harvard University Press.<br />

Rada, Roy (1995). Interactive Media. New York: Springer-Verlag.<br />

Ryan, Marie-Laure (1997). Interactive Drama: Narrativity in a Highly Interactive<br />

Environment. Modern Fiction Studies 43, 3: 677–707.<br />

Wolf, Mark J., (ed.) (<strong>2001</strong>). The Medium of the Video Game. Austin: University of Texas<br />

Press.<br />

His<strong>to</strong>ry of Computing and Multimedia<br />

A<strong>to</strong>mic Rom. Writing for Multimedia: Great Moments in Multimedia. URL:<br />

http://www.home.earthlink.net/~a<strong>to</strong>mic_rom/moments.htm. Accessed December 2002.<br />

Brand, Stewart (1987). The Media Lab: Inventing the Future at MIT. New York: Viking.<br />

Ceruzzi, Paul E. (1983). Reckoners: The Prehis<strong>to</strong>ry of the Digital Computer, from Relays<br />

<strong>to</strong> the S<strong>to</strong>red Program Concept, 1935–1945. Westport, CT: Greenwood Press.<br />

Ceruzzi, Paul E. (1998). A His<strong>to</strong>ry of Modern Computing. Cambridge, MA: MIT Press.<br />

Freiberger, Paul and Michael Swaine (1984). Fire in the Valley: The Making of the<br />

Personal Computer. Berkeley, CA: Osborne/McGraw-Hill.<br />

Ifrah, Georges (2000). The Universal His<strong>to</strong>ry of Computing: From the Abacus <strong>to</strong> the<br />

Quantum Computer, tr. E. F. Harding. New York: John Wiley.<br />

Kahney, Leander. HyperCard Forgotten, but not Gone. Wired News (August 14, 2002).<br />

URL: http://www.wir(ed.)com/news/mac/0,2125,54365,00.html. Accessed December<br />

2002.<br />

Rheingold, Howard (1985). Tools far Thought: The People and Ideas behind the Next<br />

Computer Revolution. New York: Simon and Schuster.<br />

Digital Art and Design<br />

Lunenfeld, Peter (2000). Snap <strong>to</strong> Grid: A User's Guide <strong>to</strong> Arts, Media, and Cultures.<br />

Cambridge, MA: MIT Press.<br />

Marcus, A. (1991). Graphic Design for Electronic Documents and User Interfaces. New<br />

York: ACM Press/Addison-Wesley.


New Media Encyclopedia. URL: http://www.newmediaarts.org/sommaire/english/sommaire.htm.<br />

Accessed December 2002.<br />

Rokeby, David, http://homepage.mac.com/davidrokeby/home.html. Accessed December<br />

2002.<br />

Rush, Michael (1999). New Media in Late 20th-century Art. New York: Thames and<br />

Hudson.<br />

Schwarz, Hans-Peter (1997). Media-Art-His<strong>to</strong>ry. Munich: Prestel-Verlag.<br />

Velthoven, Willem and Jorinde Seijdel, (eds.) (1996). Multimedia Graphics: The Best of<br />

Global Hyperdesign. San Francisco: Chronicle Books.<br />

Wilson, Stephen (2002). Information Arts: Intersections of Art, Science, and Technology.<br />

Cambridge, MA: MIT Press.<br />

Cyberculture and Multimedia<br />

Gibson, William (1984). Neuromancer. New York: Ace Science Fiction Books.<br />

Gilder, George (1992). Life after Television: The Coming Transformation of Media and<br />

American Life. New York: W. W. Nor<strong>to</strong>n.<br />

Gray, Chris H., (ed.) (1995). The Cyborg Handbook. New York: Routledge.<br />

Heim, Michael (1993). The Metaphysics of Virtual Reality. Oxford: Oxford University<br />

Press.<br />

Negroponte, Nicholas (1995). Being Digital. New York: Alfred A. Knopf.<br />

Postman, Neil (1993). Technopoly: The Surrender of Culture <strong>to</strong> Technology. New York:<br />

Vintage Books.<br />

Rheingold, Howard (1994). Virtual Communities: Homesteading on the Electronic<br />

Frontier. New York: Harper Perennial.<br />

Woolley, Benjamin (1992). Virtual Worlds: A Journey in Hype and Hyperreality.<br />

Oxford: Blackwell.<br />

Selected Technologies and Multimedia Works Mentioned<br />

Adobe. GoLive. URL: http://www.adobe.com/products/golive/main.html. Accessed<br />

December 2002.<br />

Apple. QuickTime. URL: http://www.apple.com/quicktime/. Accessed December 2002.


Click2learn. Toolbook. URL: http://www.asymetrix.com/en/<strong>to</strong>olbook/index/asp.<br />

Accessed December 2002. (Formerly Asymetrix Toolbook.).<br />

Cyan. Myst. URL: http://www.riven.com/home.html. Accessed December 2002.<br />

Macromedia. Flash File Format (SWF). URL:<br />

http://www.macromedia.com/software/flash/open/licensing/fileformat/. Accessed<br />

December 2002.<br />

Macromedia (for information on Dreamweaver, Flash, and Direc<strong>to</strong>r). URL:<br />

http://www.macromedia.com. Accessed December 2002.<br />

Voyager. A Hard Day's Night. URL:<br />

http://voyager.learntech.com/cdrom/catalogpage.cgiPahdn. Accessed December 2002.<br />

11.<br />

Performing Arts<br />

David Z. Saltz<br />

Computers and the performing arts make strange bedfellows. Theater, dance, and<br />

performance art persist as relics of liveness in a media-saturated world. As such, they<br />

stand in defiant opposition <strong>to</strong> the computer's rapacious tendency <strong>to</strong> translate everything<br />

in<strong>to</strong> disembodied <strong>digital</strong> data. Nonetheless, a number of theorists have posited an inherent<br />

kinship between computer technology and the performing arts (Laurel 1991; Saltz 1997).<br />

While "old" media such as print, film, and television traffic in immaterial representations<br />

that can be reproduced endlessly for any number of viewers, the interactivity of "new"<br />

media draws them closer <strong>to</strong> live performance. Every user's interaction with a computer is<br />

a unique "performance", and moreover it is one that, like theater, typically involves an<br />

element of make-believe. When I throw a computer file in the "trash" or "recycling bin", I<br />

behave much like an ac<strong>to</strong>r, performing real actions within an imaginary framework. I<br />

recognize that the "trash bin" on my screen is no more real than a cardboard dagger used<br />

in a play; both are bits of virtual reality. Indeed, theater theorist An<strong>to</strong>nin Artaud coined<br />

the term "virtual reality" <strong>to</strong> describe the illusory nature of characters and objects in the<br />

theater over fifty years before Jaron Lanier first used that term in its computer-related<br />

sense (Artaud 1958: 49).<br />

It is no wonder, then, that performance scholars and practitioners have looked <strong>to</strong> <strong>digital</strong><br />

technology <strong>to</strong> solve age-old problems in scholarship, pedagogy, and creative practice.<br />

This chapter will begin with a review of significant scholarly and pedagogical<br />

applications of computers <strong>to</strong> performance, and then turn <strong>to</strong> artistic applications.<br />

Database Analysis


In 1970, with the number of computers in the world still <strong>to</strong>taling less than 50,000,<br />

Lawrence McNamee <strong>to</strong>uted the potential of computers <strong>to</strong> aid theater research. The<br />

examples that he points <strong>to</strong> as models for the future now seem mundane: a concordance of<br />

Eugene O'Neill's plays and a database of theater dissertations, both published in 1969-<br />

Twelve years later, in his Presidential Address <strong>to</strong> the American Society of Theater<br />

Research, Joseph Donohue highlighted the still largely hypothetical opportunities<br />

computers offered <strong>to</strong> theater scholars, emphasizing the ability of computers <strong>to</strong> order<br />

textual data in<strong>to</strong> "clear, pertinent and discrete" patterns (1981: 139). In the 1980s theater<br />

his<strong>to</strong>rians increasingly turned <strong>to</strong> computers <strong>to</strong> help organize and analyze his<strong>to</strong>rical data.<br />

Edward Mullaly, for example, describes his use of an Apple lie and the database software<br />

DB Master <strong>to</strong> reconstruct the career of a minor nineteenth-century American ac<strong>to</strong>r/<br />

manager from newspaper accounts (1987: 62). These early attempts by performance<br />

scholars <strong>to</strong> tap the power of computers relied on computers' ability <strong>to</strong> crunch textual and<br />

numeric data, such as dramatic dialogue, critical texts about performance, names,<br />

locations, and dates. Such applications do not tackle challenges specific <strong>to</strong> the performing<br />

arts. Using a computer <strong>to</strong> analyze a play or a performance event is no different from<br />

using it <strong>to</strong> analyze any other kind of literary work or his<strong>to</strong>rical phenomenon.<br />

Hypermedia<br />

The performing arts, however, are not exclusively, or even primarily, textual. A work of<br />

dance, theater, or performance art is a visual, audi<strong>to</strong>ry, and, most of all, corporeal event.<br />

Only in the 1980s, when low-cost personal computers acquired the ability <strong>to</strong> s<strong>to</strong>re and<br />

manipulate images, sounds, and finally video, did computers begin <strong>to</strong> offer an effective<br />

way <strong>to</strong> represent the phenomenon of performance. Larry Friedlander's Shakespeare<br />

Project anticipated many subsequent applications of <strong>digital</strong> technology <strong>to</strong> performance<br />

pedagogy. Friedlander began <strong>to</strong> develop the Shakespeare Project in 1984 using an IBM<br />

Info Window system. He adopted HyperCard in 1987 while the software was still in<br />

development at Apple. Because personal computers then had very crude graphics<br />

capabilities and no video, Friedlander adopted a two-screen solution, with the computer<br />

providing random access <strong>to</strong> media s<strong>to</strong>red on a laserdisk. The laserdisk contained<br />

hundreds of still images and, more important, six video segments, including two<br />

contrasting filmed versions of one scene each from Hamlet, Macbeth, and King Lear. The<br />

Shakespeare Project used this video material in three ways. In a Performance area,<br />

students could read the Shakespearean text alongside the video, switch between film<br />

versions at any time, jump <strong>to</strong> any point in the text, and alternate between a film's original<br />

audio track and a recording of Friedlander's interpretation of the ac<strong>to</strong>rs'"subtext." In a<br />

Study area, students participated in interactive tu<strong>to</strong>rials covering aspects of<br />

Shakespearean performance such as characterization and verse. Finally, in a Notebook<br />

area, students could extract <strong>digital</strong> video excerpts <strong>to</strong> incorporate in<strong>to</strong> their own essays. In<br />

each case, the computer made it possible for students <strong>to</strong> read a performance almost as<br />

closely and flexibly as they could a printed text.<br />

CD-ROM editions of plays released in the 1990s – most notably the 1994 Voyager<br />

edition of Macbeth and the 1997 Annenberg/CPB edition of Ibsen's A Doll's House -<br />

incorporate core elements of Friedlander's design, keying the play's script <strong>to</strong> multiple


filmed versions of select scenes. In addition, these CD-ROMs provide a rich assortment<br />

of critical resources and still images. The Voyager Macbeth also includes an audio<br />

recording of the entire play by the Royal Shakespeare Company and a karaoke feature<br />

that allows the user <strong>to</strong> perform a role opposite the audio. These CD-ROMs take<br />

advantage of the ability acquired by personal computers in the 1990s <strong>to</strong> display video<br />

directly, obviating the need for a laserdisk player and second moni<strong>to</strong>r. This approach is<br />

far more elegant, compact, and cost-efficient than using laserdisks, but the video in these<br />

early CD-ROM titles is much smaller and lower in quality than that of a laserdisk. By<br />

2000, faster computer processors and video cards, along with more efficient video<br />

compression schemes and widespread DVD technology, had finally closed the gap<br />

between personal computers and laserdisk players.<br />

Theater Models<br />

The projects considered so far rely on the multimedia capabilities of computers; that is, a<br />

computer's ability <strong>to</strong> s<strong>to</strong>re and retrieve text, images, and audio. Other projects have<br />

exploited the power of computers <strong>to</strong> generate complex simulations of 3-D reality.<br />

Performance scholars began <strong>to</strong> explore the use of 3-D modeling in the mid-1980s <strong>to</strong><br />

visualize hypotheses about his<strong>to</strong>rical theater buildings and staging practices. In 1984,<br />

Robert Golder constructed a 3-D computer model of the 1644 Théatre du Marais, and<br />

Robert Sarlós used computer models <strong>to</strong> visualize staging strategies for a real-world<br />

recreation of the medieval Passion Play of Lucerne. This technique became more<br />

common in the 1990s when high-end computer-assisted design (CAD) software became<br />

available for personal computers. Theater his<strong>to</strong>rians used 3-D modeling software <strong>to</strong><br />

reconstruct structures such as the fifth-century bce Theater of Dionysos (Didaskalia<br />

website) and Richelieu's Palais Cardinal theater (Williford 2000). One of the most<br />

ambitious of these projects is an international effort, led by Richard Beacham and James<br />

Packer, <strong>to</strong> reconstruct the 55 bce Roman theater of Pompey. The computer model is<br />

painstakingly detailed, with every con<strong>to</strong>ur of every column and frieze being modeled in<br />

three dimensions. As a result, even using state-of-the-art graphics workstations, a single<br />

frame takes approximately one hour <strong>to</strong> render at screen resolution (Denard 2002: 36).<br />

None of the 3-D modeling projects described above allow a user <strong>to</strong> navigate the virtual<br />

spaces in real time; the models are experienced only as a series of still images or prerendered<br />

animations. These projects are geared <strong>to</strong>ward research, with the goal of<br />

generating new his<strong>to</strong>rical knowledge and testing hypotheses. Consequently, the quality of<br />

the data is more important than the experience of the user. When the emphasis is on<br />

teaching rather than research, however, the tendency is <strong>to</strong> make the opposite trade-off,<br />

favoring interactivity over detail and precision. The most significant effort along those<br />

lines is the THEATRON project, also under the direction of Richard Beacham, with<br />

funding from the European Commission. THEATRON uses Virtual Reality Modeling<br />

Language (VRML) <strong>to</strong> allow people <strong>to</strong> explore models of his<strong>to</strong>rically significant theater<br />

structures over the Web. The first set of walkthroughs, including such structures as the<br />

Ancient Greek theater of Epidauros, Shakespeare's Globe and the Bayreuth Festspielhaus,<br />

became commercially available in 2002.


The THEATRON walkthroughs provide an experience of immersion, conveying a clear<br />

sense of the scale and configuration of the theater spaces. These spaces, however, are<br />

empty and static, devoid of any sign of performance. Frank Mohler, a theater his<strong>to</strong>rian<br />

and designer, has adopted an approach that focuses not on the architecture per se, but on<br />

technologies used for changing scenery. Mohler has made effective use of simple<br />

animations <strong>to</strong> simulate the appearance and functioning of Renaissance and Baroque stage<br />

machinery.<br />

Performance Simulations<br />

Models of theaters and scenery, no matter how detailed, immersive, or interactive,<br />

simulate only the environment within which performances take place. There have also<br />

been attempts <strong>to</strong> use computer animation techniques <strong>to</strong> simulate the phenomenon of<br />

performance itself, both for pedagogical and scholarly purposes. Again, Larry<br />

Friedlander produced one of the earliest examples, a program called TheaterGame<br />

created in conjunction with the Shakespeare Project. This software was innovative for its<br />

time and attracted a good deal of press attention. TheaterGame allowed students <strong>to</strong><br />

experiment with staging techniques by selecting crude two-dimensional human figures,<br />

clothing them from a limited palette of costumes, positioning set pieces on a virtual stage,<br />

and finally moving the virtual ac<strong>to</strong>rs around the stage and positioning their limbs <strong>to</strong> form<br />

simple gestures. The goal was <strong>to</strong> allow students with no theater experience or access <strong>to</strong><br />

real ac<strong>to</strong>rs <strong>to</strong> investigate the effects of basic staging choices.<br />

At the same time Friedlander was developing TheaterGame, Tom Calvert began <strong>to</strong><br />

develop a similar, but vastly more sophisticated, application geared <strong>to</strong>ward<br />

choreographers. The project started in the 1970s as a dance notation system called<br />

Compose that ran on a mainframe computer and output its data <strong>to</strong> a line printer. In the<br />

1980s, Calvert replaced abstract symbols describing motions with 3-D human animations<br />

and dubbed the new program LifeForms. The human models in LifeForms are featureless<br />

wireframes, but the movements are precise, flexible, and ana<strong>to</strong>mically correct. LifeForms<br />

was designed as a kind of word processor for dance students and practicing<br />

choreographers, a <strong>to</strong>ol for composing dances. In 1990, the renowned choreographer<br />

Merce Cunningham adopted the software, bringing it <strong>to</strong> international attention. In the<br />

early 1990s, LifeForms became a commercial product.<br />

Motion capture technology offers a very different approach <strong>to</strong> performance simulation.<br />

Motion capture uses sensors <strong>to</strong> track a performer's movements in space and then maps<br />

those movements on<strong>to</strong> a computer-generated model. While applications such as Life-<br />

Forms are authoring <strong>to</strong>ols for virtual performances, motion capture provides a <strong>to</strong>ol for<br />

archiving and analyzing real performances. The Advanced Computer Center for Art and<br />

Design (ACCAD) at Ohio State University maintains a high-end optical motion capture<br />

system dedicated <strong>to</strong> research in the performing arts. In <strong>2001</strong>, ACCAD began <strong>to</strong> build an<br />

archive of dance and theater motion data by capturing two performances by the legendary<br />

mime Marcel Marceau. These data, which include subtle details such as the performer's<br />

breathing, can be transferred on<strong>to</strong> any 3-D model and analyzed in depth.


I am currently working with a team of researchers at seven universities on a National<br />

Science Foundation project that combines many elements of the projects discussed<br />

above: 3-D modeling of theater architecture, animated scenery, performance simulation<br />

using motion capture, along with simulations of audience interactions and hypermedia<br />

content. This project, called Virtual Vaudeville, is creating a computer simulation of late<br />

nineteenth-century American vaudeville theater. The goal is <strong>to</strong> develop reusable<br />

strategies for using <strong>digital</strong> technology <strong>to</strong> reconstruct and archive his<strong>to</strong>rical performance<br />

events. Virtual Vaudeville strives <strong>to</strong> produce the sensation of being surrounded by human<br />

activity on stage, in the audience, and backstage. Viewers enter the virtual theater and<br />

watch the animated performances from any position in the audience, and are able <strong>to</strong><br />

interact with the animated specta<strong>to</strong>rs around them. Professional ac<strong>to</strong>rs are recreating the<br />

stage performances, and these performances are being transferred <strong>to</strong> 3-D models of the<br />

nineteenth-century performers using motion and facial capture technology. The program<br />

is being built with a high-performance game engine of the sort usually used <strong>to</strong> create<br />

commercial 3-D action games.<br />

Computer simulations of performance spaces and performers are powerful research and<br />

teaching <strong>to</strong>ols, but carry inherent dangers. Performance reconstructions can encourage a<br />

positivist conception of his<strong>to</strong>ry (Denard 2002: 34). A compelling computer simulation<br />

conceals the hypothetical and provisional nature of his<strong>to</strong>rical interpretation. Moreover,<br />

vividly simulated theaters and performances produce the sensation that the viewer has<br />

been transported back in time and is experiencing the performance event "as it really<br />

was." But even if all of the physical details of the simulation are accurate, a present-day<br />

viewer's experience will be radically different from that of the original audience because<br />

the cultural context of its reception has changed radically. Some projects, such as the<br />

Pompey Project and Virtual Vaudeville, are making a concerted effort <strong>to</strong> counteract these<br />

positivistic tendencies, primarily by providing hypermedia notes that supply contextual<br />

information, provide the his<strong>to</strong>rical evidence upon which the reconstructions are based,<br />

and offer alternatives <strong>to</strong> the interpretations of and extrapolations from the his<strong>to</strong>rical data<br />

used in the simulation. Whether such strategies will prove sufficient remains <strong>to</strong> be seen.<br />

Computers in Performance<br />

My focus so far has been on applications of computers <strong>to</strong> teaching and research in the<br />

performing arts. Digital technology is also beginning <strong>to</strong> have a significant impact on the<br />

way those art forms are being practiced. For example, the computer has become a routine<br />

part of the design process for many set and lighting designers. Throughout the 1990s, a<br />

growing number of designers adopted CAD software <strong>to</strong> draft blueprints and light plots<br />

and, more recently, employed 3-D modeling software (sometimes integrated in<strong>to</strong> the<br />

CAD software) <strong>to</strong> produce pho<strong>to</strong>realistic visualizations of set and lighting designs.<br />

Computers are also being incorporated in<strong>to</strong> the performances themselves. The earliest<br />

and most fully assimilated example is computer-controlled stage lighting. Computerized<br />

light boards can s<strong>to</strong>re hundreds of light cues for a single performance, au<strong>to</strong>matically<br />

adjusting the intensity, and in some cases the color and direction, of hundreds of lighting<br />

instruments for each cue. This technology was introduced in the late 1970s, and by the


1990s had become commonplace even in school and community theaters. Similarly, set<br />

designers have used computerized motion control systems <strong>to</strong> change scenery on stage –<br />

though this practice is still rare and sometimes has disastrous results. For example, the<br />

initial pre-Broadway run of Disney's stage musical Aida featured a six-<strong>to</strong>n robotic<br />

pyramid that changed shape under computer control <strong>to</strong> accommodate different scenes.<br />

The pyramid broke down on opening night and repeatedly thereafter (Elliott 1998).<br />

Disney jettisoned the high-tech set, along with the production's direc<strong>to</strong>r and designer,<br />

before moving the show <strong>to</strong> Broadway.<br />

Computer-controlled lighting and scenery changes are simply au<strong>to</strong>mated forms of precomputer<br />

stage technologies. A growing number of dance and theater artists have<br />

incorporated interactive <strong>digital</strong> media in<strong>to</strong> live performance events. Such performances<br />

can have a profound impact on the way the art forms are conceived, collapsing the neat<br />

on<strong>to</strong>logical divide that once separated (or seemed <strong>to</strong> separate) the live performing arts<br />

from reproductive media such as film and video. The Digital Performance Archive, a<br />

major research project conducted by Nottingham Trent University and the University of<br />

Salford, has created a web-based database documenting hundreds of dance and theater<br />

performances produced in the 1990s that combined <strong>digital</strong> media with live performance.<br />

George Coates's Performance Works in San Francisco was one of the first and most<br />

prominent theater companies <strong>to</strong> combine <strong>digital</strong> media with live performance <strong>to</strong> create<br />

stunning, poetic visual spectacles. In 1989, George Coates founded SMARTS (Science<br />

Meets the Arts), a consortium including companies such as Silicon Graphics, Sun<br />

Microsystems, and Apple Computer, <strong>to</strong> acquire the high-end technology required for his<br />

productions. In a series of productions starting with Invisible Site: A Virtual Sbo in 1991,<br />

Coates perfected a technique for producing the vivid illusion of live performers fully<br />

integrated in<strong>to</strong> a rapidly moving 3-D virtual environment. The specta<strong>to</strong>rs wear polarized<br />

glasses <strong>to</strong> view huge, high-intensity stereographic projections of <strong>digital</strong> animations. The<br />

projections that surround the revolving stage cover not only the back wall but the stage<br />

floor and transparent black scrims in front of the performers. The <strong>digital</strong> images are<br />

manipulated interactively during the performances <strong>to</strong> maintain tight synchronization<br />

between the live performers and the media.<br />

Another pioneer in the use of virtual scenery is Mark Reaney, founder of the Institute for<br />

the Exploration of Virtual Realities (i.e. VR) at the University of Kansas (Reaney 1996;<br />

Gharavi 1999). In place of physical scenery, Reaney creates navigable 3-D computer<br />

models that he projects on<strong>to</strong> screens behind the performers. The perspective on Reaney's<br />

virtual sets changes in relation <strong>to</strong> the performers' movements, and a computer opera<strong>to</strong>r<br />

can instantly transform the <strong>digital</strong> scenery in any way Reaney desires. In 1995, i.e. VR<br />

presented its first production, Elmer Rice's expressionist drama The Adding Machine. For<br />

this production, Reaney simply rear-projected the virtual scenery. For Wings in 1996,<br />

Reaney had the specta<strong>to</strong>rs wear low-cost head-mounted displays that allowed them <strong>to</strong> see<br />

stereoscopic virtual scenery and the live ac<strong>to</strong>rs simultaneously. Starting with Telsa<br />

Electric in 1998 Reaney adopted an approach much like Coates's, projecting stereoscopic<br />

images for the audience <strong>to</strong> view through polarized glasses. Nonetheless, Reaney's<br />

approach differs from Coates's in a number of important ways. While Coates authors his


own highly associative works, Reaney usually selects pre-existing plays with linear<br />

narratives. Reaney's designs, while containing stylized elements, are far more literal than<br />

Coates's, and the technology he employs, while more advanced than that available <strong>to</strong><br />

most university theaters, is far more affordable than the state-of-the-art technology at<br />

Coates's disposal.<br />

A number of dance performances have experimented with interactive 3-D technology<br />

similar <strong>to</strong> that used by Coates and Reaney. One of the earliest and most influential<br />

examples is Dancing with the Virtual Dervish/Virtual Bodies, a collaboration between<br />

dancer and choreographer Yacov Sharir, visual artist Diane Gromala, and architect<br />

Marcos Novak first presented at the Banff Center for the Arts in 1994. For this piece,<br />

Sharir dons a head-mounted display and enters a VR simulation of the interior of a<br />

human body, constructed from MRI images of Gromala's own body. The images that<br />

Sharir sees in the display are projected on<strong>to</strong> a large screen behind him as he dances.<br />

The three examples of <strong>digital</strong>ly-enhanced performance considered above are radically<br />

different in their aesthetics and artistic goals, but all establish the same basic relationship<br />

between the media and live performers: in each case, the media functions as virtual<br />

scenery, in other words, as an environment within which a live performance occurs.<br />

There are, however, many other roles that media can assume in a performance 1 . For<br />

example, the media can play a dramatic role, creating virtual characters who interact with<br />

the live performers. A number of choreographers, including prominent figures such as<br />

Merce Cunningham and Bill T Jones, have enlisted motion capture technology <strong>to</strong> lend<br />

subtle and expressive movements <strong>to</strong> virtual dance partners (Dils 2002). Often, as in the<br />

case of both Cunningham's and Jones's work, the computer models themselves are highly<br />

abstract, focusing the specta<strong>to</strong>rs' attention on the motion itself. In a 2000 production of<br />

The Tempest at the University of Georgia, the spirit Ariel was a 3-D computer animation<br />

controlled in real time by a live performer using motion capture technology (Saltz<br />

<strong>2001</strong>b). Claudio Pinhanez has applied artificial intelligence and computer-vision<br />

techniques <strong>to</strong> create fully au<strong>to</strong>nomous computer characters. His two-character play It/I,<br />

presented at MIT in 1997, pitted a live ac<strong>to</strong>r against a <strong>digital</strong> character (Pinhanez and<br />

Bobick 2002).<br />

A key goal of Pinhanez's work is <strong>to</strong> produce an unmediated interaction between the live<br />

performer and <strong>digital</strong> media. While this goal is unusual in theater, it is becoming<br />

increasingly common in dance, where there is less pressure <strong>to</strong> maintain a coherent<br />

narrative, and so creating an effective interaction between the performer and media does<br />

not require sophisticated artificial intelligence techniques. Electronic musicians created a<br />

set of technologies useful for creating interactive dance in the 1980s in the course of<br />

exploring sensors for new musical instrument interfaces. The Studio for Electro-<br />

Instrumental Music, or Steim, in the Netherlands was an early center for this research,<br />

and continues <strong>to</strong> facilitate collaborations between dancers and electronic musicians. A<br />

number of dancers have created interactive performances using the Very Nervous System<br />

(VNS), a system first developed by the Canadian media artist David Rokeby in 1986. The<br />

VNS uses video cameras <strong>to</strong> detect very subtle motions that can trigger sounds or video.<br />

One of the first dance companies formed with the explicit goal of combining dance with


interactive technology was Troika Ranch, founded in 1993 by Mark Coniglio and Dawn<br />

S<strong>to</strong>ppiello. Troika Ranch has developed a wireless system, MidiDancer, which converts a<br />

dancer's movements in<strong>to</strong> MIDI (Musical Instrument Digital Interface) data, which can be<br />

used <strong>to</strong> trigger sounds, video sequences, or lighting.<br />

One of the most complex and successful applications of this kind of sensor technology<br />

was the <strong>2001</strong> production L'Universe (pronounced "loony verse") created by the Flying<br />

Karamazov Brothers, a troupe of comedic jugglers, in conjunction with Neil Gershenfeld<br />

of the Physics and Media Group at MIT. Gershenfeld created special juggling clubs with<br />

programmable displays, and used sonar, long-range RF links and computer vision <strong>to</strong> track<br />

the positions and movements of the four performers. This technology was used <strong>to</strong> create a<br />

complex interplay between the performers and media, with the jugglers' actions<br />

au<strong>to</strong>matically triggering sounds and altering the color of the clubs.<br />

Brenda Laurel and Rachel Strickland's interactive drama Placeholder is one of the bestknown<br />

attempts <strong>to</strong> create a performance in which the specta<strong>to</strong>rs interact directly with the<br />

technology. As two specta<strong>to</strong>rs move through a ten-foot diameter circle wearing headmounted<br />

displays, they interact with a series of <strong>digital</strong> characters, with a character called<br />

the Goddess controlled by a live performer, and, <strong>to</strong> a limited extent, with each other<br />

(Ryan 1997: 695–6). The challenge of creating a rich and compelling narrative within this<br />

kind of interactive, open-ended structure is immense. While a number of writers have<br />

tackled this challenge from a theoretical perspective (see, for example, Ryan 1997;<br />

Murray 1997), the promise of this new dramatic medium remains largely unfulfilled.<br />

Telematic Performance<br />

In 1932, in a remarkable anticipation of Internet culture, theater theorist and playwright<br />

Ber<strong>to</strong>lt Brecht imagined a future in which radio would cease <strong>to</strong> be merely a one-way<br />

"apparatus for distribution" and become "the finest possible communication apparatus in<br />

public life, a vast network of pipes" (Brecht 1964: 52). By the early 1990s, it had become<br />

possible <strong>to</strong> stream video images over the Internet at very low cost, and performance<br />

groups were quick <strong>to</strong> exploit video streaming technologies <strong>to</strong> create live multi-site<br />

performance events. In a 1994 production of Nowhere Band, George Coates used free<br />

CU-SeeMe video-conferencing software <strong>to</strong> allow three band members at various<br />

locations in the Bay Area <strong>to</strong> perform live with a Bulgarian bagpipe player in Australia for<br />

an audience in San Francisco (Illingworth 1995). In 1995, Cathy Weiss used the same<br />

software <strong>to</strong> create an improvised dance performance at The Kitchen in New York with<br />

the real-time participation of a video artist in Prague and a DJ in (Santa Monica Saltz<br />

<strong>2001</strong>a). In 1999 the Australian Company in Space created a live duet between a dancer in<br />

Arizona and her partner in Australia (Birringer 1999: 368–9). In <strong>2001</strong>, the opera The<br />

Technophobe and Madman <strong>to</strong>ok advantage of the new high-speed Internet2 network <strong>to</strong><br />

create a multi-site piece of theater. Half of the performers performed at Rensselaer<br />

Polytechnic University, while the other half performed 160 miles away at New York<br />

University. Two separate audiences, one at each location, watched the performance<br />

simultaneously, each seeing half the performers live and the other half projected on a<br />

large projection screen (see Mirapaul <strong>2001</strong>). The Gertrude Stein Reper<strong>to</strong>ry Theater


(GSRT) is a company dedicated <strong>to</strong> developing new technologies for creating theater. In<br />

their production of The Making of Americans (2003), adapted from Gertrude Stein's<br />

novel, performers in remote locations work <strong>to</strong>gether in real time <strong>to</strong> create live<br />

performances in both locations simultaneously, with the faces and bodies of ac<strong>to</strong>rs in one<br />

location being projected via videoconferencing on masks and costumes worn by ac<strong>to</strong>rs in<br />

the second location. The GSRT draws a parallel between this process, which they call<br />

"Distance Puppetry", and Japanese performance traditions such as bunraku and ningyo<br />

buri that also employ multiple performers <strong>to</strong> portray individual characters.<br />

Telematic performance acquires its greatest impact when specta<strong>to</strong>rs interact directly with<br />

people at the remote site and experience the uncanny collapse of space first-hand. In the<br />

1990s, new media artist Paul Sermon created a series of interactive art installations<br />

joining physically separated viewers. For example, in Telematic Dreaming a viewer lies<br />

down on one side of a bed and on the other side sees a real-time video projection of a<br />

participant lying down on a second, identical bed in a remote location. Other installations<br />

place the remote participants on a couch, around a dining room table and at a seance table<br />

(Birringer 1999: 374). A more provocative example of a performance event that joins a<br />

live performer <strong>to</strong> the Internet is Stelarc's 1996 Ping Body: An Internet Actuated and<br />

Uploaded Performance, in which a muscle stimula<strong>to</strong>r sent electric charges of 0–60 volts<br />

in<strong>to</strong> Stelarc <strong>to</strong> trigger involuntary movements in his arms and legs proportionate <strong>to</strong> the<br />

ebb and flow of Internet activity.<br />

Conclusions and Queries<br />

The use of computers in the performing arts does not merely add a new <strong>to</strong>ol <strong>to</strong> an old<br />

discipline. It challenges some of our most basic assumptions about performance. First, it<br />

blurs the boundaries between performance disciplines. When we watch a performer<br />

au<strong>to</strong>matically triggering recorded fragments of dialogue as she moves across a stage, are<br />

we witnessing a piece of music, dance, or theater? Second, it blurs the boundaries<br />

between scholarship and creative practice. Is someone who extrapolates a complete set<br />

design, script, and performance from shreds of his<strong>to</strong>rical evidence <strong>to</strong> create a virtual<br />

performance simulation an artist or a scholar? When someone develops new artificial<br />

intelligence algorithms in order <strong>to</strong> create a dramatic interaction between a <strong>digital</strong><br />

character and a live ac<strong>to</strong>r, is that person functioning as a computer scientist or an artist<br />

whose medium just happens <strong>to</strong> be computers? Finally, <strong>digital</strong> technology is challenging<br />

the very distinction between "liveness" and media. When a live performer interacts with a<br />

computer-generated animation, is the animation "live"? Does the answer depend on<br />

whether the animation was rendered in advance or is being controlled in real time via<br />

motion capture or with artificial intelligence software? Or do we now live in a world, as<br />

performance theorist Philip Auslander suggests, where the concept of liveness is losing<br />

its meaning?<br />

Note


1 Elsewhere I have distinguished between twelve types of relationships a production can<br />

define between <strong>digital</strong> media and live performers (Saltz <strong>2001</strong>b).<br />

References for Further Reading<br />

Artaud, An<strong>to</strong>nin (1958). The Theater and Its Double, tr. Mary Caroline Richards. New<br />

York: Grove Weidenfeld. (Original work published 1938).<br />

Birringer, Johannes (1998). Media and Performance: Along the Border. Baltimore: Johns<br />

Hopkins University Press.<br />

Birringer, Johannes (1999). Contemporary Performance/Technology. Theater Journal 51:<br />

361–81.<br />

Brecht, Ber<strong>to</strong>lt (1964). The Radio as an Apparatus of Communications. In John Willett<br />

(ed. and tr.), Brecht on Theater (pp. 51–3). New York: Hill and Wang.<br />

Denard, Hugh (2002). Virtuality and Performativity: Recreating Rome's Theater of<br />

Pompey. Performing Arts Journal 70: 25–43. At<br />

http://muse.jhu.edu/journals/performing_arts_journal/<strong>to</strong>c/paj24.1.html.<br />

Didaskalia. Recreating the Theater of Dionysos in Athens. Accessed April 19, 2004. At<br />

http://didaskalia.open.ac.uk/index.shtml.<br />

Digital Performance Archive. Accessed April 19, 2004. At http://dpa.ntu.ac.uk/dpa_site/.<br />

Dils, Ann (2002). The Ghost in the Machine: Merce Cunningham and Bill T. Jones.<br />

Performing Arts Journal 24, 1.<br />

Dixon, Steve (1999). Digits, Discourse and Documentation: Performance Research and<br />

Hypermedia. TDR: The Drama Review 43, 1: 152–75.<br />

Dixon, Steve (<strong>2001</strong>). Virtual Theatrics: When Computers Step in<strong>to</strong> the Limelight. In<br />

Loren K. Ruff (ed.), Theater: The Reflective Template, 2nd edn (pp. 216). Dubuque, IA:<br />

Kendall/Hunt Publishing.<br />

Donohue, Joseph (1981). Theater Scholarship and Technology: A Look at the Future of<br />

the Discipline. Theater Survey 22, 2: 133–9.<br />

Elliott, Susan (1998). Disney Offers an "Aida" with Morphing Pyramid. New York Times<br />

(Oc<strong>to</strong>ber 9): E3.<br />

Friedlander, Larry (1991). The Shakespeare Project: Experiments in Multimedia. In<br />

George Landow and Paul Delany (eds.), Hypermedia and Literary Studies (pp. 257–71).<br />

Cambridge, MA: MIT Press.


Gharavi, Lance (1999). i.e. VR: Experiments in New Media and Performance. In Stephen<br />

A. Schrum (ed.), Theater in Cyberspace (pp. 249–72). New York: Peter Lang Publishing.<br />

Golder, John (1984). The Theatre du Marais in 1644: A New Look at the Old Evidence<br />

Concerning France's Second Public Theater. Theater Survey 25: 146.<br />

Gromala, Diane J. and Yacov Shark (1996). Dancing with the Virtual Dervish: Virtual<br />

Bodies. In M. A. Moser and D. MacLeod (eds.), Immersed in Technology: Art and<br />

Virtual Environments (pp. 281–6). Cambridge, MA: MIT Press.<br />

Illingworth, Monteith M. (1995). George Coats [sic]: Toast of the Coast. Cyberstage. At<br />

http://www.cyberstage.org/archive/cstage12/coats12.htm.<br />

Laurel, Brenda (1991). Computers as Theater. Reading, MA: Addison-Wesley.<br />

Laurel, Brenda, Rachel Strickland, and Rob Tow (1994). Placeholder: Landscape and<br />

Narrative in Virtual Environments. ACM Computer Graphics Quarterly 28, 2: 118–26.<br />

McNamee, Lawrence F. (1970). Computers and the Theater. Quarterly J ournal of<br />

Speech 56: 315–19.<br />

Meisner, Sanford (1987). On Acting. New York: Vintage Books.<br />

Menicacci, Armando and Emanuele Quinz (<strong>2001</strong>). La scena <strong>digital</strong>s: nuovi media per la<br />

danza [The <strong>digital</strong> scene: new media in dance]. Venice: Marsilio.<br />

Mirapaul, Matthew (<strong>2001</strong>). How Two Sites Plus Two Casts Equals One Musical. New<br />

York Times (February 19): E2.<br />

Mohler, Frank (1999). Computer Modeling as a Tool for the Reconstruction of His<strong>to</strong>ric<br />

Theatrical Production Techniques. Theater Journal 51, 4: 417–31.<br />

Mullaly, Edward (1987). Computers and Theater Research. Theater Survey 28, 1: 59–70.<br />

Murray, Janet H. (1997). Hamlet on the Holodeck: The Future of Narrative in<br />

Cyberspace. New York: Free Press.<br />

Pinhanez, Claudio S. and Aaron F. Bobick (2002). "It/I": A Theater Play Featuring an<br />

Au<strong>to</strong>nomous Computer Character. Presence: Teleopera<strong>to</strong>rs and Virtual Environments<br />

11, 5: 536–48.<br />

Reaney, Mark (1996). Virtual Scenography: The Ac<strong>to</strong>r, Audience, Computer Interface.<br />

Theater Design and Technology 32: 36–43.<br />

Ryan, Marie-Laure (1997). Interactive Drama: Narrativity in a Highly Interactive<br />

Environment. Modern Fiction Studies 42, 2: 677–707.


Saltz, David Z. (1997). The Art of Interaction: Interactivity, Performativity and<br />

Computers. Journal of Aesthetics and Art Criticism 55, 2: 117–27.<br />

Saltz, David Z. (<strong>2001</strong>a). The Collaborative Subject: Telerobotic Performance and<br />

Identity. Performance Research 6, 4: 70–83.<br />

Saltz, David Z. (<strong>2001</strong>b). Live Media: Interactive Technology and Theater. Theater Topics<br />

11, 2: 107–30.<br />

Sarlós, Robert K. (1989). Performance Reconstruction: The Vital Link between Past and<br />

Future. In Bruce A. McConachie and Thomas Postlewait (eds.), Interpreting the<br />

Theatrical Past (pp. 198–229). Iowa City: University of Iowa Press.<br />

Schrum, Stephen A., (ed.) (1999). Theater in Cyberspace: Issues of Teaching, Acting,<br />

and Directing. New York: Peter Lang.<br />

THEATRON. Accessed April 19, 2004. At http://www.theatron.org.<br />

Watts, Allan (1997). Design, Computers, and Teaching. Canadian Theater Review 91:<br />

18–21.<br />

Williford, Christa (2000). A Computer Reconstruction of Richelieu's Palais Cardinal<br />

Theater, 1641. Theater Research International 25, 3: 233–47.<br />

12.<br />

"Revolution? What Revolution?" Successes and Limits<br />

of Computing Technologies in Philosophy and Religion<br />

Charles Ess<br />

Who can foresee the consequences of such an invention?<br />

Lady Ada Lovelace<br />

Introduction<br />

Computing technologies – like other technological innovations in the modern West – are<br />

inevitably introduced with the rhe<strong>to</strong>ric of "revolution." Especially during the 1980s (the<br />

PC revolution) and 1990s (the Internet and Web revolutions), enthusiasts insistently<br />

celebrated radical changes – changes ostensibly inevitable and certainly as radical as<br />

those brought about by the invention of the printing press, if not the discovery of fire.


These enthusiasms now seem very "1990s" – in part as the revolution stumbled with the<br />

dot.com failures and the devastating impacts of 9/11. Moreover, as I will sketch out<br />

below, the patterns of diffusion and impact in philosophy and religion show both<br />

tremendous success, as certain revolutionary promises are indeed kept – as well as<br />

(sometimes spectacular) failures. Perhaps we use revolutionary rhe<strong>to</strong>ric less frequently –<br />

because the revolution has indeed succeeded: computing technologies, and many of the<br />

powers and potentials they bring us as scholars and religionists have become so<br />

ubiqui<strong>to</strong>us and normal that they no longer seem "revolutionary" at all. At the same time,<br />

many of the early hopes and promises – instantiated in such specific projects as Artificial<br />

Intelligence and anticipations of virtual religious communities – have been dashed<br />

against the apparently intractable limits of even these most remarkable technologies.<br />

While these failures are usually forgotten, they leave in their wake a clearer sense of what<br />

these new technologies can, and cannot do.<br />

To see this, I highlight his<strong>to</strong>rical and current examples of how computing technologies<br />

are used in philosophy and religion. We will see that philosophers have been engaged<br />

with computing from its beginnings in the dreams of Leibniz in the seventeenth century 1<br />

and the earliest implementations of electronic computers in the 1940s, 1950s, and 1960s.<br />

And, perhaps because of the clear connections between computing technologies and a<br />

range of classical philosophical practices (logic) and fields (epistemology, on<strong>to</strong>logy,<br />

ethics, political philosophy, etc.), computation has enjoyed an increasingly central place<br />

in the philosophical literature of the past fifty years. Indeed, many philosophers speak of<br />

a "computational turn" – referring <strong>to</strong> ways in which computing technologies have given<br />

philosophers new kinds of labora<strong>to</strong>ries for testing and refining classical debates and<br />

hypotheses.<br />

Similarly, religious studies scholars learned early <strong>to</strong> exploit the new <strong>to</strong>ols – beginning<br />

with Father Rober<strong>to</strong> Busa's pioneering use of computers in the 1940s <strong>to</strong> analyze complex<br />

texts. More sophisticated versions of these early innovations gradually developed and<br />

became commonplace on <strong>to</strong>day's desk<strong>to</strong>ps and even palm-held computers. In addition,<br />

the impact of computation in religion seems still more powerful in the larger domain of<br />

religious practice. In ways consistent with earlier technological innovations – especially<br />

such mass media as the radio and television – it is especially the religiously marginalized<br />

and proselytizers who have benefited from computation as instantiated in computer<br />

networks, i.e., the Internet and the World Wide Web.<br />

In both domains we will see the same general pattern: an early period of enthusiasm, one<br />

that rode high on revolutionary – even apocalyptic – promises of radical transformation,<br />

followed by a quieter period of diffusion and incorporation of computing technologies<br />

within philosophical and religious domains. Hegel would remind us that it is in this<br />

relative quiet, aided now by a rich his<strong>to</strong>ry of examples, that we are able <strong>to</strong> evaluate more<br />

critically and carefully the strengths and limits of these remarkable technologies.<br />

Philosophy


"love of wisdom" (ϕιλoσoϕ α) – systematic and rational inquiry in<strong>to</strong> what is<br />

(metaphysics, on<strong>to</strong>logy), how we may know (epistemology), how we may think cogently<br />

and avoid error (logic), and, especially given some account of who we are as human<br />

beings and our relation(s) <strong>to</strong> the larger world (possibly including divinity/ies), how we<br />

should behave individually and in community (ethics, politics).<br />

The computational turn: logic, AI, ethics<br />

Arguably, computing is philosophy – specifically, the branch of philosophy concerned<br />

with logic. Our computing devices depend on the logical system developed by George<br />

Boole in the 1840s and 1850s; prior <strong>to</strong> Boole, a number of philosophers in the modern era<br />

– most notably, Leibniz – concerned themselves with the possibilities of machines that<br />

might au<strong>to</strong>mate reasoning as the manipulation of symbols (Dipert 2002: 148). Hence, it is<br />

perhaps not surprising that some of the earliest applications of computing technology in<br />

philosophy were precisely in the area of logic – in efforts both <strong>to</strong> exploit the computer as<br />

a logical calculating device (e.g., for assessing argument validity and generating valid<br />

conclusions from specified premises; ibid.) as well as <strong>to</strong> au<strong>to</strong>mate the teaching of logic<br />

(perhaps most notably by Patrick Suppes, beginning in 1963: see Suppes, Home page).<br />

Moreover, as computers au<strong>to</strong>mate and expand our ability <strong>to</strong> undertake logical analyses,<br />

they not only offer new ways of accomplishing classical logical tasks (from truth-table<br />

analysis through proofs <strong>to</strong> advanced logical applications – e.g., Tarski's World, etc.: see<br />

Barwise and Etchemendy 1999), they further open up distinctively new ways of<br />

exploring classical philosophical questions. One of the earliest, and most obvious,<br />

examples of this "computational turn" in philosophy is the rise of Artificial Intelligence<br />

(AI) – the effort <strong>to</strong> replicate human consciousness and reasoning through computing<br />

devices, prominent especially in the 1950s through the early 1990s. Computers provided<br />

philosophers with a labora<strong>to</strong>ry in which <strong>to</strong> empirically test and refine hypotheses about<br />

the nature of reason and consciousness. Initially, so-called "hard" AI proponents believed<br />

and argued that machines would quickly outstrip human intelligence. It appears, however,<br />

that the hard AI agenda has largely moved <strong>to</strong> the margins – primarily because of repeated<br />

failures <strong>to</strong> live up <strong>to</strong> early promises. As in the natural sciences, however, both successes<br />

and failures are instructive: the successful efforts <strong>to</strong> instantiate at least some components<br />

of human reasoning in machines has helped philosophers sharpen their sense of how far<br />

computation and consciousness overlap – while the failures help demarcate how<br />

computation and human reasoning remain intractably distinct from one another. Indeed,<br />

there is a recent turn in philosophy – in fields as diverse as hermeneutics,<br />

phenomenology, and feminism – <strong>to</strong>wards embodiment as a key theme of exploration:<br />

philosophers such as Albert Borgmann and Hubert Dreyfus use an understanding of who<br />

we are as embodied creatures <strong>to</strong> explore the strengths and limits of technology, including<br />

computing technologies, in contemporary life. In particular, Dreyfus recognizes the<br />

(limited) achievements of AI and the (limited) pedagogical advantages of distance<br />

learning via the Web and the Internet, but he further argues that these technologies do not<br />

help us acquire our most distinctive and important human capacities – those of making<br />

commitments, taking risks, and exercising phronesis (Aris<strong>to</strong>tle's term for practical<br />

judgment and wisdom) in our ethical and political lives.


Indeed, the crea<strong>to</strong>rs of modern computing technology have been aware from the outset<br />

that these new devices raise, precisely, questions of ethics with regard <strong>to</strong> the use of these<br />

systems, politics with regard <strong>to</strong> their potential social and political impacts, etc. As Bynum<br />

points out (<strong>2001</strong>), Norbert Wiener, in what amounts <strong>to</strong> the first book of computer ethics<br />

(Wiener 1950), recognized that computing technologies will impact, for better and for<br />

worse, life and health, knowledge and science, work and wealth, creativity and happiness,<br />

democracy and freedom, and inter/national concerns with peace and security. In Bynum's<br />

view, moreover, Wiener sets the agenda for the field of computer ethics – a field that<br />

begins <strong>to</strong> emerge only slowly in the 1960s and 1970s, but now includes an extensive<br />

literature, including the journal Ethics and Information Technology. More broadly, the<br />

computational turn is now well documented not only in AI and ethics, but also, for<br />

example, in philosophy of mind, philosophy of science, epistemology, on<strong>to</strong>logy, and so<br />

forth (Bynum and Moor 1998, 2003). Indeed, the late 1990s and early twenty-first<br />

century saw the emergence of philosophy of information as one way <strong>to</strong> provide a<br />

systematic overview of the impact of computing in virtually every domain of philosophy<br />

(Floridi 2003).<br />

Hypertext<br />

In the 1980s, both philosophers and other <strong>humanities</strong> scholars were excited by a new<br />

form of "non-linear" text made possible by computing technologies – hypertext. While<br />

the theoretical reflections and software experiments of such hypertext pioneers as Jay<br />

David Bolter and George Landow were most prominent – at least some philosophers<br />

were likewise quite interested in the potential of hypertext <strong>to</strong> open up at least alternative<br />

forms of argument. The best known of these, David Kolb, has in fact argued that<br />

hypertexts will make possible the recovery of argument forms (e.g., Hegel's dialectic)<br />

which are only awkwardly expressed in the (largely) linear frameworks of print.<br />

CD-ROM databases (ethics, his<strong>to</strong>ry of philosophy)<br />

After appearing in the 1980s, CD-ROMs seemed the perfect medium for realizing<br />

extensive hypermedia programs and databases. Two examples can be mentioned here.<br />

First, effective hypermedia resources have been developed for exploring and teaching<br />

ethics (e.g., Cavalier et al. 1998). These resources are especially useful as they provide<br />

video documentary and interviews <strong>to</strong> bring more abstract discussion of general principles<br />

down <strong>to</strong> the fine-grained contexts of real persons facing specific ethical problems in<br />

particular contexts.<br />

Second, while originally defined as a Greek classics project – i.e., focused on collecting<br />

primary and secondary literary and his<strong>to</strong>rical texts, as well as architectural, his<strong>to</strong>rical, and<br />

cultural resources (including extensive visual images of important sites, artifacts, etc.) -<br />

the second edition of the Perseus Project (Perseus 2.0 1996) also included both the Greek<br />

and English texts of Pla<strong>to</strong> and Aris<strong>to</strong>tle. Like its counterparts in Biblical studies, the<br />

Perseus CD-ROM not only provides an electronic library of these primary texts, but also<br />

allows for text searches and analyses. For example, I can find – more or less instantly –<br />

every example of Pla<strong>to</strong>'s use of the term ("cybernetes", a steersman, or


pilot, and thus figuratively, the Divine as directing the destiny of humans – and the root,<br />

not coincidentally, of "cybernetics") as well as related terms. Through hypertext linking,<br />

the program can further take me <strong>to</strong> each example in the Pla<strong>to</strong>nic text, as well as provide<br />

morphological analyses. Such searches, of course, accomplish in minutes what might<br />

otherwise take weeks, months, or years of reading.<br />

At the same time, however, while these are extraordinary resources and <strong>to</strong>ols, they are<br />

also quite limited. While the primary and secondary texts made available on CD-ROM<br />

are extensive – and supplemented, of course, by a constantly growing electronic library<br />

on the Web – only a relatively small percentage of the literatures important <strong>to</strong><br />

philosophers has been digitized. Moreover, word searches and morphological analyses<br />

are important components of scholarship – but they are only a very small component of<br />

our research, reflection, and writing. As with the specific his<strong>to</strong>ry of AI, philosophers<br />

(and, it appears, religion scholars) are gaining a more realistic and nuanced appreciation<br />

of the strengths and limits of computing technologies in their disciplines, precisely as the<br />

successful resources and <strong>to</strong>ols simultaneously mark out what the technologies cannot do<br />

(at least so far).<br />

Computer-mediated communication<br />

The sorts of communication most familiar <strong>to</strong> us in terms of the Internet and the Web also<br />

serve as a philosophical labora<strong>to</strong>ry, one that allows philosophers <strong>to</strong> revisit classical<br />

questions in the domains of on<strong>to</strong>logy, epistemology (including semiotics, hypertext, and<br />

logic), the meaning of identity and personhood (including issues of gender and<br />

embodiment), and ethical and political values (especially those clustering about the claim<br />

that these technologies will issue in a global democracy vs. the correlative dangers of<br />

commercialization and a "computer-mediated colonization": see Ess 2003).<br />

The Internet and the World Wide Web<br />

Interestingly enough, major projects in both philosophy (Perseus) and religious studies<br />

(Diana Eck's On Common Ground) that began on CD-ROM have migrated <strong>to</strong> the Web <strong>to</strong><br />

join additional "electronic libraries" and online search engines. Two significant examples<br />

of online philosophical resources are the Hippias search engine and website, and<br />

Lawrence Hinman's Ethics Updates site.<br />

Finally, these computer networks have made possible what is now the presumed<br />

environment of philosophical scholarship – namely e-mail, listservs, extensive online<br />

databases and search engines, and countless websites (of varying quality) that collect and<br />

"publish" (often) significant resources. It is commonly observed that, in contrast with the<br />

high-profile projects and experiments, in the end, it is the relatively low-tech <strong>to</strong>ols that<br />

make the most difference – in this case, e-mail and listservs. Despite the explosion of<br />

scholarly resources and specialized <strong>to</strong>ols available online and on CD-ROM, it remains<br />

debatable as <strong>to</strong> whether the computer revolution has markedly improved philosophical<br />

debate, scholarship, or insight (Dipert 2002). Nonetheless, the ability for scholars,<br />

geographically remote from one another and positioned at institutions of varying


esources and prestige, <strong>to</strong> communicate directly with one another through e-mail and<br />

listservs organized by interests and specialties is arguably "revolutionary" indeed. At the<br />

same time, however, this ability is now a commonplace.<br />

Religion<br />

"rebinding" – religio – between the human and the sacred, expressed both individually<br />

and in community in terms of beliefs, values, practices, rituals, etc.; religious studies –<br />

academic studies of multiple aspects of religion, including studies of scriptures and<br />

beliefs, as well as through a range of disciplines such as sociology, psychology,<br />

philosophy, philology, his<strong>to</strong>ry, etc.<br />

While philosophers were closely involved with the development and early uses of<br />

computing technologies because of their disciplinary focus on logic, religious scholars<br />

were among the first <strong>to</strong> explore applications of computing in textual analysis. More<br />

recently, however – beyond the use of the Web and the Internet by believers and<br />

discussants – there appear <strong>to</strong> be comparatively fewer religiously oriented computing<br />

projects (perhaps because religious studies are among the most marginalized and<br />

underfunded in the academy?). As a representative example: the Institute for Advanced<br />

Technology in the Humanities at the University of Virginia – arguably one of the most<br />

important centers for developing computing-based <strong>humanities</strong> applications – lists only<br />

two projects (out of some fifty or so) that focus on some aspect of religious studies. On<br />

the other hand, the open communicative environments of the Internet and the Web, while<br />

appropriated in much the same way among religious scholars as among philosophers, by<br />

contrast have been taken up with explosive energy by more or less every religious<br />

tradition whose representatives enjoy Internet access.<br />

The use of computing technologies in religious scholarship<br />

In the 1940s, Father Rober<strong>to</strong> Busa began developing techniques for encoding complex<br />

texts in ways that could be manipulated by computers – first of all, <strong>to</strong> develop a<br />

concordance for the corpus of Thomas Aquinas, followed by a more extensive project <strong>to</strong><br />

develop a hypertextual edition of Aquinas that allowed for multiple levels of linking and<br />

comparison (Thomae Aquinatis Opera Omnia cum hypertextibus in CD-ROM). At least<br />

one Bible concordance (for the then new Revised Standard Version of the Bible) was also<br />

generated in the 1950s using a computer. As computers became (somewhat) less<br />

expensive and more widespread through the 1960s and 1970s, religious scholars began <strong>to</strong><br />

develop ways of exploiting the computer's s<strong>to</strong>rage and processing abilities <strong>to</strong> undertake<br />

complex text analyses (Harbin 1998). Just as philosophers hoped <strong>to</strong> make the machine<br />

take over foundational but repetitive tasks of logical calculation, so their colleagues in<br />

religious studies sought <strong>to</strong> use the machine <strong>to</strong> take on the tedium of word counts and<br />

comparisons.<br />

As microcomputers and CD-ROMs made processing and s<strong>to</strong>rage increasingly<br />

inexpensive in the 1980s, descendants of these early text-base and concordance projects<br />

moved <strong>to</strong> the desk<strong>to</strong>ps and lap<strong>to</strong>ps of religious scholars. First of all, Bible database


projects flourished – i.e., collections of Bible texts, translations, and commentaries, with<br />

basic computing features (indexing, word searches, note-taking) that support both<br />

elementary and more advanced sorts of grammatical and textual analysis. These range<br />

from the his<strong>to</strong>rically oriented His<strong>to</strong>ry of the English Bible (Beam and Gagos 1997) <strong>to</strong><br />

such significant resources as Bible Works. These new technologies further allow other<br />

sorts of databases – for example, Diana Eck's On Common Ground CD-ROM and<br />

subsequent website that documents religious diversity in the USA.<br />

On the one hand, these resources fulfill some of the fondest dreams of scholars.<br />

BibleWorks, for example, allows for complex searches through multiple Bibles –<br />

translations as well as critical editions in Greek and Hebrew. A scholar can accomplish in<br />

minutes an inquiry that would otherwise take weeks or months. This is the analogue <strong>to</strong><br />

the way computing technologies have indeed revolutionized mathematics, the sciences,<br />

and logic by turning over <strong>to</strong> the machine repetitive tasks that would otherwise take<br />

humans months, years, and lifetimes <strong>to</strong> perform (cf. Hardmeier 2000). On the other hand,<br />

as in philosophy, the impact of these new resources and abilities on the quality and<br />

quantity of scholarship remains a very open question. In particular, computer-adept<br />

scholars observe that these resources exploit only the most basic potentials of the<br />

computer, and – echoing the Socratic critique of the technology of writing as leading <strong>to</strong><br />

the appearance, but not the substance, of wisdom – run the danger of giving the untu<strong>to</strong>red<br />

amateur the appearance, if not conviction, that s/he now knows as much as any Bible<br />

scholar.<br />

More fundamentally, the emergence of these new technologies themselves – again,<br />

especially as interpreted through the lenses of postmodernism – all but required scholars<br />

in religion <strong>to</strong> consider and respond <strong>to</strong> what many began <strong>to</strong> see as the emerging<br />

"secondary orality of electronic culture" (so one of the premier theologically oriented<br />

theorists of new media: Walter Ong, 1988: 135–8, cited in O'Leary and Brasher 1996:<br />

246). In response, the American Bible Society (after bringing out one of the first CD-<br />

ROM Bible databases) under<strong>to</strong>ok an ambitious series of projects <strong>to</strong> "transmediate"<br />

important Christian narratives – i.e., <strong>to</strong> translate these, using scholarly approaches and<br />

principles of translation, in<strong>to</strong> the multimedia environments made possible by computing<br />

technologies; <strong>to</strong> synthesize music, visual images, and scholarly resources as an<br />

environment for "telling the s<strong>to</strong>ry" in a way intended <strong>to</strong> be attractive especially <strong>to</strong><br />

younger people, who are oriented primarily <strong>to</strong> electronic visual media (American Bible<br />

Society 1995). Despite considerable investment and remarkable talent, however, these<br />

projects have met with only limited success in the religious marketplace.<br />

Religious scholarship on the Web<br />

As for philosophers, the Internet and the Web have diffused in<strong>to</strong> the commonplace<br />

practices and environment of religious scholars. Beyond the explosive development of<br />

sites on the Web by diverse faith communities, there is also <strong>to</strong> be found a reasonably<br />

extensive but largely pedestrian use of the Internet and the Web <strong>to</strong> support<br />

• listservs on scholarly <strong>to</strong>pics of interest <strong>to</strong> religion scholars and lay persons;


• sites for religious studies professionals (e.g., the American Academy of Religion) that<br />

offer relatively little in terms of use or pointers <strong>to</strong>wards use of computing technolo gies,<br />

but rather make use of the Web – sensibly – as an easily updateable archive for such<br />

things as their online syllabus project, etc.; and<br />

• portal sites (e.g., Oxford University [Fraser 2000], the Society of Biblical Literature's<br />

Electronic Publications and Technology Resources for Biblical Studies) that list both<br />

institu tionally based resources and often very rich and helpful sites put up by individual<br />

scholars.<br />

As with philosophers, these now relatively low-tech uses of the computing technology<br />

may have the most significance as they expand access <strong>to</strong> resources and the ability <strong>to</strong><br />

communicate with geographically distant scholars who share similar interests.<br />

The Internet, the Web, and religious communities<br />

Whatever the potentials and impacts of computing technologies for religious scholars,<br />

religious communities have exploited the Internet and the Web with extraordinary<br />

energy. This religious colonization of cyberspace began first of all as the Internet and the<br />

Web afforded safe havens for those otherwise at the margins of North American religious<br />

life, e.g., Wiccans, Pagans, New Age seekers, etc., as well as (if more gradually)<br />

representatives of the world traditions such as Islam, Buddhism, Hinduism, etc. (O'Leary<br />

and Brasher 1996; Larsen <strong>2001</strong>).<br />

This enthusiasm was fueled (as elsewhere) by the rise of postmodernism, especially as<br />

postmodernism was theoretically conjoined with the technologies of hypertext and then<br />

the Web through such theorists as Jay David Bolter and George Landow. As<br />

postmodernism made its way in<strong>to</strong> religious scholarship and theology, it was embraced<br />

especially by Evangelicals and Pentecostals, as postmodernism promised <strong>to</strong> unseat<br />

modern rationalism – and thus shift epistemological legitimacy and authority <strong>to</strong> emotive<br />

experience (e.g., the feeling of being saved), and undermine rationalist approaches <strong>to</strong><br />

Scripture such as the his<strong>to</strong>rical-critical method, thereby eliminating the chief nemesis of<br />

Fundamentalist interpretation. More broadly, this trajec<strong>to</strong>ry has led <strong>to</strong> a number of<br />

insightful analyses of the relationship between new media and religion – including texts<br />

that both celebrate the media's ostensibly libera<strong>to</strong>ry/revolutionary potentials (e.g.,<br />

Careaga <strong>2001</strong>) and those that directly challenge postmodernist communication theories<br />

by documenting how traditional religious beliefs and assumptions have shaped and<br />

constrained the development and use of the new technologies (e.g., Davis 1998; see Ess<br />

<strong>2001</strong>, 2004 for an overview of this development and relevant literature).<br />

Currently, the plethora of sites on the Web devoted <strong>to</strong> religion staggers the imagination: a<br />

Google search on "religion", for example, will turn up some 13–8 million hits. As yet,<br />

there is no authoritative study of this massive inven<strong>to</strong>ry. Two studies in progress suggest,<br />

however, an interesting pattern:


Evangelical/Pentecostal/Fundamentalist sites more fully exploit the interactive nature of<br />

online communication <strong>to</strong> proselytize; while sites representing the online "face" of more<br />

mainstream and conservative traditions – including the Roman Catholic Church and the<br />

Greek Orthodox Church – largely provide extensive databases of authoritative texts and<br />

pronouncements, and relatively little interactive opportunity.<br />

This pattern, moreover, is consistent with other studies that show, for example, that initial<br />

grassroots efforts <strong>to</strong> exploit the Internet and the Web for political activism are soon<br />

squeezed out as extant power centers learn, if somewhat more slowly, how <strong>to</strong> use the<br />

Web and the Net <strong>to</strong> re-establish their dominance and centrality online. Similarly, more<br />

ecumenical proponents of online religion hope that a global Internet may facilitate a<br />

global, dialogical culture that fosters the many voices of diverse religious traditions in a<br />

new pluralism. But the emerging patterns of use of the Internet, while giving a certain<br />

advantage <strong>to</strong> previously marginalized traditions, rather largely reflect and preserve the<br />

existing religious landscape, i.e., one marked far more often by allegiance <strong>to</strong> one's own<br />

tradition and proselytizing on its behalf.<br />

Concluding Remarks<br />

Should this overview be even approximately correct, it is then notable for two reasons.<br />

First of all, through the larger perspectives of research (including cross-cultural research)<br />

on computer-mediated communication (CMC), this curve from initial enthusiasm <strong>to</strong> more<br />

pedestrian applications fits the larger pattern of development from the 1980s and 1990s <strong>to</strong><br />

the late 1990s and early noughties (so the British say) – i.e., from the heyday of<br />

postmodernism <strong>to</strong> a "post-post-modern" period that represents more of a hybrid between<br />

postmodernism and whatever came before it. Secondly, this pattern suggests that, indeed,<br />

the revolution has succeeded in certain remarkable ways – so much so that we no longer<br />

regard computer-based resources and <strong>to</strong>ols as "revolutionary", but simply as "normal"<br />

elements of our lives – while at the same time, the multiple failures in philosophy and<br />

religion <strong>to</strong> exploit computing technologies have left a significant portion of our work and<br />

lives relatively un<strong>to</strong>uched.<br />

In my own case: I still marvel at having nearly instantaneous access <strong>to</strong> a remarkable<br />

library of important texts – both Biblical and philosophical – not only on my desk<strong>to</strong>p<br />

computer, but also my palm-held computer, and that I can undertake searches that help<br />

me locate a familiar quote, and perhaps uncover new patterns and insights. In these ways,<br />

these computer-based resources and <strong>to</strong>ols certainly enhance my scholarship and teaching.<br />

At the same time, as a number of contemporary observers of technology caution, the<br />

affordances of these technologies – what they make easy for me <strong>to</strong> do – thereby<br />

encourage me <strong>to</strong> pursue the paths they facilitate, and perhaps thereby discourage other<br />

aspects of scholarship and teaching that, as yet unreduced <strong>to</strong> computational algorithms,<br />

remain comparatively more difficult. As well, the very ubiquity and commonplace<br />

character of these technologies may discourage us from attending more carefully <strong>to</strong><br />

whatever more subtle and long-term consequences they may have for us as scholars and<br />

as human beings.


But even these critical reflections, finally, may be an indication of the success and<br />

maturity of the computing revolution in philosophy and religion: perhaps like other<br />

revolutions we now take for granted, the computing revolution has proceeded far enough<br />

along <strong>to</strong> allow us <strong>to</strong> critically evaluate both its strengths and its limits.<br />

See also chapter 4: Classics and the Computer; chapter 8: Literary Studies; chapter 10:<br />

Multimedia.<br />

Notes<br />

1 In the Science Museum (London) exhibition "Charles Babbage and his Calculating<br />

Engine", mounted on the occasion of the bicentennial of Babbage's birth and the<br />

Museum's realization of Babbage's "Difference Engine No. 2", the following is attributed<br />

<strong>to</strong> Leibniz in 1685: "It is unworthy for excellent men <strong>to</strong> lose hours like slaves in the labor<br />

of calculation which could safely be relegated <strong>to</strong> anyone else if machines were used."<br />

References for Further Reading<br />

American Bible Society (1995). A Father and Two Sons: Luke 15.11–32, CD-ROM. New<br />

York: American Bible Society.<br />

American Philosophical Association. Web Resources. Accessed Oc<strong>to</strong>ber 26, 2002. At<br />

http://www.apa.udel.edu/apa/resources/.<br />

Barwise, Jon and John Etchemendy (1999). Language, Proof, and Logic (text/software<br />

package). New York: Seven Bridges Press.<br />

Beam, Kathryn and Traianos Gagos, (eds.) (1997). The Evolution of the English Bible:<br />

From Papyri <strong>to</strong> King James, CD-ROM. Ann Arbor: University of Michigan Press.<br />

BibleWorks (2002). CD-ROM, 2 disks. Norfolk, VA: BibleWorks. At:<br />

http://www.bibleworks.com.<br />

Bolter, Jay David (1991). Writing Space: The Computer, Hypertext, and the His<strong>to</strong>ry of<br />

Writing. Hillsdale, NJ: Lawrence Erlbaum.<br />

Borgmann, Albert (1999). Holding on<strong>to</strong> Reality: The Nature of Information at the Turn of<br />

the Millennium. Chicago: University of Chicago Press.<br />

Brasher, Brenda (<strong>2001</strong>). Give Me that Online Religion. San Francisco: Jossey-Bass.<br />

Bynum, Terrell Ward (<strong>2001</strong>). Computer Ethics: Its Birth and Its Future. Ethics and<br />

Information Technology 3: 109–12.


Bynum T. W. and J. H. Moor, (eds.) (1998). The Digital Phoenix: How Computers are<br />

Changing Philosophy. Oxford: Blackwell.<br />

Bynum T. W. and J. H. Moor, (eds.) (2003). Cyberphilosophy. Oxford: Blackwell.<br />

Careaga, Andrew (<strong>2001</strong>). eMinistry: Connecting with the Net Generation. Grand Rapids,<br />

MI: Kregel.<br />

Cavalier, Robert, Pres<strong>to</strong>n Covey, Elizabeth A. Style, and Andrew Thompson (1998). The<br />

Issue of Abortion in America, CD-ROM. London and New York: Routledge.<br />

Davis, Erik (1998). Techgnosis: Myth, Magic + Mysticism in the Age of Information.<br />

New York: Three Rivers Press.<br />

Dipert, Randall R. (2002). The Substantive Impact of Computers on Philosophy:<br />

Prolegomena <strong>to</strong> a Computational and Information-theoretic Metaphysics.<br />

Metaphilosophy 33, 1/2 (January): 146–57.<br />

Dreyfus, Hubert (<strong>2001</strong>). On the Internet. London and New York: Routledge.<br />

Eck, Diana (2002). On Common Ground: World Religions in America, CD-ROM, 2nd<br />

edn. Affiliated web site: http://www.fas.harvard.edu/~pluralsm/.<br />

Ess, Charles (<strong>2001</strong>). The Word Online? Text and Image, Authority and Spirituality in the<br />

Age of the Internet. Mots Pluriels 9 (Oc<strong>to</strong>ber). At<br />

http://www.arts.uwa.edu.au/MotsPluriels/MP1901ce.html.<br />

Ess, Charles (2003). Philosophy of Computer-mediated Communication. In Luciano<br />

Floridi (ed.), The Blackwell Guide <strong>to</strong> the Philosophy of Information and Computing.<br />

Oxford: Blackwell.<br />

Ess, Charles (2004). Critical Thinking and the Bible in the Age of New Media. Lanham,<br />

MD: University Press of America.<br />

Floridi, Luciano, (ed.) (2003). The Blackwell Guide <strong>to</strong> the Philosophy of Computing and<br />

Information. Oxford: Blackwell.<br />

Fraser, Michael and the Centre for Humanities Computing (2000). Computing Resources<br />

for Theology: An Introduction. Humanities Computing Unit, Oxford University.<br />

Accessed Oc<strong>to</strong>ber 26, 2002. At http://info.ox.ac.uk/ctitext/theology/theolit.html.<br />

Harbin, Duane (1998). Fiat Lux: The Electronic Word. In Formatting the Word of God<br />

(Exhibition catalogue, Bridwell Library, Dallas, Texas). Accessed Oc<strong>to</strong>ber 31, 2002. At<br />

http://www.smu.edu/bridwell/publications/ryrie_catalog/xiii_1.htm.


Hardmeier, Chris<strong>to</strong>f (2000). Was ist Computerphilologie? Theorie, Anwendungsfelder<br />

und Methoden - eine Skizze [What is computer philology? A sketch of theories, fields of<br />

application, and methods]. In C. Hardmeier and W.-D. Syring (eds.), Ad Fontes! Quellen<br />

erfassen - lesen - deuten. Was ist Computerphilologie? Ansatzpunkte und Methodologie -<br />

Instrumente und Praxis [To the fonts! Understanding - reading - pointing <strong>to</strong> sources.<br />

What is computer philology? Beginning points and methodologies - instruments and<br />

praxis] (pp. 9–31). Amsterdam: VU University Press.<br />

Hinman, Lawrence. Ethics Updates. Accessed Oc<strong>to</strong>ber 26, 2002. At<br />

http://ethics.acusd.edu/index.html.<br />

Hippias Limited Area Search of Philosophy on the Internet. Accessed Oc<strong>to</strong>ber 26, 2002.<br />

At http://hippias.evansville.edu/.<br />

Kolb, David (1996). Discourse across Links. In Charles Ess (ed.), Philosophical<br />

Perspectives on Computer-mediated Communication (pp. 15–26). Albany, NY: State<br />

University of New York Press.<br />

Landow, George (1992). Hypertext: The Convergence of Contemporary Critical Theory<br />

and Technology. Baltimore: Johns Hopkins University Press.<br />

Larsen, Elena (<strong>2001</strong>). Cyberfaith: How Americans Pursue Religion Online. Pew Internet<br />

and American Life Project. At<br />

http://www.pewinternet.org/reports/<strong>pdf</strong>s/PIP_CyberFaith_Report.<strong>pdf</strong>.<br />

Lawrence, Bruce B. (2000). The Complete Idiot's Guide <strong>to</strong> Religions Online.<br />

Indianapolis: Macmillan.<br />

O'Leary, Stephen D. and Brenda E. Brasher (1996). The Unknown God of the Internet. In<br />

Charles Ess (ed.), Philosophical Perspectives on Computer-mediated Communication<br />

(pp. 233–69). Albany, NY: State University of New York Press.<br />

Ong, W. J. (1988). Orality and Literacy: The Technologizing of the Word. London:<br />

Routledge.<br />

Perseus 2.0. (1996). CD-ROM. New York and London: Yale University Press.<br />

Society of Biblical Literature. Electronic Publications and Technology Resources for<br />

Biblical Studies. Accessed Oc<strong>to</strong>ber 26, 2002. At http://www.sbl-site.org/e-resources.html.<br />

Suppes, Patrick. Home page. Accessed Oc<strong>to</strong>ber 26, 2002. Philosophy Department,<br />

Stanford University. At http://www.stanford.edu/~psuppes/.<br />

Wiener, Norbert (1950). The Human Use of Human Beings: Cybernetics and Society.<br />

Bos<strong>to</strong>n: Hough<strong>to</strong>n Mifflin.


13.<br />

How the Computer Works<br />

Andrea Laue<br />

Computing humanists engage <strong>digital</strong> technologies in their studies of humanistic artifacts.<br />

In this chapter, we will engage the computer as artifact, seeking a better understanding of<br />

how his<strong>to</strong>ries of culture and of technology converge in our personal computers (PCs).<br />

First, we will explore how computers "work", how their largely electrical mechanisms<br />

process data. Second, we will consider how computers "function", how in theory and in<br />

practice humans figure their working relationships with these machines. The computer<br />

was built <strong>to</strong> be a sophisticated device for the manipulation of symbols, and, at several<br />

levels, it is just that. But this description does not acknowledge that computers are also<br />

symbols, nor does it reveal the extent <strong>to</strong> which computers employ technologies symbolic<br />

of larger his<strong>to</strong>rical and cultural trends. In order <strong>to</strong> understand how computers function,<br />

we must understand how humans use computers <strong>to</strong> manipulate symbols but also how the<br />

mechanisms within the machine manipulate human users. I will consider here "Wintel"<br />

machines, desk<strong>to</strong>p PCs running on technology largely developed by Microsoft and Intel.<br />

These systems draw extensively on John von Neumann's s<strong>to</strong>red-program architecture<br />

(1945) and incorporate primarily Douglas Engelbart's input devices. While popular<br />

rhe<strong>to</strong>ric encourages users <strong>to</strong> imagine <strong>digital</strong> technologies outside of various his<strong>to</strong>ries of<br />

technology, of culture and of labor, I argue here that those his<strong>to</strong>ries are essential <strong>to</strong> a<br />

proper understanding of how a computer functions.<br />

How the Computer Works<br />

In 1945 John von Neumann described a radical new architecture for computers, and<br />

although he was not imagining personal computers, his design largely informs the<br />

machines we use <strong>to</strong>day. I will adopt his concept of the five logical parts of a computer –<br />

control, processing, s<strong>to</strong>rage, output, and input – and discuss them in the context of a more<br />

mundane narrative – just what happens when one types "Hi!" as the first line of an email.<br />

There is not a one-<strong>to</strong>-one correspondence between von Neumann's logical parts and<br />

the hardware components present in modern machines, so I will demarcate a section for<br />

each logical part, providing a basic explanation of the role of that part, and explain in<br />

greater detail how one or two hardware components perform the operations characteristic<br />

of that logical part. My choice of components and sequence of discussion will be guided<br />

by the framing narrative, the s<strong>to</strong>ry of how one types and sends an e-mail that says "Hi!"<br />

but I will occasionally offer asides rather than omit an essential component.<br />

Input<br />

Contemporary computers offer a range of input devices, including keyboards, the mouse,<br />

floppy disk drives, modems, and USB (universal serial bus) ports. Such devices are often<br />

called peripherals, a term that describes any component, internal or external, which adds


hardware capabilities <strong>to</strong> the basic design of the system. Each of these peripherals offers a<br />

means of transferring data <strong>to</strong> the machine, an operation that may seem mundane now but<br />

was key for von Neumann as previous computers used only punchcards and paper tape.<br />

Modern input devices utilize several different standards and media, but the principles<br />

behind the transfer of data share fundamental characteristics. Since the user is likely <strong>to</strong><br />

use a keyboard <strong>to</strong> input the characters "H", "I", and "!", we will start there.<br />

Although most keyboards look quite similar from the outside, there are at least five<br />

different "insides" common <strong>to</strong>day. These differ in the mechanisms engaged by the<br />

striking of a key, and the most common setup, the rubber dome keyboard, will be<br />

described here. Looking inside this keyboard we find three plastic sheets: one with<br />

conductive tracks, one that acts as a separa<strong>to</strong>r, and a final conductive layer. Two of these<br />

<strong>to</strong>gether form the key matrix, a circuit board in two layers – a bot<strong>to</strong>m layer containing<br />

gaps that are completed when the <strong>to</strong>p layer is pressed down on <strong>to</strong> it. The third layer is<br />

shaped like the keyboard, with pliable rubber domes corresponding <strong>to</strong> each rigid plastic<br />

key. These domes contain a hard carbon center that, when depressed, completes a circuit.<br />

When released, the rubber dome regains its original shape, breaking the circuit.<br />

Keyboards have simple microprocessors that constantly scan the circuit board looking for<br />

completed circuits. These detect both an increase and a decrease in current, the pressing<br />

and the releasing of a key. When a completed circuit is found, the microprocessor<br />

compares the completed circuit with its programmed character map, a sort of seating<br />

chart of characters. If more than one circuit is completed, the microprocessor checks <strong>to</strong><br />

see if that combination of keys is recorded in the character map. The character map points<br />

<strong>to</strong> a specific scan code for the completed circuit, and it is this scan code that is sent <strong>to</strong> the<br />

computer for translation in<strong>to</strong> the appropriate ASCII bit string. The scan codes are<br />

standard, but the character map is specific <strong>to</strong> a particular keyboard. The real significance<br />

of the character map lies in the fact that it separates the function of the keyboard from the<br />

letters printed on the plastic keys. If one reprograms the character map, striking the key<br />

labeled "H" might actually signal the microprocessor <strong>to</strong> send the code for the letter "o" <strong>to</strong><br />

the machine. This means that computer keyboards can be reprogrammed for almost any<br />

arrangement of keys; although the peripheral is stamped with a QWERTY keyboard,<br />

nothing prevents one from actually using a Dvorak arrangement.<br />

The keyboard s<strong>to</strong>res the scan code in its memory while it sends an interrupt <strong>to</strong> the<br />

computer's BIOS (basic input/output system). This interrupt tells the computer <strong>to</strong> s<strong>to</strong>p<br />

what it is doing and <strong>to</strong> receive the queued signals from the keyboard. The keyboard<br />

transmits the scan code <strong>to</strong> the keyboard controller, a circuit in the BIOS that forwards<br />

keyboard data <strong>to</strong> the operating system (OS). It is the keyboard controller that decides if<br />

the signal represents a system-level command or data intended for the application<br />

currently in use. In our case, the OS will send our input <strong>to</strong> an e-mail client.<br />

But what exactly is sent between the keyboard and the BIOS, between the BIOS and the<br />

OS? While the characters appearing on the screen as you type match the characters<br />

printed on the keyboard, the machine never knows them as letters and punctuation marks.<br />

All data pass through the computer as nothing more than pulses of electricity; the


machine recognizes two states: with charge and without charge. We often represent these<br />

two states with two digits, "1" and "0" respectively. (This actually reverses the his<strong>to</strong>rical<br />

representation: logical models of computing devices used binary logic previous <strong>to</strong> the<br />

actual construction of electromechanical or purely electrical machines.) This is also the<br />

origin of the term "bit" and by extension "byte." One bit of information is one switch, one<br />

1 or 0, one "place"; eight bits is one byte. So our keyboards and our computers<br />

communicate via electricity, and mathematicians and logicians have developed methods<br />

for translating language (symbols) and many thought processes in<strong>to</strong> binary code.<br />

The most pertinent example for our purposes is the American Standard Code for<br />

Information Interchange (ASCII). The ASCII character set maps the 256 most common<br />

characters <strong>to</strong> binary equivalents, a unique string of eight digits. This is also the origin of<br />

the byte as a measurement of s<strong>to</strong>rage capacity, as one byte of information is sufficient for<br />

representing any character in the ASCII set. Thus, ASCII translates the Western alphabet<br />

in<strong>to</strong> code that can be transmitted as pulses of electricity. While you type in<strong>to</strong> your e-mail<br />

client, the characters are actually being s<strong>to</strong>red as bit strings, as sequences of Is and Os.<br />

Let us return <strong>to</strong> our example, "Hi!"; in binary code: 101000 1101001 0100001.<br />

Standards-based e-mail clients and terminals understand all characters as bit strings, and<br />

it is bit strings that are transmitted from the keyboard <strong>to</strong> the computer.<br />

Control<br />

We will begin our discussion of the control unit at the motherboard, the physical<br />

component that houses the control and processing units. Most of the control unit is<br />

located in the chipset, a collection of microprocessors contained in the CPU (central<br />

processing unit). This chipset includes most essential components of the control unit and<br />

is usually visible on the motherboard. The motherboard is an intricate printed circuit<br />

board with miles of copper circuit paths called traces and numerous sockets for<br />

connecting peripherals. Together the traces constitute the bus, the metaphorical highway<br />

system that carries electricity between the processor, the memory, and the various<br />

peripherals. Buses are the physical connections that transfer data as electronic pulses, and<br />

these connections may be a fraction of an inch in length within the CPU or several inches<br />

long when stretching from the processor <strong>to</strong> an expansion slot. As we will learn later, the<br />

bus is an important fac<strong>to</strong>r of a computer's overall processing speed.<br />

Let us return <strong>to</strong> our depressed keys, our scan codes waiting in the BIOS, the component<br />

responsible for making sure that the various parts of the computer work <strong>to</strong>gether.<br />

Actually a bit of software code, the BIOS is s<strong>to</strong>red in the chipset. Most computers<br />

employ three levels of software: the OS; the device drivers that include instructions for<br />

operating particular peripherals; and the BIOS that translates and transmits instructions<br />

and data between the OS and the drivers. The BIOS handles the transmission of the<br />

electrical signals sent between the various peripherals. So the BIOS receives our scan<br />

code, translates it <strong>to</strong> the appropriate ASCII code, and sends it <strong>to</strong> the OS.<br />

Before we move on from the BIOS, let us review its role in the boot process, the routine<br />

followed each time the computer is powered on or restarted. Most likely, the boot process


is the only time you'll have any direct experience with the BIOS, as you can watch its<br />

work on the screen as your computer starts up. During the boot process, the PC checks its<br />

components and prepares the machine <strong>to</strong> load the OS. The first step in the boot process is<br />

the power-on self-test (POST), which begins when the pressing of the power but<strong>to</strong>n sends<br />

the first electrical pulse <strong>to</strong> the microprocessor, telling it <strong>to</strong> reset its registers and <strong>to</strong> initiate<br />

the boot program s<strong>to</strong>red in the BIOS. This boot program invokes a series of system<br />

checks, testing <strong>to</strong> see that components such as the system bus and clock are functioning<br />

properly. Next, the BIOS contacts the CMOS (complementary metal oxide<br />

semiconduc<strong>to</strong>r), a tiny bit of memory located on the motherboard containing basic<br />

information about the components of the computer. After that, the video adapter is<br />

checked and, if all is well, something appears on the moni<strong>to</strong>r. Next is the test of the<br />

system RAM, during which the machine will seem <strong>to</strong> count each bit of memory. Then the<br />

POST continues through the various system components, checking functionality and,<br />

when appropriate, loading a device driver in<strong>to</strong> the BIOS, ensuring proper and rapid<br />

functioning when that device is called by the OS. In most cases, the BIOS will then print<br />

<strong>to</strong> the screen a summary of the system configuration, which will flash by just before the<br />

OS starts. Then the POST contacts the CMOS again looking for s<strong>to</strong>rage devices tagged as<br />

boot devices, pieces of hardware that might contain an OS. Finally, the BIOS checks the<br />

boot sec<strong>to</strong>r of those devices for a master boot record that will take control of the PC for<br />

the remainder of the startup.<br />

Another important element of the control unit is the system clock. The clock is really an<br />

oscilla<strong>to</strong>r, a tiny quartz crystal that vibrates when electricity is applied <strong>to</strong> it. Thus the<br />

fundamental mechanism is the same as that which drives most <strong>digital</strong> watches and<br />

electronic clocks. The vibrations of the clock cause pulses of electricity <strong>to</strong> flow through<br />

the motherboard, moving bits from place <strong>to</strong> place. Thus the rate of vibration is the<br />

shortest time in which an operation can be executed, and this rate is often called the clock<br />

speed. Many modern computers operate at speeds greater than 1 GHz (gigahertz), or<br />

more than 1 billion cycles per second. The clock is necessary because different parts of<br />

the computer, even different operations within the processor, operate at different rates. It<br />

is the job of the clock <strong>to</strong> act as a master control, <strong>to</strong> synchronize the flow of bits through<br />

the machine. The bits that compose our greeting will be transmitted according <strong>to</strong> the<br />

rhythm enforced by the oscilla<strong>to</strong>r, by the "ticking" of the clock. Our three bytes of<br />

information will take a minimum of twenty-four cycles – an unimaginably small fraction<br />

of a second – for each "step" between the keyboard and the modem.<br />

Processor<br />

The microprocessor is probably the most recognizable component of the computer. Intel's<br />

first microprocessor, the 8080 introduced in 1974, had 6,000 transis<strong>to</strong>rs; Intel's most<br />

recent microprocessor, the Pentium 4, includes 42 million transis<strong>to</strong>rs. The<br />

microprocessor, also called the central processing unit (CPU), has three basic<br />

capabilities: <strong>to</strong> perform basic mathematical operations, move data from one memory<br />

location <strong>to</strong> another, make decisions <strong>to</strong> jump from one instruction <strong>to</strong> another. At the heart<br />

of the processor is the arithmetic/logic unit (ALU), which performs very simple<br />

operations such as adding, subtracting, multiplying, dividing, and completing some basic


logical operations. And each of these operations is performed on bit strings like the one<br />

that constitutes our e-mail greeting. Data and instructions are all translated in<strong>to</strong> Is and Os,<br />

pulses of electricity, and the processor is a mass of transis<strong>to</strong>rs, of switches, that can hold<br />

and manipulate these charges. Our e-mail greeting will flow through the CPU even<br />

though it will not be manipulated during its passing.<br />

S<strong>to</strong>rage<br />

When we use the word "s<strong>to</strong>rage" we usually mean hard drives, floppy disks, or other<br />

relatively permanent but slow means of s<strong>to</strong>ring data. There are four major types of<br />

internal s<strong>to</strong>rage in modern computers, listed in order of decreasing speed and cost:<br />

registers, cache, random access memory (RAM), and hard disk drives (HDD). Registers<br />

and cache are located in the CPU or on the motherboard and come in relatively small<br />

amounts, say 512 MB (megabytes). In contrast, hard drives often have 80 GB (gigabytes)<br />

of s<strong>to</strong>rage for a fraction of the cost of the cache. We will learn a bit about RAM and hard<br />

drives, two locations where our greeting might be s<strong>to</strong>red.<br />

A RAM chip is an integrated circuit made of millions of pairs of transis<strong>to</strong>rs and<br />

capaci<strong>to</strong>rs that are linked <strong>to</strong> create a memory cell. Each cell can hold one bit of data. A<br />

capaci<strong>to</strong>r holds a charge, and the transis<strong>to</strong>r acts as a switch, allowing the control circuitry<br />

<strong>to</strong> change the state of the capaci<strong>to</strong>r. These capaci<strong>to</strong>rs and transis<strong>to</strong>rs are arranged as a<br />

grid, with the columns connected by bitlines and the rows connected by wordlines. In<br />

both cases, a "line" is a microscopic strand of electrically conductive material etched in<strong>to</strong><br />

the RAM chip. When a burst of electricity is transmitted through a bitline, every closed<br />

transis<strong>to</strong>r, each switch that is "on", charges its capaci<strong>to</strong>r. The combination of bitlines and<br />

wordlines defines an address for each switch and capaci<strong>to</strong>r. This also accounts for why<br />

this type of memory is called random access: you can access any memory cell directly if<br />

you know its address. Most RAM is volatile, or dynamic, meaning that the capaci<strong>to</strong>rs are<br />

always leaking their charge and thus must be refreshed constantly. RAM is inserted<br />

directly in<strong>to</strong> the motherboard, most often in the form of a dual in-line memory module<br />

(DIMM) that holds up <strong>to</strong> 128 MB of memory.<br />

A hard disk drive is one of the few computer components that is mechanical as well as<br />

electronic. Hard drives contain a collection of platters and heads arranged so that the gaps<br />

between them are smaller than a human hair; thus hard drives are very sensitive <strong>to</strong> dust<br />

and jostling. The round platters are made of aluminum or glass coated with a magnetic<br />

material. The surface of the platter is divided two ways, as a series of concentric circles<br />

and a collection of wedges, and the resulting pieces are called sec<strong>to</strong>rs. Each sec<strong>to</strong>r<br />

contains a set number of bytes, usually 256 or 512. A single hard drive may contain as<br />

many as eight platters, and platters are stacked on a central mo<strong>to</strong>rized spindle that spins<br />

the platters so that the disk heads can read and write data. The heads write data by<br />

aligning the magnetic particles on the surface of the platter; the heads read data by<br />

detecting the magnetic charge of particles that have already been aligned. Again, all data<br />

is s<strong>to</strong>red as a charge, as a 0 or a 1.


How does the drive manage those eight disks, keeping track of where files are s<strong>to</strong>red and<br />

not inadvertently overwriting important data? Every Windows hard drive has a virtual file<br />

allocation table (VFAT), a sort of database of addresses for files. The VFAT records the<br />

sec<strong>to</strong>rs holding any saved file and offers empty sec<strong>to</strong>rs for writing. When you delete a<br />

file, the hard drive doesn't actually modify the data sec<strong>to</strong>rs. Rather, it edits the VFAT,<br />

removing the entry for that file. Thus the file is invisible but not erased. Depending on<br />

our e-mail client, our greeting has either been saved <strong>to</strong> RAM or <strong>to</strong> the hard drive while it<br />

waits <strong>to</strong> be sent.<br />

Output<br />

Output devices include moni<strong>to</strong>rs, printers, modems, Ethernet cards, and speakers. We<br />

will forgo a discussion of moni<strong>to</strong>rs in favor of a brief account of modems, the hardware<br />

device through which most of us were introduced <strong>to</strong> the Internet. The modem is<br />

particularly interesting because it is a point of union, of conversion between the <strong>digital</strong><br />

and the analogue, between new and (relatively) old technologies of communication. The<br />

modem gets its name from its primary tasks: modulating and demodulating. As we<br />

continue <strong>to</strong> discover, the computer operates exclusively with Is and Os, with a <strong>digital</strong><br />

code. By contrast, phone lines work with an analogue signal, an electronic current with<br />

variable frequency and strength. It is the job of the modem <strong>to</strong> demodulate data from a<br />

phone line, <strong>to</strong> convert the wave <strong>to</strong> a series of Is and Os, and <strong>to</strong> modulate the data<br />

generated by the computer, <strong>to</strong> convert the Is and Os <strong>to</strong> a continuous wave. This process<br />

accounts in part for the limited speed of modems. The analogue signal in the phone line<br />

cannot change frequencies as quickly as a computer can send bits of information, yet the<br />

signal needs <strong>to</strong> change for every bit sent. Clever engineers devised techniques for<br />

avoiding this bottleneck, including defining four frequencies and four associated<br />

combinations of bits – 00, 11, 01, 10 – and experimenting with variations in amplitudes,<br />

but the physical medium of the telephone line still poses material constraints on the speed<br />

of transmissions. This means that our e-mail greeting, s<strong>to</strong>red temporarily in RAM or<br />

perhaps saved <strong>to</strong> the hard drive, and displayed on the screen, will need <strong>to</strong> be translated <strong>to</strong><br />

analogue before being sent <strong>to</strong> a friend.<br />

Review<br />

Let us review what just happened. We press a few keys on our keyboards, and each strike<br />

of a key completes a circuit mapped on the key matrix. Comparing the completed circuit<br />

<strong>to</strong> the key matrix, the keyboard's microprocessor selects a scan code for each character<br />

and transmits it <strong>to</strong> the computer's BIOS for decoding. After deciding that the keyboard<br />

data are bound for our e-mail client, the BIOS transmits the ASCII bit string for each<br />

character. These bit strings are s<strong>to</strong>red temporarily in RAM or, if the e-mail has been<br />

saved, more permanently on our hard drive. When we send our e-mail, the ASCII bit<br />

strings will be sent <strong>to</strong> the modem, which will modulate the data before transferring it <strong>to</strong><br />

the phone line.<br />

Ending our discussion of how a computer works with the modem emphasizes the<br />

significance of the peripheral as a transitional device. Modems operate at the boundary


etween analogue and <strong>digital</strong> technologies, between nineteenth- and twentieth-century<br />

technologies of communication, between transmission and simulation. At multiple levels<br />

the machine is involved in sophisticated translations of information, as is evidenced by<br />

the fact that you really do not need <strong>to</strong> understand how a computer works <strong>to</strong> use a<br />

computer. A great accomplishment of software developers is the extent <strong>to</strong> which we can<br />

"talk" <strong>to</strong> computers using our own language, the extent <strong>to</strong> which the code, the Is and Os,<br />

the electric pulses are hidden. Many pioneers in computing technologies aspired <strong>to</strong> build<br />

machines capable of sophisticated symbol manipulation, as they unders<strong>to</strong>od that such<br />

skill – basically the ability <strong>to</strong> use language – is essential <strong>to</strong> human intelligence. Yet in<br />

simulating this capacity they added several layers of symbols, several additional levels of<br />

abstraction: printed letter <strong>to</strong> ASCII codes <strong>to</strong> a series of pulses.<br />

How the Computer Functions<br />

To begin, I will restate my guiding distinction: "work" describes the electronic processing<br />

performed by the machine; "function" refers <strong>to</strong> the theoretical and practical relations<br />

between the computer and its human opera<strong>to</strong>rs. To explore this distinction, I focus on the<br />

system clock and the keyboard, but my evidence could have included several other<br />

components, most notably computer memory (Douwe 2000; Locke 2000; Williams<br />

1997). I discuss the his<strong>to</strong>ry of the adoption of and adaptation <strong>to</strong> these technologies,<br />

looking at how the computer as artifact reproduces metaphors of mind and incorporates<br />

bodies in<strong>to</strong> mechanized routines. Relating the system clock and the computer keyboard <strong>to</strong><br />

previous implementations of similar technologies, I explore motivations and rationales<br />

for the choices behind these particular mechanisms. Pioneers of <strong>digital</strong> technologies often<br />

thought their designs radical and their machines revolutionary, yet these inventions<br />

incorporate conventions and habits embedded in previous technologies. In this sense,<br />

<strong>digital</strong> technologies can be seen as extensions of numerous his<strong>to</strong>rical trends.<br />

System clock<br />

Let us begin by returning <strong>to</strong> John von Neumann, focusing this time on the biological<br />

metaphors that frame his understanding of his radical architecture. Trained in logic and<br />

mathematics, von Neumann's interest in information processing was strongly influenced<br />

by the work of Warren McCulloch and Walter Pitts, in particular the article "A Logical<br />

Calculus of the Ideas Immanent in Nervous Activity" (1943). Although he acknowledged<br />

that the McCulloch-Pitts model oversimplified neural activity, von Neumann was<br />

intrigued by the parallels proposed between man-made and organic systems (Aspray<br />

1990: 180–1). In 1944 von Neumann joined the likes of McCulloch, Pitts, and Norbert<br />

Wiener <strong>to</strong> form the Teleological Society, a research group "devoted on the one hand <strong>to</strong><br />

the study of how purpose is realized in human and animal conduct and on the other hand<br />

how purpose can be imitated by mechanical and electrical means" (Aspray 1990: 315).<br />

Although the Teleological Society disbanded, many of its founding members initiated the<br />

Macy Conferences, probably the most famous early discussions of artificial intelligence.<br />

Looking at von Neumann's First Draft of a Report on the EDVAC (1945), the first<br />

publication describing the s<strong>to</strong>red-program architecture, one notices his analogies <strong>to</strong>


organic systems. For instance, in this report the s<strong>to</strong>rage unit of a computer is first called<br />

"memory", and other logical parts are called "organs." Von Neumann unders<strong>to</strong>od organic<br />

systems <strong>to</strong> be highly complex, <strong>to</strong> employ parallel processing and <strong>to</strong> be adept at switching<br />

between analogue and <strong>digital</strong> modes of computation, and he planned <strong>to</strong> simulate the<br />

organic by implementing a few key principles – process hierarchies, serialization,<br />

regulation, and au<strong>to</strong>mation. Along with others in his field, von Neumann rationalized the<br />

organic and made discrete the continuous in an effort <strong>to</strong> mimic the intellect of man.<br />

Aware that current technology couldn't accomplish the sophistication of organic systems,<br />

von Neumann intended <strong>to</strong> design an architecture that could evolve sufficient complexity.<br />

He modeled the workings of his computer after the brain, hoping that the machine would<br />

function like – or at least with – man. In an attempt <strong>to</strong> replicate the functioning of the<br />

human mind, von Neumann built a model of the brain that worked in ways quite different<br />

than its organic precursor. The mechanics of the computer introduce several technologies<br />

foreign <strong>to</strong> the mind yet not so unfamiliar <strong>to</strong> the his<strong>to</strong>ry of its acculturation and regulation.<br />

Our brief investigation of these mechanics will focus on the system clock, a device<br />

central both <strong>to</strong> the work and <strong>to</strong> the function of the computer. His<strong>to</strong>rians of technology<br />

often point <strong>to</strong> the clock as a mechanism central <strong>to</strong> the modernization of the West. In his<br />

his<strong>to</strong>ry of technology Technics and Civilization, Lewis Mumford (1934: 14) writes: "The<br />

clock, not the steam-engine, is the key-machine of the modern industrial age. For every<br />

phase of its development the clock is both the outstanding fact and the typical symbol of<br />

the machine: even <strong>to</strong>day no other machine is so ubiqui<strong>to</strong>us." Mumford published his<br />

book a decade before von Neumann designed his s<strong>to</strong>red-program architecture, the first<br />

computer architecture regulated by a control unit called a "clock." In his Report, von<br />

Neumann describes the electromechanical devices used <strong>to</strong> control computers previous <strong>to</strong><br />

the EDVAC and then proposes an alternative for his new machine: "Alternatively,<br />

[computing instruments] may have their timing impressed by a fixed clock, which<br />

provides certain stimuli that are necessary for its functioning at definite periodically<br />

recurrent moments" (1945: 24). Even at this early stage von Neumann suggested that this<br />

"clock" might actually be an electronic oscilla<strong>to</strong>r controlled by a crystal. Two years later,<br />

von Neumann, Goldstine, and Burks described dynamic circuits, regulated by electric<br />

pulses emitted from a central source called a "clock": "Since the timing of the entire<br />

computer is governed by a single pulse source, the computer circuits will be said <strong>to</strong><br />

operate as a synchronized system" (Burks et al. 1947: 131). Thus, the clock is essential <strong>to</strong><br />

the proper working of the computer, and in this section I will argue that it is key <strong>to</strong><br />

understanding the functioning of the computer as well.<br />

The system clock lacks most mechanisms associated with clocks: it has neither gears nor<br />

hands, neither a face nor an audible signal <strong>to</strong> mark the minutes. Many advanced computer<br />

science books refer initially <strong>to</strong> a clock but later drop the metaphor and describe the<br />

oscilla<strong>to</strong>r, the tiny crystal whose vibrations regulate the flow of electricity through the<br />

motherboard. The etymology of "clock" relates our modern machines <strong>to</strong> medieval bells,<br />

in particular <strong>to</strong> bells that were struck <strong>to</strong> mark the hours (OED 1989). Samuel Macey<br />

extends this relation, figuring the clock as a sort of au<strong>to</strong>ma<strong>to</strong>n: "the earliest mechanical<br />

clocks may be considered au<strong>to</strong>mated versions of the keeper of the bell or Glocke, which<br />

gave us the term clock in the first place" (Macey 1989: 1 1). Gerhard Dohrn-van Rossum,


his<strong>to</strong>rian and author of His<strong>to</strong>ry of the Hour (1996), traces the his<strong>to</strong>ry of man's relation <strong>to</strong><br />

time-keeping devices, focusing less on the technology underlying the machines and more<br />

on man's evolving relation <strong>to</strong> the device. I draw on Dohrn-van Rossum's account,<br />

foregrounding his argument that time-keeping devices have functioned <strong>to</strong> synchronize,<br />

regulate, and abstract time.<br />

Well before mechanical clocks were sufficiently precise <strong>to</strong> reliably mark the hours, bells<br />

divided the day and synchronized the activities of a populace. Mumford points <strong>to</strong> the<br />

monasteries as the origin of modern notions of time-keeping, arguing that the<br />

Benedictines "helped <strong>to</strong> give human enterprise the regular collective beat and rhythm of<br />

the machine; for the clock is not merely a means of keeping track of the hours, but of<br />

synchronizing the actions of men" (1934: 14). Church bells also marked the hours outside<br />

the monastery, alerting laypersons <strong>to</strong> the hours for prayers and for mass, and by the<br />

fifteenth century many municipalities had acquired communal bells <strong>to</strong> signal secular<br />

meetings and events. Through variations in <strong>to</strong>ne, duration, and number of rings, churches<br />

and municipalities developed elaborate codes <strong>to</strong> signal everything from mass <strong>to</strong> changing<br />

of the guards, from noontime break <strong>to</strong> opening of the courts. In both contexts – religious<br />

and secular – these bells announced hours that were irregular according <strong>to</strong> modern<br />

standards but nevertheless succeeded in improving the frequency with which monks and<br />

citizens were "on time" for various functions. In both contexts, the bells were functional<br />

well before devices for keeping time really worked.<br />

After bells, civic officials developed several time-keeping technologies <strong>to</strong> regulate<br />

everything from the marketplace <strong>to</strong> the postal system. Beginning in the thirteenth century,<br />

officials regulated the market with clocks, setting times at which particular merchants<br />

could do business and hours during which locals could shop. Previous <strong>to</strong> the fifteenth<br />

century, universities varied the time and duration of lectures according <strong>to</strong> content. By the<br />

fifteenth century, particularly in secondary schools, days were divided in<strong>to</strong> discrete<br />

blocks during which specific subjects were taught, and these hours were measured by<br />

bells and sandglasses. In addition <strong>to</strong> measuring and dividing the day, timekeeping devices<br />

increasingly aided the rationalizing of space. Between the fifteenth and eighteenth<br />

centuries, postal and courier services increasingly abstracted distance by measuring it in<br />

units of time. Delivery times once measured in days were soon measured in hours, and<br />

soon promptness became the measure of quality: the normative codes "should thus be<br />

unders<strong>to</strong>od as concepts for systematically organized transportation and communication<br />

links, which, if we trace them over a longer period, reveal the growing importance of the<br />

techniques of time measurement and time control" (Dohrn-van Rossum 1996: 343). Even<br />

before railroads, the technology most associated with the strict enforcement of<br />

standardized time, the markets, the schools, and the postal systems were regulated by<br />

increasingly abstract notions of time (and space) (Dohrn-van Rossum 1996).<br />

It was not until the mid-seventeenth century that these functional time-keeping devices<br />

developed in<strong>to</strong> working clocks. In 1656, Christiaan Huygens patented a pendulum clock<br />

that was accurate <strong>to</strong> within one minute a day, and by the late seventeenth century he had<br />

developed a balance wheel and spring assembly that could operate a pocket watch with<br />

an accuracy of 10 minutes per day. A substantial improvement over previous time-


keeping mechanisms, the pendulum was subject <strong>to</strong> inaccuracies due <strong>to</strong> changes in<br />

temperature, air pressure, and gravity. Some 200 years later, quartz crystals replaced<br />

pendulums in the 1920s, further enhancing the accuracy of clocks. Using an electric<br />

circuit <strong>to</strong> cause a crystal <strong>to</strong> vibrate, quartz clocks shed the gears and escarpments – the<br />

tangible mechanisms – that often led <strong>to</strong> variations in time. Just thirty some years after the<br />

quartz clock, an a<strong>to</strong>mic clock that was accurate <strong>to</strong> one second in 2,000 years appeared in<br />

1955. The original definition of "second" was 1/86,400 of a mean solar day; in 1967 the<br />

natural frequency of the cesium a<strong>to</strong>m – 9,192,631,770 oscillations of the a<strong>to</strong>m's resonant<br />

frequency – was declared the international unit of time. Thus, by the late twentieth<br />

century measures of time had been abstracted <strong>to</strong> the point that they seemed <strong>to</strong> have little<br />

relation <strong>to</strong> the daily rhythms – solar or social – of man.<br />

Modern system clocks utilize quartz crystals, the same technology that runs many <strong>digital</strong><br />

watches, and von Neumann first suggested using these oscilla<strong>to</strong>rs in his 1945 Report.<br />

Following a brief overview of his entire architecture, von Neumann began his section<br />

titled "Elements, Synchronism Neuron Analogy" with a description of the control unit<br />

(1945: 23). He argued that all <strong>digital</strong> computing devices must have control "elements"<br />

with discrete equilibrium and that a synchronous "element" – a system clock – is the<br />

preferable control mechanism. Going <strong>to</strong> the trouble <strong>to</strong> define "element" as a synchronous<br />

device with discrete equilibrium, he proceeded <strong>to</strong> a biological analogy: "It is worth<br />

mentioning, that the neurons of the higher animals are definitely elements in the above<br />

sense. They have all-or-none character, that is two states: Quiescent and excited" (1945:<br />

24). Drawing on the work of MacCulloch and Pitts, von Neumann acknowledged that<br />

neurons were probably asynchronous but proposed that their behavior be simulated using<br />

synchronous systems (ibid.). Thus, von Neumann chose <strong>to</strong> implement a system clock<br />

because he was attempting <strong>to</strong> mimic the functioning of the human brain. Unable <strong>to</strong> build<br />

a machine that worked like a brain, he imagined constructing a computer that (apparently)<br />

functioned like a mind: "Von Neumann, especially, began <strong>to</strong> realize that what they were<br />

talking about was a general-purpose machine, one that was by its nature particularly well<br />

suited <strong>to</strong> function as an extension of the human mind" (Rheingold 2000: 86). Essential <strong>to</strong><br />

this machine was a clock, a metronome that made regular, synchronous, and rational this<br />

model of the mind. Attempting <strong>to</strong> extend the capacities of the human mind, von<br />

Neumann's design actually recalled technologies previously used <strong>to</strong> control it.<br />

Paradoxically, von Neumann's architecture is at once flexible and rigid, is both an<br />

extension of and a constraint on our conceptual, cultural, and practical understanding of<br />

our minds.<br />

QWERTY keyboard<br />

In this section we will investigate the QWERTY keyboard, using a distinction between<br />

"machine" and "<strong>to</strong>ol" <strong>to</strong> frame our discussion. Karl Marx in Capital (1867) and later<br />

Lewis Mumford in Technics and Civilization (1934) define these terms in reference <strong>to</strong> the<br />

his<strong>to</strong>ry of labor and production. Without referring directly <strong>to</strong> Marx or Mumford, J. C. R.<br />

Licklider and Douglas Engelbart, early designers of computer interface and input devices,<br />

later defined a plan for man-computer symbiosis that echoes this tradition. Marx writes<br />

that the origin or impetus of movement is the essential difference between a <strong>to</strong>ol and a


machine: with <strong>to</strong>ols, movement originates in the laborer; with machines, movement<br />

issues from the mechanism itself (1867: 409). Working in an environment with machines<br />

requires education, that the body be trained <strong>to</strong> move along with the uniform and relentless<br />

motion of the mechanism (1867: 408). Conversely, the <strong>to</strong>ol allows the laborer freedom of<br />

movement, an opportunity for motion independent of the mechanism. Working from<br />

Marx, Mumford associates <strong>to</strong>ols with flexibility and machines with specialization. The<br />

essential difference, according <strong>to</strong> Mumford, is the degree of independence of operation.<br />

To summarize, a man works with a <strong>to</strong>ol as an extension of his own body (and perhaps his<br />

own mind); in contrast, a machine employs a man as an extension of its own mechanism.<br />

Licklider (1960) essentially fears that computers are becoming machines rather than<br />

<strong>to</strong>ols. Describing computers built previous <strong>to</strong> 1960 as "semi-au<strong>to</strong>matic systems", he<br />

laments the trend <strong>to</strong>ward au<strong>to</strong>mation at the expense of augmentation. He sees man<br />

incorporated in<strong>to</strong> the system rather than being extended by it: "In some instances,<br />

particularly in large computer-centered information and control systems, the human<br />

opera<strong>to</strong>rs are responsible mainly for functions that it proved infeasible <strong>to</strong> au<strong>to</strong>mate"<br />

(1960: 60). The human appears as a compromise required by the current state of<br />

technology, as a part that will be replaced once the memory is faster or the arithmetic unit<br />

more efficient. However, computers might act as extensions of humans, as devices for the<br />

augmentation of the intellect, as components in human-machine "symbiotic" systems. His<br />

proposed symbiosis is peculiar because he wants <strong>to</strong> introduce intuition and "trial-anderror<br />

procedures" in<strong>to</strong> the system; as later expressed by Douglas Engelbart (1963), "we<br />

refer <strong>to</strong> a way of life in an integrated domain where hunches, cut-and-try, intangibles, and<br />

the human 'feel for a situation' usefully coexist with powerful concepts, streamlined<br />

terminology and notation, sophisticated methods, and high-powered electronic aids"<br />

(1963: 72). Licklider and Engelbart differ from von Neumann in that they do not want the<br />

computer <strong>to</strong> imitate the intellect but rather <strong>to</strong> enhance it. For both Licklider and<br />

Engelbart, organization is a component of intelligence, a function that could perhaps be<br />

better handled by a computer. But the essence of intelligence is a sort of synergy, a<br />

cooperative interplay between the human capabilities of organization and intuition,<br />

analysis and synthesis. While the language of progress and evolution employed by<br />

Licklider and Engelbart may be troubling, their visions of computers as extensions rather<br />

than imitations of man, as <strong>to</strong>ols rather than machines, was revolutionary and remains<br />

relevant <strong>to</strong>day.<br />

The computer keyboard exemplifies this tug-of-war between machine and <strong>to</strong>ol. Most<br />

know that the modern computer keyboard continues the tradition of the typewriter and its<br />

less-than-intuitive arrangement of keys. Often referred <strong>to</strong> as the "QWERTY" keyboard,<br />

this arrangement of keys was devised by Chris<strong>to</strong>pher Sholes, and the first Reming<strong>to</strong>n<br />

typewriters sold in 1873 implemented his design. For decades after the first Reming<strong>to</strong>n,<br />

manufacturers experimented with alternative arrangements, including a circular layout<br />

and an arrangement that featured keys in alphabetical order. Probably the most famous<br />

redesign was the Dvorak Simplified Keyboard, patented in 1936. August Dvorak claimed<br />

that his keyboard led <strong>to</strong> increased speed, reduced fatigue, and more rapid learning. There<br />

is much dispute about the data used <strong>to</strong> support these claims (Liebowitz and Margolis<br />

1990), and more recent experimental evidence suggests that the Dvorak keyboard leads <strong>to</strong>


only a 10 percent improvement in speed (Norman 1990: 147). Yet the QWERTY<br />

keyboard remains a standard even though there exists an empirically superior option. Don<br />

Norman argues that such a small improvement does not warrant the institutional costs of<br />

conversion: "millions of people would have <strong>to</strong> learn a new style of typing. Millions of<br />

typewriters would have <strong>to</strong> be changed. The severe constraints of existing practice prevent<br />

change, even where the change would be an improvement" (1990: 150). Others add <strong>to</strong><br />

this interpretation, citing the Dvorak keyboard as an example of the failure of the market<br />

economy (David 1985). Rather than entering this debate, I suggest that we accept parts of<br />

each argument – the Dvorak keyboard is not a marvelous improvement over the<br />

QWERTY arrangement, and the institutional cost of conversion is great – and then<br />

investigate the process through which QWERTY gained form and acceptance.<br />

While the QWERTY arrangement might not be the ideal arrangement of keys<br />

independent of mechanism used <strong>to</strong> print those keys, the arrangement was the most<br />

functional of Sholes's designs. The design problem faced by Sholes was that of<br />

overlapping rods: if a typist struck in sequence two adjacent keys, their rods would<br />

intersect and jam. Then the opera<strong>to</strong>r would have <strong>to</strong> s<strong>to</strong>p typing and dislodge the<br />

mechanism, losing valuable time. Sholes's keyboard minimized the likelihood of<br />

jamming by separating keys for letters that frequently appear adjacent <strong>to</strong> one another in<br />

common words. The arrangement aims <strong>to</strong> achieve maximum speed through maximum<br />

synchronization of opera<strong>to</strong>r and machine. It is often stated that Sholes designed the<br />

keyboard <strong>to</strong> slow down opera<strong>to</strong>rs of the typewriter; in fact, Sholes designed the keyboard<br />

so as <strong>to</strong> better coordinate the actions of the machine and its opera<strong>to</strong>r, not just <strong>to</strong> slow the<br />

opera<strong>to</strong>r but <strong>to</strong> develop a sequence and a rhythm that was manageable by both human and<br />

machine. Bliven (1954) comments on the sound of a room full of typists, on the particular<br />

rhythm distinct <strong>to</strong> adept typists. Sholes's design of the machine was the engineering of<br />

this rhythm, a reorganization of the interface of the typewriter <strong>to</strong> facilitate the most<br />

efficient use of the mechanism.<br />

Today we assume that good typists use all fingers, but this technique was not common<br />

until fifteen years after the typewriter went <strong>to</strong> market. In fact, there was much debate<br />

surrounding the feasibility of teaching this skill. Early typists used the "hunt and peck"<br />

method with remarkable success. In 1882, Mrs L. V. Longley issued a pamphlet<br />

proposing a method of typing that used all ten fingers. A typing instruc<strong>to</strong>r in Cincinnati,<br />

Mrs Longley met with great opposition in the trade press. Even five years after the<br />

publication of her pamphlet, trade magazines carried s<strong>to</strong>ries arguing the impracticality of<br />

her proposal, citing a lack of strength and dexterity in some fingers. Five years later,<br />

Frank McGurrin, a court stenographer from St Louis, again proposed a ten-finger method<br />

of typing. He added <strong>to</strong> his technique the skill of "<strong>to</strong>uch typing", the ability <strong>to</strong> type without<br />

looking at the keyboard. He also met with resistance but silenced his adversaries when, in<br />

1888, he won a typing contest using his new method. McGurrin used a Reming<strong>to</strong>n<br />

typewriter, and his vic<strong>to</strong>ry led <strong>to</strong> the adoption of two standards: the QWERTY keyboard<br />

and the <strong>to</strong>uch-typing method (Norman 1990: 147). Following McGurrin's vic<strong>to</strong>ry, the<br />

same publication that berated Mrs Longley endorsed <strong>to</strong>uch-typing. It was accepted that<br />

typewriters might be more than twice as fast as writing by hand, and typists were<br />

retrained <strong>to</strong> use all fingers.


In the early twentieth century, typewriter companies employed championship typists <strong>to</strong><br />

compete in national competitions. These racers were given a racing typewriter and a<br />

salary for practicing their skills eight hours a day. Racing typists netted 120 error-free<br />

words per minute in front of crowds of businessmen at the various trade shows in<br />

Madison Square Garden. Once most speed typists mastered the art of <strong>to</strong>uch-typing, other<br />

inefficiencies in the process were targeted. Charles Smith, the coach of the famous<br />

Underwood Speed Training Group, developed the "right-hand carriage draw", a new<br />

method of exchanging paper that saved as much as a full second. This method required<br />

the typist <strong>to</strong> grab the roller knob and the filled sheet with one hand while grabbing a fresh<br />

sheet from a prepared stack of papers and inserting it without ever looking away from the<br />

copy. All of this in one-third of a second (Bliven 1954: 125)! Techniques introduced in<br />

the racing circuit gradually infiltrated the business world, producing typists with elegant<br />

and precise routines <strong>to</strong> synchronize their movements with the machine. Rhythm was key:<br />

"Above all, the mark of good typing is a continuous, even rhythm. An expert keeps<br />

going" (Bliven 1954: 143). Thus the <strong>to</strong>uch-typing system prescribed routines for striking<br />

the keys and for manipulating the impacted media, thereby rendering the typists<br />

extensions of their typewriters, parts of their machines.<br />

In the late nineteenth century, business required improvements in technologies of<br />

communication and record-keeping, and commercial culture embraced the typewriter and<br />

its promise of increased speed and accuracy of transcription. But other cultures of use<br />

emerged as well. George Flanagan writes: "the idea of Art in typewriting has been sadly<br />

ignored" (1938: vi). Calling the machine "mechanically perfect" and typists "marvelous[ly]<br />

skilled", he writes a treatise in hopes of diverting attention from efficiency<br />

<strong>to</strong>wards aesthetics. Flanagan recommends harmony and balance through the judicious<br />

arrangement of space and compactness (1938: 2). Utility should not be achieved at the<br />

expense of elegance; speed can be sacrificed for ornament. Inserting typewriting in<strong>to</strong> the<br />

tradition of printing, Flanagan advises the typists <strong>to</strong> plan the layout of the page just as one<br />

plans the content of the paragraphs. He provides sample layouts – spacing, alignment,<br />

sizes for paragraphs – and includes formulas – combinations and sequences of keystrokes<br />

– for generating ornamental drop caps and borders. Although he recognizes the pressures<br />

of an office and the primacy of speed, Flanagan argues that typists should not ignore<br />

design. They might even profit from it: "Your work will become an endless romance, and<br />

we even suspect that your pay envelopes will grow fatter" (1938: vii). Seeing the typist as<br />

a craftsman rather than a laborer, Flanagan reshapes the machine as a <strong>to</strong>ol. Although<br />

entertaining, the point isn't the aesthetics of ASCII art; rather, it is the revision of what<br />

had become a mechanistic relationship between a human and a machine. The idea of<br />

typing contests and of ornamental typing might seem equally trite <strong>to</strong>day, but their<br />

his<strong>to</strong>rical value lies in the relationships they posit between typist and typewriter, between<br />

opera<strong>to</strong>r and <strong>to</strong>ol or machine.<br />

Although most computer keyboards follow Sholes's design, Engelbart experimented with<br />

many different input devices that drew on various traditions of symbol manipulation.<br />

Most interesting in this context is the chord keyset, a group of five piano-like keys that<br />

were operated by one hand. The keys were not printed with any characters, and opera<strong>to</strong>rs<br />

used single keys and combinations of keys <strong>to</strong> produce up <strong>to</strong> 32 different codes, each of


which represented a different character. Drawing on technologies developed by<br />

telegraphers, Engelbart expected that computer users would be trained in this code and<br />

would eventually surpass the speeds of typists using QWERTY keyboards. The chord<br />

keyset would also allow users <strong>to</strong> manipulate a keyset and a mouse at the same time<br />

without ever looking at their hands. Engelbart imagined a reincarnation of the Reming<strong>to</strong>n<br />

racer who could change paper without ever losing her place in the copy. His input devices<br />

would become extensions of the hands, and users would manipulate them without the aid<br />

of the other senses. Yet such facility would require learning another language, a new<br />

code, an additional symbol set that in its abstraction matched closely the binary set used<br />

by the computer itself. The boundary between man and machine seemed permeable, but<br />

man seemed <strong>to</strong> be the one crossing the line, the one adopting the characteristics of the<br />

other.<br />

Engelbart argued in general terms that users would need <strong>to</strong> be trained <strong>to</strong> use computers,<br />

and the retraining necessary <strong>to</strong> use the chord keyset was just one step in that process. Of<br />

course, the typist's keyboard was eventually chosen over the telegrapher's, largely due <strong>to</strong><br />

the anticipated institutional costs of retraining, and Engelbart's demo now seems rather<br />

quaint. But the choice <strong>to</strong> adopt the QWERTY keyboard demonstrates a few interesting<br />

points. First, we no longer recognize the extent <strong>to</strong> which typing – even the hunt-and-peck<br />

method – requires training; the typist's keyboard seems somehow natural. Second, while<br />

we resist new technologies that require abstract, mechanistic motion, we willfully<br />

embrace old ones that require equally abstract and mechanistic routines. If the QWERTY<br />

keyset casts the computer as machine, perhaps it's difficult <strong>to</strong> argue that the chord keyset<br />

does otherwise. The latter requires complicated and coordinated movements and an<br />

internalized understanding of an abstract code. Yet Engelbart, seeking augmentation<br />

rather than au<strong>to</strong>mation, argued that this degree of training is necessary before the<br />

computer may function as a <strong>to</strong>ol. Engelbart's notions of training, and the boundaries<br />

between the training of man and the training of the computer, deserve further exploration.<br />

Returning <strong>to</strong> Marx and Mumford, we must work <strong>to</strong> cast the computer as <strong>to</strong>ol rather than<br />

as machine.<br />

Conclusion<br />

Despite the apparent contradictions in Engelbart's efforts <strong>to</strong> build input devices that<br />

would augment the human intellect, he and Licklider provide inspiration for, if not<br />

demonstration of, what the computer might someday be. Engelbart in his 1961 report<br />

"Program on Human Effectiveness" describes his plan for the augmentation of the<br />

individual problem solver. His objective is <strong>to</strong> "develop means for making humans<br />

maximally effective as comprehending solvers of problems", and he plans <strong>to</strong> do this by<br />

inserting computers in<strong>to</strong> various levels of our hierarchy of problem-solving <strong>to</strong>ols. Here<br />

again he returns <strong>to</strong> issues of training, proposing that humans be conditioned both <strong>to</strong> use<br />

his input devices and <strong>to</strong> understand the logical operations that make the computer work.<br />

Engelbart wants <strong>to</strong> build <strong>to</strong>ols even if he must first build machines; he wants humans <strong>to</strong><br />

train computers, but for now he must settle for computers that can train humans. Today,<br />

children use computers from such a young age that manipulating a mouse seems


"intuitive" and programming languages seem "natural." Half of Engelbart's program has<br />

been accomplished: we have been trained <strong>to</strong> use the computer.<br />

Even in 1960 Licklider was troubled by the extent <strong>to</strong> which scientists had been trained <strong>to</strong><br />

use the computer. Disappointed that the computer was being used primarily <strong>to</strong> solve<br />

"preformulated problems" with "predetermined procedures" – "It is often said that<br />

programming for a computing machine forces one <strong>to</strong> think clearly, that it disciplines the<br />

thought process. If the user can think his problem through in advance, symbiotic<br />

association with a computing machine is not necessary" (1960: 61) – Licklider proposed<br />

that computers be employed <strong>to</strong> ask questions, not just <strong>to</strong> provide answers. Given the<br />

limits <strong>to</strong> technology, the first attempts at symbiotic systems will rely on the human<br />

opera<strong>to</strong>rs <strong>to</strong> set the goals, formulate hypotheses, devise models, and so on (1960: 64).<br />

The opera<strong>to</strong>rs will evaluate the output, judge the contribution of the machine, and fill in<br />

the gaps when the machine lacks the proper routine for handling some problem. Although<br />

Licklider imagines much more, his first iteration of the symbiotic system is an advanced<br />

information-handling machine. Consistent with his colleague, Engelbart also framed the<br />

problem of symbiosis as one of information management: a primary objective for his H-<br />

LAM/T project was "<strong>to</strong> find the fac<strong>to</strong>rs that limit the effectiveness of the individual's<br />

basic information-handling capabilities in meeting the various needs of society for<br />

problem solving in the most general sense" (1963: 73). In practice, the symbiotic machine<br />

became a problem-solving rather than a problem-posing device. For the most part, that is<br />

how the computer continues <strong>to</strong> function. Licklider's dream remains largely unfulfilled.<br />

Perhaps transforming the computer from machine <strong>to</strong> <strong>to</strong>ol, from a device that au<strong>to</strong>mates<br />

mundane mental tasks <strong>to</strong> one that augments critical and creative thought, is the task now<br />

facing computing humanists.<br />

See also chapter 29: Speculative Computing.<br />

References for Further Reading<br />

Aspray, William (1990). John von Neumann and the Origins of Modern Computing.<br />

Cambridge, MA: MIT Press.<br />

Bliven, Bruce (1954). The Wonderful Writing Machine. New York: Random House.<br />

Burks, Arthur, Herman H. Goldstine, and John von Neumann (1947). Preliminary<br />

Discussion of the Logical Design of an Electronic Computing Instrument, 2nd edn. In<br />

William Aspray and Arthur Burks (eds.), The Papers of John Von Neumann on<br />

Computing and Computer Theory (1986). Charles Babbage Institute Reprint Series for<br />

the His<strong>to</strong>ry of Computing, vol. 12 (pp. 97–144). Cambridge, MA: MIT Press.<br />

Campbell-Kelly, Martin, and William Aspray (1996). Computer: A His<strong>to</strong>ry of the<br />

Information Machine. New York: Basic Books.<br />

David, Paul A. (1985). Clio and the Economics of QWERTY. American Economic Review<br />

75, 2: 332–7.


Dohrn-van Rossum, Gerhard (1996). His<strong>to</strong>ry of the Hour: Clocks and Modern Temporal<br />

Orders, tr. Thomas Dunlap. Chicago: University of Chicago Press.<br />

Douwe, Draaisma (2000). Metaphors of Memory, tr. Paul Vincent. Cambridge:<br />

Cambridge University Press.<br />

Engelbart, Douglas C. (1961). Proposal for Participation in the Program on Human<br />

Effectiveness. Accessed Oc<strong>to</strong>ber 1, 2002. At<br />

http://sloan.stanford.edu/mousesite/EngelbartPapers/B4_F12_HuEff3.html.<br />

Engelbart, Douglas C. (1963). A Conceptual Framework for the Augmentation of Man's<br />

Intellect. In P. W. Hower<strong>to</strong>n and D. C. Weeks (eds.), The Augmentation of Man's<br />

Intellect by Machine. In Paul A. Mayer (ed.), Computer Media and Communication<br />

(1999). New York: Oxford University Press.<br />

Engelbart, Douglas C. (1968). Demo of the Augment Research Center. Fall Joint<br />

Computer Conference. Accessed Oc<strong>to</strong>ber 1, 2002. At<br />

http://sloan.stanford.edu/mousesite/1968Demo.html.<br />

Englander, Irv (1996). The Architecture of Computer Hardware and Systems Software.<br />

New York: John Wiley.<br />

Flanagan, George A. (1938). A Treatise on Ornamental Typewriting. New York: Gregg<br />

Publishing.<br />

HowStuffWorks (2002). http://www.howstuffworks.com, by Marshall Brain.<br />

HowStuffWorks, Inc..<br />

Kozierok, Charles M. (<strong>2001</strong>). The PC Guide. At www.pcguide.com.<br />

Licklider, J. C. R. (1960). Man-Computer Symbiosis: IRE Transactions on Human<br />

Fac<strong>to</strong>rs in Electronics. In Paul A. Mayer (ed.), Computer Media and Communication<br />

(1999). New York: Oxford University Press.<br />

Liebowitz, S. J. and Stephen E. Margolis (1990). The Fable of the Keys. Journal of Law<br />

and Economics 33, 1: 1–26.<br />

Locke, Chris (2000). Digital Memory and the Problem of Forgetting. In Susannah<br />

Rads<strong>to</strong>ne (ed.), Memory and Methodology (pp. 25–36). New York: Berg.<br />

McCullough, Warren and Walter Pitts (1943). A Logical Calculus of the Ideas Immanent<br />

in Nervous Activity. Bulletin of Mathematical Biophysics, 5: 115–33.<br />

Macey, Samuel L. (1989). The Dynamics of Progress: Time, Method, and Measure.<br />

Athens, GA: University of Georgia Press.


Marx, Karl (1867). Capital, Volume One. In Robert C. Tucker (ed.), The Marx-Engels<br />

Reader, 2nd edn. (1978). New York: W W Nor<strong>to</strong>n.<br />

Mumford, Lewis (1934). Technics and Civilization. New York: Harcourt, Brace and<br />

Company.<br />

Norman, Don (1990). The Design of Everyday Things. New York: Doubleday.<br />

OED (1989). "clock, n." Oxford English Dictionary, ed. J. A. Simpson and E. S. C.<br />

Weiner, 2nd edn. Oxford: Clarendon Press. Electronic Text Center, University of<br />

Virginia. Accessed Oc<strong>to</strong>ber 1, 2002. At http://etext.virginia.edu/oed.html.<br />

Petzold, Charles (2000). Code. Redmond, WA: Microsoft Press.<br />

Rheingold, Howard (2000). Tools for Thought. Cambridge, MA: MIT Press.<br />

von Neumann, John (1945). First Draft of a Report on the EDVAC. In William Aspray<br />

and Arthur Burks (eds.), The Papers of John Von Neumann on Computing and Computer<br />

Theory (1986). Charles Babbage Institute Reprint Series for the His<strong>to</strong>ry of Computing,<br />

vol. 12 (pp. 17–82). Cambridge, MA: MIT Press.<br />

White, Ron (<strong>2001</strong>). How Computers Work, 6th edn. Indianapolis, IN: Que Publishers.<br />

Williams, Michael R. (1997). A His<strong>to</strong>ry of Computing Technology, 2nd edn. Los<br />

Alami<strong>to</strong>s, CA: IEEE Computer Society Press.<br />

14.<br />

Classification and its Structures<br />

C. M. Sperberg-McQueen<br />

Classification is, strictly speaking, the assignment of some thing <strong>to</strong> a class; more<br />

generally, it is the grouping <strong>to</strong>gether of objects in<strong>to</strong> classes. A class, in turn, is a<br />

collection (formally, a set) of objects which share some property.<br />

For example, a his<strong>to</strong>rian preparing an analysis of demographic data transcribed from<br />

census books, parish records, and city direc<strong>to</strong>ries might classify individuals by sex, age,<br />

occupation, and place of birth. Places of birth might in turn be classified as large or small<br />

cities, <strong>to</strong>wns, villages, or rural parishes. A linguist studying a text might classify each<br />

running word of text according <strong>to</strong> its part of speech, or each sentence according <strong>to</strong> its<br />

structure. Linguists, literary scholars, or social scientists might classify words occurring<br />

in a text by semantic category, organizing them in<strong>to</strong> semantic nets. Classification serves<br />

two purposes, each important: by grouping <strong>to</strong>gether objects which share properties, it<br />

brings like objects <strong>to</strong>gether in<strong>to</strong> a class; by separating objects with unlike properties in<strong>to</strong>


separate classes, it distinguishes between things which are different in ways relevant <strong>to</strong><br />

the purpose of the classification. The classification scheme itself, by identifying<br />

properties relevant for such judgments of similarity and dissimilarity, can make explicit a<br />

particular view concerning the nature of the objects being classified.<br />

Scope<br />

Since a classification may be based on any set of properties that can be attributed <strong>to</strong> the<br />

objects being classified, classification in the broad sense involves the correct<br />

identification of the properties of the objects of study and is hardly distinguishable from<br />

coherent discourse in general. (Like coherence, systematic classification is sometimes<br />

eschewed for aesthetic or ideological reasons and if overrated may descend in<strong>to</strong><br />

pedantry.) Information retrieval systems may be regarded, and are often described, as<br />

classifying records in<strong>to</strong> the classes "relevant" and "not relevant" each time a user issues a<br />

query. Norms and standards like the XML 1.0 specification or Unicode may be<br />

unders<strong>to</strong>od as classification schemes which assign any data stream or program either <strong>to</strong><br />

the class "conforming" or <strong>to</strong> the class "non-conforming." Laws may be interpreted as<br />

classifying acts as legal or illegal, censors as classifying books, records, performances,<br />

and so on. Any characteristic of any kind of thing, using any set of concepts, may be<br />

viewed as classifying things of that kind in<strong>to</strong> classes corresponding <strong>to</strong> those concepts. In<br />

the extreme case, the property associated with a class may be vacuous: the members may<br />

share only the property of membership in the class. In general, classification schemes are<br />

felt more useful if the classes are organized around properties relevant <strong>to</strong> the purpose of<br />

the classification. Details of the concepts, categories, and mechanisms used in various<br />

acts of classification may be found in other chapters in this volume: see, for example,<br />

chapters 1, 7, 15, and 17.<br />

In the narrower sense, for computer applications in the <strong>humanities</strong> classification most<br />

often involves either the application of pre-existing classification schemes <strong>to</strong>, or the post<br />

hoc identification of clusters among a sample of, for example, texts (e.g. <strong>to</strong> describe the<br />

samples in language corpora), parts of texts (e.g., <strong>to</strong> mark up the structural constituents or<br />

non-structural features of the text), bibliography entries (for subject description in<br />

enumerative bibliographies or specialized libraries), words (for semantic characterization<br />

of texts), or extra-textual events or individuals (e.g., for his<strong>to</strong>rical work). The best known<br />

of these among the readers of this work are perhaps the classification systems used in<br />

libraries and bibliographies for classifying books and articles by subject; in what follows,<br />

examples drawn from these will be used where possible <strong>to</strong> illustrate important points, but<br />

the points are by no means relevant only <strong>to</strong> subject classification.<br />

Since classification relies on identifying properties of the object being classified, perfect<br />

classification would require, and a perfect classification scheme would exhibit, perfect<br />

knowledge of the object. Because a perfect subject classification, for example, locates<br />

each <strong>to</strong>pic in a field in an n-dimensional space near other related <strong>to</strong>pics and distant from<br />

unrelated <strong>to</strong>pics, a perfect subject classification represents a perfect map of the<br />

intellectual terrain covered in the area being classified. For this reason, classification<br />

schemes can carry a great deal of purely theoretical interest, in addition <strong>to</strong> their practical


utility. Classification schemes necessarily involve some theory of the objects being<br />

classified, if only in asserting that the objects possess certain properties. Every on<strong>to</strong>logy<br />

can be interpreted as providing the basis for a classification of the entities it describes.<br />

And conversely, every classification scheme can be interpreted with more or less ease, as<br />

the expression of a particular on<strong>to</strong>logy. In practice, most classification schemes intended<br />

for general use content themselves with representing something less than a perfect image<br />

of the intellectual structure of their subject area and attempt with varying success <strong>to</strong> limit<br />

their theoretical assumptions <strong>to</strong> those most expected users can be expected <strong>to</strong> assent <strong>to</strong>.<br />

At the extreme, the assumptions underlying a classification scheme may become<br />

effectively invisible and thus no longer subject <strong>to</strong> challenge or rethinking; for purposes of<br />

scholarly work, such invisibility is dangerous and should be avoided.<br />

This chapter first describes the abstract structures most often used in classification, and<br />

describes the rules most often thought <strong>to</strong> encourage useful classification schemes. It then<br />

gives a purely formal account of classification in terms of set theory, in order <strong>to</strong> establish<br />

that no single classification scheme can be exhaustive, and indeed that there are infinitely<br />

more ways of classifying objects than can be described in any language. Finally, it turns<br />

<strong>to</strong> various practical questions involved in the development and use of classification<br />

systems.<br />

One-dimensional Classifications<br />

Very simple classification schemes (sometimes referred <strong>to</strong> as nominal classifications,<br />

because the class labels used are typically nouns or adjectives) consist simply of a set of<br />

categories: male and female; French, German, English, and other; noun, verb, article,<br />

adjective, adverb, etc. In cases like these, some characteristic of the object classified may<br />

take any one of a number of discrete values; formally, the property associated with the<br />

class is that of having some one particular value for the given characteristic. The different<br />

classes in the scheme are not ordered with respect <strong>to</strong> each other; they are merely discrete<br />

classes which, taken <strong>to</strong>gether, subdivide the set of things being classified.<br />

In some classifications (sometimes termed ordinal), the classes used fall in<strong>to</strong> some sort of<br />

sequencing or ordering with respect <strong>to</strong> each other: first-year, second-year, third-year<br />

student; folio, quar<strong>to</strong>, octavo, duodecimo; upper-class, middle-class, lower-class.<br />

In still other cases, the underlying characteristic may take a large or even infinite number<br />

of values, which have definite quantitative relations <strong>to</strong> each other: age, height, number of<br />

seats in parliament, number of pages, price, etc. For analytic purposes, it may be<br />

convenient or necessary <strong>to</strong> clump (or aggregate) sets of distinct values in<strong>to</strong> single classes,<br />

as when age given in years is reduced <strong>to</strong> the categories infant, child, adult, or <strong>to</strong> under-18,<br />

18–25, 25–35, over-35.<br />

All of the cases described so far classify objects based on the value of a single<br />

characteristic attributed <strong>to</strong> the object. In the ideal case, the characteristic can be readily<br />

and reliably evaluated, and the values it can take are discrete. The more borderline cases<br />

there are, the harder it is likely <strong>to</strong> be <strong>to</strong> apply the classification scheme, and the more


information is likely <strong>to</strong> be lost by analyses which rely on the classified data rather than<br />

the original data.<br />

Classification Schemes as N-dimensional Spaces<br />

In less simple classification schemes, multiple characteristics may be appealed <strong>to</strong>. These<br />

may often be described as involving a hierarchy of increasingly fine distinctions. The<br />

Dewey Decimal Classification, for example, assigns class numbers in the 800s <strong>to</strong> literary<br />

works. Within the 800s, it assigns numbers in the 820s <strong>to</strong> English literature, in the 830s <strong>to</strong><br />

German literature, the 840s <strong>to</strong> French, etc. Within the 820s, the number 821 denotes<br />

English poetry, 822 English drama, 823 English fiction, and so on. Further digits after the<br />

third make even finer distinctions; as a whole, then, the classification scheme may be<br />

regarded as presenting the classifier and the user with a tree-like hierarchy of classes and<br />

subclasses, with smaller classes branching off from larger ones.<br />

In the case of the Dewey classification of literature, however, the second and third digits<br />

are (almost) wholly independent of each other: a third digit 3 denotes fiction whether the<br />

second digit is 1 (American), 2 (English), 3 (German), 4 (French), 5 (Italian), 6 (Spanish),<br />

7 (Latin), or 8 (Classical Greek), and 2 as a third digit similarly denotes drama,<br />

independent of language.<br />

We can imagine the literature classification of the Dewey system as describing a plane,<br />

with the second digit of the Dewey number denoting positions on the x axis, and the third<br />

digit denoting values along the y axis. Note that neither the sequence and values of the<br />

genre numbers, nor those of the language numbers, have any quantitative significance,<br />

although the sequence of values is in fact carefully chosen.<br />

Generalizing this idea, classification schemes are often regarded as identifying locations<br />

in an n-dimensional space. Each dimension is associated with an axis, and the set of<br />

possible values along any one axis is sometimes referred <strong>to</strong> as an array. Many salient<br />

characteristics of classification schemes may be described in terms of this n-dimensional<br />

spatial model.<br />

It should be noted that, unlike the dimensions of a Cartesian space, the different<br />

characteristics appealed <strong>to</strong> in a classification scheme are not always wholly independent<br />

of each other. A medical classification, for example, may well subdivide illnesses or<br />

treatments both by the organ or biological system involved and by the age, sex, or other<br />

salient properties of the patient. Since some illnesses afflict only certain age groups or<br />

one sex or the other, the two axes are not wholly independent. A classification of dialects<br />

based on the pronunciation of a given lexical item can only apply <strong>to</strong> dialects in which that<br />

lexical item exists. A facet in a social classification that distinguishes hereditary from<br />

non-hereditary titles is only relevant <strong>to</strong> that part of the population which bears titles, and<br />

only in countries with a nobility. The digits 2 for drama and 3 for fiction have these<br />

meanings in the Dewey classification for literature, but they are not applicable <strong>to</strong> the 900s<br />

(his<strong>to</strong>ry) or the 100s (philosophy). And so on.


The idea of a classification as describing an n-dimensional Cartesian space is thus in<br />

many cases a dramatic simplification. It is nonetheless convenient <strong>to</strong> describe each<br />

characteristic or property appealed <strong>to</strong> in a classification as determining a position along<br />

an axis, even if that axis has no meaning for many classes in the scheme. Those offended<br />

by this inexactitude in the metaphor may amuse themselves by thinking of the logical<br />

space defined by such a classification not as a Cartesian or New<strong>to</strong>nian one but as a<br />

relativistic space with a non-Euclidean geometry.<br />

Some Distinctions among Classification Schemes<br />

When the axes of the logical space are explicitly identified in the description of the<br />

classification scheme, the scheme is commonly referred <strong>to</strong> as a faceted classification, and<br />

each axis (or the representation of a given class's value along a specific axis) as a facet.<br />

The concept of facets in classification schemes was first systematized by Ranganathan,<br />

though the basic phenomena are visible in earlier systems, as the example from the<br />

Dewey classification given above illustrates.<br />

Faceted schemes are typically contrasted with enumerative schemes, in which all classes<br />

in the system are exhaustively enumerated in the classification handbook or schedule. In<br />

a typical faceted scheme, a separate schedule is provided for each facet and the facets are<br />

combined by the classifier according <strong>to</strong> specified rules; because the classifier must create<br />

or synthesize the class number, rather than looking it up in an enumeration, faceted<br />

schemes are sometimes also called synthetic (or, <strong>to</strong> emphasize that the task of synthesis<br />

must be preceded by analysis of the relevant properties of the object, analytico-synthetic)<br />

schemes. Both because of their intellectual clarity and because they can readily exploit<br />

the strengths of electronic database management systems, faceted classification schemes<br />

have become increasingly popular in recent years.<br />

Some classification schemes provide single expressions denoting regions of the logical<br />

space; in what follows these are referred <strong>to</strong> as (class) formulas. Formulas are convenient<br />

when the objects classified must be listed in a single one-dimensional list, as on the<br />

shelves of a library or the pages of a classified bibliography. In such schemes, the order<br />

in which axes are represented may take on great importance, and a great deal of ingenuity<br />

can be devoted <strong>to</strong> deciding whether a classification scheme ought <strong>to</strong> arrange items first by<br />

language, then by genre, and finally by period, or in some other order.<br />

In computerized systems, however, particularly those using database management<br />

systems, it is normally easier <strong>to</strong> vary the order of axes and often unnecessary <strong>to</strong> list every<br />

object in the collection in a single sequence, and so the order of axes has tended <strong>to</strong><br />

become somewhat less important in multidimensional classification schemes intended for<br />

computer use. The provision of unique class formulas for each point in the scheme's<br />

logical space has correspondingly declined in importance, and much of the discussion of<br />

notation in pre-electronic literature on classification has taken on an increasingly quaint<br />

air. For those who need <strong>to</strong> devise compact symbolic formulas for the classes of a scheme,<br />

the discussions of notation in Ranganathan's Prolegomena (1967) are strongly<br />

recommended.


When each axis of the logical space can be associated with a particular part of the<br />

formula denoting a class, and vice versa, the notation is expressive (as in the portion of<br />

the Dewey system mentioned above). Fully expressive notations tend <strong>to</strong> be longer than<br />

would otherwise be necessary, so some classification schemes intentionally use<br />

inexpressive or incompletely expressive notation, as in most parts of the Library of<br />

Congress classification system. Expressive notations are advantageous in computer-based<br />

applications, since they make it easy <strong>to</strong> perform searches in the logical space by means of<br />

searches against class symbols. A search for dramas in any language, for example, can be<br />

performed by searching for items with a Dewey class number matching the regular<br />

expression "8.2." No similarly simple search is possible in inexpressive notations.<br />

Some classification systems describe classes using natural-language phrases, rather than<br />

by assigning them <strong>to</strong> specific locations in a class hierarchy; library subject headings are a<br />

well-known example, but there are many others. (Some classification theorists distinguish<br />

such alphabetical systems as indexing systems, as opposed <strong>to</strong> classification systems in the<br />

strict sense, restricting the latter term <strong>to</strong> systems that provide a formal notation other than<br />

natural language for their class formulas.) Typically, such systems arrange <strong>to</strong>pics in<br />

alphabetical order, rather than a systematic order imposed by the structure of the<br />

classification scheme. At one extreme, such a system may use free-form textual<br />

descriptions of objects <strong>to</strong> "classify" them. Most alphabetically organized classification<br />

systems, however, differ from wholly free-form indices in one or more ways. First, in<br />

order <strong>to</strong> avoid or minimize the inconsistencies caused by the use of different but<br />

synonymous descriptions, such systems normally use controlled vocabularies rather than<br />

unconstrained natural-language prose: descrip<strong>to</strong>rs other than proper nouns must be<br />

chosen from a closed list. In the ideal case, the controlled vocabulary has exactly one<br />

representative from any set of synonyms in the scope of the classification scheme.<br />

Second, as part of the vocabulary control alphabetic systems often stipulate that certain<br />

kinds of phrases should be "inverted", so that the alphabetical listing will place them near<br />

other entries. In some schemes, particular types of descrip<strong>to</strong>rs may be subdivided by<br />

other descrip<strong>to</strong>rs in a hierarchical fashion. Thus the Library of Congress subject heading<br />

for Beowulf will be followed by "Beowulf - Adaptations", "Beowulf - Bibliography",<br />

"Beowulf - Criticism, textual", "Beowulf - Study and teaching", "Beowulf - Translations –<br />

Bibliographies", "Beowulf- Translations – His<strong>to</strong>ry and criticism", and so on. The phrases<br />

after the dashes are, in effect, an array of possible subdivisions for anonymous literary<br />

works; the Library of Congress Subject Headings (LCSH) provide a prescribed set of<br />

such expansions for a variety of different kinds of object: anonymous literary works,<br />

individuals of various kinds, theological <strong>to</strong>pics, legislative bodies, sports, industries,<br />

chemicals, and so on. Third, most systems which use controlled vocabularies also provide<br />

a more or less systematic set of cross-references among terms. At a minimum, these<br />

cross-references will include see references from unused terms <strong>to</strong> preferred synonyms. In<br />

more elaborate cases, see-also references will be provided <strong>to</strong> broader terms, narrower<br />

terms, coordinate terms (i.e., other terms with the same broader term), partial synonyms,<br />

genus/species terms, and so on. The links <strong>to</strong> broader and narrower terms allow the<br />

alphabetically arranged scheme <strong>to</strong> provide at least some of the same information as a<br />

strictly hierarchical scheme. Like the LCSH, the New York Times Thesaurus of<br />

Descrip<strong>to</strong>rs (1983) described by Mills provides a useful model for work of this kind.


The fineness of distinction carried by the classification – that is, the size of the regions in<br />

the logical space that the classification allows us <strong>to</strong> distinguish – is called (mixing<br />

metaphors) the depth of the classification scheme. Some classification schemes provide a<br />

fixed and unvarying depth; others allow variable depth. Depth may be added either by<br />

adding more axes <strong>to</strong> the classification, as when a library using the Dewey system<br />

subdivides 822 (English drama) by period, or by adding more detail <strong>to</strong> the specification<br />

of the value along an axis already present. Faceted classification schemes often allow<br />

facets <strong>to</strong> vary in length, so as <strong>to</strong> allow the depth of classification <strong>to</strong> be increased by<br />

providing a more precise value for any facet. Notations with fixed-length facets, by<br />

contrast, like the part of Dewey described above, cannot increase the specificity of facets<br />

other than the last without creating ambiguity.<br />

Whether they use expressive notation or not, some classification schemes provide<br />

notations for each node in their hierarchy (e.g., one formula for "literature" and another<br />

for "English literature", and so on); in such cases, the categories of the classification are<br />

not, strictly speaking, disjoint: the broader classes necessarily subsume the narrower<br />

classes arranged below them. One advantage of expressive notation is that it makes this<br />

relationship explicit. Other schemes provide notations only for the most fully specified<br />

nodes of the hierarchy: the hierarchical arrangement may be made explicit in the<br />

description of the scheme, but is collapsed in the definition of the notation, so that the<br />

classification gives the impression of providing only a single array of values. Commonly<br />

used part-of-speech classification systems often collapse their hierarchies in this way:<br />

each tag used <strong>to</strong> denote word-class and morphological information denotes a complete<br />

packet of such information; there is no notation for referring <strong>to</strong> more general classes like<br />

"noun, without regard for its specific morphology." Markup languages similarly often<br />

provide names only for the "leaves" of their tree-like hierarchies of element types; even<br />

when a hierarchy of classes is an explicit part of the design, as in the Text Encoding<br />

Initiative (TEI), there may be no element types which correspond directly <strong>to</strong> classes in<br />

the hierarchy.<br />

When combinations of terms from different axes are specified in advance, as part of the<br />

process of classifying or indexing an object, we speak of a pre-coordinate system. When<br />

a classification system limits itself <strong>to</strong> identifying the appropriate values along the various<br />

axes, and values may be combined at will during a search of the classification scheme,<br />

we speak of a post-coordinate system. Printed indices that list all the subject descrip<strong>to</strong>rs<br />

applied <strong>to</strong> the items in a bibliography, in a fixed order of axes, for example, present a<br />

kind of pre-coordinate classification scheme. Online indices that allow searches <strong>to</strong> be<br />

conducted along arbitrary combinations of axes, by contrast, provide a post-coordinate<br />

scheme. It is possible for printed indices <strong>to</strong> provide free combination of terms, but postcoordinate<br />

indexing is easier for computer systems. Post-coordinate indexing allows<br />

greater flexibility and places greater demands on the intelligence of the user of the index.<br />

When the axes and the values along each axis are specified in advance, and items are<br />

classified in terms of them, we can speak of an a priori system. When the axes and their<br />

values are derived post hoc from the items encountered in the collection of objects being<br />

classified, we may speak of an a posteriori or data-driven system. Author-specified


keywords and free-text searching are simple examples of data-driven classification.<br />

Citation analysis, and in particular the study of co-citation patterns in scholarly literature,<br />

as described by Garfield, is another.<br />

In some cases, the identification of axes in a data-driven system may involve<br />

sophisticated and expensive statistical analysis of data. The technique of latent semantic<br />

analysis is an example: initially, the occurrence or non-occurrence of each word in the<br />

vocabulary of all the documents in the collection being indexed is treated as an axis, and<br />

a statistical analysis is performed <strong>to</strong> collapse as many of these axes <strong>to</strong>gether as possible<br />

and identify a useful set of axes which are as nearly orthogonal <strong>to</strong> each other as the data<br />

allow. In a typical application, latent-semantic analysis will identify documents in a space<br />

of 200 or so dimensions. It is sometimes possible <strong>to</strong> examine the dimensions and<br />

associate meaning with them individually, but for the most part data-driven statistical<br />

methods do not attempt <strong>to</strong> interpret the different axes of their space individually. Instead,<br />

they rely on conventional measures of distance in n-dimensional spaces <strong>to</strong> identify items<br />

which are near each other; when the classification has been successful, items which are<br />

near each other are similar in ways useful for the application, and items which are distant<br />

from each other are dissimilar.<br />

A priori systems may also be interpreted as providing some measure of similarity among<br />

items, but it is seldom given a numerical value.<br />

Unconscious, naive, or pre-theoretic classification (as seen, for example, in naturallanguage<br />

terminology for colors) may be regarded as intermediate between the a priori<br />

and a posteriori types of classification systems described above.<br />

Some data-driven systems work by being given samples of pre-classified training<br />

material and inducing some scheme of properties which enables them <strong>to</strong> match, more or<br />

less well, the classifications given for the training material. Other data-driven systems<br />

work without overt supervision, inducing classifications based solely on the observed<br />

data.<br />

A priori systems require more effort in advance than data-driven systems, both in the<br />

definition of the classification scheme and in its application by skilled classifiers. The<br />

costs of data-driven systems are concentrated later in the his<strong>to</strong>ry of the classification<br />

effort, and tend <strong>to</strong> involve less human effort and more strictly computational effort. Datadriven<br />

classification schemes may also appeal <strong>to</strong> scholars because they are free of many<br />

of the obvious opportunities for bias exhibited by a priori schemes and thus appear more<br />

nearly theory-neutral. It must be stressed, therefore, that while the theoretical<br />

assumptions of data-driven systems may be less obvious and less accessible <strong>to</strong> inspection<br />

by those without a deep knowledge of statistical techniques, they are nonetheless<br />

necessarily present.<br />

Rules for Classification


Some principles for constructing classification schemes have evolved over the centuries;<br />

they are not always followed, but are generally <strong>to</strong> be recommended as leading <strong>to</strong> more<br />

useful classification schemes.<br />

The first of these is <strong>to</strong> avoid cross-classification: a one-dimensional classification should<br />

normally depend on the value of a single characteristic of the object classified, should<br />

provide for discrete (non-overlapping) values, and should allow for all values which will<br />

be encountered: perhaps the best-known illustration of this rule lies in its violation in the<br />

fictional Chinese encyclopedia imagined by Jorge Luis Borges, in which<br />

it is written that animals are divided in<strong>to</strong>: (a) those that belong <strong>to</strong> the Emperor, (b)<br />

embalmed ones, (c) those that are trained, (d) suckling pigs, (e) mermaids, (f) fabulous<br />

ones, (g) stray dogs, (h) those that are included in this classification, (i) those that tremble<br />

as if they were mad, (j) innumerable ones, (k) those drawn with a very fine camel's-hair<br />

brush, (1) others, (m) those that have just broken a flower vase, (n) those that resemble<br />

flies from a distance.<br />

(Borges 1981)<br />

One apparent exception <strong>to</strong> this rule is often found in schemes which seek <strong>to</strong> minimize the<br />

length of their class formulas: often two characteristics are collapsed in<strong>to</strong> a single step in<br />

the classification hierarchy, as when a demographic classification has the classes infant<br />

(sex unspecified), infant male, infant female, child (sex unspecified), boy, girl, adult (sex<br />

unspecified), man, woman.<br />

Other desirable attributes of a classification scheme may be summarized briefly (I<br />

abbreviate here the "canons" defined by Ranganathan). Each characteristic used as the<br />

basis for an axis in the logical space should:<br />

1 distinguish some objects from others: that is, it should give rise <strong>to</strong> at least two<br />

subclasses;<br />

2 be relevant <strong>to</strong> the purpose of the classification scheme (every classification scheme has<br />

a purpose; no scheme can be unders<strong>to</strong>od fully without reference <strong>to</strong> that purpose);<br />

3 be definite and ascertainable; this means that a classification scheme cannot be<br />

successfully designed or deployed without taking in<strong>to</strong> account the conditions under<br />

which the work of classification is <strong>to</strong> be performed;<br />

4 be permanent, so as <strong>to</strong> avoid the need for constant reclassification;<br />

5 have an enumerable list of possible values which exhausts all possibilities. Provision<br />

should normally be made for cases where the value is not ascertainable after all: it is<br />

often wise <strong>to</strong> allow values like unknown or not specified. In many cases several distinct<br />

special values are needed; among those sometimes used are: unknown (but applicable),<br />

does-not-apply, any (data compatible with all possible values for the field), approximate


(estimated with some degree of imprecision), disputed, uncertain (classifier is not certain<br />

whether this axis is applicable; if it is applicable, the value is unknown).<br />

In classification schemes which provide explicit class symbols, it is useful <strong>to</strong> provide a<br />

consistent sequence of axes in the construction of the class symbol (if the subject<br />

classification for literature divides first by country or language and then by period, it is<br />

probably wise for the subject classification for his<strong>to</strong>ry <strong>to</strong> divide first by country and then<br />

by period, rather than vice versa). The sequence of values within an array of values for a<br />

given axis should also be made helpful, and consistent in different applications. Patterns<br />

often suggested include arranging the sequence for increasing concreteness, increasing<br />

artificiality, increasing complexity, increasing quantity, chronological sequence,<br />

arrangement by spatial contiguity, arrangement from bot<strong>to</strong>m up, left-<strong>to</strong>-right<br />

arrangement, clockwise sequence, arrangement following a traditional canonical<br />

sequence, arrangement by frequency of values (in bibliographic contexts this is called<br />

literary warrant), or as a last resort alphabetical sequence.<br />

Many classification schemes appeal, at some point, <strong>to</strong> one of a number of common<br />

characteristics in order <strong>to</strong> subdivide a class which otherwise threatens <strong>to</strong> become <strong>to</strong>o<br />

large (in bibliographic practice, it is often advised <strong>to</strong> subdivide a class if it would<br />

otherwise contain more than twenty items). Subdivision by chronology, by geographic<br />

location, or by alphabetization are all commonly used; standard schedules for subdivision<br />

on chronological, geographic, linguistic, genre, and other grounds can be found in<br />

standard classification schemes and can usefully be studied, or adopted wholesale, in the<br />

creation of new schemes.<br />

Classification schemes intended for use by others do well <strong>to</strong> allow for variation in the<br />

depth of classification practiced. Library classification schemes often achieve this by<br />

allowing class numbers <strong>to</strong> be truncated (for coarser classification) or extended (for finer);<br />

markup languages may allow for variable depth of markup by making some markup<br />

optional and by providing element types of varying degrees of specificity.<br />

It is also desirable, in schemes intended for general use, <strong>to</strong> provide for semantic extension<br />

and the addition of new concepts; this is not always easy. Library classification schemes<br />

often attempt <strong>to</strong> achieve this by providing standard schedules for subdividing classes by<br />

chronology, geographic distribution, and so on, <strong>to</strong> be applied according <strong>to</strong> the judgment<br />

of the classifier; the Colon Classification goes further by defining an array of abstract<br />

semantic concepts which can be used when subdivision by other standard axes is not<br />

feasible or appropriate. It provides a good illustration of the difficulty of providing useful<br />

guidance in areas not foreseen by the devisers of the classification scheme:<br />

1 unity, God, world, first in evolution or time, one-dimension, line, solid state, …<br />

2 two dimensions, plane, cones, form, structure, ana<strong>to</strong>my, morphology, sources of<br />

knowledge, physiography, constitution, physical anthropology, …


3 three dimensions, space, cubics, analysis, function, physiology, syntax, method, social<br />

anthropology, …<br />

4 heat, pathology, disease, transport, interlinking, synthesis, hybrid, salt, …<br />

5 energy, light, radiation, organic, liquid, water, ocean, foreign land, alien, external,<br />

environment, ecology, public controlled plan, emotion, foliage, aesthetics, woman, sex,<br />

crime, …<br />

6 dimensions, subtle, mysticism, money, finance, abnormal, phylogeny, evolution, …<br />

7 personality, on<strong>to</strong>geny, integrated, holism, value, public finance, …<br />

8 travel, organization, fitness.<br />

In markup languages, semantic extension can take the form of allowing class or type<br />

attributes on elements: for any element type e, an element instance labeled with a class or<br />

type attribute can be regarded as having a specialized meaning. In some markup<br />

languages, elements with extremely general semantics are provided (such as the TEI div,<br />

ab, or seg elements, or the HTML div and span elements), in order <strong>to</strong> allow the greatest<br />

possible flexibility for the use of the specialization attributes.<br />

Any new classification scheme, whether intended for general use or for use only by a<br />

single project, will benefit from clear documentation of its purpose and (as far as they can<br />

be made explicit) its assumptions. For each class in the scheme, the scope of the class<br />

should be clear; sometimes the scope is sufficiently clear from the name, but very often it<br />

is essential <strong>to</strong> provide scope notes describing rules for determining whether objects fall<br />

in<strong>to</strong> the class or not. Experience is the best teacher here; some projects, like many large<br />

libraries, keep master copies of their classification schemes and add annotations or<br />

additional scope notes whenever a doubtful case arises and is resolved.<br />

A Formal View<br />

From a purely formal point of view, classification may be regarded as the partition of<br />

some set of objects (let us call this set0) in<strong>to</strong> some set of classes (let us call this set of<br />

classes C, or the classification scheme).<br />

In simple cases (nominal classifications), the classes of C have no identified relation <strong>to</strong><br />

each other but serve merely as bins in<strong>to</strong> which the objects in 0 are sorted. For any finite 0,<br />

there are a finite number of possible partitions of 0 in<strong>to</strong> non-empty pair-wise disjoint<br />

subsets of 0. As a consequence, there are at most a finite number of extensionally distinct<br />

ways <strong>to</strong> classify any finite set 0 in<strong>to</strong> classes; after that number is reached, any new<br />

classification must reconstitute a grouping already made by some other classification and<br />

thus be extensionally equivalent <strong>to</strong> it. Such extensionally equivalent classifications need<br />

not be intensionally equivalent: if we classify the four letters a, b, l, e according <strong>to</strong> their<br />

phonological values, we might put a and e <strong>to</strong>gether as vowels, and b and l as consonants.


If we classed them according <strong>to</strong> whether their letter forms have ascenders or not, we<br />

would produce the same grouping; the two classifications are thus extensionally<br />

equivalent, though very different in intension. In practice, the extensional equivalence of<br />

two classifications may often suggest some relation among the properties appealed <strong>to</strong>, as<br />

when classifying the syllables of German according <strong>to</strong> their lexicality and according <strong>to</strong><br />

their stress.<br />

In some cases, the classes of C can be related by a proximity measure of some kind. In<br />

such a classification, any two adjacent classes are more similar <strong>to</strong> each other than, say, a<br />

pair of non-adjacent classes. If such a classification scheme relies on a single scalar<br />

property, its classes may be imagined as corresponding <strong>to</strong> positions on, or regions of, a<br />

line. If the classification schema relies on two independent properties, the classes will<br />

correspond <strong>to</strong> points or regions in a plane. In practice, practical classification schemes<br />

often involve arbitrary numbers of independent properties; if n properties are used by a<br />

classification scheme, individual classes may be identified with positions in an ndimensional<br />

space. The rules of Cartesian geometry may then be applied <strong>to</strong> test similarity<br />

between classes; this is simplest if the axes are quantitative, or at least ordered, but<br />

suitably modified distance measures can be used for purely nominal (unordered,<br />

unquantitative) classifications as well: the distance along the axis may be 0, for example,<br />

if two items have the same value for that axis, and 1 otherwise.<br />

If we imagine some finite number of classes, and conceive of a classification scheme as<br />

being defined by some finite-length description (say, in English or any other natural<br />

language) of how <strong>to</strong> apply those classes <strong>to</strong> some infinite set of objects, then it may be<br />

noted that there are an infinite number of possible groupings which will not be generated<br />

by any classification scheme described in our list. The proof is as follows:<br />

1 Let us label the classes with the numbers 1 <strong>to</strong> n, where n is the number of classes.<br />

2 Let us assume that the objects <strong>to</strong> be classified can be placed in some definite order; the<br />

means by which we do this need not concern us here.<br />

3 Then let us place the descriptions of possible classifications also in<strong>to</strong> a definite order; it<br />

is easy <strong>to</strong> see that the list of descriptions is likely <strong>to</strong> be infinite, but we can nevertheless<br />

place them in<strong>to</strong> a definite order. Since we imagine the descriptions as being in English or<br />

some other natural language, we can imagine sorting them first by length and then<br />

alphabetically. In practice, there might be some difficulty deciding whether a given text<br />

in English does or does not count as a description of a classification scheme, but for<br />

purposes of this exercise, we need not concern ourselves with this problem: we can list all<br />

English texts, and indeed all sequences of letters, spaces, and punctuation, in a definite<br />

sequence. (If we cannot interpret the sequence of letters as defining a rule for assigning<br />

objects <strong>to</strong> classes, we can arbitrarily assign every object <strong>to</strong> class 1.)<br />

4 Now let us imagine a table, with one row for each description of a classification scheme<br />

and one column for each object <strong>to</strong> be classified. In the cell corresponding <strong>to</strong> a given


scheme and object, we write the number of the class assigned <strong>to</strong> that object by that<br />

classification scheme. Each row thus describes a grouping of the objects in<strong>to</strong> classes.<br />

5 Now, we describe a grouping of the objects in<strong>to</strong> classes which differs from every<br />

grouping in our list:<br />

(a) Starting in the first row and the first column, we examine the number written there. If<br />

that number is less than n, we add one <strong>to</strong> it; if it is equal <strong>to</strong> n, we subtract n - 1 from it.<br />

(b) Next, we go <strong>to</strong> the next row and the next column, and perform the same operation.<br />

(c) We thus describe a diagonal sequence of cells in the table, and for each column we<br />

specify a class number different from the one written there. The result is that we have<br />

assigned each object <strong>to</strong> a class, but the resulting grouping does not correspond <strong>to</strong> any<br />

grouping listed in the table (since it differs from each row in at least one position).<br />

We are forced, then, <strong>to</strong> conclude that even though our list of finite-length descriptions of<br />

classification schemes was assumed <strong>to</strong> be infinite, there is at least one assignment of<br />

objects <strong>to</strong> classes that does not correspond <strong>to</strong> any classification scheme in the list. (The<br />

list contains only the schemes with finite-length descriptions, but the classification we<br />

have just described requires an infinitely large table for its description, so it does not<br />

appear in the list.) There are, in fact, not just the one but an infinite number of such<br />

classifications which are not in the list.<br />

Since the list contains, by construction, every classification scheme that has a finitelength<br />

description, we must infer that the classifications described by the diagonal<br />

procedure outlined above do not have any finite-length description; let us call them, for<br />

this reason, ineffable classifications.<br />

The existence of ineffable classifications is not solely of theoretical interest; it may also<br />

serve as a salutary reminder that no single classification scheme can be expected <strong>to</strong> be<br />

"complete" in the sense of capturing every imaginable distinction or common property<br />

attributable <strong>to</strong> the members of 0. A "perfect" classification scheme, in the sense described<br />

above of a scheme that perfectly captures every imaginable similarity among the objects<br />

of 0, is thus a purely imaginary construct; actual classification schemes necessarily<br />

capture only a subset of the imaginable properties of the objects, and we must choose<br />

among them on pragmatic grounds.<br />

Make or Find?<br />

Whenever systematic classification is needed, the researcher may apply an existing<br />

classification scheme or else devise a new scheme for the purpose at hand. Existing<br />

schemes may be better documented and more widely unders<strong>to</strong>od than an ad hoc scheme<br />

would be; in some cases they will have benefited from more sustained attention <strong>to</strong><br />

technical issues in the construction of a scheme than the researcher will be able <strong>to</strong> devote<br />

<strong>to</strong> a problem encountered only incidentally in the course of a larger research project.


Being based on larger bodies of material, they may well provide better coverage of<br />

unusual cases than the researcher would otherwise manage; they may thus be more likely<br />

<strong>to</strong> provide an exhaustive list of possible values for each axis. And the use of a standard<br />

classification scheme does allow more direct comparison with material prepared by<br />

others than would otherwise be possible.<br />

On the other hand, schemes with broad coverage may often provide insufficient depth for<br />

the purposes of specialized research (just as the thousand basic categories of the Dewey<br />

Decimal System will seldom provide a useful framework for a bibliography of secondary<br />

literature on a single major work or author), and the studied theoretical neutrality of<br />

schemes intended for wide use may be uncongenial <strong>to</strong> the purpose of the research.<br />

In the preparation of resources intended for use by others, the use of standard existing<br />

classification schemes should generally be preferred <strong>to</strong> the ad hoc concoction of new<br />

ones. Note that some existing classification schemes are proprietary and may be used in<br />

publicly available material only by license; before using an established classification<br />

scheme, researchers should confirm that their usage is authorized.<br />

For work serving a particular research agenda, no general rule is possible; the closer the<br />

purpose of the classification <strong>to</strong> the central problem of the research, the more likely is a<br />

cus<strong>to</strong>m-made classification scheme <strong>to</strong> be necessary. Researchers should not, however,<br />

underestimate the effort needed <strong>to</strong> devise a coherent scheme for systematic classification<br />

of anything.<br />

Some Existing Classification Schemes<br />

Classification schemes may be needed, and existing schemes may be found, for objects of<br />

virtually any type. Those mentioned here are simply samples of some widely used kinds<br />

of classification: classification of documents by subject or language variety, classification<br />

of words by word class or semantics, classification of extra-textual entities by socioeconomic<br />

and demographic properties, and classification of images.<br />

The best-known subject classification schemes are those used in libraries and in major<br />

periodical bibliographies <strong>to</strong> provide subject access <strong>to</strong> books and articles. The Dewey<br />

Decimal Classification (DDC) and its internationalized cousin the Universal Decimal<br />

Classification (UDC) are both widely used, partly for his<strong>to</strong>rical reasons (the Dewey<br />

system was the first widely promoted library classification scheme), partly owing <strong>to</strong> their<br />

relatively convenient decimal notation, and because their classification schedules are<br />

regularly updated. In the USA, the Library of Congress classification is now more widely<br />

used in research libraries, in part because its notation is slightly more compact than that<br />

of Dewey.<br />

Less widely used, but highly thought of by some, are the Bliss Bibliographic<br />

Classification, originally proposed by Henry Evelyn Bliss and now thoroughly revised,<br />

and the Colon Classification devised by Shiyali Ramamrita Ranganathan, perhaps the<br />

most important theorist of bibliographic classification in his<strong>to</strong>ry (Melvil Dewey is surely


more influential but can hardly be described as a theorist). Both are fully faceted<br />

classification schemes.<br />

The controlled vocabulary of the Library of Congress Subject Headings may also be<br />

useful; its patterns for the subdivision of various kinds of subjects provide useful arrays<br />

for subordinate axes.<br />

Researchers in need of specialized subject classification should also examine the subject<br />

classifications used by major periodical bibliographies in the field; Balay (1996) provides<br />

a useful source for finding such bibliographies.<br />

The crea<strong>to</strong>rs of language corpora often wish <strong>to</strong> classify their texts according <strong>to</strong> genre,<br />

register, and the demographic characteristics of the author or speaker, in order <strong>to</strong><br />

construct a stratified sample of the language varieties being collected and <strong>to</strong> allow users<br />

<strong>to</strong> select subcorpora appropriate for various tasks. No single classification scheme<br />

appears <strong>to</strong> be in general use for this purpose. The schemes used by existing corpora are<br />

documented in their manuals; that used by the Brown and the Lancaster-Oslo/Bergen<br />

(LOB) corpora is in some ways a typical example. As can be seen, it classifies samples<br />

based on a mixture of subject matter, genre, and type of publication:<br />

• A Press: reportage<br />

• B Press: edi<strong>to</strong>rial<br />

• C Press: reviews<br />

• D Religion<br />

• E Skills, trades, and hobbiesz<br />

• F Popular lore<br />

• G Belles lettres, biography, essays<br />

• H Miscellaneous (government documents, foundation reports, industry reports,<br />

college catalogue, industry house organ)<br />

• J Learned and scientific writings<br />

• K General fiction<br />

• L Mystery and detective fiction<br />

• M Science fiction<br />

• N Adventure and western fiction


• P Romance and love s<strong>to</strong>ry<br />

• R Humor<br />

Several recent corpus projects have produced, as a side effect, thoughtful articles on<br />

sampling issues and the classification of texts. Biber (1993) is an example. (See also<br />

chapter 21, this volume.) Some recent corpora, for example the British National Corpus,<br />

have not attempted <strong>to</strong> provide a single text classification in the style of the Brown and<br />

LOB corpora. Instead, they provide descriptions of the salient features of each text,<br />

allowing users <strong>to</strong> select subcorpora by whatever criteria they choose, in a kind of postcoordinate<br />

system.<br />

Some language corpora provide word-by-word annotation of their texts, most usually<br />

providing a single flat classification of words according <strong>to</strong> a mixture of word-class and<br />

inflectional information (plural nouns and singular nouns, for example, thus being<br />

assigned <strong>to</strong> distinct classes). A variety of word-class tagging schemes is in use, but for<br />

English-language corpora the point of reference typically remains the tag set defined by<br />

the Brown Corpus of Modern American English, as refined by the Lancaster-Oslo/Bergen<br />

(LOB) Corpus, and further refined through several generations of the CLAWS<br />

(Constituent Likelihood Au<strong>to</strong>matic Word-tagging System) tagger developed and<br />

maintained at the University of Lancaster (Garside and Smith 1997). When new wordclass<br />

schemes are devised, the detailed documentation of the tagged LOB corpus<br />

(Johansson 1986) can usefully be taken as a model.<br />

Semantic classification of words remains a <strong>to</strong>pic of research; the classifications most<br />

frequently used appear <strong>to</strong> be the venerable work of Roget's Thesaurus and the newer<br />

more computationally oriented work of Miller and colleagues on WordNet (on which see<br />

Fellbaum 1998) and their transla<strong>to</strong>rs, imita<strong>to</strong>rs, and analogues in other languages (on<br />

which see Vossen 1998).<br />

In his<strong>to</strong>rical work, classification is often useful <strong>to</strong> improve the consistency of data and<br />

allow more reliable analysis. When systematic classifications are applied <strong>to</strong> his<strong>to</strong>rical<br />

sources such as manuscript census registers, it is generally desirable <strong>to</strong> retain some<br />

account of the original data, <strong>to</strong> allow consistency checking and later reanalysis (e.g.,<br />

using a different classification scheme). The alternative, pre-coding the information and<br />

recording only the classification assigned, rather than the information as given in the<br />

source, was widely practiced in the early years of computer applications in his<strong>to</strong>ry, since<br />

it provides for more compact data files, but it has fallen out of favor because it makes it<br />

more difficult or impossible for later scholars <strong>to</strong> check the process of classification or <strong>to</strong><br />

propose alternative classifications.<br />

His<strong>to</strong>rians may find the industrial, economic, and demographic classifications of modern<br />

governmental and other organizations useful; even where the classifications cannot be<br />

used unchanged, they may provide useful models. Census bureaus and similar<br />

governmental bodies, and archives of social science data, are good sources of information<br />

about such classification schemes. In the anglophone world, the most prominent social


science data archives may be the Inter-university Consortium for Political and Social<br />

Research (ICPSR) in Ann Arbor aa(http://www.icpsr.umich.edu/) and the UK Data<br />

Archive at the University of Essex (http://www.data-archive.ac.uk/). The Council of<br />

European Social Science Data Archives (http://www.nsd.uib.no/cessda/index.html)<br />

maintains a list of data archives in various countries both inside and outside Europe.<br />

With the increasing emphasis on image-based computing in the <strong>humanities</strong> and the<br />

creation of large electronic archives of images, there appears <strong>to</strong> be great potential utility<br />

in classification schemes for images. If the class formulas of an image classification<br />

scheme are written in conventional characters (as opposed, say, <strong>to</strong> being themselves<br />

thumbnail images), then collections of images can be made accessible <strong>to</strong> search and<br />

retrieval systems by indexing and searching the image classification formulas, and then<br />

providing access <strong>to</strong> the images themselves. Existing image classification schemes<br />

typically work with controlled natural-language vocabularies; some resources use<br />

detailed descriptions of the images in a rather formulaic English designed <strong>to</strong> improve the<br />

consistency of description and make for better retrieval. The Index of Christian Art at<br />

Prince<strong>to</strong>n University (http://www.prince<strong>to</strong>n.edu/~ica/) is an example.<br />

The difficulties of agreeing on and maintaining consistency in keyword-based<br />

classifications or descriptions of images, however, have meant that there is lively interest<br />

in au<strong>to</strong>matic recognition of similarities among graphic images; there is a great deal of<br />

proprietary technology in this area. Insofar as it is used for search and retrieval, image<br />

recognition may be thought of as a specialized form of data-driven classification,<br />

analogous <strong>to</strong> au<strong>to</strong>matic statistically based classification of texts.<br />

References for Further Reading<br />

Anderson, James D. (1979). Contextual Indexing and Faceted Classification for<br />

Databases in the Humanities. In Roy D. Tally and Ronald R. Deultgen (eds.),<br />

Information Choices and Policies: Proceedings of the ASIS Annual Meeting, vol. 16 (pp.<br />

194–201). White Plains, NY: Knowledge Industry Publications.<br />

Balay, Robert, (ed.) (1996). Guide <strong>to</strong> Reference Books, 11th edn. Chicago, London:<br />

American Library Association.<br />

Biber, Douglas (1993). Representativeness in Corpus Design. Literary and Linguistic<br />

Computing 8, 4: 243–57.<br />

Borges, Jorge Luis (1981). The Analytical Language of John Wilkins, tr. Ruth L. C.<br />

Simms. In E. R. Monegal and A. Reid (eds.), Borges: A Reader (pp. 141–3). New York:<br />

Dut<strong>to</strong>n.<br />

Bowker, Geoffrey C. and Susan Leigh Star (1999). Sorting Things Out: Classification<br />

and its Consequences. Cambridge, MA: MIT Press.


Deerwester, Scott et al. (1990). Indexing by Latent Semantic Analysis. Journal of the<br />

American Society for Information Science 41, 6: 391–407.<br />

Fellbaum, Christiane, (ed.) (1998). WordNet: An Electronic Lexical Database.<br />

Cambridge, MA: MIT Press.<br />

Floud, Roderick (1979). An Introduction <strong>to</strong> Quantitative Methods for His<strong>to</strong>rians. London<br />

and New York: Methuen.<br />

Foskett, A. C. (1996). The Subject Approach <strong>to</strong> Information, [n.p.]: Linnet Books and<br />

Clive Bingley, 1969. (4th edn. 1982. 5th edn. London: Library Association).<br />

Garfield, Eugene (1979). Citation Indexing: Its Theory and Application in Science,<br />

Technology, and Humanities. New York: Wiley.<br />

Garside, R. and N. Smith (1997). A Hybrid Grammatical Tagger: CLAWS4. In R.<br />

Garside, G. Leech, and A. McEnery (eds.). Corpus Annotation: Linguistic Information<br />

from Computer Text Corpora (pp. 102–21). London: Longman.<br />

Johansson, Stig, in collaboration with Eric Atwell, Roger Garside, and Geoffrey Leech<br />

(1986). The Tagged LOB Corpus. Bergen: Norwegian Computing Centre for the<br />

Humanities.<br />

Kuhn, Thomas (1977). Second Thoughts on Paradigms. In Frederick Suppe (ed.), The<br />

Structure of Scientific Theories, 2nd edn. (pp. 459–82). Urbana: University of Illinois<br />

Press.<br />

Library of Congress, Cataloging Policy and Support Office (1996). Library of Congress<br />

Subject Headings, 19th edn., 4 vols. Washing<strong>to</strong>n, DC: Library of Congress.<br />

Mills, Harlan (1983). The New York Times Thesaurus of Descrip<strong>to</strong>rs. In Harlan Mills,<br />

Software Productivity (pp. 31–55). Bos<strong>to</strong>n, Toron<strong>to</strong>: Little, Brown.<br />

Mills, Jack, and Vanda Brough<strong>to</strong>n (1977–). Bliss Bibliographic Classification, 2nd edn.<br />

London: Butterworth.<br />

Ranganathan, S[hiyali] R[amamrita] (1967). Prolegomena <strong>to</strong> Library Classification, 3rd<br />

edn. Bombay: Asia Publishing House.<br />

Ranganathan, Shiyali Ramamrita (1989). Colon Classification, 7th edn. Basic and Depth<br />

version. Revised and edited by M. A. Gopinath. Vol. 1, Schedules for Classification.<br />

Bangalore: Sarada Ranganathan Endowment for Library Science.<br />

Svenonius, Elaine (2000). The Intellectual Foundation of Information Organization.<br />

Cambridge, MA: MIT Press.


Vossen, Piek (1998). Introduction <strong>to</strong> EuroWordNet. Computers and the Humanities 32:<br />

73–89.<br />

15.<br />

Databases<br />

Stephen Ramsay<br />

Introduction<br />

Databases are an ubiqui<strong>to</strong>us feature of life in the modern age, and yet the most allencompassing<br />

definition of the term "database" – a system that allows for the efficient<br />

s<strong>to</strong>rage and retrieval of information – would seem <strong>to</strong> belie that modernity. The design of<br />

such systems has been a mainstay of humanistic endeavor for centuries; the seeds of the<br />

modern computerized database being fully evident in the many text-based taxonomies<br />

and indexing systems which have been developed since the Middle Ages. Whenever<br />

humanists have amassed enough information <strong>to</strong> make retrieval (or comprehensive<br />

understanding) cumbersome, technologists of whatever epoch have sought <strong>to</strong> put forth<br />

ideas about how <strong>to</strong> represent that information in some more tractable form.<br />

The computerized database, while a new development in this broad his<strong>to</strong>rical schema,<br />

nonetheless appeared more or less simultaneously with the early use of computers in<br />

academic and commercial environments. In such contexts, the essential problem of<br />

organization and efficient retrieval (usually unders<strong>to</strong>od as falling under the rubric of data<br />

structures and algorithms respectively) is complicated by the need for systems which<br />

facilitate interaction with multiple end users, provide platform-independent<br />

representations of data, and allow for dynamic insertion and deletion of information. The<br />

use of database technology among humanists has been invigorated by the realization –<br />

common, perhaps, <strong>to</strong> many other similar convergences – that a number of fascinating<br />

problems and intellectual opportunities lurk beneath these apparently practical matters.<br />

The inclusion of certain data (and the attendant exclusion of others), the mapping of<br />

relationships among entities, the often collaborative nature of dataset creation, and the<br />

eventual visualization of information patterns, all imply a hermeneutics and a set of<br />

possible methodologies that are themselves worthy objects for study and reflection.<br />

This chapter provides an introduction <strong>to</strong> these issues by working through the design and<br />

implementation of a simple relational database (our example s<strong>to</strong>res basic information<br />

about books in print). The intent, of course, is <strong>to</strong> remove some of the complexities and<br />

idiosyncrasies of real-world data in order that the technical and conceptual details of<br />

database design might more readily emerge. The data <strong>to</strong> which humanist scholars are<br />

accus<strong>to</strong>med – literary works, his<strong>to</strong>rical events, textual recensions, linguistic phenomena –<br />

are, of course, rarely simple. We would do well, however, <strong>to</strong> bear in mind that what<br />

might be viewed as a fundamental inadequacy has often proved <strong>to</strong> be the primary<br />

attraction of relational database systems for the humanist scholar. Rather than exploiting


a natural congruity between relational on<strong>to</strong>logies and humanistic data, scholars have<br />

often sought insight in the many ways in which the relational structure enforces a certain<br />

estrangement from what is natural. The terms we use <strong>to</strong> describe books in a books<strong>to</strong>re<br />

(authors, works, publishers) and the relationships among them (published by, created by,<br />

published in) possess an apparent stability for which the relational model is ideally suited.<br />

The most exciting database work in <strong>humanities</strong> computing necessarily launches upon less<br />

certain terri<strong>to</strong>ry. Where the business professional might seek <strong>to</strong> capture airline ticket sales<br />

or employee data, the humanist scholar seeks <strong>to</strong> capture his<strong>to</strong>rical events, meetings<br />

between characters, examples of dialectical formations, or editions of novels; where the<br />

accountant might express relations in terms like "has insurance" or "is the supervisor of",<br />

the humanist interposes the suggestive uncertainties of "was influenced by", "is<br />

simultaneous with", "resembles", "is derived from."<br />

Such relationships as these hold out the possibility not merely of an increased ability <strong>to</strong><br />

s<strong>to</strong>re and retrieve information, but of an increased critical and methodological selfawareness.<br />

If the database allows one <strong>to</strong> home in on a fact or relationship quickly, it<br />

likewise enables the serendipi<strong>to</strong>us connection <strong>to</strong> come forth. Relational databases in<br />

humanistic study are, in this sense, not so much pre-interpretative mechanisms as parainterpretative<br />

formations. As with so many similar activities in <strong>digital</strong> <strong>humanities</strong>, the act<br />

of creation is often as vital <strong>to</strong> the experiential meaning of the scholarly endeavor as the<br />

use of the final product.<br />

Relational database management systems (RDBMS) represent the most popular way of<br />

creating searchable on<strong>to</strong>logies both among computing humanists and among<br />

professionals in other areas of research and industry, and so this chapter will be<br />

concerned primarily with the design and implementation of database systems using the<br />

relational model. Still, the modern database landscape continues <strong>to</strong> evolve. Some<br />

consideration of where databases may yet be going (and where humanists may be going<br />

with database technology) is therefore apposite as well.<br />

The Relational Model<br />

E. F. Codd first proposed the relational model in a 1970 article in Communications of the<br />

ACM entitled "A Relational Model of Data for Large Shared Databanks." Codd's<br />

proposal endeavored <strong>to</strong> overcome the limitations of previous systems, which had suffered<br />

from difficulties related both <strong>to</strong> inefficient (which is <strong>to</strong> say slow) access and unwieldy<br />

s<strong>to</strong>rage mechanisms – inefficiencies that often resulted from redundancies in the<br />

underlying data representation. Codd's model made great strides forward in both areas,<br />

and yet his achievement is perhaps more acutely evident in the mathematical presentation<br />

of his ideas. One researcher, who refers <strong>to</strong> the 1970 paper as "probably the most famous<br />

paper in the entire his<strong>to</strong>ry of database management", notes:<br />

It was Codd's very great insight that a database could be thought of as a set of relations,<br />

that a relation in turn could be thought of as a set of propositions …, and hence that all of<br />

the apparatus of formal logic could be directly applied <strong>to</strong> the problem of database access<br />

and related problems.


(Date <strong>2001</strong>)<br />

This fundamental idea has spawned a vast literature devoted <strong>to</strong> database theory, and<br />

while there have been several major additions <strong>to</strong> the relational model, the relational<br />

databases of <strong>to</strong>day continue <strong>to</strong> operate on the basis of Codd's insights.<br />

Database Design<br />

The purpose of a database is <strong>to</strong> s<strong>to</strong>re information about a particular domain (sometimes<br />

called the universe of discourse) and <strong>to</strong> allow one <strong>to</strong> ask questions about the state of that<br />

domain. Let us suppose, for example, that we are creating a database that will contain<br />

information about current editions of American novels. Our goal will be <strong>to</strong> create a<br />

system that can s<strong>to</strong>re information about authors, works, and publishers, and allow us <strong>to</strong><br />

ask questions like "What publications did Modern Library produce in 1992?" and "Which<br />

works by Herman Melville are currently in print?" The simplest database of all would<br />

simply list the data in tabular form (see table 15.1).<br />

This database might be expanded <strong>to</strong> include a vast collection of authors and works. With<br />

the addition of some mechanism with which <strong>to</strong> s<strong>to</strong>re and query the data, we can easily<br />

imagine a system capable of answering the questions we would like <strong>to</strong> pose. Yet the<br />

inefficiencies, which the relational model endeavors <strong>to</strong> overcome, are evident even in this<br />

simple example. A search for "Mark Twain" will require that the system continue<br />

iterating through the rows after it has found its first hit in order <strong>to</strong> ensure that the relevant<br />

matches have been found. This is because our data model allows – and indeed, demands<br />

– that the name of the author be entered in<strong>to</strong> every row in which a new work is<br />

introduced. Similar redundancies occur with dates of publication, publisher names, and<br />

publisher addresses. Moreover, a change <strong>to</strong> an author's name (for example, the decision<br />

<strong>to</strong> enter "Samuel Clemens" in place of the author's pen name) will require that we update<br />

all fields in which the original name appears. Even if we devise some mechanism for<br />

ensuring a certain vigilance on the part of the machine, we are still left with a version of<br />

the same problem: having <strong>to</strong> go <strong>to</strong> "places instead of just one. In our example, the<br />

redundancy seems unproblematic – any machine can make quick work of a database with<br />

six items". In a system containing thousands or perhaps millions of items, the extra time<br />

and space required <strong>to</strong> perform search algorithms can become a severe liability.<br />

15.1 Table<br />

Last First YOB YOD Title<br />

Pub<br />

Year<br />

Publisher<br />

Pub<br />

Address<br />

Twain Mark 1835 1910 Huckleberry 1986 Penguin USA New York<br />

Twain Mark 1835 1910 Tom Sawyer 1987 Viking New York<br />

Cather Willa 1873 1947 My An<strong>to</strong>nia 1995<br />

Library of<br />

America<br />

New York


Hemingway Ernest 1899 1961 The Sun Also Rises 1995 Scribner New York<br />

Wolfe<br />

Look Homeward,<br />

Thomas 1900 1938<br />

Angel<br />

1995 Scribner New York<br />

Faulkner<br />

The Sound and the<br />

William 1897 1962<br />

Furry<br />

1990 Random House New York<br />

Relational modeling attempts <strong>to</strong> fac<strong>to</strong>r these redundancies out of the system. We can<br />

begin modifying our original design by isolating the individual entities in the domain at a<br />

more abstract level; that is, by locating the types of information that vary independently<br />

of one another. The preliminary outline shown in figure 15.1 might emerge as one<br />

possible representation of the domain.<br />

Each of the boxes in figure 15.1 illustrates a particular entity with a set of associated<br />

attributes. We have retained all the information from the original design, but have<br />

sequestered the various entities from one another according <strong>to</strong> certain logical groupings:<br />

authors (who have last names, first names, and dates of birth and death), works (which<br />

have titles and years of publication), and publishers (which have names and cities where<br />

they are headquartered). Often, the nouns we use <strong>to</strong> describe the domain and the<br />

recurrence of the word "have" helps <strong>to</strong> establish these entities and their attributes. To this<br />

basic outline we may also add a set of verb phrases describing the nature of the<br />

relationships among the entities. For example, authors create works, and works are, in<br />

turn, published by publishers. The addition of this information can then give rise <strong>to</strong> what<br />

is called an entity relationship (ER) diagram (see figure 15.2).<br />

This diagram captures the basic relationships we have isolated, but it remains <strong>to</strong> say how<br />

many instances of a single entity can be associated with other entities in the model. For<br />

example, one author may contract with several publishers, and a publisher may offer<br />

many different works by multiple authors. There are several ways <strong>to</strong> capture these<br />

features diagrammatically 1 . We will simply use the number "1" <strong>to</strong> indicate a single<br />

instance and an "M" <strong>to</strong> indicate multiple instances (see figure 15.3). We may then read<br />

the relationship line connecting authors and works <strong>to</strong> mean "One author has many<br />

works."


15.1 Figure<br />

Thus far, we have been pursuing the logical design of our database – a design entirely<br />

independent of both the eventual machine representation of the data and of the end user's<br />

view of the information contained within it. We would do well at this point <strong>to</strong> imagine<br />

how this logical structure might be populated with the particular instances from the first<br />

model. To do this, we need <strong>to</strong> make a subtle mental shift in the way we view the entity<br />

relationship diagram. We might be tempted <strong>to</strong> see the various boxes as s<strong>to</strong>rage areas that<br />

can hold instances of authors' last names, work titles, and so forth. However, we have<br />

been really modeling the generic form that the particular instances of data will take. The<br />

populated database is properly conceived of as a set of tables with rows and columns, in<br />

which each row corresponds <strong>to</strong> the entities and each column <strong>to</strong> the attributes in the ER<br />

diagram. These rows are usually referred <strong>to</strong> as records and the intersection of rows and<br />

columns as fields. Table 15.2, for example, is a mock-up of a populated database built<br />

according <strong>to</strong> the terms of the ER diagram.<br />

This more concrete view of our database captures the entities, but it makes no mention of<br />

the relationships. In order <strong>to</strong> represent the relationships between records, we need <strong>to</strong><br />

introduce some variable that can hold these connections.<br />

Our ability <strong>to</strong> do this will be significantly enhanced if we can devise some way <strong>to</strong> refer <strong>to</strong><br />

each individual instance of a particular entity as a unique datum. After all, the final<br />

database will not merely connect authors <strong>to</strong> works in some generic way, but will reflect<br />

the fact that, for example, the author Mark Twain created both Huckleberry Finn and<br />

Tom Sawyer. The usual method for establishing this uniqueness is <strong>to</strong> create a primary key<br />

for each record – a unique value associated with each individual record in a table 2 . This<br />

value is simply a new attribute which can be added <strong>to</strong> our ER diagram, and by extension,<br />

a new column in the final database for each record type. The author entity, for example,<br />

may be modified as shown in figure 15.4.


15.2 Figure<br />

15.3 Figure<br />

15.2 Table<br />

AUTHORS<br />

Last Name First Name Year of Birth Year of Death<br />

Twain Mark 1835 1910<br />

Cather Willa 1873 1947<br />

Hemingway Ernest 1899 1961<br />

Wolfe Thomas 1900 1935<br />

Faulkner William 1897 1962


WORKS<br />

Title PubYear<br />

The Adventures of Huckleberry Finn 1986<br />

Tom Sawyer 1987<br />

My An<strong>to</strong>nia 1995<br />

The Sun also Rises 1995<br />

Look Homeward, Angel 1995<br />

The Sound and the Fury 1990<br />

PUBLISHERS<br />

Name City<br />

Penguin USA New York<br />

Library of America New York<br />

Schribner New York<br />

Viking New York<br />

Random House New York<br />

15.4 Figure<br />

The resulting database table would then include a new column <strong>to</strong> hold this value (see<br />

table 15.3). With the other entities similarly modified, we now have a way of referring <strong>to</strong><br />

each individual record in a table without ambiguity.<br />

The next step is really at the heart of the relational model. In order <strong>to</strong> capture the one-<strong>to</strong>many<br />

(1:M) relationship between authors and works, we introduce a second key attribute<br />

<strong>to</strong> the entities on the "many" side of the relationship – one that can hold a reference (or<br />

pointer) back <strong>to</strong> the entity on the "one" side of the relationship. This reference is, like the<br />

primary key, simply another attribute called a foreign key. Table 15.4, for example,<br />

shows how the "Works" table would look with additional fields for primary and foreign<br />

keys.<br />

The foreign key field contains the primary key of the record with which it is associated.<br />

Thus the records for The Adventures of Huckleberry Finn and Tom Sawyer (which have<br />

been assigned primary keys 1 and 2, respectively) now contain foreign key references <strong>to</strong>


the record in the "Authors" table which bears primary key 1 (the record for "Mark<br />

Twain"). In this way, the database is able <strong>to</strong> retain one reference for the author "Mark<br />

Twain." The redundancy that hindered the original design has been eliminated.<br />

Unfortunately, the database still contains other instances of redundancy. For example,<br />

every publisher in our database is located in New York, which means this information is<br />

repeated for all six publisher records. Theoretically, this situation could have been<br />

avoided if we had held ourselves <strong>to</strong> a very strict interpretation of "types of information<br />

that vary independently of one another" in our initial ER diagram. In practice, such<br />

redundancies are often difficult <strong>to</strong> discern upon initial analysis of the domain. It may<br />

even be that some redundancies only appear after a considerable amount of data has<br />

already been entered in<strong>to</strong> a pro<strong>to</strong>type system.<br />

15.3 Table<br />

AUTHORS<br />

Author ID Last Name First Name Year of Birth Year of Death<br />

1 Twain Mark 1835 1910<br />

2 Cather Willa 1873 1947<br />

3 Hemingway Ernest 1899 1961<br />

4 Wolfe Thomas 1900 1938<br />

5 Faulkner William 1897 1962<br />

15.4 Table<br />

WORKS<br />

Work ID Title PubYear Author ID<br />

1 The Adventures of Huckleberry Finn 1986 1<br />

2 Tom Sawyer 1987 1<br />

3 My An<strong>to</strong>nia 1995 2<br />

4 The Sun Also Rises 1995 3<br />

5 Look Homeward, Angel 1995 4<br />

6 The Sound and the Fury 1990 5<br />

In any event, eliminating the redundancy in the "Publishers" table is simply a matter of<br />

breaking the "City" attribute off in<strong>to</strong> its own table, assigning a primary key value <strong>to</strong> each<br />

record instance, and providing a new foreign key field in the "Publishers" table which can


hold a reference <strong>to</strong> the correct city. In other words, we need <strong>to</strong> take one of our attributes<br />

and elevate it <strong>to</strong> the status of entity (see table 15.5).<br />

Since primary key values in one table can be referenced from multiple tables as foreign<br />

keys, this restructuring may have a useful side effect if we ever decide <strong>to</strong> add "Place of<br />

Birth" <strong>to</strong> the Authors table.<br />

15.5 Table<br />

PUBLISHERS<br />

Pub ID Name City ID<br />

1 Penguin USA 1<br />

2 Library of America 1<br />

3 Scribner 1<br />

4 Viking 1<br />

5 Random House 1<br />

CITIES<br />

CityID City<br />

1 New York<br />

There is another kind of relationship in this domain which isn't represented; namely, the<br />

many-<strong>to</strong>-many (M:M) relationship. This situation might easily arise if several publishers<br />

release editions of the same work. We would naturally describe this relationship as being,<br />

like all the relationships in the current design, one-<strong>to</strong>-many, but in this case, there is<br />

already a one-<strong>to</strong>-many relationship going the other way (from "Publishers" <strong>to</strong> "Works").<br />

One might be tempted simply <strong>to</strong> introduce a foreign key pointing <strong>to</strong> "Works" from<br />

"Publishers" <strong>to</strong> complement the foreign key pointing from "Publishers" <strong>to</strong> "Works."<br />

However, the more apt solution is <strong>to</strong> abstract the relationship in<strong>to</strong> a new entity called an<br />

association (or junction table). An association is simply a new table which contains the<br />

two related foreign keys (see table 15.6). This association captures the fact that Penguin<br />

USA (Pub ID 1) publishes The Adventures of Huckleberry Finn (Work ID 1) and an<br />

edition of Tom Sawyer (Work ID 2).<br />

Each record in an association may be assigned a primary key, but in this case (and in the<br />

case of most associations) the combination of the two primary keys is unders<strong>to</strong>od <strong>to</strong><br />

represent a unique combination. Most RDBMSs therefore allow one <strong>to</strong> declare the pair of<br />

values <strong>to</strong> be the primary key for that record (the creation of these compound keys will be<br />

discussed in the next section).<br />

We have now analyzed our domain with entity-relationship modeling and have begun <strong>to</strong><br />

fac<strong>to</strong>r out the major redundancies in the model. Readers interested in the formal


explication of these methodologies will find abundant resources in the scholarly literature<br />

of the field, and while it is unnecessary <strong>to</strong> go in<strong>to</strong> these matters here, at least one aspect<br />

of the more technical discussion deserves mention even in an introduc<strong>to</strong>ry context.<br />

Database theorists (and serious designers) often speak of databases as being in one of five<br />

normal forms. The normal forms may be stated in set theoretic terms, but they may also<br />

be stated more simply as design criteria by which <strong>to</strong> judge the soundness of one's design.<br />

For example, one practically minded book paraphrases first normal form by stating that<br />

"at each row-and-column intersection, there must be one and only one value, and that<br />

value must be a<strong>to</strong>mic: there can be no repeating groups in a table that satisfies first<br />

normal form" (Bowman et al. 1999).<br />

By the time a database is in fifth normal form, all redundancy has been removed; as<br />

Bowman puts it, "Tables normalized <strong>to</strong> this extent consist of little more than primary<br />

keys" (Bowman et al. 1999). This has the advantage of making it easier for the RDBMS<br />

<strong>to</strong> ensure the overall integrity of the data, but one may find that queries on those data<br />

become rather confusing <strong>to</strong> compose. As with most matters related <strong>to</strong> computer<br />

programming, one needs <strong>to</strong> balance the goals of correctness against the practical<br />

exigencies of the system and its users.<br />

15.6 Table<br />

PUBLISHER-WORKS TABLE<br />

Pub ID Work ID<br />

1 1<br />

1 2<br />

2 2<br />

3 3<br />

4 4<br />

Schema Design<br />

Up until now, our meditations on database design have been confined <strong>to</strong> what one might<br />

normally undertake at the whiteboard. The actual implementation of the design is much<br />

more akin <strong>to</strong> programming. Fortunately, there is little in the implementation stage that<br />

requires new concepts; for the most part, we simply need <strong>to</strong> translate our design in<strong>to</strong> a<br />

representation intelligible <strong>to</strong> the machine. This representation is usually referred <strong>to</strong> as a<br />

database schema, and is created using Structured Query Language (SQL) 3 .<br />

There are a number of excellent SQL-compliant RDBMSs available <strong>to</strong> the humanist<br />

scholar. The majority of <strong>humanities</strong> computing projects that use databases employ free


(open source) systems, of which the most popular are MySQL, mSQL, and Post-greSQL.<br />

There are also a number of commercial systems in use (Oracle, Microsoft Access, IBM<br />

DB2). The systems differ somewhat in their feature sets and in the amount of<br />

documentation available, but all provide efficient, feature-rich implementations of the<br />

relational model with advanced security and management functions. For our database, we<br />

will use PostgreSQL – a free, well-supported RDBMS for Unix-like systems which can<br />

be downloaded over the Internet. 4<br />

The database schema is nothing more than a machine-readable representation of the ER<br />

diagram. We may begin by laying out the various entities and their attributes, but since<br />

we are now giving instructions <strong>to</strong> the machine, we need <strong>to</strong> be more specific about the<br />

precise nature of the data <strong>to</strong> be entered:<br />

Like most programming languages, SQL includes the notion of a datatype. Datatype<br />

declarations help the machine <strong>to</strong> use space more efficiently and also provide a layer of<br />

verification for when the actual data is entered (so that, for example, a user cannot enter<br />

character data in<strong>to</strong> a date field). In this example, we have specified that the last_name,<br />

first_name, title, city, and name fields will contain character data of varying length (not<br />

<strong>to</strong> exceed 80 characters), and that the year_of_birth, year_of_cleath, and pub_year fields<br />

will contain integer data. Other possible datatypes include DATE (for day, month, and<br />

year data), TEXT (for large text blocks of undetermined length), and BOOLEAN (for<br />

true/ false values). Most of these can be further specified <strong>to</strong> account for varying date<br />

formats, number bases, and so forth. PostgreSQL, in particular, supports a wide range of<br />

datatypes, including types for geometric shapes, Internet addresses, and binary strings.<br />

Databases often differ in the way they represent (and ensure the integrity of) primary<br />

keys. In PostgreSQL, the usual method is <strong>to</strong> create a separate mechanism for generating<br />

and keeping track of unique numeric values, and <strong>to</strong> have the tables containing the entities


etrieve a value from that mechanism each time a new record is created. So, for example,<br />

<strong>to</strong> create a primary key for the authors table, we first create what is called a sequence<br />

table that will hold the set of unique values for that table:<br />

We can then add a new field <strong>to</strong> the author table which will have as its default value the<br />

next value called from the appropriate sequence:<br />

This new line thus amounts <strong>to</strong> the following declaration about the author_icl field: the<br />

author_icl will have as its default value an integer that corresponds <strong>to</strong> the next value<br />

provided by the sequence named author_seq. 5<br />

We also want <strong>to</strong> declare this value specifically as the primary key for the table. This is<br />

accomplished with the following addition:<br />

A foreign key is simply another data field, which, like the primary key field, has been<br />

specifically designated as a key. In order <strong>to</strong> capture the one-<strong>to</strong>-many relationship between<br />

works and authors, for example, we would modify the works table as follows:


The PRIMARY KEY () specifier also makes it easy <strong>to</strong> declare the publisher and work ids<br />

in our publisher/work association as a compound key:<br />

The designation of primary and foreign keys is one of the most important aspects of<br />

schema design, because it helps the system <strong>to</strong> achieve what is called referential integrity.<br />

Referential integrity is compromised when we delete a record that contains a reference <strong>to</strong><br />

another record. So, for example, if we were <strong>to</strong> delete an author from the authors table, we<br />

might leave a reference <strong>to</strong> that author in the works table without a referent. A good<br />

RDBMS will use the primary and foreign key references either <strong>to</strong> prevent this from<br />

occurring or <strong>to</strong> warn the database administra<strong>to</strong>r that the operation will result in a dangling<br />

reference (sometimes called a null pointer).<br />

The method for creating an empty database and getting this schema in<strong>to</strong> the RDBMS<br />

varies from system <strong>to</strong> system. Most RDBMSs provide command-line <strong>to</strong>ols for setting up<br />

databases and for executing commands through an interactive command interpreter. The<br />

documentation for the particular RDBMS will discuss such matters in detail 6 .<br />

Importing Data<br />

Importing data in<strong>to</strong> this schema requires the use of another set of SQL commands. The<br />

most useful of these is the INSERT keyword, which adds a record <strong>to</strong> one of the tables<br />

specified in the schema:<br />

The structure of this command is such that the target fields are declared in the first set of<br />

parentheses and the actual values in the second (the structure of the DELETE command,


for removing data from a database, uses the same syntax). In most systems, character<br />

strings must be enclosed in quotation marks. Notice also that the INSERT statement does<br />

not include a primary key because our database schema has already instructed the system<br />

<strong>to</strong> use the next value in the sequence table as the default. A work record may be added in<br />

much the same way:<br />

In this case, however, we need <strong>to</strong> add a foreign key value that contains a primary key<br />

from the authors table. There are three ways <strong>to</strong> accomplish this. The first is simply <strong>to</strong><br />

look up the appropriate record in the authors table before executing the INSERT<br />

statement for the work, note its primary key, and add this <strong>to</strong> the statement. The second<br />

way is <strong>to</strong> perform an UPDATE statement after the record has been created, which adds<br />

the primary key for the author <strong>to</strong> the foreign key field author_icl of the work record. The<br />

third, and perhaps the most efficient way, is <strong>to</strong> embed the statement that looks up the<br />

appropriate primary key in<strong>to</strong> the INSERT statement for adding a work record. All three<br />

of these methods require that we have an understanding of how <strong>to</strong> query the database for<br />

information, so let us defer this discussion for a moment while we explore the query<br />

commands of SQL.<br />

Database Queries<br />

The database administra<strong>to</strong>r can query the database using the SELECT command. To<br />

retrieve all the last names currently in the authors table, for example, one would execute<br />

the following command:<br />

Most RDBMSs will produce output that looks something like this:<br />

In most circumstances, however, we want <strong>to</strong> create a more complex query that will let us<br />

home in on a particular record. Here, for example, is a SELECT statement that only<br />

selects authors born after 1890:


This statement begins by isolating two fields which should be returned <strong>to</strong> the user, but<br />

then adds a WHERE clause which qualifies the search according <strong>to</strong> particular criteria – in<br />

this case, <strong>to</strong> records in which the integer value contained in the year_of_birth field is<br />

greater than (or equal <strong>to</strong>) 1890. The result is a list of authors:<br />

Assuming we have already created records for Mark Twain and for The Adventures of<br />

Huckleberry Finn, we can now use the query syntax <strong>to</strong> discover the primary key for Mark<br />

Twain and use it <strong>to</strong> fill in the foreign key field for his works. The first is accomplished<br />

with a SELECT statement:<br />

The system returns the value "1." We can now use an UPDATE statement <strong>to</strong> modify the<br />

appropriate work record:<br />

The more efficient method alluded <strong>to</strong> above – loading a record in<strong>to</strong> the works table with<br />

the appropriate foreign key in one statement – uses what is called a subselect:<br />

This statement follows the usual syntax of the INSERT statement, except that the last<br />

value in the second set of parentheses (the one pertaining <strong>to</strong> author_id) is itself a<br />

SELECT statement designed <strong>to</strong> return a single value: the primary key for the author<br />

record with the word "Twain" in the last_name field.<br />

Database queries can reach a significant level of complexity; the WHERE clause can<br />

accept Boolean opera<strong>to</strong>rs (e.g.''WHERE year_of_birth > 1890 AND year_of_-birth <<br />

1900''), and most SQL implementations offer keywords for changing the ordering of the<br />

output (e.g. ORDER BY year_of_birth). Taking full advantage of the relational model,<br />

moreover, requires that we be able <strong>to</strong> gather information from several tables and present<br />

it in one result list. This operation is called a join.<br />

Let us suppose that we want <strong>to</strong> return a single result set that lists the authors' names,<br />

titles, and publishers of all works in the database. Doing so will require that we gather


information from the author, publisher, and work tables, and also from the association<br />

that links publishers with works. The syntax for constructing this query follows the<br />

general template for SELECT … FROM … WHERE … but with a twist:<br />

Any database of non-trivial complexity will have equivalent column names spread across<br />

several tables. Such equivalence will naturally occur with columns representing common<br />

concepts (such as "date"), but it will occur inevitably with the keys (since the primary key<br />

label in one table will often occur as the foreign key label in several others). The key <strong>to</strong><br />

understanding this join query lies in the FROM clause, where each table participating in<br />

the join is aliased <strong>to</strong> a short variable. In this query, the variable a will stand in for the<br />

authors table, w for works, and so on. Thus we may read the first part of the query as<br />

saying "Select the last_name field from the authors table, the first_name field from the<br />

authors table, the name field from the publishers table, and the title field from the works<br />

table." The WHERE clause then tries <strong>to</strong> match up the appropriate key columns: the<br />

publisher_id in the publishers and pub_works tables must match, the work_id in the<br />

works and pub_works tables must match, and the author_ids in the authors and works<br />

tables must match. The RDBMS will then return all records that match all of these<br />

criteria in the order in which they were requested in the first part of the query:<br />

Constraining this query <strong>to</strong> a single author is simply a matter of adding another constraint<br />

<strong>to</strong> the WHERE clause: au . last_name = d 'Twain'.<br />

The necessity for join operations (and subselects, which can often be used <strong>to</strong> accomplish<br />

the same thing) increases as the databases approach full normalization. Database<br />

administra<strong>to</strong>rs wishing <strong>to</strong> take full advantage of a good relational design will want <strong>to</strong><br />

study the particular features of their SQL implementation closely.<br />

Database Management


A good RDBMS will include all the facilities necessary for implementing a schema,<br />

importing data in<strong>to</strong> the system, and performing queries. However, such facilities<br />

represent only part of the overall set of functions and capabilities necessary for creating a<br />

production system.<br />

Prudence would suggest that only a few privileged users should possess the ability <strong>to</strong><br />

create and destroy databases, and that a broader (but not unlimited) set of users be able <strong>to</strong><br />

add data <strong>to</strong> existing databases. There may also be administrative data s<strong>to</strong>red in the<br />

database that are not intended for ordinary users of the system. All of the major RDBMSs<br />

provide facilities – ranging from simple permissions files <strong>to</strong> elaborate security systems –<br />

for limiting user access <strong>to</strong> the system. Most simply use a separate database (integral <strong>to</strong> the<br />

RDBMS itself) that permits or denies the ability <strong>to</strong> run specific commands (often using<br />

the GRANT and REVOKE keywords from SQL). Since RDBMSs usually run as network<br />

processes (in order <strong>to</strong> allow for remote connections), some mechanism should also be in<br />

place for limiting connections <strong>to</strong> particular network domains. Unlike simple user<br />

applications (which usually come with a reasonable set of security defaults), advanced<br />

multi-user applications often assume that the administra<strong>to</strong>r will take the time <strong>to</strong> study the<br />

security model carefully and implement the appropriate procedures. The assumption that<br />

users will do this is perhaps naive, but the assumption that the developers have already<br />

done everything <strong>to</strong> make the particular installation secure can be reckless indeed.<br />

Transaction management represents another related set of concerns. Suppose that we had<br />

created a large file full of INSERT statements for the initial load of our data, but had left<br />

off the semicolon at the end of one of the statements. In all likelihood, the RDBMS<br />

(interpreting each INSERT statement as a discrete SQL operation) would execute the first<br />

n number of statements, inserting the data for each one, only <strong>to</strong> break in the middle of the<br />

list. We could perhaps delete the data already inserted, fix the error in the file, and reload<br />

the entire file, or perhaps we could erase the INSERT statements from the file that we<br />

have already executed and start anew. We would do better, however, if we had some way<br />

<strong>to</strong> declare all the INSERT statements in the file as an "all-or-nothing" block; the database<br />

has <strong>to</strong> perform all of these operations successfully or else it must exit with a warning,<br />

having performed none of them. Most systems provide some method for declaring a set<br />

of statements as a transaction block of this kind. In PostgreSQL, for example, the<br />

keywords are BEGIN (for signaling the beginning of a transaction), COMMIT (for<br />

executing the statements in the event that the proposed operations would generate no<br />

errors), and ROLLBACK (for undoing a set of operations). 7<br />

While the relational model lends itself <strong>to</strong> efficient s<strong>to</strong>rage and retrieval mechanisms, the<br />

sheer number of insertions, deletions, and updates on a production database system will<br />

often result in a certain fragmentation of the data (much in the way a hard drive can<br />

become fragmented after many frequent read-write operations). Much research has been<br />

devoted <strong>to</strong> finding new ways <strong>to</strong> defragment and otherwise optimize s<strong>to</strong>rage and retrieval<br />

in heavily used RDBMSs. While some of these operations occur behind the scenes,<br />

others are designed <strong>to</strong> be cus<strong>to</strong>mized by the administra<strong>to</strong>r for local circumstances. Such<br />

optimizations become more and more pertinent as a database (or a userbase) grows in<br />

size, and administra<strong>to</strong>rs will want <strong>to</strong> avail themselves of these facilities when appropriate.


Databases and Software Development<br />

Database administra<strong>to</strong>rs (particularly those working in a Unix environment) will often<br />

find that the humble command-line interface represents the most efficient way <strong>to</strong> create<br />

database schema, load large files full of SQL statements, check the validity of queries,<br />

and perform the various tasks associated with user accounts and security. However, many<br />

database applications in <strong>humanities</strong> computing are intended for use outside the group of<br />

scholars and developers implementing the system. In such cases, low-level access <strong>to</strong> the<br />

RDBMS is rarely desirable. Indeed, it is often useful <strong>to</strong> design a system that shields users<br />

from SQL entirely.<br />

The construction of middleware systems intended <strong>to</strong> provide such abstraction is among<br />

the more common software development tasks in <strong>humanities</strong> computing. In the case of<br />

both standalone and web-based applications, the usual goal is <strong>to</strong> place the execution of<br />

SQL queries and commands "in between" the user interface and the underlying RDBMS.<br />

Such a design (often called an n-tiered design) has the additional virtue of allowing the<br />

interface, database, and intervening programming logic <strong>to</strong> vary independently of one<br />

another; in a well-designed system, a change <strong>to</strong> the user interface (for example) would<br />

ideally imply only minor changes <strong>to</strong> the other tiers. Fortunately, nearly every<br />

programming language in common use among computing humanists provides elaborate<br />

mechanisms for executing SQL commands against a database from within an application.<br />

The DBI (or DataBase independent) module is an extension <strong>to</strong> the Perl programming<br />

language which provides a set of functions and variables for communicating with an<br />

RDBMS. In addition <strong>to</strong> allowing one <strong>to</strong> pass SQL statements <strong>to</strong> a database, retrieve<br />

results, and restructure data for delivery (<strong>to</strong>, for example, a web browser), DBI also<br />

provides an extremely powerful plug-in architecture that allows one <strong>to</strong> use the same<br />

program with several database implementations. So, for example, a program using the<br />

DBI interface could retrieve values from a web form, embed those values in a series of<br />

SQL statements, and pass them off <strong>to</strong> the RDBMS. Instead of going directly <strong>to</strong> the<br />

RDBMS, however, the DBI system would look for a driver (or DBD module)<br />

corresponding <strong>to</strong> the particular RDBMS being used, and effectively translate the DBI<br />

SQL syntax in<strong>to</strong> the dialect of the particular database. Switching from PostgreSQL <strong>to</strong><br />

MySQL (or perhaps running the two simultaneously) is therefore simply a matter of<br />

adding a new driver. DBI is also fairly mature; there are DBD modules corresponding <strong>to</strong><br />

nearly every major database implementation, and all are freely available. 8<br />

Java programmers can find the same concept implemented in the JDBC (or Java.<br />

DataBase Connectivity) library. JDBC is one of the more comprehensive<br />

implementations of SQL, and drivers are also available for all the major RDBMSs. Java<br />

also has the advantage of avoiding the overhead of CGI (Common Gateway interface)<br />

upon which Perl and other similar scripting languages depend. CGI usually requires that a<br />

separate copy of the language interpreter be loaded in<strong>to</strong> memory each time the program is<br />

invoked. The Java servlet architecture (intended for creating server-side applications)<br />

avoids this overhead by having a copy of the Java Virtual Machine (JVM) running<br />

constantly in memory with various caching facilities for the individual programs running


on it 9 . This method is generally superior <strong>to</strong> CGI and can leverage the considerable power<br />

and flexibility of the Java programming language, but one sacrifices the simplicity of the<br />

CGI in the process.<br />

Another technology that deserves mention is PHP (PHP Hypertext Processor). PHP is a<br />

programming language that can be embedded in ordinary HTML pages and interpreted<br />

by a program embedded in<strong>to</strong> the web server itself (thus aligning itself with Java servlets<br />

in its avoidance of CGI). PHP does not employ the concept of a single interface with<br />

multiple drivers, but instead provides built-in function sets for all the major RDBMSs. Of<br />

the three middleware technologies here discussed, PHP is perhaps the easiest <strong>to</strong> use, but it<br />

is unfortunately the least "middle" of the three. Embedding code in<strong>to</strong> an HTML page<br />

implies mingling the logic of the user interface with the programming logic – a<br />

convergence that can lead <strong>to</strong> systems which are difficult <strong>to</strong> maintain. Often, the size (and<br />

predicted mutability) of the system will determine whether the simplicity of PHP<br />

outweighs the potential for confusion later on. 10<br />

Alternative Models<br />

The relational database model has been an extremely successful one. Previous models –<br />

hierarchical databases and network databases – are seldom used <strong>to</strong>day, and it seems clear<br />

that the relational model will continue <strong>to</strong> be the dominant one for many years <strong>to</strong> come 11 .<br />

However, there are a number of competing models which have been the subject of active<br />

research in computer science and information technology circles over the last two<br />

decades. While none are as widely employed as the relational model, the usual proclivity<br />

for exploration and early adoption among computing humanists may well serve <strong>to</strong> bring<br />

these models in<strong>to</strong> prominence in <strong>humanities</strong> research.<br />

The most active contender for the prominence of the relational model is the objec<strong>to</strong>riented<br />

(OO) database model. The impetus for its creation lies in the widespread use of<br />

the object-oriented programming paradigm among software engineers. The details of this<br />

paradigm are beyond the scope of this discussion, but some sense of the model may be<br />

gleaned from the observation that relational schemas rely upon an older conception of<br />

programming in which the data and the procedures for manipulating those data are kept<br />

separate from one another. In our database, for example, the data pertaining <strong>to</strong> authors are<br />

entirely separate from the various operations (queries) we would like <strong>to</strong> perform on those<br />

data. The object-oriented model proposes that the data and the procedures be refac<strong>to</strong>red<br />

in<strong>to</strong> discrete objects. The data for the author table and the elements of the operations that<br />

can be performed on them would therefore belong <strong>to</strong> the same basic structure. Partisans<br />

of OO argue that this model facilitates maintenance by creating fewer dependencies<br />

between data items, allows for reusable data modules that can be easily moved from one<br />

database context <strong>to</strong> another, and creates a certain semantic richness (through the creation<br />

of inheritance hierarchies) in the data unattainable with more conventional methods.<br />

While the more mainstream database developers have been reluctant <strong>to</strong> embrace this<br />

model fully, the relational database landscape has begun <strong>to</strong> see implementations which<br />

bill themselves as object-relational database management systems (ORDBMS). Such


implementations typically fall short of the fully object-oriented databases envisioned by<br />

researchers, but borrow the notions of inheritance (having one table "inherit" properties<br />

from another without duplication) and complex object creation commonly associated<br />

with object-oriented systems. It seems clear that the trend <strong>to</strong>ward such hybrid systems<br />

will continue.<br />

Databases and the Humanist<br />

Whether it is the his<strong>to</strong>rian attempting <strong>to</strong> locate the causes of a military conflict, the<br />

literary critic teasing out the implications of a metaphor, or the art his<strong>to</strong>rian tracing the<br />

development of an artist's style, humanistic inquiry reveals itself as an activity<br />

fundamentally dependent upon the location of pattern. Dealing with patterns necessarily<br />

implies the cultivation of certain habits of seeing; as one critic has averred: "Recognizing<br />

a pattern implies remaining open <strong>to</strong> gatherings, groupings, clusters, repetitions, and<br />

responding <strong>to</strong> the internal and external relations they set up" (Hunter 1990). Of all the<br />

technologies in use among computing humanists, databases are perhaps the best suited <strong>to</strong><br />

facilitating and exploiting such openness. To build a database one must be willing <strong>to</strong><br />

move from the forest <strong>to</strong> the trees and back again; <strong>to</strong> use a database is <strong>to</strong> reap the benefits<br />

of the enhanced vision which the system affords.<br />

Humanists have used relational databases as the engines behind complex visualization<br />

systems, text archives, and multimedia works. In most cases the intent has been merely <strong>to</strong><br />

leverage the efficiencies of the relational model as a means for s<strong>to</strong>ring and retrieving the<br />

information needed <strong>to</strong> populate a map, load a list of hits, or assemble a website.<br />

However, even in these relatively simple applications it becomes clear that the underlying<br />

on<strong>to</strong>logy has considerable intellectual value. A well-designed database that contains<br />

information about people, buildings, and events in New York City contains not static<br />

information, but an entire set of on<strong>to</strong>logical relations capable of generating statements<br />

about a domain. A truly relational database, in other words, contains not merely "Central<br />

Park", "Frederick Law Olmstead", and "1857", but a far more suggestive string of logical<br />

relationships (e.g., "Frederick Law Olmstead submitted his design for Central Park in<br />

New York during 1857").<br />

One possible path for the future may seek <strong>to</strong> exploit further the implications of<br />

collaborative database creation. A database, as we have seen, can be set up in such a way<br />

as <strong>to</strong> allow multiple users access <strong>to</strong> the insert mechanisms of the system. The challenges<br />

proposed by this possibility are substantial, since the robustness of a system largely<br />

depends upon its consistency. Enforceable rules mechanisms (far above the level of mere<br />

transaction management) would need <strong>to</strong> be devised <strong>to</strong> ensure such consistency. The<br />

successful employment of such systems in humanistic contexts, however, would expand<br />

the possibilities of knowledge representation considerably. Since the data would enter<br />

in<strong>to</strong> the system from a number of different sources, the logical statements that would<br />

flow from that on<strong>to</strong>logy would necessarily exceed the knowledge of any one individual.<br />

The power of relational databases <strong>to</strong> enable the serendipi<strong>to</strong>us apprehension of<br />

relationships would be that much more increased.


There is, of course, ample precedent for using complex, collaboratively managed data<br />

structures in humanistic inquiry. The earliest concordances were nominally produced <strong>to</strong><br />

assist scholars in locating passages in the Bible, and one of the earliest uses of computers<br />

in humanistic study was a concordance of Thomas Aquinas. In both cases, the ultimate<br />

goal was not efficient retrieval, but interpretative insight. It only seems appropriate that<br />

after we have designed and implemented relational systems, and reaped the benefits of<br />

the efficiencies they grant, we consider the role they may play in the varied pursuits<br />

which have descended from what was once called – appropriately – the higher criticism.<br />

Note<br />

1 Entity Relationship Modeling (developed by Peter Chen in the mid-1970s) is by far the<br />

most popular diagramming notation; even people unacquainted with formal<br />

methodologies will often adopt something like it when working through an initial design.<br />

Unfortunately, there is no standard outlining the precise notation (and thus enabling<br />

designers <strong>to</strong> communicate in a common format). Other, perhaps more sophisticated<br />

notations include Object-Role Modeling (ORM) and (for object- oriented database<br />

design) Unified Modeling Language (UML). See Halpin (<strong>2001</strong>).<br />

2 Some database systems can generate a unique value for each record in the entire<br />

database. While this ensures an additional level of integrity useful (and perhaps even<br />

requisite) for other aspects of database management, it is not strictly necessary for the<br />

relational model as such.<br />

3 SQL is specified in a standards document first issued by the American National<br />

Standards Institute in 1983, later revised in 1992, and again in 1999- While this standard<br />

has gone a long way <strong>to</strong>ward creating a common <strong>to</strong>ngue among database programmers,<br />

companies and developer groups that create RDBMSs continue <strong>to</strong> introduce subtle<br />

dialectical differences between implementations. Fortunately, these differences are<br />

relatively minor; the programmer searching for a particular semantic construct in one<br />

implementation will usually be able <strong>to</strong> find its counterpart in another with relative ease.<br />

4 PostgreSQL (available at http://www.postgresql.org) is an open source application that<br />

can be modified and distributed freely (in source or binary form) under the terms of the<br />

BSD License. While it is intended for Unix-like systems (Linux, Solaris, AIX,* etc.),<br />

PostgreSQL can be run under Microsoft Windows using the Cygwin <strong>to</strong>ols (available at<br />

http://www.cygwin.com). See Geschwinde (<strong>2001</strong>) and Stinson (<strong>2001</strong>).<br />

5 The creation of unique values for use as primary keys is one of the areas in which<br />

RDBMSs differ the most, and most databases provide facilities for ensuring uniqueness<br />

that amount <strong>to</strong> extensions <strong>to</strong> the SQL-92 standard. Such is the case with the PostgreSQL<br />

sequence construct and with, for example, MySQL's use of the (somewhat less flexible)<br />

AUTO_INCREMENT keyword.<br />

6 In PostgreSQL, the command for creating a database is simply createdb [database<br />

name]. The interactive command interpreter can be launched with pgsql [data base name]


. While it is possible <strong>to</strong> build up a schema by entering commands one line at a time in<strong>to</strong><br />

the interpreter, the more common method is <strong>to</strong> use the interpreter's built-in command for<br />

importing a file containing SQL commands.<br />

7 All the major RDBMSs provide some level of built-in transaction management <strong>to</strong><br />

prevent concur rent users from executing incompatible commands.<br />

8 DBI-style modules exist for all the major scripting languages, including Python, Ruby,<br />

and Tel.<br />

9 Most of the major scripting languages have modules (usually intended <strong>to</strong> be used with<br />

the Apache web server) that allow the target interpreter <strong>to</strong> be used in the same manner.<br />

10 PHP does provide facilities for moving the bulk of the programming logic in<strong>to</strong><br />

separate libraries on the server – a facility that can at least help <strong>to</strong> minimize the effect of<br />

embedding code in web pages.<br />

11 The most famous hierarchical database system was undoubtedly IMS (Information<br />

Management System). This highly successful product, developed jointly by IBM and<br />

North American Rockwell in the late 1960s, was the dominant DBMS for commercial<br />

accounting and inven<strong>to</strong>ry for many years (Elmasri and Navanthe 1994). Its design – and<br />

even more so, the design of its query language, DL/1 – had a substantial influence on the<br />

development of later systems.<br />

References for Further Reading<br />

Bowman, J. S., et al. (1999). The Practical SQL Handbook. Reading, MA: Addison-<br />

Wesley.<br />

Codd, E. F. (1970). A Relational Model of Data for Large Shared Data Banks.<br />

Communications of the Association for Computing Machinery 13: 377–87.<br />

Date, C. J. (<strong>2001</strong>). The Database Relational Model: A Retrospective Review and<br />

Analysis. Reading: Addison-Wesley.<br />

Elmasri, R. and S. Navanthe (1994). Fundamentals of Database Systems. Redwood City:<br />

Benjamin/Cummings.<br />

Geschwinde, E. (<strong>2001</strong>). PostgreSQL Developer's Handbook. Indianapolis: SAMS.<br />

Halpin, T. (<strong>2001</strong>). Information Modeling and Relational Databases: From Conceptual<br />

Analysis <strong>to</strong> Logical Design. San Francisco: Morgan-Kaufmann.<br />

Hunter, L. (1990). Fact - Information - Data - Knowledge: Databases as a Way of<br />

Organizing Knowledge. Literary and Linguistic Computing 5: 49–57.


Postgresql. Vers. 7.2.3. Accessed April 22, 2004. At http://www.postgresql.org.<br />

Stinson, B. (<strong>2001</strong>). PostgreSQL Essential Reference. Indianapolis: New Riders.<br />

16.<br />

Marking Texts of Many Dimensions<br />

Jerome McGann<br />

Bring out number weight & measure in a year of death<br />

William Blake<br />

A sign is something by knowing which we know something more.<br />

C. S. Peirce<br />

Introduction: What is Text?<br />

Although "text" has been a "Keyword" in clerical and even popular discourse for more<br />

than fifty years, it did not find a place in Raymond Williams's important (1976) book<br />

Keywords. This strange omission may perhaps be explained by the word's cultural<br />

ubiquity and power. In that lexicon of modernity Williams called the "Vocabulary of<br />

Culture and Society", "text" has been the "one word <strong>to</strong> rule them all." Indeed, the word<br />

"text" became so shape-shifting and meaning-malleable that we should probably label it<br />

with Tolkien's full rubrication: "text" has been, and still is, the "one word <strong>to</strong> rule them all<br />

and in the darkness bind them."<br />

We want <strong>to</strong> keep in mind that general context when we address the issues of digitized<br />

texts, text markup, and electronic editing, which are our specialized concerns here. As we<br />

lay foundations for translating our inherited archive of cultural materials, including vast<br />

corpora of paper-based materials, in<strong>to</strong> <strong>digital</strong> deposi<strong>to</strong>ries and forms, we are called <strong>to</strong> a<br />

clarity of thought about textuality that most people, even most scholars, rarely undertake.<br />

Consider the phrase "marked text", for instance. How many recognize it as a redundancy?<br />

All text is marked text, as you may see by reflecting on the very text you are now<br />

reading. As you follow this conceptual exposition, watch the physical embodiments that<br />

shape the ideas and the process of thought. Do you see the typeface, do you recognize it?<br />

Does it mean anything <strong>to</strong> you, and if not, why not? Now scan away (as you keep reading)<br />

and take a quick measure of the general page layout: the font sizes, the characters per<br />

line, the lines per page, the leading, the headers, footers, margins. And there is so much<br />

more <strong>to</strong> be seen, registered, unders<strong>to</strong>od simply at the documentary level of your reading:<br />

paper, ink, book design, or the markup that controls not the documentary status of the text<br />

but its linguistic status. What would you be seeing and reading if I were addressing you


in Chinese, Arabic, Hebrew – even Spanish or German? What would you be seeing and<br />

reading if this text had been printed, like Shakespeare's sonnets, in 1609?<br />

We all know the ideal reader of these kinds of traditional documents. She is an actual<br />

person, like the texts this person reads and studies. He writes about her readings and<br />

studies under different names, including Randall McLeod, Randy Clod, Random Cloud,<br />

etc. She is the Dupin of the textual mysteries of our exquisite and sophisticated<br />

bibliographical age.<br />

Most important <strong>to</strong> realize, for this book's present purposes, is that <strong>digital</strong> markup schemes<br />

do not easily – perhaps do not even naturally - map <strong>to</strong> the markup that pervades paperbased<br />

texts. Certainly this is the case for every kind of electronic markup currently in use:<br />

from simple ASCII, <strong>to</strong> any inline SGML derivative, <strong>to</strong> the recent approaches of standoff<br />

markup (see Berrie website; Thompson and McKelvie 1997). The symp<strong>to</strong>ms of this<br />

discrepancy are exemplified in the AI (Artificial Intelligence) community's struggles <strong>to</strong><br />

simulate the complex processes of natural language and communicative exchange.<br />

Stymied of success in achieving that goal, these efforts have nonetheless been singularly<br />

fruitful for giving us a clearer view of the richness and flexibility of traditional textual<br />

machineries.<br />

How, then, are traditional texts marked? If we could give an exhaustive answer <strong>to</strong> that<br />

question we would be able <strong>to</strong> simulate them in <strong>digital</strong> forms. We cannot complete an<br />

answer for two related reasons: first, the answer would have <strong>to</strong> be framed from within the<br />

discourse field of textuality itself; and second, that framework is dynamic, a continually<br />

emerging function of its own operations, including its explicitly self-reflexive operations.<br />

This is not <strong>to</strong> say that markup and theories of markup must be "subjective." (It is also not<br />

<strong>to</strong> say – see below – that they must not be subjective.) It is <strong>to</strong> say that they are and must<br />

be social, his<strong>to</strong>rical, and dialectical, and that some forms have greater range and power<br />

than others, and that some are useful exactly because they seek <strong>to</strong> limit and restrict their<br />

range for certain special purposes.<br />

Au<strong>to</strong>poietic Systems and Co-dependency<br />

Describing the problems of electronic texts in her book Electronic Texts in the<br />

Humanities, Susan Hockey laconically observes that "There is no obvious unit of<br />

language" (2000: 20). Hockey is reflecting critically on the ordinary assumption that this<br />

unit is the word. Language scholars know better. Words can be usefully broken down in<strong>to</strong><br />

more primitive parts and therefore unders<strong>to</strong>od as constructs of a second or even higher<br />

order. The view is not unlike the one continually encountered by physicists who search<br />

out basic units of matter. Our analytic tradition inclines us <strong>to</strong> understand that forms of all<br />

kinds are "built up" from "smaller" and more primitive units, and hence <strong>to</strong> take the selfidentity<br />

and integrity of these parts, and the whole that they comprise, for objective<br />

reality.<br />

Hockey glances at this problem of the text-unit in order <strong>to</strong> clarify the difficulties of<br />

creating electronic texts. To achieve that, we instruct the computer <strong>to</strong> identify (the) basic


elements of natural language text and we try <strong>to</strong> ensure that the identification has no<br />

ambiguities. In natural language, however, the basic unit – indeed, all divisioning of any<br />

kind – is only procedurally determinate. The units are arbitrary. More, the arbitrary units<br />

themselves can have no absolute self-identity. Natural language is rife with redundancy<br />

and ambiguity at every unit and level and throughout its operating relations. A long<br />

his<strong>to</strong>ry of analytic procedures has evolved certain sets of best practices in the study of<br />

language and communicative action, but even in a short run, terms and relations of<br />

analysis have changed.<br />

Print and manuscript technology represent efforts <strong>to</strong> mark natural language so that it can<br />

be preserved and transmitted. It is a technology that constrains the shapeshiftings of<br />

language, which is itself a special-purpose system for coding human communication.<br />

Exactly the same can be said of electronic encoding systems. In each case constraints are<br />

installed in order <strong>to</strong> facilitate operations that would otherwise be difficult or impossible.<br />

In the case of a system like the Text Encoding Initiative (TEI), the system is designed <strong>to</strong><br />

"disambiguate" entirely the materials <strong>to</strong> be encoded.<br />

The output of TEI's markup constraints differs radically from the output generated by the<br />

constraints of manuscript and print technology. Whereas redundancy and ambiguity are<br />

expelled from TEI, they are preserved – are marked - in manuscript and print. While print<br />

and manuscript markups don't "copy" the redundancies of natural language, they do<br />

construct systems that are sufficiently robust <strong>to</strong> develop and generate equivalent types of<br />

redundancy. This capacity is what makes manuscript and print encoding systems so much<br />

more resourceful than any electronic encoding systems currently in use. ("Natural<br />

language" is the most complex and powerful reflexive coding system that we know of.) 1<br />

Like biological forms and all living systems, not least of all language itself, print and<br />

manuscript encoding systems are organized under a horizon of co-dependent relations.<br />

That is <strong>to</strong> say, print technology – I will henceforth use that term as shorthand for both<br />

print and manuscript technologies – is a system that codes (or simulates) what are known<br />

as au<strong>to</strong>poietic systems. These are classically described in the following terms:<br />

If one says that there is a machine M in which there is a feedback loop through the<br />

environment so that the effects of its output affect its input, one is in fact talking about a<br />

larger machine M 1 which includes the environment and the feedback loop in its defining<br />

organization.<br />

(Maturana and Varela 1980: 78)<br />

Such a system constitutes a closed <strong>to</strong>pological space that "continuously generates and<br />

specifies its own organization through its operation as a system of production of its own<br />

components, and does this in an endless turnover of components" (Maturana and Varela<br />

1980: 79). Au<strong>to</strong>poietic systems are thus distinguished from allopoietic systems, which are<br />

Cartesian and which "have as the product of their functioning something different from<br />

themselves" (1980: 80).


In this context, all coding systems appear <strong>to</strong> occupy a peculiar position. Because "coding<br />

… represents the interactions of [an] observer" with a given system, the mapping stands<br />

apart from "the observed domain" (Maturana and Varela 1980: 135). Coding is a function<br />

of "the space of human design" operations, or what is classically called "hetero-poietic"<br />

space. Positioned thus, coding and markup appear allopoietic.<br />

As machines of simulation, however, coding and markup (print or electronic) are not like<br />

most allopoietic systems (cars, flashlights, a road network, economics). Coding functions<br />

emerge as code only within an au<strong>to</strong>poietic system that has evolved those functions as<br />

essential <strong>to</strong> the maintenance of its life (its dynamic operations). Language and print<br />

technology (and electronic technology) are second- and third-order au<strong>to</strong>poietic systems –<br />

what McLuhan famously, expressively, if also somewhat mislead-ingly, called<br />

"extensions of man." Coding mechanisms – proteins, print technology – are generative<br />

components of the <strong>to</strong>pological space they serve <strong>to</strong> maintain. They are folded within the<br />

au<strong>to</strong>poietic system like membranes in living organisms, where distinct components<br />

realize and execute their extensions of themselves.<br />

This general frame of reference is what makes Maturana and Varela equate the "origin"<br />

of such systems with their "constitution" (1980: 95). This equation means that codependency<br />

pervades an au<strong>to</strong>poietic structure of relations. All components of the system<br />

arise (so <strong>to</strong> speak) simultaneously and they perform integrated functions. The system's<br />

life is a morphogenetic passage characterized by various dynamic mutations and<br />

transformations of the local system components. The purpose or goal of these processes<br />

is au<strong>to</strong>poietic – self-maintenance through self-transformation – and their basic element is<br />

not a system component but the relation (co-dependence) that holds the mutating<br />

components in changing states of dynamic stability. The states generate measurable codependency<br />

functions both in their periods (or basins) of stability and in their unique<br />

moments of catastrophic change.<br />

Marking the Text: A Necessary Distinction<br />

At the 2002 Extreme Markup Conference, Michael Sperberg-McQueen offered these<br />

observations on the problem of overlapping structures for SGML-based markup systems.<br />

It is an interesting problem because it is the biggest problem remaining in the residue. If<br />

we have a set of quantitative observations, and we try <strong>to</strong> fit a line <strong>to</strong> them, it is good<br />

practice <strong>to</strong> look systematically at the difference between the values predicted by our<br />

equation (our theory) and the values actually observed; the set of these differences is the<br />

residue …. In the context of SGML and XML, overlap is a residual problem. 2<br />

But in any context other than SGML and XML, this formulation is a play of wit, a kind<br />

of joke – as if one were now <strong>to</strong> say that the statistical deviations produced by New<strong>to</strong>nian<br />

mathematical calculations left a "residue" of "interesting" matters <strong>to</strong> be cleared up by<br />

further, deeper calculations. But those matters are not residual, they are the hem of a<br />

quantum garment.


My own comparison is itself a kind of joke, of course, for an SGML model of the world<br />

of textualities pales in comprehensiveness before the New<strong>to</strong>nian model of the physical<br />

world. But the outrageousness of the comparison in each case helps <strong>to</strong> clarify the<br />

situation. No au<strong>to</strong>poietic process or form can be simulated under the horizon of a<br />

structural model like SGML, not even <strong>to</strong>pic maps. We see this very clearly when we<br />

observe the inability of a derivative model like TEI <strong>to</strong> render the forms and functions of<br />

traditional textual documents. The latter, which deploy markup codes themselves, supply<br />

us with simulations of language as well as of many other kinds of semeiotic processes, as<br />

Peirce called them. Textualized documents restrict and modify, for various kinds of<br />

reflexive purposes, the larger semeiotic field in which they participate. Nonetheless, the<br />

procedural constraints that traditional textualities lay upon the larger semeiotic field that<br />

they model and simulate are far more pragmatic, in a full Peircean sense, than the<br />

electronic models that we are currently deploying.<br />

Understanding how traditional textual devices function is especially important now when<br />

we are trying <strong>to</strong> imagine how <strong>to</strong> optimize our new <strong>digital</strong> <strong>to</strong>ols. Manuscript and print<br />

technologies – graphical design in general – provide arresting models for information<br />

technology <strong>to</strong>ols, especially in the context of traditional <strong>humanities</strong> research and<br />

education needs. To that end we may usefully begin by making an elementary distinction<br />

between the archiving and the simulating functions of textual (and, in general, semeiotic)<br />

systems. Like gene codes, traditional textualities possess the following as one of their<br />

essential characteristics: that as part of their simulation and generative processes, they<br />

make (of) themselves a record of those processes. Simulating and record keeping, which<br />

are co-dependent features of any au<strong>to</strong>poietic or semeiotic system, can be distinguished<br />

for various reasons and purposes. A library processes traditional texts by treating them<br />

strictly as records. It saves things and makes them accessible. A poem, by contrast,<br />

processes textual records as a field of dynamic simulations. The one is a machine of<br />

memory and information, the other a machine of creation and reflection. Each may be<br />

taken as an index of a polarity that characterizes all semeoitic or au<strong>to</strong>poietic systems.<br />

Most texts – for instance, this chapter you are reading now – are fields that draw upon the<br />

influence of both of those polarities.<br />

The power of traditional textualities lies exactly in their ability <strong>to</strong> integrate those different<br />

functions within the same set of coding elements and procedures.<br />

SGML and its derivatives are largely, if not strictly, coding systems for s<strong>to</strong>ring and<br />

accessing records. They possess as well certain analytic functions that are based in the<br />

premise that text is an "ordered hierarchy of context objects." This conception of<br />

textuality is plainly non-comprehensive. Indeed, its specialized understanding of "text"<br />

reflects the pragmatic goal of such a markup code: <strong>to</strong> s<strong>to</strong>re objects (in the case of TEI,<br />

textual objects) so that they can be quickly accessed and searched for their informational<br />

content – or more strictly, for certain parts of that informational content (the parts that fall<br />

in<strong>to</strong> a hierarchical order modeled on a linguistic analysis of the structure of a book).<br />

These limitations of electronic markup codes are not <strong>to</strong> be lamented, but for humanist<br />

scholars they are <strong>to</strong> be clearly unders<strong>to</strong>od. A markup code like TEI creates a record of a


traditional text in a certain form. Especially important <strong>to</strong> see is that, unlike the textual<br />

fields it was designed <strong>to</strong> mark up, TEI is an allopoietic system. Its elements are<br />

unambiguously delimited and identified a priori, its structure of relations is precisely<br />

fixed, it is non-dynamical, and it is focused on objects that stand apart from itself. Indeed,<br />

it defines what it marks not only as objective, but as objective in exactly the<br />

unambiguous terms of the system's a priori categories. This kind of machinery will<br />

therefore serve only certain, very specific, purposes. The au<strong>to</strong>poietic operations of textual<br />

fields – operations especially pertinent <strong>to</strong> the texts that interest <strong>humanities</strong> scholars – lie<br />

completely outside the range of an order like the TEI.<br />

For certain archival purposes, then, structured markup will serve. It does not unduly<br />

interfere with, or forbid implementing, some of the searching and linking capacities that<br />

make <strong>digital</strong> technology so useful for different types of comparative analysis. Its strict<br />

formality is abstract enough <strong>to</strong> permit implementation within higher-order formalizations.<br />

In these respects it has greater flexibility than a stand-off approach <strong>to</strong> text markup, which<br />

is more difficult <strong>to</strong> integrate in<strong>to</strong> a dispersed online network of different kinds of<br />

materials. All that having been recognized and said, however, these allopoietic textprocess<br />

ing systems cannot access or display the au<strong>to</strong>poietic character of textual fields.<br />

Digital <strong>to</strong>ols have yet <strong>to</strong> develop models for displaying and replicating the self-reflexive<br />

operations of bibliographical <strong>to</strong>ols, which alone are operations for thinking and<br />

communicating – which is <strong>to</strong> say, for transforming data in<strong>to</strong> knowledge.<br />

We have <strong>to</strong> design and build <strong>digital</strong> environments for those purposes. A measure of their<br />

capacity and realization will be whether they can integrate data-function mechanisms like<br />

TEI in<strong>to</strong> their higher-order operations. To achieve that will entail, I believe, the<br />

deployment of dynamic, <strong>to</strong>pological models for mapping the space of <strong>digital</strong> operations.<br />

But these models will have <strong>to</strong> be reconceived, as one can see by reflecting on a remark<br />

about textual interpretation that Stanley Fish liked <strong>to</strong> make years ago. He would point out<br />

that he was able <strong>to</strong> treat even the simplest text – road signage, for example – as a poem<br />

and thus develop from his own "response" and commentary its au<strong>to</strong>poietic potential. The<br />

remark underscores a basic and almost entirely neglected (undertheorized) feature of<br />

discourse fields: that <strong>to</strong> "read" them – <strong>to</strong> read "in" them at any point – one must regard<br />

what we call "the text" and "the reader" as co-dependent agents in the field. You can't<br />

have one without the other.<br />

Fish's observation, therefore, while true, signals a widespread theoretical and<br />

methodological weakness in our conceptions of textuality, traditional or otherwise. This<br />

approach figures "text" as a heuristic abstraction drawn from the larger field of discourse.<br />

The word "text" is used in various ways by different people – Barthes's understanding is<br />

not the same as a TEI understanding – but in any case the term frames attention on the<br />

linguistic dimension of a discourse field. Books and literary works, however, organize<br />

themselves along multiple dimensions of which the linguistic is only one.<br />

Modeling <strong>digital</strong> simulations of a discourse field requires that a formal set of dimensions<br />

be specified for the field. This is what TEI provides a priori, though the provision, as we<br />

know, is minimal. Our received scholarly traditions have in fact passed down <strong>to</strong> us an


understanding of such fields that is both far more complex and reasonably stable.<br />

Discourse fields, our textual condition, regularly get mapped along six dimensions (see<br />

below, and Appendix B). Most important of all in the present context, however, are the<br />

implications of cognizing a discourse field as au<strong>to</strong>poietic. In that case the field<br />

measurements will be taken by "observers" positioned within the field itself. That<br />

intramural location of the field interpreter is in truth a logical consequence of the codependent<br />

character of the field and its components. "Interpretation" is not undertaken<br />

from a position outside the field; it is an essential part of a field's emergence and of any<br />

state that its emergence might assume.<br />

This matter is crucial <strong>to</strong> understand when we are reaching for an adequate formalizing<br />

process for textual events like poetry or other types of orderly but discontinuous<br />

phenomena. Rene Thorn explains very clearly why <strong>to</strong>pological models are preferable <strong>to</strong><br />

linear ones in dynamic systems:<br />

it must not be thought that a linear structure is necessary for s<strong>to</strong>ring or transmitting<br />

information (or, more precisely, significance); it is possible that a language, a semantic<br />

model, consisting of <strong>to</strong>pological forms could have considerable advantages from the<br />

point of view of deduction, over the linear language that we use, although this idea is<br />

unfamiliar <strong>to</strong> us. Topological forms lend themselves <strong>to</strong> a much richer range of<br />

combinations…than the mere juxtaposition of two linear sequences.<br />

(Thorn 1975: 145)<br />

These comments distinctly recall Peirce's exploration of existential graphs as sites of<br />

logical thinking. But Thorn's presentation of <strong>to</strong>pological models does not conceive field<br />

spaces that are au<strong>to</strong>poietic, which seems <strong>to</strong> have been Peirce's view. Although Thorn's<br />

approach generally eschews practical considerations in favor of theoretical clarity, his<br />

models assume that they will operate on data carried in<strong>to</strong> the system from some external<br />

source. If Thorn's "data" come in<strong>to</strong> his studies in a theoretical form, then, they have been<br />

theorized in traditional empirical terms. The <strong>to</strong>pological model of a s<strong>to</strong>rm may therefore<br />

be taken either as the description of the s<strong>to</strong>rm and/or a prediction of its future behavior.<br />

But when a model's data are taken <strong>to</strong> arise co-dependently with all the other components<br />

of its system, a very different "result" ensues. Imagined as applied <strong>to</strong> textual au<strong>to</strong>poiesis,<br />

a <strong>to</strong>pological approach carries itself past an analytic description or prediction over <strong>to</strong> a<br />

form of demonstration or enactment.<br />

The view taken here is that no textual field can exist as such without "including" in itself<br />

the reading or measurement of the field, which specifies the field's dataset from within.<br />

The composition of a poem is the work's first reading, which in that event makes a call<br />

upon others. An extrinsic analysis designed <strong>to</strong> specify or locate a poetic field's selfreflexiveness<br />

commonly begins from the vantage of the rhe<strong>to</strong>rical or the social dimension<br />

of the text, where the field's human agencies (efficient causes) are most apparent. The<br />

past century's fascination with structuralist approaches <strong>to</strong> cultural phenomena produced,<br />

as we know, a host of analytic procedures that chose <strong>to</strong> begin from a consideration of


formal causation, and hence from either a linguistic or a semiotic vantage. Both<br />

procedures are analytic conventions based in empirical models.<br />

Traditional textuality provides us with au<strong>to</strong>poietic models that have been engineered as<br />

effective analytic <strong>to</strong>ols. The codex is the greatest and most famous of these. Our problem<br />

is imagining ways <strong>to</strong> recede them for <strong>digital</strong> space. To do that we have <strong>to</strong> conceive<br />

formal models for au<strong>to</strong>poietic processes that can be written as computer software<br />

programs.<br />

Field Au<strong>to</strong>poiesis: from IVANHOE <strong>to</strong> 'Patacriticism<br />

Let's recapitulate the differences between book markup and TEI markup. TEI defines<br />

itself as a two-dimensional generative space mapped as (1) a set of defined "content<br />

objects" (2) organized within a nested tree structure. The formality is clearly derived<br />

from an elementary structuralist model of language (a vocabulary + a syntax, or a<br />

semantic + a syntagmatic dimension). In the SGML/TEI extrusion, both dimensions are<br />

fixed and their relation <strong>to</strong> each other is defined as arbitrary rather than co-dependent. The<br />

output of such a system is thus necessarily symmetrical with the input (cf. Curie's<br />

principle of causes and effects). Input and output in a field of traditional textuality works<br />

differently. Even in quite restricted views, as we know, the operations of natural language<br />

and communicative exchange generate incommensurable effects. The operations exhibit<br />

behavior that <strong>to</strong>polo-gists track as bifurcation or even generalized catastrophe, whereby<br />

an initial set of structural stabilities produces morphogenetic behaviors and conditions<br />

that are unpredictable. 3 This essential feature of "natural language" – which is <strong>to</strong> say, of<br />

the discourse fields of communicative exchange – is what makes it so powerful, on one<br />

hand, and so difficult <strong>to</strong> model and formalize, on the other.<br />

In these circumstances, models like TEI commend themselves <strong>to</strong> us because they can be<br />

classically quantified for empirical – numerable – results. But as Thorn observed long<br />

ago, there is no such thing as "a quantitative theory of catastrophes of a dynamical<br />

system" like natural language. To achieve such a theory, he went on <strong>to</strong> say, "it would be<br />

necessary <strong>to</strong> have a good theory of integration on function spaces" (Thorn 1975: 321),<br />

something that Thorn could not conceive.<br />

That limitation of qualitative mathematical models did not prevent Thorn from<br />

vigorously recommending their study and exploration. He particularly criticized the<br />

widespread scientific habit of "tak[ing] the main divisions of science, the[ir] taxonomy…<br />

as given a priori" rather than trying <strong>to</strong> re-theorize taxonomies as such (1975: 322). In this<br />

frame of reference we can see (1) that textualization in print technology is a qualitative<br />

(rather than a taxonomic) function of natural language, and (2) that textualization<br />

integrates function spaces through demonstrations and enactments rather than<br />

descriptions. This crucial understanding – that print textuality is not language but an<br />

operational (praxis-based) theory of language – has stared us in the face for a long time,<br />

but seeing we have not seen. It has taken the emergence of electronic textualities, and in<br />

particular operational theories of natural language like TEI, <strong>to</strong> expose the deeper truth<br />

about print and manuscript texts. SGML and its derivatives freeze (rather than integrate)


the function spaces of discourse fields by reducing the field components <strong>to</strong> abstract forms<br />

– what Coleridge called "fixities and definites." This approach will serve when the object<br />

is <strong>to</strong> mark textual fields for s<strong>to</strong>rage and access.<br />

Integration of dynamic functions will not emerge through such abstract reductions,<br />

however. To develop an effective model of an au<strong>to</strong>poietic system requires an analysis<br />

that is built and executed "in the same spirit that the author writ." That formulation by<br />

Alexander Pope expresses, in an older dialect, what we have called in this century "the<br />

uncertainty principle", or the co-dependent relation between measurements and<br />

phenomena. An agent defines and interprets a system from within the system itself – at<br />

what Dante Gabriel Rossetti called "an inner standing point." What we call "scientific<br />

objectivity" is in one sense a mathematical function; in another, it is a useful method for<br />

controlling variables. We use it when we study texts as if they were objective things<br />

rather than dynamic au<strong>to</strong>poietic fields.<br />

Traditional textual conditions facilitate textual study at an inner standing point because<br />

all the activities can be carried out – can be represented – in the same field space,<br />

typically, in a bibliographical field. Subject and object meet and interact in the same<br />

dimensional space – a situation that gets reified for us when we read books or write about<br />

them. Digital operations, however, introduce a new and more abstract space of relations<br />

in<strong>to</strong> the study-field of textuality. This abstract space brings the possibility of new and in<br />

certain respects greater analytic power <strong>to</strong> the study of traditional texts. On the downside,<br />

however, digitization – at least <strong>to</strong> date, and typically – situates the critical agent outside<br />

the field <strong>to</strong> be mapped and re-displayed. Or – <strong>to</strong> put this crucial point more precisely<br />

(since no measurement has anything more than a relative condition of objectivity) –<br />

digitization situates the critical agent within levels of the textual field's dimensionalities<br />

that are difficult <strong>to</strong> formalize bibliographically.<br />

To exploit the power of those new formalizations, a <strong>digital</strong> environment has <strong>to</strong> expose its<br />

subjective status and operation. (Like all scientific formalities, <strong>digital</strong> procedures are<br />

"objective" only in relative terms.) In the present case – the <strong>digital</strong> marking of textual<br />

fields – this means that we will want <strong>to</strong> build <strong>to</strong>ols that foreground the subjectivity of any<br />

measurements that are taken and displayed. Only in this way will the au<strong>to</strong>poietic<br />

character of the textual field be accurately realized. The great gain that comes with such a<br />

<strong>to</strong>ol is the ability <strong>to</strong> specify – <strong>to</strong> measure, display, and eventually <strong>to</strong> compute and<br />

transform – an au<strong>to</strong>poietic structure at what would be, in effect, quantum levels.<br />

A series of related projects <strong>to</strong> develop such <strong>to</strong>ols is under way at University of Virginia's<br />

Speculative Computing Labora<strong>to</strong>ry (Speclab). The first of these, IVANHOE, is an online<br />

gamespace being built for the imaginative reconstruction of traditional texts and<br />

discourse fields. Players enter these works through a <strong>digital</strong> display space that encourages<br />

players <strong>to</strong> alter and transform the textual field. The game rules require that<br />

transformations be made as part of a discourse field that emerges dynamically through<br />

the changes made <strong>to</strong> a specified initial set of materials. 4


As the IVANHOE project was going forward, a second, related project called Time<br />

Modelling was being taken up by Bethany Nowviskie and Johanna Drucker. The project<br />

was begun "<strong>to</strong> bring visualization and interface design in<strong>to</strong> the early content modeling<br />

phase" of projects like IVANHOE, which pursue interpretation through transformational<br />

and even deformative interactions with the primary data. IVANHOE's computer is<br />

designed <strong>to</strong> s<strong>to</strong>re the game players' performative interpretational moves and then produce<br />

algorithmically generated analyses of the moves after the fact. The chief critical function<br />

thus emerges after-the-fact, in a set of human reflections on the differential patterns that<br />

the computerized analyses expose. In the Time Modelling device, however, the<br />

performative and the critical actions are much more closely integrated because the human<br />

is actively involved in a deliberated set of <strong>digital</strong> transformations. The Time Modelling<br />

device gives users a set of design functions for reconstructing a given lineated timeline of<br />

events in terms that are subjective and hypothetical. The specified field of event-related<br />

data is brought forward for transformation through editing and display mechanisms that<br />

emphasize the malleability of the initial set of field relations. The project stands,<br />

conceptually, somewhere between design programs (with their sets of <strong>to</strong>ols for making<br />

things) and complex websites like The Rossetti Archive (with their hypertextual datasets<br />

organized for on-the-fly search and analysis). It is a set of editing and display <strong>to</strong>ols that<br />

allows users <strong>to</strong> design their own hypothetical (re)formulations of a given dataset.<br />

The frankly experimental character of Time Modelling's data (re)constructions has led <strong>to</strong><br />

an important reimagining of the original IVANHOE project. From the outset of that<br />

project we intended <strong>to</strong> situate the "interpreter" within the discourse field that was the<br />

subject of interpretive transformation. Our initial conception was <strong>to</strong>ward what we called<br />

"Ultimate IVANHOE", that is, <strong>to</strong>ward a playspace that would be controlled by emergent<br />

consciousness software. With the computer an active agent in an IVANHOE session,<br />

players could measure and compare their own understandings of their actions against a<br />

set of computer generated views. This prospect for IVANHOE's development remains,<br />

but the example of Time Modelling exposed another way <strong>to</strong> situate the human interpreter<br />

at an inner standing point of an au<strong>to</strong>poietic system.<br />

If'Pataphysics is, in the words of its origina<strong>to</strong>r, "the science of exceptions", the project<br />

here is <strong>to</strong> reconceive IVANHOE under the rubric of'Patacriticism, or the theory of<br />

subjective interpretation. The theory is implemented through what is here called the<br />

dementianal method, which is a procedure for marking the au<strong>to</strong>poietic features of textual<br />

fields. The method works on the assumption that such features characterize what <strong>to</strong>pologists<br />

call a field of general catastrophe. The dementianal method marks the dynamic<br />

changes in au<strong>to</strong>poietic fields much as Thorn's <strong>to</strong>pological models allow one <strong>to</strong> map forms<br />

of catastrophic behavior. The'Patacritical model differs from Thorn's models because the<br />

measurements of the au<strong>to</strong>poietic field's behaviors are generated from within the field<br />

itself, which only emerges as a field through the action of the person interpreting – that is<br />

<strong>to</strong> say, marking and displaying – the field's elements and sets of relations. The field arises<br />

co-dependently with the acts that mark and measure it. In this respect we wish <strong>to</strong><br />

characterize its structure as dementianal rather than dimensional.


As the device is presently conceived, readers engage au<strong>to</strong>poietic fields along three<br />

behavior dementians: transaction, connection, resonance. A common transaction of a<br />

page space moves diagonally down the page, with regular deviations for horizontal line<br />

transactions left <strong>to</strong> right margin, from the <strong>to</strong>p or upper left <strong>to</strong> the bot<strong>to</strong>m at lower right.<br />

Readers regularly violate that pattern in indefinite numbers of ways, often being called <strong>to</strong><br />

deviance by how the field appears marked by earlier agencies. Connections assume, in<br />

the same way, multiple forms. Indeed, the primal act of au<strong>to</strong>poietic connection is the<br />

identification or location of a textual element <strong>to</strong> be "read." In this sense, the transaction<br />

of an au<strong>to</strong>poietic field is a function of the marking of connections of various kinds, on<br />

one hand, and of resonances on the other. Resonances are signals that call attention <strong>to</strong> a<br />

textual element as having a field value – a potential for connectivity – that appears and<br />

appears unrealized.<br />

Note that each of these behavior dementians exhibit co-dependent relations. The field is<br />

transacted as connections and resonances are marked; the connections and resonances are<br />

continually emergent functions of each other; and the marking of dementians<br />

immediately reorders the space of the field, which itself keeps re-emerging under the sign<br />

of the marked alteration of the dynamic fieldspace and its various elements.<br />

These behavioral dementians locate an au<strong>to</strong>poietic syntax, which is based in an<br />

elementary act or agenting event: G. Spencer Brown's "law of calling", which declares<br />

that a distinction can be made. From that law comes the possibility that elements of<br />

identities can be defined. They emerge with the co-dependent emergence of the textual<br />

field's control dimensions, which are the field's au<strong>to</strong>poietic semantics. (For further<br />

discussion of these matters see below, Appendix A and Appendix B.)<br />

Writing and Reading in Au<strong>to</strong>poietic Fields<br />

This'Patacritical approach <strong>to</strong> textual dementians is a meta-theory of textual fields, a<br />

pragmatistic conception of how <strong>to</strong> expose discontinuous textual behaviors ("natural<br />

language" so called, or what Habermas (1984) has better called "communicative action").<br />

Integration of the dynamic functions begins not by abstracting the theory away from a<br />

target object – that is, the method of a taxonomic methodology – but by integrating the<br />

meta-theoretical functions within the discourse space itself.<br />

Informational discourse fields function well precisely by working <strong>to</strong> limit redundancy<br />

and concurrent textual relations. Because poetry – or imaginative textuality broadly<br />

conceived – postulates much greater freedom of expressive exchange, it exhibits a special<br />

attraction for anyone wishing <strong>to</strong> study the dynamics of textuality. Aris<strong>to</strong>tle's studies of<br />

semiotic systems preserve their foundational character because they direct their attention<br />

<strong>to</strong> au<strong>to</strong>poietic rather than allopoietic discourse fields. His studies pursue a taxonomy for<br />

the dynamic process of making and exchanging (remaking) simulations.<br />

Pla<strong>to</strong>'s dialogues, by contrast, situate – or, more precisely, generate – their critical<br />

reflections at a standing point inside the textualities they are themselves unfolding. In this<br />

respect they have much in common with Wittgenstein's critical colloquies in the


Philosophical Investigations or with Montaigne's Essais. But the dynamic play of even<br />

these textual fields remain, from the point of view of their readers, exemplary exercises.<br />

This situation prevails in all modes of critical reflection which assume <strong>to</strong> preserve the<br />

integrity and self-identity of the textual fields they study. Two forms of critical reflection<br />

regularly violate the sanctity of such self-contained textual spaces: translation and<br />

editing. The idea that an object of criticism like a textual field is an object can be<br />

maintained either as an heuristic procedure or as an on<strong>to</strong>logical illusion. Consequently,<br />

acts of translation and editing are especially useful forms of critical reflection because<br />

they so clearly invade and change their subjects in material ways. To undertake either,<br />

you can scarcely not realize the performative – even the deformative - character of your<br />

critical agency.<br />

At this point let me exemplify the general markup model for au<strong>to</strong>poietic textualities. This<br />

comes as the following hypothetical passage through an early poem by Robert Creeley,<br />

"The Innocence." Because imaginative textuality is, in this view, an exemplary kind of<br />

au<strong>to</strong>poietic process, any poetical work would do for a narrative demonstration. I choose<br />

"The Innocence" because it illustrates what Creeley and others called "field poetics." As<br />

such, it is especially apt for clarifying the conception of the au<strong>to</strong>poietic model of<br />

textuality being offered here. "Composition by field" poetics has been much discussed,<br />

but for present purposes it suffices <strong>to</strong> say that it conceives poetry as a self-unfolding<br />

discourse. "The poem" is the "field" of action and energy generated in the poetic<br />

transaction of the field that the poem itself exhibits. "Composition by field", whose<br />

theoretical foundations may be usefully studied through Charles Olson's engagements<br />

with contemporary philosophy and science, comprised both a method for understanding<br />

(rethinking) the entire inheritance of poetry, and a program for contemporary and future<br />

poetic discourse (its writing and its reading).<br />

The text chosen is taken from Donald Allen's famous anthology (first published in 1960)<br />

The New American Poetry in its 1999 University of California Press reprinting.<br />

The Innocence<br />

Looking <strong>to</strong> the sea, it is a line<br />

of unbroken mountains.<br />

It is the sky.<br />

It is the ground. There<br />

we live, on it.<br />

It is a mist<br />

now tangent <strong>to</strong> another<br />

quiet. Here the leaves<br />

come, there<br />

is the rock in evidence


or evidence.<br />

What I come <strong>to</strong> do<br />

is partial, partially kept<br />

Before tracing a model for this poetic field we want <strong>to</strong> bear two matters in mind. First,<br />

the field we are transacting is localized in relation <strong>to</strong> this documentary instance of "the<br />

text." One of the most persistent and misleading procedures in traditional hermeneutics is<br />

<strong>to</strong> take the object of study as something not only abstract and disembodied, but as<br />

something lying outside the field space – itself specific and material – of the act of<br />

critical representation. Second, the sequence of readings (below) consciously assumes a<br />

set of previous readings whereby certain elementary forms of order – by no means<br />

insignificant forms – have been integrated in<strong>to</strong> the respective textual dementians. All<br />

such forms are extrusions from the elementary semiotic move, which is Spencer Brown's<br />

basic law of form: that a distinction can be drawn (as a dementian, or within and between<br />

dementians). Thus the readings below assume that each dementian is oriented <strong>to</strong> a set of<br />

established formal objects which get called and then crossed (transformed) in the<br />

transaction of the field.<br />

That said, let me transact the poetic field through the initial textual model supplied above.<br />

A First Reading:<br />

I mark the following elements in the first line group (and in that act I mark as well the<br />

presence of (a) lines and (b) line groups): "Looking" as a dangling participle; "it" (line 1)<br />

as ambiguously pronominal; "line" as a word play referencing (first) this line of verse I<br />

am transacting, and (second) a landscape of "unbroken mountains" (<strong>to</strong> be marked as such<br />

only with the marking of the final line in the group). All of these are defined (connected<br />

<strong>to</strong> the fieldspace) as textual elements with marked resonances (anticipations and clear if<br />

inchoate recollections) as well as several manifest, second-order connections (e.g., "sea",<br />

"line", and "mountains" as objects in a landscape).<br />

Line group 2 emerges <strong>to</strong> connect a network of "it" words as well as <strong>to</strong> settle the<br />

dominance of a linguistic gravity field centered in the initially marked "landscape" (a<br />

linguistic dementian subdomain). As line group 3 continues <strong>to</strong> elaborate the "landscape<br />

field", several distinctly new elements emerge and get marked. They center in the words<br />

"tangent", "quiet", "evidence", the notable enjambment at the end of the line group, and<br />

the deictics "Here and "there." The first four resonate by the differences they make with<br />

the previous elements I had defined in my transaction of the field. The deictics connect<br />

back <strong>to</strong> the second linguistic dementian subdomain (the self-reflexive set of textual<br />

elements marked in line 1 as the dangling participle and the final word "line"). The fourth<br />

and last line group is itself marked as strongly resonant in itself because of the emergence<br />

within it of the unique "I" and the startling repetitions ("evidence", "partial"/ "partially").<br />

So the field transaction is marked geometrically as a complete and continuous passage<br />

from upper left <strong>to</strong> lower right and proceeding line by line left <strong>to</strong> right. That passage of the<br />

textspace marks out two control dementians, linguistic and graphical, as well as several


distinct basins of order within them. In the graphical dementian we see an array of<br />

marked words, lines, and line groups. In the linguistic dementian I have marked two<br />

distinct subdomains, one referential (the set of "landscape" semantics), one a subdomain<br />

of pure signifiers (proliferating from line 1 through the deictic markers "Here" and<br />

"there."<br />

A Second Reading:<br />

I mark the title as strongly resonant and I find myself scanning the poem rather than<br />

reading it linearly, and marking elements unnoticed in the first reading. I re-mark the<br />

array of "it" words and connect all of them <strong>to</strong> the title, marking thereby another linguistic<br />

subdomain. I mark as resonant the striking idea of "a mist / now tangent <strong>to</strong> another /<br />

quiet", and I mark a distinction in the linguistic subdomain (of "landscape") between<br />

different sensory aspects of a "landscape." I mark as resonant the equally striking final<br />

sentence and the phrase "the rock in evidence / or evidence."<br />

A Third Reading:<br />

This is a sequential transaction through the poem as in the first reading. It is largely<br />

devoted <strong>to</strong> marking connections between the various elements already marked with<br />

resonance values. The wordplay in "line" is marked as a strongly resonant locus of<br />

fieldspace connections across the several linguistic subdomains. This connective<br />

fieldspace is especially resonant as the connection between the words "line" and<br />

"tangent." I mark all of the previously marked textual elements as connected <strong>to</strong> each other<br />

in a broadly dispersed semiotic dementian because I am seeing that elements in different<br />

fieldspace dementians and domains (e.g., "mist" and "quiet") are connected <strong>to</strong> each other.<br />

A Fourth Reading:<br />

A sequential reading leads <strong>to</strong> marking the final sentence as a dramatic locus of a<br />

rhe<strong>to</strong>rical dementian in the fieldspace. The construction of the textspace is "What I come<br />

<strong>to</strong> do." The emergence of this idea allows me <strong>to</strong> mark the poem as a deliberated<br />

sequential organization that exposes itself in certain telling (marked) moments and<br />

textual elements: "Looking", "line", "tangent", the deictic words, the previously<br />

unmarked "we" (line 5), the enjambment between the third and fourth line groups. In all<br />

these I mark a rhe<strong>to</strong>rical organization tied most centrally <strong>to</strong> the phrase "What I come <strong>to</strong><br />

do." I mark that these marks unfold as a relation that must be seen as sequenced: "I" in<br />

the present tense here is always the present tense in the linguistic dementian of this work.<br />

Marking the verb tense in that way immediately produces the first, remarkable emergence<br />

in this reading process of the work's social dementian. "I" comes <strong>to</strong> write this poem,<br />

which is marked thereby as an event in the world and as objective as any material thing<br />

(these material things, the "landscape" things, first marked in the linguistic dementian). In<br />

that rhe<strong>to</strong>rical dementian I mark as well a key element of this work's social dementian<br />

first marked in the linguistic dementian: the relation between the "we" and the "I." The<br />

phrase "is partial, partially kept" is marked now as an element in the social dementian of<br />

the textspace – as if one were <strong>to</strong> say, interpretatively, that the "doing" of the poem is only


one event in a larger field that the poem is part of and points <strong>to</strong>ward. My acts of marking<br />

the poem fall in<strong>to</strong> both the local fieldspace and the larger discourse field marked by this<br />

local poetical field. And I further mark the social space by connecting the textspace <strong>to</strong> the<br />

book in which the text is printed – for that book (the polemic it made) marks this specific<br />

text in the strongest way. At this point the sixth dementian of the fieldspace begins <strong>to</strong> get<br />

marked, the material dementian. I mark three documentary features in particular: the<br />

placement of the text in the book, the organ of publication, the date of publication. I mark<br />

as well the fact that these material features of the work are, like the word "line", doublemeaninged<br />

(or double dementianed), having as well a clear placement in the work's social<br />

dementian as well.<br />

A Fifth Reading:<br />

I mark new elements in the six marked dementians that emerge in a widespread process<br />

of subdividing and proliferating. Elements defined in one dementian or subdomain get<br />

marked in another (for instance, "I" began in the rhe<strong>to</strong>rical, reappeared in the social, and<br />

now gets marked in all the other dementians as well); unmarked textual features, like the<br />

letter "t", get marked as resonant; the shape of the textspace from word <strong>to</strong> line <strong>to</strong> word<br />

group is marked as a linked set of spare elements. These additional markings lead <strong>to</strong><br />

other, previously unseen and unmarked relations and elements. The spare graphical<br />

dementian gets linked <strong>to</strong> the linguistic dementian ("The Innocence") and <strong>to</strong> the social and<br />

rhe<strong>to</strong>rical dementians (the graphical spareness is only markable in relation <strong>to</strong> the<br />

absent/present discourse field in which this poetical work stands and declares its<br />

comparative allegiance.<br />

A Sixth Reading:<br />

This is a reading that poses significant theoretical and practical issues. Time-stamped two<br />

weeks after the previous readings, this reading was negotiated in my mind as I recalled<br />

the his<strong>to</strong>ry of my readings of the poem. It is thus a reading <strong>to</strong> be <strong>digital</strong>ly marked afterthe-fact.<br />

Focused on the final line group, it also marks the entirety of the au<strong>to</strong>poietic field.<br />

The reading marks the "I" as a figure in the social dementian, the poet (Creeley) who<br />

composed the poem. In that linking, however, I as reader become linked <strong>to</strong> the linguistic<br />

"I" that is also a social "I." This linkage gets enforced by marking a set of "partial" agents<br />

who "come <strong>to</strong> do" part of the continuous marking of the au<strong>to</strong>poietic field. (Creeley does<br />

what he does, I do what I do, and we both inhabit a space resonant with other, as yet<br />

unspecified, agents.)<br />

Conclusion<br />

What we theorize here and propose for a <strong>digital</strong> practice is a science of exceptions, a<br />

science of imaginary (subjective) solutions. The markup technology of the codex has<br />

evolved an exceedingly successful instrument for that purpose. Digital technology ought<br />

<strong>to</strong> be similarly developed. Organizing our received <strong>humanities</strong> materials as if they were<br />

simply information deposi<strong>to</strong>ries, computer markup as currently imagined handicaps or<br />

even baffles al<strong>to</strong>gether our moves <strong>to</strong> engage with the well-known dynamic functions of


textual works. An alternative approach <strong>to</strong> these matters through a formal reconception of<br />

textspace as <strong>to</strong>pological offers distinct advantages. Because this space is au<strong>to</strong>poietic,<br />

however, it does not have what mathematicians would normally call dimensionality. As<br />

au<strong>to</strong>poietic, the model we propose establishes and measures its own dimensions au<strong>to</strong>telically,<br />

as part of its self-generative processes. Furthermore, space defined by pervasive<br />

co-dependencies means that any dimension specified for the system might be formally<br />

related <strong>to</strong> any other. This metamorphic capacity is what translates the concept of a<br />

dimension in<strong>to</strong> the concept of a dementian.<br />

This model of text-processing is open-ended, discontinuous, and non-hierarchical. It<br />

takes place in a fieldspace that is exposed when it is mapped by a process of "reading." A<br />

<strong>digital</strong> processing program is <strong>to</strong> be imagined and built that allows one <strong>to</strong> mark and s<strong>to</strong>re<br />

these maps of the textual fields and then <strong>to</strong> study the ways they develop and unfold and<br />

how they compare with other textual mappings and transactions. Constructing textualities<br />

as field spaces of these kinds short-circuits a number of critical predilections that inhibit<br />

our received, commonsense wisdom about our textual condition. First of all, it escapes<br />

crippling interpretative dicho<strong>to</strong>mies like text and reader, or textual "subjectivity" and<br />

"objectivity." Reader-response criticism, so-called, intervened in that space of problems<br />

but only succeeded in reifying even further the primary distinctions. In this view of the<br />

matter, however, one sees that the distinctions are purely heuristic. The "text" we "read"<br />

is, in this view, an au<strong>to</strong>poietic event with which we interact and <strong>to</strong> which we make our<br />

own contributions. Every textual event is an emergence imbedded in and comprising a set<br />

of complex his<strong>to</strong>ries, some of which we each partially realize when we participate in<br />

those textual his<strong>to</strong>ries. Interestingly, these his<strong>to</strong>ries, in this view, have <strong>to</strong> be grasped as<br />

fields of action rather than as linear unfoldings. The fields are <strong>to</strong>pological, with various<br />

emergent and dynamic basins of order, some of them linear and hierarchical, others not.<br />

Appendix A: The 'Pataphysics of Text and Field<br />

Markup<br />

Texts and their field spaces are au<strong>to</strong>poietic scenes of co-dependent emergence. As such,<br />

their primal state is dynamic and has been best characterized by G. Spencer Brown's<br />

Laws of Form (1969), where "the form of distinction" – the act of making indications by<br />

drawing a distinction – is taken as "given" and primal (1). This means that the elementary<br />

law is not the law of identity but the law of non-identity (so that we must say that "a<br />

equals a if and only if a does not equal a)." Identities emerge as distinctions are drawn<br />

and redrawn, and the acts of drawing out distinctions emerge as co-dependent responses<br />

<strong>to</strong> the field identities that the form of distinction calls <strong>to</strong> attention.<br />

Spencer Brown supplies a formal demonstration of what Alfred Jarry called'pataphysics<br />

and that he and his OULIPian inheri<strong>to</strong>rs demonstrated in forms of traditional textual<br />

practice (i.e., in forms of "literature").'Pataphysics is a general theory of au<strong>to</strong>poietic<br />

systems (i.e., a general theory of what we traditionally call "imaginative literature"), and<br />

Laws of Form is a specifically 'patapbysical event because it clearly gives logical priority<br />

<strong>to</strong> the unique act and practice of its own theoretical thought. The fifth "Chant" of Lautrea-


mont's Les chants de Maldoror, Jarry's Gestes et opinions du docteur<br />

Faustroll,'patapbysicien, and all the descendants of those self-conscious works – Laura<br />

Riding's s<strong>to</strong>ries are the earliest English-language examples – are the "literary" equivalents<br />

of Spencer Brown's Laws of Form.<br />

In this view of any systematics, the taxonomy of a system is a derivation of what Peirce<br />

called an initial abduction. The abduction is an hypothesis of the <strong>to</strong>tal semeiotic integrity<br />

of the system. The hypothesis is tested and transformed (internally as well as externally)<br />

in a dialectical process – ultimately endless – of representation and reflection.<br />

Appendix B: Control Dementians for a 'Patacriticism of<br />

Textualities<br />

The transaction of textual fields proceeds by a series of moves (field behaviors) that<br />

proliferate from an elementary modal distinction between what have been specified<br />

(above) as connections and resonances, which are the elementary behavioral forms of the<br />

textual transaction. These modes correspond <strong>to</strong> what traditional grammarians define as<br />

an indicative and a subjunctive verbal mood. (In this view, interrogative and interjective<br />

moods are derivatives of these two primary categories.) Emerging co-dependently with<br />

these behavioral dementians is an elementary taxonomy of control dementians that are<br />

called in<strong>to</strong> form and then internally elaborated.<br />

The his<strong>to</strong>ry of textual studies has evolved a standard set of field formalities that may be<br />

usefully analyzed in six distinct parts. These correspond <strong>to</strong> an elemental set of<br />

dimensions for textual fields (or, in fields conceived as au<strong>to</strong>poietic systems, an elemental<br />

set of six dementians). These control dementians locate what grammarians designate as<br />

the semantics of a language.<br />

Let it be said here that these behavioral and control dementians, like their allopoietic<br />

dimensions, comprise a set of categories that recommend themselves through an evolved<br />

his<strong>to</strong>ry of previous use. Other dimensions (and dementians) might be proposed or<br />

imagined. However, since the proposals being advanced here are all conceived within a<br />

pragmatistic frame of reference, the categories bring with them the strong authority of a<br />

habitual usefulness.<br />

The Linguistic Dimension/Dementian. This aspect of the textual condition has been the<br />

principal focus of attention in the West. It represents a high-order framework of<br />

conceptual markers or distinctions that unfold and multiply from an initial pair of<br />

categories, the semantic and the grammatical. The former is an elemental category, the<br />

latter is a relational one, and the two <strong>to</strong>gether epi<strong>to</strong>mize the structure of co-dependency<br />

that pervades and in a sense defines all textual processes at every dimension. That is <strong>to</strong><br />

say, neither marker or category has conceptual priority over the other, they generate<br />

meaning <strong>to</strong>gether in a co-dependent and dialectical process. However, <strong>to</strong> specify their codependence<br />

requires that one adopt a pragmatistic or performative approach such as we<br />

see in Maturana, Spencer Brown, and Peirce.


The Graphical/Auditional Dimension/Dementian. Some kind of graphical and/or auditional<br />

state of affairs is a prerequisite for any appearance or functional operation of a<br />

Linguistic Dimension, and that state must be formally constrained. In Western attempts <strong>to</strong><br />

clarify language and textuality, these forms are defined in the systematic descrip<strong>to</strong>rs of<br />

morphology and phonology, which are co-dependent subcategories of the Linguistic<br />

Dimension.<br />

This Graphical/Auditional Dimension comprises the set of a text's codes of materiality<br />

(as opposed <strong>to</strong> the specific material state of a particular document). In print and<br />

manuscript states, the dimension includes various subsets of bibliographical codes and<br />

paratexts: typography, layout, book design, and the vehicular components of those forms.<br />

(If we are considering oral texts, the material assumes auditional forms, which can have<br />

visual components as well.)<br />

Documentary Dimension/Dementian. This comprises the physical incarnation – the "real<br />

presence", so <strong>to</strong> speak – of all the formal possibilities of the textual process. We<br />

recognize it as a bibliographical or palaeographical description of some specific object, or<br />

as a library or archival record of an object's his<strong>to</strong>rical passage (transmission his<strong>to</strong>ry).<br />

Note that this dimension does not simply constitute some brute chemical or physical<br />

thing – what Coleridge referred <strong>to</strong> when he spoke as the "object as object", which he<br />

called "fixed and dead." Coleridge's "object as object" is a negative abstraction – that's <strong>to</strong><br />

say, a certain formal conception of the documentary dimension that sets it apart (a priori)<br />

from any place in a study or interpretation of textuality. A document can and – in any<br />

comprehensive approach <strong>to</strong> textuality – should be maintained as an integral function of<br />

the textual process.<br />

A document is a particular object that incarnates and constrains a specific textual process.<br />

In terms of print and manuscript texts, it is a specific actualized state of the<br />

Graphical/Auditional Dimension.<br />

Semiotic Dimension/Dementian. This dimension defines the limit state of any text's<br />

formal possibilities. It postulates the idea of the complete integration of all the elements<br />

and dynamic relations in a field of discourse. In this dimension we thus cognize a textual<br />

process in holistic terms. It is a purely formal perspective, however, and as such stands as<br />

the mirrored antithesis of the document per se, whose integrity is realized as a<br />

phenomenal event. The document is the image of the hypothesis of <strong>to</strong>tal form; it appears<br />

at (or as) a closure of the dynamic process set in perpetual motion by the hypothesis at<br />

the outset.<br />

We register the semiotic dimension as a pervasiveness of patterned relations throughout<br />

the textual system – both within each part of the system and among the parts. The<br />

relations emerge in distinct types or modes: elements begin and end; they can be<br />

accumulated, partitioned, and replicated; they can be anchored somewhere, linked <strong>to</strong><br />

other elements, and relayed through the system


The first of those late systems of analysis called by Herbert Simon "Sciences of the<br />

Artificial", the science of semiotics, labels itself as an heuristic mechanism. The<br />

pervasive order of a textual process's semiotic dimension thus emerges as a function of<br />

the formal categories, both system elements and system processes, that are consciously<br />

specified by the system's agents. Order is constructed from the systemic demand for<br />

order. As a result, the forms of order can be of any type – hierarchical or nonhierarchical,<br />

continuous or discontinuous.<br />

Rhe<strong>to</strong>rical Dimension/Dementian. The dominant form of this dimension is genre, which<br />

is a second-order set of textual forms. Genre calls in<strong>to</strong> play poems, mathematical proofs,<br />

novels, essays, speeches, dramas, and so forth. The function of this dimension is <strong>to</strong><br />

establish forms of readerly attention – <strong>to</strong> select and arrange textual materials of every<br />

kind in order <strong>to</strong> focus the interest of the reader (audience, user, listener) and establish a<br />

ground for response.<br />

Readers and writers (speakers and listeners) are rhe<strong>to</strong>rical functions. (Writers' first<br />

readers are themselves in their act of composition.) Bakhtin's celebrated studies of textual<br />

polyvalence and heteroglossia exemplify the operative presence of this textual dimension.<br />

Social Dimension/Dementian. This is the dimension of a textual production and of<br />

reception his<strong>to</strong>ries. It is the dimension of the object as subject: that is <strong>to</strong> say, of a<br />

determinate set of textual elements arrayed under names like "writer", "printer",<br />

"publisher", "reader", "audience", "user." It is the dimension that exposes the temporality<br />

function which is an inalienable feature of all the dimensions of the textual condition.<br />

The social dimension of textuality unfolds a schedule of the uses <strong>to</strong> which its works are<br />

put beyond what New Critics liked <strong>to</strong> call "the poem itself." It is the dimension in which<br />

the dynamic and non-self-identical character of textual works is most plainly disclosed.<br />

In most traditional theories of textuality, the social dimension is not considered an<br />

intrinsic textual feature or function. Framed under the sign "context", it is seen as the<br />

environment in which texts and documents stand. Until the recent emergence of more<br />

holistic views of environments – notably in the work of Donald McKenzie – this way of<br />

seeing textuality's social dimension forced severe restrictions on our ability <strong>to</strong><br />

comprehend and study the dynamic character of textual processes.<br />

Acknowledgment<br />

"The Innocence" by Robert Creeley is reprinted from The Collected Poems of Robert<br />

Creeley, 1945–1975 with kind permission of the University of California Press and<br />

Robert Creeley.<br />

Note<br />

1 See Maturana and Varela (1992).


2 "What Matters?" (at: http://www.w3.org/People/cmsmcq/2002/whatmatters.html).<br />

3 As the terms in this sentence indicate, I am working in terms laid down thirty years ago<br />

by Rene Thorn in his classic study Structural Stability and Morphogenesis (1975).<br />

4 For essays describing IVANHOE see the Special issue of TEXT Technology devoted <strong>to</strong><br />

it (vol. 12, no. 2, 2003).<br />

References for Further Reading<br />

Allen, Donald, (ed.) (1999). The New American Poetry 1945–1960, with a new<br />

Foreword. Berkeley and Los Angeles: University of California Press.<br />

Bellman, Rickard (1961). Adaptive Control Processes: A Guided Tour. Prince<strong>to</strong>n, NJ:<br />

Prince<strong>to</strong>n University Press.<br />

Berrie, Phillip William (1999). Just in Time Markup for Electronic Editions. Accessed<br />

April 19, 2004. At http://idun.itsc.adfa.edu.au/ASEC/PWB_REPORT/Index.html.<br />

Birnbaum, David J. (<strong>2001</strong>). The Relationship between General and Specific DTDs:<br />

Criticizing TEI Critical Editions. Markup Languages: Theory and Practice 3, 1: 17–53.<br />

Bornstein, George and Teresa Tinkle, (eds.) (1998). The Iconic Page in Manuscript,<br />

Print, and Digital Culture. Ann Arbor: University of Michigan Press.<br />

Brown, G. Spencer (1969). Laws of Form. London: George Allen and Unwin.<br />

Buzzetti, Dino (2002). Digital Representation and the Text Model. New Literary His<strong>to</strong>ry<br />

33, 1: 61–88.<br />

Casati, Rober<strong>to</strong> and Achille C. Varzi (1999). Parts and Places: The Structures of Spatial<br />

Representation. Cambridge, MA: MIT Press.<br />

Ca<strong>to</strong>n, Paul (<strong>2001</strong>). Markup's Current Imbalance. Markup Languages: Theory and<br />

Practice 3, 1: 1–13.<br />

Chandrasekaran, B., J. Glasgow, and N. H. Narayanan, (eds.) (1995). Diagrammatic<br />

Reasoning: Cognitive and Computational Perspective. Cambridge, MA: MIT Press, and<br />

Menlo Park: AAAI Press.<br />

Drucker, Johanna (1998). Figuring the Word: Essays on Books, Writing, and Visual<br />

Poetics. New York: Granary Books.<br />

Elkins, James (1998). On Pictures and Words that Fail Them. Cambridge: Cambridge<br />

University Press.


Elkins, James (1999). The Domain of Images. Ithaca, NY: Cornell University Press.<br />

Engell, James and W. Jackson Bate (1983). Samuel Taylor Coleridge. Biographia<br />

Literaria, 2 vols. Prince<strong>to</strong>n, NJ: Prince<strong>to</strong>n University Press.<br />

Fraenkel, Ernest (1960). Les dessins trans-conscients de Stéphane Mallarmé, à propos de<br />

la typographic de Un coup de dés, avant-propos par Étienne Souriau [Subconscious<br />

Drawings of Stephane Mallarmé, in connection with the typography of Un coup de dés,<br />

foreword by Étienne Souriau]. Paris: Librairie Nizet.<br />

Habermas, Jiirgen (1984). The Theory of Communicative Action, tr. Thomas McCarthy.<br />

Bos<strong>to</strong>n: Beacon Press.<br />

Hardwick, Charles, (ed.) (1977). Semiotic and Signifies: The Correspondence between<br />

Charles S. Peirce and Vic<strong>to</strong>ria, Lady Welby. Blooming<strong>to</strong>n: Indiana University Press.<br />

Hauser, Nathan and Christian Kloesel, (eds.) (1992). The Essential Peirce: Selected<br />

Philosophical Writings, 2 vols. Blooming<strong>to</strong>n: Indiana University Press.<br />

Hockey, Susan (2000). Electronic Texts in the Humanities. Oxford: Oxford University<br />

Press.<br />

Luhmann, Niklas (1998). Observations on Modernity, tr. William Whobrey. Stanford,<br />

CA: Stanford University Press.<br />

Maturana, Humber<strong>to</strong> and Francisco Varela (1980). Au<strong>to</strong>poiesis and Cognition: The<br />

Realization of Living. Bos<strong>to</strong>n: D. Reidel.<br />

Maturana, Humber<strong>to</strong> and Francisco Varela (1992). The Tree of Knowledge: Biological<br />

Roots of Human Understanding, tr. Robert Paolucci. New York: Random House.<br />

McCarty, Willard (2002a). Computing the Embodied Idea: Modeling in the Humanities.<br />

Körper - Verkörperung - Entkörperung I Body - Embodiment - Disembodiment. 10.<br />

Internationaler Kongress, Deutsche Gesellschaft fur Semiotik, Universität Kassel, July<br />

19, 2002. At (http://www.kcl.ac.uk/<strong>humanities</strong>/cch/wlm/essays/kassel/).<br />

McCarty, Willard (2002b). Humanities Computing: Essential Problems, Experimental<br />

Practice. Accessed April 19, 2004. At<br />

http://www.kcl.ac.uk/<strong>humanities</strong>/cch/wlm/essays/stanford/.<br />

McDonald, Peter D. and Suarez, SJ Michael, (eds.) (2002). Making Meaning: "Printers<br />

of the Mind" and Other Essays. D. F. McKenzie. Amherst: University of Massachusetts<br />

Press.<br />

McGann, Jerome (<strong>2001</strong>). Radiant Textuality: Literature after the World Wide Web. New<br />

York and London: Palgrave/St Martins.


McGann, Jerome, (ed.) (2004). The Complete Writings and Pictures of Dante Gabriel<br />

Rossetti. A Hypermedia Research Archive. Accessed April 19, 2004. At<br />

http://jefferson.village.virginia.edu/rossetti/.<br />

McKenzie, D. F. (1986). Bibliography and the Sociology of Texts: The Panizzi Lectures,<br />

1985. London: British Library.<br />

McLuhan, Marshall (1964). Understanding Media: The Extensions of Man. New York:<br />

McGraw-Hill.<br />

Mineau, G., B. Moulin, and J. Sowa, (eds.) (1993). Conceptual Graphs for Knowledge<br />

Representation. Berlin: Springer-Verlag.<br />

Omnès, Roland (1999). Understanding Quantum Mechanics. Prince<strong>to</strong>n, NJ: Prince<strong>to</strong>n<br />

University Press.<br />

Shin, Sun-Joo (2002). The Iconic Logic of Peirce's Graphs. Cambridge, MA: MIT Press.<br />

Simon, Herbert (1981). The Sciences of the Artificial, 2nd edn., rev. and enlarged.<br />

Cambridge, MA: MIT Press.<br />

Sontag, Susan, (ed.) (1982). A Barthes Reader. New York: Noonday Press.<br />

Sperberg-McQueen, C. M. (2002). What Matters? At<br />

http://www.w3.org/People/cmsmcq/2002/whatmatters.html.<br />

Sperberg-McQueen, C. M., Claus Huitfeldt, and Allen Renear (2000). Meaning and<br />

Interpretation of Markup. Markup Languages 2, 3: 215–34.<br />

Thorn, Rene (1975). Structural Stability and Morphogenesis: An Outline of a General<br />

Theory of Models, tr. D. H. Fowler, with a Foreword by C. H. Wadding<strong>to</strong>n. Reading,<br />

MA: W. A. Benjamin.<br />

Thompson, Henry S. and David McKelvie (1997). Hyperlink Semantics for Standoff<br />

Markup of Read-Only Documents. At http://www.ltg.ed.ac.uk/~ht/sgmleu97.html.<br />

Varela, Francisco J., Evan Thompson, and Eleanor Rosch (1991). The Embodied Mind:<br />

Cognitive Science and Human Experience. Cambridge, MA: MIT Press.<br />

17.<br />

Text Encoding<br />

Allen H. Renear


Before they can be studied with the aid of machines, texts must be encoded in a machinereadable<br />

form. Methods for this transcription are called, generically, "text encoding<br />

schemes"; such schemes must provide mechanisms for representing the characters of the<br />

text and its logical and physical structure … ancillary information achieved by analysis or<br />

interpretation [may be also added] …<br />

Michael Sperberg-McQueen, Text Encoding and Enrichment. In The Humanities<br />

Computing Yearbook 1989–90, ed. Ian Lancashire (Oxford: Oxford University Press,<br />

1991)<br />

Introduction<br />

Text encoding holds a special place in <strong>humanities</strong> computing. It is not only of<br />

considerable practical importance and commonly used, but it has proven <strong>to</strong> be an exciting<br />

and theoretically productive area of analysis and research. Text encoding in the<br />

<strong>humanities</strong> has also produced a considerable amount of interesting debate – which can be<br />

taken as an index of both its practical importance and its theoretical significance.<br />

This chapter will provide a general orientation <strong>to</strong> some of the his<strong>to</strong>rical and theoretical<br />

context needed for understanding both contemporary text encoding practices and the<br />

various ongoing debates that surround those practices. We will be focusing for the most<br />

part, although not exclusively, on "markup", as markup-related techniques and systems<br />

not only dominate practical encoding activity, but are also at the center of most of the<br />

theoretical debates about text encoding. This chapter provides neither a survey of markup<br />

languages nor a tu<strong>to</strong>rial introduction <strong>to</strong> the practice of markup. The reader new <strong>to</strong><br />

SGML/XML text encoding should read this chapter of the Companion concurrently with<br />

the short (21-page) second chapter of the TEI Guidelines, "A Gentle Introduction <strong>to</strong><br />

XML", online at http://www.tei-c.org/P4X/SG.html. The justly renowned "Gentle<br />

Introduction" remains the best brief presentation of SGML/XML text encoding and it<br />

provides a necessary complement of specific description <strong>to</strong> the background and theory<br />

being presented here. For a good general introduction <strong>to</strong> text encoding in the <strong>humanities</strong>,<br />

see Electronic Texts in the Humanities: Theory and Practice, by Susan Hockey (Hockey<br />

<strong>2001</strong>).<br />

In accordance with the general approach of this Companion, we understand text encoding<br />

in the "<strong>digital</strong> <strong>humanities</strong>" in a wide sense. Traditional <strong>humanities</strong> computing<br />

(particularly when related <strong>to</strong> literature and language) typically emphasized either<br />

analytical procedures on encoded textual material – such as, for instance, stylometric<br />

analysis <strong>to</strong> support authorship or seriation studies – or the publishing of important<br />

traditional genres of scholarship such as critical and variorum editions, indexes,<br />

concordances, catalogues, and dictionaries. But text encoding is no less important <strong>to</strong><br />

<strong>digital</strong> <strong>humanities</strong> broadly conceived, in the sense which includes the creation and study<br />

of new cultural products, genres, and capabilities, such as those involving hypertext,<br />

multimedia, interactivity, and networking – cultural products that are often called "new<br />

media." In order <strong>to</strong> be presented using computers, such material must be encoded in<br />

machine-readable form. Although the presentation below does not take up hypertext or


"new media" <strong>to</strong>pics directly, we believe the background presented nevertheless provides<br />

a useful background for text encoding in general, new media applications as well as<br />

traditional <strong>humanities</strong> computing and publishing. For more specific treatment of<br />

encoding issues, hypertext, multimedia, and other new media <strong>to</strong>pics, see chapters 10, 28,<br />

29, and 30. For discussions of traditional <strong>humanities</strong> computing applications exploiting<br />

text encoding, see chapters 20, 21, 22, and 35.<br />

What follows then is intended as a background for understanding the text encoding as a<br />

representation system for textually based cultural objects of all kinds.<br />

Markup<br />

Introduction<br />

Markup, in the sense in which we are using the term here, may be characterized, at least<br />

provisionally, as information formally distinct from the character sequence of the <strong>digital</strong><br />

transcription of a text, which serves <strong>to</strong> identify logical or physical features or <strong>to</strong> control<br />

later processing. In a typical unformatted view of a <strong>digital</strong> representation of a text such<br />

markup is visibly evident as the more or less unfamiliar expressions or codes that are<br />

intermixed with the familiar words of the natural-language writing system. The term<br />

markup comes, of course, from traditional publishing, where an edi<strong>to</strong>r marks up a<br />

manuscript by adding annotations or symbols on a paper copy of text indicating either<br />

directly (e.g., "center") or indirectly ("heading") how something is <strong>to</strong> look in print<br />

(Spring 1989); Chicago 1993.<br />

Many markup theorists have found it instructive <strong>to</strong> conceptualize markup as a very<br />

general phenomenon of human communication. Extending the notion of markup in<br />

straightforward and natural ways, one can easily use this concept <strong>to</strong> illuminate aspects of<br />

the general nature and his<strong>to</strong>ry of writing systems and printing, particularly in the areas<br />

such as page layout, typography, and punctuation (Coombs et al. 1987).<br />

In addition, other fields and disciplines related <strong>to</strong> communication, such as rhe<strong>to</strong>ric,<br />

bibliography, textual criticism, linguistics, discourse and conversation analysis, logic, and<br />

semiotics also seem <strong>to</strong> be rich sources of related findings and concepts that generalize our<br />

understanding of markup practices and make important connections between markup<br />

practices narrowly unders<strong>to</strong>od and other bodies of knowledge and technique.<br />

However, although such a broad perspective can be illuminating, the significance of<br />

markup for <strong>humanities</strong> computing is best approached initially by considering markup's<br />

origin and development in computer-based typesetting and early text processing.<br />

Although markup might arguably be considered part of any communication system, and<br />

of fundamental theoretical significance, it is with straightforward applications in <strong>digital</strong><br />

text processing and typesetting in the 1960s, 1970s, and 1980s that the use of markup, in<br />

our sense, first becomes explicit and begins <strong>to</strong> undergo deliberate and self-conscious<br />

development (Goldfarb 1981; Spring 1989; SGML Users' Group 1990).


Emergence of descriptive markup<br />

The use of computers <strong>to</strong> compose text for typesetting and printing was common by the<br />

mid-1960s and the general process was more or less the same regardless of the specific<br />

technology. Typically, the first step was <strong>to</strong> create and s<strong>to</strong>re in a computer file a<br />

representation of the text <strong>to</strong> be printed. This representation consisted of both codes for the<br />

individual characters of the textual content and codes for formatting commands, the<br />

commands and the text being distinguished from each other by special characters or<br />

sequences of characters serving as "delimiters." The file would then be processed by a<br />

software application that acted on the formatting instructions <strong>to</strong> create data which could<br />

in turn be directly read and further processed by the pho<strong>to</strong>typesetter or computer printer –<br />

creating formatted text as final output (Seybold 1977; Furuta et al. 1982).<br />

The 1960s, 1970s, and 1980s saw extensive development of these document markup<br />

systems as software designers worked <strong>to</strong> improve the efficiency and functionality of<br />

<strong>digital</strong> typesetting and text processing software (IBM 1967; Ossanna 1976; Lesk 1978;<br />

Goldfarb 1978; Reid 1978; Knuth 1979; Lamport 1985).<br />

One natural improvement on the approach just described was <strong>to</strong> replace the long strings<br />

of complex formatting codes with simpler abbreviations that could be au<strong>to</strong>matically<br />

expanded in<strong>to</strong> the formatting commands being abbreviated. The composi<strong>to</strong>r would enter<br />

just the abbreviation instead of the entire string of commands. In many typesetting<br />

systems in the 1960s and 1970s these abbreviations were called "macros", a term drawn<br />

from assembly language programming where it referred <strong>to</strong> higher-level symbolic<br />

instructions which would be expanded, before program execution, in<strong>to</strong> sequences of<br />

lower-level primitive instructions.<br />

In typesetting and text processing these macros had the obvious immediate advantage of<br />

easier and more reliable data entry. However, during their early use it wasn't always<br />

entirely clear whether a macro got its primary identity from the text component (e.g., a<br />

caption, extract, or heading) whose formatting it was controlling, or whether it was<br />

simply a short name for the combination of the specific formatting codes it was<br />

abbreviating, with no other meaning or identity. The distinction is subtle but important. If<br />

a macro is truly just an abbreviation of a string of formatting codes then it can be<br />

appropriately used wherever those formatting codes would be used. So if, for instance,<br />

figure captions and third-level headings happen <strong>to</strong> have the same design specifications,<br />

then the same macro could reasonably and appropriately be used for both. In such a case<br />

the macro, as a mere abbreviation, gets its entire meaning and identity from the<br />

formatting commands it abbreviates and it would be natural for the macro name <strong>to</strong> then<br />

simply indicate the appearance of the formatted text (e.g., ":SmallCenteredBold;"), or be<br />

an arbitrary expression (e.g., ":format!7;") rather than have a name that suggested an<br />

intrinsic relationship with the text component being formatted (such as, ":figurecaption")<br />

(Goldfarb 1997).<br />

Although macros used as described above obviously provided some advantages over<br />

entering strings of formatting codes, it was a natural next step <strong>to</strong> see that many more


advantages could be achieved by understanding the presence of the macro name in the<br />

file <strong>to</strong> be identifying the occurrence of a particular text component – a third-level<br />

heading, caption, stanza, extract, title, etc. – rather than just being an abbreviation for a<br />

string of formatting commands. On this new approach, figure captions, for instance,<br />

would be identified with one code (say, ":FigureCaption;"), and third-level headings<br />

would be identified with another (say, ":Heading3;" even if, according <strong>to</strong> the page design<br />

specification currently being applied, these codes were mapped <strong>to</strong> the same set of<br />

formatting commands.<br />

The first advantage <strong>to</strong> this new approach is that it is now possible <strong>to</strong> globally alter the<br />

formatting of figure captions (by simply updating the formatting commands associated<br />

with ":FigureCaption;") without necessarily changing the formatting of the third-level<br />

headings (identified by the macro name ":Heading3;"). In addition, authoring and even<br />

composition can now take place without the author or composi<strong>to</strong>r needing <strong>to</strong> know how<br />

the different components are <strong>to</strong> be formatted. As this approach <strong>to</strong> typesetting and text<br />

processing began <strong>to</strong> be systematically applied it became quickly apparent that there were<br />

a great many other advantages. So many advantages, in fact, and such a diversity of<br />

advantages, that the descriptive markup approach began <strong>to</strong> appear <strong>to</strong> be somehow the<br />

fundamentally correct approach <strong>to</strong> organizing and processing text (Goldfarb 1981; Reid<br />

1981; Coombs et al. 1987).<br />

The promotion of descriptive markup as the fundamentally correct systematic approach<br />

in <strong>digital</strong> publishing and text processing is usually traced <strong>to</strong> three events: (i) a<br />

presentation made by William Tunnicliffe, chairman of the Graphic Communications<br />

Association's Composition Committee, at the Canadian Government Printing Office in<br />

September 1967; (ii) book designer Stanley Rice's project, also in the late 1960s, of<br />

developing a universal catalogue of "edi<strong>to</strong>rial structure" tags that would simplify book<br />

design and production; and (iii) early work on the text processing macro language<br />

"GML", led by Charles Goldfarb, at IBM in 1969 (SGML Users' Group 1990; Goldfarb<br />

1997). In the late 1970s these events would lead <strong>to</strong> an effort <strong>to</strong> develop SGML, a standard<br />

for machine-readable definitions of descriptive markup languages. Other examples of<br />

early use of descriptive markup in <strong>digital</strong> text processing include the Brown University<br />

PRESS hypertext system (Carmody et al. 1969; DeRose and van Dam 1999), and, later in<br />

the 1970s, Brian Reid's SCRIBE (Reid 1978, 1981). In addition, the seminal work on text<br />

processing by Douglas Engelbart in the 1960s should probably also be seen as exhibiting<br />

some of the rudiments of this approach (Engelbart et al. 1973).<br />

Nature and advantages of descriptive markup (adapted from DeRose<br />

et al. 1990)<br />

Early experience with descriptive markup encouraged some text processing researchers <strong>to</strong><br />

attempt <strong>to</strong> develop a general theoretical framework for markup and <strong>to</strong> use that framework<br />

<strong>to</strong> support the development of high-function text processing systems. Some of the<br />

research and analysis on markup systems was published in the scientific literature (Goldfarb<br />

1981; Reid 1981; Coombs et al. 1987), but most was recorded only in the working


documents and products of various standards bodies, and in the manuals and technical<br />

documentation of experimental systems.<br />

At the heart of this effort <strong>to</strong> understand markup systems was the distinction between<br />

"descriptive" and "procedural" markup originally put forward by Goldfarb. Descriptive<br />

markup was typically said <strong>to</strong> "identify" or "describe" the "parts" of a document, whereas<br />

procedural markup was a "command" or "instruction" invoking a formatting procedure. It<br />

was also often said that descriptive markup identified the "logical" or "edi<strong>to</strong>rial" parts or<br />

"components" of a document, or a text's "content objects" or its "meaningful structure" –<br />

emphasizing the distinction between the intrinsic ("logical") structure of the document<br />

itself, and the varying visual, graphic features of a particular presentation of that<br />

document. (For recent arguments that descriptive markup can be further divided in<strong>to</strong> the<br />

genuinely descriptive and the "performative", see Renear 2000.)<br />

Several particular advantages of descriptive markup, such as simplified composition and<br />

systematic control over formatting, have already been alluded <strong>to</strong>, but in order <strong>to</strong><br />

appreciate how the descriptive markup motivated a new theory of the nature of text it is<br />

useful <strong>to</strong> rehearse the number, diversity, and value of the advantages of descriptive<br />

markup in overview; so we present a categorized summary below.<br />

Advantages for authoring, composition, and transcription<br />

• Composition is simplified. With descriptive markup, intended formatting considerations<br />

make no claim on the attention of the author, composi<strong>to</strong>r, or transcriber, whereas with<br />

procedural markup one must remember both (i) the style conventions that are intended,<br />

and (ii) the specific commands required by the formatting software <strong>to</strong> get those effects.<br />

With descriptive markup one simply identifies each text component for what it is and the<br />

appropriate formatting takes place au<strong>to</strong>matically. (Particularly striking is how descriptive<br />

markup allows the author <strong>to</strong> work at an appropriate "level of abstraction" – identifying<br />

something as a quotation, paragraph, or caption is a natural authorial task, while knowing<br />

whether <strong>to</strong>, and how <strong>to</strong>, format that text a certain way is not.)<br />

• Structure-oriented editing is supported. Descriptive markup supports "structureoriented<br />

edi<strong>to</strong>rs" who "know" about what patterns of components can be found in a<br />

particular genre of document and who use this knowledge <strong>to</strong> assist the author or<br />

composi<strong>to</strong>r. For instance, if a date component must always follow a title component then<br />

the software, upon detecting that a title component has just been entered by an author,<br />

can au<strong>to</strong>matically add the required date markup and prompt the author <strong>to</strong> enter the actual<br />

date. If either date or status is allowed after a title then the author will be presented with a<br />

choice. During editing the cursor location can be used <strong>to</strong> identify and present <strong>to</strong> the<br />

author the list of components that may be added or deleted at that point. For complicated<br />

document genres this means that there is much less for the author <strong>to</strong> remember and fewer<br />

possibilities for error.<br />

• More natural editing <strong>to</strong>ols are supported. Moves and deletes, for example, can take as<br />

their targets and scope the natural meaningful parts of the text (words, sentences, para


graphs, sections, extracts, equations, etc.) rather than relying on the mediation of<br />

accidental features (such as current lineation) or arbitrarily marked regions.<br />

• Alternative document views are facilitated. An outline view of a text, for instance, can<br />

be done au<strong>to</strong>matically, by taking advantage of the descriptive markup for chapters,<br />

sections, and headings. Or a more sophisticated and specialized display of portions of<br />

documents can be effected using identified discipline-specific components: such as<br />

equations, examples, cautions, lines spoken by a particular character in a play script, and<br />

so on.<br />

Advantages for publishing<br />

• Formatting can be generically specified and modified. When procedural markup is<br />

being used, the appearance of paragraphs can only be modified by editing the formatting<br />

commands preceding each actual occurrence of a paragraph in the source file, whereas<br />

with descriptive markup only the rule associating formatting commands with the<br />

descriptive markup for paragraph needs <strong>to</strong> be updated – and if these rules are s<strong>to</strong>red<br />

separately there may be no need <strong>to</strong> even alter the files containing text in order <strong>to</strong> make<br />

formatting alterations. Obviously, controlling formatting with descriptive markup is<br />

easier, less error-prone, and ensures consistency.<br />

• Apparatus can be au<strong>to</strong>mated. Descriptive markup supports the creation of indexes,<br />

appendices, and such. For instance, if stanzas and verse lines are explicitly identified, the<br />

creation of an index of first lines of stanzas (or second lines or last lines) is a matter of<br />

simple programming, and does not require that a human edi<strong>to</strong>r again laboriously identify<br />

verses and lines. Similarly, generating tables for equations, plates, figures, examples, is<br />

also easy, as are indexes of place names, personal names, characters, medium, authors,<br />

periods, and so on.<br />

• Output device support is enhanced. When coding is based on logical role rather than<br />

appearance, output-device specific support for printers, typesetters, video display<br />

terminals and other output devices can be maintained separately, logically and physically,<br />

from the data with the convenient result that the data files themselves are output-device<br />

independent while their processing is efficiently output-device sensitive.<br />

• Portability and interoperability are maximized. Files that use descriptive markup, rather<br />

than complicated lists of application-specific formatting instructions, <strong>to</strong> identify<br />

components are much easier <strong>to</strong> transfer <strong>to</strong> other text processing systems. In some cases<br />

little more than a few simple systematic changes <strong>to</strong> alter the delimiter conventions,<br />

substitute one mnemonic name for another, and a translation of format ting rules in<strong>to</strong><br />

those for the new system, are all that is necessary.<br />

Advantages for archiving, retrieval, and analysis<br />

• Information retrieval is supported. Descriptive markup allows documents <strong>to</strong> be treated<br />

as a database of fielded content that can be systematically accessed. One can request all


equations, or all headings, or all verse extracts; or one can request all titles that contain a<br />

particular personal name, place name, chemical, disease, drug, or therapy. This can<br />

facilitate not only personal information retrieval functions, such as the generation of<br />

alternative views, but also a variety of finding aids, navigation, and data retrieval<br />

functions. It may seem that in some cases, say where equations are uniquely formatted, it<br />

is not necessary <strong>to</strong> identify them, as the computer could always be programmed <strong>to</strong> exploit<br />

the formatting codes. But in practice it is unlikely that equations will always be<br />

consistently formatted, and even more unlikely that they will be uniquely formatted.<br />

Similarly, it might be thought that a string search might retrieve all references <strong>to</strong> the city<br />

Chicago, but without markup one cannot distinguish the city, the rock band, and the<br />

artist.<br />

• Analytical procedures are supported. Computer-based analysis (stylometrics, content<br />

analysis, statistical studies, etc.) can be carried out more easily and with better results if<br />

features such as sentences, paragraphs, stanzas, dialogue lines, stage directions, and so on<br />

have been explicitly identified so that the computer can au<strong>to</strong>matically distin guish them.<br />

Consider some examples: If the style of spoken language in a play is being analyzed, the<br />

text that is not speech (stage directions, notes, etc.) must not be conflated with the<br />

dialogue. If the speech of a particular character is being studied it must be distinguished<br />

from the speech of other characters. If proximity is being studied, the word ending one<br />

paragraph should perhaps not be counted as collocated with the word beginning the next,<br />

and so on. Descriptive markup identifies these important components and their<br />

boundaries, supporting easier, more consistent, and more precise au<strong>to</strong>matic processing.<br />

The OHCO view of "what text really is"<br />

That the descriptive markup approach had so many advantages, and so many different<br />

kinds of advantages, seemed <strong>to</strong> some people <strong>to</strong> suggest that it was not simply a handy<br />

way of working with text, but that it was rather in some sense deeply, profoundly,<br />

correct, that "descriptive markup is not just the best approach … it is the best imaginable<br />

approach" (Coombs et al. 1987).<br />

How could this be? One answer is that the descriptive markup approach, and only the<br />

descriptive markup approach, reflects a correct view of "what text really is" (DeRose et<br />

al. 1990). On this account, the concepts of descriptive markup entail a model of text, and<br />

that model is more or less right. The model in question postulates that text consists of<br />

objects of a certain sort, structured in a certain way. The nature of the objects is best<br />

suggested by example and contrast. They are chapters, sections, paragraphs, titles,<br />

extracts, equations, examples, acts, scenes, stage directions, stanzas, (verse) lines, and so<br />

on. But they are not things like pages, columns, (typographical) lines, font shifts, vertical<br />

spacing, horizontal spacing, and so on. The objects indicated by descriptive markup have<br />

an intrinsic direct connection with the intellectual content of the text; they are the<br />

underlying "logical" objects, components that get their identity directly from their role in<br />

carrying out and organizing communicative intention. The structural arrangement of<br />

these "content objects" seems <strong>to</strong> be hierarchical – they nest in one another without<br />

overlap. Finally, they obviously also have a linear order as well: if a section contains


three paragraphs, the first paragraph precedes the second, which in turn precedes the<br />

third.<br />

On this account then text is an "Ordered Hierarchy of Content Objects" (OHCO), and<br />

descriptive markup works as well as it does because it identifies that hierarchy and makes<br />

it explicit and available for systematic processing. This account is consistent with the<br />

traditional well-unders<strong>to</strong>od advantages of "indirection" and "data abstraction" in<br />

information science.<br />

A number of things seem <strong>to</strong> fall in<strong>to</strong> place from this perspective. For one thing, different<br />

kinds of text have different kinds of content objects (compare the content objects in<br />

dramatic texts with those in legal contracts), and typically the patterns in which content<br />

objects can occur is at least partially constrained: the parts of a letter occur in a certain<br />

order, the lines of a poem occur within, not outside of a stanza, and so on. Presentational<br />

features make it easier for the reader <strong>to</strong> recognize the content objects of the text.<br />

There are alternative models of text that could be compared with the OHCO model. For<br />

instance, one could model text as a sequence of graphic characters, as in the "plain vanilla<br />

ASCII" approach of Project Gutenberg; as a combination of procedural coding and<br />

graphic characters, as in a word processing file; as a complex of geometric shapes, as in<br />

"vec<strong>to</strong>r graphics" format of an image of a page on which the text is written; as a pure<br />

image, as in a raster image format (JPEG, GIF, etc.); or in a number of other ways.<br />

However, implementations based on these alternative models are all clearly inferior in<br />

functionality <strong>to</strong> implementations based on an OHCO model, and in any case can be easily<br />

and au<strong>to</strong>matically generated from an OHCO format (DeRose et al. 1990; Renear et al.<br />

1996).<br />

The OHCO account of "what text really is" has not gone uncriticized, but many have<br />

found it a compelling view of text, one that does explain the effectiveness of descriptive<br />

markup and provides a general context for the systematization of descriptive markup<br />

languages with formal metalanguages such as SGML.<br />

SGML and XML<br />

As we indicated above, this chapter is not designed <strong>to</strong> introduce the reader <strong>to</strong> text<br />

encoding, but <strong>to</strong> provide in overview some useful his<strong>to</strong>rical and theoretical background<br />

that should help support a deeper understanding. In this section we present some of the<br />

principal themes in SGML and XML, and take up several <strong>to</strong>pics that we believe will be<br />

useful, but for a complete presentation the reader is again directed <strong>to</strong> the "Gentle<br />

Introduction" and the other readings indicated in the references below.<br />

His<strong>to</strong>ry of SGML and XML: Part I<br />

Norman Scharpf, direc<strong>to</strong>r of the Graphics Communication Association, is generally<br />

credited with recognizing the significance of the work, mentioned above, of Tunnicliffe,<br />

Rice, and Goldfarb, and initiating, in the late 1960s, the GCA "GenCode" project, which


had the goal of developing a standard descriptive markup language for publishing<br />

(SGML Users' Group 1990; Goldfarb 1997). Soon much of this activity shifted <strong>to</strong> the<br />

American National Standards Institute (ANSI), which selected Charles Goldfarb <strong>to</strong> lead<br />

an effort for a text description language standard based on GML and produced the first<br />

working draft of SGML in 1980. These activities were reorganized as a joint project of<br />

ANSI and the International Organization for Standardization (ISO) and in 1986 ISO<br />

published ISO 8879: Information Processing – Text and Office Systems – Standard<br />

Generalized Markup Language (SGML) (ISO 1986; Goldfarb 1991). Later events in the<br />

his<strong>to</strong>ry of SGML, including the development of XML and the profusion of SGML/XML<br />

standards that began in the mid-1990s, are discussed below. We will later describe more<br />

precisely the relationship between SGML and XML; at this point in the chapter the reader<br />

only needs <strong>to</strong> know that XML is, roughly, just a simpler version of SGML, and as a<br />

consequence almost everything we say about SGML applies <strong>to</strong> XML as well.<br />

The basic idea: a metalanguage for defining descriptive markup<br />

languages<br />

Although for a variety of reasons the SGML standard itself can be difficult <strong>to</strong> read and<br />

understand, the basic idea is quite simple. SGML is a language for creating machinereadable<br />

definitions of descriptive markup languages. As such it is called a<br />

"metalanguage", a language for defining a language. The SGML standard provides a way<br />

<strong>to</strong> say things like this: these are the characters I use for markup tag delimiters; these are<br />

the markup tags I am using; these are the acceptable arrangements that the components<br />

identified by my markup tags may take in a document; these are some characteristics I<br />

may be asserting of those components, these are some abbreviations and shortcuts I'll be<br />

using, and so on. An alternative characterization of SGML is as a "metagrammar", a<br />

grammar for defining other grammars; here "grammar" is used in a technical sense<br />

common in linguistics and computer science. (Although its origin is as a metalanguage<br />

for defining document markup languages, SGML can be, and is, used <strong>to</strong> define markup<br />

languages for things other than documents, such as languages for data interchange or<br />

interprocess communication, or specific data management applications.)<br />

An almost inevitable first misconception is that SGML is a document markup language<br />

with a specific set of document markup tags for features such as paragraphs, chapters,<br />

abstracts, and so on. Despite its name ("Standard Generalized Markup Language"),<br />

SGML is not a markup language in this sense and includes no document markup tags for<br />

describing document content objects. SGML is a metalanguage, a language for defining<br />

document markup languages. The confusion is supported by the fact that both markup<br />

metalanguages like SGML and XML and also SGML-based markup languages like<br />

HTML and XHTML (technically applications of SGML) are "markup languages" in<br />

name. In each case the "ML" stands for "markup language", but SGML and HTML are<br />

markup languages in quite different senses. The characterization of SGML as a document<br />

markup language is not entirely unreasonable of course, as it is in a sense a language for<br />

marking up documents, just one without a predefined set of markup for content objects.<br />

Whether or not SGML should be called a markup language is his<strong>to</strong>rically moot, but <strong>to</strong><br />

emphasize the distinction between markup languages that are actual sets of markup tags


for document content objects (such as HTML, XHTML, TEI, DocBook, etc.) and<br />

metalanguages for defining such markup languages (such as SGML and XML) we will,<br />

in what follows, reserve the term markup language for languages like HTML and TEI,<br />

excluding the metalanguages used <strong>to</strong> define them.<br />

Standardizing a metalanguage for markup languages rather than a markup language was a<br />

key strategic decision. It was an important insight of the GCA GenCode Committee that<br />

any attempt <strong>to</strong> define a common markup vocabulary for the entire publishing industry, or<br />

even common vocabularies for portions of the industry, would be extremely difficult, at<br />

best. However, a major step <strong>to</strong>wards improving the interoperability of computer<br />

applications and data could be achieved by standardizing a metalanguage for defining<br />

markup languages. Doing this would obviously be an easier task: (i) it would not require<br />

the development of a markup language vocabulary with industry-wide acceptance and (ii)<br />

it would not assume that the industry would continue <strong>to</strong> conform, even as circumstances<br />

changed and opportunities for innovations arose, <strong>to</strong> the particular markup language it had<br />

agreed on.<br />

At the same time this approach still ensured that every markup language defined using<br />

SGML, even ones not yet invented, could be recognized by any computer software that<br />

unders<strong>to</strong>od the SGML metalanguage. The idea was that an SGML software application<br />

would first process the relevant SGML markup language definition, and then could go<br />

on, as a result of having processed that definition, <strong>to</strong> process the marked-up document<br />

itself – now being able <strong>to</strong> distinguish tags from content, expand abbreviations, and verify<br />

that the markup was correctly formed and used and that the document had a structure that<br />

was anticipated by the markup language definition. Because the SGML markup language<br />

definition does not include processing information such as formatting instructions,<br />

software applications would also require instructions for formatting that content; shortly<br />

after the SGML project began, work started on a standard for assigning general<br />

instructions for formatting and other processing <strong>to</strong> arbitrary SGML markup. This was the<br />

Document Style and Semantics Specification Language (DSSSL). The World Wide Web<br />

Consortium (W3C)'s Extensible Stylesheet Language (XSL) is based on DSSSL, and the<br />

W3C Cascading Style Sheet specification performs a similar, although more limited<br />

function.<br />

This strategy supports technological innovation by allowing the independent<br />

development of better or more specialized markup, without the impediment of securing<br />

antecedent agreement from software developers that they will manufacture software that<br />

can process the new markup – all that is required is that the software understand SGML,<br />

and not a specific SGML markup language.<br />

Elements and element types<br />

Up until now we have been using the phrase content object for the logical parts of a<br />

document or text. SGML introduces a technical term, element, that roughly corresponds<br />

<strong>to</strong> this notion, developing it more precisely. The first bit of additional precision is that<br />

where we have used "content object" ambiguously, sometimes meaning the specific titles,


paragraphs, extracts and the like that occur in actual documents, and sometimes meaning<br />

the general kind of object (title, paragraph, extract) of which those specific occurrences<br />

are instances, SGML has distinct terminology for these two senses of "content object": an<br />

actual component of a document is an "element", and the type of an element (title,<br />

paragraph, extract, etc.) is an "element type." However, apart from the standard itself, and<br />

careful commentary on the standard, it is common, and usually not a problem, for<br />

"element" <strong>to</strong> be used in the sense of "element type" as well. This kind of ambiguity is<br />

common enough in natural language and almost always easily and unconsciously<br />

resolved by context.<br />

The second bit of additional precision is that where we used "content object" <strong>to</strong> refer <strong>to</strong><br />

the parts of texts unders<strong>to</strong>od in the usual way as abstract cultural objects, independent of,<br />

and prior <strong>to</strong>, any notion of markup languages, the most exact use of "element" in the<br />

technical SGML sense is <strong>to</strong> refer <strong>to</strong> the combination of SGML markup tags and enclosed<br />

content that is being used <strong>to</strong> represent (for instance, in a computer file) these familiar<br />

abstract textual objects. Again the distinction is subtle and "element" is routinely used in<br />

both senses (arguably even in the standard itself). Typically, no damage is done, but<br />

eventually encoding cruxes, design quandaries, or theoretical disputes may require appeal<br />

<strong>to</strong> the full range of distinctions that can be made when necessary.<br />

Document types and document instances<br />

Fundamental <strong>to</strong> SGML is the notion of the document type, a class of documents with a<br />

particular set of content objects that can occur in some combinations but not in others.<br />

Some examples of document types might be: novel, poem, play, article, essay, letter,<br />

contract, proposal, receipt, catalogue, syllabus, and so on. Some examples of combination<br />

rules: in a poem, lines of verse occur within stanzas, not vice versa; in a play, stage<br />

directions may occur either within or between speeches but not within other stage<br />

directions (and not, normally, in the table of contents); in a catalogue, each product<br />

description must contain exactly one part number and one or more prices for various<br />

kinds of cus<strong>to</strong>mers. There are no predefined document types in SGML of course; the<br />

identification of a document type at an appropriate level of specificity, (e.g., poem,<br />

sonnet, petrarchan sonnet, sonnet-by-Shakespeare) is up <strong>to</strong> the SGML markup language<br />

designer.<br />

A document type definition (DTD) defines an SGML markup language for a particular<br />

document type. Part of a document type definition is given in a document type<br />

declaration that consists of a series of SGML markup declarations that formally define<br />

the vocabulary and syntax of the markup language. These markup declarations specify<br />

such things as what element types there are, what combinations elements can occur in,<br />

what characteristics may be attributed <strong>to</strong> elements, abbreviations for data, and so on.<br />

The SGML standard is clear that there is more <strong>to</strong> a document type definition than the<br />

markup declarations of the document type declaration: "Parts of a document type<br />

definition can be specified by a SGML document type declaration, other parts, such as<br />

the semantics of elements and attributes or any application conventions … cannot be


expressed formally in SGML" (ISO 1986: 4.105). That is, the document type<br />

declarations can tell us that a poem consists of exactly one title followed by one or more<br />

lines and that lines can't contain titles, but it does not have any resources for telling us<br />

what a title is. The document type definition encompasses more. However, all that the<br />

SGML standard itself says about how <strong>to</strong> provide the required additional information<br />

about "semantics … or any application conventions" is: "Comments may be used … <strong>to</strong><br />

express them informally" (ISO 1986: 4.105). In addition, it is generally thought that<br />

properly written prose documentation for a markup language is an account of the<br />

"semantics … or any application conventions" of that markup language.<br />

Obviously these two expressions "document type definition" and "document type<br />

declaration", which are closely related, sound similar, and have the same initials, can<br />

easily be confused. But the difference is important, if irregularly observed. We reiterate<br />

that: it is the document type definition that defines a markup language; part of that<br />

definition is given in the markup declarations of the document type declaration; and an<br />

additional portion, which includes an account of the meaning of the elements' attributes,<br />

is presented informally in comments or in the prose documentation of the markup<br />

language. We note that according <strong>to</strong> the SGML standard the acronym "DTD" stands for<br />

"document type definition" which accords poorly with the common practice of referring<br />

<strong>to</strong> the markup declarations of the document type declaration as, collectively, the "DTD",<br />

and using the extension "DTD" for files containing the markup declarations – these being<br />

only part of the DTD.<br />

A document instance is "the data and markup for a hierarchy of elements that conforms <strong>to</strong><br />

a document type definition" (ISO 1986: 4.100, 4.160). Typically, we think of the file of<br />

text and markup, although in fact the definition, like SGML in general, is indifferent <strong>to</strong><br />

the physical organization of information: an SGML document instance may be organized<br />

as a "file", or it may be organized some other way.<br />

Strictly speaking (i.e., according <strong>to</strong> ISO 1986, definitions 4.100 and 4.160), a document<br />

instance by definition actually conforms <strong>to</strong> the document type definition that it is<br />

putatively an instance of. This means that (i) the requirements for element and attribute<br />

use expressed in the markup declarations are met; and (ii) any other expressed semantic<br />

constraints or applications are followed. However, actual usage of "document instance"<br />

deviates from this definition in three ways: (i) it is common <strong>to</strong> speak of text and markup<br />

as a "document instance" even if it in fact fails <strong>to</strong> conform <strong>to</strong> the markup declarations of<br />

the intended DTD (as in "The document instance was invalid"); (ii) text plus markup is<br />

referred <strong>to</strong> as a document instance even if it is not clear what if any DTD it conforms <strong>to</strong>;<br />

and (iii) even a terminological purist wouldn't withhold the designation "document<br />

instance" because of a failure <strong>to</strong> conform <strong>to</strong> the "semantics and application conventions"<br />

of the document type definition – partly, no doubt, because there is no established way <strong>to</strong><br />

au<strong>to</strong>matically test for semantic conformance.


His<strong>to</strong>ry of SGML and XML: Part II<br />

Although SGML was soon used extensively in technical documentation and other largescale<br />

publishing throughout the world, it did not have the broad penetration in<strong>to</strong><br />

consumer publishing and text processing that some of its proponents had expected. Part<br />

of the problem was that it was during this same period that "WYSIWIG" word processing<br />

and desk<strong>to</strong>p publishing emerged, and had rapid and extensive adoption. Such software<br />

gave users a sense that they were finally (or once again?) working with "the text itself",<br />

without intermediaries, and that they had no need for markup systems; they could, again,<br />

return <strong>to</strong> familiar typewriter-based practices, perhaps enhanced with "cut-and-paste." This<br />

was especially frustrating for SGML enthusiasts because there is no necessary real<br />

opposition between WYSIWYG systems and descriptive markup systems and no reason,<br />

other than current market forces and perhaps technical limitations, why a WYSIWYG<br />

system couldn't be based on the OHCO model of text, and s<strong>to</strong>re its files in SGML. Most<br />

of what users valued in WYSIWYG systems – immediate formatting, menus, interactive<br />

support, etc. – was consistent with the SGML/OHCO approach. Users would of course<br />

need <strong>to</strong> forgo making formatting changes "directly", but on the other hand they would get<br />

all the advantages of descriptive markup, including simpler composition, context-oriented<br />

support for selecting document components, global formatting, and so on. There would<br />

never, needless <strong>to</strong> say, be any need <strong>to</strong> remember, type, or even see, markup tags. The<br />

"WYSIWYG" classification was misleading in any case, since what was at issue was<br />

really as much the combination of editing conveniences such as interactive formatting,<br />

pull-down menus, and graphical displays, as it was a facsimile of how the document<br />

would look when printed – which would typically vary with the circumstances, and often<br />

was not up <strong>to</strong> the author in any case.<br />

But very few SGML text processing systems were developed, and fewer still for general<br />

users. Several explanations have been put forward for this. One was that OHCO/SGML -<br />

based text processing was simply <strong>to</strong>o unfamiliar, even if more efficient and more<br />

powerful, and the unfamiliarity, and prospect of spending even an hour or two learning<br />

new practices posed a marketing problem for salesmen trying <strong>to</strong> make a sale "in three<br />

minutes on the floor of Radio Shack." Another was that OHCO/SGML software was just<br />

<strong>to</strong>o hard <strong>to</strong> develop: the SGML standard had so many optional features, as well as several<br />

characteristics that were hard <strong>to</strong> program, that these were barriers <strong>to</strong> development of<br />

software. Several OHCO/SGML systems were created in the 1980s, but these were not<br />

widely used.<br />

The modest use of SGML outside of selected industries and large organizations changed<br />

radically with the emergence of HTML, the HyperText Markup Language, on the World<br />

Wide Web. HTML was designed as an SGML language (and with tags specifically<br />

modeled on the "starter tag set" of IBM DCF/GML). However, as successful as HTML<br />

was it became apparent that it had many problems, and that the remedies for these<br />

problems could not be easily developed without changes in SGML.<br />

From the start HTML's relationship with SGML was flawed. For one thing HTML began<br />

<strong>to</strong> be used even before there was a DTD that defined the HTML language. More


important, though, was the fact that HTML indiscriminately included element types not<br />

only for descriptive markup, but also procedural markup as well ("font", "center", and so<br />

on). In addition there was no general stylesheet provision for attaching formatting or<br />

other processing HTML. Finally none of the popular web browsers ever validated the<br />

HTML they processed. Web page authors had only the vaguest idea what the syntax of<br />

HTML actually was and had little motivation <strong>to</strong> learn, as browsers were forgiving and<br />

always seemed <strong>to</strong> more or less "do the right thing." DTDs seemed irrelevant and<br />

validation unnecessary. HTML files were in fact almost never valid HTML.<br />

But perhaps the most serious problem was that the HTML element set was impoverished.<br />

If Web publishing were going <strong>to</strong> achieve its promise it would have <strong>to</strong> accommodate more<br />

sophisticated and specialized markup languages. That meant that it would need <strong>to</strong> be<br />

easier <strong>to</strong> write the software that processed DTDs, so that arbitrary markup vocabularies<br />

could be used without prior coordination. And it was also obvious that some provision<br />

needed <strong>to</strong> be made for allowing more reliable processing of document instances without<br />

DTDs.<br />

This was the background for the development of XML 1.0, the "Extensible Markup<br />

Language." XML was developed within the World Wide Web Consortium (W3C) with<br />

the first draft being released in 1996, the "Proposed Recommendation" in 1997, and the<br />

"Recommendation" in 1998. The committee was chaired by Jon Bosak of Sun<br />

Microsystems, and technical discussions were conducted in a working group of 100 <strong>to</strong><br />

150 people, with final decisions on the design made by an edi<strong>to</strong>rial review board with<br />

eleven members. The edi<strong>to</strong>rs, leading a core working group of eleven people, supported<br />

by a larger group of 150 or so experts in the interest group, were Michael Sperberg-<br />

McQueen (University of Illinois at Chicago and the Text Encoding Initiative), Tim Bray<br />

(Textuality and Netscape), and Jean Paoli (Microsoft).<br />

XML<br />

The principal overall goal of XML was <strong>to</strong> ensure that new and specialized markup<br />

languages could be effectively used on the Web, without prior coordination between<br />

software developers and content developers. One of the intermediate objectives <strong>to</strong>wards<br />

this end was <strong>to</strong> create a simpler, more constrained version of SGML, so that with fewer<br />

options <strong>to</strong> support and other simplifications, it would be easier for programmers <strong>to</strong><br />

develop SGML/XML software, and then more SGML software would be developed and<br />

used, and would support individualized markup languages. This objectively was<br />

achieved: the XML specification is about 25 pages long, compared <strong>to</strong> SGML's 155 (664<br />

including the commentary in Charles Goldfarb's SGML Handbook), and while the<br />

Working Group did not achieve their stated goal that a graduate student in computer<br />

science should be able <strong>to</strong> develop an XML parser in a week, the first graduate student <strong>to</strong><br />

try it reported success in a little under two weeks.<br />

Another objective was <strong>to</strong> allow for processing of new markup languages even without a<br />

DTD. This requires several additional constraints on the document instance, one of which<br />

is illustrative and will be described here. SGML allowed markup tags <strong>to</strong> be omitted when


they were implied by their markup declarations. For instance if the element type<br />

declaration for paragraph did not allow a paragraph inside of a paragraph, then the starttag<br />

of a new paragraph would imply the closing of the preceding paragraph even without<br />

an end-tag, and so the end-tag could be omitted by the content developer as it could be<br />

inferred by the software. But without a DTD it is not possible <strong>to</strong> know what can be nested<br />

and what can't – and so in order <strong>to</strong> allow "DTD-less" processing XML does not allow<br />

tags <strong>to</strong> be omitted. It is the requirement that elements have explicit end-tags that is<br />

perhaps the best-known difference between documents in an SGML language like HTML<br />

and an XML language like XHTML.<br />

Whereas SGML has just one characterization of conformance for document instances,<br />

XML has two: (i) all XML documents must be well-formed; (ii) an XML document may<br />

or may not be valid (vis-a-vis a DTD). To be well-formed, a document instance need not<br />

conform <strong>to</strong> a DTD, but it must meet other requirements, prominent and illustrative among<br />

them: no start-tags or end-tags may be omitted, elements must not "overlap", attribute<br />

values must be in quotation marks, and case must be consistent. These requirements<br />

ensure that software processing the document instance will be able <strong>to</strong> unambiguously<br />

determine a hierarchy (or tree) of elements and their attribute assignments, even without a<br />

DTD♭.<br />

The Text Encoding Initiative<br />

Background<br />

The practice of creating machine-readable texts <strong>to</strong> support <strong>humanities</strong> research began<br />

early and grew rapidly. Literary text encoding is usually dated from 1949, when Father<br />

Rober<strong>to</strong> Busa began using IBM punched-card equipment for the Index Thomisticus - in<br />

other words, literary text encoding is almost coeval with <strong>digital</strong> computing itself. By the<br />

mid-1960s there were at least three academic journals focusing on <strong>humanities</strong> computing,<br />

and a list of "Literary Works in Machine-Readable Form" published in 1966 already<br />

required 25 pages (and included some projects, listed modestly as single items, that were<br />

encoding an entire authorial oeuvre) (Carlson 1967). It is tempting <strong>to</strong> speculate that<br />

efforts <strong>to</strong> standardize encoding practices must have begun as soon as there was more than<br />

one encoding project. Anxiety about the diversity of encoding systems appears early –<br />

one finds that at a 1965 conference on computers and literature for instance, an<br />

impromptu meeting was convened <strong>to</strong> discuss "the establishment of a standard format for<br />

the encoding of text … a matter of great importance." It obviously wasn't the first time<br />

such a concern had been expressed, and it wouldn't be the last (Kay 1965).<br />

At first, of course, the principal problem was consistent identification and encoding of the<br />

characters needed for the transcription of literary and his<strong>to</strong>rical texts. But soon it included<br />

encoding of structural and analytic features as well. A standard approach <strong>to</strong> literary text<br />

encoding would have a number of obvious advantages; it would make it easier for<br />

projects and researchers <strong>to</strong> share texts, possible <strong>to</strong> use the same software across textual<br />

corpora (and therefore more economical <strong>to</strong> produce such software), and it would simplify<br />

the training of encoders – note that these advantages are similar <strong>to</strong> those, described


above, which motivated the effort <strong>to</strong> develop standards in commercial electronic<br />

publishing. And one might hope that without the disruptive competitive dynamic of<br />

commercial publishing where formats are sometimes aspects of competitive strategy, it<br />

would be easier <strong>to</strong> standardize. But there were still obstacles. For one thing, given how<br />

specialized text encoding schemes sometimes were, and how closely tied <strong>to</strong> the specific<br />

interests and views – even disciplines and theories – of their designers, how could a<br />

single common scheme be decided on?<br />

Origins<br />

The TEI had its origins in early November, 1987, at a meeting at Vassar College,<br />

convened by the Association for Computers in the Humanities and funded by the<br />

National Endowment for the Humanities. It was attended by thirty-two specialists, from<br />

many different disciplines and representing professional societies, libraries, archives, and<br />

projects in a number of countries in Europe, North America, and Asia. There was a sense<br />

of urgency, as it was felt that the proliferation of needlessly diverse and often poorly<br />

designed encoding systems threatened <strong>to</strong> block the development of the full potential of<br />

computers <strong>to</strong> support <strong>humanities</strong> research.<br />

The resulting "Poughkeepsie Principles" defined the project of developing a set of text<br />

encoding guidelines. This work was then undertaken by three sponsoring organizations:<br />

The Association for Computers in the Humanities, the Association for Literary and<br />

Linguistic Computing, and the Association for Computational Linguistics. A Steering<br />

Committee was organized and an Advisory Board of delegates from various professional<br />

societies was formed. Two edi<strong>to</strong>rs were chosen: Michael Sperberg-McQueen of the<br />

University of Illinois at Chicago, and Lou Burnard of Oxford University. Four working<br />

committees were formed and populated. By the end of 1989 well over fifty scholars were<br />

already directly involved and the size of the effort was growing rapidly.<br />

The first draft ("PI") of the TEI Guidelines was released in June, 1990. Another period of<br />

development followed (the effort now expanded and reorganized in<strong>to</strong> 15 working groups)<br />

releasing revisions and extensions throughout 1990–93. Early on in this process a number<br />

of leading <strong>humanities</strong> textbase projects began <strong>to</strong> use the draft Guidelines, providing<br />

valuable feedback and ideas for improvement. At the same time workshops and seminars<br />

were conducted <strong>to</strong> introduce the Guidelines as widely as possible and ensure a steady<br />

source of experience and ideas. During this period, comments, corrections, and requests<br />

for additions arrived from around the world. After another round of revisions and<br />

extensions the first official version of the Guidelines ("P3") was released in May, 1994.<br />

The TEI Guidelines have been an enormous success and <strong>to</strong>day nearly every <strong>humanities</strong><br />

textbase project anywhere in the world uses TEI. In 2002 the newly formed TEI<br />

Consortium released "P4", a revision that re-expressed the TEI Guidelines in XML, and<br />

<strong>to</strong>day the Consortium actively continues <strong>to</strong> evolve the Guidelines and provide training,<br />

documentation, and other support. After HTML, the TEI is probably the most extensively<br />

used SGML/XML text encoding system in academic applications.


The his<strong>to</strong>ry of the development of the TEI Guidelines, briefly <strong>to</strong>ld here, makes evident an<br />

important point. The Guidelines were, and continue <strong>to</strong> be, the product of a very large<br />

international collaboration of scholars from many disciplines and nations, with many<br />

different interests and viewpoints, and with the active participation of other specialists<br />

from many different professions and institutions. The work was carried out over several<br />

years within a disciplined framework of development, evaluation, and coordination, and<br />

with extensive testing in many diverse projects. All of this is reflected in a text encoding<br />

system of extraordinary power and subtlety.<br />

General nature and structure<br />

The TEI Guidelines take the form of several related SGML document type definitions,<br />

specified both in formal markup declarations and in English prose. The main DTD, for<br />

the encoding of conventional textual material such as novels, poetry, plays, dictionaries,<br />

or term-bank data, is accompanied by a number of auxiliary DTDs for specialized data:<br />

tag-set documentation, feature-system declarations used in structured linguistic or other<br />

annotation, writing-system declarations, and free-standing bibliographic descriptions of<br />

electronic texts.<br />

The main DTD itself is partitioned in<strong>to</strong> eighteen discrete tag sets or modules, which can<br />

be combined in several thousand ways <strong>to</strong> create different views of the main DTD. Some<br />

tag sets are always included in a view (unless special steps are taken <strong>to</strong> exclude them),<br />

some (the "base" tag sets) are mutually exclusive, and some (the "additional" tag sets)<br />

may be combined with any other tag sets.<br />

In addition <strong>to</strong> the flexibility of choosing among the various tag sets, the main TEI DTD<br />

takes a number of steps <strong>to</strong> provide flexibility for the encoder and avoid a Procrustean<br />

rigidity that might interfere with the scholarly judgment of the encoder. Individual<br />

element types may be included or excluded from the DTD, additional element types may<br />

be introduced (as long as they are properly documented), and the names of element types<br />

and attributes may be translated in<strong>to</strong> other languages for the convenience of the encoder<br />

or user.<br />

It cannot be repeated enough that the apparent complexity of the TEI is, in a sense, only<br />

apparent. For the most part only as much of the TEI as is needed will be used in any<br />

particular encoding effort; the TEI vocabulary used will be exactly as complex, but no<br />

more complex, than the text being encoded. Not surprisingly, a very small TEI<br />

vocabulary, known as TEI Lite (http://www.tei.org), is widely used for simple texts.<br />

Larger significance of the TEI<br />

The original motivation of TEI was <strong>to</strong> develop interchange guidelines that would allow<br />

projects <strong>to</strong> share textual data (and theories about that data) and promote the development<br />

of common <strong>to</strong>ols. Developing such a language for the full range of human written culture,<br />

the full range of disciplinary perspectives on those objects, and the full range of<br />

competing theories was a daunting task.


It is easy <strong>to</strong> talk about accommodating diversity, about interdisciplinarity, about multiculturalism,<br />

about communication across various intellectual gaps and divides. But few<br />

efforts along these lines are more than superficial. The experience of the TEI makes it<br />

evident why this is so. Not only do different disciplines have quite different interests and<br />

perspectives, but, it seems, different conceptual schemes: fundamentally different ways<br />

of dividing up the world. What is an object of critical contest and debate for one<br />

discipline, is theory-neutral data for another, and then completely invisible <strong>to</strong> a third.<br />

What is structured and composite for one field is a<strong>to</strong>mic for another and an irrelevant<br />

conglomeration <strong>to</strong> a third. Sometimes variation occurred within a discipline, sometimes<br />

across his<strong>to</strong>rical periods of interest, sometimes across national or professional<br />

communities. Practices that would seem <strong>to</strong> have much in common could vary radically –<br />

and yet have enough in common for differences <strong>to</strong> be a problem! And even where<br />

agreement in substance was obtained, disagreements over nuances, of terminology for<br />

instance, could derail a tenuous agreement.<br />

(Mylonas and Renear 1999)<br />

The TEI made several important decisions that gave the enterprise at least the chance of<br />

success. For one thing the TEI determined that it would not attempt <strong>to</strong> establish in detail<br />

the "recognition criteria" for TEI markup; it would be the business of the encoderscholar,<br />

not that of the TEI, <strong>to</strong> specify precisely what criteria suffice <strong>to</strong> identify a<br />

paragraph, or a section, or a technical term, in the text being transcribed. Examples of<br />

features commonly taken <strong>to</strong> indicate the presence of an object were given for illustration,<br />

but were not defined as normative. In addition the TEI, with just a few trivial exceptions,<br />

does not specify what is <strong>to</strong> be identified and encoded. Again, that is the business of the<br />

scholar or encoder. What the TEI does is provide a language <strong>to</strong> be used when the encoder<br />

(i) recognizes a particular object, and (ii) wishes <strong>to</strong> identify that object. In this sense, the<br />

TEI doesn't require antecedent agreement about what features of a text are important and<br />

how <strong>to</strong> tell whether they are present; instead, it makes it possible, when the scholar<br />

wishes, <strong>to</strong> communicate a particular theory of what the text is. One might say that the TEI<br />

is an agreement about how <strong>to</strong> express disagreement.<br />

The principal goal of the TEI, developing an interchange language that would allow<br />

scholars <strong>to</strong> exchange information, was ambitious enough. But the TEI succeeded not only<br />

in this, but at a far more difficult project, the development of a new data description<br />

language that substantially improves our ability <strong>to</strong> describe textual features, not just our<br />

ability <strong>to</strong> exchange descriptions based on current practice. The TEI Guidelines represent<br />

an elucidation of current practices, methods, and concepts, [that] opens the way <strong>to</strong> new<br />

methods of analysis, new understandings, and new possibilities for representation and<br />

communication. Evidence that this is indeed a language of new expressive capabilities<br />

can be found in the experience of pioneering textbase projects which draw on the<br />

heuristic nature of the TEI Guidelines <strong>to</strong> illuminate textual issues and suggest new<br />

analyses and new techniques.<br />

(Mylonas and Renear 1999)


Finally, we note that the TEI is now itself a research community,<br />

connecting many professions, disciplines, and institutions in many countries and defining<br />

itself with its shared interests, concepts, <strong>to</strong>ols, and techniques. Its subject matter is textual<br />

communication, with the principal goal of improving our general theoretical<br />

understanding of textual representation, and the auxiliary practical goal of using that<br />

improved understanding <strong>to</strong> develop methods, <strong>to</strong>ols, and techniques that will be valuable<br />

<strong>to</strong> other fields and will support practical applications in publishing, archives, and<br />

libraries. It has connections <strong>to</strong> knowledge representation systems (formal semantics and<br />

on<strong>to</strong>logy, objection orientation methodologies, etc.), new theorizing (non-hierarchical<br />

views of text, antirealism, etc.), and new applications and <strong>to</strong>ols. ♭ providing new insights<br />

in<strong>to</strong> the nature of text, and new techniques for exploiting the emerging information<br />

technologies.<br />

(Mylonas and Renear 1999)<br />

Concluding Remarks<br />

Text encoding has proven an unexpectedly exciting and illuminating area of activity, and<br />

we have only been able <strong>to</strong> <strong>to</strong>uch on a few aspects here. Ongoing work is of many kinds<br />

and taking place in many venues, from experimental hypermedia systems, <strong>to</strong> improved<br />

techniques for literary analysis, <strong>to</strong> philosophical debates about textuality. The references<br />

below should help the reader explore further.<br />

For Further Information<br />

SGML, XML, and related technologies<br />

The best introduction <strong>to</strong> XML is the second chapter of the TEI Guidelines, "A Gentle<br />

Introduction <strong>to</strong> XML", online at .<br />

An excellent single-volume presentation of a number of XML-related standards, with<br />

useful practical information, is Harold and Means (<strong>2001</strong>). See also Bradley (<strong>2001</strong>).<br />

The SGML standard is available as Goldfarb (1991), a detailed commentary that includes<br />

a cross-reference annotated version of the specification. A treatment of a number of<br />

subtleties is DeRose (1997).<br />

A good detailed presentation of the XML specification is Graham and Quin (1999).<br />

There are two extremely important resources on the Web for anyone interested in any<br />

aspect of contemporary SGML/XML text encoding. One is the extraordinary "Cover<br />

Pages", http://xml.coverpages.org, edited by Robin Cover and hosted by OASIS, an<br />

industry consortium. These provide an unparalleled wealth of information and news<br />

about XML and related standards. The other is the W3C site itself, http://www.w3c.org.


These will provide an entry <strong>to</strong> a vast number of other sites, mailing lists, and other<br />

resources. Two other websites are also valuable: Cafe con Leche<br />

http://www.ibiblio.org/xml/, and XML.com http://www.xml.com. The most advanced<br />

work in SGML/ XML markup research is presented at IDEAlliances' annual Extreme<br />

Markup Languages conference; the proceedings are available online, at<br />

http://www.idealliance.org/papers/extreme03/.<br />

Document analysis, data modeling, DTD design, project design<br />

Unfortunately there is a lack of intermediate and advanced material on the general issues<br />

involved in analyzing document types and developing sound DTD designs. There is just<br />

one good book-length treatment: Maler and Andaloussi (1996), although there are<br />

introduc<strong>to</strong>ry discussions from various perspectives in various places (e.g., Bradley <strong>2001</strong>).<br />

Other relevant discussions can be found in various places on the Web (see, for example,<br />

the debate over elements vs. attributes at<br />

http://xml.coverpages.org/elementsAndAttrs.html) and in various discussion lists and the<br />

project documentation and reports of encoding projects. (There are however very good<br />

resources for learning the basic techniques of TEI-based <strong>humanities</strong> text encoding; see<br />

below.)<br />

Humanities text encoding and the TEI<br />

For an excellent introduction <strong>to</strong> all aspects of contemporary <strong>humanities</strong> text encoding see<br />

Electronic Texts in the Humanities: Theory and Practice, by Susan Hockey (Hockey<br />

<strong>2001</strong>).<br />

The TEI Consortium, http://www.tei-c.org, periodically sponsors or endorses workshops<br />

or seminars in text encoding using the Guidelines; these are typically of high quality and<br />

a good way <strong>to</strong> rapidly get up <strong>to</strong> speed in <strong>humanities</strong> text encoding. There are also selfpaced<br />

tu<strong>to</strong>rials and, perhaps most importantly, a collection of project descriptions and<br />

encoding documentation from various TEI projects. See: http://www.tei-c.org/Talks/ and<br />

http://www.tei-c.org/Tu<strong>to</strong>rials.index.html. Anyone involved in TEI text encoding will<br />

want <strong>to</strong> join the TEI-L discussion list.<br />

Theoretical issues<br />

Text encoding has spawned an enormous number and wide variety of theoretical<br />

discussions and debates, ranging from whether the OHCO approach neglects the<br />

"materiality" of text, <strong>to</strong> whether TEI markup is excessively "interpretative"; <strong>to</strong> whether<br />

texts are really "hierarchical", <strong>to</strong> the political implications of markup, <strong>to</strong> whether<br />

SGML/XML is flawed by the lack of an algebraic "data model." From within the<br />

<strong>humanities</strong> computing community some of the influential researchers on these <strong>to</strong>pics<br />

include Dino Buzzetti, Paul Ca<strong>to</strong>n, Glaus Huitfeldt, Julia Flanders, Jerome McGann,<br />

Michael Sperberg-McQueen, Alois Pichler, and Wendell Piez, among others. A<br />

compendium of these <strong>to</strong>pics, with references <strong>to</strong> the original discussions, may be found at<br />

http://www.isrl.uiuc.edu/eprg/markuptheoryreview.html.


One theoretical <strong>to</strong>pic promises <strong>to</strong> soon have a broad impact beyond as well as within the<br />

<strong>digital</strong> <strong>humanities</strong> and deserves special mention. There has recently been a renewed<br />

concern that SGML/XML markup itself is not a "data model" or a "conceptual model", or<br />

does not have a "formal semantics." This is a criticism that on this view SGML/XML<br />

markup serializes a data structure, but it does not express the meaning of that data<br />

structure (the specification of the conceptual document) in a sufficiently formal way. For<br />

an introduction <strong>to</strong> this <strong>to</strong>pic see Renear et al. (2002) and Robin Cover (1998, 2003). For a<br />

related discussion from the perspective of literary theory see Buzzetti (2002). Obviously<br />

these issues are closely related <strong>to</strong> those currently being taken up in the W3C Semantic<br />

Web Activity (http://www.w3.org/<strong>2001</strong>/sw/Activity).<br />

Acknowledgments<br />

The author is deeply indebted <strong>to</strong> Michael Sperberg-McQueen for comments and<br />

corrections. Remaining errors and omissions are the author's alone.<br />

Bibliography<br />

Bradley, N. (<strong>2001</strong>). The XML Companion. Reading, MA: Addison-Wesley.<br />

Buzzetti, D. (2002). Digital Representation and the Text Model. New Literary His<strong>to</strong>ry 33:<br />

61–88.<br />

Carlson, G. (1967). Literary Works in Machine-Readable Form. Computers and the<br />

Humanities 1: 75–102.<br />

Carmody, S., W Gross, T. H. Nelson, D. Rice, and A. van Dam (1969). A Hypertext<br />

Editing System for the 360. In M. Faiman and J. Nievergelt (eds.), Pertinent Concepts in<br />

Computer Graphics (pp. 291–330). Champaign: University of Illinois Press.<br />

Chicago (1993). The Chicago Manual of Style: The Essential Guide for Writers, Edi<strong>to</strong>rs,<br />

and Publishers, 14th edn. Chicago: University of Chicago Press.<br />

Coombs, J.-H., A.-H. Renear, and S.-J. DeRose (1987). Markup Systems and the Future<br />

of Scholarly Text Processing. Communications of the Association for Computing<br />

Machinery 30: 933–47.<br />

Cover, R. (1998). XML and Semantic Transparency. OASIS Cover Pages. At<br />

http://www.oasis-open.org/cover/xmlAndSemantics.html.<br />

Cover, R. (2003). Conceptual Modeling and Markup Languages. OASIS Cover Pages. At<br />

http://xml.coverpages.org/conceptualModeling.html.<br />

DeRose, S. J. (1997). The SGML FAQ Book: Understanding the Foundation of HTML<br />

and XML. Bos<strong>to</strong>n: Kluwer.


DeRose, S.-J., and A. van Dam (1999). Document Structure and Markup in the PRESS<br />

Hypertext System. Markup Languages: Theory and Practice 1: 7–32.<br />

DeRose, S.-J., D. Durand, E. Mylonas, and A.-H. Renear (1990). What Is Text, Really?<br />

Journal of Computing in Higher Education 1: 3–26. Reprinted in the ACM/SIGDOC<br />

*Journal of Computer Documentation 21,3: 1–24.<br />

Engelbart, D. C., Watson, R. W. and Nor<strong>to</strong>n, J. C. (1973). The Augmented Knowledge.<br />

Workshop: AFIPS Conference Proceedings, vol. 42, National Computer Conference,<br />

June 4–8, 1973, pp. 9–21.<br />

Furuta, R., J. Scofield, and A. Shaw (1982). Document Formatting Systems: Survey,<br />

Concepts, and Issues. ACM Computing Surveys 14: 417–72.<br />

Goldfarb, C.-F. (1978). Document Composition Facility: Generalized Markup Language<br />

(GML) Users Guide. IBM General Products Division.<br />

Goldfarb, C.-F. (1981). A Generalized Approach <strong>to</strong> Document Markup. [Proceedings of<br />

the (ACMj (SIGPLAN- SIGOA] Symposium on Text Manipulation (pp. 68–73). New<br />

York: ACM.<br />

Goldfarb, C.-F. (1991). The SGML Handbook Guide. Oxford: Oxford University Press.<br />

Goldfarb, C.-F. (1997). SGML: The Reason Why and the First Publishing Hint. Journal<br />

of the American Society for Information Science 48.<br />

Graham, I. G. and L. Quin (1999). The XML Specification Guide. New York: John Wiley.<br />

Harold, E. R. and W. S. Means (<strong>2001</strong>). XML in a Nutshell: A Quick Desk<strong>to</strong>p Reference.<br />

Sebas<strong>to</strong>pol, CA: O'Reilly & Associates.<br />

Hockey, S. M. (<strong>2001</strong>). Electronic Texts in the Humanities: Theory and Practice. Oxford:<br />

Oxford University Press.<br />

IBM (1967). Application Description, System/360 Document Processing System. White<br />

Plains, NY: IBM.<br />

Ide, N.-M. and C.-M. Sperberg-McQueen (1997). Toward a Unified Docuverse:<br />

Standardizing Document Markup and Access without Procrustean Bargains. Proceedings<br />

of the 60th Annual Meeting of the American Society for Information Science, ed. C.<br />

Schwartz and M. Rorvig (pp. 347–60). Medford, NJ: Learned Information Inc..<br />

ISO (1986). Information Processing - Text and Office Systems - Standard Generalized<br />

Markup Language (SGML). ISO 8879–1986 (E). Geneva: International Organization for<br />

Standardization.


Kay, M. (1965). Report on an Informal Meeting on Standard Formats for Machinereadable<br />

Text. In J. B. Bessinger and S. M. Parrish (eds), Literary Data Processing<br />

Conference Proceedings (pp. 327–28). White Plains, NY: IBM.<br />

Knuth, D. E. (1979). TEX and Metafont. New Directions in Typesetting. Bedford, MA:<br />

Digital Press.<br />

Lamport, L. (1985). LaTeX - A Document Preparation System. Reading, MA: Addison-<br />

Wesley.<br />

Lesk, M.-E. (1978). Typing Documents on the UNIX System: Using the -ms Macros with<br />

Troft and Nroff. Murray Hill, NJ: Bell Labora<strong>to</strong>ries.<br />

Maler, E. and E. L. Andaloussi (1996). Developing SGML DTDs: From Text <strong>to</strong> Model <strong>to</strong><br />

Markup. Englewood Cliffs, NJ: Prentice Hall.<br />

Mylonas, M. and Renear, A. (1999). The Text Encoding Initiative at 10: Not Just an<br />

Interchange Format Anymore - but a New Research Community. Computers and the<br />

Humanities 33: 1–9.<br />

Ossanna, J.-F. (1976). NROFF/TROFF User's Manual (Tech. Rep.-54). Murray Hill, NJ:<br />

Bell Labora<strong>to</strong>ries.<br />

Reid, B.-K. (1978). Scribe Introduc<strong>to</strong>ry User's Manual. Pittsburgh, PA: Carnegie-Mellon<br />

University, Computer Science Department.<br />

Reid, B.-K. (1981). Scribe: A Document Specification Language and its Compiler. PhD<br />

thesis. Pittsburgh, PA: Carnegie-Mellon University. Also available as Technical Report<br />

CMU-CS-81–100.<br />

Renear, A. H. (2000). The Descriptive/Procedural Distinction is Flawed. Markup<br />

Languages: Theory and Practice 2: 411–20.<br />

Renear, A. H., E. Mylonas, and D.-G. Durand (1996). Refining Our Notion of What Text<br />

Really Is: The Problem of Overlapping Hierarchies. In S. Hockey and N. Ide (eds.),<br />

Research in Humanities Computing 4: Selected Papers from the ALLC/ACH Conference,<br />

Christ Church, Oxford, April, 1992 (pp. 263–80).<br />

Renear, A. H., D. Dubin, C. M. Sperberg-McQueen, and C. Huitfeldt (2002). Towards a<br />

Semantics for XML Markup. In R. Furuta, J. I. Maletic, and E. Munson (eds.),<br />

Proceedings of the 2002 ACM Symposium on Document Engineering (pp. 119–26).<br />

McLean, VA, November. New York: Association for Computing Machinery.<br />

Seybold, J. W. (1977). Fundamentals of Modern Composition. Media, PA: Seybold<br />

Publications.


SGML Users' Group (1990). A Brief His<strong>to</strong>ry of the Development of SGML. At<br />

http://www.sgmlsource.com/his<strong>to</strong>ry/sgmlhist.htm.<br />

Sperberg-McQueen, C.-M. (1991). Text Encoding and Enrichment. In Ian Lancashire<br />

(ed.), The Humanities Computing Yearbook 1989–90. Oxford: Oxford University Press.<br />

Sperberg-McQueen, M. and L. Burnard, (eds.) (1994). Guidelines for Text Encoding and<br />

Interchange (TEI P3). Chicago, Oxford: ACH/ALLC/ACL Text Encoding Initiative.<br />

Sperberg-McQueen, M. and L. Burnard, (eds.) (2002). Guidelines for Text Encoding and<br />

Interchange (TEI P4). Oxford, Providence, Charlottesville, Bergen: ACH/ALLC/ACL<br />

Text Encoding Initiative.<br />

Spring, M. B. (1989). The Origin and Use of Copymarks in Electronic Publishing.<br />

Journal of Documentation 45: 110–23.<br />

18.<br />

Electronic Texts: Audiences and Purposes<br />

Perry Willett<br />

His<strong>to</strong>ry<br />

… <strong>to</strong>day, even the most reluctant scholar has at least an indefinite notion of the<br />

computer's ability <strong>to</strong> s<strong>to</strong>re, manipulate and analyse natural language texts, whilst his less<br />

wary colleagues are likely <strong>to</strong> meet with a sympathetic reception if they approach their<br />

local computing centre with proposals for literary or linguistic research – some<br />

universities, indeed, have set up institutes and departments whose special task it is <strong>to</strong><br />

facilitate work of this kind.<br />

Wisbey, The Computer in Literary and Linguistic Research: Papers from a Cambridge<br />

Symposium, vii<br />

Claims of the ubiquity of electronic text, such as this one published in 1971, may seem<br />

exaggerated for an era before personal computers, before standards for character or text<br />

encoding, before the World Wide Web, yet they have come more or less true. What is<br />

remarkable about this and other writing on electronic texts in the <strong>humanities</strong> published<br />

before 1995 is how well they predicted the uses of electronic text, before many people<br />

were even aware that text could be electronic. It is important <strong>to</strong> remember the<br />

introduction of the World Wide Web in the mid-1990s in evaluating the statements of<br />

computing humanists, for prior <strong>to</strong> the Web's arrival, while a great deal was written on the<br />

uses of and audiences for electronic text, almost no one foresaw such a powerful <strong>to</strong>ol for<br />

the wide distribution of electronic texts, or that wide distribution for a general reading<br />

public would become the most successful use made of electronic texts.


It is well documented that the his<strong>to</strong>ry of electronic text is almost as long as the his<strong>to</strong>ry of<br />

electronic computing itself. Vannevar Bush famously imagined vast libraries available<br />

via the new technology in 1945. Father Rober<strong>to</strong> Busa began his pioneering effort <strong>to</strong><br />

create the Index Thomisticus, the monumental index <strong>to</strong> the works of St Thomas Aquinas,<br />

in 1946 (Busa 1950, 1992). Other pioneers adopted computers and advocated their use in<br />

literary and linguistic research, with electronic texts as the necessary first ingredient. The<br />

same early study quoted above divides the world of <strong>humanities</strong> computing in<strong>to</strong> several<br />

categories, as shown in its table of contents:<br />

1 Lexicographical, textual archives, and concordance making<br />

2 Textual editing and attribution studies<br />

3 Vocabulary studies and language learning<br />

4 Stylistic analysis and au<strong>to</strong>mated poetry generation<br />

This list encompasses all aspects of <strong>humanities</strong> scholarship, from literary analysis and<br />

author attribution studies, <strong>to</strong> scholarly editing, <strong>to</strong> rhe<strong>to</strong>ric and language studies, <strong>to</strong> the<br />

creation of archives of electronic text. While au<strong>to</strong>mated poetry generation has yet <strong>to</strong> find<br />

its audience, the other categories seem remarkably prescient. Humanities computing now<br />

includes media in other formats such as <strong>digital</strong> images, audio and video, yet early studies<br />

viewed electronic text as the starting point. Electronic text has indeed become common in<br />

<strong>humanities</strong> research and publishing. Many other early studies (Lusignan and North 1977;<br />

Hockey 1980; Bailey 1982) enthusiastically describe the potential that computers and<br />

electronic text hold for literary studies, and describe similar audiences.<br />

These scholars developed a vision of the importance of electronic texts for the<br />

<strong>humanities</strong>, and developed the standards by which they are created. Humanists, precisely<br />

because of their sophisticated understanding of text, have a central role in the<br />

development of standards driving the World Wide Web, with far-reaching implications<br />

for what can be done <strong>to</strong>day on the Internet.<br />

Still, the kind of environment described by Wisbey in 1971, with computing centers<br />

sympathetic <strong>to</strong> humanists' concerns, existed at only a handful of universities and research<br />

centers at that time. One was more likely <strong>to</strong> find scientists and social scientists at such<br />

computer centers than literary scholars or his<strong>to</strong>rians. Universities might have had at most<br />

one or two <strong>humanities</strong> professors interested in electronic texts, leaving computing<br />

humanists largely isolated, with annual conferences and specialist journals the only<br />

opportunities <strong>to</strong> discuss issues, ideas, and developments with like-minded colleagues.<br />

Skepticism about the use of electronic texts in <strong>humanities</strong> research has a long his<strong>to</strong>ry<br />

also, and not just among traditionalists. Despite the best efforts of computing humanists,<br />

electronic text remained the domain of a few specialists in<strong>to</strong> the early 1990s. In 1993,<br />

Mark Olsen wrote:


[c]omputer processing of textual data in literary and his<strong>to</strong>rical research has expanded<br />

considerably since the 1960s. In spite of the growth of such applications, however, it<br />

would seem that computerized textual research has not had a significant influence on<br />

research in humanistic disciplines and that literature research has not been subject <strong>to</strong> the<br />

same shift in perspective that accompanied computer-assisted research in the social<br />

science oriented disciplines, such as his<strong>to</strong>ry. … In spite of the investment of significant<br />

amounts of money and time in many projects, the role of electronic text in literary<br />

research remains surprisingly limited.<br />

(Olsen 1993–4: 309)<br />

He goes on <strong>to</strong> list several reasons <strong>to</strong> support this assertion. Most importantly, Olsen notes<br />

a distrust of computing methodology among humanists at large, and believes that most<br />

research involving computing is <strong>to</strong>o narrowly focused <strong>to</strong> one or two authors. Olsen spoke<br />

not as a traditionalist, for he was (and remains) the direc<strong>to</strong>r of the pioneering collection of<br />

electronic text, the American and French Research on the Treasury of the French<br />

Language (ARTFL). He describes the potential of harnessing computers <strong>to</strong> large<br />

collections of literature, and the ability <strong>to</strong> trace concepts and important keywords over<br />

time both within and among various authors' works, along with more sophisticated kinds<br />

of analysis. He notes that in at least some disciplines, in the era prior <strong>to</strong> the World Wide<br />

Web, a critical mass of primary texts was available, pointing <strong>to</strong> the Thesaurus Linguae<br />

Graecae (TLG) for classics and ARTFL for modern French. Yet, he still concludes that<br />

the potential goes largely ignored. Alison Finch in 1995 displayed even more skepticism<br />

in her analysis of computing humanists, as she explored the heroic narratives invoked by<br />

literary researchers who use computers, while she claimed that these studies lacked<br />

importance <strong>to</strong> the larger field of literary studies.<br />

Other commenta<strong>to</strong>rs, less technologically savvy, see darker implications. The best known<br />

of these critics, Sven Birkerts in The Gutenberg Elegies (1994) and Nicholson Baker in<br />

Double Fold (<strong>2001</strong>), mourn the changes wrought by the development of electronic media,<br />

and fear that books, once decoupled from their physical presence, will lose their meaning<br />

and his<strong>to</strong>rical importance. Birkerts, also writing pre-World Wide Web, in particular fears<br />

that the digitization of books may lead <strong>to</strong> the "erosion of language" and a "flattening of<br />

his<strong>to</strong>rical perspectives" (Birkerts 1994: 128–9). He does not consider other possible<br />

outcomes, such as one in which general readers and scholars alike have a better sense of<br />

the concerns and ideas of peoples and his<strong>to</strong>rical periods with increased access <strong>to</strong> works<br />

otherwise available in only a few libraries. The development of <strong>digital</strong> collections does<br />

not require the destruction of books; instead, it may provoke more interest in their<br />

existence and provide different opportunities for their study through keyword and<br />

structured searching.<br />

More nuanced critiques recognize the disadvantages of <strong>digital</strong> technology, while<br />

exploring its uses and potentials. As Bornstein and Tinkle noted:<br />

We agree with the proposition that the shift from print <strong>to</strong> <strong>digital</strong> culture has analogies in<br />

scale and importance <strong>to</strong> the shift from manuscript <strong>to</strong> print culture beginning in the


Renaissance. Further, the change enables us <strong>to</strong> consider more strikingly than ever before<br />

the diverse characteristics of manuscript, print, and electronic cultures. This is<br />

particularly important now that new electronic technologies are making possible the<br />

production, display, and transmission of texts in multiple forms that far exceed the more<br />

fixed capacity of traditional codex books.<br />

(Bornstein and Tinkle 1998: 2)<br />

Most of these studies were written prior <strong>to</strong> widespread popularization of the World Wide<br />

Web. As a <strong>to</strong>ol, the Web does not solve all of the problems surrounding the creation,<br />

s<strong>to</strong>rage, delivery and display of electronic texts, but certainly simplifies many of them,<br />

particularly in comparison with the older technologies used for electronic text s<strong>to</strong>rage and<br />

delivery such as ftp, listserv, and Gopher. The early adopters, enthusiasts, and critics<br />

viewed electronic texts as the domain of scholars and researchers, and used electronic<br />

texts <strong>to</strong> assist in the traditional work of humanists as outlined in the table of contents<br />

reproduced above. They did not foresee the World Wide Web's capacity <strong>to</strong> reach a wider<br />

reading public.<br />

Has the world of electronic texts changed with the ubiquity of the World Wide Web, or<br />

are these criticisms still accurate? Put another way, who is the audience for electronic<br />

texts <strong>to</strong>day? A recent book on electronic text (Hockey 2000) provides a current view on<br />

their audience. Hockey lists these categories in her table of contents:<br />

1 Concordance and Text Retrieval Programs<br />

2 Literary Analysis<br />

3 Linguistic Analysis<br />

4 Stylometry and Attribution Studies<br />

5 Textual Critical and Electronic Editions<br />

6 Dictionaries and Lexical Databases<br />

On the whole, this list is not significantly different from the categories of uses and users<br />

in the book edited by Wisbey almost 30 years earlier. Hockey believes that "most of the<br />

present interest in electronic texts is focused on access" (Hockey 2000: 3), that is, the use<br />

of computers <strong>to</strong> s<strong>to</strong>re and deliver entire texts. In her view, access is the least interesting<br />

aspect of electronic texts, for it leaves largely unexploited their real power: the ability for<br />

texts <strong>to</strong> be searched and manipulated by computer programs.<br />

What is Electronic Text?<br />

Answering this simple question could involve textual and literary theory and their<br />

intersection with <strong>digital</strong> technology, as discussed elsewhere in this volume. I wish <strong>to</strong>


focus instead on practical considerations: what forms do electronic texts in the <strong>humanities</strong><br />

take?<br />

The first type <strong>to</strong> consider is an electronic transcription of a literary text, in which<br />

characters, punctuation, and words are faithfully represented in a computer file, allowing<br />

for keyword or contextual searching. Just what is meant by a "faithful representation" of a<br />

printed book is at the crux of a long debate, again explored elsewhere in this volume (see<br />

chapters 16, 17, and 22, for discussions of theoretical issues, and chapter 32 for more<br />

practical matters). A transcription in the form of a computer text file is the most compact<br />

form for electronic text, an important consideration given the severe constraints on<br />

computer s<strong>to</strong>rage and bandwidth that still exist for some people. In addition, this form<br />

provides greater ease of manipulation, for searching, for editing, and has significant<br />

advantages for people with visual impairments. Another basic form is a <strong>digital</strong> image of a<br />

physical page, allowing readers <strong>to</strong> see a representation of the original appearance, of<br />

great importance for textual scholars. This once seemed impractical given the amount of<br />

s<strong>to</strong>rage necessary for the hundreds of image files needed for even one book. Now, with<br />

ever greater s<strong>to</strong>rage available even on desk<strong>to</strong>p computers, this is much less of an issue<br />

(while bandwidth remains an issue). Some collections, such as those available through<br />

<strong>digital</strong> libraries at the Library of Congress, the University of Virginia, and the University<br />

of Michigan, combine the two forms.<br />

Another category is encoded text. For a multitude of reasons, many scholars and edi<strong>to</strong>rs<br />

believe that encoding schemes, such as that developed by the Text Encoding Initiative<br />

(TEI), provide the best and fullest representation of text in all its complexity, allowing the<br />

crea<strong>to</strong>r or edi<strong>to</strong>r an opportunity <strong>to</strong> encode the hierarchical structures and multitude of<br />

features found in text. In addition <strong>to</strong> the belief that explicit structural markup along with<br />

robust metadata will allow for greater longevity of electronic texts, those who choose the<br />

TEI use its many elements and features for sophisticated analysis and repurposing of<br />

electronic texts for electronic and print publishing. The people developing the TEI have<br />

used open standards such as SGML and XML, as well as providing extensive<br />

documentation <strong>to</strong> assist in the use of the standard, as a way of encouraging and promoting<br />

the preservation and interchange of e-texts.<br />

Others, notably people associated with Project Gutenberg, distrust all encoding schemes,<br />

believing that the lack of encoding will better allow e-texts <strong>to</strong> survive changes in<br />

hardware, operating systems, and application software. No special software and little<br />

training are required for contribu<strong>to</strong>rs <strong>to</strong> Project Gutenberg, and a volunteer ethos prevails.<br />

They will have over 10,000 titles at the end of 2003, but little can be said for the accuracy<br />

of the transcriptions, for there is no central edi<strong>to</strong>rial control. The lack of encoding means<br />

it would be impossible, for example, <strong>to</strong> separate notes from text, or <strong>to</strong> determine quickly<br />

where chapters begin and end, or <strong>to</strong> indicate highlighting and font shifts within the text.<br />

Still, Project Gutenberg's founder Michael Hart has tapped in<strong>to</strong> a volunteer spirit that<br />

drives most open source projects.<br />

Encoding and edi<strong>to</strong>rial standards remain expensive, and the more encoding <strong>to</strong> be added<br />

and editing performed, the higher the level of expertise about the original document and/


or the technology for searching and display is required. Projects such as the Women<br />

Writers Online project at Brown University and the Model Editions Partnership provide<br />

extensive documentation about edi<strong>to</strong>rial standards and practices. Other projects listed on<br />

the TEI Consortium website provide examples of equally detailed documentation and<br />

exacting edi<strong>to</strong>rial standards.<br />

Some projects have avoided this additional cost by converting page images <strong>to</strong> text using<br />

optical character recognition (OCR) software, and allowing readers <strong>to</strong> perform keyword<br />

searches against the unedited and unencoded text files. The results are displayed as <strong>digital</strong><br />

images of the original page, with the uncorrected text hidden. While imperfections<br />

impede precise and complete results from keyword searches, proponents believe that with<br />

adequate OCR accuracy, the results will still prove useful, and at a significantly lower<br />

overall cost <strong>to</strong> produce than highly accurate transcriptions. Some extremely large e-text<br />

collections with tens of thousand of volumes and millions of pages, such as the Making of<br />

America and American Memory collections, are based upon this idea, pioneered by John<br />

P. Wilkin, called "rough OCR" or "dirty OCR."<br />

Several people have pursued the idea of creating an archive of works containing<br />

manuscripts and variant editions, meant for both sophisticated researchers and general<br />

readers. In this type of publication, researchers would have access <strong>to</strong> all manuscript<br />

versions, allowing them <strong>to</strong> trace the development of a work through its variants, while a<br />

general audience could use a preferred reading created by the edi<strong>to</strong>rs and derived directly<br />

from the sources. This kind of archive, proposed for authors such as Yeats, Hardy, and<br />

others, would be a boon <strong>to</strong> both audiences, but has proven <strong>to</strong> be very difficult. Outside of<br />

a few notable exceptions such as the Canterbury Tales Project, this idea has rarely been<br />

realized, indicating the enormous commitment and difficult work required.<br />

A very different category is hypertext. Works of this type are generally original and not<br />

representations of previously published works. Hypertext would seem <strong>to</strong> have great<br />

potential for expression, allowing for a multiplicity of narrative choices and making the<br />

reader an active participant in the reading experience. Hypertextual content is generally<br />

bound inextricably with hypertext software, making productions in this format even more<br />

ephemeral than other kinds of electronic text. Some hypertext authors, such as Michael<br />

Joyce and Stuart Mouthrop, recognize the impermanence inherent in the genre and<br />

incorporate it in<strong>to</strong> their hyperfictions, but this fundamental aspect of its nature will make<br />

it difficult or impossible for libraries <strong>to</strong> collect hyperfictions for the long term.<br />

Eventually, even if the medium on which the hypertext is s<strong>to</strong>red remains viable, the<br />

software on which it relies will no longer run. While early critics such as Robert Coover<br />

believed (with some hope) that hypertext would lead <strong>to</strong> "the end of books", others, such<br />

as Tim Parks, dismiss the genre as "hype."<br />

Creating Electronic Text<br />

Scholars, students, librarians, computing professionals, and general enthusiasts of all<br />

kinds create and publish electronic texts. Commercial publishers, notably those<br />

previously known for publishing microfilm collections, are digitizing and licensing


access <strong>to</strong> ambitiously large collections of tens of thousands of volumes. A few projects,<br />

such as the University of Virginia's collaboration with Chadwyck-Healey <strong>to</strong> publish<br />

Early American Fiction, and the Text Creation Partnerships formed for Early English<br />

Books Online and the Evans Early American Imprints, are examples of partnerships<br />

between libraries and publishers.<br />

How do <strong>humanities</strong> scholars and students approach the creation of electronic texts? One<br />

axis stretches from a traditional <strong>humanities</strong> approach, taking time and great pains <strong>to</strong><br />

create well-considered, well-documented, well-edited electronic works, such as the<br />

Women Writers Online project at Brown University, the archive of Henrik Ibsen's<br />

writings, sponsored by the National Library of Norway and hosted by the universities of<br />

Oslo and Bergen, and the archive of Sir Isaac New<strong>to</strong>n's manuscripts at Imperial College,<br />

London, <strong>to</strong> less rigorous efforts, relying on volunteers and enthusiasts, with the bestknown<br />

examples being Project Gutenberg and Distributed Proofreaders. This difference<br />

in approaches mirrors various edi<strong>to</strong>rial approaches in the past century, with the same<br />

difference occurring between scholarly editions and reading editions, each with different<br />

purposes and audiences. For some scholars, the latter type suffices for classroom use or<br />

quick consultation, while for others, the extensive documentation and edi<strong>to</strong>rial standards<br />

of the former are of paramount importance.<br />

Interestingly, skepticism may play a central role in either attitude. On the one hand,<br />

Project Gutenberg enthusiasts, most notably project leader Michael Hart, express<br />

skepticism about the added software and expertise needed <strong>to</strong> create and read encoded<br />

electronic texts. His concerns, forged by painful hardware and software changes<br />

beginning in the mainframe era, are founded in the belief that the lowest common<br />

denomina<strong>to</strong>r of encoding will ensure the largest readership and longevity of the texts.<br />

Others, equally concerned about longevity and the demands created by changing<br />

hardware and software, believe that encoding systems based on international standards<br />

such as SGML, XML, and Unicode allow for the best representation of complex textual<br />

structures and writing systems, and these formats will prove most useful for scholars and<br />

general readers, as well as being the most promising means for the long-term archiving of<br />

these works. It should be noted that even some Project Gutenberg texts are available in<br />

HTML, PDF, or Microsoft Ebook reader formats, presumably for the added functionality<br />

or characters available beyond plain ASCII. Still, both sides view the other with<br />

suspicion.<br />

High-cost equipment is not required for creating electronic text – it can be achieved with<br />

the simplest of computers and word processing software. Scanning equipment and OCR<br />

software continue <strong>to</strong> drop in price. However, our expectations of accuracy in printed texts<br />

are high – we are generally disappointed <strong>to</strong> find any errors, and certainly disappointed if<br />

we discover an error every 10 <strong>to</strong> 20 pages. An accurate transcription requires skill and<br />

concentration, either in keying it in or in proofreading, yet the standard acceptable<br />

accuracy rate used by many large e-text projects is 99.995 percent, or 1 error in every<br />

20,000 characters. Given that an average page has 1,000–2,000 characters, this works out<br />

<strong>to</strong> 1 error every 10 <strong>to</strong> 20 pages. Spellcheckers, while useful in catching some<br />

misspellings, cannot catch typographic errors or inaccurate OCR that result in correctly


spelled words, and they are also little help with dialect or creative misspellings that<br />

authors employ. Spellcheckers exist for only a limited number of languages as well,<br />

making them generally not useful for the work of correcting electronic texts. Instead, it is<br />

common for large projects <strong>to</strong> outsource the creation of electronic text <strong>to</strong> vendors. Many<br />

of these vendors use a technique whereby two or three typists work on the same text. The<br />

typists' work is collated, with any discrepancies between the versions being used <strong>to</strong> find<br />

errors. This method produces a high accuracy rate, because the chance of two or three<br />

typists making the same error in the same spot in a text is statistically low.<br />

Using Electronic Text<br />

The use of online collections, even collections of relatively obscure writers, remains<br />

surprisingly high. As noted above, one group overlooked by most early computing<br />

humanists is the general reader, someone willing and interested <strong>to</strong> read something online,<br />

remarkable considering the relatively primitive displays, unsophisticated interfaces, and<br />

slow connections available. These general readers may have many different reasons for<br />

using electronic texts – they may lack access <strong>to</strong> research libraries; they may prefer <strong>to</strong> find<br />

and use books on their computers, in their homes or offices; they may live in countries<br />

without access <strong>to</strong> good collections in foreign languages; the books themselves may be<br />

fairly rare and physically available at only a handful of libraries. Whatever the motivation<br />

may be, these readers are finding and using electronic texts via the World Wide Web.<br />

Jerome McGann, noted scholar, edi<strong>to</strong>r, theorist, and crea<strong>to</strong>r of the Dante Gabriel Rossetti<br />

Archives, places much more value on this aspect of access than does Hockey. A great<br />

change has swept through literary studies in the past twenty years, as scholars re-evaluate<br />

writers and works overlooked by previous generations of literary critics. The immediate<br />

challenge for anyone interested in re-evaluating a little-known text will be <strong>to</strong> find a copy,<br />

for it may be available in only a few libraries. For McGann, collections of electronic texts<br />

may be most valuable for holding works by those writers who are less well known, or<br />

who are considered minor. In analyzing the Chadwyck-Healey English Poetry Database,<br />

McGann explains it this way:<br />

For research purposes the database grows less and less useful for those authors who<br />

would be regarded, by traditional measures, as the more or the most important writers. It's<br />

most useful for so-called minor writers. This paradox comes about for two reasons. On<br />

one hand, the poetical works of "minor" writers are often hard <strong>to</strong> obtain since they exist<br />

only in early editions, which are typically rare and can be quite expensive. By providing<br />

electronic texts of those hard-<strong>to</strong>-acquire books, "The English Poetry Database" supplies<br />

scholars with important primary materials. On the other hand, the policy of the Database<br />

is – wherever possible – <strong>to</strong> print from collected editions of the poets as such editions<br />

exist. The better-known the poet, the more likely there will be collected edition(s) ….<br />

The Database would have done much better <strong>to</strong> have printed first editions of most of its<br />

authors, or at any rate <strong>to</strong> have made its determinations about editions on scholastic rather<br />

than economic grounds. But it did not do this.


[…] Speaking for myself, I now use the Database in only two kinds of operation: as a<br />

vast concordance, and as an initial source for texts that we don't have in our library. In the<br />

latter case I still have <strong>to</strong> find a proper text of the work I am dealing with.<br />

(McGann 1995: 382–3)<br />

The initial hindrances <strong>to</strong> reading works by lesser-known writers, perhaps insurmountable<br />

in the past, can be much more easily overcome in this new medium. It should be noted<br />

that even such a <strong>digital</strong>ly adept scholar as McGann prefers printed sources over electronic<br />

versions (or at least, still did in 1996). He used the online collection for discovery and<br />

research, but turned <strong>to</strong> the printed copy for verification and citation.<br />

Many problems face a scholar wishing <strong>to</strong> use electronic texts for research. The immediate<br />

problem is in discovering just what is available. Traditionally, a researcher would check<br />

the library catalogue <strong>to</strong> discover whether a particular title or edition is available in the<br />

stacks. Given the growing number of online collections, it is impossible <strong>to</strong> be aware of all<br />

relevant sources for any given research <strong>to</strong>pic, and even the specialized portals such as<br />

Voice of the Shuttle have fallen behind. As it stands now for most online projects,<br />

researchers must remember that a particular website has works by Charles Dickens or<br />

Margaret Oliphant <strong>to</strong> find electronic editions by these authors, for they may be found<br />

neither in online catalogues nor by using search <strong>to</strong>ols such as Google. Libraries could<br />

have records for electronic texts in their online catalogues also and link directly <strong>to</strong> the<br />

electronic editions. However, few libraries have included records for all titles available<br />

in, for instance, the English Poetry Database from Chadwyck-Healey even if they have<br />

acquired the full-text database for their communities. This situation is worse for those<br />

texts that are freely available, for the job of discovering and evaluating electronic texts,<br />

and then creating and maintaining links <strong>to</strong> them, is overwhelming.<br />

There is a great deal of interest in improving this situation. Libraries are beginning <strong>to</strong><br />

include records for electronic texts in their online catalogues. Other developments, such<br />

as the Open Archives Initiative, could allow the discovery of the existence of electronic<br />

texts much more readily than at present. Methods are under development for searching<br />

across collections s<strong>to</strong>red and maintained at different institutions, meaning that someone<br />

interested in nineteenth-century American his<strong>to</strong>ry could perform one search that would<br />

be broadcast <strong>to</strong> the many sites with collections from this period, with results collected<br />

and presented in a single interface.<br />

Another problem is the artificial divide that exists between those collections available<br />

from commercial publishers and those that are locally created. Generally, these two<br />

categories of materials use different interfaces and search systems. This distinction is<br />

completely irrelevant <strong>to</strong> researchers and students, but systems and interfaces that allow<br />

for searching both categories of materials simultaneously and seamlessly, while available,<br />

are still rare, due <strong>to</strong> the complexities of authentication and authorization.<br />

Another type of divide exists as well. With the growing body of collections available<br />

from commercial publishers, the divide between the haves and have-nots in this area is


growing. At a recent conference on nineteenth-century American literature, it was notable<br />

that graduate students at research universities had access <strong>to</strong> a wide range of commercially<br />

published electronic text collections, while many of their colleagues, recently graduated<br />

with first jobs at smaller institutions, did not. These untenured scholars may not need <strong>to</strong><br />

travel <strong>to</strong> institutions <strong>to</strong> see original documents any more, but they will continue <strong>to</strong> need<br />

support <strong>to</strong> travel <strong>to</strong> institutions that have access <strong>to</strong> licensed collections of electronic texts.<br />

There is hope for these scholars, however, as libraries and museums digitize wider<br />

portions of their collections and make them publicly available.<br />

A much more complex problem is the limited range of electronic texts that are available.<br />

A crazy patchwork quilt awaits any researcher or reader willing <strong>to</strong> use electronic texts,<br />

and as McGann points out, the selection of proper editions may not be given much<br />

thought. The situation resembles a land rush, as publishers, libraries, and individuals seek<br />

<strong>to</strong> publish significant collections. The number of freely available texts, from projects such<br />

as Making of America, <strong>to</strong> the Wright American Fiction project, <strong>to</strong> the University of<br />

Virginia E-Text Center, <strong>to</strong> the Library of Congress's American Memory, is growing at a<br />

phenomenal pace.<br />

Commercial publishers are now digitizing large microfilm collections such as Pollard and<br />

Redgrave, and Wing (Early English Books, published by Bell and Howell/UMI), and<br />

Evans (Early American Imprints, by Readex), and Primary Source Media has recently<br />

announced its intention <strong>to</strong> digitize the massive Eighteenth Century collection. The<br />

English Poetry Database, first published in 1992 and one of the earliest efforts, used the<br />

New Cambridge Bibliography of English Literature, first published in 1966 based on an<br />

earlier edition from the 1940s, as the basis for inclusion. The collections listed above are<br />

based upon bibliographies begun in the 1930s. Those collections of early literature such<br />

as the Early English Books Online can claim comprehensive coverage of every known<br />

publication from Great Britain before 1700, but as these online collections include<br />

publications closer <strong>to</strong> the present time, the less inclusive they can be, given the<br />

exponential growth of publishing. Thus, large collections such as the English Poetry<br />

Database have <strong>to</strong> employ selection criteria. Critics have debated these traditional<br />

collections, known informally as the "canon", and argued for broader inclusion of women<br />

or authors in popular genres. These electronic collections, while very large and inclusive,<br />

reflect the values and selection criteria of fifty years ago or more.<br />

Some libraries are developing their own <strong>digital</strong> collections, and McGann suggests it is no<br />

coincidence. He sees in literary studies two movements: a return <strong>to</strong> a "bibliographical<br />

center" and a "return <strong>to</strong> his<strong>to</strong>ry":<br />

The imperatives driving libraries and museums <strong>to</strong>ward greater computerization are not<br />

the same as those that have brought the now well-known "return <strong>to</strong> his<strong>to</strong>ry" in literary<br />

and <strong>humanities</strong> scholarship. Nevertheless, a convergence of the twain has come about,<br />

and now the two movements – the computerization of the archives, and the rehis<strong>to</strong>ricization<br />

of scholarship – are continually stimulating each other <strong>to</strong>ward new<br />

ventures.


(McGann 1995: 380)<br />

Even with this phenomenal growth, one realizes immediately the inadequacy of using<br />

these collections for comprehensive research. Scholars still cannot assume that their<br />

fields have an adequate collection of electronic texts available, nor can they assume that<br />

those collections that do exist will reflect current thinking about inclusion. McGann<br />

stated in 1996, something still true, "the Net has not accumulated those bodies of content<br />

that we need if we are <strong>to</strong> do our work" (p. 382). The collection size of a research library<br />

numbers in millions of volumes. The size of the collective electronic text collections<br />

available through both commercial publishers and freely available websites probably<br />

exceeds 200,000, but not by much. Scholars, students, and librarians are learning that the<br />

collections will have <strong>to</strong> grow considerably in order <strong>to</strong> reliably meet the needs of a broad<br />

range of humanists.<br />

As McGann states, these collections and editions were chosen in large part because they<br />

are in the public domain and free of copyright restrictions. The commercial publishers<br />

listed above started as microfilm publishers, and their microfilm collections were formed<br />

by the same principle. Copyright is the hidden force behind most electronic text<br />

collections. Very few electronic text collections, even those from commercial publishers,<br />

contain publications under copyright. This has two main effects on electronic collections.<br />

First and foremost, it means that most <strong>digital</strong> collections consist of authors who lived and<br />

published up <strong>to</strong> the twentieth century; the works of writers after that may still be under<br />

copyright, and therefore more difficult and perhaps expensive <strong>to</strong> copy and republish.<br />

Second, new editions of these pre-twentieth-century writers are generally excluded, with<br />

projects and publishers selecting those in the public domain. Finally, contemporary works<br />

of literary criticism, biography, and theory, that could provide needed context and<br />

interpretation <strong>to</strong> the primary literature, also remain largely excluded. The possibilities<br />

inherent in the medium, for providing a rich context for the study of primary literary texts<br />

and his<strong>to</strong>rical documents, have not yet been realized.<br />

The effect of copyright means that researchers and students interested in twentiethcentury<br />

and contemporary writing are largely prevented from using electronic text. A<br />

quick check of the online MLA Bibliography shows the number of articles published after<br />

1962 dealing with twentieth-century authors is nearly double that of all other centuries<br />

combined, covering 1200–1900 ce. With the majority of researchers and students<br />

interested in writers and their work after 1900, it is no wonder that they may consider<br />

electronic text largely irrelevant <strong>to</strong> their studies.<br />

In addition, many of the markers traditionally used by scholars <strong>to</strong> determine the merit of<br />

any given electronic text are missing. There may be no recognizable publisher, no edi<strong>to</strong>r,<br />

no preface or statement of edi<strong>to</strong>rial principles, which may cause scholars <strong>to</strong> shy away.<br />

They are left <strong>to</strong> their own devices <strong>to</strong> judge the value of many resources. On the other<br />

hand, in this unfamiliar terrain, there may be a tendency <strong>to</strong> trust the technology in the<br />

absence of such markers. Electronic text collections do not receive reviews as frequently<br />

as those published commercially, perhaps for some of these same reasons, although they<br />

may receive more use. Even in those reviews, one notes an uncritical trust on the part of


eviewers. For instance, I assert on the website of the Vic<strong>to</strong>rian Women Writers Project<br />

(VWWP) that it is "devoted <strong>to</strong> highly accurate transcriptions" of the texts. I am<br />

alternately amused and alarmed <strong>to</strong> see that phrase quoted verbatim in reviews and<br />

websites that link <strong>to</strong> the VWWP (e.g., Burrows 1999: 155; Hanson 1998; McDermott<br />

<strong>2001</strong>) without any test of its accuracy. Scholars who use these collections are generally<br />

appreciative of the effort required <strong>to</strong> create these online resources and reluctant <strong>to</strong><br />

criticize, but one senses that these resources will not achieve wider acceptance until they<br />

are more rigorously and systematically reviewed.<br />

It is difficult <strong>to</strong> find scholarly articles that cite electronic text collections as sources, or<br />

discuss the methodology of creating or using e-texts, outside of journals for computing<br />

humanists. Humanists have been slow <strong>to</strong> accept electronic texts for serious research, for<br />

the reasons explained above, particularly their scattershot availability, and the inadequate<br />

documentation of sources and edi<strong>to</strong>rial practices used in creating them. In an article taken<br />

from a major scholarly journal almost at random (Knowles <strong>2001</strong>), the author investigates<br />

the use of the word "patriot" throughout the seventeenth century. He cites familiar<br />

authors such as Mil<strong>to</strong>n, Dryden, Bacon, Jonson, Sir Wins<strong>to</strong>n Churchill, as well as less<br />

familiar authors. Electronic text collections such as Chadwyck-Healey's English Poetry<br />

Database or ProQuest's Early English Books Online would be perfectly suited for this<br />

purpose – McGann's "vast concordance", with the ability <strong>to</strong> search across thousands of<br />

works, leading <strong>to</strong> both familiar and unfamiliar works containing the term. However,<br />

Knowles does not mention these, although it is possible that he used them behind the<br />

scenes. At any rate, it is rare <strong>to</strong> find them cited in scholarly articles or books, and thus<br />

their use and importance goes unnoticed and untested. Scholars need <strong>to</strong> hear more from<br />

their peers about the use of these resources in their research.<br />

Electronic texts have an important place in classroom instruction as well. Julia Flanders<br />

points <strong>to</strong> a survey conducted by Women Writers Online in which a majority of professors<br />

responded that they "were more likely <strong>to</strong> use electronic <strong>to</strong>ols in their teaching than in<br />

their research" (Flanders <strong>2001</strong>: 57). The same problems described above would apply,<br />

but she posits that for students, unfamiliarity is an opportunity rather than an obstacle,<br />

giving them the chance <strong>to</strong> interact directly with the texts independent of contextual<br />

information and prior interpretation. The form of interaction offered by e-texts, with the<br />

opportunity <strong>to</strong> quickly explore words and terms across a large body of works, is vastly<br />

different than that offered by print texts. This approach <strong>to</strong> literary studies, closely related<br />

<strong>to</strong> traditional philology developed in the nineteenth century, is more common in classical<br />

studies, and may account for the successful adoption of computers and <strong>digital</strong> resources<br />

among classicists and medievalists in comparison <strong>to</strong> those who study and teach later<br />

periods.<br />

One assumption underlying much of the criticism of Olsen and Finch noted above is that<br />

the computer will in some way aid in literary criticism. Beginning with Father Busa, up<br />

<strong>to</strong> the creation of the World Wide Web, computer-aided analysis was the entire purpose<br />

of electronic texts, as there was no simple way for people <strong>to</strong> share their works. More<br />

recent books (Sutherland 1997; Hockey 2000) have shifted focus <strong>to</strong> the methods needed<br />

<strong>to</strong> create and publish electronic texts, leaving aside any discussion of their eventual use.


Still, Hockey and others point <strong>to</strong> the inadequacies of the <strong>to</strong>ols available as another<br />

limitation <strong>to</strong> wider scholarly acceptance of electronic texts. Early electronic texts, such as<br />

the WordCruncher collection of texts, were published on CD-ROM along with <strong>to</strong>ols for<br />

their use. While there are a multitude of problems associated with this kind of bundling, it<br />

certainly made entrance <strong>to</strong> use very easy. Today, many collections of electronic texts<br />

contain <strong>to</strong>ols for simple navigation and keyword searching, but little else. As Hockey<br />

notes (2000: 167), "[c]omplex <strong>to</strong>ols are needed, but these <strong>to</strong>ols must also be easy for the<br />

beginner <strong>to</strong> use", which explains in part why few exist. Another researcher, Jon K.<br />

Adams, writes:<br />

Like many researchers, I have found the computer both a fascinating and frustrating <strong>to</strong>ol.<br />

This fascination and frustration, at least in my case, seems <strong>to</strong> stem from a common<br />

source: we should be able <strong>to</strong> do much more with the computer in our research on literary<br />

texts than we actually do.<br />

(Adams 2000: 171)<br />

In some sense, while the World Wide Web has sparked development and distribution of<br />

electronic texts in numbers unthinkable before it, the <strong>to</strong>ols available for using these texts<br />

are of lesser functionality than those available through the previous means of publishing,<br />

such as CD-ROMs. This is perhaps due <strong>to</strong> the lack of consensus about the uses of<br />

electronic texts, as well as the difficulty of creating generalized text analysis <strong>to</strong>ols for use<br />

across a wide range of collections. The acceptance of more sophisticated au<strong>to</strong>mated<br />

analysis will remain limited until more sophisticated <strong>to</strong>ols become more widely available.<br />

Until then, the most common activities will be fairly simple keyword searches, and text<br />

retrieval and discovery. These types of research can be powerful and important for many<br />

scholars, but do not begin <strong>to</strong> tap the potential of <strong>humanities</strong> computing.<br />

Still, even noting their inadequacies, these collections are finding audiences. The E-text<br />

Center at the University of Virginia reports that their collections receive millions of uses<br />

every year. Making of America at the University of Michigan reports similar use. The<br />

Vic<strong>to</strong>rian Women Writers Project (VWWP), which contains about 200 works by lesserknown<br />

nineteenth-century British women writers, receives over 500,000 uses per year.<br />

The most popular works in the VWWP collection include such books as Caroline<br />

Nor<strong>to</strong>n's study English Laws for Women, Vernon Lee's gothic s<strong>to</strong>ries Hauntings, and<br />

Isabella Bird's travel writings. These writers are better known than most of the others in<br />

the collection. Nevertheless, the most heavily used work in the collection is Maud and<br />

Eliza Keary's Enchanted Tulips and Other Verse for Children, which receives thousands<br />

of uses per month. This is one indication that scholars are not the only audience for<br />

electronic text. In the end, the general reader, students, and elementary and secondary<br />

school teachers, particularly those in developing countries without access <strong>to</strong> good<br />

libraries, may be the largest and most eager audience for these works.<br />

Electronic texts give humanists access <strong>to</strong> works previously difficult <strong>to</strong> find, both in terms<br />

of locating entire works, with the Internet as a distributed interconnected library, and in<br />

access <strong>to</strong> the terms and keywords within the works themselves, as a first step in analysis.


As reliable and accurate collections grow, and as humanists come <strong>to</strong> understand their<br />

scope and limitations, the use of e-texts will become recognized as a standard first step in<br />

<strong>humanities</strong> research. As <strong>to</strong>ols for the use of e-texts improve, humanists will integrate etexts<br />

more deeply and broadly in<strong>to</strong> their subsequent research steps. Until then, electronic<br />

texts will remain, in Jerome McGann's words, a "vast concordance" and library with great<br />

potential.<br />

Bibliography<br />

Adams, Jon K. (2000). Narrative Theory and the Executable Text. Journal of Literary<br />

Semantics 29, 3: 171–81.<br />

American Memory. Washing<strong>to</strong>n, DC: Library of Congress. Accessed Oc<strong>to</strong>ber 13, 2003.<br />

At http://memory.loc.gov.<br />

ARTFL: Project for American and French Research on the Treasury of the French<br />

Language. Mark Olsen, ed. University of Chicago. Accessed Oc<strong>to</strong>ber 13, 2003. At<br />

http://<strong>humanities</strong>.uchicago.edu/orgs/ARTFL.<br />

Bailey, Richard W. (ed.) (1982). Computing in the Humanities: Papers from the Fifth<br />

International Conference on Computing in the Humanities. Amsterdam: North Holland.<br />

Baker, Nicholson (<strong>2001</strong>). Double Fold: Libraries and the Assault on Paper. New York:<br />

Random House.<br />

Birkerts, Sven (1994). The Gutenberg Elegies: The Fate of Heading in an Electronic Age.<br />

Bos<strong>to</strong>n: Faber and Faber.<br />

Bornstein, George and Theresa Tinkle (1998). Introduction. In The Iconic Page in<br />

Manuscript, Print and Digital Culture, ed. Bornstein and Tinkle (pp. 2–6). Ann Arbor:<br />

University of Michigan Press.<br />

Burrows, Toby (1999). The Text in the Machine: Electronic Texts in the Humanities.<br />

New York: Haworth Press.<br />

Busa, Rober<strong>to</strong> (1950). Complete Index Verborum of Works of St Thomas. Speculum 25, 3:<br />

424–5.<br />

Busa, Rober<strong>to</strong> (1992). Half a Century of Literary Computing: Towards a "New"<br />

Philology. Literary and Linguistic Computing 7, 1: 69–72.<br />

Bush, Vannevar (1945). As We May Think. The Atlantic Monthly (July): 101–8.<br />

Canterbury Tales Project. Peter Robinson, ed. De Montfort University. Accessed<br />

Oc<strong>to</strong>ber 13, 2003. At http://www.cta.dmu.ac.uk/projects/ctp/.


Coover, Robert (1992). The End of Books. The New York Times Book Review (June 21):<br />

1, 4.<br />

Early American Fiction. ProQuest/Chadwyck-Healey. Accessed Oc<strong>to</strong>ber 13, 2003. At<br />

http://www.e-text.virginia.edu/eaf/.<br />

Early English Books Online Text Creation Partnership. University of Michigan Library.<br />

Accessed Oc<strong>to</strong>ber 13, 2003. At http://www.lib.umich.edu/eebo/.<br />

Electronic Text Center. Alderman Library, University of Virginia. Accessed Oc<strong>to</strong>ber 13,<br />

2003. At http://e-text.virginia.edu.<br />

Evans Early American Imprint Collection Text Creation Partnership. University of<br />

Michigan Library. Accessed Oc<strong>to</strong>ber 13, 2003. At http://www.lib.umich.edu/evans/.<br />

Finch, Alison (1995). The Imagery of a Myth: Computer-Assisted Research on Literature.<br />

Style 29, 4: 511–21.<br />

Flanders, Julia (<strong>2001</strong>). Learning, Reading, and the Problem of Scale: Using Women<br />

Writers Online. Pedagogy 2, 1: 49–59.<br />

Hanson, R. (1998). Review, Vic<strong>to</strong>rian Women Writers Project. Choice 35 (Supplement):<br />

88.<br />

Hockey, Susan (1980). A Guide <strong>to</strong> Computer Applications in the Humanities. Baltimore:<br />

Johns Hopkins University Press.<br />

Hockey, Susan (2000). Electronic Texts in the Humanities: Principles and Practices.<br />

Oxford: Oxford University Press.<br />

Knowles, Ronald (<strong>2001</strong>). The "All-At<strong>to</strong>ning Name": the Word "Patriot" in Seventeenthcentury<br />

England. Modern Language Review 96, 3: 624–43.<br />

Lusignan, Serge and John S. North, (eds.) (1977). Computing in the Humanities:<br />

Proceedings of the Third International Conference on Computing in the Humanities.<br />

Waterloo, ON: University of Waterloo Press.<br />

Making of America. Ann Arbor: University of Michigan. Accessed Oc<strong>to</strong>ber 15, 2003. At<br />

http://moa.umdl.umich.edu.<br />

McDermott, Irene (<strong>2001</strong>). Great Books Online, Searcher 9, 8: 71–7.<br />

McGann, Jerome (1996). Radiant Textuality. Vic<strong>to</strong>rian Studies 39, 3: 379–90.<br />

Model Editions Partnership. David Chestnut, Project Direc<strong>to</strong>r, University of South<br />

Carolina. Oc<strong>to</strong>ber 13, 2003 http://mep.cla.sc.edu.


Olsen, Mark (1993–4). Signs, Symbols, and Discourses: A New Direction for Computeraided<br />

Literature Studies. Computers and the Humanities 27, 5–6: 309–14.<br />

Parks, Tim (2002). Tales Told by a Computer. New York Review of Books 49:16 (Oc<strong>to</strong>ber<br />

24): 49–51.<br />

Schaffer, Talia (1999). Connoisseurship and Concealment in Sir Richard Calmady.<br />

Lucas Malet's Strategic Aestheticism. In Talia Schaffer and Kathy Alexis Psomiades<br />

(eds.), Women and British Aestheticism. Charlottesville: University Press of Virginia.<br />

Sutherland, Kathryn (ed.) (1997). Electronic Text: Investigations in Method and Theory.<br />

New York: Oxford University Press.<br />

Text Encoding Initiative. University of Oxford, Brown University, University of Virginia,<br />

University of Bergen. Accessed Oc<strong>to</strong>ber 13, 2003. At http://www.tei-c.org.<br />

Thesaurus Linguae Graecae (1999). CD-ROM. Irvine, CA: University of California,<br />

Irvine.<br />

Vic<strong>to</strong>rian Women Writers Project. Perry Willett, ed. Indiana University. Accessed<br />

Oc<strong>to</strong>ber 13, 2003. At http://www.indiana.edu/~letrs/vwwp.<br />

Wilkin, John P. (1997). Just-in-time Conversion, Just-in-case Collections: Effectively<br />

Leveraging Rich Document Formats for the WWW. D-Lib Magazine (May). Accessed<br />

Oc<strong>to</strong>ber 13, 2002. At http://www.dlib.org/dlib/may97/michigan/05pricewilkin.html.<br />

Wisbey, R. A., (ed.) (1971). The Computer in Literary and Linguistic Research: Papers<br />

from a Cambridge Symposium. Cambridge: Cambridge University Press.<br />

Women Writers Online Project. Julia Flanders, ed. Brown University. Accessed Oc<strong>to</strong>ber<br />

13, 2003. At http://www.wwp.brown.edu.<br />

The WordCrumher Disk (1990). CD-ROM. Orem, Utah: Electronic Text Corporation.<br />

19.<br />

Modeling: A Study in Words and Meanings<br />

Willard McCarty<br />

Out on site, you were never parted from your plans. They were your Bible. They got dogeared,<br />

yellowed, smeared with mud, peppered with little holes from where you had<br />

unrolled them on the ground. But although so sacred, the plans were only the start. Once<br />

you got out there on the site everything was different. No matter how carefully done, the


plans could not foresee the variables. It was always interesting, this moment when you<br />

saw for the first time the actual site rather than the idealised drawings of it.<br />

He knew men who hated the variables. They had their plans and by golly they were<br />

going <strong>to</strong> stick <strong>to</strong> them. If the site did not match the drawings it was like a personal insult.<br />

He himself liked the variables best. He liked the way that the solution <strong>to</strong> one problem<br />

created another problem further down the line, so that you had <strong>to</strong> think up something<br />

else, and that in turn created another problem <strong>to</strong> solve. It was an exchange, backwards<br />

and forwards. Some men thought of it as a war, but <strong>to</strong> him it was more like a<br />

conversation.<br />

Kate Grenville, The Idea of Perfection (Sydney: Picador, 1999): 62–3<br />

Introduction<br />

The question of modeling arises naturally for <strong>humanities</strong> computing from the prior<br />

question of what its practitioners across the disciplines have in common. What are they<br />

all doing with their computers that we might find in their diverse activities indications of<br />

a coherent or cohesible practice? How do we make the best, most productive sense of<br />

what we observe? There are, of course, many answers: practice varies from person <strong>to</strong><br />

person, from project <strong>to</strong> project, and ways of construing it perhaps vary even more. In this<br />

chapter I argue for modeling as a model of such a practice. I have three confluent goals:<br />

<strong>to</strong> identify <strong>humanities</strong> computing with an intellectual ground shared by the older<br />

disciplines, so that we may say how and <strong>to</strong> what extent our field is of as well as in the<br />

<strong>humanities</strong>, how it draws from and adds <strong>to</strong> them; at the same time <strong>to</strong> reflect experience<br />

with computers "in the wild"; and <strong>to</strong> aim at the most challenging problems, and so the<br />

most intellectually rewarding future now imaginable.<br />

My primary concern here is, as Confucius almost said, that we use the correct word for<br />

the activity we share lest our practice go awry for want of understanding (Analects 13.3).<br />

Several words are on offer. By what might be called a moral philology I examine them,<br />

arguing for the most popular of these, "modeling." The nominal form, "model", is of<br />

course very useful and even more popular, but for reasons I will adduce, its primary<br />

virtue is that properly defined it defaults <strong>to</strong> the present participle, its semantic lemma.<br />

Before getting <strong>to</strong> the philology I discuss modeling in the light of the available literature<br />

and then consider the strong and learned complaints about the term.<br />

Background<br />

Let me begin with provisional definitions 1 . By "modeling" I mean the heuristic process of<br />

constructing and manipulating models', a "model" I take <strong>to</strong> be either a representation of<br />

something for purposes of study, or a design for realizing something new. These two<br />

senses follow Clifford Geertz's analytic distinction between a denotative "model of" such<br />

as a grammar describing the features of a language, and an exemplary "model for" such


as an architectural plan (Geertz 1973: 93) 2 . In both cases, as the literature consistently<br />

emphasizes, a model is by nature a simplified and therefore fictional or idealized<br />

representation, often taking quite a rough-and-ready form: hence the term "tinker <strong>to</strong>y"<br />

model from physics, accurately suggesting play, relative crudity, and heuristic purpose<br />

(Cartwright 1983: 158). By nature modeling defines a ternary relationship in which it<br />

mediates epistemologically, between modeler and modeled, researcher and data or theory<br />

and the world (Morrison and Morgan 1999). Since modeling is fundamentally relational,<br />

the same object may in different contexts play either role: thus, e.g., the grammar may<br />

function prescriptively, as a model for correct usage, the architectural plan descriptively,<br />

as a model of an existing style. The distinction also reaches its vanishing point in the<br />

convergent purposes of modeling: the model of exists <strong>to</strong> tell us that we do not know, the<br />

model for <strong>to</strong> give us what we do not yet have. Models realize.<br />

Perhaps the first question <strong>to</strong> ask is what such a process has <strong>to</strong> do with computing, since as<br />

the examples suggest neither of the two senses of "model" assumes it unless the<br />

definition is further qualified. In his<strong>to</strong>ry, for example, Gordon Leff has argued that<br />

models have always been implicit in scholarly practice (Leff 1972). Leff cites, e.g., the<br />

his<strong>to</strong>ric-graphical notion of "epoch", but any well-articulated idea would qualify as a<br />

model of its subject. Nevertheless, Leff notes that as M. I. Finley said in Ancient His<strong>to</strong>ry:<br />

Evidence and Models, "model-construction is rare among all but economic his<strong>to</strong>rians";<br />

Finley recommends Max Weber's parallel concept of "ideal types", which "expresses<br />

clearly the nature and function of models in his<strong>to</strong>rical inquiry" (1986: 60f). Explicit<br />

model-construction is still rare in mainstream <strong>humanities</strong> scholarship. Even for noncomputational<br />

research in the social sciences, it is more common, as Finley's demarcation<br />

suggests. For example, political schemes by nature model for a better or at least different<br />

world, even if like Marx's his<strong>to</strong>riography they begin as models of it; delineating them as<br />

models is the scholar's obvious work (Mironesco 2002). Nevertheless, outside<br />

computationally affected scholarly practice Marvin Minsky's simple, straightforward<br />

definition remains alien in style and in thought: "To an observer B, an object A* is a<br />

model of an object A <strong>to</strong> the extent that B can use A* <strong>to</strong> answer questions that interest him<br />

about A" (Minsky 1995).<br />

A strong temptation for us here is <strong>to</strong> dismiss the residual alienness of Minsky's<br />

formulation and <strong>to</strong> accept, as we have accepted computing, the reified, explicit "model"<br />

of Minsky's definition as what we really have been doing all along. This would, however,<br />

be a serious error. As with the relationship of hypertext <strong>to</strong> earlier ways of referring<br />

(McCarty 2002), the new form of expression, with its vocabulary and <strong>to</strong>ols, means an<br />

altered way of thinking. A his<strong>to</strong>rical imagination is required <strong>to</strong> see what this means.<br />

Two effects of computing make the distinction between "idea" or other sort of mental<br />

construct on the one hand, and on the other "model" in the sense we require: first, the<br />

demand for computational tractability, i.e., for complete explicitness and absolute<br />

consistency; second, the manipulability that a computational representation provides.<br />

The first effects a sea-change by forcing us <strong>to</strong> confront the radical difference between<br />

what we know and what we can specify computationally, leading <strong>to</strong> the epistemological


question of how we know what we know. On the one hand, as Michael Polanyi observed,<br />

"we can know more than we can tell" (1966: 4–5). Computational form, which accepts<br />

only that which can be <strong>to</strong>ld explicitly and precisely, is thus radically inadequate for<br />

representing the full range of knowledge – hence useful for isolating the tacit or inchoate<br />

kinds. On the other hand, we need <strong>to</strong> trust what we somehow know, at least provisionally,<br />

in order not <strong>to</strong> lose all that goes without saying or cannot be said in computational form.<br />

Take, for example, knowledge one might have of a particular thematic concentration in a<br />

deeply familiar work of literature. In modeling one begins by privileging this knowledge,<br />

however wrong it might later turn out <strong>to</strong> be, then building a computational representation<br />

of it, e.g., by specifying a structured vocabulary of word-forms in a text-analysis <strong>to</strong>ol. In<br />

the initial stages of use, this model would be almost certain <strong>to</strong> reveal trivial errors of<br />

omission and commission. Gradually, however, through perfective iteration trivial error<br />

is replaced by meaningful surprise. There are in general two ways in which a model may<br />

violate expectations and so surprise us: either by a success we cannot explain, e.g.,<br />

finding an occurrence where it should not be; or by a likewise inexplicable failure, e.g.,<br />

not finding one where it is otherwise clearly present. In both cases modeling<br />

problematizes. As a <strong>to</strong>ol of research, then, modeling succeeds intellectually when it<br />

results in failure, either directly within the model itself or indirectly through ideas it<br />

shows <strong>to</strong> be inadequate. This failure, in the sense of expectations violated, is, as we will<br />

see, fundamental <strong>to</strong> modeling.<br />

The second quality of "model" that distinguishes it from "idea" is manipulability, i.e., the<br />

capability of being handled, managed, worked, or treated by manual and, by extension,<br />

any mechanical means (OED: la.). Change is common <strong>to</strong> both models and ideas, but at<br />

greater or lesser metaphorical distance, "model" denotes a concrete, articulated plan<br />

inviting the etymological sense of action-by-hand (L. manus) in response. Manipulation<br />

in turn requires something that can be handled (physical objects, diagrams, or symbols of<br />

a formal language) – and a time-frame sufficiently brief that the emphasis falls on the<br />

process rather than its product. In other words, the modeling system must be interactive.<br />

Manipulable objects from the physical <strong>to</strong> the metaphorical have characterized<br />

mathematics, engineering, the physical sciences, and the arts ab wo, but with exceptions<br />

the necessary time-frame, allowing for interactivity, has been possible only with<br />

computing. With its advent, Minsky has noted, models could be "conceived, tested, and<br />

discarded in days or weeks instead of years" (1991). Computing met research easily in<br />

fields where modeling was already an explicit method because, Brian Cantwell Smith has<br />

pointed out, models are fundamental <strong>to</strong> computing: <strong>to</strong> do anything useful at all a<br />

computer must have a model of something, real or imaginary, in software. But in the<br />

context of computing, models per se are not the point. What distinguishes computers<br />

from other kinds of machines, Smith notes, is that "they run by manipulating<br />

representations, and representations are always formulated in terms of models" (Smith<br />

1995/1985: 460; cf. Fetzer 1999: 23).<br />

In other words, computational models, however finely perfected, are better unders<strong>to</strong>od as<br />

temporary states in a process of coming <strong>to</strong> know rather than fixed structures of<br />

knowledge. It is of course possible <strong>to</strong> argue ideologically, as some still do, that we are


converging on and will achieve such structures, 3 but in any case these structures would<br />

not then be models and would no longer have reason <strong>to</strong> exist in software. (Note that the<br />

his<strong>to</strong>ry of computing is the s<strong>to</strong>ry of ever more complex and extensive software, not less,<br />

despite the fact that implementations in hardware are faster and can be cheaper.) For the<br />

moment and the foreseeable future, then, computers are essentially modeling machines,<br />

not knowledge jukeboxes. To think of them as the latter is profoundly <strong>to</strong> misunderstand<br />

human knowledge – and so <strong>to</strong> constrict it <strong>to</strong> the narrow scope of the achievably<br />

mechanical.<br />

In analytical terms, as I have suggested, modeling has two phases: first, construction;<br />

second, manipulation. Examples come readily <strong>to</strong> mind from ordinary technical practice,<br />

e.g., building a relational database, then querying the data thus shaped <strong>to</strong> explore<br />

emergent patterns. As experience with databases shows, the two phases often blur in<strong>to</strong><br />

each other especially in the early stages when use uncovers faults or suggests<br />

improvements that direct redesign. A model 0/and a model for may be distinct types –<br />

because in our terms they are fixed objects. But modeling of something readily turns in<strong>to</strong><br />

modeling for better or more detailed knowledge of it; similarly, the knowledge gained<br />

from realizing a model for something feeds or can feed in<strong>to</strong> an improved version. This<br />

characteristic blurring of design in<strong>to</strong> use and use in<strong>to</strong> (re)design is what denies modeling<br />

of any sense of closure. Modeling for, U<strong>to</strong>pian by definition, is denied it in any case.<br />

Learned Complaints<br />

So far so good – but at the cost of averting our gaze from the problems with the word<br />

"model." Indeed, the extensive and growing literature on the <strong>to</strong>pic may seem adrift in a<br />

hopeless muddle. "I know of no model of a model", physicist H. J. Groenewold declared<br />

many years ago (1960: 98). The philosopher Peter Achinstein has warned us away even<br />

from attempting a systematic theory (1968: 203). The word itself is indeed as<strong>to</strong>nishingly<br />

polysemous – or promiscuous, as Nelson Goodman puts it. "Model", he complains, can<br />

be used <strong>to</strong> denote "almost anything from a naked blonde <strong>to</strong> a quadratic equation" (1976:<br />

171). Nevertheless, the word is often used as if its semantic complexity either did not<br />

exist or could be safely ignored. The muddle of partly overlapping, partly contradic<strong>to</strong>ry<br />

senses is proof enough that we ignore it at our peril. Nor can we simply avoid the<br />

problem by dismissing "model" al<strong>to</strong>gether, as Goodman and others recommend, without<br />

(as I will argue) hobbling our ability <strong>to</strong> understand inter alia those aspects of computing<br />

most important <strong>to</strong> research – one might even say, as I do, its essence. Despite several<br />

other, supposedly less confusing terms on offer, the word remains stubbornly popular in<br />

the literature of the social and physical sciences, the his<strong>to</strong>ry and philosophy of science,<br />

cognitive science, artificial intelligence, and related areas.<br />

Theoretical physicist John Ziman and philosopher Stephen Toulmin, for example,<br />

recommend "map" on the basis of its conceptual clarity and fitness for describing the<br />

relationship between theoretical knowledge and reality (Ziman 2000: 126–38, 147–50;<br />

Toulmin 1953: 94–109). Goodman would have us collapse modeling in<strong>to</strong> diagramming,<br />

which he thinks less troublesome (1976: 171–3). But preference needs <strong>to</strong> be based on<br />

more than such criteria; "map", for example, serves experiment much less well than


theory, as I will show. We require a close and careful look at the semantic fields of all<br />

major alternatives, including "map", for their disjunctions and overlaps. We need <strong>to</strong><br />

scrutinize each of these, asking what does it denote and connote that the others do not,<br />

and vice versa? What do all have in common? What are their individual tendencies of<br />

mind, and which of these best suits computing as we are learning <strong>to</strong> conceive it?<br />

Philological Analysis of Related Terms<br />

So far I have used the term "model" as the default, partly for purposes of convenience,<br />

partly because, as I argue, it is right for the job. To answer the learned criticisms and<br />

further clarify our <strong>to</strong>pic, however, I propose <strong>to</strong> question it by comparison against the<br />

major alternatives: "analogy", "representation", "diagram", "map", "simulation", and<br />

"experiment." As we have seen, the two decisive criteria are that the thing named by the<br />

chosen term be computationally tractable and manipulable. Tractability in turn requires<br />

complete explicitness and absolute consistency; manipulability resolves in<strong>to</strong> mechanical<br />

action and interactivity. Hence the term must denote a continual process of coming <strong>to</strong><br />

know, not an achievement but an approximation. As I have argued, it is from the<br />

difference between the approximation and the reality approximated – which ultimately<br />

for the <strong>humanities</strong> is our humanly known apprehension of that reality – that we learn.<br />

For each of the alternative terms I ask whether and <strong>to</strong> what degree the word normally<br />

denotes a dynamic process and whether it refers <strong>to</strong> a concrete, i.e. manipulable, form –<br />

the requirements of anything whose function is fulfilled through being changed.<br />

Corresponding <strong>to</strong> the two senses of "model" I identified earlier, the denotative model of<br />

and the exemplary model for, I also ask whether each word tends <strong>to</strong> the mimetic<br />

(imitation) or proleptic (anticipation). The distinction helps in determining whether the<br />

action denoted by a term may be said <strong>to</strong> be bounded, either by a fixed form, as is the case<br />

with "analogy", or by an inherent tendency <strong>to</strong> reach a definitive or satisfac<strong>to</strong>ry<br />

conclusion, as in "representation."<br />

Thus bringing "model" in<strong>to</strong> focus against the semantic background of these other terms<br />

will show that the problem has not so much been <strong>to</strong>o many meanings for "model" as use<br />

without regard <strong>to</strong> any of them, often as if the sense of it simply goes without saying. It<br />

doesn't. But perhaps the most important lesson we learn from seeing the word in the<br />

context of its synonym set is not the range and variety of its meanings; rather, again, its<br />

strongly dynamic potential. Apart from the popularity of "model" taken at face-value, the<br />

word would have little <strong>to</strong> recommend it (and, as the complainers say, much against it) but<br />

for the open-ended present-participial strength of "modeling." 4 Indeed, the manifest<br />

confusion in the literature on the <strong>to</strong>pic may be primarily due <strong>to</strong> a mistaken preference for<br />

the noun – as if getting a model right, and so promoting it <strong>to</strong> the status of theory, were the<br />

point. Modeling has an entirely different role <strong>to</strong> play. There are several better terms if<br />

what one wants is <strong>to</strong> name a stable conceptualization.


Analogy<br />

"Analogy" (Gk. ναλoγ α, "equality of ratios, proportion") is, like "model", a highly<br />

polysemous term with a long and complex career 5 . John Stuart Mill complained that "no<br />

word … is used more loosely, or in a greater variety of senses, than Analogy" (A System<br />

of Logic, 1882). Yet Dr Johnson's pithy definition, "resemblance of things with regard <strong>to</strong><br />

some circumstances or effects", and Mill's even pithier one, "resemblance of relations",<br />

give us an idea of why it is so fruitful. From its original meaning in Greek mathematics,<br />

analogy specifies a structured relationship between pairs of related quantities, which for<br />

convenience may be represented in the form of an equation, "A/B = C/D", read "as A is<br />

<strong>to</strong> B, so C is <strong>to</strong> D." Extended beyond mathematics <strong>to</strong> other modes of reasoning, analogy<br />

yields a powerful (but still poorly unders<strong>to</strong>od) means of inferring from the one relation <strong>to</strong><br />

the other. Like most of the words in our domain, "analogy" is proleptic, a means of<br />

inference, based on conjecture, <strong>to</strong> something unknown or uncertain. Examples in the<br />

his<strong>to</strong>ry of science are plentiful, e.g., Kepler's discovery of the vis matrix, or cause of<br />

planetary motion, by reasoning that as the sun radiated light, so it must also radiate this<br />

motive power (Gentner 2002).<br />

Here I wish only <strong>to</strong> argue two points. The first is that analogy is basic <strong>to</strong> the entire<br />

vocabulary. Although not every model is as strictly based on an analogy as Kepler's,<br />

modeling is inherently analogical, with just the features that make the idea attractive for<br />

our purposes. Thus we require a structured correspondence between model and artifact,<br />

so that by playing with one we can infer facts about the other. (For example, by adjusting<br />

choice of words and weightings for a distribution-display across a textual corpus, one can<br />

investigate the effect of vocabulary on the interplay of meanings in that corpus.) The<br />

second point is that "analogy" is inherently static: it means either a type of relationship or<br />

an instance of one, never an object and not, literally or directly, a process. Action is<br />

implied in the ratio of quantities – thus Kepler's "as A does B, so C does D" – but acting<br />

is not denoted by the analogy. The word has no commonly used verbal form ("analogize"<br />

and "analogizing" are rare if not strange). Although an analogy may be algebraically or<br />

geometrically expressed and may refer <strong>to</strong> concrete objects, it itself is abstract.<br />

Because analogy works so well as a way of describing how we often think, efforts <strong>to</strong><br />

understand acquisition of new knowledge tend <strong>to</strong> engage with theories of analogy and <strong>to</strong><br />

propose many mechanisms, e.g., in cognitive science, educational theory, and artificial<br />

intelligence (Hoffman 1995). Because modeling is analogical, this work is potentially<br />

relevant <strong>to</strong> questions raised in computing the artifacts of the <strong>humanities</strong>. We need <strong>to</strong> pay<br />

attention here.<br />

Representation<br />

"Representation" in Nelson Goodman's terms is defined by a symbolic denotative<br />

correspondence, not likeness or imitation (1976: 3–41) 6 . In a less philosophically precise<br />

sense, however, we may say that "representation" displays strong mimetic tendencies,<br />

e.g., in the definition given by the OED: "An image, likeness, or reproduction in some<br />

manner of a thing …. A material image or figure; a reproduction in some material or


tangible form; in later use esp. a drawing or painting (of a person or thing)." The his<strong>to</strong>ry<br />

of aesthetics from earliest times, in fits and starts of fashion, demonstrates that the copytheory<br />

of representation, though its falsity names its achievement in a trompe I'oeil,<br />

remains a habit of mind. If in aesthetics, why not in computer science?<br />

A well-attested present participle and a full complement of verbal forms establishes the<br />

action of representing, but semantically this action is bounded by its relationship <strong>to</strong> the<br />

represented object, whether this be symbolic or imitative.<br />

As with "analogy", the semantic fields of "model" and "representation" clearly overlap,<br />

but the situation is more complex because of the mimetic and proleptic kinds of "model."<br />

Hence (Pla<strong>to</strong>nism aside) we may say that modeling of is representational but not<br />

modeling for. In fact a model of is a manipulable variety of representation – which any<br />

representation in software would of course be. The crucial difference between model<br />

0/and representation is the quality of the action implied. Unlike representing, modeling of<br />

is denied closure, as I noted earlier. It has no satisfac<strong>to</strong>ry trompe I'oeil or symbolizing<br />

conclusion. If the researcher calls a halt, then the last state of the system, as it were, is<br />

better called a "representation."<br />

In the context of computing, the meaning of "representation" is dominated by the subfield<br />

of artificial intelligence known as "knowledge representation" (KR). Given the scope of<br />

this essay I can do little more than make a few observations, chiefly about the<br />

assumptions built in<strong>to</strong> the name and apparently in<strong>to</strong> tendencies in some KR work. In<br />

brief, my argument concerning KR is that it needs <strong>to</strong> be unders<strong>to</strong>od as a particularly<br />

rigorous control on model-building suitable <strong>to</strong> that which can be stated in prepositional<br />

form. 7<br />

To the connotations of "representation" I have already reviewed, KR adds the demarcational<br />

term "knowledge." The point I wish <strong>to</strong> draw from current epistemology is a simple<br />

one, for which I quote Michael Williams – at length, because the issues are consequential<br />

for us:<br />

"Knowledge" is an honorific title we confer on our paradigm cognitive achievements ….<br />

More generally, "know" is a "success-term", like "win" or "pass" (a test). Knowledge is<br />

not just a factual state or condition but a particular normative status. Such statuses are<br />

related <strong>to</strong> appropriate factual states: winning depends on crossing the line before any<br />

other competi<strong>to</strong>r. But they also depend on meeting certain norms or standards which<br />

define, not what you do do, but what you must or ought <strong>to</strong> do. To characterize someone's<br />

claim as expressing or not expressing knowledge is <strong>to</strong> pass judgement on it. Epistemic<br />

judgements are thus a particular kind of value-judgement ….<br />

This normative dimension distinguishes philosophical theories ofknowledge from<br />

straightforwardly factual inquiries and explains why demarcational (and related<br />

methodological) issues are so significant. Because epistemological distinctions are<br />

invidious, ideas about epistemological demarcation always involve putting some claims<br />

or methods above others: mathematics above empirical science, empirical science above


metaphysics or religion, logic above rhe<strong>to</strong>ric, and so on. Demarcational projects use<br />

epistemological criteria <strong>to</strong> sort areas of discourse in<strong>to</strong> factual and non-factual, truthseeking<br />

and merely expressive, and, at the extreme, meaningful and meaningless. Such<br />

projects amount <strong>to</strong> proposals for a map of culture: a guide <strong>to</strong> what forms of discourse are<br />

"serious" and what are not. Disputes about demarcation – induding disputes about<br />

whether demarcational projects should be countenanced at all – are disputes about the<br />

shape of our culture and so, in the end, of our lives.<br />

(<strong>2001</strong>: 11–12)<br />

Projects such as Cyc, based on what Northrop Frye characterized as the discredited<br />

Wissenscbaft-theory of knowledge – that its accumulation in vast quantities will one day,<br />

somehow, result in understanding 8 – clearly assume if not perfect closure, then a<br />

threshold beyond which lack of perfection ceases <strong>to</strong> matter. But <strong>to</strong> whom, and for what<br />

purposes? 9 Apart from such questions, and the serious doubts within computer science on<br />

the wisdom of building massive knowledge-bases for expert systems 10 – there are, again,<br />

the very serious demarcational issues. When, for example, one of the leading theorists of<br />

KR writes in passing that, "Perhaps there are some kinds of knowledge that cannot be<br />

expressed in logic" (Sowa 2000: 12), our intellectual claustrophobia tells an important<br />

tale. Not, of course, the only one. If the point of modeling is <strong>to</strong> fail well, then KR has a<br />

vital quality-control function <strong>to</strong> serve.<br />

Diagram<br />

A diagram (Gk. δι γραµµα, "that which is marked out by lines, a geometrical figure,<br />

written list, register, the gamut or scale in music") 11 is an analogical drawing, "a figure<br />

drawn in such a manner that the geometrical relations between the parts of the figure<br />

illustrate relations between other objects": thus the physicist James Clerk Maxwell on the<br />

graphic, symbolic, and hybrid kinds (1911). Such a diagram ranges from the precisely<br />

drawn schematic, whose measurements are significant, <strong>to</strong> the rough sketch intended <strong>to</strong><br />

express symbolic relations only. But what makes a graphic a diagram, properly so called,<br />

is the way in which it is read, not its resemblance <strong>to</strong> anything 12 . Reviel Netz argues the<br />

point for the lettered diagram in Greek mathematical texts: "It is only the diagram<br />

perceived in a certain way which may function alongside the text" – irrelevant<br />

discrepancies are overlooked, the important matters denoted by the lettering: "All<br />

attention is fixed upon the few intersecting points, which are named" (1999: 33–5). Even<br />

when the diagrammed entity is a physical object, the diagram represents structure and<br />

interrelation of essential parts, foregrounding interpretative choice and conscious<br />

purpose. The ability <strong>to</strong> manipulate structures and parts may be implied.<br />

The word "diagram" doubles as noun and verb and has a full range of verbal inflections.<br />

Like "represent", its action is bounded, but more by ideas than appearance, even when<br />

that appearance is precisely delineated. As Maxwell notes for physics, diagrams often<br />

represent force or movement, even if only implicitly, though the form is static. As a<br />

means of communication, e.g., in a lecture or discussion between collabora<strong>to</strong>rs,


diagramming is the point, not the static trace left behind. That trace may in fact be<br />

unintelligible apart from the discussion of which it was a dynamic part.<br />

The focus on ideas rather than things per se, the role of manipulation where diagramming<br />

is heuristic and the kinaesthetics of the action suggest the close relationship between<br />

diagramming and modeling for which Goodman argues. Models of, he declares, "are in<br />

effect diagrams, often in more than two dimensions and with working parts … [and]<br />

diagrams are flat and static models" (1972: 172–3). Nevertheless, the two-dimensionally<br />

graphic, geometrical, and finally static qualities of the diagram define it, not "model",<br />

which has considerably broader applications.<br />

Sun-Joo Shin and Oliver Lemon note that although diagramming is very likely one of the<br />

oldest forms of communication, modern logicians and philosophers have tended until<br />

recently <strong>to</strong> regard it as only of marginal importance. 13 That is changing very rapidly now.<br />

As a cognitive, reasoning process it is studied in relation <strong>to</strong> Greek mathematics and<br />

geometry in particular (Netz 1999). Modern philosophical attention can be traced from<br />

Descartes's "La Geometric" (1637) and Kant's Critique of Pure Reason II.1.1 (1781) <strong>to</strong><br />

Peirce's "existential graphs" in the late nineteenth century, significantly as part of his<br />

much broader interest in scientific discovery, <strong>to</strong> which I will return. His work is now<br />

central <strong>to</strong> much current research.<br />

Shin and Lemon delineate three branches of research since the mid-1990s: (1) multimodal<br />

and especially non-linguistic reasoning, in the philosophy of mind and cognitive<br />

science; (2) logical equivalence of symbolic and diagrammatic systems, in logic; and (3)<br />

heterogeneous systems implementing theories of multi-modal reasoning, in computer<br />

science. The close relationship of diagramming and modeling make this research<br />

immediately relevant.<br />

Map<br />

A map may be defined as a schematic spatial representation, or following Maxwell, a<br />

diagram of "anything that can be spatially conceived." 14 Indeed, if not for the<br />

geographical focus of mapping, the semantic fields of "map" and "diagram" would<br />

completely overlap: both are fully verbal, their action bounded by a graphical<br />

representation that has both strongly mimetic and ideational aspects; both manipulate<br />

data for specific purposes and introduce fictional elements <strong>to</strong> serve these purposes. But<br />

the long his<strong>to</strong>ry of mapping the physical world for exploration and description gives<br />

"map" specific (and evidently powerful) connotations.<br />

Mapping is particularly characteristic of an early, explora<strong>to</strong>ry period, when a terri<strong>to</strong>ry is<br />

unknown <strong>to</strong> its discoverers (or conquerors). Mapping constructs the world it represents,<br />

selectively, therefore shaping thought and guiding action. It orients the newcomer, giving<br />

him or her control of the mapped terrain, at the same time expressing, though perhaps<br />

covertly, a perspective, a set of interests, and a his<strong>to</strong>ry. Mapping attaches meaning <strong>to</strong><br />

place. Like modeling it can be either of or for a domain, either depicting the present


landscape or specifying its future – or altering how we think about it, e.g., by renaming<br />

its places. A map is never entirely neutral, politically or otherwise.<br />

As I have noted, John Ziman, following Stephen Toulmin, has argued persuasively for<br />

describing scientific research as mapping – hence the immediate relevance of the<br />

car<strong>to</strong>graphic imagination <strong>to</strong> my project. Mapping particularly fits a Kuhnian view of<br />

research: long periods of stable activity within a settled terrain interspersed by<br />

revolutionary, and perhaps incommensurable, reconceptualizations of that terrain.<br />

Mapping is for our purposes more <strong>to</strong> the point than representation, because we always<br />

know that there can be many maps for any terri<strong>to</strong>ry and that all of them have a fictional<br />

character (which is why the map is a standard example of an interpretative genre for<br />

data 15 ). But because its action is bounded and its result defines a world, mapping better<br />

suits the theoretician's than the experimenter's view.<br />

In computer science, mapping is used in knowledge and argument representation and<br />

implementation of schemes for depicting cyberspace in general and the Web in particular.<br />

The term surfaces in "Topic Maps", "Concept Maps", and is implicit in talk about, e.g.,<br />

"semantic networks." This interest seems <strong>to</strong> originate with Toulmin's mapping of<br />

argument (Toulmin 1958), which suggests techniques of au<strong>to</strong>mated inferencing in AI.<br />

(Maps of argument look like flowcharts.) As a form of data-visualization, mapping also<br />

connects with a very strong, recent interest in <strong>humanities</strong> computing (Kirschenbaum<br />

2002), and so connects this interest with modeling.<br />

Simulation<br />

"Simulation" is "The technique of imitating the behaviour of some situation or process …<br />

by means of a suitably analogous situation or apparatus" (OED). 16 Its mimetic tendencies<br />

and so bounded action are perhaps most emphatic among the terms we are considering.<br />

Again, <strong>to</strong>tal replication is not at issue; a simulation attends <strong>to</strong> selected details of the<br />

world, thus can be explora<strong>to</strong>ry, as when relevant conditions of flight are simulated for<br />

purposes of study or training. Simulation also relies on abstraction from the original <strong>to</strong><br />

the analogue system (Simon 1969: 15–18), which makes it a kind of representation,<br />

subject <strong>to</strong> the same philosophical argument including the caveat respecting mimesis. But<br />

in usage, the connotation, if not denotation, of an exact correspondence between<br />

simulation and original remains paradoxically alongside knowledge of real difference. 17<br />

That knowledge, however uneasily, can be put aside, as in what we now call "virtual<br />

reality" (VR). 18<br />

In its current form VR is of course quite a recent phenomenon, but the essential<br />

movement of simulation on which it is based, from self-conscious imitation <strong>to</strong><br />

displacement of reality, is attested from the get-go of applied computing, in the weapons<br />

research in nuclear physics immediately following World War II. "Proven on the most<br />

complex physical problem that had ever been undertaken in the his<strong>to</strong>ry of science", Peter<br />

Galison notes, simulation came <strong>to</strong> replace experimental reality, thus blurring multiple<br />

boundaries that had previously defined research, redefining it in new terms (1997: 690f).<br />

Since then the turn away from traditional analytic methods <strong>to</strong> simulation has spread <strong>to</strong>


several other fields (Burch 2002). As the biologist Glenn W Rowe points out, with this<br />

turn has come the realization that "a great many systems seem <strong>to</strong> have an inherent<br />

complexity that cannot be simplified" 19 – and so must be studied as simulations. Thus<br />

simulation has opened our eyes <strong>to</strong> the new problems with which it can deal. In the<br />

<strong>humanities</strong> we have known for some years that computer-based simulations, in the form<br />

of pedagogical games, can play a role in teaching. An old but very good example is The<br />

Would-Be Gentleman, a re-creation of economic and social life in seventeenth-century<br />

France in which the student-player must realize and put aside his or her modern<br />

preconceptions in order <strong>to</strong> win (Lougee 1988). In other words he or she must become a<br />

seventeenth-century Frenchman mentally and emotionally. From more recent and far<br />

more technically advanced VR applications, such as Richard Beacham's and Hugh<br />

Denard's reconstruction of the theater of Pompey in Rome (Beacham and Denard 2003),<br />

one can predict a scholarly future for simulation in many areas of humanistic research.<br />

Simulation, like game-playing, tends <strong>to</strong> forgetfulness of the mechanism by which it is<br />

created, so long as its terms of engagement (expressed in parameters and algorithms) are<br />

fixed. Unfix them – e.g., in The Would-Be Gentleman by allowing the player <strong>to</strong> change<br />

the encoded attitude <strong>to</strong>ward marrying above or below one's station – and the simulation<br />

becomes a modeling exercise directed <strong>to</strong> exploring the question of that attitude. Thus<br />

simulation crosses over in<strong>to</strong> modeling when the constants of the system become<br />

variables. Modeling, one might say, is a self-conscious simulation, and simulation an<br />

assimilated modeling.<br />

Experiment<br />

In common usage "experiment" (L. experiri, <strong>to</strong> try) is either "An action or operation<br />

undertaken in order <strong>to</strong> discover something unknown …" or "The action of trying<br />

anything, or putting it <strong>to</strong> proof; a test, trial …" (OED). In its broadest sense, the word<br />

embraces "modeling", indeed any heuristic experience of the world, especially that which<br />

involves conscious purpose, defined procedure, or its material instantiation in equipment.<br />

Like "modeling", "experiment" refers <strong>to</strong> a process whose ending is constructed rather<br />

than given: as Peter Galison has argued for the physical sciences, the experimenter<br />

decides if and when the attempt has succeeded or failed, in "that fascinating moment …<br />

when instrumentation, experience, theory, calculation, and sociology meet" (Galison<br />

1987: 1). Modeling and experimenting are by nature open-ended; indeed they are often at<br />

the time ill-defined, even quite messy. 20<br />

The semantic overlap of "modeling" and "experiment" is so close that the two can be<br />

quite difficult <strong>to</strong> separate (Guala 2002). Mary S. Morgan, writing about modeling in<br />

economics, has argued that they may be discriminated by the degree <strong>to</strong> which the former<br />

involves hypothesis, the latter reality (2002: 49). But, as she notes, hybrids provide<br />

exceptions and thought experiments – "devices of the imagination used <strong>to</strong> investigate<br />

nature" (Brown 2002) – a very close analogue <strong>to</strong> modeling. 21 Perhaps the relationship is<br />

most clearly stated by saying that in the context of research a model is an experimental<br />

device, modeling an experimental technique.


The point of considering "experiment" here is, however, primarily <strong>to</strong> locate our <strong>to</strong>pic<br />

within the context of a particular his<strong>to</strong>ry of ideas and so <strong>to</strong> engage with large and<br />

important areas of his<strong>to</strong>rical and philosophical research. Indeed, as an experimental<br />

technique modeling has shared the fate of experiment in the specialist literature, and so<br />

also in the popular understanding. Allow me briefly <strong>to</strong> summarize the background. 22<br />

By the mid-nineteenth century, understanding of scientific work had begun <strong>to</strong> polarize<br />

in<strong>to</strong> two epistemological conditions, which physicist and philosopher Hans Reichenbach<br />

later famously named "the context of discovery" and "the context of justification." 23 By<br />

the early <strong>to</strong> mid-twentieth century, attention had shifted almost completely <strong>to</strong> justification<br />

and so, as Thomas Nickels has said, discovery was expelled from mainstream<br />

epistemology (2000: 87). Experiment, the means of discovery, was in consequence also<br />

demoted and theory, the focus of justification, promoted. "The asymmetric emphasis on<br />

theory in the his<strong>to</strong>rical literature", Peter Galison explains, meant that attention was<br />

confined "<strong>to</strong> the invention and testing of theories" (1987: 8), the actual conduct and<br />

intellectual role of experiments being largely overlooked. In philosophy <strong>to</strong>o, "experiment<br />

for theory" dominated until (in Ian Hacking's paraphrase of Nietzsche) Thomas Kuhn's<br />

The Structure of Scientific Revolutions "unwrapped the mummy of science" by<br />

his<strong>to</strong>ricizing it (Hacking 1983: If) – What had actually always been happening in<br />

experimental work could then become a proper subject of investigation. Furthermore,<br />

Kuhn's ample demonstration of "the essential role theory plays in the conduct of<br />

experimentation, the interpretation of data, and in the definition of 'relevant' phenomena"<br />

depolarized theory and experiment (Galison 1987: 80- In other words, from an entirely<br />

subordinate and observational role, experiment emerged alongside theory as an<br />

interdependent partner.<br />

Subsequently, through the work of Hacking, Feyerabend, Galison, and several others, the<br />

fiction of a unitary "scientific method", in which theory cleanly defines the role of<br />

experiment, has been dispelled. As Hacking says, calling for a "Back-<strong>to</strong>-Bacon"<br />

movement, "Experiment has a life of its own" (1983: 150), sometimes preceded by<br />

theory, sometimes not. But the consequent liberation of experiment from the debilitating<br />

pretense of grubby handmaiden <strong>to</strong> "pure" theory has at the same time directed attention<br />

back <strong>to</strong> the very hard, old problem of discovery: how does this happen?<br />

Conclusion<br />

Why do we need an answer <strong>to</strong> this question? Because, I have argued, ours is an<br />

experimental practice, using equipment and instantiating definite methods, for the skilled<br />

application of which we need <strong>to</strong> know what we are doing as well as it can be known. I<br />

have labeled the core of this practice "modeling", and suggested how, properly<br />

unders<strong>to</strong>od, modeling points the way <strong>to</strong> a computing that is of as well as in the<br />

<strong>humanities</strong>: a continual process of coming <strong>to</strong> know by manipulating representations. We<br />

are, I have suggested, in good epistemological company. But this only sharpens the<br />

epistemological question. The signified of modeling vanishes in<strong>to</strong> the murk because we<br />

lack a disciplined way of talking about it. Methods are explicit, actions definite, results<br />

forthcoming, yet we have been unable fully and persuasively <strong>to</strong> articulate the intellectual


case for the means by which these results are produced. Hence the just-a-<strong>to</strong>ol status of<br />

computing, the not-a-discipline slur, the tradesman's entrance or other back door in<strong>to</strong> the<br />

academy. No one doubts the usefulness of the practice. Rather it's the intellection of<br />

praxis <strong>to</strong> which the next stage in the argument I have begun here must turn.<br />

Note<br />

1 My definitions reflect the great majority of the literature explicitly on modeling in the<br />

his<strong>to</strong>ry and philosophy of the natural sciences, especially of physics (Bailer-Jones 1999).<br />

The literature tends <strong>to</strong> be concerned with the role of modeling more in formal scientific<br />

theory than in experiment. The close relationship between modeling and experimenting<br />

means that the rise of a robust philosophy of experiment since the 1980s is directly<br />

relevant <strong>to</strong> our <strong>to</strong>pic; see Hacking (1983); Gooding (2000). Quite helpful in rethinking<br />

the basic issues for the <strong>humanities</strong> are the writings from the disciplines other than<br />

physics, e.g., Clarke (1972) on archaeology; Wimsatt (1987) on biology; Del Re (2000)<br />

on chemistry; and on the social sciences, the essays by de Callatay, Mironesco, Burch,<br />

and Gardin in Franck (2002). For interdisciplinary studies see Shanin (1972) and<br />

Morrison and Morgan (1999), esp. "Models as Mediating Instruments" (pp. 10–37). For<br />

an overview see Lloyd (1998).<br />

2 Cf. Goodman's distinction between "denotative" and "exemplary" models, respectively<br />

(1976: 172–3); H. J. Groenewold's "more or less poor substitute" and "more or less<br />

exemplary ideal" (1960: 98). Similar distinctions are quite common in the literature.<br />

3 This is usually done in "the rhe<strong>to</strong>ric of technohype … the idiom of grant proposals and<br />

of interviews in the Tuesday New York Science Times: The breakthrough is at hand; this<br />

time we've got it right; theory and practice will be forever altered; we have really made<br />

fantastic progress, and there is now general agreement on the basics; further funding is<br />

required" (Fodor 1995). More serious criticism is leveled by Terry Winograd (1991: 207–<br />

8); see below.<br />

4 I have in mind the present-participial imagination described by Greg Dening, with<br />

which we may "return <strong>to</strong> the past the past's own present, a present with all the<br />

possibilities still in it, with all the consequences of actions still unknown" (Dening 1998:<br />

48; see also Dening 1996: 35–63).<br />

5 For the concept in imaginative language and thought see Gibbs (1994; reviewed by<br />

Turner 1995), Turner (1996); in computer science, Hoffman (1995) – whose summary of<br />

research is quite valuable; in cognitive science, including psychology, Mitchell (1993),<br />

Holyoak and Thagard (1997); in the philosophy of science, Achinstein (1968),<br />

Leatherdale (1974), Gentner (2002), Shelley (2002); in relation <strong>to</strong> modeling, Bailer-Jones<br />

(1999), Bailer-Jones and Bailer-Jones (2002). I do not deal here with metaphor in relation<br />

<strong>to</strong> modeling, for which see Black (1979), Johnson (2002).<br />

6 Goodman dismantles the copy-theory of representation, arguing that representation is<br />

not mimetic but symbolic: object X is always represented as Y, which means that Y is


selective with respect <strong>to</strong> X and stands in symbolic relationship <strong>to</strong> it. See also Elgin<br />

(1998), Hopkins (2000).<br />

7 Possibly the best and least problematic view is afforded by Davis et al. (1993); see also<br />

Sowa (2000); Barr and Feigenbaum (1981). Lenat (1998) illustrates the problematic<br />

tendencies in this field; Winograd (1991) and Dreyfus (1985) provide the antidote.<br />

8 Frye (1991: 4), <strong>to</strong> which compare Winograd's analysis of the "almost childish leap of<br />

faith" made, e.g., by Marvin Minsky in his "Society of Mind" thesis that "the modes of<br />

explanation that work for the details of [the artificial micro-worlds thus represented] will<br />

be adequate for understanding conflict, consciousness, genius, and freedom of will"<br />

(Winograd 1991: 204–7) – as the ambitious claim; see also Winder (1996).<br />

9 Note the boast that "Cyc knows that trees are usually outdoors, that once people die<br />

they s<strong>to</strong>p buying things, and that glasses of liquid should be carried rightside-up" (Cycorp<br />

Company Overview, at http://www.cyc.com/overview.html, accessed September 22,<br />

2003.<br />

10 Winograd and Flores (1986: 97–100, 131–3, 174–7); Dreyfus (1985). See also Brooks<br />

(1991).<br />

11 See, however, the discussion in Netz (1999: 35–8).<br />

12 Goodman (1976: 170f), who distinguishes between analogue and <strong>digital</strong> diagrams. As<br />

Netz explains, the lettered diagram provides a good example of the latter (1999: 34f).<br />

13 Shin and Lemon (2002); note the extensive bibliography.<br />

14 Robinson and Petchenik (1976); see also Monmonier (1996); Wood (1992). Turnbull<br />

(1994) argues specifically for the link between maps and theories.<br />

15 See Bateson (2002: 27–8), who cites Alfred Korzybski's principle that "the map is not<br />

the terri<strong>to</strong>ry" (Korzybski 1933) and points out that "the natural his<strong>to</strong>ry of human mental<br />

process" nevertheless tells a different tale: part of us in fact regularly identifies map and<br />

terri<strong>to</strong>ry, name and thing named. See also Goodman (1972: 15); Kent (2000/1978: xix).<br />

16 On the semantic overlap of "simulation" with "experiment" and "model", see Guala<br />

(2002), who also stresses the necessity of including the designer or initia<strong>to</strong>r as part of the<br />

simulation.<br />

17 Hence, perhaps, the deception attributed <strong>to</strong> the word: "intent <strong>to</strong> deceive" in a "false<br />

assumption or display, a surface resemblance or imitation …" (OED 1 .a., 2) – an<br />

animated trompe l'oeil.<br />

18 Recent research in psychology and cognitive science, working with the<br />

representational model of mind, might be summarized by the proposition that reality as


we know and participate in it is simulated. Thus mental simulation is used <strong>to</strong> explain<br />

aspects of cognition (see, e.g., Markman and Dietrich 2000; Davies and S<strong>to</strong>ne 2000).<br />

Especially relevant here is the idea that perceptual simulations play a significant role in<br />

cognition, as when the replay of a kinaesthetic memory, awakened by some<br />

corresponding movement or gesture, lends meaning <strong>to</strong> a diagram or physical model. This<br />

is why animations can in principle be more effective than static diagrams: they are more<br />

knowledgeable (Craig et al. 2002)<br />

19 Rowe (1994), quoted by Burch (2002: 245); see also Winsberg (<strong>2001</strong>).<br />

20 This is true for scientific experiment much more often and more significantly than<br />

popular and earlier philosophical accounts would have us believe. Ian Hacking illustrates<br />

the point in an illuminating discussion of the famous Michelson-Morley experiment, "a<br />

good example of the Baconian exploration of nature" (1983: 254); see his discription<br />

(1983: 253–61), and esp. the book as a whole. See also Gooding (2000) and Morrison<br />

(1998) for an overview of experiment in current philosophy of science; the triplet of<br />

articles presented at a symposium on "The Philosophical Significance of<br />

Experimentation", Hacking (1988), Heelan (1988), and Galison (1988); and Franklin<br />

(2002) for physics in particular.<br />

21 See also Brown (2000), Gooding (1998) and note esp. the careful argument in Kuhn<br />

(1964) <strong>to</strong>ward an understanding of how thought experiment can lead <strong>to</strong> new knowledge,<br />

not simply expose logical contradictions or confusions.<br />

22 I rely primarily on Galison (1987), Hacking (1983) and Nickels (2000).<br />

23 Reichenbach introduced the distinction <strong>to</strong> mark "the well-known difference" between<br />

"the form in which thinking processes are communicated <strong>to</strong> other persons [and] the form<br />

in which they are subjectively performed", i.e., justification and discovery, respectively<br />

(Reichenbach 1938, chapter 1); compare Feyerabend (1993: 147–58). This distinction<br />

involves a long-standing argument that goes back <strong>to</strong> the debate between William<br />

Whewell (first <strong>to</strong> use the term "philosophy of science") and John Stuart Mill; it was then<br />

taken up by Charles Sanders Peirce on the one hand and Karl Popper on the other.<br />

Popper, in Logik der Forschung (translated as The Logic of Scientific Discovery)<br />

characteristically ruled discovery out of court by identifying it as a matter for psychology,<br />

recalling Duhem's and Reichenbach's use of that term: "The question of how it happens<br />

that a new idea occurs <strong>to</strong> a man … may be of great interest <strong>to</strong> empirical psychology; but<br />

it is irrelevant <strong>to</strong> the logical analysis of scientific knowledge" (Popper 1959/1935: 7). See<br />

Nickels (2000).<br />

Works Cited<br />

Achinstein, Peter (1968). Concepts of Science: A Philosophical Analysis. Baltimore:<br />

Johns Hopkins University Press.


Bailer-Jones, Daniela M. (1999). Tracing the Development of Models in the Philosophy<br />

of Science. In Magnani, Nersessian and Thagard (1999): 23–40.<br />

Bailer-Jones, Daniela M. 2002. Models, Metaphors and Analogies. In Machamer and<br />

Silberstein (2002): 108–27.<br />

Bailer-Jones, Daniela M. and Coryn A. L. Bailer-Jones (2002). Modeling Data:<br />

Analogies in Neural Networks, Simulated Annealing and Genetic Algorithms. In Magnani<br />

and Nersessian (2002): 147–65.<br />

Barr, Avron and Edward A. Feigenbaum (1981). Representation of Knowledge. In The<br />

Handbook of Artificial Intelligence. Volume I, ed. Avron Barr and Edward A.<br />

Feigenbaum (pp. 141–222). Los Al<strong>to</strong>s, CA: Morgan Kaufmann.<br />

Bateson, Gregory (2002). Mind and Nature: A Necessary Unity. Cresskill, NJ: Hamp<strong>to</strong>n<br />

Press.<br />

Beacham, Richard, and Hugh Denard (2003). The Pompey Project: Digital Research and<br />

Virtual. Reconstruction of Rome's First Theatre. Computers and the Humanities 37, 1:<br />

129–39.<br />

Black, Max (1979). More About Metaphor. In Metaphor and Thought, ed. Andrew<br />

Or<strong>to</strong>ny (pp. 19–39). Cambridge: Cambridge University Press.<br />

Brooks, Rodney A. (1991). Intelligence without Representation. Artificial Intelligence<br />

Journal 47: 139–59.<br />

Brown, James Robert (2000). Thought Experiments. In New<strong>to</strong>n-Smith (2000): 528–31.<br />

Brown, James Robert 2002. Thought Experiments. Stanford Encyclopedia of Philosophy.<br />

Accessed September 12, 2003. At .<br />

Burch, Thomas K. (2002). Computer Modeling of Theory: Explanation for the 21st<br />

Century. In Franck (2002): 245–65.<br />

Cartwright, Nancy (1983). How the Laws of Physics Lie. Oxford: Clarendon Press.<br />

Clarke, David L. (1972). Models and Paradigms in Contemporary Archaeology. In<br />

Models in Archaeology, ed. David L. Clarke (pp. 1–60). London: Methuen.<br />

Craig, David L., Nancy J. Nersessian, and Richard Catrambone (2002). Perceptual<br />

Simulation in Analogical Problem Solving. In Magnani and Nersessian (2002): 167–89.<br />

Davies, Martin and Tony S<strong>to</strong>ne (2000). Simulation Theory. In Routledge Encyclopedia of<br />

Philosophy. London: Routledge.


Davis, Randall, Howard Shrobe, and Peter Szolovits (1993). What Is Knowledge<br />

Representation? AI Magazine(Spring): 17–33. Accessed September 18, 2003. At<br />

.<br />

Del Re, Giuseppe (2000). Models and Analogies in Science. Hyle - International Journal<br />

for the Philosophy of Chemistry 6, 1: 5–15. Accessed September 18, 2003. At<br />

http://www.hyle.Org/journal/issues/6/delre.htm.<br />

Dening, Greg (1996). Performances. Chicago: University of Chicago Press.<br />

Dening, Greg (1998). Readings/Writings. Melbourne: University of Melbourne Press.<br />

Dreyfus, Hubert L. (1985). From Micro-Worlds <strong>to</strong> Knowledge Representation: AI at an<br />

Impasse. In Readings in Knowledge Representation, ed. Ronald J. Brachman and Hec<strong>to</strong>r<br />

Levesque (pp. 71–94). Los Al<strong>to</strong>s, CA: Morgan Kaufmann.<br />

Elgin, Catherine Z. (1998). Goodman, Nelson. In Routledge Encyclopedia of Philosophy.<br />

London: Routledge.<br />

Fetzer, James H. (1999). The Role of Models in Computer Science. The Monist 82, 1: 20–<br />

36.<br />

Feyerabend, Paul (1993). Against Method, 3rd edn. London: Verso.<br />

Finley, M. I. (1986). Ancient His<strong>to</strong>ry: Evidence and Models. New York: Viking.<br />

Fodor, Jerry (1995). West Coast Fuzzy. Review of Paul M. Churchland, The Engine of<br />

Reason, the Seat of the Soul. Times Literary Supplement (August 25).<br />

Franck, Robert, (ed.) (2002). The Explana<strong>to</strong>ry Power of Models. Methodos Series, vol. 1.<br />

Dordrecht: Kluwer Academic.<br />

Franklin, Allan (2002). Experiment in Physics. The Stanford Encyclopedia of Philosophy.<br />

Accessed September 12, 2003. At .<br />

Frye, Northrop (1991). Literary and Mechanical Models. In Research in Humanities<br />

Computing 1. Papers from the 1989 ACH-ALLC Conference, ed. Ian Lancashire (pp. 1–<br />

12). Oxford: Clarendon Press.<br />

Galison, Peter (1987). How Experiments End. Chicago: University of Chicago Press.<br />

Galison, Peter (1988). Philosophy in the Labora<strong>to</strong>ry. Journal of Philosophy 85, 10: 525–<br />

7.<br />

Galison, Peter (1997). Image and Logic: A Material Culture of Microphysics. Chicago:<br />

University of Chicago Press.


Geertz, Clifford (1973). Religion as a Cultural System. In The Interpretation of Cultures:<br />

Selected Essays. New York: Basic Books. (Reprint: London: Fontana Press, 1993.).<br />

Gentner, Dedre (2002). Analogy in Scientific Discovery: The Case of Johannes Kepler. In<br />

Magnani and Nersessian (2002): 21–39.<br />

Gibbs, Raymond W, Jr (1994). The Poetics of Mind: Figurative Thought, Language, and<br />

Understanding. Cambridge: Cambridge University Press.<br />

Gooding, David C. (1998). Thought Experiments. Routledge Encyclopedia of Philosophy.<br />

London: Routledge.<br />

Gooding, David C. (2000). Experiment. In New<strong>to</strong>n-Smith (2000): 117–26.<br />

Goodman, Nelson (1972). Problems and Projects. Indianapolis, IN: Bobbs-Merrill.<br />

Goodman, Nelson (1976). The Languages of Art. Indianapolis, IN: Hackett.<br />

Groenewold, H. J. (1960). The Model in Physics. In The Concept and the Role of the<br />

Model in Mathematics and the Natural Sciences, ed. Hans Friedenthal (pp. 98–103).<br />

Dordrecht: D. Reidel.<br />

Guala, Francesco (2002). Models, Simulations, and Experiments. In Magnani and<br />

Nersessian (2002): 59–74.<br />

Hacking, Ian (1983). Representing and Intervening: Introduc<strong>to</strong>ry Topics in the<br />

Philosophy of Natural Science. Cambridge: Cambridge University Press.<br />

Hacking, Ian (1988). On the Stability of the Labora<strong>to</strong>ry Sciences. Journal of Philosophy<br />

85, 10: 507–14.<br />

Heelan, Patrick A. (1988). Experiment and Theory: Constitution and Reality. Journal of<br />

Philosophy 85, 10: 515–27.<br />

Hoffmann, Robert R. (1995). Monster Analogies. AI Magazine 16, 3: 11–35. Accessed<br />

September 18, 2003. At .<br />

Holyoak, Keith J. and Paul Thagard (1997). The Analogical Mind. American<br />

Psychologist 52, 1: 35–44.<br />

Hopkins, R. D. (2000). Depiction. In Routledge Encyclopedia of Philosophy. London:<br />

Routledge.<br />

Johnson, Mark (2002). Metaphor-Based Values in Scientific Models. In Magnani and<br />

Nersessian (2002): 1–19.


Kent, William (2000/1978). Data and Reality, 2nd edn. Blooming<strong>to</strong>n, IN: 1st Books.<br />

(Originally published in Amsterdam: North Holland.).<br />

Kirschenbaum, Matthew, (ed.) (2002). Image-Based Humanities Computing. Special<br />

issue of Computers and the Humanities 36, 3.<br />

Korzybski, Alfred (1933). Science and Sanity: An Introduction <strong>to</strong> Non-Aris<strong>to</strong>telian<br />

Systems and General Semantics. Lancaster, PA: International Non-Aris<strong>to</strong>telian Library<br />

Publishing.<br />

Kuhn, Thomas (1964). A Function for Thought Experiments. In L'aventure de la science,<br />

Mélanges Alexandre Koyré, vol. 2 (pp. 307–34). Paris: Hermann.<br />

Leatherdale, W. H. (1974). The Role of Analogy, Model and Metaphor in Science.<br />

Amsterdam: North Holland Publishing.<br />

Leff, Gordon (1972). Models Inherent in His<strong>to</strong>ry. In Shanin, ed.(1972):148–74.<br />

Lenat, Douglas B. (1998). From <strong>2001</strong> <strong>to</strong> <strong>2001</strong>: Common Sense and the Mind of HAL. In<br />

HAL's Legacy: <strong>2001</strong>'s Computer as Dream and Reality, ed. David G. S<strong>to</strong>rk (pp. 193–<br />

209). Cambridge, MA: MIT Press.<br />

Lloyd, Elisabeth A. (1998). Models. In Routledge Encyclopedia of Philosophy. London:<br />

Routledge.<br />

Lougee, Carolyn (1988). The Would-Be Gentleman: a His<strong>to</strong>rical Simulation of the<br />

France of Louis XIV. His<strong>to</strong>ry Microcomputer Review 4: 7–14.<br />

McCarty, Willard (2002). A Network with a Thousand Entrances: Commentary in an<br />

Electronic Age? In The Classical Commentary: His<strong>to</strong>ries, Practices, Theory, ed. Roy K.<br />

Gibson and Christina Shuttleworth Kraus (pp. 359–402). Leiden: Brill.<br />

Machamer, Peter and Silberstein, Michael, eds. (2002). The Blackwell Guide <strong>to</strong> the<br />

Philosophy of Science. Oxford: Blackwell Publishers.<br />

Magnani, Lorenzo and Nancy J. Nersessian, (eds.) (2002). Model-Eased Reasoning:<br />

Science, Technology, Values. New York: Kluwer Academic/Plenum.<br />

Magnani, Lorenzo, Nancy J. Nersessian, and Paul Thagard, (eds.) (1999). Model-Eased<br />

Reasoning in Scientific Discovery. New York: Kluwer Academic/Plenum.<br />

Markman, Arthur B. and Eric Dietrich (2000). Extending the Classical View of<br />

Representation. Trends in Cognitive Sciences 4, 12: 470–5.<br />

Maxwell, James Clerk (1911). Diagram. In Encyclopaedia Britannica, 11th edn. (pp.<br />

146–9).


Minsky, Marvin (1991). Conscious Machines. Proceedings, National Research Council<br />

of Canada, 75th Anniversary Symposium on Science in Society (June). Accessed April<br />

22, 2004. At .<br />

Minsky, Marvin L. (1995). Matter, Mind and Models. Accessed January 20, 2003. At<br />

http://medg.lcs.mit.edu/people/doyle/gallery/minsky.mmm.html>. (Rev. version of the<br />

essay in Semantic Information Processing, ed. Marvin Minsky. Cambridge, MA: MIT<br />

Press, 1968.).<br />

Mironesco, Christine (2002). The Role of Models in Comparative Politics. In Franck<br />

(2002): 181–95.<br />

Mitchell, Melanie (1993). Analogy-Making as Perception. Cambridge, MA: MIT Press.<br />

Monmonier, Mark (1996). How <strong>to</strong> Lie with Maps, 2nd edn. Chicago: University of<br />

Chicago Press.<br />

Morgan, Mary S. (2002). Model Experiments and Models in Experiments. In Magnani<br />

and Nersessian (2002): 41–58.<br />

Morrison, Margaret (1998). Experiment. In Routledge Encyclopedia of Philosophy.<br />

London: Routledge.<br />

Morrison, Margaret and Mary S. Morgan, (eds.) (1999). Models as Media<strong>to</strong>rs:<br />

Perspectives on Natural and Social Science. Ideas in Context, 52. Cambridge: Cambridge<br />

University Press.<br />

Netz, Reviel (1999). The Shaping of Deduction in Greek Mathematics: A Study in<br />

Cognitive His<strong>to</strong>ry. Ideas in Context, 51. Cambridge: Cambridge University Press.<br />

New<strong>to</strong>n-Smith, W. H. ed. (2000). A Companion <strong>to</strong> the Philosophy of Science. Oxford:<br />

Blackwell.<br />

Nickels, Thomas (2000). Discovery. In New<strong>to</strong>n-Smith (2000): 85–96.<br />

Polanyi, Michael (1966). The Tacit Dimension. New York: Doubleday.<br />

Popper, Karl R. (1959/1935). The Logic of Scientific Discovery. London: Hutchinson.<br />

(Originally published as Logik der Forschung. Vienna: Verlag von Julius Springer.).<br />

Popper, Karl R. (1979). Objective Knowledge. An Evolutionary Approach. Oxford:<br />

Clarendon Press.<br />

Reichenbach, Hans (1938). Experience and Prediction: An Analysis of the Foundations<br />

and the Structure of Knowledge. Chicago: University of Chicago Press. (A transcription


of Chapter 1 is available at http://www.ditext.com/reichenbach/reicho.html. Accessed<br />

September 12, 2003.).<br />

Robinson, Arthur H. and Barbara Petchenik (1976). The Nature of Maps. Chicago:<br />

University of Chicago Press.<br />

Rowe, Glenn W. (1994). Theoretical Models in Biology: The Origin of Life, the Immune<br />

System, and the Brain. Oxford: Oxford University Press.<br />

Shanin, Teodor (1972). Models in Thought. In Shanin, ed. (1972): 1–22.<br />

Shanin, Teodor, (ed.) (1972). Rules of the Game: Cross-Disciplinary Essays on Models in<br />

Scholarly Thought. London: Tavis<strong>to</strong>ck.<br />

Shelley, Cameron (2002). Analogy Counterarguments and the Acceptability of<br />

Analogical Hypotheses. British Journal for the Philosophy of Science 53: 477–96.<br />

Shin, Sun-Joo and Oliver Lemon (2002). Diagrams. Stanford Encyclopedia of<br />

Philosophy. Accessed September 11, 2003. At http://pla<strong>to</strong>.stanford.edu/entries/diagrams/.<br />

Simon, Herbert (1969). The Sciences of the Artificial. Cambridge, MA: MIT Press.<br />

Smith, Brian Cantwell (1995/1985). Limits of Correctness in Computers. In Computers,<br />

Ethics and Social Values, ed. Deborah G. Johnson and Helen Nissenbaum (pp. 456–69).<br />

Englewood Cliffs, NJ: Prentice Hall. (Originally published in Computers and Society<br />

(ACM SIGCAS) 14/15: 18–26.).<br />

Sowa, John F. (2000). Knowledge Representation: Logical, Philosophical, and<br />

Computational Foundations. Pacific Grove, CA: Brooks/Cole.<br />

Toulmin, Stephen (1953). The Philosophy of Science. London: Hutchinson University<br />

Library.<br />

Toulmin, Stephen (1958). The Uses of Argument. Cambridge: Cambridge University<br />

Press.<br />

Turnbull, David (1994). Maps are Terri<strong>to</strong>ries. Science is an Atlas. Chicago: University of<br />

Chicago Press.<br />

Turner, Mark (1995). Review of The Poetics of Mind: Figurative Thought, Language, and<br />

Understanding, by Raymond W. Gibbs. Pragmatics and Cognition 3, 1: 179–85.<br />

Turner, Mark (1996). Literary Mind: The Origins of Thought and Language. Oxford:<br />

Oxford University Press.


Williams, Michael (<strong>2001</strong>). Problems of Knowledge: A Critical Introduction <strong>to</strong><br />

Epistemology. Oxford: Oxford University Press.<br />

Wimsatt, William C. (1987). False Models as Means <strong>to</strong> Truer Theories. In Neutral<br />

Models in Biology, ed. Matthew H. Nitecki and An<strong>to</strong>ni Hoffman (pp. 23–55). Oxford:<br />

Oxford University Press.<br />

Winder, William (1996). Textpert Systems. Text Technology 6, 3. Rpt. 1997, Computing<br />

in the Humanities Working Papers B.35. Accessed September 11, 2003. At<br />

http://www.chass.u<strong>to</strong>ron<strong>to</strong>.ca/epc/chwp/winder2/.<br />

Winograd, Terry (1991). Thinking Machines: Can There Be? Are We? In The Boundaries<br />

of Humanity: Humans, Animals, Machines, ed. James J. Sheehan and Mor<strong>to</strong>n Sosna (pp.<br />

198–223). Berkeley: University of California Press.<br />

Winograd, Terry and Fernando Flores (1986). Understanding Computers and Cognition:<br />

A New Foundation for Design. Bos<strong>to</strong>n: Addison-Wesley.<br />

Winsberg, Eric (<strong>2001</strong>). Simulations, Models, and Theories: Complex Physical Systems<br />

and Their Representations. Philosophy of Science (Proceedings) 68: S442–S454.<br />

Wood, Denis (1992). The Power of Maps. Mappings: Society/Theory/Space. London:<br />

Guilford Press.<br />

Ziman, John (2000). Real Science: What it Is, and What it Means. Cambridge:<br />

Cambridge University Press.<br />

20.<br />

Stylistic Analysis and Authorship Studies<br />

Hugh Craig<br />

Introduction<br />

Stylistics and authorship studies are siblings with obvious differences and important<br />

underlying similarities. Stylistic analysis is open-ended and explora<strong>to</strong>ry. It aims <strong>to</strong> bring<br />

<strong>to</strong> light patterns in style which influence readers' perceptions and relate <strong>to</strong> the disciplinary<br />

concerns of literary and linguistic interpretation. Authorship studies aim at "yes or no"<br />

resolutions <strong>to</strong> existing problems, and avoid perceptible features if possible, working at the<br />

base strata of language where imitation or deliberate variation can be ruled out.<br />

Authorship attribution has a forensic aspect – evidence of authorship based on Stylistics<br />

has been accepted in British courtrooms in particular – and this notion of determining a<br />

legally enforceable responsibility has given a particular intensity <strong>to</strong> scrutiny of its<br />

reliability. Yet stylistic analysis needs finally <strong>to</strong> pass the same tests of rigor, repeatability,


and impartiality as authorship analysis if it is <strong>to</strong> offer new knowledge. And the measures<br />

and techniques of authorship studies must ultimately be explained in stylistic terms if<br />

they are <strong>to</strong> command assent.<br />

In this chapter I offer first an example of a study in computational Stylistics <strong>to</strong> set the<br />

scene. Then I discuss the most important theoretical challenge <strong>to</strong> Stylistics, and go on <strong>to</strong><br />

sketch a methodological basis for the practice. Turning <strong>to</strong> authorship attribution, I treat<br />

some underlying ideas about authorship itself, then theoretical issues in attribution<br />

studies, and finally some practical considerations.<br />

Computational Stylistics aims <strong>to</strong> find patterns in language that are linked <strong>to</strong> the processes<br />

of writing and reading, and thus <strong>to</strong> "style" in the wider sense, but are not demonstrable<br />

without computational methods. We might, for instance, wish <strong>to</strong> examine patterns of<br />

association and difference among Shakespeare plays, based on their spoken dialogue. In<br />

these plays there are the obvious groupings of genre, and chronological divisions in<strong>to</strong><br />

early, middle, and late. There are clusters which have been intensively discussed by<br />

critics at various times, like the "problem plays" and the "four great tragedies." But how<br />

would the plays arrange themselves if only internal and empirical evidence was used?<br />

With the idea of starting with as few presumptions as possible, we take twenty-five<br />

Shakespeare plays (a good proportion of the thirty-eight complete plays normally<br />

included in a Complete Shakespeare) and calculate which are the dozen most common<br />

words in them overall. We can make a table of counts of each of these twelve words in<br />

each of the twenty-five plays.<br />

Principal components analysis (PCA) is a statistical technique much used in<br />

computational stylistics for analyzing the variation in a table like this. It creates new<br />

composite variables which are combinations of the original ones, with each of the latter<br />

given a separate weighting. The aim is <strong>to</strong> simplify the data by finding a few new<br />

variables which account for most of the relationships revealed by the table. To take a<br />

commonplace example, it might turn out that after collecting statistics on individuals'<br />

height and weight in a sample the two prove <strong>to</strong> be so closely related that a single new<br />

variable, size – height times weight, a composite of the original two – represents most of<br />

the variation. PCA vec<strong>to</strong>rs, "principal components", are an extension of this principle.<br />

The weightings in a component are worked out so as <strong>to</strong> create, first, the vec<strong>to</strong>r which<br />

accounts for the greatest proportion of the variance in the original table. Then the<br />

procedure finds a second independent vec<strong>to</strong>r which accounts for the next greatest<br />

proportion, and so on. If there are strong associations between variables, then the first<br />

few of these new composite variables will account for most of what is going on. In the<br />

Shakespeare plays table the first principal component accounts for 33 percent of the <strong>to</strong>tal<br />

variance, and the second for 19 percent, so we are justified in thinking that there are<br />

strong associations and contrasts in the way the variables behave. If there were not, and<br />

the variation was purely random, then we could expect each of the twelve principal<br />

components <strong>to</strong> have shared the variation equally, i.e., <strong>to</strong> account for around 8 percent.<br />

The first step is <strong>to</strong> look at the variable scores relative <strong>to</strong> each other on the first and second<br />

principal components, as in Figure 20.1.


Looking at the X axis first, the highest score is for frequencies of I; those of is, you, and it<br />

are most closely associated with it. These variables evidently behave like each other – in<br />

plays where one of them is notably frequent, the others tend <strong>to</strong> be frequent <strong>to</strong>o, and where<br />

one of them is notably scarce, the others will tend <strong>to</strong> be scarce also. At the other end, of is<br />

the extreme, associated with and and the. The first vec<strong>to</strong>r would seem <strong>to</strong> be a contrast<br />

between the variables which are associated with interactive dialogue, where pronouns<br />

like / and you are very common, and those associated with description and narration. The<br />

plays can also be given scores on the same principal components, as in Figure 20.2. The<br />

contrast along the X axis here emerges as one between the his<strong>to</strong>ry plays and the others.<br />

All the his<strong>to</strong>ry plays included are <strong>to</strong> the left of the rest. This fits well with the idea that<br />

the first principal component, the most important line of difference through the plays, is a<br />

contrast between plays in which description and its associated function words is stronger<br />

and those where interaction between characters predominates. Comedy and tragedy do<br />

not therefore present the biggest contrast in Shakespearean dialogue in these terms. Both<br />

the comedy and the tragedy groups are well mixed on the first axis. Within the comedy<br />

group the range is from Love's Labour's Lost (most disquisi<strong>to</strong>ry) <strong>to</strong> Two Gentlemen of<br />

Verona (most interactive); within the tragedies the range is from Hamlet <strong>to</strong> Othello.<br />

(Here, as very often in computational stylistics, we are relating two kinds of variables:<br />

fixed, externally determined attributes of the particular text like genre, known as<br />

independent variables, and counts of internal features, which are dependent variables.<br />

Correlation of the first with the second is the primary <strong>to</strong>ol of computational stylistics.)


20.1 Figure Variable weightings in a PCA of 12 word-variables in 25 shakespeare<br />

plays<br />

The second component, the vertical axis in the graphs, runs from <strong>to</strong> <strong>to</strong> a and from<br />

Richard II <strong>to</strong> Merry Wives of Windsor. High frequencies of <strong>to</strong> indicate a formal style<br />

(infinitives and prepositional uses contribute equally, and one arises through more<br />

complex constructions, the other by more precise specification), and high frequencies of<br />

a are associated with a more casual one. Once the disquisi<strong>to</strong>ry-interactive differences are<br />

accounted for, it seems that a high-style-low-style contrast is next most important in the<br />

plays. Along this axis the his<strong>to</strong>ry plays are strung out rather than bunched <strong>to</strong>gether,<br />

Richard II most formal of all and Henry IV Part 1 close <strong>to</strong> the opposite, informal<br />

extreme. The second component is not associated with the regular generic categories.<br />

Figure 20.1 and Figure 20.2 represent a sketch-map of the style of Shakespeare's plays,<br />

and of course raise more questions than they answer. In taking such a study beyond the<br />

sketch stage one would want <strong>to</strong> know how constant these patterns are when a play is<br />

added <strong>to</strong> the set or taken away, for instance, or if different variables are used, for instance<br />

by splitting <strong>to</strong> in<strong>to</strong> infinitive, prepositional, and adverbial uses. In such work, a little<br />

paradoxically, one wants <strong>to</strong> be reassured by seeing patterns already familiar from the way<br />

the texts are usually discussed, yet also <strong>to</strong> be surprised so that they seem more than a<br />

restatement of the obvious. If the dispositions of these graphs do stand up <strong>to</strong> scrutiny,<br />

then they invite further exploration. The first principal component has many similarities<br />

with the fac<strong>to</strong>r that emerged as the strongest in Biber's 1988 study of a large mixed<br />

sample of modern English. Biber labels the fac<strong>to</strong>r "high informational density and exact<br />

informational content versus affective, interactional, and generalized content" (1988:<br />

107). A similar contrast is dominant in a number of different studies of this kind, across<br />

many periods and text types. Should we think of this opposition as the fundamental one<br />

in categorizing texts? In Shakespeare studies the findings might be compared with<br />

Lancashire (1997), and the suggestion that the pattern of phrasal repetends indicates a<br />

segmentation of semantic knowledge in Shakespeare's mind between comedies and<br />

tragedies on the one hand and poetry and the his<strong>to</strong>ries on the other (1997: 182). Then,<br />

more locally still, one might ask what it is exactly that brings the two Richard plays<br />

<strong>to</strong>gether, which are from different tetralogies and not often discussed alongside each<br />

other, at the bot<strong>to</strong>m of Figure 20.2. Again, what does it mean that six of the tragedies lie<br />

in a close cluster, occupying quite a narrow range on both of these components? The<br />

outliers on both axes are his<strong>to</strong>ries or comedies: tragedies are more middling in style, or<br />

more mixed (an important difference, which might be explored by looking at segments<br />

from within the plays).<br />

Beyond these are questions of the design of the experiment. Why choose words, and why<br />

these words? Why choose plays, when one might choose oeuvres or periods as larger<br />

aggregations, or characters or scenes as segmentations, or some combination? How<br />

would the main lines of differentiation in a larger set of Elizabethan and Jacobean plays,<br />

going beyond Shakespeare, compare?


These are questions that are internal <strong>to</strong> stylistics, and can be answered in its own terms.<br />

There are other more fundamental ones. What, for example, is the status and nature of the<br />

axes of differences that the procedure has derived? They have a precise definition – each<br />

variable and each case has a score for each vec<strong>to</strong>r – but this says nothing about the<br />

stylistic context. Should an axis of this kind be regarded as a stylistic reality, figuring as<br />

part of the writing or the reading or hearing process, or just as an artifact of the method?<br />

Then, even if an axis of this kind is allotted some importance in relation <strong>to</strong> style, one<br />

must be cautious about describing it. Any departure from the purely enumerative ("I has<br />

the highest score on the first principal component") is an act of judgment and is open <strong>to</strong><br />

question. Then, if we accept that the vec<strong>to</strong>r has an authentic relationship <strong>to</strong> style, and is<br />

adequately interpreted, one can still ask whether anything of interest has been gained in<br />

relation <strong>to</strong> the questions that interest those who read and study Shakespeare plays.<br />

This example will indicate what is at stake in computational stylistics: at best, a powerful<br />

new line of evidence in long-contested questions of style; at worst, an elaborate display<br />

of meaningless patterning, and an awkward mismatch between words and numbers and<br />

the aesthetic and the statistical. For the moment we can note that this sort of work<br />

remains under challenge, and is still largely ignored by mainstream <strong>humanities</strong><br />

disciplines. It is worth reflecting also that though all its procedures individually predate<br />

the <strong>digital</strong> age, the combination in this sort of analysis is inconceivable without the<br />

computer. The labor involved with counting features like instances of word-types by<br />

hand and processing the results with a multivariate procedure like PCA the same way,<br />

would rule out embarking on such a project. Indeed, with no way of assembling and<br />

manipulating counts of word-variables, researchers in earlier periods assumed that very<br />

common words varied in frequency only in trivial ways among texts (Ellegard 1962: 15–<br />

16).<br />

A Challenge <strong>to</strong> Stylistics<br />

Stanley Fish has provided the most root-and-branch challenge <strong>to</strong> stylistics. He finds the<br />

whole project of identifying formal structures in a text and then interpreting them in<br />

relation <strong>to</strong> its meaning so flawed as <strong>to</strong> make stylistics entirely vacuous. A central problem<br />

for Fish is the assumption that meaning resides within the text rather than being created<br />

as it is read. He argues that the formal features described by stylisticians are meaningless<br />

except in relation <strong>to</strong> the reader's perception of them within a reading situation. When<br />

abstracted from this setting they refer <strong>to</strong> nothing but themselves and so any further<br />

analysis of patterns within their use, or comparison with the use of others or in other<br />

texts, or relating of them <strong>to</strong> meaning, is entirely pointless. Fish does not deny that there<br />

may be formal features which can enable categorical classification such as authorship<br />

(and so he is less hostile <strong>to</strong> authorship studies using computational stylistics), but he<br />

insists that these features cannot offer any information of interest about the text – just as a<br />

fingerprint may efficiently identify an individual but reveal nothing about his or her<br />

personality.<br />

The two essays on stylistics reprinted in Fish's 1980 book discuss work by stylisticians<br />

which he declares <strong>to</strong> be either circular, so that findings based on the analysis of formal


features are simply restated in "slightly different terms" (1980: 71–2), or arbitrary,<br />

because there is no principled way <strong>to</strong> link the meaning asserted with the feature. This<br />

rules out any claims <strong>to</strong> objectivity, since either there is no necessary connection between<br />

the features observed and the interpretation asserted for them, or the interpretation is built<br />

in<strong>to</strong> the choice and description of the features themselves. He can envisage a pure<br />

stylistics which confines itself <strong>to</strong> the description of formal features, or a confessedly<br />

impure stylistics which would make no claim <strong>to</strong> objectivity, but suggests that as practiced<br />

in the studies he analyzes it is worthless.<br />

Fish should be compulsory reading for any beginning stylistician. His papers belong with<br />

Schoenbaum's (1966) book as sources for cautionary tales in poor method. Any worker in<br />

the field will have felt tempted <strong>to</strong> do as one study Fish describes does, and slide from<br />

labeling a verb "active" <strong>to</strong> calling it "dynamic." Fish's comment is that the first "locates<br />

the verb in a system of formal differences", while the second "semanticizes, even<br />

moralizes" it (1980: 81). Again, it does seem illegitimate <strong>to</strong> move from "syntactic<br />

preferences" <strong>to</strong> "habits of meaning" or "conceptual orientation" (1980: 75), because, as<br />

Fish demonstrates, the step from preferences <strong>to</strong> significance can be made in any number<br />

of equally legitimate directions.<br />

Fish is speaking as a humanist: he sees stylistics as an attempt <strong>to</strong> make interpretation<br />

mechanical and thus <strong>to</strong> exclude the human. His affective stylistics, later modified so as <strong>to</strong><br />

make the agent interpretative communities rather than individual readers, is intended <strong>to</strong><br />

locate the study of meaning in the proper crea<strong>to</strong>rs, readers, or groups of readers, rather<br />

than in the linguistic features themselves. This is an attempt <strong>to</strong> incorporate the insight that<br />

"formal units are always a function of the interpretive model one brings <strong>to</strong> bear (they are<br />

not 'in the text')" (1980: 13).<br />

Fish wants <strong>to</strong> do more than ridicule the occasional excesses of stylistics: he wants <strong>to</strong><br />

demolish it entirely. The urgency in his analysis comes from his distrust of the claims of<br />

stylistics <strong>to</strong> a special "scientific" validity and of what he sees as its underlying antihumanist<br />

motivation, the exclusion of the human fac<strong>to</strong>r in reading and interpretation. It is<br />

possible, however, <strong>to</strong> propose an alternative motivation for stylistics, that is, the<br />

uncovering of patterns of language use which because of their "background" quality, or<br />

their emergence on a superhumanly wide scale, would otherwise not be noticed; and the<br />

testing of hypotheses about language use where some empirical validation seems possible<br />

and appropriate. An example of the former would be patterns of use of thou and you<br />

forms within one dramatist; an example of the latter would be curiosity about whether<br />

men and women write differently from one another. This need not be anti-humanist, since<br />

it would recognize the role of the interpreter in relating any empirically derived results <strong>to</strong><br />

readers' experiences, and <strong>to</strong> the myriad ways in which texts participate in a culture.<br />

A Methodological Basis for Stylistics<br />

A well-founded computational stylistics works with tendencies rather than rules; the<br />

semantic operation of language is so variable that the relationship of feature <strong>to</strong> meaning<br />

can never be fixed. Each element, and each combination of elements, can be used in new


contexts <strong>to</strong> mean new things. Yet there are also continuities and patterns. These require a<br />

mixed theory of meaning. A textual form has meaning in its local context, but also as part<br />

of a collective. Thus an instance of/ has a local meaning, a deixis in which the referent is<br />

the speaker, but also has a meaning as part of a wider abundance or scarcity of this<br />

pronoun in the text, reflecting a discourse more or less frequently framed as first-person<br />

expression.<br />

Tendencies are best observed in multiple instances, and it is here that the superhuman<br />

reach and memory of the computer can be of most assistance. Computational stylistics is<br />

thus extensive by nature, and it may be that the greatest potential for important<br />

discoveries lies in large-scale comparisons, rather than in the intensive study which is the<br />

staple of traditional philology. An obvious example would be change in language over<br />

time, within a writer's career or more largely over a collective artistic enterprise, or<br />

between periods. There have been promising forays also in<strong>to</strong> groupings of texts by<br />

gender and nationality.<br />

Stylistics may be thought of as the epidemiology of textual study: its methods allow<br />

general conclusions about the relationship between variables. Those which have high<br />

counts <strong>to</strong>gether form a syndrome. Once this is identified, the researcher can seek out a<br />

mechanism <strong>to</strong> explain it, as medical researchers might isolate the cell pathology which<br />

lies behind the epidemiological link between the incidence of smoking and that of lung<br />

cancer.<br />

Stylistics in its explora<strong>to</strong>ry form has had a great deal in common with sociolinguistics,<br />

which relies on correlations between the frequency of linguistic features and categorical<br />

independent variables like class and gender, with the more or less explicit assumption<br />

that language patterns are explained by these independent variables. Individualist<br />

linguistics and cognitive Stylistics, on the other hand, are not entirely sympathetic with<br />

strictly quantitative work, since they share an underlying belief that the most important<br />

dimensions of individuality cannot be captured in a common grid of frequencies.<br />

Johns<strong>to</strong>ne, for example, argues for the value of the qualitative as against the quantitative,<br />

for small as against large amounts of data, greater as against lesser detail, and the<br />

methods of cases and interpretation as against those of rules and instances (1996: 23–4).<br />

Stylistics is perhaps best seen as the offspring of Saussurean linguistics, prizing what she<br />

calls "the knowledge that can be modelled with rules and conventions" (1996: 188) both<br />

for categorical assignations (as in authorship studies) and for explorations of relations of<br />

form and meaning (as in Stylistics). The tension can be seen as one between an interest in<br />

the "more tractable explana<strong>to</strong>ry variables" (1996: 15–16) favored for sociolinguistic<br />

study and a sense that these are not adequate <strong>to</strong> explain language behavior. For<br />

individualists, a full explanation returns <strong>to</strong> the irreducibly unique self-expression of the<br />

human agent. This kind of humanism links them with Fish, who finds the particular<br />

instance always triumphant in its circumstantial complexity over the rule which is<br />

imposed on it. In their practice both Johns<strong>to</strong>ne and a practitioner of cognitive Stylistics<br />

like Lancashire favor a more conventional explication de texte, an intensive examination<br />

of a single passage, albeit against the background of a larger work or oeuvre. This<br />

Johns<strong>to</strong>ne calls "modern philology" (1996: 180–8), and consciously aligns with the


practice of literary criticism. In their practice the new individualists create a foundation<br />

for authorship attribution (via the motivation and practice of individual language users, or<br />

the nature of the cognitive process) but this is so extremely individualist that it militates<br />

against the kind of systematic comparison which has been the basic method of<br />

stylometrics and the more recent computational Stylistics.<br />

There must be room both for the insight that any text, and any collection of texts, has<br />

elements which will never be reducible <strong>to</strong> tabular form, as well as for the knowledge that<br />

many of the elements of the individual case will form part of a pattern. There is a strong<br />

instinct in human beings <strong>to</strong> reduce complexity and <strong>to</strong> simplify: this is a survival<br />

mechanism. Rules of thumb can save a great deal of time and effort. Stylistics is born of<br />

this instinct. What seems hard <strong>to</strong> explain in the individual case may be easier <strong>to</strong><br />

understand when it is seen in a larger context. If these elements can be put in tabular<br />

form, then one can harness the power of statistics and of numerous visualizing and<br />

schematizing <strong>to</strong>ols <strong>to</strong> help in the process of finding patterns.<br />

With a statistical method comes the possibility of falsifiability. This can be thought of in<br />

terms of a wager. If one researcher has a hunch about a literary problem – say, that drama<br />

written for the private theaters in the Shakespearean period is distinct in its style from<br />

that written for the public theaters – and bets a colleague that this is true, the wager can<br />

only be decided under strict conditions. The two will have <strong>to</strong> agree on how <strong>to</strong> judge<br />

difference; on which are clearly private-theater, and which clearly public-theater plays;<br />

and on how many of them amount <strong>to</strong> an adequate sample. Then the threshold conditions<br />

have <strong>to</strong> be agreed. What test will be used on what data from the plays, and which range of<br />

values will constitute difference, and which sameness, for the purposes of awarding the<br />

wager <strong>to</strong> one side or the other? Here random selection is a useful <strong>to</strong>ol, in cases where it is<br />

not possible <strong>to</strong> use all the possible data, and both sides of a question will want <strong>to</strong> escape<br />

preconceptions about relevant attributes and groups.<br />

This follows the lines of a more traditional, hypothesis-driven design. The alternative<br />

approach is through explora<strong>to</strong>ry data analysis, in which the researcher changes all<br />

possible parameters in the search for a revealing finding. Performed with due cautions,<br />

this may lead <strong>to</strong> discoveries that might be obscured by the starting conditions of a more<br />

fixed study. As the cost in time of collecting data and manipulating it and presenting it<br />

visually has come down, the attractiveness of explora<strong>to</strong>ry analysis has increased.<br />

In a sense (as is often pointed out by critics) computational stylistics merely creates<br />

another text, which, far from doing away with or au<strong>to</strong>mating interpretation, itself requires<br />

it. But this secondary text is related <strong>to</strong> the first in a systematic and repeatable way, and<br />

may be revealing about it. It offers a new perspective on it, like a spectroscopic analysis.<br />

The unfamiliar and often elaborate and arcane presentation of the results of statistical<br />

study in stylistics should not obscure what it has in common with traditional practices. It<br />

is common in any analysis <strong>to</strong> move from relatively uncontroversial <strong>to</strong> relatively<br />

controversial observations. This definition of some readily agreed-on aspects of a text is a<br />

prelude <strong>to</strong> interpretation. These results may be gathered by hand, as in the observation<br />

that plays written by women tend <strong>to</strong> have more women characters, and that women


characters in them are more likely <strong>to</strong> start and end a scene. In principle, work in<br />

computational stylistics is no different.<br />

In recent years the most-used markers for computational stylistics have been function<br />

words. Other very common words which do not appear <strong>to</strong> be unduly sensitive <strong>to</strong> subject<br />

matter are also attractive. Then there are combinations of word-types: in Lancashire's<br />

terminology, the fixed phrase, collocation (pair of words within a certain number of<br />

words of each other), and word cluster (combination of word phrase and collocation)<br />

(Lancashire 1997: 172). Lancashire's understanding of language production as mostly<br />

instinctive, and based on an associative memory, leads him <strong>to</strong> highlight phrases as the<br />

key markers of idiolect – phrases of up <strong>to</strong> seven words since that is the limit of short-term<br />

or working memory. "We speak and write in chunks", he says (1997: 178–80). The<br />

existence of large commercial online full-text databases makes it now possible <strong>to</strong> use this<br />

kind of authorship marker against the background of a large corpus, which can show if<br />

any parallels between the phrases of a doubtful text and those of a target author are truly<br />

unusual. There are some sample studies in Jackson (2002). Among the other most<br />

important style markers have been choices among pairs of words which can be readily<br />

substituted for each other in a sentence, such as while and whilst, has and hath, on and<br />

upon, and the words used <strong>to</strong> begin sentences. McMenamin details a vast range of these<br />

markers; the range itself, and the dangers of arbitrary choices among them, have<br />

contributed <strong>to</strong> doubts about the overall validity of such studies (Furbank and Owens<br />

1991).<br />

The best method of guarding against these dangers is careful testing of the method<br />

alongside its application <strong>to</strong> the immediate problem. If the markers and procedures chosen<br />

provide an efficient separation of samples in<strong>to</strong> authorial or other groups in a closely<br />

comparable set – ideally, drawn at random from the main set and then reserved for testing<br />

– then that provides a guide <strong>to</strong> the reliability of the method when it comes <strong>to</strong> the<br />

contested texts. Discriminant analysis, a method for data reduction with some similarities<br />

<strong>to</strong> PCA, provides a good illustration. Discriminant analysis provides a weighted vec<strong>to</strong>r<br />

which will maximize the separation between two groups of samples named in advance.<br />

The vec<strong>to</strong>r can then be held constant and a score can be worked out for a mystery<br />

segment. The procedure even supplies a probability that any given segment belongs <strong>to</strong><br />

one group or the other. This seems ideal for authorship problems: simply provide some<br />

samples of author A, some of author B, calculate the Discriminant function, and then test<br />

any doubtful segment. Yet Discriminant results are no<strong>to</strong>riously optimistic: the separation<br />

of the segments in<strong>to</strong> named groups is maximized, but we have no way of knowing if they<br />

are representative of the population of the works of Author A or B – which, in the fullest<br />

sense, is the range of works it was (or is) possible for Author A or B <strong>to</strong> write. There is the<br />

danger of "overtraining" – the method will work superbly for these particular samples,<br />

but what it is providing is exactly a separation of those samples, which may be a very<br />

different matter from a true authorial separation. Good practice, therefore, is <strong>to</strong> reserve<br />

some samples (as many as 10 percent) as test samples. They do not contribute <strong>to</strong> the<br />

"training" exercise, the formation of the function, and therefore are in the same position<br />

as the doubtful sample or samples. If these are correctly assigned <strong>to</strong> their group, or most<br />

of them are, then one can have some confidence that the result on the mystery sample is


eliable. "Testing the test" in this way should be the first move in any statistical authorial<br />

experiment. The ease of use of statistical packages makes it possible <strong>to</strong> do a large amount<br />

of testing even with scarce data by "bootstrapping" – a single item can be withdrawn<br />

from the training set and tested, then returned <strong>to</strong> the set while a second is extracted and<br />

tested, and so on.<br />

The claim for computational stylistics is that it makes available a class of evidence not<br />

otherwise accessible (i.e., not <strong>to</strong> the naked eye). This evidence is comparable <strong>to</strong> the<br />

evidence interpreters always use, if not <strong>to</strong> form their views, then at least <strong>to</strong> persuade<br />

others that they are true.<br />

The computational varieties of stylistic analysis and authorship studies require some<br />

considerable immersion in traditional <strong>humanities</strong> disciplines (scholarly, i.e.,<br />

bibliographical, literary-his<strong>to</strong>rical and textual; and critical, i.e., interpretative and<br />

theoretical); in <strong>humanities</strong> computing generally (in particular the compiling and editing<br />

of literary texts and assembling them in<strong>to</strong> corpora); and statistical (involving the<br />

understanding of statistical techniques and the practicalities of statistics and spreadsheet<br />

programs). All three present enough challenges and large enough bodies of knowledge <strong>to</strong><br />

occupy the working lifetimes of individual researchers by themselves. Most commonly,<br />

single practitioners are experts in one or possibly two areas, very rarely in three, and have<br />

<strong>to</strong> make do in various more or less satisfac<strong>to</strong>ry ways in a third. Often the requisite skills<br />

have been assembled in two practitioners (as in the case of Burrows and Love).<br />

Authorship<br />

Authorship attribution is as old as writing itself, and its his<strong>to</strong>ry displays a fascinating<br />

variety of problems and solutions. Groupings of texts (Homer, the Bible, Shakespeare)<br />

may have been created at times when their coherence was not especially significant, but<br />

later generations have placed enormous importance on the difference between the<br />

canonical and the apocryphal in each case. The modern interest in authorship attribution<br />

derives from the Renaissance, when the availability of texts made comparative study<br />

possible, and a new critical spirit went with the linguistic and textual disciplines of<br />

Humanism. The demonstration by Lorenzo Valla in the fifteenth century that the<br />

Donation of Constantine, which gave the western part of the Roman Empire <strong>to</strong> Pope<br />

Sylvester, was a forgery, is perhaps the most famous example (Love 2002: 18–19).<br />

In the modern era most texts come securely attributed on external evidence. The title<br />

page of the first edition announces the author, or some knowledgeable contemporary<br />

assigns the work <strong>to</strong> the author. Some exuberant authors include their name in the text<br />

itself, as Ben Jonson does in his ode <strong>to</strong> Lucius Gary and Henry Morison. These<br />

attributions may be a little deceptive in their straightforwardness – there are<br />

collabora<strong>to</strong>rs, edi<strong>to</strong>rs, and writers of source materials <strong>to</strong> complicate matters – but for<br />

reasons of economy of effort such approximations (Charles Dickens wrote Great<br />

Expectations) are allowed <strong>to</strong> stand.


Then for texts without this explicit and decisive external evidence there are numerous<br />

considerations of <strong>to</strong>pic, approach, attitudes, imagery, turns of phrase, and so on which<br />

have always served scholars as foundations for attributions. The balance between the<br />

credibility of this internal evidence and the external kind has swung back and forth. The<br />

1960s were a low point for internal evidence, as a reaction <strong>to</strong> the undisciplined<br />

accumulation of parallel passages and the wholesale "disintegration" of canons like<br />

Shakespeare's. This skepticism about internal evidence can be seen in Schoenbaum's<br />

(1966) book and in the 1966 Erdman and Fogel collection.<br />

Humanities computing is involved with a second generation of attribution based on<br />

internal evidence, depending on measuring stylistic features in the doubtful and other<br />

texts, and comparing the results. An interesting case study is the debate over attributing<br />

the Funerall Elegie for William Peter by W. S. <strong>to</strong> Shakespeare, which began with Foster's<br />

(1989) book. The only evidence in favor of the attribution was internal, and stylistic in<br />

the specialized sense, being quantitative on the one hand and concerned with largely<br />

unnoticed idiosyncracies of language on the other. Both sides of the debate agreed that<br />

the poem's style, in the more usual sense of the imagery, diction, and language use<br />

noticed by readers, was unlike canonical Shakespeare. The proponents of the attribution<br />

argued that since the quantitative and stylistic evidence was so strong, the generally held<br />

view of the "Shakespearean" would just have <strong>to</strong> change. (Recently, another candidate,<br />

John Ford, has been proposed, for whom there were none of the cognitive dissonances<br />

just mentioned, and whose work appears <strong>to</strong> satisfy both readers and stylisticians as a<br />

good match for the Elegie.)<br />

Underpinning any interest in assigning authorship is a model of the author. Since the<br />

1970s the traditional scholarly activity of determining authorship has been conducted<br />

with a certain unease, resulting from the work of the French post-structuralists Roland<br />

Barthes, Michel Foucault, and Jacques Derrida, who, in undermining what they saw as<br />

the bourgeois individual subject, displaced the author as the primary source of meaning<br />

for texts. In the literary sphere this individual subject had reached an apogee in the<br />

Romantic idea of the heroic individual author. Since structuralism, there has been more<br />

interest in the role of discourse, culture, and language itself in creating texts.<br />

There are, as Furbank and Owens (1988) elaborate, special cautions which should attach<br />

<strong>to</strong> the activity of adding items <strong>to</strong> a canon. The more items included, the wider the canon,<br />

and the easier it is <strong>to</strong> add further ones (1988: 4). There is no such thing as adding a work<br />

temporarily <strong>to</strong> a canon (1988: 30). They urge resistance <strong>to</strong> the pressure <strong>to</strong> assign authors,<br />

since doing this has so many consequences for the canon <strong>to</strong> which the work is added:<br />

after all, "works do not need <strong>to</strong> be assigned <strong>to</strong> authors and it does not matter that<br />

anonymous works should remain anonymous" (1988: 30). It is not enough that a work is<br />

a "plausible" addition <strong>to</strong> a canon, since once added, works are very hard <strong>to</strong> subtract<br />

(1988: 15).<br />

There is also a more traditional objection <strong>to</strong> the energy that has gone in<strong>to</strong> attribution. In<br />

Erasmus's words in his edition of St Jerome, "What is so important about whose name is<br />

on a book, provided it is a good book?" (1992: 75). Erasmus's answer is that this may


indeed not matter in the case of a mere playwright like Plautus, whose works he has<br />

discussed earlier (1992: 71), but is vital in the case of "sacred writers and pillars of the<br />

church" like Jerome. If "nonsense" by others is presented as the work of such writers, and<br />

the imposture is not detected, then readers are forced <strong>to</strong> remain silent about any doubts or<br />

<strong>to</strong> accept falsehood (1992: 76). (Erasmus's attribution methods are discussed in Love<br />

2002: 19–22). A modern critic might argue that authorship is important even in the case<br />

of a dramatist like Plautus. Attaching a name <strong>to</strong> a work "changes its meaning by changing<br />

its context … certain kinds of meaning are conferred by its membership and position in<br />

the book or oeuvre. Hamlet by William Shakespeare is a different play from Hamlet by<br />

the Earl of Oxford or Francis Bacon" (Love 2002: 46). Kermode suggests that "a<br />

different and special form of attention" is paid <strong>to</strong> the work of a famous writer (quoted in<br />

Furbank and Owens 1988: 44). Even within the world of scholarship, discovery that an<br />

essay was not after all by an important thinker like Foucault would make a considerable<br />

difference (Love 2002: 96–7).<br />

The signs are that models of authorship are evolving from the post-structuralist "author<br />

function", an intersection of discourses, <strong>to</strong>ward a concept more influenced by the<br />

workings of cognitive faculties. If the long-term memory s<strong>to</strong>re of human beings works<br />

not systematically but by "casting out a line for any things directly or indirectly<br />

associated with the object of our search", then "the organisation of memories … reflects<br />

the person's own past experience and thought rather than a shared resource of cultural<br />

knowledge" and this would imply a "unique idiolect" for each individual's speech or<br />

writing (Lancashire 1997: 178).<br />

Then there has been a renewed interest in the linguistics of the individual speaker, whose<br />

differences from other speakers, according <strong>to</strong> Johns<strong>to</strong>ne, should be accorded<br />

"foundational status" (1996: 21). She shows that, even in highly constraining situations<br />

like academic discourse or conducting or answering a telephone survey, speakers tend <strong>to</strong><br />

create an individual style, and <strong>to</strong> maintain this style across different discourse types.<br />

Sociolinguistics explains difference through social categories (most often class, gender,<br />

and race) or rhe<strong>to</strong>rical ones (purpose and audience), but Johns<strong>to</strong>ne argues that these<br />

should be seen as resources from which the individual constructs difference rather than<br />

the determinants of it (1996: ix-x). She is prepared <strong>to</strong> envisage a return <strong>to</strong> the Romantic<br />

view of the importance of the individual in language, overturning the highly influential<br />

arguments of Saussure that language study should concern itself with langue, the system<br />

of a language, rather than parole, the individual instance of language production (1996:<br />

20) and challenging the prestige of abstract scientific laws which have meant that, in the<br />

words of Edward Sapir, "[t]he laws of syntax acquire a higher reality than the immediate<br />

reality of the stammerer who is trying '<strong>to</strong> get himself across'" (quoted in Johns<strong>to</strong>ne 1996:<br />

20).<br />

Theoretical Considerations for Attribution<br />

The practicalities of attribution by stylistic means hinge on the question of the variation<br />

in the style of an author. In Sonnet 76 Shakespeare's speaker laments that he cannot vary<br />

his style in line with the poetic fashion, with the result that everything he writes is


hopelessly easy <strong>to</strong> identify as his own, because it is "So far from variation or quick<br />

change." (There is a special reason in this case: he always writes on the same subject, on<br />

his beloved and his love.) In the eighteenth century, Alexander Pope thought that<br />

attributing authorship by style was foolish – on the grounds, it seems, that it was <strong>to</strong>o easy<br />

for an author <strong>to</strong> "borrow" a style (quoted in Craig 1992: 199). Variation, in other words,<br />

was unlimited. Samuel Johnson later in the same century <strong>to</strong>ok the opposite point of view.<br />

His friend and biographer James Boswell asked him if everyone had their own style, just<br />

as everyone has a unique physiognomy and a unique handwriting. Johnson answered<br />

emphatically in the affirmative: "Why, Sir, I think every man whatever has a peculiar<br />

style, which may be discovered by nice examination and comparison with others: but a<br />

man must write a great deal <strong>to</strong> make his style obviously discernible" (quoted in Love<br />

2002: 7).<br />

The evidence from the vast activity in empirical work on authorship supports a qualified<br />

version of Johnson's view. An attribution <strong>to</strong> satisfy most standards of proof is possible on<br />

internal grounds provided the doubtful sample is of sufficient length, and sufficient<br />

samples for comparison in similar text types by candidate writers are available. It is<br />

reasonable <strong>to</strong> say that the extreme skeptics about "stylometry" or "non-traditional<br />

authorship attribution studies", those who suggested that it had similar claims <strong>to</strong> useful<br />

information about authorship <strong>to</strong> those of phrenology about personality (Love 2002: 155),<br />

have been proved wrong.<br />

Statistics depends on structured variation – on finding patterns in the changes of items<br />

along measurable scales. It is easy <strong>to</strong> see that language samples must vary in all sorts of<br />

ways as different messages are composed and in different modes and styles. The claims<br />

of statistical authorship attribution rest on the idea that this variation is constrained by the<br />

cognitive faculties of the writer. The writer will compose various works in the same or<br />

different genres and over a more or less extended career. Language features must vary in<br />

frequency within his or her output. Chronology and genre are readily detectable as<br />

sources of change; readers might expect <strong>to</strong> tell the difference between early and late<br />

Henry James, or between the writing in a comic novel and a serious essay by the same<br />

writer. Play dialogue presents an expected sharp variation in style within the same work,<br />

which may contain old and young speakers, men and women, the rich and the poor, the<br />

witty and the dull, and so on. The approach <strong>to</strong> authorial idiolect through cognitive science<br />

and neurology offers its own reinforcement of the notion that different genres are treated<br />

differently, and that word-patterns can be acquired and also lost over a lifetime<br />

(Lancashire 1997: 182, and 1999: 744).<br />

There is, however, evidence which suggests that authorial consistency is quite a strong<br />

fac<strong>to</strong>r in most eras, so that one can expect <strong>to</strong> find other sources of systematic variation<br />

nested within it. Early James and late James are different, but not so different as <strong>to</strong><br />

override the difference between James and Thomas Hardy. The characters created by a<br />

single dramatist scatter on many variables, but the scatter may still be constrained enough<br />

so that their multidimensional "terri<strong>to</strong>ry" does not overlap with a second dramatist of the<br />

same period writing in the same genres. There is contradic<strong>to</strong>ry evidence on this matter<br />

from empirical studies. Burrows (1991) shows that Henry Fielding's style in his parody of


Samuel Richardson remained close enough <strong>to</strong> his usual pattern <strong>to</strong> group his parody,<br />

Sbamela, with his other writing, though it did move some way <strong>to</strong>ward the style of<br />

Richardson himself. On the other side of the ledger is the case of Remain Gary, who in<br />

the 1970s wrote two novels under a pseudonym in an effort <strong>to</strong> escape the reception which<br />

he felt his established reputation influenced <strong>to</strong>o heavily. These novels were successful,<br />

and indeed after the publication of the second of them Gary won a second Prix Goncourt<br />

under his nom de plume, Emile Ajar. Tirvengadum (1998) reports that in the second of<br />

the Ajar novels Gary was able <strong>to</strong> change his style so radically as <strong>to</strong> make the profile of<br />

his common-word usage distinct from that in his other work.<br />

The search for stylistic markers which are outside the conscious control of the writer has<br />

led <strong>to</strong> a divergence between literary interpretation and stylometry, since, as Hor<strong>to</strong>n puts<br />

it, "the textual features that stand out <strong>to</strong> a literary scholar usually reflect a writer's<br />

conscious stylistic decisions and are thus open <strong>to</strong> imitation, deliberate or otherwise"<br />

(quoted in Lancashire 1998: 300). In support of these unconscious style markers are the<br />

studies that show that much of language production is done by parts of the brain which<br />

act in such swift and complex ways that they can be called a true linguistic unconscious<br />

(Crane <strong>2001</strong>: 18). Lancashire (1997: 177) adds: "This is not <strong>to</strong> say that we cannot<br />

ourselves form a sentence mentally, edit it in memory, and then speak it or write it, just<br />

that the process is so arduous, time-consuming, and awkward that we seldom strive <strong>to</strong> use<br />

it."<br />

Naturally it is incumbent on the researchers <strong>to</strong> show their audiences how markers<br />

invisible <strong>to</strong> writers or readers can be operative in texts at more than a narrowly statistical<br />

level, <strong>to</strong> avoid what Furbank and Owens call "the spectre of the meaningless" (1988:<br />

181). As Love argues, a stylistic explanation for grouping or differentiating texts is<br />

always <strong>to</strong> be preferred <strong>to</strong> a stylometric or "black-box" one (2002: 110). Unlike most<br />

research in the <strong>humanities</strong>, results of the purely computational kind cannot be checked<br />

against the reader's recollection or fresh study of the texts: in principle, the only check is<br />

a replication of the statistical tests themselves.<br />

Practical Considerations in Attribution<br />

The first fac<strong>to</strong>r <strong>to</strong> consider in an authorship question is the number of candidates<br />

involved. There may be no obvious candidates indicated by external evidence; a group of<br />

candidates, from two <strong>to</strong> a very large but still defined number; or there may be a single<br />

candidate with an unlimited group of other possible authors. Commenta<strong>to</strong>rs have<br />

generally been pessimistic about all but two-author problems. Furbank and Owens<br />

(1988), assessing the potential usefulness of quantitative stylistics in testing attributions<br />

in the very large group of texts associated with Daniel Defoe, concluded that these new<br />

methods "were not yet in a position <strong>to</strong> replace more traditional approaches" (1988: 183).<br />

They proceeded <strong>to</strong> work on external and purely qualitative internal evidence in<br />

establishing a revised Defoe canon. The most-cited successful attribution on quantitative<br />

measures, the assignation of a group of essays in The Federalist, is an adjudication<br />

between two candidates. The disputed papers could only have been written by Alexander


Madison or John Hamil<strong>to</strong>n; the problem was solved by Mosteller and Wallace (1964)<br />

using function-word markers.<br />

There are grounds for optimism, on the other hand, that with advances in technique and<br />

in the availability of good-quality, well-suited data, attribution by computational stylistics<br />

can make a contribution even in complicated cases. Burrows (2002), for instance, reports<br />

success with a recently developed technique adapted for multiple-candidate problems. It<br />

establishes multiple authorial profiles and determines a distance from each of these <strong>to</strong> the<br />

target text, as a measure of which author's work the mystery piece is "least unlike."<br />

The way forward would seem <strong>to</strong> be by a double movement: calibrating the reliability of<br />

attribution methods by good technique and sheer collective experience, and on the other<br />

hand, with the increasing availability of text for comparison, advancing on even the most<br />

difficult problems (especially those involving a single author as against an unlimited<br />

group of others, and searching for good candidates from a very large pool). It may be<br />

misguided <strong>to</strong> aim for complete certainty in attaching a single name <strong>to</strong> a doubtful text.<br />

There is, after all, much <strong>to</strong> be gained from eliminating one or more candidates, while<br />

remaining uncertain about the actual author, or narrowing authorship <strong>to</strong> a small group of<br />

names.<br />

It may be useful <strong>to</strong> conclude with a brief list of commonly encountered pitfalls in<br />

authorial attribution by computational stylistics:<br />

1 Assuming that, if groups separate according <strong>to</strong> author, the separation is authorial.<br />

This is obvious when one authorial group is all tragedies and the second comedies. Even<br />

if all texts are from the same genre, there must always be the question whether the<br />

difference is really between (say) comedies set in the city and those set in the court.<br />

2 Assuming that parallels between elements of a doubtful text and a given author<br />

provide an argument for authorship when those elements may be found in any number of<br />

other authors outside the sample tested. This was the downfall of what Schoenbaum calls<br />

the "parallelographic school" (1966: 94).<br />

3 Selecting those techniques, and those features, which favor one case, while ignoring<br />

others ("cherry-picking"). Furbank and Owens (1991: 243) put it this way: "any liberty <strong>to</strong><br />

choose one test, rather than another for a given authorship problem will lay you open <strong>to</strong><br />

incalculable, perhaps subliminal, temptations <strong>to</strong> choose the one that will help you <strong>to</strong><br />

prove what you want <strong>to</strong> prove."<br />

4 Assuming that an author cannot vary from his or her normal style when carrying out<br />

a particular assignment, or under a particular internal or external influence.<br />

5 Using a new attribution, not yet generally accepted, as the basis for further attribu<br />

tions – what has been called in art-his<strong>to</strong>rical studies "the forging of chains", Friedlander's<br />

term for works in the visual arts added <strong>to</strong> an oeuvre on the basis of connoisseurship and<br />

then used <strong>to</strong> justify the addition of further works (cited in Furbank and Owens 1988: 47).


More positively, one might consider a set of ideal conditions for an attribution on stylistic<br />

grounds:<br />

1 The doubtful text is long. (In the experiment in assigning Res<strong>to</strong>ration poems <strong>to</strong> their<br />

authors described in Burrows (2002), reliability increases with sample length.)<br />

2 There is a small number of candidates (two is ideal, as in the case of the federalist<br />

papers).<br />

3 There is no significant involvement in the doubtful text by other writers as collabor<br />

a<strong>to</strong>rs or revisers, and very few changes have been made by third parties such as printers<br />

and edi<strong>to</strong>rs between composition and the surviving texts.<br />

4 There is a lot of securely attributed text by the candidates, in a similar genre and<br />

from a similar period <strong>to</strong> the doubtful text.<br />

Purists would say that all these conditions must be met before a secure attribution can be<br />

made, but it may be more realistic <strong>to</strong> see a continuum of reliability <strong>to</strong> which all these<br />

conditions contribute. Attribution work may still yield valuable results, in eliminating<br />

some of the candidates, for example, where there are significant deficits in one or more of<br />

these conditions.<br />

Conclusion<br />

Computer-assisted stylistics and authorship studies are at an interesting stage. A<br />

remarkable range of studies has been built up; methods have been well calibrated by a<br />

variety of researchers on a variety of problems; and even if some studies have proved<br />

faulty, the vigorous discussion of their shortcomings is a resource for those who follow.<br />

Well-tested methods can now provide multiple approaches <strong>to</strong> a given problem, so that<br />

results can be triangulated and cross-checked. There is an ever-increasing volume of text<br />

available in machine-readable form: this means that a large purpose-built corpus can now<br />

be assembled quite quickly in many areas. There are enough successes <strong>to</strong> suggest that<br />

computational stylistics and non-traditional attribution have become essential <strong>to</strong>ols, the<br />

first places one looks <strong>to</strong> for answers on very large questions of text patterning, and on<br />

difficult authorship problems. It is worth noting, <strong>to</strong>o, that the lively debate provoked by<br />

computational work in stylistics and authorship is an indication that these activities are<br />

playing a significant part in a much wider contemporary discussion about the relations of<br />

the human and the mechanical.<br />

References for Further Reading<br />

Biber, D. (1988). Variation across Speech and Writing. Cambridge: Cambridge<br />

University Press.


Burrows, J. F. (1987). Computation in<strong>to</strong> Criticism: A Study of Jane Austen and an<br />

Experiment in Method. Oxford: Clarendon Press.<br />

Biber, D. (1991). "I lisp'd in numbers": Fielding, Richardson and the Appraisal of<br />

Statistical Evidence. Scriblerian and the Kit-Kats 23: 234–41.<br />

Biber, D. (2002). "Delta": A Measure of Stylistic Difference and a Guide <strong>to</strong> Likely<br />

Authorship. Literary and Linguistic Computing 17: 267–87.<br />

Burrows, J. F., and H. Love (1998). The Role of Stylistics in Attribution: Thomas<br />

Shadwell and "The Giants' War". Eighteenth Century Life 22: 18–30.<br />

Craig, D. H. (1992). Authorial Styles and the Frequencies of Very Common Words:<br />

Jonson, Shakespeare, and the Additions <strong>to</strong> "The Spanish Tragedy." Style 26: 199–220.<br />

Crane, M. T. (<strong>2001</strong>). Shakespeare's Brain: Reading with Cognitive Theory. Prince<strong>to</strong>n,<br />

NJ: Prince<strong>to</strong>n University Press.<br />

Ellegård, Alvar (1962). A Statistical Method for Determining Authorship. Gothenburg:<br />

University of Gothenburg.<br />

Erasmus, Desiderius (1992). Collected Works of Erasmus. Vol. 61. Patristic Scholarship:<br />

The Edition of St Jerome, ed. J. F. Brady and John C. Olin. Toron<strong>to</strong>: University of<br />

Toron<strong>to</strong> Press.<br />

Erdman, David V. and Ephim G. Fogel, (eds.) (1966). Evidence for Authorship: Essays<br />

on Problems of Attribution. Ithaca, NY: Cornell University Press.<br />

Fish, S. (1980). Is there a Text in this Class? The Authority of Interpretive Communities.<br />

Cambridge, MA: Harvard University Press.<br />

Foster, Donald W (1989). Elegy by W. S.: A Study in Attribution. Newark, DE: University<br />

of Delaware Press.<br />

Furbank, P. N. and W. R. Owens (1988). The Canonisation of Daniel Defoe. New Haven,<br />

CT: Yale University Press.<br />

Furbank, P. N. and W. R. Owens (1991). Dangerous Relations. Scriblerian and the Kit-<br />

Kats 33: 242–4.<br />

Jackson, MacD. P. (2002). Determining Authorship: A New Technique. Research<br />

Opportunities in Renaissance Drama 41: 1–14.<br />

Johns<strong>to</strong>ne, B. (1996). The Linguistic Individual: Self-Expression in Language and<br />

Linguistics. New York: Oxford University Press.


Lancashire, I. (1997). Empirically Determining Shakespeare's Idiolect. Shakespeare<br />

Studies 25: 171–85.<br />

Lancashire, I. (1998). Paradigms of Authorship. Shakespeare Studies 26: 298–301.<br />

Lancashire, I. (1999). Probing Shakespeare's Idiolect in Troilus and Cressida, 1.3.1–29.<br />

University of Toron<strong>to</strong> Quarterly 68: 728–67.<br />

Love, Harold (2002). Attributing Authorship: An Introduction. Cambridge: Cambridge<br />

University Press.<br />

McMenamin, G. (1993). Forensic Stylistics. Amsterdam: Elsevier.<br />

Milic, Louis T. (1991). Progress in Stylistics: Theory, Statistics, Computers. Computers<br />

and the Humanities 25: 393–400.<br />

Mosteller, F. and D. L. Wallace (1964). Inference and Disputed Authorship: The<br />

Federalist. Reading, MA: Addison-Wesley.<br />

Schoenbaum, S. (1966). Internal Evidence and Elizabethan Dramatic Authorship: An<br />

Essay in Literary His<strong>to</strong>ry and Method. London: Arnold.<br />

Tirvengadum, Vina (1998). Linguistic Fingerprints and Literary Fraud. Computing in<br />

the Humanities Working Papers. Accessed November 1, 2002. At<br />

http://www.chass.u<strong>to</strong>ron<strong>to</strong>.ca/epc/chwp/tirven/index.html.<br />

20.2 Figure Scores for texts in a PCA of 12 word-variables in 25 Shakespeare plays<br />

21.<br />

Preparation and Analysis of Linguistic Corpora<br />

Nancy Ide<br />

The corpus is a fundamental <strong>to</strong>ol for any type of research on language. The availability of<br />

computers in the 1950s immediately led <strong>to</strong> the creation of corpora in electronic form that


could be searched au<strong>to</strong>matically for a variety of language features, and compute<br />

frequency, distributional characteristics, and other descriptive statistics. Corpora of<br />

literary works were compiled <strong>to</strong> enable stylistic analyses and authorship studies, and<br />

corpora representing general language use became widely used in the field of<br />

lexicography. In this era, the creation of an electronic corpus required entering the<br />

material by hand, and the s<strong>to</strong>rage capacity and speed of computers available at the time<br />

put limits on how much data could realistically be analyzed at any one time. Without the<br />

Internet <strong>to</strong> foster data sharing, corpora were typically created, and processed at a single<br />

location. Two notable exceptions are the Brown Corpus of American English (Kucera<br />

and Francis 1967) and the London/Oslo/Bergen (LOB) corpus of British English<br />

(Johansson et al. 1978); both of these corpora, each containing 1 million words of data<br />

tagged for part of speech, were compiled in the 1960s using a representative sample of<br />

texts produced in the year 1961. For several years, the Brown and LOB were the only<br />

widely available computer-readable corpora of general language, and therefore provided<br />

the data for numerous language studies.<br />

In the 1980s, the speed and capacity of computers increased dramatically, and, with more<br />

and more texts being produced in computerized form, it became possible <strong>to</strong> create<br />

corpora much larger than the Brown and LOB, containing millions of words. The<br />

availability of language samples of this magnitude opened up the possibility of gathering<br />

meaningful statistics about language patterns that could be used <strong>to</strong> drive language<br />

processing software such as syntactic parsers, which sparked renewed interest in corpus<br />

compilation within the computational linguistics community. Parallel corpora, which<br />

contain the same text in two or more languages, also began <strong>to</strong> appear; the best known of<br />

these is the Canadian Hansard corpus of Parliamentary debates in English and French.<br />

Corpus creation still involved considerable work, even when texts could be acquired from<br />

other sources in electronic form. For example, many texts existed as typesetter's tapes<br />

obtained from publishers, and substantial processing was required <strong>to</strong> remove or translate<br />

typesetter codes.<br />

The "golden era" of linguistic corpora began in 1990 and continues <strong>to</strong> this day. Enormous<br />

corpora of both text and speech have been, and continue <strong>to</strong> be, compiled, many by<br />

government-funded projects in Europe, the USA, and Japan. In addition <strong>to</strong> monolingual<br />

corpora, several multilingual parallel corpora covering multiple languages have also been<br />

created. A side effect of the growth in the availability and use of corpora in the 1990s was<br />

the development of au<strong>to</strong>matic techniques for annotating language data with information<br />

about its linguistic properties. Algorithms for assigning part of speech tags <strong>to</strong> words in a<br />

corpus and aligning words and sentences in parallel text (i.e., associating each word or<br />

sentence with its translation in the parallel version) were developed in the 1990s that<br />

achieve 95–98 percent accuracy. Au<strong>to</strong>matic means <strong>to</strong> identify syntactic configurations<br />

such as noun phrases, and proper names, dates, etc., were also developed.<br />

There now exist numerous corpora, many of which are available through the Linguistic<br />

Data Consortium (LDC) (http://www.ldc.upenn.edu) in the USA and the European<br />

Language Resources Association (ELRA) (http://www.elra.org) in Europe, both of which<br />

were founded in the mid-1990s <strong>to</strong> serve as reposi<strong>to</strong>ries and distribu<strong>to</strong>rs of corpora and


other language resources such as lexicons. However, because of the cost and difficulty of<br />

obtaining some types of texts (e.g., fiction), existing corpora vary considerably in their<br />

composition; very few efforts have been made <strong>to</strong> compile language samples that are<br />

"balanced" in their representation of different genres. Notable exceptions (apart from the<br />

early Brown and LOB) are the British National Corpus (BNC)<br />

(http://www.hcu.ox.ac.uk/BNC/) and the American National Corpus (ANC)<br />

(http://www.Amer-icanNationalCorpus.org), as well as (<strong>to</strong> some extent) the corpora for<br />

several western European languages produced by the PAROLE project. In fact the<br />

greatest number of existing text corpora are composed of readily available materials such<br />

as newspaper data, technical manuals, government documents, and, more recently,<br />

materials drawn from the World Wide Web. Speech data, whose acquisition is in most<br />

instances necessarily controlled, are more often representative of a specific dialect or<br />

range of dialects.<br />

Many corpora are available for research purposes by signing a license and paying a small<br />

reproduction fee. Other corpora are available only by paying a (sometimes substantial)<br />

fee; this is the case, for instance, for many of the holdings of the LDC, making them<br />

virtually inaccessible <strong>to</strong> humanists.<br />

Preparation of Linguistic Corpora<br />

The first phase of corpus creation is data capture, which involves rendering the text in<br />

electronic form, either by hand or via optical character recognition (OCR), acquisition of<br />

word processor or publishing software output, typesetter tapes, PDF files, etc. Manual<br />

entry is time-consuming and costly, and therefore unsuitable for the creation of very large<br />

corpora. OCR output can be similarly costly if it requires substantial post-processing <strong>to</strong><br />

validate the data. Data acquired in electronic form from other sources will almost<br />

invariably contain formatting codes and other information that must be discarded or<br />

translated <strong>to</strong> a representation that is processable for linguistic analysis.<br />

Representation formats and surrounding issues<br />

At this time, the most common representation format for linguistic corpora is XML.<br />

Several existing corpora are tagged using the EAGLES XML Corpus Encoding Standard<br />

(XCES) (Ide et al. 2000), a Text Encoding Initiative (TEI) – compliant XML application<br />

designed specifically for linguistic corpora and their annotations. The XCES introduced<br />

the notion of stand-off annotation, which requires that annotations are encoded in<br />

documents separate from the primary data and linked <strong>to</strong> them. One of the primary<br />

motivations for this approach is <strong>to</strong> avoid the difficulties of overlapping hierarchies, which<br />

are common when annotating diverse linguistic features, as well as the unwieldy<br />

documents that can be produced when multiple annotations are associated with a single<br />

document. The stand-off approach also allows for annotation of the same feature (e.g.,<br />

part of speech) using alternative schemes, as well as associating annotations with other<br />

annotations rather than directly <strong>to</strong> the data. Finally, it supports two basic notions about<br />

text and annotations outlined in Leech (1993): it should be possible <strong>to</strong> remove the


annotation from an annotated corpus in order <strong>to</strong> revert <strong>to</strong> the raw corpus; and, conversely,<br />

it should be possible <strong>to</strong> extract the annotations by themselves from the text.<br />

The use of stand-off annotation is now widely accepted as the norm among corpus and<br />

corpus-handling software developers; however, because mechanisms for inter-document<br />

linking have only recently been developed within the XML framework, many existing<br />

corpora include annotations in the same document as the text.<br />

The use of the stand-off model dictates that a distinction is made between the primary<br />

data (i.e., the text without additional linguistic information) and their annotations, in<br />

particular, what should and should not be marked in the former. The XCES identifies two<br />

types of information that may be encoded in the primary data:<br />

1 Gross structure: universal text elements down <strong>to</strong> the level of the paragraph, which is<br />

the smallest unit that can be identified language-independently; for example:<br />

• structural units of text, such as volume, chapter, etc., down <strong>to</strong> the level of paragraph;<br />

also footnotes, titles, headings, tables, figures, etc.;<br />

• features of typography and layout, for previously printed texts: e.g., list item markers;<br />

• non-textual information (graphics, etc.).<br />

2 Segmental structure: elements appearing at the sub-paragraph level which are usually<br />

signaled (sometimes ambiguously) by typography in the text and which are languagedependent;<br />

for example:<br />

• orthographic sentences, quotations;<br />

• orthographic words;<br />

• abbreviations, names, dates, highlighted words.<br />

Annotations (see next section) are linked <strong>to</strong> the primary data using XML conventions<br />

(XLink, Xpointer).<br />

Speech data, especially speech signals, are often treated as "read-only", and therefore the<br />

primary data contain no XML markup <strong>to</strong> which annotations may be linked. In this case,<br />

stand-off documents identify the start and end points (typically using byte offsets) of the<br />

structures listed above, and annotations are linked indirectly <strong>to</strong> the primary data by<br />

referencing the structures in these documents. The annotation graphs representation<br />

format used in the ATLAS project, which is intended primarily <strong>to</strong> handle speech data,<br />

relies entirely on this approach <strong>to</strong> link annotations <strong>to</strong> data, with no option for referencing<br />

XML-tagged elements.


Identification of segmental structures<br />

Markup identifying the boundaries of gross structures may be au<strong>to</strong>matically generated<br />

from original formatting information. However, in most cases the original formatting is<br />

presentational rather than descriptive; for example, titles may be identifiable because<br />

they are in bold, and therefore transduction <strong>to</strong> a descriptive XML representation may not<br />

be straightforward. This is especially true for sub-paragraph elements that are in italic or<br />

bold font; it is usually impossible <strong>to</strong> au<strong>to</strong>matically tag such elements as emphasis, foreign<br />

word, etc.<br />

Creation of linguistic corpora almost always demands that sub-paragraph structures such<br />

as sentences and words, as well as names, dates, abbreviations, etc., are identified.<br />

Numerous programs have been developed <strong>to</strong> perform sentence "splitting" and word<br />

<strong>to</strong>kenization, many of which are freely available (see, for example, the <strong>to</strong>ols listed in the<br />

Natural Language Software Registry (http://www.dfki.de/lt/registry/) or the SIL Software<br />

Catalogue (http://www.sil.org). These functions are also embedded in more general<br />

corpus development <strong>to</strong>ols such as GATE (Cunningham 2002). Sentence splitting and<br />

<strong>to</strong>kenization are highly language-dependent and therefore require specific information<br />

(e.g., abbreviations for sentence splitting, clitics and punctuation conventions for<br />

<strong>to</strong>kenization) for the language being processed; in some cases, language-specific software<br />

is developed, while in others a general processing engine is fed the language-specific<br />

information as data and can thus handle multiple languages. Languages without wordboundary<br />

markers, such as Chinese and Japanese, and continuous speech represented by<br />

phoneme sequences, require an entirely different approach <strong>to</strong> segmentation, the most<br />

common of which is a dynamic programming algorithm <strong>to</strong> compute the most likely<br />

boundaries from a weighted transition graph. This, of course, demands that the<br />

probabilities of possible symbol or phoneme sequences are available in order <strong>to</strong> create the<br />

weighted graph.<br />

Within the computational linguistics community, software <strong>to</strong> identify so-called "named<br />

entities" (proper names designating people, places, organizations, events, documents,<br />

etc.), as well as dates and other time expressions, has been developed, and many of these<br />

<strong>to</strong>ols are freely available for research. However, most of these <strong>to</strong>ols have been developed<br />

using existing corpora which consist of newspapers and government reports and are<br />

unlikely <strong>to</strong> perform well on the kinds of data that interest humanists, such as literary<br />

works, his<strong>to</strong>rical documents, etc. This is just one example of the broader situation in the<br />

development of corpus-handling <strong>to</strong>ols: the available data, which are often highly skewed<br />

for genre, drive <strong>to</strong>ol, and algorithm development, and therefore applicability <strong>to</strong> more<br />

generalized corpora is often limited.<br />

Corpus annotation<br />

For computational linguistics research, which has driven the bulk of corpus creation<br />

efforts over the past decade, corpora are typically annotated with various kinds of<br />

linguistic information. The following sections outline the major annotation types.


Morpho-syntactic annotation<br />

By far the most common corpus annotation is morpho-syntactic annotation (part-ofspeech<br />

tagging), primarily because several highly accurate au<strong>to</strong>matic taggers have been<br />

developed over the past fifteen years. Part-of-speech tagging is a disambiguation task: for<br />

words that have more than one possible part of speech, it is necessary <strong>to</strong> determine which<br />

one, given the context, is correct. Although, in English, close <strong>to</strong> 90 percent of the words<br />

in the language have only one part of speech, in actual use (e.g., in a corpus), part of<br />

speech for as many as 40 percent of the word instances may be ambiguous (DeRose<br />

1988), due in large part <strong>to</strong> ambiguity for a handful of high-frequency words such as<br />

"that", which can be either a determiner ("that girl") or a complementizer ("He heard that<br />

you came home"). Beyond this, the most common ambiguity in English is between verb<br />

and noun, e.g., "hide", "dog", "brand", "felt", etc.<br />

Taggers fall in<strong>to</strong> two general classes: rule-based (e.g. ENGTWOL, Voutilainen, 1995),<br />

which use manually generated rules <strong>to</strong> assign part of speech, and s<strong>to</strong>chastic, which rely<br />

on probabilities for n-grams – sequences of n (usually, 2 or 3) tags that are known <strong>to</strong><br />

occur in real data. S<strong>to</strong>chastic taggers learn these probabilities by being "trained" on<br />

previously tagged data that have been hand-validated for correctness. A third type of<br />

tagger, often called the "Brill tagger" after its developer (Brill 1995), uses a hybrid<br />

approach which learns its tagging rules from a previously tagged training corpus.<br />

Obviously, the more accurately tagged data a tagger can use for training, the more likely<br />

its probabilities will be correct. This fact has led <strong>to</strong> the creation of corpora that have been<br />

hand-annotated (or in which au<strong>to</strong>matically produced tags have been hand-validated)<br />

specifically intended for training, in order <strong>to</strong> enable au<strong>to</strong>matic taggers <strong>to</strong> produce even<br />

more tagged corpora.<br />

The tags generated by a part-of-speech tagger provide more information than simple<br />

word class (noun, verb, adjective, etc.). Various levels of morpho-syntactic information<br />

can be represented; the more detailed the information, the bigger the tagset, and the<br />

bigger the tagset, the less accurate an au<strong>to</strong>matic tagger will be. For this reason, tagsets<br />

including 50–100 tags that collapse or eliminate detailed morpho-syntactic information –<br />

such as the information in the morpho-syntactic specifications for western and eastern<br />

European languages produced by the EAGLES project<br />

(http://www.ilc.pi.cnr.it/EAGLES/home.html) – are most typically used in au<strong>to</strong>matic<br />

tagging.<br />

There are several tagsets in common use for English, most of which evolved from the 87<br />

tags used in the Brown corpus (http://www.hit.uib.no/icame/brown/bcm.html). Probably<br />

the most widely used is the 45-tag set of the Penn Treebank project (Marcus et al. 1993),<br />

of which only 36 are actual morpho-syntactic categories (the rest are for punctuation, list<br />

markers, etc.). The Penn tagset is a variant of the Brown tagset that eliminates<br />

information retrievable from the form of the lexical item. It therefore includes only one<br />

tag for different forms of the verbs "be", "have", and "do", whereas the Brown (and other<br />

common tagsets for English) provide a different tag for each of these forms. Another<br />

well-known tagset for English is the 61-tag C5 tagset of the CLAWS (Constituent


Likelihood Au<strong>to</strong>matic Word-tagging System) tagger, developed at the University of<br />

Lancaster (Garside and Smith 1997) and used <strong>to</strong> tag the British National Corpus.<br />

Part-of-speech taggers rely on lexicons that provide all of the possible part of speech<br />

assignments for a lexical item found in the input, from which it must choose the most<br />

probable given the immediate context. The information in the lexicons must therefore<br />

match – or at least, be mappable <strong>to</strong> – the tagset used by the tagger. It is an unfortunate<br />

fact that it is often extremely difficult and sometimes impossible <strong>to</strong> map one tagset <strong>to</strong><br />

another, which has resulted in much re-creation of lexical information <strong>to</strong> suit the needs of<br />

a particular tagger.<br />

Many lexicons include lemmas (root forms), and morpho-syntactic annotation may<br />

produce lemmas as well as part-of-speech tags in their output. The presence of lemmas in<br />

an annotated text enables extraction of all orthographic forms associated with a given<br />

lemma (e.g., "do", "does", "doing, "did" for the lemma "do"). However, although<br />

relatively easily produced, many existing corpora do not include lemmas; notable<br />

exceptions are the Journal of the Commission corpus, the Orwell multilingual corpus, and<br />

the SUSANNE corpus.<br />

Parallel alignment<br />

In addition <strong>to</strong> algorithms for morpho-syntactic annotation, reliable algorithms for<br />

alignment of parallel texts – i.e., texts for which there exist translations in two or more<br />

languages – have been developed. Probability information about correspondences<br />

between words, phrases, and other structures derived from aligned corpora are used <strong>to</strong><br />

choose from multiple possible translations that may be generated by a machine<br />

translation system. Parallel corpora have also been used <strong>to</strong> au<strong>to</strong>matically generate<br />

bilingual dictionaries (e.g., Dagan and Church 1997) and, more recently, as a means <strong>to</strong><br />

achieve au<strong>to</strong>matic sense tagging of corpora (e.g., Ide et al. <strong>2001</strong>).<br />

Two types of parallel alignment are common: sentence alignment and word alignment.<br />

Sentence alignment is by far the easier and more accurate of the two, the major problem<br />

being <strong>to</strong> determine cases in which one-<strong>to</strong>-many or partial mappings exist. Many sentence<br />

and word-aligned corpora exist, the vast majority of which include only two languages.<br />

Probably the best known is the English-French Hansard Corpus of Canadian<br />

Parliamentary debates, which has served as the basis for numerous translation studies.<br />

Multilingual parallel corpora are much more rare; the difficulty is not in the alignment<br />

itself, but rather in the availability of texts in multiple translations (in particular, texts that<br />

are not bound by copyright or other restrictions). Existing multilingual aligned corpora<br />

include the United Nations Parallel Text Corpus (English, French, Spanish), the Journal<br />

of the Commission (JOC) corpus (English, French, German, Italian, Spanish), the Orwell<br />

1984 Corpus (Bulgarian, Czech, English, Es<strong>to</strong>nian, Hungarian, Latvian, Lithuanian,<br />

Romanian, Serbo-Croat, Slovene), Pla<strong>to</strong>'s Republic (Bulgarian, Chinese, Czech, English,<br />

German, Latvian, Polish, Romanian, Slovak, Slovene) and the Bible (Chinese, Danish,<br />

English, Finnish, French, Greek, Indonesian, Latin, Spanish, Swahili, Swedish,<br />

Vietnamese).


Syntatic annotation<br />

There are two main types of syntactic annotation in linguistic corpora: noun phrase (NP)<br />

bracketing or "chunking", and the creation of "treebanks" that include fuller syntactic<br />

analysis. Syntactically annotated corpora serve various statistics-based applications, most<br />

notably, by providing probabilities <strong>to</strong> drive syntactic parsers, and have been also used <strong>to</strong><br />

derive context-free and unification-based grammars (Charniak 1996; Van Genabith et al.<br />

1999). Syntactically annotated corpora also provide theoretical linguists with data <strong>to</strong><br />

support studies of language use.<br />

The best known and most widely used treebank is the Penn Treebank for English (Marcus<br />

et al. 1993). Ongoing projects exist for the development of treebanks for other languages,<br />

including German (the NEGRA corpus, Brants et al. 2003; Stegman et al. 1998), Czech<br />

(The Prague Dependency Treebank, Hajic 1998), French (Abeille et al. 2000), and<br />

Chinese (Xue et al. 2002). The Penn Treebank and the Chinese Treebank, both created at<br />

the University of Pennsylvania, use LISP-like list structures <strong>to</strong> specify constituency<br />

relations and provide syntactic category labels for constituents, as shown below:<br />

Although they differ in the labels and in some cases<br />

the function of various nodes in the tree, many treebank annotation schemes provide a<br />

similar constituency-based representation of relations among syntactic components. In<br />

contrast, dependency schemes do not provide a hierarchical constituency analysis, but<br />

rather specify grammatical relations among elements explicitly; for example, the sentence<br />

"Paul intends <strong>to</strong> leave IBM" could be represented as follows:<br />

Here, the predicate is the relation type, the first<br />

argument is the head, the second the dependent, and additional arguments may provide<br />

category-specific information (e.g., introducer for prepositional phrases). Finally, socalled<br />

"hybrid systems" combine constituency analysis and functional dependencies,<br />

usually producing a shallow constituent parse that brackets major phrase types and<br />

identifies the dependencies between heads of constituents (e.g., the NEGRA Corpus).


Although most modern treebanks utilize an SGML or XML encoding rather than list<br />

structures, syntactic annotation is invariably interspersed with the text itself. This makes<br />

it difficult or impossible <strong>to</strong> add other kinds of annotations <strong>to</strong> the data, or <strong>to</strong> provide<br />

alternative syntactic annotations. As noted earlier, the use of stand-off annotations is<br />

increasingly encouraged. A stand-off scheme for syntactic annotations, which also serves<br />

as a. "pivot" format for representing different types of syntactic annotation with a<br />

common scheme, has been developed by Ide and Romary (<strong>2001</strong>).<br />

Semantic annotation<br />

Semantic annotation can be taken <strong>to</strong> mean any kind of annotation that adds information<br />

about the meaning of elements in a text. Some annotations that may be considered <strong>to</strong><br />

provide semantic information – for example, "case role" information such as agent,<br />

instrument, etc. – are often included in syntactic annotations. Another type of semantic<br />

annotation common in literary analysis (especially in the 1960s and 1970s) marks words<br />

or phrases in a text as representative of a particular theme or concept. At present, the<br />

most common type of semantic annotation is "sense tagging": the association of lexical<br />

items in a text with a particular sense or definition, usually drawn from an existing sense<br />

inven<strong>to</strong>ry provided in a dictionary or online lexicon such as WordNet (Miller et al. 1990).<br />

The major difficulty in sense annotation is <strong>to</strong> determine an appropriate set of senses.<br />

Simply examining the differences in sense distinctions made from one dictionary <strong>to</strong><br />

another demonstrates the difficulty of this task. To solve the problem, some attempts<br />

have been made <strong>to</strong> identify the kinds of sense distinctions that are useful for au<strong>to</strong>matic<br />

language processing: for example, tasks such as information retrieval may require only<br />

very coarse-grained sense distinctions – the difference between "bank" as a financial<br />

institution vs. a river bank – whereas others, and in particular machine translation, require<br />

finer distinctions – say, between "bank" as a financial institution and as a building.<br />

However, by far the most common source of sense tags used for semantic annotation is<br />

WordNet, an online dictionary that in addition <strong>to</strong> providing sense lists, groups words in<strong>to</strong><br />

"synsets" of synonymous words. WordNet has been updated several times, and as a<br />

result, its sense list may differ for a particular word depending on the WordNet version<br />

used for the tagging.<br />

It is widely acknowledged that the sense distinctions provided in WordNet are far from<br />

optimal. However, this resource, which is among the most heavily used in natural<br />

language processing research over the past decade, will likely continue <strong>to</strong> serve as a basis<br />

for sense tagging for at least the foreseeable future, if for no other reason than that it<br />

continues <strong>to</strong> be the only freely available, machine-tractable lexicon providing extensive<br />

coverage of English. The Euro WordNet project has produced WordNets for most<br />

western European languages linked <strong>to</strong> WordNet 1.5 (the current version of the English<br />

WordNet is 2.0), and WordNets for other languages (e.g., Balkan languages) are under<br />

development, which is likely <strong>to</strong> extend the research community's reliance on its sense<br />

inven<strong>to</strong>ry. In any case, no clearly superior alternative source of sense distinctions has<br />

been proposed.


Because sense-tagging requires hand annotation, and because human annota<strong>to</strong>rs will<br />

often disagree on sense assignments even given a predefined sense inven<strong>to</strong>ry, very few<br />

sense-tagged corpora exist. Examples include the Semantic Concordance Corpus<br />

(SemCor; Miller et al. 1993), produced by the WordNet project, which assigns WordNet<br />

sense tags for all nouns, verbs, and adjectives in a 250,000-word corpus drawn primarily<br />

from the Brown Corpus; the DSO Corpus, containing sense-tags for 121 nouns and 70<br />

verbs in about 192,800 sentences taken from the Brown Corpus and the Wall Street<br />

Journal; and Hec<strong>to</strong>r, containing about 200,000 tagged instances of 300 words in a corpus<br />

of British English.<br />

Au<strong>to</strong>matic means <strong>to</strong> sense-tag data have been sought ever since machine-readable texts<br />

became available. This area of research, called "word sense disambiguation", remains <strong>to</strong><br />

this day one of the more difficult problems in language processing. Although rule-based<br />

approaches have been developed, the most common approaches <strong>to</strong> word sense<br />

disambiguation over the past decade are statistics-based, relying on the frequency with<br />

which lexical items (or categories of lexical items) in the context under consideration<br />

have been found in the context of a word in a given sense. Some recent research (Ide<br />

2000; Ide et al. <strong>2001</strong>; Doab and Resnik 2002) has explored the use of information<br />

gleaned from parallel translations <strong>to</strong> make sense distinctions. For a comprehensive<br />

overview of approaches <strong>to</strong> word sense disambiguation see Ide and Veronis (1998).<br />

Discourse-level annotation<br />

There are three main types of annotation at the level of discourse: <strong>to</strong>pic identification, coreference<br />

annotation, and discourse structure.<br />

Topic ientification<br />

(also called "<strong>to</strong>pic detection") annotates texts with information about the events or<br />

activities described in the text. Au<strong>to</strong>matic means for <strong>to</strong>pic detection over streams of data<br />

such as broadcast news and newswires is under development, primarily in the context of<br />

the DARPA-sponsored Topic Detection Task. A subtask of this kind of annotation is<br />

detection of boundaries between s<strong>to</strong>ries/texts, which may also be included in the<br />

annotation.<br />

Co-reference annotation<br />

Links referring objects (e.g., pronouns, definite noun phrases) <strong>to</strong> prior elements in a<br />

discourse <strong>to</strong> which they refer. This type of annotation is invariably performed manually,<br />

since reliable software <strong>to</strong> identify co-referents is not available. For example, the MUG 7<br />

(Message Understanding Conference) corpus contains hand-generated co-reference links<br />

(Hirschman and Chinchor 1997) marked in a non-stand-off SGML format similar <strong>to</strong> the


following:<br />

Discourse structure annotation<br />

Identifies multi-level hierarchies of discourse segments and the relations between them,<br />

building upon low-level analysis of clauses, phrases, or sentences. There are several<br />

theoretical approaches <strong>to</strong> the analysis of discourse structure, differing in terms of the span<br />

of text (clause, sentence, etc.) that is considered <strong>to</strong> be the a<strong>to</strong>mic unit of analysis, and in<br />

the relations defined <strong>to</strong> exist between units and higher-level structures built from them.<br />

Most common approaches are based on "focus spaces" (Grosz and Sidner 1986) or<br />

Rhe<strong>to</strong>rical Structure Theory (Mann and Thompson 1988). To date, annotation of<br />

discourse structure is almost always accomplished by hand, although some software <strong>to</strong><br />

perform discourse segmentation has been developed (e.g., Marcu 1996).<br />

Annotation of speech and spoken data<br />

The most common type of speech annotation is an orthographic transcription that is timealigned<br />

<strong>to</strong> an audio or video recording. Often, annotation demarcating the boundaries of<br />

speech "turns" and individual "utterances" is included. Examples of speech corpora in<br />

this type of format include the Child Language Data Exchange System (CHILDES)<br />

corpus (MacWhinney 1995), the TIMIT corpus of read speech (Lamel et al. 1990), and<br />

the LACITO Linguistic Data Archiving Project (Jacobson and Michailovsky 2000).<br />

The orthographic transcription of speech data can be annotated with any of the kinds of<br />

linguistic information outlined in previous sections (e.g., part of speech, syntax, coreference,<br />

etc.), although this is not common. In addition, speech data may be annotated<br />

with phonetic segmentation and labeling (phonetic transcription), prosodic phrasing and<br />

in<strong>to</strong>nation, disfluencies, and, in the case of video, gesture.<br />

Most types of annotation particular <strong>to</strong> speech data are time-consuming <strong>to</strong> produce<br />

because they must be generated by hand, and because the annota<strong>to</strong>rs must be skilled in<br />

the recognition and transcription of speech sounds. Furthermore, speech annotation is<br />

problematic: for example, phonetic transcription works on the assumption that the speech<br />

signal can be divided in<strong>to</strong> single, clearly demarcated sounds, but these demarcations are<br />

often rather unclear. Prosodic annotation is even more subjective, since the decision<br />

about the exact nature of a <strong>to</strong>ne movement often varies from annota<strong>to</strong>r <strong>to</strong> annota<strong>to</strong>r. The<br />

kinds of phenomena that are noted include onset, rising <strong>to</strong>ne, falling <strong>to</strong>ne, rising/falling<br />

<strong>to</strong>ne, level <strong>to</strong>ne, pause, overlap, etc. The notations for prosodic annotation vary widely


and they are not typically rendered in a standard format such as XML. As a result, the<br />

few existing corpora annotated for prosody are widely inconsistent in their format. One of<br />

the best known such corpora is the London-Lund Corpus of Spoken English (Svartvik<br />

1990).<br />

Corpus annotation <strong>to</strong>ols<br />

Over the past decade, several projects have created <strong>to</strong>ols <strong>to</strong> facilitate the annotation of<br />

linguistic corpora. Most are based on a common architectural model introduced in the<br />

MULTEXT project, which views the annotation process as a chain of smaller, individual<br />

processes that incrementally add annotations <strong>to</strong> the data. A similar model was developed<br />

in the TIPSTER project.<br />

Among the existing annotation <strong>to</strong>ols for language data are LT XML (University of<br />

Edinburgh), which directly implements the MULTEXT model and is especially geared<br />

<strong>to</strong>ward XML-encoded resources; GATE (University of Sheffield), based on the TIPSTER<br />

model and including <strong>to</strong>ols for <strong>to</strong>kenization, sentence splitting, named entity recognition,<br />

part of speech tagging, as well as corpus and annotation editing <strong>to</strong>ols. The Multilevel<br />

Annotation, Tools Engineering (MATE) project provides an annotation <strong>to</strong>ol suite<br />

designed especially for spoken dialogue corpora at multiple levels, focusing on prosody,<br />

(morpho-)syntax, co-reference, dialogue acts, and communicative difficulties, as well as<br />

inter-level interaction. ATLAS (Architecture and Tools for Linguistic Analysis Systems)<br />

is a joint initiative of the US National Institute for Standards and Technology (NIST),<br />

MITRE, and the LDC <strong>to</strong> build a general-purpose annotation architecture and a data<br />

interchange format. The starting point in ATLAS is the annotation graph model, with<br />

some significant generalizations.<br />

Currently, a subcommittee of the International Standards Organization (ISO) – ISO TC37<br />

SC4 – is developing a generalized model for linguistic annotations and processing <strong>to</strong>ols,<br />

based on input from developers of the annotation <strong>to</strong>ol suites mentioned above as well as<br />

from annotation scheme designers. The goal is <strong>to</strong> provide a common "pivot" format that<br />

instantiates a generalized data model of linguistic resources and their annotations <strong>to</strong> and<br />

from which existing formats – provided they are consistent with the model – can be<br />

mapped in order <strong>to</strong> enable seamless interchange. At present, data annotated within one<br />

project, using XML, annotation graphs, or a proprietary annotation scheme, are typically<br />

difficult <strong>to</strong> import for manipulation using another <strong>to</strong>ol suite, either for further annotation<br />

or analysis. The developing ISO model is intended <strong>to</strong> allow annota<strong>to</strong>rs <strong>to</strong> use any of a<br />

variety of schemes <strong>to</strong> represent their data and annotations, and map it <strong>to</strong> a common<br />

format for interchange. Users of other schemes would then import the data in the<br />

common format <strong>to</strong> their own schemes. This means, for example, that a corpus marked up<br />

using TEI or XCES conventions can be mapped <strong>to</strong> the common format and, from that,<br />

mapped again <strong>to</strong>, say, an annotation graph representation that will enable manipulation of<br />

the data by <strong>to</strong>ols implementing that model, without information loss.


The future of corpus annotation<br />

Recent developments in the XML world, primarily in the scope of work within the World<br />

Wide Web Consortium (W3C), have focused attention on the potential <strong>to</strong> build a<br />

Semantic Web. This possibility has interesting ramifications for both corpus annotation<br />

and analysis, in two (related) ways. First, the underlying technology of the Semantic Web<br />

enables definition of the kinds of relationships (links) one resource – where a "resource"<br />

can be any fragment of a document or the document as a whole – may have with another.<br />

For example, a "word" may have a link labeled "part of speech" with another resource<br />

that represents (possibly, as a simple string) "noun" or "noun singular masculine."<br />

Because annotation is, at the base, the specification of relationships between information<br />

in a corpus and linguistic information that describes it, the development of technologies<br />

such as the Resource Description Framework (RDF) can have significant impact on the<br />

way annotations are associated with primary language resources in the future.<br />

A second activity within the Semantic Web community that has ramifications for both<br />

annotation and analysis of linguistic data is the development of technologies <strong>to</strong> support<br />

the specification of and access <strong>to</strong> on<strong>to</strong>logical information. On<strong>to</strong>logies, which provide a<br />

priori information about relationships among categories of data, enable the possibility <strong>to</strong><br />

apply inferencing processes that can yield information that is not explicit in the data<br />

itself. For example, <strong>to</strong> properly parse a sentence such as "I ate a fish with a fork" – that is,<br />

<strong>to</strong> attach the prepositional phrase "with a fork" <strong>to</strong> the verb "eat" and not as a modifier of<br />

"fish" – we can check an on<strong>to</strong>logy that specifies that "fork" IS-A sub-class of<br />

"instrument", and that "eat" has a USES-A relation with things of type "instrument."<br />

More generally, we may be able <strong>to</strong> identify the <strong>to</strong>pic of a given document by consulting<br />

an on<strong>to</strong>logy. For example, the on<strong>to</strong>logy may specify that the word "bank" can represent a<br />

sub-class of "financial institution" or "geological formation"; if the document contains<br />

several other words that are also related <strong>to</strong> "financial institution" such as "money",<br />

"account", etc., we may conclude that the document has financial institutions as one of its<br />

<strong>to</strong>pics (and, as a side effect, that the word "bank" is being used in that sense in this<br />

document).<br />

In order <strong>to</strong> be able <strong>to</strong> exploit on<strong>to</strong>logical information, it must be created and s<strong>to</strong>red in a<br />

universally accessible form. The W3C group developing the On<strong>to</strong>logical Web Language<br />

(OWL) is providing a standard representation format. It is now up <strong>to</strong> the computational<br />

linguistics community <strong>to</strong> instantiate the relevant on<strong>to</strong>logical information and use it <strong>to</strong><br />

annotate and analyze language data. Development and use of on<strong>to</strong>logies has in fact been<br />

a part of the field for several years – in particular, the on<strong>to</strong>logy in the WordNet lexicon<br />

has been widely used for annotation and analysis of language data. Semantic Web<br />

technologies will enable development of common and universally accessible on<strong>to</strong>logical<br />

information.<br />

Analysis of Linguistic Corpora


There are two major uses for linguistic corpora: statistics-gathering <strong>to</strong> support natural<br />

language processing; and language analysis <strong>to</strong> support language learning and dictionary<br />

creation.<br />

Statistics gathering<br />

A corpus provides a bank of samples that enable the development of numerical language<br />

models, and thus the use of corpora goes hand in hand with empirical methods. In the late<br />

1980s, the increased availability of large amounts of electronic text enabled, for the first<br />

time, the full-scale use of data-driven methods <strong>to</strong> attack generic problems in<br />

computational linguistics, such as part-of-speech identification, prepositional phrase<br />

attachment, parallel text alignment, word sense disambiguation, etc. The success in<br />

treating at least some of these problems with statistical methods led <strong>to</strong> their application <strong>to</strong><br />

others, and by the mid-1990s, statistical methods had become a staple of computational<br />

linguistics work.<br />

The key element for many statistics-based approaches <strong>to</strong> language processing (in<br />

particular, so-called "supervised" methods) is the availability of large, annotated corpora,<br />

upon which annotation algorithms are trained <strong>to</strong> identify common patterns and create<br />

transformations or rules for them. The s<strong>to</strong>chastic part-of-speech taggers described above<br />

are probably the best-known application relying on previously annotated corpora, but<br />

similar approaches are also used for word sense disambiguation, probabilistic parsing,<br />

and speech recognition. In word sense disambiguation, for example, statistics are<br />

gathered reflecting the degree <strong>to</strong> which other words are likely <strong>to</strong> appear in the context of<br />

some previously sense-tagged word in a corpus. These statistics are then used <strong>to</strong><br />

disambiguate occurrences of that word in untagged corpora, by computing the overlap<br />

between the unseen context and the context in which the word was seen in a known<br />

sense.<br />

Speech recognition is, in fact, the area of computational linguistics in which corpora were<br />

first put in<strong>to</strong> large-scale use in support of language processing, beginning with the<br />

utilization of Hidden Markov Models (HMMs) in the 1970s. This paradigm required data<br />

<strong>to</strong> statistically train an acoustic model <strong>to</strong> capture typical sound sequences and a language<br />

model <strong>to</strong> capture typical word sequences, and produced results that were far more<br />

accurate and robust than the traditional methods. It was not until the late 1980s that the<br />

statistical approach was applied <strong>to</strong> other areas, one of the first of which was machine<br />

translation. Following the same approach as the speech recognition systems, researchers<br />

au<strong>to</strong>matically trained a French-English correspondence model (the Translation Model) on<br />

3 million sentences of parallel French and English from the Canadian Parliamentary<br />

records, and also trained a Language Model for English production from Wall Street<br />

Journal data. To translate, the former model was used <strong>to</strong> replace French words or phrases<br />

by the most likely English equivalents, and then the latter model ordered the English<br />

words and phrases in<strong>to</strong> the most likely sequences <strong>to</strong> form output sentences.<br />

Probabilistic parsing is one of the more recent applications of statistical methods <strong>to</strong><br />

language processing tasks. Again, large bodies of data previously annotated and validated


for syntactic structure are required in order <strong>to</strong> provide statistics concerning the probability<br />

that a given syntactic construction is the correct one in its context. Syntactic structure can<br />

be very ambiguous; traditional parsers often produced numerous alternative structural<br />

analyses for an input sentence. A probabilistic parser uses previously gathered statistics<br />

<strong>to</strong> choose the most probable interpretation.<br />

Issues for corpus-based statistics gathering<br />

In order <strong>to</strong> be representative of any language as a whole, it is necessary that a corpus<br />

include samples from a variety of texts that reflect the range of syntactic and semantic<br />

phenomena across that language. This demands, first of all, that the data be adequately<br />

large in order <strong>to</strong> avoid the problem of data sparseness that plagues many statistical<br />

approaches. For example, for tasks such as word sense disambiguation, data must be<br />

extensive enough <strong>to</strong> ensure that all senses of a polysemous word are not only represented,<br />

but represented frequently enough so that meaningful statistics can be compiled.<br />

Although it has been extensively used for natural language processing work, the million<br />

words of a corpus such as the Brown Corpus are not sufficient for <strong>to</strong>day's large-scale<br />

applications: many word senses are not represented; many syntactic structures occur <strong>to</strong>o<br />

infrequently <strong>to</strong> be significant, and the corpus is far <strong>to</strong>o small <strong>to</strong> be used for computing the<br />

bi-gram and tri-gram probabilities that are necessary for training language models for<br />

speech recognition.<br />

Unfortunately, in the main, the large corpora freely available for research consist of texts<br />

that can be easily acquired and are available for redistribution without undue problems of<br />

copyright, etc. Because of this, corpora used for statistics gathering for language<br />

processing are vastly over-representative of certain genres, in particular newspaper<br />

samples, which constitute the greatest percentage of texts currently available from, for<br />

example, the LDC, and which also dominate the training data available for speech<br />

recognition purposes. Other available corpora typically consist of technical reports,<br />

transcriptions of parliamentary and other proceedings, short telephone conversations, and<br />

the like. The upshot of this is that corpus-based natural language processing has relied<br />

heavily on language samples representative of usage in a handful of limited and<br />

linguistically specialized domains. This can lead <strong>to</strong> drastically skewed results: for<br />

example, in newspaper data, there is a disproportionate number of complex NP<br />

complements for some verbs, which appear in sentences typical of newspaper style, such<br />

as "The price rose two percent <strong>to</strong> 102 dollars per share from 100 dollars per share."<br />

Similar problems arise in work on word sense disambiguation: it has been noted that for<br />

some typical test words such as "line", certain senses (for example, the common sense of<br />

"line" as in the sentence, "He really handed her a line") are absent entirely from resources<br />

such as the Wall Street Journal.<br />

The problem of balance is acute in speech recognition. Speech recognition systems are<br />

no<strong>to</strong>riously dependent on the characteristics of their training corpora. Corpora large<br />

enough <strong>to</strong> train the tri-gram language models of modern speech recognizers (many tens<br />

of millions of words) are invariably composed of written rather than spoken texts. But the<br />

differences between written and spoken language are even more severe than the


differences between balanced corpora like the Brown and newspaper corpora like the<br />

Wall Street Journal. Therefore, whenever a state-of-the-art speech recognition research<br />

effort moves <strong>to</strong> a new domain, a new large training corpus of speech must be collected,<br />

transcribed at the word level, and the transcription aligned <strong>to</strong> the speech.<br />

Language analysis<br />

The gathering of authentic language data from corpora enables a description of language<br />

that starts from the evidence rather than from imposing some theoretical model. Because<br />

speakers and writers produce language with real communicative goals, corpora of nativespeaker<br />

texts provide, in principle, samples of genuine language. For this reason, one of<br />

the oldest uses of corpora is for dictionary-making, or lexicography, and in particular,<br />

lexicography with the goal of producing so-called "learners' dictionaries" designed for<br />

those learning a new language.<br />

The COBUILD corpus was compiled starting in the early 1980s, at which time it<br />

included about 7 million words. This corpus was used <strong>to</strong> create the Collins COBUILD<br />

English Dictionary, the first dictionary relying fully on corpora for its creation. Following<br />

this lead, over the course of the next decade most British dictionary publishers began <strong>to</strong><br />

use corpora as the primary data source for their dictionaries, although interestingly,<br />

American dictionary publishers are only now beginning <strong>to</strong> rely on corpora <strong>to</strong> guide<br />

lexicography.<br />

The basic lexicographic <strong>to</strong>ol for analyzing a corpus is a concordancer, a program that<br />

displays occurrences of a word in the middle of a line of context from the corpus.<br />

However, the huge increase in available data has led <strong>to</strong> a situation where lexicographers<br />

are presented with hundreds or even thousands of concordance lines for a single word. As<br />

a result, lexicographers have begun <strong>to</strong> rely on techniques derived from computational<br />

linguistics <strong>to</strong> summarize concordance data. The most common is the "Mutual<br />

Information" (MI) score, a statistical measure that shows how closely one word is<br />

associated with others based on the regularity with which they co-occur in context. For<br />

example, in English, the mutual information score for the words "strong" and "tea" is<br />

much higher than the score for "powerful" and "tea", despite the fact that "strong" and<br />

"powerful" have similar meanings. This kind of information is invaluable for<br />

lexicographers, especially for the creation of language learners' dictionaries that must<br />

provide this kind of distinction.<br />

Most recently, dictionary makers have teamed with researchers in the field of<br />

computational linguistics <strong>to</strong> glean even more precise information from corpus data. For<br />

example, MI scores can show that "strong" collocates with northerly, showings, believer,<br />

currents, supporter, and odor, while "powerful" collocates with words such as <strong>to</strong>ol,<br />

minority, neighbor, symbol, figure, weapon, and post; but the MI score does not indicate<br />

the kinds of grammatical relations that exist among a word and its collocates.<br />

Supplemental grammatical information can provide even more precise understanding of<br />

word usage. For example, the word "tea" collocates with words like spoon, milk, and<br />

sugar when these words appear as the object of the preposition "with"; with mug, saucer,


and cup when they are objects of the preposition "in<strong>to</strong>"; with coffee, <strong>to</strong>ast, sugar, etc.,<br />

when joined by "and" or "or"; and "tea" is most commonly the object of verbs such as<br />

drink, sip, pour, finish, and make.<br />

To gather this information, relatively sophisticated language-processing software is<br />

required that can annotate the data for one or more of the types of information outlined in<br />

the previous sections. The need for more informative results <strong>to</strong> serve the needs of<br />

lexicography has therefore led <strong>to</strong> increased collaboration with computational linguists.<br />

We see, as a result, two groups of researchers who had previously worked independently<br />

coming <strong>to</strong>gether <strong>to</strong> tackle common problems.<br />

In general, researchers in the <strong>humanities</strong> and researchers in computational linguistics<br />

have not collaborated, despite their common problems and goals. With the advent of the<br />

World Wide Web, this should change soon. Humanists have increased access <strong>to</strong><br />

information about work in computational linguistics as well as <strong>to</strong> <strong>to</strong>ols and resources<br />

developed by that community. Computational linguists, on the other hand, are likely <strong>to</strong><br />

face new language processing challenges due <strong>to</strong> the need <strong>to</strong> handle a greater variety of<br />

web-accessible materials, including literary works, his<strong>to</strong>rical documents, and the like.<br />

The eventual collaboration of these two groups should lead, in the end, <strong>to</strong> vastly<br />

increased capabilities for both.<br />

References for Further Reading<br />

Abeillé, A., L. Clément, and A. Kinyon (2000). Building a Treebank for French.<br />

Proceedings of the Second International Conference on Language Resources and<br />

Evaluation: 87–94.<br />

Brants, T., W. Skut, and H. Uszkoreit (in press 2003). Syntactic Annotation of a German<br />

Newspaper Corpus. In A. Abeillé (ed.), Building and Using Syntactically Annotated<br />

Corpora. Dordrecht: Kluwer Academic.<br />

Brill, E. (1995). Transformation-based Error-driven Learning and Natural Language<br />

Processing: A Case Study in Part of Speech Tagging. Computational Linguistics 21:<br />

543–65.<br />

Charniak, E. (1996). Tree-bank Grammars. AAAI-96: Proceedings of the Thirteenth<br />

National Conference on Artificial Intelligence: 598–603.<br />

Cunningham, H. (2002). GATE, a General Architecture for Text Engineering. Computers<br />

and the Humanities 36: 223–54.<br />

Dagan, I. and K. Church (1997). Termight: Coordinating Humans and Machines in<br />

Bilingual Terminology Translation. MachineTranslation 12: 89–107.<br />

DeRose, S. (1988). Grammatical Category Disambiguation by Statistical Optimization.<br />

Computational Linguistics 14: 31–9.


Doab, M. and P. Resnik (2002). An Unsupervised Method for Word Sense Tagging Using<br />

Parallel Corpora. Proceedings of the 40th Annual Meeting of the Association for<br />

Computational Linguistics: 255–61.<br />

Garside, R. and Smith, N. (1997). A Hybrid Grammatical Tagger: CLAWS4. In R.<br />

Garside, G. Leech, and A. McEnery (eds.), Corpus Annotation: Linguistic Information<br />

from Computer Text Corpora (pp. 102–21). London: Longman.<br />

Grosz, B. and C. Sidner (1986). Attention, Intention and the Structure of Discourse.<br />

Computational Linguistics 12: 175–204.<br />

Hajič, J. (1998). Building a Syntactically Annotated Corpus: The Prague Dependency<br />

Treebank. In E. Hajicova (ed.), Issues of Valency and Meaning: Studies in Honour of<br />

jarmila Panevova. Prague: Charles University Press.<br />

Hirschman, L. and N. Chinchor (1997). MUC-7 Co-reference Task Definition Version<br />

3.0. At<br />

http://www.itl.nist.gov/iaui/894.02/related_projects/muc/proceedings/co_task.html.<br />

Ide, N. (2000). Cross-lingual Sense Determination: Can It Work? Computers in the<br />

Humanitie 34: 1–2. (Special Issue on the Proceedings of the SIGLEX/SENSEVAL<br />

Workshop, ed. A. Kilgarriff and M. Palmer: 223–34.).<br />

Ide, N. and L. Romary (<strong>2001</strong>). A Common Framework for Syntactic Annotation.<br />

Proceedings of ACL'<strong>2001</strong>: 298–305.<br />

Ide, N. and J. Veronis (1998). Introduction <strong>to</strong> the Special Issue on Word Sense<br />

Disambiguation: The State of the Art. Computational Linguistics 24: 1–40.<br />

Ide, N., P. Bonhomme, and L. Romary (2000). XCES: An XML-based Standard for<br />

Linguistic Corpora. Proceedings of the Second Language Resources and Evaluation<br />

Conference: 825–30.<br />

Ide, N., T. Erjavec, and D. Tufis (<strong>2001</strong>). Au<strong>to</strong>matic Sense Tagging Using Parallel<br />

Corpora. Proceedings of the Sixth Natural Language Processing Pacific Rim<br />

Symposium: 83–9.<br />

Jacobson, M., and B. Michailovsky (2000). A Linguistic Archive on the Web.<br />

IRCS/lSLE/Talkbank Linguistic Exploration Workshop: Web-Based Language<br />

Documentation and Description. At<br />

http://www.ldc.upenn.edu/exploration/expl2000/papers/michailovsky/index.htm.<br />

Johansson, S., G. Leech, and H. Goodluck (1978). Manual of Information <strong>to</strong> Accompany<br />

the Lancaster-Oslo/Bergen Corpus of British English, for Use with Digital Computers.<br />

Department of English, University of Oslo.


Kucera, H. and W. N. Francis (1967). Computational Analysis of Present-Day American<br />

English. Providence, RI: Brown University Press.<br />

Lamel, L. F., J. Garafolo, J. Fiscus, W. Fisher, and D. S. Fallen (1990). TIMIT: The<br />

Darpa Acoustic-phonetic Speech Corpus. Technical Report NISTIR 4930, National<br />

Institute of Standards and Technology.<br />

Leech, G. (1993). Corpus Annotation Schemes. Literary and Linguistic Computing 8:<br />

275–81.<br />

MacWhinney, B. (1995). The CHILDES Project, 2nd edn. Mahwah, NJ: Lawrence<br />

Erlbaum.<br />

Mann, W. C. and S. A. Thompson (1988). Rhe<strong>to</strong>rical Structure Theory: A Theory of Text<br />

Organization. Text 8: 243–81.<br />

Marcu, D. (1996). Building Up Rhe<strong>to</strong>rical Structure Trees. Proceedings of the Thirteenth<br />

National Conference on Artificial Intelligence 2: 1069–74.<br />

Marcus, M., B. San<strong>to</strong>rini, and M. A. Marcinkiewicz (1993). Building a Large Annotated<br />

Corpus of English: The Penn Treebank. Computational Linguistics 19, 2.<br />

Miller, G., C. Leacock, T. Randee, and R. Bunker (1993). A Semantic Concordance.<br />

Proceedings of the 3rd DARPA Workshop on Human Language Technology. 303–8.<br />

Miller, G. A., R. Beckwith, C. Fellbaum, D. Gross, and K. Miller (1990). WordNet: An<br />

On-line Lexical Database. International Journal of Lexicography 3: 235–44.<br />

Ramshaw, L. and M. Marcus (1995). Text Chunking Using Transformation-based<br />

Learning. Proceedings of the Third ACL Workshop on Very Large Corpora: 82–94.<br />

Srinivas, B. (1997). Performance Evaluation of Supertagging for Partial Parsing. In H.<br />

Blunt and A. Nijholt (eds.), Advances in Probabilistic and other Parsing Technologies<br />

(pp. 203–20). The Hague: Kluwer Academic.<br />

Stegman, R., H. Schulz, and E. Hinrichs (1998). Stylebook for the German Treebank in<br />

VERBMOBIL. Manuskript Universität Tübingen. At http://verbmobil.dfki.de/verbmobil/.<br />

Svartvik, J., (ed.) (1990). The London Corpus of Spoken English: Description and<br />

Research. Lund Studies in English 82. Lund: Lund University Press.<br />

Van Genabith, J., A Way, and L. Sadler (1999). Semi-au<strong>to</strong>matic Generation of f-<br />

Structures from Tree Banks. In M. Butt and T. H. King (eds.), Proceedings of the LFG99<br />

Conference, Manchester University, July 19–21. CSLI Online, From Treebank Resources<br />

<strong>to</strong> LFG Structures. Stanford, CA. At http://www-csli.stanford.edu/publications/.


Voutilainen, A. (1995). A Syntax-based Part-of-speech Analyzer. Proceedings of the<br />

European Chapter of the Association for Computational Linguistics: 157–64.<br />

Xue, N., F. Chiou, and M. Palmer (2002). Building a Large-scale Annotated Chinese<br />

Corpus. Proceedings of the 19th International Conference on Computational Linguistics<br />

(COLING).<br />

22.<br />

Electronic Scholarly Editing<br />

Martha Nell Smith<br />

Though they are not bound <strong>to</strong>gether in this multi-authored, multi-edited volume and are<br />

not likely <strong>to</strong> be in any other, this essay is the second in a series I am writing musing on<br />

the effects and meanings of the kinds of electronic scholarly editing, work extending well<br />

beyond the spatial and typographically fixed limitations of the codex, which I and others<br />

have been doing and that has become increasingly visible over the past decade. Johanna<br />

Drucker's observations about the nature of writing as she reflects upon the word itself<br />

convey well, in order <strong>to</strong> reflect upon editing, its nature and character. Writing (editing) is<br />

"a noun as well as a verb, an act and a product, a visual and verbal form, the composition<br />

of a text and trace of the hand"; "letters, words, and pic<strong>to</strong>rial elements all participate in<br />

producing a work with complex textual value. At its most fundamental writing is<br />

inscription, a physical act which is the foundation of literary and symbolic activity"<br />

(Drucker 1998: 3). Editing is a physical as well as a philosophical act, and the medium in<br />

which an edition is produced (or an edition's place in the material world) is both part of<br />

and contains the message of the edi<strong>to</strong>rial philosophies at work. As Adrienne Rich<br />

remarked about poetry (in "North American Time"), so editing "never s<strong>to</strong>od a chance / of<br />

standing outside his<strong>to</strong>ry." Indeed, we have entered a different edi<strong>to</strong>rial time, one that<br />

demands the conscious cultivation of and by many hands, eyes, ears, and voices. While<br />

print editions are containers for static objects, artifacts that are by definition<br />

unchangeable once produced, the world of <strong>digital</strong> surrogates practically demands new<br />

models for edi<strong>to</strong>rial praxes in which edi<strong>to</strong>rs and readers work <strong>to</strong>gether. Such models are<br />

encouraged by the fact that in a world with access <strong>to</strong> pho<strong>to</strong>graphic copies of texts and<br />

images, no one has <strong>to</strong> bear the burden of forging the perfect linguistic description of the<br />

artifact, and by the fact that <strong>digital</strong> artifacts are by definition alterable once produced.<br />

After all, <strong>digital</strong> surrogates featuring high-quality color images of a writer's manuscripts<br />

offer a more ample sense of their textual conditions, including the conditions of the<br />

writing scene in which they were produced. Informed more fully about textual conditions,<br />

readers can collaborate with the postulating edi<strong>to</strong>r in making the edi<strong>to</strong>rial artifacts for<br />

electronic media in ways not possible when decisions have already been made <strong>to</strong> exclude<br />

or include data and seal the resulting artifact in<strong>to</strong> a printed state.<br />

That this series of my essays examining and hypothesizing about these new edi<strong>to</strong>rial<br />

praxes does not share the binding of a book's spine is appropriate for a critique of work


on scholarly projects that do not have clear end points – the boundaries that variorums<br />

and reading editions (such as the printed volume of Emily Dickinson's writings on which<br />

I collaborated) do. The first essay – "Computing: What's American Literary Study Got <strong>to</strong><br />

Do with IT?" – was directed <strong>to</strong>ward my Americanist colleagues who have not embraced<br />

new media for their scholarly work. The present essay is directed <strong>to</strong>ward upper-level<br />

undergraduate and graduate students and instruc<strong>to</strong>rs, seasoned scholars and edi<strong>to</strong>rs, as<br />

well as other interested readers, many of whom may have embraced new media for their<br />

academic work while others may just be exploring the terri<strong>to</strong>ry, as it were. Like that first<br />

essay, this one is concerned with the changing sociologies surrounding new media critical<br />

endeavors, and is focused particularly on the impact of those sociologies on the work and<br />

workers involved in electronic scholarly editing and the development of its principles<br />

(described in Part Two of this book). The changes evident in <strong>digital</strong> <strong>humanities</strong> in how<br />

we work in, as, with, and for groups constitute a profound shift in <strong>humanities</strong> knowledge<br />

production. The new edi<strong>to</strong>rial praxes made possible, indeed demanded, by the critical<br />

environments that new media call in<strong>to</strong> being are pivotal <strong>to</strong> that shift. Those edi<strong>to</strong>rial<br />

praxes are not only made visible, but are constituted by some of the new technologies in<br />

<strong>digital</strong> <strong>humanities</strong>, technologies that have gone under-remarked, even by specialists.<br />

These praxes and technologies are this essay's primary foci, because they make possible<br />

the <strong>digital</strong> resources that, as I remarked in my earlier piece, are more than advantages for<br />

<strong>humanities</strong> knowledge production: they are necessities. As such, these praxes and<br />

technologies should be taken in<strong>to</strong> account by anyone working or interested in <strong>humanities</strong><br />

knowledge production.<br />

Technologies bring scientific knowledge <strong>to</strong> bear on practical purposes in a particular<br />

field, and are thus the means by which we accomplish various ends – the <strong>to</strong>ols and<br />

devices on which our critical suppositions rely. Some <strong>digital</strong> technologies, the machines<br />

and software with which we produce and consume scholarly editions, are prey <strong>to</strong><br />

obsolescence (HyperCard, for example, is for all practical purposes already unusable, as<br />

are DOS-based machines; the SGML document delivery system DynaWeb has<br />

limitations, at least aesthetically, that are much lamented; examples abound). By contrast,<br />

all but one of the technologies under review here were already used in some form for<br />

book production and are obviously not susceptible <strong>to</strong> such rapid obsolescence. In fact, in<br />

migrating <strong>to</strong> the electronic realm the technologies have become much more visible and<br />

productively exploitable. Constitutive of <strong>digital</strong> <strong>humanities</strong>, the technologies are: access,<br />

multimedia study objects, collaboration, and increased self-consciousness. Central <strong>to</strong> the<br />

formulations and speculations of this critique is evaluating the effects of a technology<br />

implicitly invoked by all four, that of audience. If one considers a sense of audience a<br />

technology (with explanation and performance as kinds of knowledge application), then<br />

the technology of audience provides analytical perspectives that would not have been<br />

obtained had I been writing with only one audience in mind. Those different viewpoints<br />

have in turn deepened my understanding of just how profound are the shifts under way in<br />

scholarly applications of <strong>digital</strong> resources <strong>to</strong> the study of arts and <strong>humanities</strong>. That I and<br />

other enthusiasts have underestimated the degree of that profundity is remarkable.<br />

Finally, this chapter makes explicit the importance of standards, as well as the importance<br />

of querying the prides and prejudices of those standards and the implications for<br />

interpretation made possible in this new edi<strong>to</strong>rial work, invoking as a representative


<strong>to</strong>uchs<strong>to</strong>ne the new Guidelines for Electronic Editions just approved for use by the<br />

Modern Language Association (MLA)'s Committee on Scholarly Editions.<br />

The fruits of several evolutions converged <strong>to</strong> produce a mode of editing suited <strong>to</strong> the new<br />

media: post-structuralist re-inflections of theories of editing; the proliferation of<br />

affordable, portable personal computers; the networking, on an unprecedented scale, of<br />

homes, individuals, institutions of all sorts (educational, governmental, commercial,<br />

religious, and medical), and nations. A fantasy unimaginable as part of the real world just<br />

a very short time ago, such networking and the accelerated communication it brings are<br />

now facts of everyday life throughout much of the world. Concomitantly, academics have<br />

moved beyond worried resistance <strong>to</strong> the work made possible by new media and computer<br />

technologies, and worked through much of the reactionary skepticism so widely and<br />

forcefully articulated in the 1990s, in books like Sven Birkerts's The Gutenberg Elegies<br />

(1994). By contrast, at the time of writing, under-informed skepticism has been replaced<br />

by the realization that critical engagements with new technologies are the best hope for<br />

advancing knowledge production in the <strong>humanities</strong>. That these advances are within our<br />

grasp is abundantly clear in the field of scholarly editing.<br />

For example, one happy consequence of new media revolutions is that breathtaking<br />

advances in networking and communication have enabled edi<strong>to</strong>rial praxes previously<br />

thought <strong>to</strong>o impractical <strong>to</strong> enact, whatever their intellectual value. Instead of being bound<br />

<strong>to</strong> representative samples, editions can now include images of all primary documents<br />

included in an edition. The surrogates possible in <strong>digital</strong> <strong>humanities</strong> are more than just<br />

singular pho<strong>to</strong>graphic examples meant <strong>to</strong> represent hundreds or even thousands of<br />

instances. Thus edi<strong>to</strong>rial claims <strong>to</strong> being "comprehensive" (a staple of the print variorum<br />

relying on representative examples) accrue whole new meanings in the extraordinarily<br />

capacious <strong>digital</strong> realm. Ironically, though, electronic textual edi<strong>to</strong>rs would not be likely<br />

<strong>to</strong> make such claims, realizing just how vast is the information one could choose <strong>to</strong><br />

include in a "comprehensive" edition. Those same edi<strong>to</strong>rs would nevertheless demand<br />

inclusion of images of all primary documents (manuscripts or corrected proofs, for<br />

example) rather than only the analytical descriptions thereof one finds in the edi<strong>to</strong>rial<br />

apparatus of print translations (iterations mechanically reproduced for public distribution<br />

and annotated for scholarly consumption). Also, editions can now implement new kinds<br />

of scholarly annotation by including, where applicable, sound and even video<br />

reproductions simply impossible <strong>to</strong> contain as a constitutive part of a book (though a CD<br />

might be slipped inside the cover as a <strong>companion</strong> or a chip ingeniously packaged and<br />

placed for a reader's play).<br />

Networking and communication have also made imaginable and agreeable edi<strong>to</strong>rial<br />

praxes traditionally thought impracticable or undesirable. These practices modify the<br />

very nature of scholarly editing itself. For example, if an edi<strong>to</strong>r chooses, the much<br />

theorized common reader can now provide immediate and actual (in contrast <strong>to</strong><br />

imagined) critical feedback that might in turn be usefully incorporated in<strong>to</strong> the work of a<br />

scholarly edition, making the collaboration between reader and author that is<br />

characteristic of reading itself something more than the "creative regeneration" [of texts]<br />

by readers invested with and in private intentions such as those D. F. McKenzie describes


in Bibliography and the Sociology of Texts (1999). All of these fac<strong>to</strong>rs will be examined<br />

in this essay as examples of under-analyzed technologies, but first some observations<br />

about editing and the phrase that provides my title are in order.<br />

Distinguished editing of literary and artistic work is a matter of the heart (emotions), and<br />

at least four of our five senses (physical perceptions), as well as of the head (logic).<br />

Astute edi<strong>to</strong>rs take advantage of the acumen offered by these diverse ways of<br />

apprehending the world, and though it is in one sense bound by algorithms, <strong>digital</strong> editing<br />

is no exception in terms of its characteristics (of the heart, the physical senses, the head).<br />

Reflecting on the individual terms embodied by the phrase, and the traditions they<br />

bespeak, deepens and broadens understandings of "electronic scholarly editing", both in<br />

its present forms and in the theories about those praxes and the possibilities that inhere in<br />

multimedia. Editing makes works (poems, plays, fiction, film footage, musical<br />

performances, and artistic and documentary material) publishable (in books, films,<br />

television and radio, and recordings) by eliminating unwanted material and organizing<br />

what remains for optimal and intelligible presentation <strong>to</strong> audiences. In other words,<br />

editing translates raw creative work in<strong>to</strong> an authoritative (not <strong>to</strong> be confused with<br />

definitive or authoritarian) form. Scholarly editing is editing performed under the aegis of<br />

research, learning, sustained instruction, mastery, knowledge building, standard setting.<br />

Electronic scholarly editing consciously incorporates phenomena associated with the<br />

movement and manipulation of electrons, those indivisible charges of negative electricity,<br />

through wires and radio waves on<strong>to</strong> screens and through speakers. Though it incorporates<br />

the principles of conventional scholarly editing (that is, in and for books) in<strong>to</strong> its own<br />

methods, electronic scholarly editing is not performed <strong>to</strong> be read through the machine of<br />

the book but via the hardware of computers. Other nomenclatures for this field include<br />

"<strong>digital</strong> scholarly editing", exploiting <strong>to</strong> the fullest the obvious pun that enables clear<br />

contrastive play between the manual and the au<strong>to</strong>mated, and those pertaining <strong>to</strong> editing<br />

itself, especially in the world of ubiqui<strong>to</strong>us computers and the connective possibilities of<br />

the World Wide Web. Electronic scholarly editing now inevitably involves networking,<br />

both among the texts being worked through and the workers performing the editing<br />

(including readers viewing the edited works), in ways never realized by that magnificent<br />

<strong>to</strong>ol of knowledge transmission that has served so splendidly well for the past several<br />

centuries, the book.<br />

Previously I have examined the implications for knowledge production of the fact that in<br />

the <strong>digital</strong> environment, access – both <strong>to</strong> numbers of study objects and <strong>to</strong> numbers of<br />

audience members – is facilitated on an unprecedented scale. I considered at length the<br />

possibilities for literary criticism and theory when the staggering data s<strong>to</strong>rage capacities<br />

of computers enable so much more visibility of what I refer <strong>to</strong> as BE-O objects – artifacts<br />

that have cus<strong>to</strong>marily been viewed By Experts Only. Agreeing that the skepticism with<br />

which many mainstream scholars regard "quantitative research methods" was not really<br />

misplaced "so long as the computer's primary role in the <strong>humanities</strong> was, ostensibly, <strong>to</strong><br />

compute", Matthew Kirschenbaum proclaims that access <strong>to</strong> a <strong>digital</strong> surrogate of a<br />

"Rossetti painting" or "of one of Emily Dickinson's turbulent manuscripts" erases or at<br />

least reverses that skepticism. The visibilities that enable the unprecedented access <strong>to</strong><br />

images that were previously locked away in library and museum archives for exclusive


view by a very few is probably the technological boon with which scholars and readers<br />

are most familiar, for that perpetually available and repeatable access is really quite a big<br />

deal. The ability <strong>to</strong> provide artifacts for direct examination (rather than relying on<br />

scholarly hearsay) has altered the reception of <strong>humanities</strong> computing in the disciplines of<br />

the <strong>humanities</strong> so that skepticism is "at least replaced with more <strong>to</strong>-the-point questions<br />

about image acquisition and edi<strong>to</strong>rial fidelity, not <strong>to</strong> mention scholarly and pedagogical<br />

potential." These are indeed "questions of representation, and they are eminently relevant<br />

<strong>to</strong> the work of the <strong>humanities</strong>" (Kirschenbaum 2002: 4). As Ray Siemens's introduction<br />

<strong>to</strong> A New Computer-assisted Literary Criticism? makes plain, enthusiasts of text analysis<br />

would take issue with Kirschenbaum's assertion about quantitative research methods, but<br />

the point about the power of visibility, of making accessible for as many pairs of eyes as<br />

possible, images (rather than their analytical textual descriptions) for critical analysis,<br />

cannot be overstated. When edi<strong>to</strong>rs make as much about a text visible <strong>to</strong> as wide an<br />

audience as possible, rather than silencing opposing views or establishing one definitive<br />

text over all others, intellectual connections are more likely <strong>to</strong> be found than lost.<br />

Though it will not necessarily do so, that access <strong>to</strong> primary <strong>digital</strong> artifacts can in turn<br />

create extraordinary access <strong>to</strong> the edi<strong>to</strong>rial process itself, especially if edi<strong>to</strong>rs and readers<br />

alike are proactive in taking advantage of these opportunities by building "sound<br />

infrastructure, and the organizational and financial structures … essential <strong>to</strong> consolidate,<br />

<strong>to</strong> preserve, and <strong>to</strong> maintain" the "new organizational forms" created by electronic media<br />

(D'Arms 2000). The standards described by the MLA's Guidelines for Scholarly Editing,<br />

print and electronic, are commonsensical and bound <strong>to</strong> ensure quality: they ask for<br />

"explicitness and consistency with respect <strong>to</strong> methods, accuracy with respect <strong>to</strong> texts,<br />

adequacy and appropriateness with respect <strong>to</strong> documenting edi<strong>to</strong>rial principles and<br />

practice." However practical and necessary for meeting principled goals and however<br />

derived by scholarly consensus, uses of various media constrain and enable adherence <strong>to</strong><br />

these benchmarks <strong>to</strong> different degrees. Indeed, one clear example lies in the fact that<br />

explicitness, consistency, and accuracy are all finally matters of faith in the bibliographic<br />

realm.<br />

Considering an actual instance will make clear the importance of enhanced visibility and<br />

the more critical reading it enables. My examples will be drawn from the edi<strong>to</strong>rial<br />

resources with which I have been most involved, both as a user and a maker, those of the<br />

writings of the nineteenth-century American poet Emily Dickinson. Because the practices<br />

involved in their making are so widespread in the production of scholarly editions, and<br />

similar principles <strong>to</strong> those employed in producing Dickinson are so widely accepted for<br />

making accessible the texts of other authors, these specific instances serve well as<br />

exemplifications of general rules and procedures for scholarly textual production. R. W.<br />

Franklin spent scrupulous decades producing a three-volume variorum edition of The<br />

Poems of Emily Dickinson for Harvard University Press, a publisher highly esteemed for<br />

producing study objects meeting the highest standards. Franklin's stated goal is <strong>to</strong> give a<br />

"comprehensive account" of all of Dickinson's texts, as well as any texts directly bearing<br />

on her compositional practices (1998: 28). In a comprehensive account, one would expect<br />

<strong>to</strong> see all omissions, additions, alterations, emendations of which the edi<strong>to</strong>r is aware<br />

marked in some way for his readers. Thus any ellipsis, any omission of data, would be


signaled by familiar punctuation (…) and probably explained in edi<strong>to</strong>rial notes. Yet this<br />

is not the case, and readers cannot readily know when an omission has occurred or a<br />

change silently made because their access <strong>to</strong> the objects edited is limited and the evidence<br />

of explicitness, consistency, and accuracy cannot be seen. Instead, readers have <strong>to</strong> trust<br />

that the printed texts they see are explicit, consistent, and accurate about all textual data.<br />

Thus when Franklin erases 74 words from his representation of an epis<strong>to</strong>lary manuscript<br />

involving the composition of Dickinson's poem "Safe in their Alabaster Chambers" and<br />

does so without commentary and without signifying punctuation, readers of his print<br />

variorum have no way of knowing that the omission has been made, much less of<br />

evaluating its import (FP 124C). Besides being misinformed about the nature of this text<br />

because of omission of information, access <strong>to</strong> the edi<strong>to</strong>rial process itself is limited, the<br />

implicit contract with the reader (that such treatment of documentary evidence would be<br />

signaled in some readable way) having been violated. Franklin's change might, one could<br />

argue, simply be chalked up <strong>to</strong> bad edi<strong>to</strong>rial work. Yet he silently enacts other<br />

emendations as a course of standard, widely accepted edi<strong>to</strong>rial practice. Much ink has<br />

been spilled by Dickinson and American poetry scholars about her use of punctuation<br />

marks that most have labeled "the dash." Strong positions have been taken about whether<br />

the shorter en- or the longer em-mark most faithfully translates Dickinson's mark in<strong>to</strong><br />

print, and Franklin has resolved the matter with the authoritarian stance that neither the<br />

en- nor the em-suffice: according <strong>to</strong> him, the shorter-than-either hyphen best conveys<br />

Dickinson's practice of using a horizontal or angled mark rather than a comma. The<br />

problem with staking out a hard and fast position on this matter (as many a Dickinson<br />

critic has done) is the "one size fits all" perspective practically required by print<br />

production. Examining almost any poem written during Dickinson's middle years shows<br />

that within individual poems, indeed within individual lines, she did not consign herself<br />

<strong>to</strong> a single sort of mark but used marks of varying lengths and angles (see, for example,<br />

"The name -of it -is 'Autumn'___" (FP 465; facsimile available in Franklin's Manuscript<br />

Books of Emily Dickinson, p. 494). Editing nearly 1,800 poems, Franklin makes the<br />

decision for readers that a hyphen will suffice <strong>to</strong> represent Dickinson's diverse, varied<br />

marks, and thus also decides for readers that the differences among her punctuating<br />

inscriptions are not poetically significant.<br />

Trust in any print edition, then, whether it be that made by Franklin or any other edi<strong>to</strong>r,<br />

including myself, is necessarily faith-based, for readers cannot adequately see the<br />

documentary evidence that determines everything from genre <strong>to</strong> suitability for inclusion<br />

in a scholarly edition. Indeed, such matters of representation are more than eminently<br />

relevant <strong>to</strong> <strong>humanities</strong>; they are central <strong>to</strong> <strong>humanities</strong> knowledge production itself.<br />

Opportunities for analysis of who made the scholarly objects we study and for what<br />

purposes, and of which parts of those objects are constitutive and worthy of study, are<br />

proportionate <strong>to</strong> the visibility of foundational documentary evidence enabled by the<br />

medium of representation.<br />

In contrast <strong>to</strong> the constrained visibilities of book representations, access <strong>to</strong> questions of<br />

edi<strong>to</strong>rial fidelity and therefore <strong>to</strong> the edi<strong>to</strong>rial process itself is much more obtainable in an<br />

electronic edition featuring images of all documents edited as well as their translations


in<strong>to</strong> typography. In such a realm, 74 words cannot simply be excised without<br />

commentary and marks translated with a "one size fits all" authoritarian stance and go<br />

unnoticed, because a <strong>digital</strong> image of the document is available for readers <strong>to</strong> view,<br />

assess, compare with edi<strong>to</strong>rs' representations. Representation via <strong>digital</strong> facsimiles of<br />

original documents changes access <strong>to</strong> the foundational materials of scholarly editions, the<br />

con<strong>to</strong>urs of expertise, and even standard-setting itself.<br />

Consciously foregrounding the on<strong>to</strong>logical differences between electronic and<br />

bibliographic scholarly editions is necessary for understanding ways in which this access<br />

can work, for learning how electronic editions made might best be used, and for planning<br />

and developing electronic editions that are more and more well made (in that they are<br />

designed for optimum usability as well as optimum incorporation of relevant but<br />

scattered data and commentary). Introducing Harvard University Press's latest variorum<br />

edition of The Poems of Emily Dickinson, Franklin claims that although that threevolume<br />

edition "is a printed codex", it is practically electronic, for the edition "has an<br />

electronic database" and the "poems are in bits and bytes." He then flatly states that<br />

"other outputs are possible, including other printed editions, organized or presented<br />

differently. Dickinson and Hypertext may well be matched, and images are particularly<br />

useful with an unpublished poet who left her poems unprepared for others" (1998: 27–8).<br />

Yet electronic editions are by their very constitution markedly different, both from the<br />

print edition Franklin made and the hypertext one he imagines. For one thing (which he<br />

acknowledges), the bits and bytes <strong>to</strong> which he refers do not contain images and thus do<br />

not provide multifaceted viewpoints, the visibility of <strong>digital</strong> surrogates of original<br />

manuscripts presented in an environment where the distance that separates the libraries in<br />

which they are held presents no obstacle for direct comparison of texts within each of<br />

those collections spread across oceans and nations, separated by geography, national<br />

boundaries, and disparate laws. Collected in<strong>to</strong> an electronic edition, Blake or Dickinson<br />

manuscripts logically related <strong>to</strong> one another but dispersed across various collections are<br />

gathered <strong>to</strong>gether so they can be studied and synthesized <strong>to</strong> make new knowledge;<br />

collected in<strong>to</strong> the same electronic edition, they no longer lie like scattered pieces of a<br />

puzzle, their connections splintered, difficult <strong>to</strong> see as possibly yoked, and thus next <strong>to</strong><br />

impossible <strong>to</strong> imagine. Another difference unacknowledged by Franklin is that the<br />

typographical bits and bytes <strong>to</strong> which he refers are those of word processing, which only<br />

encode the textual data found in Dickinson's works. A scholarly electronic edition would<br />

likely be prepared in XML or SGML (probably following the guidelines of the Text<br />

Encoding Initiative), and would capture logical, structural, and artifactual aspects of the<br />

original, as well as its textual content. An electronic edition also might include databases,<br />

Flash, video, and sound presentations. Different editions are not simply "outputs" of the<br />

same bits and bytes, which Franklin's declaration assumes. The statement makes plain<br />

that he envisions electronic editions as typographically edited texts with illustrations that<br />

can be accessed with the click of a mouse. But electronic scholarly editions provide much<br />

greater opportunities than texts with links <strong>to</strong> illustrations and <strong>to</strong> one another, enabling a<br />

range of manipulation not possible for those bound <strong>to</strong> the printed page.<br />

Far more than display, replications constitute the textual reproductions and<br />

representations in scholarly editions such as The Blake Archive, the Rossetti Archive, The


Walt Whitman Archive, The Canterbury Tales Project, the Dickinson Electronic<br />

Archives. Scholarly electronic editions are not simply a matter of bits and bytes and how<br />

they have been encoded <strong>to</strong> link <strong>to</strong> one another and <strong>to</strong> appear in particular browsers.<br />

Dynamic rather than static, deeply encoded, they are designed <strong>to</strong> enable readers <strong>to</strong><br />

achieve better understanding of texts and open up possibilities for more sophisticated<br />

interpretations. Susan Schreibman and others have chronicled the fact that by the mid-<br />

1980s it had become clear <strong>to</strong> numbers of scholars already involved in the creation of<br />

electronic resources for the <strong>humanities</strong> that a standard for encoding texts needed <strong>to</strong> be<br />

developed that was not dependent upon a particular platform or software package<br />

(Schreibman 2002a). Achieving platform-independence requires separating the<br />

information content of documents from their formatting, which is what led an<br />

international group of those concerned with publishing and preserving government<br />

documents <strong>to</strong> develop GML (Generalized Markup Language) in the 1970s, and then<br />

SGML (Standard Generalized Markup Language), adopted as an international standard in<br />

the mid-1980s. Working with SGML, a group of <strong>humanities</strong> scholars and librarians<br />

launched the Text Encoding Initiative (TEI) in 1987.<br />

Though markup has a performative significance, the TEI is designed <strong>to</strong> represent already<br />

existing literary texts. For electronic scholarly edi<strong>to</strong>rs, "markup is supplied as part of the<br />

transcription in electronic form of pre-existing material…. Markup reflects the<br />

understanding of the text held by the transcriber; we say that the markup expresses a<br />

claim about the text" (see Allen Renear, in chapter 17, this volume, and in writings<br />

elsewhere, and many others, for example Susan Hockey, Michael Sperberg-McQueen,<br />

Jerome McGann: see entries in the References for Further Reading). John Unsworth's<br />

claim that "any attempt <strong>to</strong> systematically express one's understanding of a text in the kind<br />

of internally consistent, explicit and unambiguous encoding that is required in the<br />

creation of a computable SGML or XML edition will produce some intellectual benefit<br />

and will ensure some degree of integrity", is indisputable (Unsworth, website). Indeed,<br />

the technology of self-consciousness required by computer encoding of texts, especially<br />

that such as TEI-conformant XML encoding, produces a healthy self-consciousness about<br />

what Bruno La<strong>to</strong>ur and Steve Woolgar describe in Labora<strong>to</strong>ry Life as "black-boxing" –<br />

which occurs when one "renders items of knowledge distinct from the circumstances of<br />

their creation" (1986: 259n). In black-boxing, critical opinion becomes "fact"; more often<br />

than not, amnesia sets in after that factual instantiation, and having been effectively<br />

black-boxed, "fact" becomes "truth." As in the case of Dickinson discussed here,<br />

wittingly or unwittingly bibliographic edi<strong>to</strong>rs sometimes remove elements that are (or<br />

well might be) constitutive of its poetics when transmitting a text from the messy<br />

circumstance of its creation <strong>to</strong> the ordered pages of print. The challenge for electronic<br />

scholarly edi<strong>to</strong>rs is not <strong>to</strong> perpetrate similar dis<strong>to</strong>rtions when enthusiastically embracing<br />

the ordering <strong>to</strong>ols of encoding <strong>to</strong> convey literary expression, which by its very nature<br />

often perplexes the orders of exposi<strong>to</strong>ry logic.<br />

Hypertext theorists and practitioners, who have been (sometimes roundly) critiqued for<br />

being <strong>to</strong>o exuberant, are sometimes echoed by the encoding specialist's enthusiasm for<br />

important developments in XML: "One of the most powerful features of the XML family<br />

of technologies is Extensible Stylesheet Language Transformations (XSLT), a language


that facilitates transformation of text rather than display (as HTML's Cascading Style<br />

Sheet language does)…. The 'infinitely recenterable system whose provisional point of<br />

focus depends on the reader' (Landow 1997, p. 36) is made realizable through XML-<br />

XSLT technology, and is much closer <strong>to</strong> the democratized space envisioned by early<br />

hypertext theorists" (Schreibman 2002b: 85). Without a doubt, though, new technologies<br />

for structuring data, coupled with the guidelines provided by the Committee on Scholarly<br />

Editions and the TEI's commitment <strong>to</strong> work out of a "broad, community-based<br />

consensus", are crucial for realizing advances in edi<strong>to</strong>rial praxes. Also necessary is the<br />

commitment <strong>to</strong> facilitate difference by evolving "guidelines … <strong>to</strong> accommodate a greater<br />

variety of edi<strong>to</strong>rial methods and a broader range of materials and periods" (MLA<br />

Committee on Scholarly Editions, website), and by the TEI's "extension mechanism, as<br />

well as its consistent insistence… on international and interdisciplinary representation in<br />

its governing bodies, its workgroups, its funding sources, and its membership"<br />

(Unsworth, website). The collaborative development of these standards is absolutely<br />

essential for any demotic ethic valued as a worthwhile goal, and the disambiguation<br />

forced upon knowledge workers who thrive on ambiguity often proves critically<br />

profound, because texts are seen as never before, both on the surface and in their deep<br />

structures.<br />

Yet none of these advances in markup, nor any of these guidelines, is robust enough <strong>to</strong><br />

accommodate all facets of the actual textual experience of edi<strong>to</strong>rs working with primary<br />

artifacts. Original documents, the raw materials with which edi<strong>to</strong>rs must work, are by<br />

their very nature queer, and must be normalized <strong>to</strong> some degree in order <strong>to</strong> be put in<strong>to</strong> an<br />

edition. In part, this is a result of the fact that "there are certain basic – irresolvable –<br />

philosophical tensions in language – most particularly between its capacity <strong>to</strong> represent<br />

knowledge systematically and its capacity <strong>to</strong> be knowledge experientially and<br />

perceptually – and this tension intensifies in an electronic environment as a tension<br />

between machine language and natural language – since it is a reductive hybrid version of<br />

the one which can be encoded/encrypted in order <strong>to</strong> serve as the basis of the other"<br />

(Drucker 1998: 219). This characteristic of language itself is much more visible in an<br />

environment in which<br />

the more sophisticated we are the more we normalize textual incommensurables. We<br />

have internalized an immensely complicated, many-leveled set of semiotic rules and<br />

signs, and we control the contradictions of actual textual circumstances by various<br />

normalizing operations. We can hope <strong>to</strong> expose these normalizations – which are<br />

themselves deformative acts – by opening the conversation… between analogue and<br />

<strong>digital</strong> readers. We begin by implementing what we think we know about the rules of<br />

bibliographical codes. The conversation should force us <strong>to</strong> see – finally, <strong>to</strong> imagine –<br />

what we don't know that we know about texts and textuality. At that point, perhaps, we<br />

may begin setting philology – "the knowledge of what is known", as it used <strong>to</strong> be called –<br />

on a new footing.<br />

(McGann <strong>2001</strong>: 207)


In a case such as the Dickinson Electronic Archives, the editions are explicitly designed<br />

not <strong>to</strong> define and normalize texts, as has been the objective of most bibliographic<br />

scholarly editions, and has been the practice of many electronic scholarly editions. Many<br />

electronic editions are still framed by the good work of the book because the dream of<br />

shedding the inertia imported from the bibliographical realm is only gradually finding its<br />

modes of practice. The TEI was developed <strong>to</strong> represent already existing literary texts in<br />

electronic media, but in its years of development it has been confronted by queerer and<br />

queerer texts and by edi<strong>to</strong>rs whose commitment <strong>to</strong> integrity means that they cannot<br />

simply normalize those texts <strong>to</strong> fit a predetermined grid. This, in turn, demands that we<br />

grasp the real significance of the truism that editing is a kind of encoding and encoding is<br />

a kind of editing, and it also requires that we probe the politics of the encoding standards<br />

we are embracing. Readers <strong>to</strong>o<br />

rarely think about the myriad of databases, standards, and instruction manuals subtending<br />

our reading lamps, much less about the politics of the electric grid that they tap in<strong>to</strong>. And<br />

so on, as many layers of technology accrue and expand over space and time. Systems of<br />

classification (and of standardization) form a juncture of social organization, moral order,<br />

and layers of technical integration. Each subsystem inherits, increasingly as it scales up,<br />

the inertia of the installed bases of systems that have come before.<br />

(Bowker and Star 1999: 33)<br />

Besides asking who made our objects of study, generative questions about standard<br />

setting, who made them, and for what purposes should be posed, and in ways, as McGann<br />

argues, not previously imagined. "There is more at stake – epistemologically, politically,<br />

and ethically – in the day-<strong>to</strong>-day work of building classification systems and producing<br />

and maintaining standards than in abstract arguments about representation" (Bowker and<br />

Star 1999: 10). Standards are of course crucial for realizing reliability, sustainability<br />

(both in terms of being intellectually substantive and in terms of residing in a preservable<br />

medium), and interoperability among different works and even different systems. Those<br />

editing for new media are carrying on this responsibility by working with and helping <strong>to</strong><br />

develop the new international standards and guidelines. In this ongoing work, we must<br />

self-consciously pose questions about the consequences of standardizing, classifying, and<br />

categorizing. In previous scholarly editions, classifications or the qualifying criteria<br />

dictating how an entity would be classified have been invisible. One benefit of encoding<br />

is "<strong>to</strong> understand the role of invisibility in the work that classification does in ordering<br />

human interaction" and thus <strong>to</strong> keep an eye on the "moral and ethical agenda in our<br />

querying of these systems. Each standard and each category valorizes some point of view<br />

and silences another. This is not inherently a bad thing – indeed it is inescapable"<br />

(Bowker and Star 1999: 5). Standards and categories become problematic when they are<br />

insufficiently critiqued, and when "a folk theory of categorization itself" prevails. That<br />

folk theory "says that things come in well-defined kinds, that the kinds are characterized<br />

by shared properties, and that there is one right taxonomy of the kinds" (Lakoff<br />

1987:121). Both the TEI Consortium and the MLA's Committee on Scholarly Editions<br />

have said that principled accommodation is necessary, and that's as it must be, if<br />

electronic scholarly editing is <strong>to</strong> realize its potential by pre-empting this compulsion


<strong>to</strong>ward standard setting as a kind of intellectual police force. The community has<br />

established standards that are flexible and adjustable, adopting what in the builder's trade<br />

is called a lesbian rule – a mason's rule of lead, which bends <strong>to</strong> fit the curves of a<br />

molding; hence, figuratively, lesbian rules are pliant and accommodating principles for<br />

judgment (OED). Commitment <strong>to</strong> such principles for judgment must be vigilantly<br />

maintained, with edi<strong>to</strong>rs relentlessly asking: "What work are the classifications and<br />

standards doing? Who does that work? What happens <strong>to</strong> the cases that do not fit the<br />

classification scheme? What happens <strong>to</strong> cases in which several classification schemes<br />

apply?" After all, au<strong>to</strong>ma<strong>to</strong>ns do not make literary texts: people do.<br />

Understanding the poetics and principles of electronic scholarly editing means<br />

understanding that the primary goal of this activity is not <strong>to</strong> dictate what can be seen but<br />

rather <strong>to</strong> open up ways of seeing. The disambiguating codes are <strong>to</strong>ols <strong>to</strong> understand texts,<br />

not <strong>to</strong> define them. Though consensus may develop around certain elements of what is<br />

seen, no consensus need be developed on how <strong>to</strong> read those elements. In fact, the<br />

different perspectives offered via the encoding, image-based <strong>digital</strong> surrogates, and direct<br />

critiques from "common" readers that are possible in electronic scholarly editing create a<br />

climate of possibility for interpretation that is precluded by the invisible controls placed<br />

on literary works and their interpretation in bibliographic editing. The edi<strong>to</strong>rial goal of<br />

electronic scholarly editing is not <strong>to</strong> find the proper literary container for "poems" (or<br />

whatever genre is being produced) but <strong>to</strong> find the medium that transmits more, rather<br />

than fewer, of the techniques inscribed and found on the literary page. Propriety (which<br />

inheres in the notion of the proper container) is not the issue, hence the TEI's extension<br />

mechanism. The issue is textual pleasure and access <strong>to</strong> greater understanding, and<br />

enabling audiences <strong>to</strong> avail themselves of as many aspects as possible of the scriptures<br />

under study.<br />

So in scholarly editing, the challenge of using these electronic <strong>to</strong>ols that create so many<br />

advantages for s<strong>to</strong>rage of data (including sound and images), retrieval, and searching is <strong>to</strong><br />

develop them so that edi<strong>to</strong>rial praxes themselves are truly advanced and the hieratic ethic<br />

of editing for books is not simply imported in<strong>to</strong> a new, more proficient medium (as was<br />

originally the case with the TEI). Derrida's deconstruction of the meanings of "archive"<br />

(Derrida 1996) and its connotations of "commandment" and "commencement" help<br />

clarify distinctions between hieratic and demotic and show why the latter is an advance<br />

when it comes <strong>to</strong> edi<strong>to</strong>rial approaches. By tracing his emphasis on a hierarchy of genres,<br />

one can see that commandment – stressing authority, social order – describes the guiding<br />

principle of Franklin's variorum. Thus in spite of his emphasis on versioning in his<br />

iteration of 1,789 Emily Dickinson poems, order is imposed on that which is otherwise<br />

unruly – the messy handwritten artifacts of poems, letters, letter-poems, scraps, notes,<br />

fragments. The idea of "poem" disciplines and contains views and the version represented<br />

of Dickinson's writings so that they conform <strong>to</strong> social order and literary law, whatever the<br />

material evidence may suggest. According <strong>to</strong> this principle of commandment, the material<br />

evidence, the manuscripts, contain the idea of "poem" and an edi<strong>to</strong>r's job is <strong>to</strong> deliver that<br />

idea in a container that makes "poem" extractable. Here textual boundaries are clear,<br />

commanded as they are by the ideas that demarcate genres for books. The temptation in<br />

disambiguating markup is <strong>to</strong> repeat that strategy of containment, for it is much easier


simply <strong>to</strong> claim that a text is prose or poetry rather than <strong>to</strong> acknowledge that it is much<br />

more complicated than that, a blend of the two, and requires innovative extensions of the<br />

predetermined TEI markup schemes.<br />

By contrast, commencement – physical, his<strong>to</strong>rical, on<strong>to</strong>logical beginning – describes the<br />

guiding principle of the Dickinson Electronic Archives' production, including our<br />

markup. Unpersuaded that "poem" is an "idea" easily separable from its artifact, the<br />

edi<strong>to</strong>rs of the electronic archives feature images of Dickinson's manuscript bodies in their<br />

multiple sizes and shapes, in all their messiness. Though our markup designates verse,<br />

letter, verse-letter, letter with embedded verse, and letter with enclosed verse, what<br />

constitutes a "poem" and poetic meanings is left up <strong>to</strong> the reader<br />

(http://jefferson.village.virginia.edu/dickinson/nosearch/documentation/). A work might<br />

sport a stamp or a cu<strong>to</strong>ut from Dickens placed by Dickinson in<strong>to</strong> the literary scene that<br />

the reader deems part of a "poem", and the edi<strong>to</strong>rs of the electronic archives refuse <strong>to</strong><br />

bind those elements as extra-literary but put them on display and include them in the<br />

claims our markup makes about the text. Finger smudges, pinholes, paste marks, coffee<br />

stains, and traces of ribbons, flowers, or other attachments offer a view in<strong>to</strong> the<br />

manuscript circulation and exchange so central <strong>to</strong> Dickinson's literary world, and we<br />

likewise put them on display and include questions about such signs in our edi<strong>to</strong>rial<br />

submission forms so that markup and notes recognize these materialities as part and<br />

parcel of literary production and circulation. Also featured are images of the printed<br />

pages, the bodies that have transmitted Dickinson's writings <strong>to</strong> the world, and in these are<br />

s<strong>to</strong>ries of Dickinson's his<strong>to</strong>ry as a poet whose writings are read and enjoyed by a wide<br />

audience. The tidy organizations of those printings bound in<strong>to</strong> The Poems of Emily<br />

Dickinson and The Letters of Emily Dickinson juxtaposed with the not fully intelligible<br />

bindings of manuscripts (in<strong>to</strong> the manuscript books found in her room or her<br />

correspondences <strong>to</strong> 99 recipients) by Dickinson herself, as well as with her many writings<br />

unbound (single sheets, notes, drafts, fragments, scraps), renew on<strong>to</strong>logical questions<br />

about the identities of these many writings. Textual boundaries are not clear, hence on<br />

our submission form for co-edi<strong>to</strong>rs, the word "genre" in the question asking them <strong>to</strong><br />

select one for encoders is in quotation marks <strong>to</strong> underscore the fact that such denotation is<br />

forced by the encoding scheme. Underscored both in the markup and the display of<br />

original documents is the fact that though an ideal idea of "poem" or "letter" must<br />

dominate for the writings <strong>to</strong> be neatly divided by bibliograph-ically determined genre,<br />

and though some denotation must be made for the meta-information of markup,<br />

electronic editing enables a much more flexible approach <strong>to</strong> understanding Dickinson's<br />

writings and her manipulations of genre. With many of the documents, readers cannot<br />

help but begin <strong>to</strong> ask "what is this?" "What is this writer doing?" The electronic archives<br />

are designed <strong>to</strong> enable much more extensive considerations of such questions, <strong>to</strong> confront<br />

the philosophical tensions in language rather than smoothing them over and masking<br />

them via edi<strong>to</strong>rial "clean-up."<br />

Thus if edi<strong>to</strong>rs self-consciously work beyond the inertia inherited from bibliographic<br />

editing, the edi<strong>to</strong>rial environment made possible by these technologies of publication can<br />

have profound implications for the interpretation of texts and for knowledge production.<br />

Though editing might pretend <strong>to</strong> be objective, it is always someone enacting his or her


critical prerogatives. Presenting artistic and critical works can never be done without<br />

some bias being conveyed. Though the fact that editing is always interpretation was<br />

realized long before their incorporation in<strong>to</strong> <strong>humanities</strong> work, the new media, and new<br />

and newly recognized technologies facilitate accommodation of this fluidity of authorial<br />

and edi<strong>to</strong>rial/authorial intentions. Edi<strong>to</strong>rial theorists as diverse as Tanselle, McGann,<br />

Bornstein, Bryant, Shillingsburg, McKenzie, and Grigely in one way or another<br />

acknowledge the philosophical tensions in language in which these fluidities inhere and<br />

point out that "writing is fundamentally an arbitrary hence unstable hence variable<br />

approximation of thought" (Bryant 2002: 1). In the context of these recognitions, all of<br />

these values are profoundly affected: authority, literariness, authenticity, sociology,<br />

access, reproductivity, original/originary texts/moments, edi<strong>to</strong>rial responsibilities,<br />

authorship, intention. Electronic editions are most valuable in their capaciousness, not<br />

only in their ability <strong>to</strong> offer different versions but also in their ability <strong>to</strong> facilitate varying<br />

views and valuations, both of what is seen and of how it is comprehended, as text or as<br />

extra-textual.<br />

Though major edi<strong>to</strong>rial projects have often involved collabora<strong>to</strong>rs, the collaborations<br />

enabled by electronic editions can be something different than previously seen in<br />

<strong>humanities</strong> research, as are the editions themselves. Because websites can be mounted by<br />

individuals and because most have not been vetted, peer review has been a contentious<br />

issue in the first generation of <strong>digital</strong> <strong>humanities</strong> production. Publishers, libraries, and<br />

other institutions invested in quality control routinely vet their <strong>digital</strong> publications in<br />

ways akin <strong>to</strong> those used for print publications: two or three readers' reports by experts in<br />

the field are presented <strong>to</strong> a board who then votes <strong>to</strong> accept or reject. This is important and<br />

valuable, but critical review need not s<strong>to</strong>p there. In electronic publishing, <strong>to</strong>ols such as<br />

dynamic databases enable a more sustained and collaborative critical engagement of<br />

reader with edi<strong>to</strong>r than has previously been seen. The Dickinson Electronic Archives has,<br />

for example, opened a critical review space<br />

, making user feedback an integral part<br />

of the edi<strong>to</strong>rial process. In this virtual space users have the opportunity <strong>to</strong> post<br />

commentary, reviews, or critical analysis of a particular section(s) or of the site as a<br />

whole. Because comments are s<strong>to</strong>red (either anonymously or with signature) in a<br />

searchable database, users have the option of reading and responding <strong>to</strong> previous<br />

commentary. Such a space enables interactions with the edi<strong>to</strong>rs and with readers and<br />

provides a generative, previously unavailable, context for critical responses. The dynamic<br />

interplay of the audience, the original writer who inscribes the marks, and the edi<strong>to</strong>rs<br />

communicating those marks <strong>to</strong> posterity is thereby more likely <strong>to</strong> open what Dickinson<br />

herself would call "doors and windows of possibility." In turn, these myriad perspectives<br />

can enable a much more sustained reflection on how our critical findings are produced<br />

and then transmitted, on the mechanisms of authorization and critical review, and on the<br />

criteria for authenticity. These assessments can then begin <strong>to</strong> penetrate critical mystiques<br />

in ways likely <strong>to</strong> expand rather than restrict knowledge, and <strong>to</strong> focus attention more on<br />

the knowledge itself than on the individual responsible for bringing it <strong>to</strong> the fore. Access<br />

<strong>to</strong> such knowledge can in turn foster a variety of new co-edi<strong>to</strong>rial collaborations among<br />

authors, edi<strong>to</strong>rs, and readers. Digital surrogates alone make definitive analytical<br />

descriptions, on which readers of scholarly editions have depended, neither possible nor


desirable. Instead, such analytical descriptions accompany images that readers can<br />

examine, make judgments about, and then use <strong>to</strong> assess the edi<strong>to</strong>rial commentary. The<br />

Dickinson Electronic Archives is complementing the Open Critical Review database by<br />

importing and making available <strong>to</strong> users the Virtual Lightbox<br />

, a software <strong>to</strong>ol via which images can be<br />

compared and evaluated online, and the Versioning Machine<br />

, a software <strong>to</strong>ol designed by a<br />

team of programmers, designers, and literary scholars for displaying and comparing<br />

multiple versions of texts. Both products are open source and have been produced and<br />

maintained at Maryland Institute for Technology in the Humanities (MITH)<br />

. The database and the sophisticated online <strong>to</strong>ols for<br />

comparing images and texts make it possible for users not only <strong>to</strong> examine the edi<strong>to</strong>rial<br />

process, but even <strong>to</strong> become part of that process. Such collaborations are ways of turning<br />

the edi<strong>to</strong>rial process inside out in order <strong>to</strong> advance its integrity, its scope, and its reach, as<br />

is the choreography of guest co-edi<strong>to</strong>rs working with a group of general edi<strong>to</strong>rs.<br />

On multivolume editions of a major author's work, guest co-edi<strong>to</strong>rs have worked under a<br />

rubric framed by (a) presiding edi<strong>to</strong>r(s), but electronic editions provide new opportunities<br />

for more extensive knowledge transfer and diverse yet coherent knowledge production in<br />

electronic scholarly editions. As do their print predecessors, electronic co-edi<strong>to</strong>rs act as<br />

second and third readers, the first readers being the writer at work on creating text <strong>to</strong> be<br />

disseminated and the general edi<strong>to</strong>r conceiving of an edition appropriately designed for<br />

the electronic field. Since not simply the originating writer but also edi<strong>to</strong>rs are creative<br />

agents, coordinating groups of qualified edi<strong>to</strong>rs <strong>to</strong> work on a vast body of work rather<br />

than relying on a single master edi<strong>to</strong>r or master pair or trio of edi<strong>to</strong>rs <strong>to</strong> bring a scholarly<br />

edition in<strong>to</strong> being practically insures, as do the disambiguating codes of TEI-conformant<br />

XML, greater integrity in textual production.<br />

As we have seen, new materialities of editing, the fact that multimedia representations<br />

make it possible for readers <strong>to</strong> examine documentary evidence previously hidden from<br />

their view, that dynamic databases make it possible for critical readers <strong>to</strong> record and s<strong>to</strong>re<br />

their feedback in a manageable format, and that the electronic work produced by edi<strong>to</strong>rs<br />

has been instantiated in forms much more manipulable than the rigid freeze of print make<br />

innovations in the work of editing itself possible. Teams of edi<strong>to</strong>rs, rather than a solitary<br />

master with her assistants, can work on projects as never before possible. However,<br />

multiply edited electronic editions run the risk of inducing a kind of interpretative vertigo<br />

in the perspective of readers. After all, though the objective is no longer <strong>to</strong> produce the<br />

definitive, immutable text, electronic editions do aim <strong>to</strong> produce greater understandings.<br />

As readers' critical responses are managed through an online database, so edi<strong>to</strong>rs' work<br />

can be managed via a database-driven edi<strong>to</strong>rial submission form such as that provided by<br />

the Dickinson Electronic Archives ,<br />

with its guidelines for encoding, statement of<br />

edi<strong>to</strong>rial principles, and documentation with links <strong>to</strong> regularization databases, lists of<br />

hands identified on the Dickinson manuscripts, lists of manuscript reposi<strong>to</strong>ries, and so<br />

forth.


Just as any edi<strong>to</strong>r has <strong>to</strong> take in<strong>to</strong> account authorial intentions, and any reader is wise <strong>to</strong><br />

consider both the writer's and subsequent edi<strong>to</strong>rs' authorial intentions, multiple edi<strong>to</strong>rs all<br />

working on parts of the same textual body must take in<strong>to</strong> account the intentions of their<br />

co-edi<strong>to</strong>rs and collabora<strong>to</strong>rs, those readers willing <strong>to</strong> commit time and energy and <strong>to</strong><br />

abide by the principles established for scholarly textual production. Such forced<br />

engagements with the sociologies of intention that frame and inhere in any and all textual<br />

productions are immensely valuable for creating edi<strong>to</strong>rial environments that are not only<br />

more trustworthy but that are bound <strong>to</strong> advance critical understandings, teaching scholars<br />

<strong>to</strong> ask questions here<strong>to</strong>fore unimagined. As I noted in an earlier essay, Lucy Suchman<br />

makes insightful observations about the conditions necessary for optimizing knowledge<br />

production. Instead of viewing the "objective knowledge" proffered by a critical edi<strong>to</strong>n<br />

"as a single, asituated, master perspective that bases its claims <strong>to</strong> objectivity in the<br />

closure of controversy", "objective knowledge" in the production of a dynamic critical<br />

edition online can more easily be seen as "multiple, located, partial perspectives that find<br />

their objective character through ongoing processes of debate." Since critical vision is<br />

parallactic rather than unidimensional, the processes of comparing and evaluating those<br />

different angles of seeing as one compares and evaluates different images or different<br />

perspectives of the same images is essential in order <strong>to</strong> see more clearly and accurately.<br />

The locus of objectivity is not "an established body of knowledge… produced or owned<br />

by anyone", but "knowledges in dynamic production, reproduction and transformation,<br />

for which we are all responsible." By contrast, the hieratic models of the master edi<strong>to</strong>rial<br />

perspective do not acknowledge how "layered and intertwined" are the "relations of<br />

human practice and technical artifact" and how such individualistically driven<br />

productions can tend <strong>to</strong> obstruct rather than facilitate intellectual connections, treating<br />

edi<strong>to</strong>rial and critical works as "finished… achievements" rather than as ongoing research<br />

activities and part of a "process of accretion" of edi<strong>to</strong>rial technique and knowledge, part<br />

of midrash, as it were (Suchman, website).<br />

Lessons learned by the makers and users of electronic scholarly editions can help answer<br />

the call of the late John D'Arms and others <strong>to</strong> "move the scholarly monograph in<strong>to</strong> the<br />

second generation of the <strong>digital</strong> age", as the electronic scholarly edition has moved in<strong>to</strong><br />

the first generation. As President of the MLA, Stephen Greenblatt issued a letter <strong>to</strong> the<br />

membership about the plight of scholarly publishing in the <strong>humanities</strong> and a call <strong>to</strong> action<br />

<strong>to</strong> address the "systemic, structural, and at base economic" crises confronting twentyfirst-century<br />

<strong>humanities</strong> scholars committed <strong>to</strong> the free flow of information that enables<br />

knowledge building ("Call for Action on Problems in Scholarly Book Publishing"<br />

); as President of the Association for Computers and the<br />

Humanities and Chair of the TEI Board of Direc<strong>to</strong>rs, John Unsworth answered<br />

Greenblatt's call in presentations such as "The Emergence of Digital Scholarship: New<br />

Models for Librarians, Scholars, and Publishers"<br />

and in his work on the MLA<br />

Committee on Scholarly Editions and the TEI.<br />

In order <strong>to</strong> move the scholarly monograph in<strong>to</strong> the <strong>digital</strong> realm, humanists need <strong>to</strong><br />

embrace the new technologies developing in <strong>digital</strong> <strong>humanities</strong> communities of practice,<br />

technologies making not only new work but new ways of working possible, especially


those that will not become obsolete (access, multimedia study objects, collaboration, selfconsciousness,<br />

audience).<br />

Thus <strong>digital</strong> <strong>humanities</strong> needs <strong>to</strong> move from producing primarily the scholarly archive<br />

(first generation of electronic literary projects) <strong>to</strong> also producing <strong>digital</strong> monographs or<br />

multiply (and <strong>to</strong> varying degrees) authored works – polygraphs. There is no reason that<br />

tightly focused critical inquiry cannot be produced by a group. Revealing and practical<br />

questions about reproductive fidelity – such as "what degree of imaging accuracy is<br />

needed for critical inquiry?" and "who should be responsible for maintaining and making<br />

available the highest quality copies?" – are crucial for such editions and have expanded<br />

group efforts. Initiatives such as the LEADERS project, which "aims <strong>to</strong> enhance remote<br />

user access <strong>to</strong> archives by providing the means <strong>to</strong> present archival source materials within<br />

their context", ensure that archivists best trained <strong>to</strong> maintain both the actual and the<br />

surrogate archival copies assume the responsibilities for doing so and work with scholarly<br />

edi<strong>to</strong>rs best trained <strong>to</strong> organize the data critically. That they can work remotely, as do<br />

multiple guest co-edi<strong>to</strong>rs discussed earlier in this chapter, can be imagined, and perhaps<br />

achieved, only in the <strong>digital</strong> realm of information organization and exchange. Similarly,<br />

<strong>to</strong> develop <strong>digital</strong> monographs the most basic question – "what lines of critical argument<br />

are possible only in a <strong>digital</strong> monograph?" – needs <strong>to</strong> be posed repeatedly. Certainly,<br />

there are adept ways of using sound and image, as well as linguistic codes, that are only<br />

possible in the <strong>digital</strong> realm, but that is only the beginning. Moving from edi<strong>to</strong>r (author<br />

surrogate) and author <strong>to</strong> reader (including edi<strong>to</strong>r and author), from enacting definitude<br />

(editing for the static printed page) <strong>to</strong> enacting fluidity (the dynamic screen), is enabling<br />

profound innovations in edi<strong>to</strong>rial praxes, changes demonstrating how vital are "recent<br />

moves <strong>to</strong> reframe objectivity from the epistemic stance necessary <strong>to</strong> achieve a definitive<br />

body of knowledge, <strong>to</strong> a contingent accomplishment of dynamic processes of knowing<br />

and acting" for enriching our intellectual commons (Suchman, website). Acknowledging<br />

the fluidity of texts instead of insisting upon single-minded, singularly-oriented texts,<br />

"learning the meaning of the revision of texts", as well as the revision of our edi<strong>to</strong>rial<br />

practices, creates an environment in which a "new kind of critical thinking based on<br />

difference, variation, approximation, intention, power, and change" can flourish and work<br />

for the common good. If we are <strong>to</strong> shed the inertia bequeathed from the bibliographic<br />

world, edi<strong>to</strong>rial integrity and fidelity created within and upon the "shifting sands of<br />

democratic life" demand a "new cosmopolitanism" in scholarly editing (Bryant 2002:<br />

177), adopting the "lesbian rule" of principled accommodation for <strong>digital</strong> <strong>humanities</strong>.<br />

References for Further Reading<br />

Bernstein, Charles (2000). The Art of Immemorability. In Jerome Rothenberg and Steven<br />

Clay (eds.), A Book of the Book: Some Works and Projections about the Book and<br />

Writing (pp. 504–17). New York: Granary Books.<br />

Bornstein, George and Theresa Tinkle, (eds.) (1998). The Iconic Page in Manuscript,<br />

Print, and Digital Culture. Ann Arbor: University of Michigan Press.


Bowker, Geoffrey C. and Susan Leigh Star (1999). Sorting Things Out: Classification<br />

audits Consequences. Cambridge, MA, and London: MIT Press.<br />

Bryant, John (2002). The Fluid Text: A Theory of Revision and Editing for Book and<br />

Screen. Ann Arbor: University of Michigan Press.<br />

Condron, Frances, Michael Fraser, and Stuart Sutherland (<strong>2001</strong>). Digital Resources for<br />

the Humanities. Morgan<strong>to</strong>wn: West Virginia University Press.<br />

D'Arms, John H. (2000). The Electronic Monograph in the 21st Century. Based on<br />

Scholarly Publishing in the 21st Century, presented at the American His<strong>to</strong>rical<br />

Association, Chicago, January, 2000. Available at http://www.acls.org/jhd-aha.htm.<br />

Derrida, Jacques (1996). Archive Fever: A Freudian Impression. Chicago and London:<br />

University of Chicago Press.<br />

Drucker, Johanna (1998). Figuring the Word: Essays on Books, Writing, and Visual<br />

Poetics. New York: Granary Books.<br />

Finneran, Richard J., (ed.) (1996). The Literary Text in the Digital Age. Ann Arbor:<br />

University of Michigan Press.<br />

Flanders, Julia (1998). Trusting the Electronic Edition. Computers and the Humanities<br />

31: 301–10.<br />

Franklin, R. W., (ed.) (1998). The Poems of Emily Dickinson: Variorum Edition, 3 vols.<br />

Cambridge, MA, and London: Harvard University Press. (References <strong>to</strong> texts in this<br />

edition will use "FP" and refer <strong>to</strong> the poem's number.).<br />

Grigely, Joseph (1995). Texualterity: Art, Theory, and Textual Criticism. Ann Arbor:<br />

University of Michigan Press.<br />

Hamill, Sam. "Open Letter" on "Who We Are", Poets Against the War. Accessed spring<br />

2003. Available at http://poetsagainstthewar.org.<br />

Hockey, Susan (2000). Electronic Texts in the Humanities. Oxford and New York:<br />

Oxford University Press.<br />

Hockey, Susan, Allen Renear, and Jerome McGann. What is Text? A Debate on the<br />

Philosophical Nature of Text in the Light of Humanities Computing Research. Accessed<br />

spring 2003. Available at<br />

http://www.<strong>humanities</strong>.ualberta.ca/Susan_Hockey/achallc99.htm.<br />

Kirschenbaum, Matthew (guest ed.) (2002). Image-based Humanities Computing. Special<br />

issue of Computers and the Humanities 36: 1–140; his introduc<strong>to</strong>ry essay (pp. 3–6).


Lakoff, George (1987). Women, Fire, and Dangerous Things: What Categories Reveal<br />

about the Mind. Chicago and London: University of Chicago Press.<br />

Landow, George (1997). Hypertext 2.0: The Convergence of Contemporary Critical<br />

Theory and Technology. Baltimore: Johns Hopkins University Press.<br />

La<strong>to</strong>ur, Bruno and Steve Woolgar (1986). Labora<strong>to</strong>ry Life: The Construction of Scientific<br />

Facts, 2nd edn. Prince<strong>to</strong>n, NJ: Prince<strong>to</strong>n University Press. (Originally published 1979.).<br />

Love, Harold (1993). Scribal Publication in Seventeenth-century England. Oxford:<br />

Clarendon Press.<br />

McGann, Jerome (<strong>2001</strong>). Radiant Textuality: Literature after the World Wide Web. New<br />

York: Palgrave.<br />

McKenzie, D. F. (1999). Bibliography and the Sociology of Texts. Cambridge:<br />

Cambridge University Press. (1986) London: British Library. (1985) The Panizzi<br />

Lectures. (1984) The Sociology of a Text: Oral Culture, Literacy and Print in early New<br />

Zealand. The Library 6 (December 1984).<br />

MLA Committee on Scholarly Editions. Revision Plan for the Guidelines for Scholarly<br />

Editions of the Committee on Scholarly Editions. Accessed spring 2003. Available at<br />

http://www.mla.org.www_mla_org/reports/reports_main.asp?. Guidelines for Electronic<br />

Scholarly Editions: Available at http://sunsite.berkeley.edu/MLA/guidelines.html.<br />

Guidelines for Edi<strong>to</strong>rs of Scholarly Editions: Available at<br />

http://jefferson.village.virginia.edu/~jmu2m/cse/CSEguidelines.html.<br />

Rockwell, Geoffrey (2002). Gore Galore: Literary Theory and Computer Games. In<br />

Siemens (ed.), A New Computer-Assisted Literary Criticism? (pp. 345–58).<br />

Schreibman, Susan (2002a). Computer-mediated Texts and Textuality: Theory and<br />

Practice. In R. G. Siemens (ed.), A New Computer-Assisted Literary Criticism? (pp. 283–<br />

93).<br />

Schreibman, Susan (2002b). The Text Ported. Literary and Linguistic Computing 17: 77–<br />

87.<br />

Siemens, Raymond G. (guest ed.) (2002). A New Computer-assisted Literary Criticism?<br />

Special issue of Computers and the Humanities 36: 255–378; his introduc<strong>to</strong>ry essay (pp.<br />

259–67).<br />

Shillingsburg, Peter L. (1996). Scholarly Editing in the Computer Age: Theory and<br />

Practice. Ann Arbor: University of Michigan Press.


Smith, Martha Nell (2002). Computing: What's American Literary Study Got <strong>to</strong> Do with<br />

IT? American Literature 74: 833–57. Online. Project Muse. Accessed spring 2003.<br />

Available at http://muse.jhu.edu/journals/american_literature/v074/74.4smith.html.<br />

Sperberg-McQueen, C. M. and L. Burnard, (eds.) (1990). Guidelines for Electronic Text<br />

Encoding and Interchange: TEI P3. Chicago and Oxford: The Text Encoding Initiative.<br />

(Reprinted 1999.) Available at http://www.tei-c.org/Guidelines/index.htm.<br />

Sperberg-McQueen, C. M., Claus Huitfeldt, and Allen Renear (2000). Meaning and<br />

Interpretation of Markup. Markup Languages 2: 215–34.<br />

Suchman, Lucy. Located Accountabilities in Technology Production. Published by the<br />

Department of Sociology, Lancaster University. Online. Accessed Spring 2003.<br />

Available at http://www.comp.lanes.ac.uk/sociology/soc0391s.html.<br />

Unsworth, John (1996). Electronic Scholarship; or, Scholarly Publishing and the Public.<br />

In R. J. Finneran (ed.), The Literary Text in the Digital Age (pp. 233–43). Ann Arbor:<br />

University of Michigan Press.<br />

Unsworth, John. Electronic Textual Editing and the TEI. Online. Accessed spring 2003.<br />

Available at http://jefferson.village.virginia.edu/~jmu2m/mla-cse.2002.html.<br />

Winder, William (2002). Industrial Text and French Neo-structuralism. In R. G. Siemens<br />

(ed.), A New Computer-Assisted Literary Criticism? (pp. 295–306).<br />

23.<br />

Textual Analysis<br />

John Burrows<br />

Preamble<br />

The object of this paper is <strong>to</strong> show that computer-assisted textual analysis can be of value<br />

in many different sorts of literary inquiry, helping <strong>to</strong> resolve some questions, <strong>to</strong> carry<br />

others forward, and <strong>to</strong> open entirely new ones. The emphasis will not be on the<br />

straightforward, albeit valuable, business of gathering specimens of chosen phenomena<br />

for close study – the business of concordances and tagged sets. It will fall rather on the<br />

form of computational stylistics in which all the most common words (whatever they<br />

may be) of a large set of texts are subjected <strong>to</strong> appropriate kinds of statistical analysis.<br />

Statistical analysis is necessary for the management of words that occur <strong>to</strong>o frequently <strong>to</strong><br />

be studied one by one. But why study such words at all? For those of a sufficiently<br />

ascetic taste, there is a certain intrinsic interest in considering why Henry Fielding should<br />

stand out among a host of others in his recourse <strong>to</strong> the plain man's very and the lawyer's


inferential use of the conjunction for. And if the substance of Aphra Behn's love poetry is<br />

epi<strong>to</strong>mized in her fondness for warm and soft, her slightly strident <strong>to</strong>ne of voice can be<br />

heard in all and none and never. On a larger scene, my datasets uphold the belief that<br />

British writers still make more use of which and should than their American or Australian<br />

contemporaries. Australian writers, however, show a preference for we/our/us in contexts<br />

where their British and American counterparts would favor limy/me. A desire <strong>to</strong> avoid<br />

any appearance of immodesty or a desire for concealment within the herd?<br />

But the real value of studying the common words rests on the fact that they constitute the<br />

underlying fabric of a text, a barely visible web that gives shape <strong>to</strong> whatever is being<br />

said. Despite a brave attempt made long ago (Bratley and Ross 1981), we have yet <strong>to</strong> find<br />

satisfac<strong>to</strong>ry ways of tracing the interconnections of all the different threads, as each of the<br />

common words occurs and recurs in the full sequence of a text. We are able, nevertheless,<br />

<strong>to</strong> show that texts where certain of these threads are unusually prominent differ greatly<br />

from texts where other threads are given more use. An appropriate analogy, perhaps, is<br />

with the contrast between handwoven rugs where the russet <strong>to</strong>nes predominate and those<br />

where they give way <strong>to</strong> the greens and blues. The principal point of interest is neither a<br />

single stitch, a single thread, nor even a single color but the overall effect. Such effects<br />

are best seen, moreover, when different pieces are put side by side. That, at all events, is<br />

the case I shall present.<br />

Classification of Texts<br />

Analyses<br />

To begin directly with the first of several case studies, let us bring a number of poems<br />

<strong>to</strong>gether, employ a statistical procedure that allows each of them <strong>to</strong> show its affinity for<br />

such others as it most resembles, and draw whatever inferences the results admit. For this<br />

purpose, I have taken forty specimens of seventeenth-century and early eighteenthcentury<br />

English verse, as listed in Table 23.1. Many are excerpts from longer works but<br />

each of them exceeds two thousand words in length. While they range in date from 1633<br />

<strong>to</strong> about 1733, most come from 1660–1700. While they run across the range of literary<br />

forms taken by long poems in that era, satire and epic figure prominently among them.<br />

The first twenty on the list are by poets whose other work contributes <strong>to</strong> the large, wideranging<br />

database listed in the Appendix. Seventeen of the second twenty are by poets<br />

whose work is not included there. Although the remaining three poems (nos. 22–24) are<br />

of doubtful authorship, they are usually attributed <strong>to</strong> Andrew Marvell, a member of the<br />

database. If these three are counted, notionally, as Marvell's, twenty of the forty poems<br />

form small authorial sets. The other twenty are independent pieces.<br />

The actual data for analysis consist of word-counts, from each of the forty specimens, of<br />

150 common words. The words themselves are the 150 most common in the<br />

aforementioned database of seventeenth-century verse. Such a list is more robust than<br />

one derived from a small, selective set of texts like the one we shall examine. The texts<br />

were prepared according <strong>to</strong> the same pro<strong>to</strong>cols as the main database. They were all<br />

modernized so as <strong>to</strong> remove the statistically misleading effects of seventeenth-century


spelling. Contracted forms like don't and I've were expanded <strong>to</strong> allow full counts of the<br />

parent words. A few common words like so and that were tagged in such a way as <strong>to</strong><br />

distinguish their main grammatical functions. With this prepara<strong>to</strong>ry work complete, the<br />

texts were subjected <strong>to</strong> our sorting and counting programs. The raw counts for each<br />

common word in each text were normalized as percentages of the <strong>to</strong>tal number of words<br />

in that text. The object is <strong>to</strong> avoid the tyranny of the larger numbers with texts of uneven<br />

length.<br />

The chief analytical procedures employed for comparing texts on the basis of the relative<br />

frequencies of many common words all depend upon the logical principle of concomitant<br />

variation. The counts for each word are treated as variables on which each of the chosen<br />

specimens (the texts under examination) lies nearest <strong>to</strong> such others as it most resembles<br />

and furthest from those it least resembles. To take examples that will easily be<br />

appreciated, texts (like plays) where / and you occur unusually often are likely <strong>to</strong> differ in<br />

many ways from texts (like prose treatises) where the and a rise well above the norm.<br />

Texts of the former sort usually run high in the simpler auxiliary verbs and in many<br />

colloquial, speech-oriented forms of expression. Texts of the latter sort tend <strong>to</strong> be<br />

oriented <strong>to</strong>wards the relationships among things and ideas rather than persons and often<br />

show high frequencies in the most common prepositions. Differences in level of<br />

discourse (as in the wide-ranging contrasts between Latinate forms like ascend and<br />

Germanic equivalents like go up); in genre (as between the past-tense verbs of<br />

retrospective narrative and the present tense of most disquisi<strong>to</strong>ry writing); and in<br />

his<strong>to</strong>rical provenance (as in the shift from thou <strong>to</strong> you) – all these and many more are<br />

reflected in systematic relationships among the frequencies of the very common words.<br />

23.1 Table Forty long poems of the late seventeenth century<br />

*1 2586 Butler, Samuel Hudibras, Can<strong>to</strong> III<br />

#2 2208<br />

Upon the Imperfection … of Human Learning, Parts<br />

I & II. 1–72<br />

*3 3373 Cot<strong>to</strong>n, Charles A Voyage <strong>to</strong> Ireland in Burlesque, Can<strong>to</strong>s 2 and 3<br />

#4 3004 Cowley, Abraham The Plagues of Egypt, 1–262<br />

#5 6812 Davideis, Book II<br />

*6 7824 Dryden, John Absalom and Achi<strong>to</strong>phel<br />

*7 19896 The Hind and the Panther<br />

*8 7817 D'Urfey, Thomas The Malecontent<br />

#9 6020 Gould, Robert To the Society of the Beaux Esprits<br />

#10 4019 The Play-House. A Satyr, Part II<br />

#11 4057 A Satyr against Man, Part I<br />

#12 4492 Presbytery Rough-drawn. A Satyr<br />

*13 27154 Mil<strong>to</strong>n, John Paradise Lost, lines 201–500 of each Book


*14 15694 Paradise Regained<br />

*15 12885 Samson Agonistes<br />

*16 2210 Oldham, John Satyr II<br />

*17 3378 The Eighth Satyr of Monsieur Boileau Imitated<br />

*18 3381 Swift, Jonathan On Poetry: A Rhapsody<br />

*19 3206 Verses in the Death of Dr. Swift, D. S. P. D.<br />

*20 2606 Waller, Edmund Instructions <strong>to</strong> a Painter<br />

#21 3547 Addison, Joseph The Campaign<br />

*22 2867 Anon. The Second Advice <strong>to</strong> a Painter<br />

*23 3638 The Third Advice <strong>to</strong> a Painter<br />

*24 7693 Last Instructions <strong>to</strong> a Painter<br />

*25 11111 Billingsley, Nicholas The World's Infancy<br />

#26 6986 Blackmore, Richard King Arthur, Book I<br />

*27 3892 Caryll, John Naboth's Vineyard<br />

#28 3617 Chamberlayne, Willi Pharonnida, Book I, Can<strong>to</strong> 1<br />

#29 2743 Chudleigh, Mary<br />

On the Death of his Highness the Duke of<br />

Gloucester<br />

#30 5167 Davenant, William Gondibert, Book I, Can<strong>to</strong>s 1 and 2<br />

#31 3892 Duke, Richard Paris <strong>to</strong> Helena<br />

#32 5935 Fletcher, Phineas The Purple Island, Can<strong>to</strong>s I and II<br />

#33 8797 Heyrick, Thomas The New Atlantis, Parts I and II<br />

#34 3731 Parnell, Thomas Homer's Battle of the Frogs and Mice, Books I-III<br />

*35 3103 Pordage, Samuel The Medal Revers'd<br />

*36 3250 Thompson, Thomas Midsummer Moon<br />

*37 2287 Tutchin, John A Search after Honesty<br />

*38 4374 Vaughan, Henry Juvenal's Tenth Satire Translated<br />

*39 2156 Wase, Chris<strong>to</strong>pher Divination<br />

*40 3321 Wild, Robert Iter Boreale<br />

Texts are complege except where specified. Texts marked* are from original or standard<br />

modern editions. Those marked # are from the Chadwyck-Healey Archive of English<br />

Poetry, <strong>to</strong> which my university subscribes. The texts were used only for the extraction of<br />

word-counts.<br />

When any large group of texts is analyzed, such manifestations of concomitant variation<br />

as these yield complex but intelligible patterns of affinity and disaffinity. They are much<br />

enriched, moreover, by the many exceptions that arise. Some of these are associated with<br />

authorial idiosyncrasies. The possible combinations from genre <strong>to</strong> genre, era <strong>to</strong> era,<br />

author <strong>to</strong> author, text <strong>to</strong> text, are almost unlimited. Effects like these, we may suppose,<br />

impinge upon the minds of good readers as part of their overall response. But they are not


easy for a reader <strong>to</strong> articulate. Statistical analysis, on the other hand, offers clear but<br />

comparatively unsubtle ways of representing them.<br />

The main analytical procedure <strong>to</strong> be employed in this chapter is that of cluster analysis,<br />

chosen because it offers rather a harsh test of the questions <strong>to</strong> be considered and also<br />

because the "family trees" in which the results are displayed speak plainly for themselves.<br />

The procedure is put <strong>to</strong> excellent use in a recent study of prose fiction. (See Hoover<br />

2002.) The principal disadvantage of cluster analysis is that the underlying word-patterns<br />

are not made visible and must therefore be approached in other ways. For those who are<br />

versed in these matters, I should add that the statistical package used for these cluster<br />

analyses is MINITAB. My many trials suggest that, for such data as we are examining,<br />

complete linkages, squared Euclidean distances, and standardized variables yield the<br />

most accurate results. This pattern of preferences avoids any undue smoothing of data<br />

whose inherent roughness reflects the complexities of the language itself. This pattern, at<br />

all events, is used throughout.<br />

Figure 23.1 represents the outcome of a cluster analysis in which our forty texts are<br />

compared with each other on the basis of the full list of the 150 most common words. It<br />

should be studied from the base upwards, taking account of the way the clusters form.<br />

The texts are identified, along the horizontal base of the figure, by the numbers attached<br />

<strong>to</strong> them in Table 23.1. The true affinities are not between entries that merely stand beside<br />

each other before separating, like nos. 2 and 4, but between those that form unions, like<br />

nos. 4 and 5. The closest affinities of all are between those pairs that unite soonest, like<br />

nos. 13 and 14, and those trios that do so, like nos. 13, 14, and 15.<br />

The most obvious feature of Figure 23.1 is that some of its members are so slow <strong>to</strong> form<br />

any union at all. The isolation of no. 35, Samuel Pordage's satire The Medal Revers'd, can<br />

readily be shown <strong>to</strong> rest on an unusual preponderance of verbs couched in the present<br />

tense. They reflect the rhe<strong>to</strong>rical stance of an ostensibly philosophic observer of affairs.<br />

No. 3, Charles Cot<strong>to</strong>n's A Voyage <strong>to</strong> Ireland in Burlesque, is isolated by its narrative<br />

mode and its colloquial speech-register. When it does finally form an affinity, it is not<br />

with the epic and heroic poems <strong>to</strong> its left but with the loose group, chiefly made up of<br />

satires, lying <strong>to</strong> its right. And even though nos. 1 and 2, the highly idiosyncratic satires of<br />

Samuel Butler, differ markedly from each other, they stand even further apart from all the<br />

rest.<br />

Considered more broadly, Figure 23.1 shows a rough differentiation between the earlierand<br />

later-born poets, with the work of the latter lying <strong>to</strong>wards the right. Near the righthand<br />

extremity, for example, the poems of Gould, Swift, Duke, and Lady Mary<br />

Chudleigh make comfortable enough chronological neighbors while Butler, Cowley,<br />

Billingsley, and Waller are well placed at the left. But this incipient pattern is severely<br />

damaged by the genre-differentiation already noted. On a. chronological basis,<br />

Blackmore (no. 26), Addison (no. 21) and Parnell (no. 34) have no place on the left while<br />

Vaughan (no. 38) and Phineas Fletcher (no. 32) lie much <strong>to</strong>o far <strong>to</strong> the right. The location<br />

of Mil<strong>to</strong>n's three poems (nos. 13–15) is anomalous in both chronology and genre.


Of the seventeen poems that make up authorial sets, only four fail <strong>to</strong> reveal their true<br />

authorial affinities. The three that are thought <strong>to</strong> be Marvell's form another little set. (In<br />

the case of Robert Gould, whose quartet forms two pairs, the success of the cluster<br />

analysis is incomplete.) Since the members of each authorial pair are locating their proper<br />

partners from a field of thirty-nine, the odds against the chance achievement of such a<br />

success rate are immense. Moreover, as readers who know these forty poems will be<br />

aware, the three anonymous painter-satires (nos. 22–24) are the only set where the effects<br />

of genre and authorship might be thought <strong>to</strong> work entirely in unison. The two Dryden<br />

poems, for instance, have little in common but their authorship; and Samson Agonistes<br />

differs markedly in genre from Mil<strong>to</strong>n's epics.<br />

Figure 23.2 is specifically designed <strong>to</strong> show why the painter-satires form a trio. It is<br />

constructed in exactly the same fashion as Figure 23.1 save that two extra specimens<br />

have been added. They are two of Andrew Marvell's best-known poems, Upon Apple<strong>to</strong>n<br />

House (no. 41) and The First Anniversary of the Government under 0. C. (no. 42). At<br />

4,845 and 3,131 words respectively, they are of appropriate length for the comparison.<br />

And they differ markedly in genre from each other and from the painter-satires. In Figure<br />

23.2, it will be seen that they form an additional authorial pair and that this pair unites<br />

with the trio of painter-satires. As in Figure 23.1, furthermore, the target of the paintersatires,<br />

Edmund Waller's Instructions <strong>to</strong> a Painter stands well away from them. Marvell's<br />

likely authorship of these three painter-satires is a matter of some importance in literary<br />

his<strong>to</strong>ry. It is pursued more thoroughly in a new article <strong>to</strong> be submitted <strong>to</strong> The Yearbook of<br />

English Studies. For our present purpose, however, Figure 23.2 makes a more general<br />

point. With the addition of the two Marvell pieces and the uniting of the Gould quartet,<br />

Figure 23.2 surpasses the strong result achieved in Figure 23.1. The 150 most common<br />

words of our main dataset can clearly be said, therefore, <strong>to</strong> offer a useful basis for testing<br />

the authorship of long poems of the late seventeenth century. The possibility that this<br />

outcome might be a statistical artifact is easily set at rest: other methods of analyzing<br />

these data yield equally accurate results.<br />

If the word list is progressively truncated from the bot<strong>to</strong>m, much shorter poems can be<br />

tested for authorship. Although the rate of success diminishes as the texts grow shorter, it<br />

still far exceeds the rate of chance success. A study of this <strong>to</strong>pic (Burrows 2002a) has<br />

recently been published. But a selective word list can be constructed in other ways. Is it<br />

possible <strong>to</strong> achieve similar levels of success more economically, working with only a<br />

chosen subset of our 150 most common words? After making many attempts, varying in<br />

rigor and subtlety but without conspicuous success, as reported elsewhere (Burrows<br />

2003: 21–3), I chose a plainer path. In Table 23.2, the 150 most common words are<br />

dispersed in three subsets. The 54 words of "domain 1" comprise the principal referential<br />

markers, namely the definite and indefinite articles and the personal pronouns, <strong>to</strong> which<br />

are added the infinitive particle <strong>to</strong> and those common verb-forms whose inflections differ<br />

in association with the articles and pronouns. The 69 words of "domain 2" comprise the<br />

common connectives (including conjunctions, prepositions, and relative pronouns) and<br />

the common intensifiers, whether adjectives or adverbs. The descriptive adjectives,<br />

however, are put aside with the nouns among the 27 words excluded as <strong>to</strong>o lexical, <strong>to</strong>o<br />

subject-oriented, <strong>to</strong> be valid markers of more formal differences. A close student of Table


23.2 will detect at least two cases, art and own, where I have sought the less bad<br />

compromise. The former is included in "domain 1" on the ground that, in seventeenthcentury<br />

verse, its role as an auxiliary verb is more potent than its role as an abstract noun.<br />

The latter is excluded on the ground that it is used less often as a verb than as an<br />

adjective. Such is the wealth of our data that a few arguable choices have little impact on<br />

the analyses.<br />

23.2 Table Experimental classification of the 150 most common words of the<br />

main dataset<br />

"Domain 1":54 words "Domain 2":69 words Exclusions: 27 words<br />

1 the 79 do 2 and 71 there 64 love<br />

4 a 80 should 3 of 72 some 70 great<br />

5 <strong>to</strong> (i) 82 let 6 in (p) 73 <strong>to</strong>o 92 own<br />

7 his 83 make 8 with 74 how 100 first<br />

10 is 84 could 9 <strong>to</strong> (p) 76 one 101 men<br />

12 he 86 must 11 but 77 never 103 man<br />

14 I 87 an 13 all 81 though 106 heaven<br />

15 it 90 us 16 as 85 those 108 good<br />

17 their 93 thee 19 not 88 where 111 well<br />

18 her 94 made 23 for (p) 89 still 114 long<br />

20 be 95 has 24 by (p) 91 here 115 old<br />

21 you 96 see 27 from 97 these 116 day<br />

22 they 122 art 28 that (rp) 98 before 121 heart<br />

25 my 124 give 29 or 99 thus 123 wit<br />

26 we 128 know 33 this 102 every 126 world<br />

30 our 129 find 34 when 104 whose 130 fate<br />

31 thy 131 its 37 at 105 out 132 eyes<br />

32 was 137 been 38 which (rp) 107 much 133 life<br />

35 are 146 come 39 no (aj) 109 while 134 vain<br />

36 your 40 what 110 so (c) 135 power<br />

43 will (v) 41 so (ad) 112 each 139 God<br />

45 can 42 that (d) 113 only 140 soul<br />

46 have 44 on (p) 117 once 141 new<br />

47 she 49 more 118 through 142 fair<br />

48 thou 50 if 119 up (ap) 147 time<br />

51 did 53 now 120 both 148 name<br />

52 would 54 who (rp) 125 till 150 last


"Domain 1":54 words "Domain 2":69 words Exclusions: 27 words<br />

57 had 55 that (c) 127 ever<br />

60 him 56 yet 136 down<br />

66 may 58 then 138 whom<br />

67 shall 59 such 143 upon (p)<br />

68 me 61 nor 144 whilst<br />

69 were 62 for (c) 145 since<br />

75 does 63 like (p) 149 over<br />

78 them 65 than<br />

Abbreviations:<br />

(p) = preposition<br />

(i) = infinitive<br />

(c) = conjunction<br />

(rp) = relative pronoun<br />

(v) = verb<br />

(ap) = adverbial particle<br />

(ad) = adverb of degree<br />

(aj) = adjective<br />

(d) = demonstrative<br />

Figure 23.3 is a cluster analysis of our original set of forty poems using only the 69 words<br />

of "domain 2." As a test of authorship, it offers much the same result as the full list of<br />

150 words used earlier. The members of the Oldham and Swift pairs still stand apart from<br />

their fellows. The Gould poems now fall in<strong>to</strong> a trio and a single<strong>to</strong>n instead of the two<br />

pairs of Figure 23.1 or the quartet of Figure 23.2. The other authorial sets hold firm. On<br />

closer study, however, it can be seen that Figure 23.3 surpasses both its predecessors<br />

because the various groups form much more quickly. For testing the authorship of this<br />

large group of poems, "domain 2", therefore, is more economical of data and more<br />

efficient in operation than the full list of 150 words.<br />

It goes almost without saying that, if these 69 words can match or outmatch the full set,<br />

the omitted words must shed little light on authorship. Figure 20.4 is the counterpart of<br />

Figure 23.3, taking the 54 words of "domain 1" as its basis. Two of the three Mil<strong>to</strong>n<br />

poems, two of the three painter-satires, and two of the four Gould poems still unite with<br />

their respective fellows. But where we had four or five errors out of twenty, we now have<br />

fourteen. Since this pattern is certainly not author-driven, the question is whether other<br />

inferences can be drawn. A close study of the way the poems cluster offers rough but<br />

suggestive distinctions between predominantly monologic, dialogic, narrative, and<br />

reflective forms of rhe<strong>to</strong>ric. Since we are working exclusively with referential markers<br />

and inflected verbs, this seems a plausible result. But the genre-differences in the poetry<br />

of the Res<strong>to</strong>ration era are <strong>to</strong>o subtle and <strong>to</strong>o impure <strong>to</strong> allow any firm conclusions <strong>to</strong> be<br />

drawn from this one analysis.


To pursue the hypothesis that "domain 1" may indeed be genre-driven, I have accordingly<br />

added a further 26 specimens, as listed in Table 23.3 – Some 16 of them (nos. 41–56) are<br />

selections from plays of the period. Most of the plays are represented by a single Act but<br />

there are cases where other considerations call for a choice of scattered scenes. (Both<br />

Rochester and Waller, for example, contributed scenes <strong>to</strong> plays by other dramatists.)<br />

Among the plays, the first ten selections are of verse drama. The next four are of prose<br />

drama and the last two, both from Southerne's The Fatal Marriage, include both verse<br />

and prose. Of the dramatists included, Dryden, Rochester, Stapyl<strong>to</strong>n, Congreve, and<br />

Southerne offer authorial pairs within their respective subsets. Sedley and Shadwell make<br />

it possible <strong>to</strong> test dramatic authorship across the differences between verse and prose.<br />

While these 16 specimens of drama are <strong>to</strong>o few <strong>to</strong> be representative, they may suffice for<br />

an instructive comparison with the poetry and the personal letters of their day.<br />

23.3 Table Twenty-six additional texts<br />

Sixteen excerpts from plays of the late seventeenth century<br />

Verse plays<br />

*41 2918 ?Dorset, Charles Sackville, Earl of Pompey the Great, Act II<br />

*42 7133 Dryden, John The Siege of Granada, Act I.i, Act V.i-ii<br />

*43 4955 Tyrannic Love, Acts I-II<br />

*44 7870 Rochester, John Wilmot, Earl of Valentinian, replacement scenes<br />

*45 2106 The Walls of China, a contributed scene<br />

*46 3177 Sedley, Sir Charles An<strong>to</strong>ny and Cleopatra, Act I.ii, Act III.i<br />

*47 2304 Shadwell, Thomas Psyche, two scenes<br />

#48 7817 Stapyl<strong>to</strong>n, Sir Robert Hero and Leander, Act II<br />

#49 3764 The Slighted Maid, Act III<br />

*50 4980 Waller, Edmund The Maid's Tragedy, replacement scenes<br />

Prose plays<br />

#51 5612 Congreve, William Love for Love, Act IV<br />

#52 5334 The Way of the World, Act IV<br />

#53 5303 Sedley, Sir Charles Bellamira, Act III<br />

*54 6144 Shadwell, Thomas The Libertine, Act III<br />

Prose and verse<br />

*55 3732 Southerne, Thomas The Fatal Marriage, Act III<br />

#56 4082 The Fatal Marriage, Act IV<br />

Ten sets of letters from the late seventeenth century<br />

*57 9615 Conway, Anne Finch, Lady Letters <strong>to</strong> her husband<br />

*58<br />

5058 King, William, Archbishop of<br />

Dublin<br />

Letters <strong>to</strong> Jonathan Swift


*59 5545 Montagu, Lady Mary Wortley<br />

Letters <strong>to</strong> Edward Montagu (before<br />

marriage)<br />

*60 5969 More, Henry Letters <strong>to</strong> Anne, Lady Conway<br />

*61 8113 Osborne, Dorothy<br />

Letters <strong>to</strong> William Temple (before<br />

marriage)<br />

*62 7193 Rochester, John Wilmot, Earl of Letters<br />

*63 6251 Savile, Henry Letters <strong>to</strong> Rochester<br />

*64 8948 Swift, Jonathan<br />

Letters <strong>to</strong> Esther Vanhomrigh<br />

("Vanessa")<br />

*65 8730 Swift, Jonathan Letters <strong>to</strong> Archbishop King<br />

*66 5418 Vanhomrigh, Esther ("Vanessa") Letters <strong>to</strong> Jonathan Swift<br />

Texts are complete except where specified. Texts marked* are from original or standard<br />

modern editions. Those marked # are from the Chadwyck-Healey Archive of English<br />

Poetry, <strong>to</strong> which my university subscribes. The texts were used only for the extraction of<br />

word-counts.<br />

The other ten specimens (nos. 57–66) are selections from the personal letters of some<br />

well-known figures of the period. All the letters within each set except for Rochester's are<br />

directed <strong>to</strong> a particular recipient. Two sets of Swift's letters are included <strong>to</strong> display the<br />

possible diversity of an epis<strong>to</strong>lary reper<strong>to</strong>ire.<br />

Whereas only tentative conclusions could be drawn from Figure 23.4, Figure 23.5 uses<br />

the 54 words of "domain 1" <strong>to</strong> quite striking effect. Some 90 percent of our 66 specimens<br />

form genre-based clusters. All ten sets of personal letters form a cluster of their own, at<br />

the extreme right. Thirty-eight of the forty poems lie <strong>to</strong> the left of any other specimens.<br />

Twenty-nine of them form one large cluster at the left with another seven clustering <strong>to</strong><br />

their immediate right. The rest are located among the verse-plays and are marked as<br />

errors. Two verse plays are marked as errors because they cross in<strong>to</strong> the remaining<br />

cluster, where they join the prose plays and the two plays in which prose and verse are<br />

mixed.<br />

From another, simpler perspective, Figure 23.5 is composed of three large clusters. To<br />

the left is a set made up entirely of poems. In the middle is a set uniting some poems of a<br />

dialogic cast with verse drama, obviously their close kin. To the right is a set of sixteen<br />

prose specimens and two stray pieces of verse drama (nos. 48 and 49). These last, both<br />

extracts from little-known plays by Sir Robert Stapyl<strong>to</strong>n, are couched in what has aptly<br />

been described (Sutherland 1969: 43) as "an easy and colloquial blank verse." Why do<br />

they cross the border between verse and prose? The best explanation, I believe, is that<br />

Stapyl<strong>to</strong>n favors a dialogue of brief, deictic interchange and only rarely offers a large,<br />

poetic set-speech. In an analysis based principally upon pronouns and auxiliary verbs,<br />

this tendency carries his plays <strong>to</strong>wards the prose drama of the period.


That explanation is supported by Figure 23.6, where the full list of 150 words is<br />

employed in an analysis of the same 66 specimens. The richer word list, less strongly<br />

influenced by I/thou, places Stapyl<strong>to</strong>n correctly in the cluster composed entirely of verse<br />

drama. No. 45, Rochester's highly poetic contribution <strong>to</strong> Sir Robert Howard's verse drama<br />

The Walls of China, is now misplaced among the 37 poems that make up the main<br />

cluster. Three poems (nos. 31, 35, and 3), which have previously shown little affinity for<br />

the others, are now misplaced among the plays. Every other entry is correctly placed in a<br />

tripartite distribution of the texts according <strong>to</strong> their genre.<br />

The foregoing series of analyses shows many points of interest. One of the more<br />

unexpected offers special food for thought. When we compared Figure 23.1 and Figure<br />

23.3, we saw that the selective word list, though scarcely less accurate than the full list,<br />

was more economical and more efficient. In the corresponding case, Figure 23.5 does<br />

offer a somewhat less accurate separation of literary genres than Figure 23.6. Yet Figure<br />

23.5 is much more economical than Figure 23.6 in its use of data and its clusters form<br />

more speedily. If the 69 words of "domain 2" and the 54 words of "domain 1" compare so<br />

favorably with the full list of 150 words as indica<strong>to</strong>rs of authorship and genre<br />

respectively, we do well <strong>to</strong> seek an explanation.<br />

Rationale<br />

As a first step, the case for excluding the most common lexical words from such analyses<br />

because they are <strong>to</strong>o subject-oriented is well unders<strong>to</strong>od and widely practiced. Even in<br />

the upper reaches of the frequency hierarchy, the lexical words favored by different<br />

authors are sometimes idiosyncratic, sometimes those of a coterie. The Christian<br />

vocabulary of seventeenth-century devotional verse stands out in contrast, for example, <strong>to</strong><br />

the obscene vocabulary that marks the clandestine verse of Rochester and his circle: some<br />

words of both kinds rank very high in appropriate texts. We do not need a computer <strong>to</strong><br />

determine that the obscene farce, Sodom, is unlikely <strong>to</strong> be the work of Henry Vaughan.<br />

And even if it were, a quite uncharacteristic set of lexical words would act only as a<br />

disguise. Since such problems cannot validly be addressed by picking and choosing<br />

among the lexical words, it is best <strong>to</strong> treat them all alike and <strong>to</strong> recognize what one is<br />

doing. And, though the boundary between lexical and function words is inexact, the<br />

upper end of the frequency range includes very few hard choices. On this basis, 27 words<br />

can be excluded from our list of 150.<br />

A good case can also be made for excluding the personal pronouns because they are <strong>to</strong>o<br />

volatile in frequency <strong>to</strong> serve as reliable measures. If that is done, the question of the<br />

inflected verbs arises with some force. It would obviously be logical <strong>to</strong> exclude them <strong>to</strong>o.<br />

But by now the increasing sacrifice of information must occasion serious concern. These<br />

three excisions would take out 81 words from the 150 of the list we have been using. A<br />

desirable rigor would be obtained at a disturbing cost.<br />

The results already offered illustrate what follows when the 27 most common lexical<br />

words are set aside while the common function-words are all retained but are separated<br />

in<strong>to</strong> two "domains." Figure 23.7 represents an attempt <strong>to</strong> understand the difference


etween these notional domains. Anybody who sets about a piece of writing may be<br />

thought of as beginning at the standpoint marked at the extreme left of the figure. Even at<br />

that point, such a person is <strong>to</strong>o much a creature of habit and background <strong>to</strong> begin with a<br />

clean slate. Even at that point, such a person will also have some notion of a subject, an<br />

intended audience, and an attitude <strong>to</strong> that audience. And, although the writer retains a<br />

good deal of latitude as the work proceeds, these broad constraints begin <strong>to</strong> strengthen<br />

from the moment the very first words are written: "Of Man's first disobedience … "; "The<br />

glories of our blood and state… "; "It is a truth universally acknowledged… "; "She<br />

waited, Kate Croy… "; "My father had a small estate in Nottinghamshire; I was the third<br />

of five sons… "; "When I consider how my light is spent… "; "I wandered lonely as a<br />

cloud… "; "Shall I compare thee <strong>to</strong> a summer's day… "; "Believe't, I will.//Thy worst. I<br />

fart at thee."<br />

The authorial I of a text is ever-present though not always visible. That I, of course,<br />

cannot simply be identified with the writer in propria persona. But four of the many roles<br />

open <strong>to</strong> that I govern the categories set out in the upper part of Figure 23.7. In category<br />

A, an I not easily distinguished from the author engages in some form of direct address <strong>to</strong><br />

a real or putative reader. Such work need not be disingenuous. Samuel Johnson offers a<br />

celebrated warning about the deceits of epis<strong>to</strong>lary correspondence and when Lady Mary<br />

Pierrepont tells her future husband, Edward Wortley Montagu, "I am going <strong>to</strong> write you a<br />

plain long letter" (Halsband 1970: 41), only the warning signs are plain. (He did not heed<br />

them.) The I of category A usually occurs freely and carries a select group of verbal<br />

inflections along with it. The thou/you who is addressed may occur just as freely but may<br />

also remain inexplicit. T. S. Eliot distinguishes the pulpit-rhe<strong>to</strong>ric of Donne from that of<br />

Lancelot Andrewes on pertinent lines: whereas the former is always in his pulpit,<br />

addressing his congregation, the latter "is wholly in his subject, unaware of anything<br />

else" (Eliot 1932: 341). While the comparative frequencies of the relevant pronouns,<br />

articles, and verbs in their sermons are not what engages a reader, they show the force of<br />

Eliot's distinction.<br />

In category B, both the authorial I and the you of the addressees are absorbed in<strong>to</strong> the we<br />

of a real or notional group. They is assigned <strong>to</strong> outsiders, whether feared, pitied, or<br />

admired. When these plural forms predominate, the singular pronouns and verbs all drop<br />

away in frequency. Wins<strong>to</strong>n Churchill's "We will fight them on the beaches … We will<br />

never surrender" takes the horta<strong>to</strong>ry form. The elegiac form is taken by Lawrence<br />

Binyon's Armistice hymn, For the Fallen: "They shall grow not old / As we that are left<br />

grow old."<br />

In category C, the authorial I is displaced by the deictic I of whichever character,<br />

dramatic or fictional, is speaking at a given time. In plays, accordingly, Thou/you is<br />

reserved for the characters then addressed. All of the personal pronouns are free <strong>to</strong> occur<br />

but I/thou/you usually predominate. The present tense verb-forms are usually in the<br />

ascendant. In the retrospective narrative "his<strong>to</strong>ries" that either constitute or are<br />

interpolated in many novels, especially those of the eighteenth and nineteenth centuries, I<br />

is cus<strong>to</strong>marily the highly visible property of the present speaker. (Whereas Lemuel<br />

Gulliver always holds the center of his stage, Henry Fielding's Man of the Hill and Mrs


Fitzpatrick or his Miss Mathews, Captain Booth, and Mrs Bennet take turns at being /.)<br />

Thou/you may either be reserved for fellow characters or extended <strong>to</strong> include us "Gentle<br />

Readers." Both he and she are likely <strong>to</strong> occur almost as freely as /, being devoted <strong>to</strong> those<br />

others who <strong>to</strong>ok part in the events now recounted. Except in those retrospective narratives<br />

where the his<strong>to</strong>ric present tense is adopted as a rhe<strong>to</strong>rical strategy, most verbs are<br />

couched in the past tense.<br />

In the many and varied forms of writing composing category D, neither I nor you need<br />

ever occur, most other pronouns may be rare, and the articles are likely <strong>to</strong> be the chief<br />

referential markers. Among the third-person pronouns, it and they are likely <strong>to</strong><br />

predominate. And though his<strong>to</strong>rical writings are often more personally oriented than most<br />

treatises and disquisitions, they do not often find much use for the feminine pronouns. It<br />

might be interesting <strong>to</strong> ascertain whether a flurry of feminine pronouns could be found in<br />

any part of Gibbon's great his<strong>to</strong>ry except the chapter treating the notable exploits of the<br />

Empress Theodora. She is certainly one of the only two women named in the extensive<br />

titles of his 71 chapters. Modern feminist his<strong>to</strong>rians are busy redressing the imbalances of<br />

the past, and altering the incidence of pronouns as they do so.<br />

With Henry Fielding's novels among my examples, Gentle Reader, I could not easily<br />

forget that many of the literary forms I have referred <strong>to</strong>, and the novel in particular, can<br />

draw on the resources of more than one of my categories. When this occurs, of course,<br />

the pattern of pronouns, articles, and auxiliary verbs is bound <strong>to</strong> be modified. My short<br />

list of distinguishable literary forms is only a scanty representation of the more obvious<br />

classes. And I am well aware that, in broadening its field of reference in recent years, the<br />

word genre has lost its former rigor. Its over-extension can be seen when it is applied, for<br />

example, <strong>to</strong> personal letters, a whole family of loosely related literary forms.<br />

With all due allowance for these reservations, Figure 23.5 still indicates that some major<br />

genre-divisions can indeed be differentiated by the statistical analysis of a select word list<br />

confined <strong>to</strong> pronouns, inflected verbs, articles, and the infinitive particle. This last acts as<br />

a marker of complex verbal compounds, often couched in the first person, in which the<br />

speaker stands back a little from the action encompassed in the main verb and expresses a<br />

self-conscious attitude <strong>to</strong> it. "I hoped/feared/wished/chose <strong>to</strong>…" and many other forms<br />

like them typify a usage which is remarkably common in some writers, Butler and<br />

Rochester among them, and which offers us an easy transition from domain 1 <strong>to</strong> domain<br />

2.<br />

From the standpoint at the left of Figure 23.7, as a writer moves forward with the<br />

emerging text, the choices made in domain 1 interweave with those of domain 2. Here,<br />

however, the more open choices of domain 1 are outweighed though not extinguished by<br />

the constraints of upbringing and habit. Predisposition or idiosyncrasy? The syntactical<br />

forms by which a writer's ideas are connected and the forms of emphasis a writer chooses<br />

all offer a new battleground for ancient debates about the nexus between idiosyncratic<br />

and socially influenced facets of our behavior. If social fac<strong>to</strong>rs were all that mattered, it<br />

would be hard <strong>to</strong> explain why men of such different backgrounds as Defoe and Henry<br />

Fielding should far outrank their contemporaries in the use of so commonplace a word as


very. If idiosyncrasy were all that mattered, it would be hard <strong>to</strong> explain why, nearly three<br />

centuries later, that same word should act as a statistically significant differentia between<br />

a large set of Australian writers and their British and American fellows. Contrasts in<br />

socially imposed roles may well explain why not, but, and never are statistically<br />

significant differentiae between large sets of female and male writers of the eighteenth<br />

and nineteenth, but not the twentieth centuries. They do not explain why Dryden and<br />

Gould should stand out among their contemporaries in their use of before and there<br />

respectively.<br />

Whether they originate in idiosyncrasy or in upbringing, such authorial habits are strong<br />

enough and consistent enough <strong>to</strong> allow the 69 words of domain 2 <strong>to</strong> act as effective<br />

authorial discrimina<strong>to</strong>rs among the forty poems of Figure 23.3 – Although cluster<br />

analysis rests upon combinations of many word-frequencies, it is possible <strong>to</strong> pick out<br />

some of the more striking. Rochester is much given <strong>to</strong> who and whose and ever, Marvell<br />

<strong>to</strong> yet and where and only. Dryden and Congreve make uncommonly frequent use of thus.<br />

Of while and whilst, Marvell and Dryden lean <strong>to</strong> the former whereas Aphra Behn prefers<br />

the latter. Whereas both Dryden and Mil<strong>to</strong>n prefer on <strong>to</strong> upon, Samuel Butler goes the<br />

other way. As these last examples suggest, the words <strong>to</strong> which particular writers have<br />

more recourse than most are matched, as discrimina<strong>to</strong>rs, by words that they eschew.<br />

Provided the texts called in<strong>to</strong> comparison are not <strong>to</strong>o dissimilar in kind, such preferences<br />

among the common connectives and modifiers that constitute our domain 2 (or the<br />

corresponding "domain 2" of any other large database) can go far <strong>to</strong>wards establishing<br />

the likely authorship of doubtful texts. And the notion of a writer finding a path forward<br />

by intermingling choices and habits helps <strong>to</strong> explain why this should be so. As the writer<br />

advances, the emerging shape of the text makes the basis of the stylistic corollaries and<br />

consequences indicated at the extreme right of Figure 23.7. Choices, constraints, further<br />

constraints following from each choice – and always some room for accident, whether<br />

inept or inspired.<br />

But Figure 23.8 is an abrupt reminder that a writer's habits are not set in s<strong>to</strong>ne. It<br />

represents the outcome of a cluster analysis of the 66 texts of Table 23.1 and Table 23.3,<br />

using the 69 words of domain 2. Some 38 of the 66 texts are eligible as members of little<br />

authorial subsets: only 17 of them find their partners. Those 17 almost exactly match the<br />

successful unions of Figure 23.3 and include, in addition, the two specimens of Dryden's<br />

verse drama (nos. 42 and 43). Samson Agonistes (no. 15) has moved away from the other<br />

Mil<strong>to</strong>n entries, but the previously misplaced Gould entry (no. 9) has rejoined the other<br />

three. Dryden's verse drama apart, the eligible specimens from among the additional texts<br />

not only fail <strong>to</strong> achieve a single cross-genre union but even fail <strong>to</strong> find their partners<br />

within the same genre. One interesting failure lies in Swift's two sets of letters (nos. 64–<br />

5), whose strikingly different recipients may explain why he can adopt such different<br />

styles. The two Acts from the same tragedy by Thomas Southerne (nos. 55–6) are<br />

distinguished from each other by their different proportions of subplot and main plot,<br />

prose and verse. The two specimens of Congreve's prose drama (nos. 51–2) are<br />

distinguished from each other by this dramatist's keen ear for different stylistic registers.<br />

The aris<strong>to</strong>crats of The Way of the World (no. 52) converse in a manner more akin <strong>to</strong> that


of their peers in Sedley's Bellamira (no. 53) than that of the (not ill-spoken) cits and<br />

squires in Love for Love (no. 51).<br />

But the strongest general contrast in Figure 23.8 is between prose and verse. Three<br />

specimens of verse drama by Rochester (no. 44), Sedley (no. 46), and – as it is thought –<br />

by Dorset (no. 41) lie among the large cluster of poems at the left. Another by Rochester<br />

(no. 45) lies among the poems at the right. In the middle of the whole group, ranging<br />

across from no. 52 <strong>to</strong> no. 65, the sixteen prose specimens form a slightly disordered but<br />

unbroken sequence. Even in domain 2, so it appears, strong differences of genre can<br />

smother authorial resemblance.<br />

There is matter here for a searching, full-scale inquiry. Even a cursory study of the wordcounts,<br />

however, is suggestive. The following examples are grounded on comparisons<br />

among the word-counts for four groups of texts: the 40 poems of Table 23.1; the 25<br />

authorial sets that make up our main database of seventeenth-century verse; a large<br />

database of first-person narratives by 55 writers of prose fiction born between 1660 and<br />

1860, the subject of the next main section of this chapter; and the ten sets of personal<br />

letters of Table 23.3. From an examination of these four large groups, I pick out the more<br />

striking contrasts between verse and prose and offer some small inferences.<br />

Among the connectives, verse scores higher than prose for or and nor, the twin staples of<br />

many an antithesis, and for the preposition like, the prelude <strong>to</strong> simile. Verse scores higher<br />

for four of the five main relative pronouns, the exception being which. The embedded<br />

clauses cus<strong>to</strong>marily introduced by which (at least in British writing) may be less<br />

congenial <strong>to</strong> the flow of verse rhythms than the corresponding appended clauses. A<br />

number of monosyllabic connectives and semi-connectives may serve the metrical needs<br />

of many poets while offering the strong rhe<strong>to</strong>rical pointing of Res<strong>to</strong>ration verse<br />

argument. They include what, yet, now, such, <strong>to</strong>o, still, while, whilst, and thus. Verse also<br />

runs <strong>to</strong> high scores in some absolute forms like all and no.<br />

The prose set employed in these comparisons is <strong>to</strong>o heterogeneous <strong>to</strong> yield examples of<br />

much interest. But it is certainly worth noting that some of our poets abandon some of<br />

their most pronounced verse habits when they turn <strong>to</strong> prose. Congreve and Swift stand<br />

out among the poets in their frequent recourse <strong>to</strong> thus and never respectively. In<br />

Congreve's prose drama and Swift's letters, their scores fall right away. Other poets, like<br />

Katherine Phillips ("Orinda"), maintain much the same stylistic reper<strong>to</strong>ire through thick<br />

and thin. On/upon is a strong verse/prose discrimina<strong>to</strong>r, possibly because on has greater<br />

metrical versatility. Of the few poets much given <strong>to</strong> upon, Samuel Butler is celebrated not<br />

for a tin ear but for his mastery of offbeat rhymes and rhythms. In general, however, it<br />

seems that metrical necessities, rhe<strong>to</strong>rical patterning, and a comparative dearth of<br />

circumstantial detail are among the fac<strong>to</strong>rs that induce writers <strong>to</strong> assume rather different<br />

habits of connection and emphasis when they turn from verse <strong>to</strong> prose.<br />

While the separation of domain 1 and domain 2 is obviously worth maintaining and<br />

deserves further investigation, it is not a Berlin wall but only an analytical convenience.<br />

The vigorous traffic in both directions represents an intermingling of stylistic forces. As a


valuable sidelight in a study of doubtful authorship (Forsyth et al. 1999: 395–6), the<br />

writers make the point that an unduly ambitious analysis can collapse under its own<br />

weight. While something of that kind occurs in Figure 23.8, the outcome is not<br />

unintelligible. It is rather that some real authorial affinities are obscured when such<br />

extreme differences of genre supervene.<br />

Classification of Authorial Groups<br />

Genre and authorship are not the only powerful stylistic differentiae. A suitable choice of<br />

specimens allows other effects <strong>to</strong> be observed. Let us turn, for this purpose, <strong>to</strong> another<br />

large database. In its current state, it runs <strong>to</strong> 1.8 million words and consists entirely of<br />

first-person, retrospective narratives from works of prose fiction by 100 authors. The<br />

earliest born of the authors is Daniel Defoe (1660) and the latest is Louise Erdrich (1954).<br />

The later sets comprise American and Australian as well as British writings. Some 48 of<br />

the authors are female. The authorial sets range in length from 4,000 <strong>to</strong> almost 76,000<br />

words and each includes at least three narratives. The small sets are included, perforce, in<br />

order <strong>to</strong> represent some authors of particular interest like Louisa Atkinson, who is<br />

accepted as the earliest native-born Australian writer. With a threshold of 4,000 words,<br />

however, the differences of size among the sets show no apparent effects. (The e-texts<br />

were created by keyboard from standard editions and from the admirable facsimiles of<br />

early British fiction published by the Garland Press, New York. Permission was obtained<br />

where necessary and the texts were used only for the extraction of word-counts.)<br />

Earlier versions of this database have been used for a study of change over time (Burrows<br />

1992) and a pilot study of national differences in the English of literary narrative<br />

(Burrows 1996). The former takes its place beside Cluett (1976) and other, more<br />

linguistically oriented work like Biber and Finegan (1989). The latter, unfortunately, still<br />

stands alone.<br />

On the present occasion, developing work I presented at ACH/ALLC 1993 but have<br />

never published, the database will be used <strong>to</strong> examine possible differences between male<br />

and female writers. To this end, for reasons that will emerge, the database is divided<br />

between 55 authors born before 1860 and 45 born afterwards. The former subset includes<br />

26 female writers, the latter 22. As Table 23.4 and Table 23.5 indicate, the female writers<br />

are placed first in each subset. (Most of the information offered in those tables speaks for<br />

itself. It should be noted, however, that, where a writer is given two nationalities, his or<br />

her "less influential" country of abode is shown in parentheses.)<br />

Distribution tests like Student's /-test and the Mann-Whitney test are among the most<br />

appropriate procedures for assessing possible differences, over a range of many variables,<br />

between any two large sets of specimens. In both tests, the underlying assumption – the<br />

null hypothesis – is that the sets do not represent distinct populations. For each variable in<br />

turn, therefore, the results show the likelihood that the assumption is false. This<br />

likelihood, expressed as a degree of probability, takes the form of a decimal fraction like<br />

0.05 or 0.005. The smaller this fraction, the more likely it is that the assumption is false.<br />

When the tests are conducted with many variables, they must be expected <strong>to</strong> yield some


chance successes: around one result in twenty should be above 0.05; around one in a<br />

hundred should be above 0.01, and so on. With our 150 variables, therefore, we need<br />

many more than seven results from levels below 0.05, or many more than two from<br />

below 0.01 before we can begin <strong>to</strong> accept that there are systematic differences between<br />

our male and female writers, that they do indeed represent different populations. (We<br />

must also stiffen our stated requirement <strong>to</strong> allow for the fact that the "two-tailed" form of<br />

these tests, as used here, doubles the probability of a chance success.) To make the<br />

requirement still more rigorous, I have chosen, finally, <strong>to</strong> ignore results that do not satisfy<br />

both of the distribution tests at levels below 0.05.<br />

As Table 23.6 indicates, there is no need <strong>to</strong> quibble about borderline cases. In the upper<br />

part of the table, labeled (a), the results for our 55 earlier-born authors include no fewer<br />

than 32 out of 150 where the probabilities fall below 0.05. Almost all of them, in fact, fall<br />

far below that threshold. In the lower part of the table, labeled (b), the results for our 45<br />

later-born authors lie well within the realm of chance, with only six words yielding<br />

probabilities from below the 0.05 threshold. Taken <strong>to</strong>gether, the two sets of results<br />

indicate that, in our earlier but not our later set, the male and female authors represent<br />

different populations. Whereas men and women used <strong>to</strong> write very differently from each<br />

other, they have long since ceased <strong>to</strong> do so.<br />

In Table 23.6, the columns headed "/-score" and "DF" represent a different way of<br />

expressing the same probabilities from those shown in the column headed "/-prob." (The<br />

discrimina<strong>to</strong>ry power of /-scores is established by consulting an appropriate manual of<br />

statistical tables.) My reason for including the /-scores here is that they distinguish<br />

between positives and negatives. The positive scores are always those of the first set<br />

entered – in this case, the set of female authors.<br />

The contrasting scores for such referential markers as l/me/sbe/ber and, on the other<br />

hand, the/a point <strong>to</strong> a straightforward opposition between more and less personally<br />

oriented rhe<strong>to</strong>rical stances among the two sets of earlier-born writers. The absence of any<br />

masculine pronouns makes this opposition a little more complex; and the opposed scores<br />

(albeit weak) for mother and man add <strong>to</strong> that effect. Female authors, it seems, had more<br />

<strong>to</strong> say of females: both male and female authors had much <strong>to</strong> say of males. For the<br />

substance of what they said, of course, one turns, with pleasure, <strong>to</strong> the texts themselves.<br />

23.4 Table Fifty-five authors born before 1860 (female authors listed first)<br />

1 2 3 4 5 6 7 8<br />

Atkinson, Austen, Braddon, Brontë;, Brontë, Brontë, Cambridge, Edgeworth,<br />

Louisa Jane Mary Anne Charlotte Emily Ada Maria<br />

1775–<br />

1834– 72<br />

1817<br />

A<br />

UK<br />

1837–1915 1820–49<br />

UK UK<br />

1816–55<br />

UK<br />

1818–48 UK 1844–1926 1767–1849<br />

UK(A) UK


4005 8889 8269 22640 15644 15743 12266 29330<br />

9 10 11 12 13 14 15 16<br />

"Eliot, Fielding, Freeman, Gaskell, Hamil<strong>to</strong>n, Haywood, Hearne, Jewett,<br />

George" Sarah Mary W. Elizabeth Elizabeth Eliza Mary Sarah O.<br />

1819–80 1710–68 1852–1930 1810–65 1758– 1693?–1756 fl. 1718 1849–1909<br />

UK UK US UK 1816 UK UK UK US<br />

22686 75982 6084 51989 16922 28633 12815 5985<br />

17 18 19 20 21 22 23 24<br />

Lennox, Manley,<br />

Praed, Rosa<br />

Charlotte Mary<br />

Radcliffe,<br />

Ann<br />

Robinson,<br />

Scott, Sarah<br />

Mary<br />

Shelley,<br />

Mary<br />

Smith,<br />

Charlotte<br />

1720– 1663–<br />

1804 1724<br />

(US) UK UK<br />

1851–1935 1764–1823 1758–<br />

1797–1851 1749–1806<br />

1723–95 UK<br />

(A) UK UK 1800 UK UK UK<br />

29037 15727 9147 12066 10891 15316 35806 28695<br />

25 26 27 28 29 30 31 32<br />

S<strong>to</strong>we,<br />

Spence,<br />

Harriet<br />

Catherine<br />

B.<br />

Bage,<br />

Robert<br />

Becke,<br />

Louis<br />

Bierce, "Boldrewood, Brockden<br />

Ambrose Rolf" Brown, C.<br />

Clarke,<br />

Marcus<br />

1825–<br />

1910<br />

(UK) A<br />

1811–96 1728–1801 1855–1913 1842– 1826–1915<br />

US UK A 1914? US UK(A)<br />

1771–1810 1846–81<br />

US UK(A)<br />

11538 10965 13969 9641 7745 17722 15923 13342<br />

33 34 35 36 37 38 39 40<br />

Cleland, Collins, Cooper, Defoe, Dickens, Favenc, Fielding, Furphy,<br />

John Wilkie James F. Daniel Charles Ernest Henry Joseph<br />

1709–89 1824–89 1789–1851 1660–1731 1812–70 1845–1908 1707–54 1843–1912<br />

UK UK US UK UK UK(A) UK A<br />

12995 61539 5760 35317 8519 4762 75532 15532<br />

41 42 43 44 45 46 47 48<br />

Godwin, Graves, Hardy,<br />

William Richard Thomas<br />

Hawthorne, Holcroft,<br />

James, Henry<br />

Nath'l Thomas<br />

Johnson,<br />

Samuel<br />

Maturin, C.<br />

R.<br />

1715–<br />

1756–<br />

1804<br />

1836 UK<br />

UK<br />

1840–1928 1804–64<br />

UK US<br />

1745– 1843–1916<br />

1809 UK US<br />

1709–84<br />

UK<br />

1782–1824<br />

UK<br />

28814 23685 12867 18154 19628 33244 16943 16581<br />

49 50 51 52 53 54 55 56<br />

Poe,<br />

Melville,<br />

Edgar<br />

Herman<br />

A.<br />

Richardson, Scott,<br />

Samuel Walter<br />

Smollett, S<strong>to</strong>ck<strong>to</strong>n,<br />

Tobias Frank R.<br />

"Mark<br />

Twain"<br />

1819–91 c1809– 1689–1761 1771–1832 1721–71 1834–1902 1835–1910<br />

US 49 US UK UK UK US US


21084 10930 18615 29156 31746 13321 10142<br />

23.5 Table Forty–five authors born after 1860 (female authors listed first)<br />

1 2 3 4 5 6 7 8<br />

Anderson, Bedford, Bowen, Carter, Cather, Erdrich, Franklin, Gardam,<br />

Jessica Jean Elizabeth Angela Willa Louise Miles Jane<br />

1920?–99 1946-<br />

A (UK(A)<br />

1899– 1940–92 1873–<br />

1973 UK UK 1947 US 1954-US<br />

1879–<br />

1954<br />

A(UK)<br />

1928-UK<br />

12825 6463 11071 15642 12051 19565 16712 10470<br />

9 10 11 12 13 14 15 16<br />

Grenville, Hazzard, Hospital, Jolley, Jong, Langley, Lurie, Mansfield,<br />

Kate Shirley Janette T. Elizabeth Erica Eve Alison Katherine<br />

1950-A<br />

1940?-<br />

1931-A(US) A(US/<br />

Can.)<br />

1923-<br />

UK(A)<br />

1942-US 1908–<br />

1974 A<br />

1888–<br />

1926-US 1923<br />

NZ(UK)<br />

14482 10359 5700 12533 13545 13345 14287 10827<br />

17 18 19 20 21 22 23 24<br />

Murdoch,<br />

Park, Ruth<br />

Iris<br />

Stead, Tennant, Whar<strong>to</strong>n, Wright,<br />

Christina Kylie Edith Judith<br />

Barth,<br />

John<br />

Bellow,<br />

Saul<br />

1919–99<br />

UK<br />

1923?-<br />

NZ(A)<br />

1902–83<br />

A(UK)<br />

1912–88 1862– 1915–<br />

A 1937 US 2000 A<br />

1930-US 1915-US<br />

21437 6755 10442 7112 19269 10012 12836 20342<br />

25 26 27 28 29 30 31 32<br />

Boothby, Boyd, Carey, Cary, Cowan, Doc<strong>to</strong>row, Faulkner, Fowles,<br />

Guy William Peter Joyce Peter E. L. William John<br />

1867–<br />

1905 (A)<br />

UK<br />

1952-UK 1943-A<br />

1888–<br />

1914-A<br />

1957 UK<br />

1931-US<br />

1897–<br />

1962 US 1926-UK<br />

13182 9871 24708 9936 13221 18701 21471 30394<br />

33 34 35 36 37 38 39 40<br />

Greene, Hemingway, Johns<strong>to</strong>n, Joyce, Lawson, Lodge, Malouf, McEwan,<br />

Graham Ernest George James Henry David David Ian<br />

1904–91<br />

A<br />

1899–1961<br />

US<br />

1912–70<br />

A<br />

1882– 1867–<br />

1941 UK 1922 A<br />

1935-UK 1934-A 1948-UK<br />

9184 14202 24729 6531 16056 8465 12261 18919<br />

41 42 43 44 45


Palmer,<br />

Vance<br />

1885–<br />

1959 A<br />

Porter, Hal Sargeson,<br />

Frank<br />

1911–84 A 1903–82<br />

NZ<br />

Waugh,<br />

Evelyn<br />

1903–66<br />

UK<br />

Wells, H.<br />

G.<br />

1866–<br />

1946 UK<br />

13256 14274 10869 11380 50612<br />

Texts are complege except where specified. Texts marked* are from original or standard<br />

modern editions. Those marked # are from the Chadwyck-Healey Archive of English<br />

Poetry, <strong>to</strong> which my university subscribes. The texts were used only for the extraction of<br />

word-counts.<br />

23.6 Table One hundred authors: significant differentiations among the 150 most<br />

common words<br />

RK. Word t-score DF t-prob.<br />

MWprob.<br />

(a) 55 authors born before 1860: 26 women vs. 29 men<br />

1 1 the −5.18 52 0.0000 0.0000<br />

2 3 I 3.40 52 0.0013 0.0038<br />

3 4 of −3.50 51 0.0010 0.0020<br />

4 5a −2.79 52 0.0073 0.0088<br />

5 6 was 2.69 49 0.0097 0.0112<br />

6 7 <strong>to</strong> (inf.) 3.33 52 0.0016 0.0023<br />

7 8 in (p.) −3.44 49 0.0012 0.0036<br />

8 14 her 3.03 52 0.0038 0.0021<br />

9 15 me 3.55 48 0.0009 0.0019<br />

10 18 not 2.91 44 0.0057 0.0068<br />

11 20 she 4.69 52 0.0000 0.0000<br />

12 21 but 3.30 43 0.0020 0.0020<br />

13 35 which (rp) −2.38 52 0.0210 0.0358<br />

14 36 would 2.89 52 0.0057 0.0065<br />

15 38 this −2.48 52 0.0160 0.0244<br />

16 41 could 2.67 50 0.0100 0.0149<br />

17 45 or −2.55 52 0.0140 0.0187<br />

18 46 one −2.09 52 0.0420 0.0330<br />

19 52 did 3.04 37 0.0043 0.0029<br />

20 54 what 3.37 39 0.0017 0.0007<br />

21 73 never 2.88 35 0.0068 0.0224


RK. Word t-score DF t-prob.<br />

MWprob.<br />

22 75 man −2.10 52 0.0410 0.0278<br />

23 82 thought 3.03 49 0.0039 0.0018<br />

24 83 do 3.35 32 0.0021 0.0033<br />

25 99 two −3.36 33 0.0020 0.0006<br />

26 101 how 3.76 46 0.0005 0.0012<br />

27 108 know 4.57 29 0.0001 0.0002<br />

28 115 mother 2.04 44 0.0470 0.0065<br />

29 122 <strong>to</strong>ld 2.58 41 0.0140 0.0123<br />

30 141 think 3.36 32 0.0021 0.0038<br />

31 142 go 2.87 47 0.0062 0.0045<br />

32 149 these −3.75 46 0.0005 0.0002<br />

(b) 45 authors born after 1860: 22 women vs. 23 men<br />

1 6 was −2.37 39 0.0230 0.0377<br />

2 59 so (av. deg.) 3.48 41 0.0012 0.0015<br />

3 62 time −2.52 41 0.0160 0.0188<br />

4 72 like (p) 2.61 35 0.0130 0.0114<br />

5 75 man −3.26 31 0.0027 0.0017<br />

6 81 before −2.59 42 0.0130 0.0121<br />

Abbreviations:<br />

Rk. = rank ex 150<br />

DF = degrees of freedom<br />

t-prob and MW-prob. = probahility as assessed on Student's t-test and Mann-<br />

Whitney test respectively<br />

inf. = infinitive particle<br />

p = preposition<br />

rp = relative pronoun<br />

av. deg. = adverb of degree<br />

Among the clausal connectives, but, what, and bow, words often introducing an<br />

exclama<strong>to</strong>ry note, are favored by the female writers. The auxiliary verbs did and do are<br />

often used for emphasis. These all stand, perhaps, with not, never, could, and would in<br />

signifying that the present indicative is not quite as these writers might wish and that a<br />

certain insistence is thought necessary if their voices are <strong>to</strong> be heard. The infinitive<br />

particle <strong>to</strong> (the marker of compound verbs) and the verbs thought, think, know, and <strong>to</strong>ld<br />

are also favored by the female writers. The point <strong>to</strong> observe here is that these words have<br />

less <strong>to</strong> do with mental activities as such than with overt reference <strong>to</strong> those activities: "I<br />

thought that…", "…, I think,…", "…, you know,…" and "as I <strong>to</strong>ld you/him/her" are<br />

typical specimens. Their presence posits a distance between speaker and deed.


Among the smaller group of words favored by the male writers, the association of the/a<br />

with of I in and this/these and even, perhaps, with one/two suggests an emphasis on<br />

things and ideas rather than persons. But, as markers of a firmly marshaled, somewhat<br />

artificial syntax, which/or may reveal more than any of the phenomena so far mentioned<br />

about the most remarkable aspect of Table 23.6.<br />

If a body of statistically significant differences between male and female writers can<br />

disappear completely over a generation or two, as these results suggest, the explanation<br />

should not be sought in biological differences between the sexes. But if differences in<br />

"gender", expressed in the little manifestations of male and female social roles <strong>to</strong>uched<br />

on above, were the whole explanation, it is unlikely that the alteration would be so nearly<br />

absolute. Irrespective of one's personal standpoint, it is hard <strong>to</strong> claim that the social roles<br />

of men and women in English-speaking societies can no longer be distinguished.<br />

Figure 23.9, the last of my cluster analyses, throws a clear light on these questions. The<br />

55 entries for the authors born before 1860 fall, chiefly by gender, in<strong>to</strong> two large groups.<br />

Six of the 29 men are located among the women. Five of the 26 women are located<br />

among the men. Of these five women, all but one are known <strong>to</strong> have been educated as if<br />

they were boys. Of these six men, all but one are known <strong>to</strong> have grown up without<br />

benefit of the classical education offered in British grammar schools. (The other two are<br />

Charlotte Lennox, who was certainly well enough educated <strong>to</strong> publish her translations<br />

from the French, and Thomas Alexander Brown – "Rolf Boldrewood" – who was born in<br />

Britain but came young <strong>to</strong> Australia.) The idea that a British classical education exerted a<br />

powerful influence on the way boys wrote while most girls continued <strong>to</strong> write something<br />

more like spoken English can be endorsed by making a similar analysis of an even earlier<br />

subgroup – those of our British writers who were born before 1800. The division between<br />

male and female is even more pronounced, but Richardson, Defoe, and Charlotte Lennox<br />

continue as exceptions. It seems clear, in short, that we began <strong>to</strong> write more like each<br />

other when we began <strong>to</strong> go <strong>to</strong> school <strong>to</strong>gether. And the very notion that a general<br />

difference between males and females was evident is transmuted in<strong>to</strong> the more plausible<br />

idea that our education deeply affects the way we use our language.<br />

Reflections<br />

In its traditional forms, textual analysis has <strong>to</strong> do with separating, distinguishing, and the<br />

like and it usually treats of single works. This sort of literary analysis often rests upon<br />

seemingly intuitive insights and discriminations, processes that may seem remote from<br />

the gathering and combining and classifying on which I have concentrated and in which<br />

computational stylistics is usually engaged. But those insights and discriminations are not<br />

ultimately intuitive because they draw, albeit covertly, upon data gathered in a lifetime's<br />

reading, s<strong>to</strong>red away in a subconscious memory bank, and put <strong>to</strong> use, as Samuel Johnson<br />

reminds us, through processes of comparison and classification, whether tacit or overt.<br />

When I. A. Richards opened Principles of Literary Criticism (1924) by saying that "A<br />

book is a machine <strong>to</strong> think with", he foreshadowed an attack upon what he regarded as<br />

the undue aestheticizing of literary criticism. So far as it bears upon books themselves,


his remark gains new force from recent changes in the technology of texts. Although my<br />

own chief premise may also offend some old-fashioned litterateurs, that is not my motive.<br />

In proposing that literary criticism and analysis are always grounded in the inductive<br />

processes of classification and differentiation, I wish <strong>to</strong> emphasize the links between the<br />

old stylistics and the new. The close readers and analysts of former years and their<br />

present counterparts, who draw upon computational methods for their textual analyses,<br />

should actually be close allies.<br />

Traditional and computational forms of textual analysis do have their distinct strengths.<br />

The parallels between the two plots of King Lear or the subtleties discerned by Coleridge<br />

in the opening lines of Hamlet lie as far beyond the reach of a computer as the ironies<br />

offered by Jane Austen or the ambiguities detected by William Empson. But computers<br />

obviously surpass our unassisted powers in managing large textual corpora, singling out<br />

unique forms or gathering all the instances of common ones.<br />

Within the realm of English studies, statistical work on single authors has usually been<br />

focused on comparisons of different texts. There are a few such studies at book length,<br />

treating respectively of Swift, Jane Austen, and Mil<strong>to</strong>n (Milic 1967; Burrows 1987;<br />

Corns 1990). For exemplary studies of authorial development at article length, see<br />

Brainerd (1980), Craig (1999), and Forsyth (1999). Among those studies of single texts in<br />

which literary and statistical analysis are associated, there is notable work on Blake's The<br />

Four Zoas (Ide 1989) and Joyce's Ulysses (McKenna and An<strong>to</strong>nia <strong>2001</strong>). In both<br />

instances, the argument relies upon internal comparisons between different facets of these<br />

large and complex texts.<br />

But the vast textual archives now available offer rich opportunities for the study of single<br />

texts in a manner that has never before been feasible. (For a brief illustration of this<br />

approach, see Burrows 2002b: 693–6.) The database of late seventeenth-century verse<br />

used at the beginning of the present chapter yields well-founded scores for as many<br />

common words as anyone might wish. Norms derived from the database make it easy <strong>to</strong><br />

study the vocabulary of any poet of the day. We are now able, however, <strong>to</strong> take a further<br />

step. The corresponding scores for any poem of adequate length can be compared not<br />

only with those same norms but also with the author's overall scores. It is therefore<br />

possible <strong>to</strong> establish that, in certain demonstrable respects, Absalom and Acbi<strong>to</strong>pbel or<br />

Upon Apple<strong>to</strong>n House is characteristic either of its author or of some specified group –<br />

male poets, say, or satirists – but that, in others, it departs from his usual practice or from<br />

theirs. At this point, I suggest, the evidence flowing from statistical comparison is more<br />

nearly complementary <strong>to</strong> the evidence offered by the traditional literary analyst than it<br />

has ever been. With due allowance for the felt needs and all the mordant joys of a<br />

continuing sibling rivalry, dare we hope <strong>to</strong> go forward hand in hand?<br />

Appendix<br />

In our main dataset, from which the list of the 150 most common words derives, the<br />

present corpus of 540,244 words ranges widely across the work of the following 25<br />

poets: Aphra Behn (1640–89) 21,705 words; Alexander Brome (1620–66) 29,539;


Samuel Butler (1612–80) 30,932; William Congreve (1670–1729) 30,917; Charles<br />

Cot<strong>to</strong>n (1630–87) 12,625; Abraham Cowley (1618–67) 19,272; Sir John Denham (1615–<br />

69) 30,092; Charles Sackville, Earl of Dorset (1638–1706) 9,586; John Dryden (1631–<br />

1700) 18,238; Thomas D'Urfey (1653–1723), 18,757; Robert Gould (1660?-1709?)<br />

29,110; Andrew Marvell (1621–78) 23,282; John Mil<strong>to</strong>n (1608–74) 18,924; John<br />

Oldham (1653–83) 32,462; Katherine Phillips (1631–64) 29,004; Matthew Prior (1664–<br />

1721) 32,000; Alexander Radcliffe (floruit 1669–96) 11,889; John Wilmot, Earl of<br />

Rochester (1648–80) 12,725; Sir Charles Sedley (1639?-1701) 10,304; Elkanah Settle<br />

(1648–1724) 24,080; Thomas Shadwell (l642?-92) 14,540; Jonathan Swift (1667–1745)<br />

30,974; Nahum Tate (1652–1715) 20,333; Edmund Waller (1606–87) 16,443; Anne<br />

Whar<strong>to</strong>n (1659–85) 12,511. Most of the corpus was prepared by John Burrows and<br />

Harold Love, assisted by Alexis An<strong>to</strong>nia and Meredith Sherlock. The Marvell subset was<br />

contributed by Chris<strong>to</strong>pher Wortham, assisted by Joanna Thompson.<br />

References for Further Reading<br />

NB. For a broader framework than the present argument requires, see the reference list in<br />

Burrows (2003).<br />

Biber, Douglas and Edward Finegan (1989). Drift and the Evolution of English Style: A<br />

His<strong>to</strong>ry of Three Genres. Language 65: 487–517.<br />

Brainerd, B. (1980). The Chronology of Shakespeare's Plays: A Statistical Study.<br />

Computers and the Humanities 14: 221–30.<br />

Bratley, Paul and Donald Ross, Jr (1981). Syllabic Spectra. ALLC Journal 2: 41–50.<br />

Burrows, J. F. (1987). Computation in<strong>to</strong> Criticism: A Study of Jane Austen's Novels and<br />

an Experiment in Method. Oxford: Clarendon Press.<br />

Burrows, J. F. (1992). Computers and the Study of Literature. In C. S. Butler (ed.),<br />

Computers and Written Texts (pp. 167–204). Oxford: Blackwell.<br />

Burrows, J. F. (1996). Tip<strong>to</strong>eing in<strong>to</strong> the Infinite: Testing for Evidence of National<br />

Differences in the Language of English Narrative. In N. Ide and S. Hockey (eds.),<br />

Research in Humanities Computing 4 (pp. 1–33). Oxford: Clarendon Press.<br />

Burrows, J. F. (2002a). "Delta": a Measure of Stylistic Difference and a Guide <strong>to</strong> Likely<br />

Authorship. Literary and Linguistic Computing 17: 267–86.<br />

Burrows, J. F. (2002b). The Englishing of Juvenal: Computational Stylistics and<br />

Translated Texts. Style 36: 677–94.<br />

Burrows, J. F. (2003). Questions of Authorship: Attribution and Beyond. Computers and<br />

the Humanities 37: 1–26.


Cluett, Robert (1976). Prose Style and Critical Reading. New York: Teachers College<br />

Press.<br />

Corns, Thomas N. (1990). The Development of Mil<strong>to</strong>n's Prose Style. London: Oxford<br />

University Press.<br />

Craig, Hugh (1999). Contrast and Change in the Idiolects of Ben Jonson Characters.<br />

Computers and the Humanities 33: 221–40.<br />

Eliot, T S. (1932). For Lancelot Andrewes. In Selected Essays. London: Faber.<br />

Forsyth, R. S. (1999). Stylochronometry with Substrings, or: a Poet Young and Old.<br />

Literary and Linguistic Computing 14: 467–77.<br />

Forsyth, Richard S., David I. Holmes, and Emily K. Tse (1999). Cicero, Sigonio, and<br />

Burrows: Investigating the Authenticity of the Consolatio. Literary and Linguistic<br />

Computing 14: 375–400.<br />

Halsband, Robert, (ed.) (1970). The Selected Letters of Lady Mary Wortley Montagu.<br />

London: Longman.<br />

Hoover, David (2002). Frequent Word Sequences and Statistical Stylistics. Literary and<br />

Linguistic Computing 17: 157–80.<br />

Ide, Nancy M. (1989). Meaning and Method: Computer-assisted Analysis of Blake. In<br />

Rosanne G. Potter (ed.), Literary Computing and Literary Criticism. Philadelphia:<br />

University of Pennsylvania Press.<br />

McKenna, C. W. F. and Alexis An<strong>to</strong>nia (<strong>2001</strong>). The Statistical Analysis of Style:<br />

Reflections on Form, Meaning, and Ideology in the "Nausicaa" Episode of Ulysses.<br />

Literary and Linguistic Computing 16: 353–74.<br />

Milic, Louis T. (1967). A Quantitative Approach <strong>to</strong> the Style of Jonathan Swift. Paris:<br />

Mou<strong>to</strong>n.<br />

Richards, I. A. (1924). Principles of Literary Criticism. London: Routledge.<br />

Sutherland, James (1969). English Literature of the Late Seventeenth Century. Oxford:<br />

Clarendon Press.


23.1 Figure Forty long Poems: cluster anlaysis based on the 150 most common<br />

words of the main corpus<br />

23.2 Figure Forty-two long poems: cluster analysis based on the 150 most common<br />

words of the main corpus<br />

23.3 Figure Forty long poems: cluster analysis based on the 69 most common<br />

words of "domain 2"<br />

23.4 Figure Forty long poems: cluster analysis based on the 54 most common<br />

words of "domain 1"


23.5 Figure Sixty-six tents: cluster analysis based on the 54 most common words of<br />

"domain 1"<br />

23.6 Figure Sixty-six texts: cluster anlaysis based on the 150 most common words<br />

of the main dataser<br />

23.7 Figure A sketch-map of relationships among the common words of English: a<br />

rationale for concomitance of frequency<br />

23.8 Figure Sixty-six texts: cluster analysis based on the 69 most common words of<br />

"domain 2"


23.9 Figure Fifty-five "hos<strong>to</strong>rians" born before 1860: cluster analysis based on<br />

the thirty-two most common gender-discriminating words of the databse<br />

24.<br />

Thematic Research Collections<br />

Carole L. Palmer<br />

Introduction<br />

The analogy of the library as the labora<strong>to</strong>ry of the <strong>humanities</strong> has always been an<br />

exaggeration. For most <strong>humanities</strong> scholars, it has been rare <strong>to</strong> find the necessary<br />

materials for a research project amassed in one place, as they are in a labora<strong>to</strong>ry setting.<br />

Thematic research collections are <strong>digital</strong> resources that come closer <strong>to</strong> this ideal. Where<br />

in the past scholars produced documents from source material held in the collections of<br />

libraries, archives, and museums, they are now producing specialized scholarly resources<br />

that constitute research collections. Scholars have recognized that information<br />

technologies open up new possibilities for re-creating the basic resources of research and<br />

that computing <strong>to</strong>ols can advance and transform work with those resources (Unsworth<br />

1996). Thematic research collections are evolving as a new genre of scholarly production<br />

in response <strong>to</strong> these opportunities. They are <strong>digital</strong> aggregations of primary sources and<br />

related materials that support research on a theme.<br />

Thematic research collections are being developed in tandem with the continuing<br />

collection development efforts of libraries, archives, and museums. These institutions<br />

have long served as s<strong>to</strong>rehouses and workrooms for research and study in the <strong>humanities</strong><br />

by collecting and making accessible large bodies of diverse material in many subject<br />

areas. Thousands of extensive, specialized research collections have been established, but<br />

they are often far removed from the scholars and students who wish <strong>to</strong> work with them.<br />

In recent years, many institutions have begun <strong>to</strong> <strong>digital</strong>ly reformat selected collections<br />

and make them more widely available on the Web for use by researchers, students, and<br />

the general public.<br />

Humanities scholars are participating in this movement, bringing their subject expertise<br />

and acumen <strong>to</strong> the collection development process. In taking a thematic approach <strong>to</strong><br />

aggregating <strong>digital</strong> research materials, they are producing circumscribed collections,<br />

cus<strong>to</strong>mized for intensive study and analysis in a specific research area. In many cases<br />

these <strong>digital</strong> resources serve as a place, much like a virtual labora<strong>to</strong>ry, where specialized<br />

source material, <strong>to</strong>ols, and expertise come <strong>to</strong>gether <strong>to</strong> aid in the process of scholarly work<br />

and the production of new knowledge.


This chapter is focused primarily on the thematic research collections created by scholars.<br />

The his<strong>to</strong>ry and proliferation of the new genre cannot be examined in full within the<br />

limits of the chapter. Instead, this essay will identify characteristics of the genre and<br />

clarify its relationship <strong>to</strong> the collection development activities that have traditionally<br />

taken place in research libraries. This relationship is central <strong>to</strong> understanding how our<br />

s<strong>to</strong>res of research materials will evolve in the <strong>digital</strong> realm, since thematic research<br />

collections are derived from and will ultimately contribute <strong>to</strong> the larger institution-based<br />

collections. The thematic collections used as examples throughout the chapter have been<br />

selected <strong>to</strong> illustrate specific features and trends but do not fully represent the variety of<br />

collections produced or under development. Although there are numerous collections<br />

available for purchase or through licensing agreements, all the examples identified here<br />

were available free on the Web as of Oc<strong>to</strong>ber 2002.<br />

Characteristics of the Genre<br />

There are no firm parameters for defining thematic research collections (hereafter<br />

referred <strong>to</strong> as thematic collections), but there are characteristics that are generally<br />

attributable <strong>to</strong> the genre. John Unsworth (2000b) describes thematic collections as being:<br />

electronic<br />

heterogeneous datatypes<br />

extensive but thematically coherent<br />

structured but open-ended<br />

designed <strong>to</strong> support research<br />

authored or multi-authored<br />

interdisciplinary<br />

collections of <strong>digital</strong> primary resources<br />

These characteristics are common <strong>to</strong> the projects developed at the Institute for Advanced<br />

Technology in the Humanities (IATH), the research and development center at the<br />

University of Virginia, previously directed by Unsworth. They are also broadly<br />

applicable <strong>to</strong> many of the research collections being developed elsewhere by scholars,<br />

librarians, and collaborative teams. With thematic collections, however, there is<br />

considerable synergy among these characteristics, and as the genre grows and matures<br />

additional characteristics are emerging that differentiate it from other types of <strong>digital</strong><br />

resources.<br />

Table 24.1 reworks Unsworth's list of descrip<strong>to</strong>rs, separating content and function aspects<br />

of the collections and adding emerging features of the genre. The first tier of features


contains basic elements that are generally shared by thematic collections. In terms of<br />

content, they are all <strong>digital</strong> in format and thematic in scope. In terms of function, they are<br />

all intentionally designed <strong>to</strong> support research. The next tier of features further defines the<br />

makeup and role of thematic collections, and <strong>to</strong>gether they reflect the unique contribution<br />

this type of <strong>digital</strong> resource is making <strong>to</strong> research in the <strong>humanities</strong>. Unlike the basic<br />

elements, these characteristics are highly variable. They are not represented in all<br />

thematic collections, and the degree <strong>to</strong> which any one is present in a given collection is<br />

varied. Collections differ in the range and depth of content and the types of functions<br />

provided. The basic elements and all the content features, outlined below, are closely<br />

aligned with Unsworth's description. The emergent function features are explicated more<br />

fully in sections that follow.<br />

Thematic collections are <strong>digital</strong> in format. While the sources may also exist as printed<br />

texts, manuscripts, pho<strong>to</strong>graphs, paintings, film, or other artifacts, the value of a thematic<br />

collection lies in the effectiveness of the <strong>digital</strong> medium for supporting research with the<br />

materials. For example, through advances in information technology, crea<strong>to</strong>rs of The<br />

William Blake Archive have been able <strong>to</strong> produce images that are more accurate in color,<br />

detail, and scale than commercially printed reproductions, and texts more faithful <strong>to</strong> the<br />

author's originals than existing printed editions (Viscomi 2002).<br />

The contents are thematic or focused on a research theme. For example, a number of the<br />

IATH collections are constructed around author-based themes, including The Complete<br />

Writings and Pictures of Dante Gabriel Rossetti: A Hypermedia Research Archive, The<br />

Dickinson Electronic Archives, and The Walt Whitman Archive. Collections can also be<br />

developed around a literary or artistic work, such as Uncle Tom's Cabin and American<br />

Culture. A collection called Hamlet on the Ramparts, designed and maintained by the<br />

MIT Shakespeare Project, is a good example of a collection based on a narrowly defined<br />

literary theme. That project aims <strong>to</strong> bring <strong>to</strong>gether texts, artwork, pho<strong>to</strong>graphs, films,<br />

sound recordings, and commentary related <strong>to</strong> a very specific literary entity, Hamlet's first<br />

encounter with the ghost. A collection theme can be an event, place, phenomenon, or any<br />

other object of study. Interesting examples outside of literary studies include the Salem<br />

Witch Trials, Pompeii Forum, and the Waters of Rome projects. Some thematic<br />

collections are embedded in larger <strong>digital</strong> resources. One example is the Wilfred Owen<br />

Multimedia Digital Archive, a core content area in the Virtual Seminars for Teaching<br />

Literature, a pedagogical resource developed at Oxford University.<br />

Like traditional library collections in the <strong>humanities</strong>, thematic collections have been built<br />

for research support, but the new genre is producing more specialized microcosms of<br />

materials that are tightly aligned with specific research interests and that aid in specific<br />

research processes. Some thematic collections have been designed as new virtual<br />

environments for scholarly work. For example, Digital Dante, a project produced at the<br />

Institute for Learning Technologies at Columbia University, has been conceived as a<br />

"place" for study and learning and a "means" of scholarly production.


24.1 Table Features of thematic research collections<br />

Content Function<br />

Basic elements<br />

* Digital<br />

* Thematic<br />

Variable characteristics<br />

Research support<br />

* Coherent Scholarly contribution<br />

* Heterogeneous Contextual mass<br />

* Structured Interdisciplinary platform<br />

* Open-ended Activity support<br />

The thematic framework allows for coherent aggregation of content. All the materials<br />

included assist in research and study on the theme. This coherence is generally anchored<br />

by a core set of primary sources. The capabilities of networked, <strong>digital</strong> technology make<br />

it possible <strong>to</strong> bring <strong>to</strong>gether extensive corpuses of primary materials and <strong>to</strong> combine those<br />

with any number of related works. Thus the content is heterogeneous in the mix of<br />

primary, secondary, and tertiary materials provided, which might include manuscripts,<br />

letters, critical essays, reviews, biographies, bibliographies, etc., but the materials also<br />

tend <strong>to</strong> be multimedia. The <strong>digital</strong> environment provides the means <strong>to</strong> integrate many<br />

different kinds of objects in<strong>to</strong> a collection. In literary studies collections, multiple<br />

versions of a given text are commonly made available <strong>to</strong> aid in comparative analysis,<br />

along with additional types of media such as maps, illustrations, and recorded music. For<br />

example, Uncle Tom's Cabin and American Culture contains different editions of the<br />

primary text along with poems, images, films, and songs that show the context and<br />

his<strong>to</strong>ry surrounding the primary work (Condron et al. <strong>2001</strong>).<br />

The individual items in a collection are structured <strong>to</strong> permit search and analysis, with<br />

most projects in the <strong>humanities</strong> adopting SGML-based markup formats. Many aspects of<br />

a source may be coded, including bibliographic information, physical features, and<br />

substantive content, <strong>to</strong> produce highly flexible and searchable <strong>digital</strong> materials. The<br />

collection as a whole is further organized in<strong>to</strong> interrelated groups of materials for display<br />

and <strong>to</strong> assist in retrieval. Libraries, archives, and museums have conventions for<br />

structuring and representing collections, which include systems and guidelines for<br />

applying metadata, classification schemes, and descrip<strong>to</strong>rs. Some of these methods are<br />

being applied in thematic collections, but scholars are also designing new approaches and<br />

developing new standards, such as the TEI (Text Encoding Initiative) markup language<br />

for tagging scholarly material, that are more attuned with scholarly practices. As a result,<br />

there is not yet uniformity in the methods used for structuring data in <strong>digital</strong> collections,<br />

but there are ongoing efforts <strong>to</strong> align standards and a growing awareness by collection<br />

developers of the value of developing interoperable systems.


Collections of all kinds can be open-ended, in that they have the potential <strong>to</strong> grow and<br />

change depending on commitment of resources from collec<strong>to</strong>rs. Most thematic<br />

collections are not static. Scholars add <strong>to</strong> and improve the content, and work on any given<br />

collection could continue over generations. Moreover, individual items in a collection can<br />

also evolve because of the inherent flexibility (and vulnerability) of "born <strong>digital</strong>" and<br />

transcribed documents. The dynamic nature of collections raises critical questions about<br />

how they will be maintained and preserved as they evolve over time.<br />

Scholarly Contribution<br />

While thematic collections support both research and pedagogy, the scholarly<br />

contribution that results from the creation and use of the resources is what qualifies them<br />

as a scholarly genre. When electronic sources are brought <strong>to</strong>gether for scholarly purposes<br />

they become a new, second-generation electronic resource (Unsworth 2000b). Scholars<br />

are not only constructing environments where more people can do research more<br />

conveniently, they are also creating new research. Like other scholarship in the<br />

<strong>humanities</strong>, research takes place in the production of the resource, and research is<br />

advanced as a result of it. Thus, scholarship is embedded in the product and its use. And<br />

like research generated in the fields of engineering, computer science, and information<br />

science, some of the research contribution lies in the technical design, functionality, and<br />

innovation that makes new kinds of research possible.<br />

Authorship is included in Unsworth's description of thematic collections, but the term<br />

does not fully capture the nature of the work involved in developing thematic collections.<br />

A collection is not always attributable <strong>to</strong> one author or even a few co-authors, and the<br />

process of production may not generate new, original content. Some of the technical<br />

work involved in creating a collection requires expertise outside that of a typical author in<br />

the <strong>humanities</strong>. Literary scholars who are assembling electronic texts and archives of<br />

multimedia objects have become "literary-encoders" and "literary-librarians"<br />

(Schreibman 2002). Moreover, as noted above, thematic collections are inherently openended<br />

and therefore can be added <strong>to</strong> and altered in dramatic ways over time by new<br />

participants in the process. After a collection has been established for some time, it may<br />

not be accurate <strong>to</strong> continue <strong>to</strong> assign complete authorship <strong>to</strong> the origina<strong>to</strong>r, and it may<br />

prove complicated <strong>to</strong> trace authorial responsibility as it evolves over years or even<br />

decades.<br />

In many ways collection work resembles that of an edi<strong>to</strong>r, but the activities of cura<strong>to</strong>rs,<br />

archivists, and compilers are also applicable. But the concept of author is useful in its<br />

ability <strong>to</strong> relate the significance of the purposeful organization of information. As<br />

Atkinson notes in reference <strong>to</strong> professional collection developers in research libraries,<br />

since "every text is <strong>to</strong> some extent a compilation of previous texts, then the collection is a<br />

kind of text – and the building of the collection is a kind of authorship" (1998: 19).<br />

Nonetheless, "crea<strong>to</strong>r" seems best for encapsulating the range of work involved in the<br />

development of thematic collections. The term "crea<strong>to</strong>r" has become common in standard<br />

schemes for describing electronic resources, such as the Dublin Core metadata element<br />

set, and it accommodates the technical, intellectual, and creative aspects of the <strong>digital</strong>


collection development process. Academic subject expertise is considered critical in the<br />

development of quality, scholarly <strong>digital</strong> resources (Crane and Rydberg-Cox 2000), and<br />

technical computing knowledge of text, image, audio and video standards, and<br />

applications is equally important. Thus, the crea<strong>to</strong>rs of scholarly collections will need <strong>to</strong><br />

be a new kind of scholar, or team, with a distinct mix of expertise in at least three areas –<br />

the specific subject matter and associated critical and analytical techniques, technical<br />

computing processes, and principles of content selection and organization.<br />

Contextual Mass<br />

The crea<strong>to</strong>rs of thematic collections are constructing research environments with<br />

contextual mass, a proposed principle for <strong>digital</strong> collection development that prioritizes<br />

the values and work practices of scholarly communities (Palmer 2000). The premise<br />

behind the principle is that rather than striving for a critical mass of content, <strong>digital</strong><br />

research libraries should be systematically collecting sources and developing <strong>to</strong>ols that<br />

work <strong>to</strong>gether <strong>to</strong> provide a supportive context for the research process. For libraries, this<br />

approach <strong>to</strong> collection development requires analysis of the materials and activities<br />

involved in the practices of the different research communities served (Brockman et al.<br />

<strong>2001</strong>). Researchers are able <strong>to</strong> more readily construct contextual mass for themselves<br />

through highly purposeful selection and organization of content directly related <strong>to</strong> their<br />

specialized areas of research.<br />

All collections are built through the process of privileging some materials over others<br />

(Buckland 1995), and the construction of contextual mass takes place through careful,<br />

purposeful privileging. Because of the specific scope and aims of thematic collections,<br />

crea<strong>to</strong>rs select materials in a highly focused and deliberate manner, creating dense,<br />

interrelated collections. By contrast, in both physical and <strong>digital</strong> libraries, materials are<br />

usually separated for reasons unimportant <strong>to</strong> the scholar. For example, primary texts may<br />

be part of an isolated rare book room or special collection, while secondary works are in<br />

separate book and journal collections, with indexes, bibliographies, and handbooks kept<br />

in reference areas. Moreover, the his<strong>to</strong>rical, literary, and cultural treatments of a <strong>to</strong>pic are<br />

likely <strong>to</strong> be further scattered across differentiated classes of subjects. When a person uses<br />

a research library collection they are interacting with a context that includes physical,<br />

institutional, and intellectual features (Lee 2000). It is a grand and scattered context<br />

compared <strong>to</strong> that of thematic collections, which tend <strong>to</strong> focus on the physical context of<br />

core primary sources and the intellectual context represented in a mix of heterogeneous<br />

but closely associated materials.<br />

Collections built on a contextual mass model create a system of interrelated sources<br />

where different types of materials and different subjects work <strong>to</strong>gether <strong>to</strong> support deep<br />

and multifaceted inquiry in an area of research. Although many of the resources<br />

referenced in this chapter contain large, complex cores of primary materials, this is not<br />

necessary <strong>to</strong> achieve contextual mass. For instance, the Decameron Web project, a<br />

collection devoted <strong>to</strong> the literary, his<strong>to</strong>rical, and cultural context of Boccaccio's famous<br />

text, contains an established critical edition with translations and a selection of related<br />

materials, such as annotations, commentaries, critical essays, maps, and bibliographies.


The pedagogical intent of the site is obvious in its content and layout, but it is<br />

simultaneously strong as a research context.<br />

A number of existing thematic collections exemplify the notion of contextual mass in<br />

their depth and complexity, as well as in their explicit goals. The core of the Rossetti<br />

archive is intended <strong>to</strong> be all of Rossetti's texts and pic<strong>to</strong>rial works, and this set of primary<br />

works is complemented by a corpus of contextual materials that includes other works<br />

from the period, family letters, biography, and contemporary secondary materials. In the<br />

Blake archive, "contextual" information is at the heart of the scholarly aims of the<br />

project. The documentation at the website explains that works of art make sense only in<br />

context. In this case creating a meaningful context involves presenting the texts with the<br />

illustrations, illuminated books in relation <strong>to</strong> other illuminated books, and putting those<br />

<strong>to</strong>gether with other drawings and paintings. All of this work is then presented in the<br />

context of relevant his<strong>to</strong>rical information.<br />

Collaboration is required in the creation of contextually rich thematic collections. Instead<br />

of being a patron of existing collections, scholars must partner with libraries, museums,<br />

and publishers <strong>to</strong> compile diverse materials that are held in different locations. For<br />

example, collections from the Bos<strong>to</strong>n Public Library, the New York Public Library, the<br />

Massachusetts His<strong>to</strong>rical Society, the Massachusetts Archives, and the Peabody Essex<br />

Museum were melded <strong>to</strong> create the Salem Witch Trials collection. The Dickinson<br />

Electronic Archives, produced by a collective that includes four general edi<strong>to</strong>rs who work<br />

collabora-tively with co-edi<strong>to</strong>rs, staff, and users, has reproduced works housed in library<br />

archives all along the northeast corridor of the United States (Smith 1999).<br />

Interdisciplinary Platform<br />

Humanities scholars have long been engaged in interdisciplinary inquiry, but the library<br />

collections they have relied on have been developed around academic structures that tend<br />

<strong>to</strong> obscure connections between fields of research. This is partly because of the large<br />

scale of research libraries, but also because of the inherent difficulties in maintaining<br />

standard organization and access systems for materials that represent a complex base of<br />

knowledge. The continuing growth of interdisciplinary research is a recognized challenge<br />

<strong>to</strong> collection development and the provision of information services in research libraries,<br />

and recent studies have identified specific practices of interdisciplinary <strong>humanities</strong><br />

scholars that need <strong>to</strong> be better supported in the next generation of <strong>digital</strong> resources<br />

(Palmer 1996; Palmer and Neumann 2002). Crea<strong>to</strong>rs of thematic collections are<br />

beginning <strong>to</strong> address some of these needs through the conscious development of<br />

interdisciplinary platforms for research.<br />

A number of collections have been explicitly designed <strong>to</strong> be conducive <strong>to</strong><br />

interdisciplinary research and have effectively incorporated the interests of diverse<br />

intellectual communities. The Thomas MacGreevy Archive aims <strong>to</strong> promote inquiry in<strong>to</strong><br />

the interconnections between literature, culture, his<strong>to</strong>ry, and politics by blurring the<br />

boundaries that separate the different fields of study. Monuments and Dust, a thematic<br />

collection focused on Vic<strong>to</strong>rian London, states its interdisciplinary intent in the project


subtitle: "New Technologies and Sociologies of Research." A stated objective of the<br />

project is <strong>to</strong> foster international collaboration and intellectual exchange among scholars<br />

in literature, architecture, painting, journalism, colonialism, modern urban space, and<br />

mass culture. The premise is that the aggregation of diverse sources – images, texts,<br />

numerical data, maps, and models – will seed intellectual interaction by making it<br />

possible <strong>to</strong> discover new visual, textual, and statistical relationships within the collection<br />

and between lines of research.<br />

Activity Support<br />

The functions of thematic collections are being greatly expanded as crea<strong>to</strong>rs add <strong>to</strong>ols <strong>to</strong><br />

support research activities. Humanities research processes involve numerous activities,<br />

and collections are essential <strong>to</strong> many of them. Scholarly information seeking is one type<br />

of activity that has been studied empirically for some time, and our understanding of it is<br />

slowly informing the development of information retrieval systems in the <strong>humanities</strong><br />

(Bates 1994). Other significant research practices, especially those involved in<br />

interpretation and analysis of sources, have received less attention.<br />

Reading is of particular importance in the <strong>humanities</strong>, and the technologies being<br />

developed for collections are beginning <strong>to</strong> address the complexities of how and why<br />

people read texts (Crane et al. <strong>2001</strong>). Scanning a text is a different activity from rereading<br />

it deeply and repeatedly over time, and it is but one stage of a larger pattern of wide<br />

reading and collecting practiced by many <strong>humanities</strong> scholars (Brockman et al. <strong>2001</strong>).<br />

Other common scholarly activities, such as the "scholarly primitives" identified by<br />

Unsworth (2000a), have yet <strong>to</strong> be adequately supported by the <strong>digital</strong> resources designed<br />

for scholars. These include the basic activities of annotating, comparing, referring,<br />

selecting, linking, and discovering that are continually carried out by scholars as part of<br />

the complex processes of reading, searching, and writing. Just as materials can be<br />

structured for scholarly purposes as we transform our bodies of texts in<strong>to</strong> <strong>digital</strong> format,<br />

<strong>to</strong>ols can be tailored for specific scholarly tasks.<br />

Searching collection content is a standard function of databases of all kinds, and among<br />

thematic collections there is considerable variation in the level of retrieval supported.<br />

Image searching has been a vital area of development, since many collections contain a<br />

significant number of facsimile images and other pic<strong>to</strong>rial works. In the Blake archive,<br />

for instance, access <strong>to</strong> images is enhanced through extensive description of each image<br />

and the application of a controlled vocabulary. Likewise, the Rossetti archive project has<br />

been dedicated <strong>to</strong> developing methods for formally coding features of images <strong>to</strong> make<br />

them amenable <strong>to</strong> full search and analysis (McGann 1996). Imagesizer, a <strong>to</strong>ol developed<br />

at IATH and provided <strong>to</strong> both the Blake and Rossetti archives, allows the user <strong>to</strong> view<br />

images in their original size or in other convenient dimensions. Inote, another IATH <strong>to</strong>ol<br />

configured for the Blake project, allows viewing of illustrations, components of<br />

illustrations, and image descriptions. It also allows users <strong>to</strong> create their own personal<br />

annotations along with saved copies of an image. This, in particular, is a key<br />

advancement in activity support for collections since it goes beyond searching, viewing,<br />

and linking <strong>to</strong> assist scholars with the basic tasks of interpretation and note-taking,


activities that have been largely neglected in <strong>digital</strong> resource development. Tools <strong>to</strong><br />

support scholarly tasks are also being developed independent of thematic collection<br />

initiatives. For example, the Version-ing Machine software designed by the Maryland<br />

Institute for Technology in the Humanities (MITH) blends features of the traditional book<br />

format with electronic publishing capabilities <strong>to</strong> enhance scholars' interpretive work with<br />

multiple texts.<br />

Hypertext has been a monumental advancement in the functionality of collections, and<br />

many current projects are working <strong>to</strong>ward extensive interlinking among aggregated<br />

materials. The Walt Whitman Hypertext Archive intends <strong>to</strong> link sources <strong>to</strong> demonstrate<br />

the numerous and complex revisions Whitman made <strong>to</strong> his poems. The items associated<br />

with a poem might include the initial notes and trial lines in a notebook, a published<br />

version from a periodical, publisher's page proofs, and various printed book versions. The<br />

Bolles Collection on the His<strong>to</strong>ry of London, part of the larger Perseus <strong>digital</strong> library,<br />

exploits hypertext by presenting full his<strong>to</strong>rical texts on London with hyperlinks from<br />

names of <strong>to</strong>wns, buildings, and people <strong>to</strong> correlate items, such as pho<strong>to</strong>graphs and maps<br />

of places and drawings and biographies of people.<br />

Mapping and modeling <strong>to</strong>ols are valuable features of a number of place-based thematic<br />

collections. The Bolles Collection provides electronic timelines for visualizing how time<br />

is represented within and across documents. Also, his<strong>to</strong>rical maps of London have been<br />

integrated with a current geographic information system (GIS) <strong>to</strong> allow users <strong>to</strong> view the<br />

same location across the maps (Crane et al. <strong>2001</strong>). Modeling adds an important layer of<br />

data <strong>to</strong> the Monuments and Dust collection. A variety of VRML (Virtual Reality<br />

Modeling Language) images of the Crystal Palace have been developed from engineering<br />

plans and other sources of data <strong>to</strong> create models of the building and small architectural<br />

features such as drains, trusses, and wall panels, as well as animations of the building's<br />

lighting and other three-dimensional replicas.<br />

The Humanities Labora<strong>to</strong>ry<br />

The thematic collections concentrating on contextual mass and activity support are<br />

coming closest <strong>to</strong> creating a labora<strong>to</strong>ry environment where the day-<strong>to</strong>-day work of<br />

scholars can be performed. As with scientific labora<strong>to</strong>ries, the most effective places will<br />

be those that contain the materials that need <strong>to</strong> be studied and consulted during the course<br />

of an investigation as well as the instrumentation <strong>to</strong> carry out the actual work. For<br />

<strong>humanities</strong> scholars, a well-equipped labora<strong>to</strong>ry would consist of the sources that would<br />

be explored, studied, annotated, and gathered in libraries and archives for an area of<br />

research and the means <strong>to</strong> perform the reading, analyzing, interpreting, and writing that<br />

would normally take place in their offices. The most successful of these sites will move<br />

beyond the thematic focus <strong>to</strong> provide contextual mass and activity support that is not only<br />

responsive <strong>to</strong> what scholars currently do, but also <strong>to</strong> the questions they would like <strong>to</strong> ask<br />

and the activities they would like <strong>to</strong> be able <strong>to</strong> undertake.<br />

In the sciences the virtual labora<strong>to</strong>ry, or collabora<strong>to</strong>ry, concept has been around for some<br />

time. Traditional labora<strong>to</strong>ries that are physically located encourage interaction and


cooperation within teams, and collabora<strong>to</strong>ries extend that dimension of research <strong>to</strong><br />

distributed groups that may be as small as a work group or as large as an international<br />

research community. Collabora<strong>to</strong>ries are designed as media-rich networks that link<br />

people <strong>to</strong> information, facilities, and other people (Finholt 2002). They are places where<br />

scientists can obtain resources, do work, interact, share data, results, and other<br />

information, and collaborate.<br />

Collaborative processes have not been a significant fac<strong>to</strong>r in technology development in<br />

the <strong>humanities</strong>, due at least in part <strong>to</strong> the prevalent notion that <strong>humanities</strong> scholars work<br />

alone. This is true <strong>to</strong> some degree. Reading, searching databases, and browsing<br />

collections are solitary activities, and most articles and books are written by individuals.<br />

However, most <strong>humanities</strong> scholars do work <strong>to</strong>gether in other important ways. They<br />

frequently share citations, ideas, drafts of papers, and converse about research in<br />

progress, and these interactions are dependent on strong relationships with others in an<br />

intellectual community. For most scholars, this type of collaborative activity is a<br />

necessary part of the daily practice of research, and it has been shown <strong>to</strong> be especially<br />

vital for interdisciplinary work (Palmer and Neumann 2002).<br />

An increasing number of thematic projects are encouraging scholars <strong>to</strong> share their<br />

research through the submission of content and by providing forums for dialogue among<br />

crea<strong>to</strong>rs and the user community. A few initiatives are aimed at enabling collaboration<br />

among scholars, as with the case of Monuments and Dust, noted above. In other projects,<br />

the corpus brought <strong>to</strong>gether is considered a resource <strong>to</strong> foster collaboration among<br />

scholars, students, and lay readers. While it is not a thematic collection per se, Collate is<br />

a unique resource that provides <strong>to</strong>ols for indexing, annotation, and other types of work<br />

with <strong>digital</strong> resources. This international project has been designed <strong>to</strong> support the<br />

development of collections of digitized his<strong>to</strong>ric and cultural materials, but its other<br />

primary goal is <strong>to</strong> demonstrate the viability of the collabora<strong>to</strong>ry in the <strong>humanities</strong>.<br />

The Process of Collocation<br />

All collections, either physical or virtual, are formed through collocation, the process of<br />

bringing <strong>to</strong>gether related information (Taylor 1999). Collocation is a useful term because<br />

it emphasizes the purpose of collection building and can be applied <strong>to</strong> the different means<br />

used <strong>to</strong> unite materials. The collocation of research materials can take many forms.<br />

Anthologies are collocations of selected works on a subject. In a traditional archive,<br />

collocation is based on the originating source – the person or institution – and these<br />

collections are acquired and maintained as a whole. Collocation is often associated with<br />

physical location, such as when materials by the same author are placed <strong>to</strong>gether on<br />

shelves in a library. It is not surprising that some thematic collections have adopted the<br />

metaphor of physical collocation. For example, the Decameron Web describes its<br />

collection as a "specialized bookshelf or mini-library" generated from Boccaccio's<br />

masterpiece. The Tibetan and Himalayan Digital Library describes its collections as the<br />

equivalent of the "stacks" in a traditional library. A library catalogue also provides<br />

collocation by bringing <strong>to</strong>gether like materials through a system of records and<br />

references.


To see a substantial portion of the works associated with a particular author or <strong>to</strong>pic, it<br />

has been necessary for scholars <strong>to</strong> consult many catalogues and indexes and travel <strong>to</strong><br />

different libraries, copying and collecting what they can along the way. In the case of<br />

fragile items, handling is limited and pho<strong>to</strong>copying or microfilming may be prohibited.<br />

With thematic collections, scholars are now exercising the power of virtual collocation.<br />

By pulling <strong>to</strong>gether materials that are part of various works and located in reposi<strong>to</strong>ries at<br />

different sites, they collocate deep, sophisticated collections of sources that can be used at<br />

a convenient time and place. For example, the direc<strong>to</strong>ry of the Rossetti archive, which<br />

lists pictures, poems, prose, illustrated texts, double works, manuscripts, books,<br />

biography, bibliography, chronology, and contexts, illustrates the diversity and richness<br />

that can be achieved through the collocation of <strong>digital</strong> materials.<br />

The physical proximity of resources becomes trivial when the material is <strong>digital</strong> and<br />

made available in a networked information system (Lagoze and Fielding 1998), but the<br />

intellectual and technical work of selecting and structuring meaningful groupings of<br />

materials remains critical. This is especially true in the <strong>humanities</strong>, where research often<br />

concentrates on the documents and artifacts created by or surrounding an object of study.<br />

Compared with other fields of study, the sources used are highly heterogeneous and<br />

wide-ranging, and their value persists over time, rather than dissipating through<br />

obsolescence. Over the course of a scholar's research project, certain manuscripts or<br />

artifacts in foreign archives may be essential, but a standard edition or a popular culture<br />

website can be equally important (Palmer and Neumann 2002). The distributed, dynamic<br />

research collections that can be created on the Web are attuned with the nature of<br />

<strong>humanities</strong> fields, which are "concerned with the construction of knowledge from sources<br />

of different types, scattered across different subject areas" (Fraser 2000: 274).<br />

The principles that guide the collocation of research collections in libraries are different<br />

from scholarly motivations for collocation. Library collections are amassed for<br />

preservation, dispensing, bibliographic, and symbolic purposes (Buckland 1992). The<br />

process of collecting is ruled first by the mission of the institution and more specifically<br />

by the selection criteria developed <strong>to</strong> support research and teaching in a given subject<br />

area. In contrast, collections produced by scholars are cus<strong>to</strong>mized <strong>to</strong> the research focus of<br />

a scholarly community or <strong>to</strong> the specific interests of the crea<strong>to</strong>rs. Thus, the "principles of<br />

inclusion" for The William Blake Archive – that designate the illuminated books as the<br />

foundation and the strategy of adding clusters of materials based on medium, theme, or<br />

his<strong>to</strong>ry – are idiosyncratic <strong>to</strong> that particular project. Schreibman suggests that the Blake<br />

archive and other early thematic collections, as well as broader collection initiatives such<br />

as the CELT Project and the Women Writers Project, have been governed by a <strong>digital</strong><br />

library model that collocates "previously published texts based on a theory of collection<br />

appropriate <strong>to</strong> the particular archive" (2002: 287). While a loose theory of collecting may<br />

be guiding crea<strong>to</strong>rs' selection of content, the criteria being used <strong>to</strong> determine what is<br />

appropriate for a collection and the long-term development principles of a project are not<br />

always clarified for users of thematic collections.<br />

The potential of <strong>digital</strong> collocation has been restrained by copyright concerns. It is much<br />

less complicated <strong>to</strong> digitize and redistribute sources that do not have copyright


estrictions, and therefore older materials in the public domain have been more widely<br />

selected for <strong>digital</strong> collections of all kinds. Increasingly, thematic collection crea<strong>to</strong>rs are<br />

working through copyright requirements for published works, as well as adding new,<br />

born <strong>digital</strong> sources, <strong>to</strong> build systematic and principled collections that meet their<br />

scholarly aims. Again, The William Blake Archive is one of the projects that offer a sound<br />

model on this front. They have gained access <strong>to</strong> valuable and important materials for their<br />

collection by working closely with museums, libraries, and collec<strong>to</strong>rs <strong>to</strong> address their<br />

copyright concerns.<br />

Research Collections Landscape<br />

Scholarly thematic collections are a new addition <strong>to</strong> the array of existing and emerging<br />

research collocations. Interestingly, many thematic collections created by scholars refer<br />

<strong>to</strong> themselves as archives, but conventional archives differ in important ways, especially<br />

in terms of their mission and the methods they use <strong>to</strong> organize materials. The collections<br />

held in archival reposi<strong>to</strong>ries document the life and work of institutions or individuals. An<br />

archive's role is <strong>to</strong> "preserve records of enduring value that document organizational and<br />

personal activities accumulated in the course of daily life or work" (Taylor 1999: 8).<br />

Archival collections are collocated according <strong>to</strong> provenance – the individual or corporate<br />

origina<strong>to</strong>r – and organized in original working order. As accumulations of materials from<br />

regular daily life, these collections may contain print and electronic documents and<br />

artifacts of any kind, including meeting minutes, annual reports, memoranda, deeds,<br />

manuscripts, pho<strong>to</strong>graphs, letters, diaries, printed books, and audio recordings.<br />

Thematic collections are more analogous <strong>to</strong> the subject collections traditionally<br />

developed in research libraries than they are <strong>to</strong> archives. Examples of research library<br />

subject collections include the Chicano Studies collection at the University of California<br />

at Berkeley and the His<strong>to</strong>rical Linguistics collection at the Newberry Library in Chicago.<br />

A standard direc<strong>to</strong>ry lists 65,818 library and museum subject collections in the United<br />

States and Canada (Ash 1993), many of which are thematic in scope. For example, the<br />

entry for William Blake lists 20 collections, two of which are contribu<strong>to</strong>rs <strong>to</strong> The William<br />

Blake Archive project. Subject collections are sometimes developed cooperatively by<br />

multiple institutions, and these tend <strong>to</strong> cover broad academic or geographic categories.<br />

Examples include the Urban Studies collection at the Center for Research Libraries, a<br />

membership organization devoted <strong>to</strong> cooperative collection programs, and the East Asian<br />

collections cooperatively developed at the University of North Carolina and Duke<br />

University.<br />

Localized collections that contain rare or valuable items may be kept as part of a special<br />

collection. Special collections departments develop concentrated subject and theme-based<br />

collections that include substantial primary materials. They are also the place where<br />

manuscripts, papers, and other unique, fragile, or valuable items are maintained and<br />

segregated for restricted access. As research libraries began <strong>to</strong> make materials accessible<br />

via the Web, the contents of special collections were often the first <strong>to</strong> be selected for<br />

digitization. Research libraries have been eager <strong>to</strong> share their treasures with a wider<br />

audience and make them more convenient <strong>to</strong> view, and offering a <strong>digital</strong> alternative


decreases handling and the wear and tear on valuable materials. The first digitized special<br />

collections released by the Library of Congress in 1994 through their American Memory<br />

project were pho<strong>to</strong>graphic collections. The initiative has since grown <strong>to</strong> offer over 100<br />

online collections, many of which are thematic and multimedia. Most are based on<br />

existing special collections within the Library of Congress, but some, such as Band Music<br />

from the Civil War Era and the American Variety Stage, are thematic collections that<br />

have been collocated for the first time specifically for online presentation. Many other<br />

institutions have selected notable special collections for digitization. For instance, the<br />

Academic Affairs Library at the University of North Carolina at Chapel Hill has<br />

produced Documenting the American South, a <strong>digital</strong> collection based on their existing<br />

Southern His<strong>to</strong>rical Collection, one of the largest s<strong>to</strong>res of Southern manuscripts in the<br />

country.<br />

As research libraries continue <strong>to</strong> undertake these projects, substantial bodies of<br />

previously hidden source material are coming in<strong>to</strong> public view. These <strong>digital</strong> collections<br />

make an important contribution <strong>to</strong> <strong>digital</strong> research library development and provide<br />

potential raw materials for the construction of thematic collections and other aggregations<br />

of <strong>digital</strong> content. Digital special collections provide an important service for researchers,<br />

but they generally do not possess the range of scholarly functions – scholarly<br />

contribution, contextual mass, interdisciplinary platform, and activity support – provided<br />

by many thematic collections.<br />

Digital Collections Terminology<br />

There is little consistency in the terms used <strong>to</strong> describe <strong>digital</strong> resources, and the number<br />

of terms and the overlap between them seem <strong>to</strong> be increasing. In addition <strong>to</strong> being a new<br />

conceptualization of research collection development, the phrase "thematic research<br />

collection" is itself a new addition <strong>to</strong> the vocabulary. As discussed above, the term<br />

"archive" is being widely applied <strong>to</strong> thematic collections, and the adoption of the word<br />

has merit for this purpose, since, like traditional archives, scholarly thematic collections<br />

tend <strong>to</strong> focus on primary sources and emphasize the importance of the physical object.<br />

For example, The Dickinson Electronic Archives prioritizes the physical object by<br />

representing Emily Dickinson's poems, letters, letter-poems, drafts, fragments, and<br />

manuscripts as facsimile images. Likewise, in the William Blake Archive the physical<br />

nature of the artifacts is a central thrust. The <strong>digital</strong> collection represents the integration<br />

of Blake's illustrations and texts and the variations among different copies of his books,<br />

features that have not been well represented in printed editions of his work. As with most<br />

thematic collections, the actual goals of the Blake project reach beyond those of a<br />

traditional archive, where the central aim would be <strong>to</strong> preserve the physical record of<br />

production. Here the notion of the archive has been extended <strong>to</strong> include the catalogue,<br />

scholarly edition, database, and <strong>to</strong>ols that work <strong>to</strong>gether <strong>to</strong> fully exploit the advantages of<br />

the <strong>digital</strong> medium (Viscomi 2002).<br />

It has been suggested that electronic archives will increasingly take the form of hypertext<br />

editions, similar <strong>to</strong> the Electronic Variorum Edition of Don Quixote being developed at<br />

the Center for the Study of Digital Libraries at Texas A & M University (Urbina et al.


2002). But, at present, many kinds of resources, including journal article pre-print servers<br />

and lists of links on a web page, are being referred <strong>to</strong> as <strong>digital</strong> or electronic archives.<br />

The traditional, professional archive still holds an important place in the array of research<br />

collections, and some of these are being <strong>digital</strong>ly reformatted while retaining their<br />

original aims and organizational methods. At the same time, colloquial applications of the<br />

term are increasing and new scholarly idealizations of the concept are evolving.<br />

The vocabulary of <strong>digital</strong> resources has been further complicated by the wide usage of the<br />

term "<strong>digital</strong> library" for all kinds of <strong>digital</strong> collections, essentially blurring the<br />

distinction between collections and libraries. Of the many definitions of <strong>digital</strong> libraries<br />

circulating in the literature, several can be readily applied <strong>to</strong> thematic collections.<br />

However, a widely accepted conception in the field of library and information science<br />

clarifies the difference:<br />

Digital libraries are organizations that provide the resources, including the specialized<br />

staff, <strong>to</strong> select, structure, offer intellectual access <strong>to</strong>, interpret, distribute, preserve the<br />

integrity of, and ensure the persistence over time of collections of <strong>digital</strong> works so that<br />

they are readily and economically available for use by a defined community or set of<br />

communities.<br />

(Waters 1998)<br />

A thematic collection is not a library in the organizational sense; it is a collection that<br />

may be developed or selected for inclusion in a <strong>digital</strong> library, or it may exist separately<br />

from any library or similar institution. A library contains a collection of collections and<br />

has an institutional commitment <strong>to</strong> services that ensure access and persistence. Because<br />

of their size and diverse user population, libraries, including <strong>digital</strong> libraries, generally<br />

lack the coherency and the functional features characteristic of thematic collections.<br />

In the <strong>humanities</strong>, the Perseus project is considered an exemplar <strong>digital</strong> library. It serves<br />

as a good, albeit complex, example of the relationship between collections and libraries.<br />

The scope of Perseus was originally disciplinary rather than thematic, providing access <strong>to</strong><br />

an immense integrated body of materials in Classics, including primary Greek texts,<br />

translations, images, and lexical <strong>to</strong>ols (Fraser 2000). As the project has grown, as both a<br />

scholarly resource and a research initiative, it has added collections outside the realm of<br />

Classics. The mix of subject collections within Perseus represents an interesting variety<br />

of collection-building approaches in terms of scope and mode of creation. Three<br />

geographically oriented collections, California, Upper Midwest, and Chesapeake Bay,<br />

have been developed in association with the Library of Congress's American Memory<br />

project. The library also contains an extensive body of primary and secondary materials<br />

covering the early modern period. The best example of a thematic collection within<br />

Perseus is the pre-twentieth-century London segment based on the Bolles Collection on<br />

the His<strong>to</strong>ry of London. It is a digitized recreation of an existing special collection that is<br />

homogeneous in theme but heterogeneous in content. As noted previously, it interlinks<br />

maps of London, relevant texts, and his<strong>to</strong>rical and contemporary illustrations of the city.


The Tibetan and Himalayan Digital Library (THDL) is another kind of <strong>digital</strong> library/<br />

thematic collection hybrid. It is being designed <strong>to</strong> capitalize on internal collocation of an<br />

underlying base of holdings <strong>to</strong> create multiple collections with different structures and<br />

perspectives. For example, the Environment and Cultural Geography collection organizes<br />

the library's texts, videos, images, maps and other types of materials according <strong>to</strong> space<br />

and time attributes. The thematic and special collections are organized by subject<br />

attributes. The categorization scheme used in the THDL is an interesting case of variant<br />

applications of <strong>digital</strong> resource terminology. Its special collections are thematic, with a<br />

focus, for example, on the life or activities of an individual, while the thematic<br />

collections integrate diverse sources within broad disciplinary units, such as Art,<br />

Linguistics, Literature, or Music. Subtheme collections, which are independent projects<br />

with their own content and goals, are nested within the thematic collections.<br />

Needless <strong>to</strong> say, there is much overlap between <strong>digital</strong> library and thematic collection<br />

efforts, and variations and hybrids will continue <strong>to</strong> evolve along with the terminology.<br />

Digital research libraries will no doubt continue <strong>to</strong> acquire the collections built by<br />

scholars, collaborative teams, and institutions, while scholars' projects grow <strong>to</strong> nest and<br />

annex <strong>digital</strong> special collections. An important outcome of this activity is that expert<br />

collocation of research materials by scholars is adding an important new layer of<br />

resources <strong>to</strong> <strong>humanities</strong> research collections.<br />

Turn in the Collection Cycle<br />

In the past, scholars used collections for their research and contributed <strong>to</strong> collections as<br />

authors, but their role as collection builders was limited. They developed significant<br />

personal collections for their own purposes, and they collocated materials by editing and<br />

publishing collected works. On the other hand, collection development has long been a<br />

significant part of the professional responsibilities of librarians, archivists, and cura<strong>to</strong>rs.<br />

The interaction between the scholarly community and collection professionals has been<br />

an important influence on the development of research library collections. As the primary<br />

constituency of research libraries, scholars' questions, requests, and ongoing research and<br />

teaching activities have guided the collection processes at research institutions. Of<br />

course, the essential contribution of scholars has been as crea<strong>to</strong>rs of intellectual works<br />

that make up a large proportion of research collections. Now scholars have also become<br />

crea<strong>to</strong>rs of research collections, and this change will have an important impact on how<br />

our vast arrays of research materials take shape in the future.<br />

Where libraries once acquired the documents authored by scholars, they now also need <strong>to</strong><br />

collect the thematic research collections created by scholars. This genre has new qualities<br />

that cannot be treated in the same way as a printed or electronic single work, and the<br />

interdisciplinary, multimedia, and open-ended characteristics of the resources, further<br />

complicate matters. Libraries are not yet systematically collecting the collections<br />

produced by scholars, in part because of the newness of the genre, but also because this<br />

type of meta-collecting is an unfamiliar practice. In fact, most research libraries do not<br />

yet collect and catalogue non-commercial, web-based <strong>digital</strong> materials of any kind. For<br />

example, at the time of this writing, WorldCat, a major bibliographic database of library


holdings, indicated that 263 libraries had purchased and catalogued a recent book of<br />

criticism on William Blake by William Vaughan. In contrast, only 26 libraries had added<br />

the William Blake Archive collection <strong>to</strong> their catalogue. As a point of reference, 34<br />

libraries had catalogued Voice of the Shuttle, a <strong>humanities</strong> gateway that is widely used<br />

but less similar <strong>to</strong> the scholarly creations traditionally collected by libraries than the<br />

materials in a typical thematic collection. As research libraries begin <strong>to</strong> regularly acquire<br />

and catalogue thematic collections, they will be interjecting a new layer of collecting<br />

activities and causing a shift in the scholarly information transfer cycle.<br />

The traditional cycle of document transfer as conceptualized before the advent of <strong>digital</strong><br />

documents (King and Bryant 1971) required publishers and libraries <strong>to</strong> take documents<br />

through most steps of the process. A large part of the production and distribution of<br />

scholarly research materials still adheres <strong>to</strong> this cycle. It begins with use of a document<br />

for scholarly work, usually in conjunction with many other documents, which leads <strong>to</strong><br />

composition by the scholar. At this point, the scholar becomes an author, in addition <strong>to</strong> an<br />

information user. A publisher handles the reproduction or printing and distribution phase.<br />

Libraries move published documents through the circuit by selecting, acquiring,<br />

organizing, and s<strong>to</strong>ring them, and then by making them accessible, usually on library<br />

shelves and through representation in catalogues and indexes. Use of the materials is<br />

enhanced by assistance from reference, instruction, and access services provided by the<br />

library. The cycle is completed when the material becomes part of the research process<br />

by being accessed and assimilated by a scholar.<br />

To some degree, libraries can treat thematic collections like the documents produced by<br />

scholars – by selecting, acquiring, and organizing them, as has been done by the 26<br />

libraries that have catalogued the Blake archive. Over time, the distribution of scholarly<br />

materials in the <strong>humanities</strong> may be greatly expanded by scholars as they take up selection<br />

and organization activities. Perhaps even more importantly, scholars are adding valuable<br />

features <strong>to</strong> collections as they cus<strong>to</strong>mize them for their scholarly purposes. While<br />

research libraries strive <strong>to</strong> meet the information needs of the communities they serve,<br />

they are not equipped or charged <strong>to</strong> fully support the scholarly process. Selection criteria<br />

for collections in research libraries emphasize how <strong>to</strong> choose the best items from the<br />

universe of publications being produced relative <strong>to</strong> the research communities being<br />

served. Measurements of satisfaction, circulation, and Web activity are combined with<br />

librarians' knowledge of their scholarly constituencies, which grows based on what<br />

scholars ask for and what they reveal <strong>to</strong> librarians about their interests and projects. Less<br />

attention has been paid <strong>to</strong> assessing how <strong>to</strong> prioritize materials in terms of what scholars<br />

do, what they value, or what would be the most likely <strong>to</strong> enhance specific research areas.<br />

Many research libraries are currently focusing on global approaches <strong>to</strong> <strong>digital</strong> collection<br />

building by producing expansive gateways for all their user communities. At the same<br />

time, researchers are creating their own reposi<strong>to</strong>ries and <strong>to</strong>ols, highly cus<strong>to</strong>mized <strong>to</strong> the<br />

scholarly work of their intellectual communities. Research libraries will need <strong>to</strong> fill the<br />

gap by developing mid-range collection services that actively collocate thematic<br />

collections within meaningful aggregations. The profile of a mid-level research collection<br />

would look quite different from the current <strong>digital</strong> research library. It would not prioritize


the <strong>to</strong>p tier of scholarly journals, the major indexes, a large general set of reference<br />

materials, or disciplinary canons. Instead, it would provide access <strong>to</strong> constellations of<br />

high-quality thematic research collections that are aligned with the scholarly activities<br />

conducted at the institution.<br />

Scholar-created research collections are likely <strong>to</strong> increase in number as the work of<br />

producing them becomes more widely accepted as legitimate scholarship. Research<br />

libraries have yet <strong>to</strong> grasp how this will impact their practices, and it may be some time<br />

before there is a confluence of scholar- and institution-generated collections. First there<br />

will need <strong>to</strong> be a wider awareness of thematic collections as an important mode of<br />

scholarly work. Scholars and scientists are producing an abundance of <strong>digital</strong> products,<br />

many of which are important, high-quality compilations, and these activities are<br />

proliferating through support from funding agencies. It will be necessary for research<br />

libraries <strong>to</strong> respond <strong>to</strong> this trend in their collection development programs. Just as<br />

importantly, as collection building grows as a form of scholarly production, universities<br />

will need <strong>to</strong> provide resources <strong>to</strong> assist in this form of research. At present, the materials<br />

and expertise required for collection building research tend <strong>to</strong> be thinly scattered across<br />

departments, libraries, and computing centers. Resources and support services would be<br />

best centralized in the library or in auxiliary research units where scholars from all fields<br />

can turn for assistance in developing content and <strong>to</strong>ols.<br />

Conclusion<br />

As scholars gain mastery in <strong>digital</strong> collocation and produce innovative research<br />

environments, they are practicing a new kind of collection development. Thematic<br />

collections are conceived not only as support for scholarship but as contributions <strong>to</strong><br />

scholarship. They provide configurations of research materials that strongly represent the<br />

relationships between different kinds of sources and different subject areas. Through<br />

contextual mass, interdisciplinary platform, and activity support, thematic collections add<br />

density, flexibility, and interactivity <strong>to</strong> previously scattered and static reposi<strong>to</strong>ries of<br />

content. They assist in the production of new research, but they also have the potential <strong>to</strong><br />

substantively improve the scholarly research process.<br />

In thematic collections, research materials are closely tied <strong>to</strong> the processes of inquiry,<br />

making the con<strong>to</strong>urs of scholarship more visible as they are inscribed in<strong>to</strong> the collection.<br />

The questions and methods that propel scholarship become part of the representation, and<br />

as scholars build the partnerships it takes <strong>to</strong> construct quality collections, the networks of<br />

researchers and institutions involved in a research area become more explicit. Thematic<br />

collections are a substantive contribution <strong>to</strong> the rebuilding of research resources in the<br />

<strong>digital</strong> age, adding richness <strong>to</strong> our expansive s<strong>to</strong>res of materials and new opportunities for<br />

<strong>humanities</strong> scholarship.<br />

See also Chapter 36: The Past, Present, and Future of Digital Libraries.<br />

References for Further Reading


Ash, L. (1993). Subject Collections, 7th edn. New Providence, NJ: R. R. Bowker.<br />

Atkinson, R. (1998). Managing Traditional Materials in an Online Environment: Some<br />

Definitions and Distinctions for a Future Collection Management. Library Resources and<br />

Technical Services 42: 7–20.<br />

Bates, M. J. (1994). The Design of Databases and Other Information Resources for<br />

Humanities Scholars: The Getty Online Searching Project Report no. 4. Online and<br />

CDROM Review 18: 331–40.<br />

Brockman, W. S., L. Neumann, C. L. Palmer, and T. Tidline (<strong>2001</strong>). Scholarly Work in<br />

the Humanities and the Evolving Information Environment. Washing<strong>to</strong>n, DC: Digital<br />

Library Federation and the Council on Library and Information Resources. Accessed<br />

November 26, 2002. At http://www.clir.org/pubs/reports/publ04/contents.html.<br />

Buckland, M. (1992). Collections Reconsidered. In Redesigning Library Services: A<br />

Manifes<strong>to</strong> (pp. 54–61). Chicago: American Library Association.<br />

Buckland, M. (1995). What Will Collection Developers Do? Information Technologies<br />

and Libraries 14: 155–9.<br />

Condron, F., M. Fraser, and S. Sutherland (<strong>2001</strong>). Oxford University Computing Services<br />

Guide <strong>to</strong> Digital Resources for the Humanities. Morgan<strong>to</strong>wn, WV: West Virginia<br />

University Press.<br />

Crane, G. and J. A. Rydberg-Cox (2000). New Technology and New Roles: The Need for<br />

"Corpus Edi<strong>to</strong>rs." Proceedings of the Fifth ACM Conference on Digital Libraries (pp.<br />

252–3), June 2–7, San An<strong>to</strong>nio, Texas.<br />

Crane, G., C. E. Wulfman, and D. A. Smith (<strong>2001</strong>). Building a Hypertextual Digital<br />

Library in the Humanities: A Case Study of London. Proceedings of the ACM/IEEE Joint<br />

Conference on Digital Libraries (pp. 426–34), June 24–28, Roanoke, Virginia.<br />

Finholt, T. (2002). Collabora<strong>to</strong>ries. Annual Review of Information Science and<br />

Technology 36: 73–107.<br />

Fraser, M. (2000). From Concordances <strong>to</strong> Subject Portals: Supporting the Text-centred<br />

Humanities Community. Computers and the Humanities 34: 265–78.<br />

King, D. W. and E. C. Bryant (1971). The Evaluation of Information Services and<br />

Products. Washing<strong>to</strong>n, DC: Information Resources Press.<br />

Lagoze, C. and D. Fielding (1998). Defining Collections in Distributed Digital Libraries.<br />

D-Lib Magazine, November. Accessed November 26, 2002. At<br />

http://www.dlib.org.dlib/november98/lagoze/11lagoze.html.


Lee, H.-L. (2000). What Is a Collection? Journal of the American Society for Information<br />

Science 51:1106–13.<br />

McGann, J. (1996). The Rossetti Archive and Image-based Electronic Editing. In R. J.<br />

Finneran (ed.), The Literary Text in the Digital Age (pp. 145–83). Ann Arbor, MI:<br />

University of Michigan Press.<br />

Palmer, C. L. (ed.) (1996). Navigating among the Disciplines: The Library and<br />

Interdisciplinary Inquiry. Library Trends 45.<br />

Palmer, C. L. (2000). Configuring Digital Research Collections around Scholarly Work.<br />

Paper presented at Digital Library Federation Forum, November 19, Chicago, Illinois.<br />

Accessed November 26, 2002. At http://www.diglib.org/forums/fallOO/palmer.htm.<br />

Palmer, C. L., and L. J. Neumann (2002). The Information Work of Interdisciplinary<br />

Humanities Scholars: Exploration and Translation. Library Quarterly 72: 85–117.<br />

Schreibman, S. (2002). Computer-mediated Texts and Textuality: Theory and Practice.<br />

Computers and the Humanities 36: 283–93.<br />

Smith, M. N. (1999). Because the Plunge from the Front Overturned Us: The Dickinson<br />

Electronic Archives Project. Studies in the Literary Imagination 32: 133–51.<br />

Taylor, A. (1999). The Organization of Information. Englewood, CO: Libraries<br />

Unlimited.<br />

Unsworth, J. (1996). Electronic Scholarship: or, Scholarly Publishing and the Public.<br />

Journal of Scholarly Publishing 28: 3–12.<br />

Unsworth, J. (2000a). Scholarly Primitives: What Methods Do Humanities Researchers<br />

Have in Common, and How Might Our Tools Reflect This? Paper presented at<br />

symposium, Humanities Computing: Formal Methods, Experimental Practice, May 13,<br />

King's College London. Accessed November 26, 2002. At<br />

http://www.iath.virginia.edu/~jmu2m/Kings.5-00/primitives.html.<br />

Unsworth, J. (2000b). Thematic Research Collections. Paper presented at Modern<br />

Language Association Annual Conference, December 28, Washing<strong>to</strong>n, DC. Accessed<br />

November 26, 2002. At http://www.iath.virginia.edu/~jmu2m/MLA.00/.<br />

Urbina, E., R. Furuta, A. Goenka, R. Kochumman, E. Melgoza, and C. Monroy (2002).<br />

Critical Editing in the Digital Age: Informatics and Humanities Research. In J. Frow<br />

(ed.), The New Information Order and the Future of the Archive. Conference<br />

proceedings, Institute for Advanced Studies in the Humanities, March 20–23, University<br />

of Edinburgh. Accessed November 26, 2002. At<br />

http://webdb.ucs.ed.ac.uk/malts/other/IASH/dsp-all-papers.cfm.


Viscomi, J. (2002). Digital Facsimiles: Reading the William Blake Archive. Computers<br />

and the Humanities 36: 27–48.<br />

Waters, D. J. (1998). What Are Digital Libraries? CLIR Issues 4 (July/August). Accessed<br />

December 23, 2002. At http://www.clir.org/pubs/issues/issues04.html.<br />

25.<br />

Print Scholarship and Digital Resources<br />

Claire Warwick<br />

Whosoever loves not picture, is injurious <strong>to</strong> truth: and all the wisdom of poetry. Picture is<br />

the invention of heaven: the most ancient, and most akin <strong>to</strong> nature. It is itself a silent<br />

work: and always of one and the same habit: yet it doth so enter, and penetrate the inmost<br />

affection (being done by an excellent artificer) as sometimes it o'ercomes the power of<br />

speech and ora<strong>to</strong>ry.<br />

Ben Jonson, Explorata or Discoveries, 11. 1882–90<br />

Introduction<br />

In the late 1990s there was a great deal of concern about the death of the book. From<br />

every corner it was possible <strong>to</strong> hear quoted Vic<strong>to</strong>r Hugo's Archbishop, complaining that<br />

"ceci, tuera cela" (Nunberg 1996). Articles and books were published on the future of the<br />

book, which was assumed <strong>to</strong> be going <strong>to</strong> be a brief one (Finneran 1996). At the same time<br />

we were being <strong>to</strong>ld that we might no longer need <strong>to</strong> commute <strong>to</strong> work, or attend or teach<br />

at real physical universities, and of course if there were no longer any books, we would<br />

only need virtual libraries from which <strong>to</strong> access our electronic documents. Just a few<br />

years later all this seems <strong>to</strong> be as misguidedly futuristic as those 1970s newspaper articles<br />

predicting that by the year 2000 we would all eat protein pills instead of food. It is clear,<br />

then, that far from being killed off, print scholarship is still very much alive and well, and<br />

that its relationship <strong>to</strong> electronic resources is a highly complex one. In this chapter I will<br />

examine this relationship, and argue that we cannot hope <strong>to</strong> understand the complexity of<br />

such a relationship without looking at scholarly practices, and the way that such resources<br />

are used. This in turn means examining how we use information, either computationally<br />

or in print, in the wider context of scholarly life. We must consider how we read texts,<br />

and indeed visual objects, <strong>to</strong> understand why print remains important, and how it may<br />

relate <strong>to</strong> the <strong>digital</strong> objects and surrogates that are its <strong>companion</strong>s.<br />

Scenes from Academic Life<br />

To exemplify some of the changes that are taking place I would like <strong>to</strong> recreate two<br />

scenes from my academic life, which are illustrative of some of these larger issues.


Thessaloniki, Greece<br />

I am in the Byzantine museum, being shown some of the treasures of this northern Greek<br />

city, which prides itself on having been longest under continuous Byzantine rule. It is full<br />

of icons, mosaics, marble carvings, and richly painted <strong>to</strong>mbs. My guide asks what I think<br />

of them, and repeatedly I marvel at their beauty. But this, it appears, is not the point. We<br />

must, I am <strong>to</strong>ld, be able <strong>to</strong> read these images and icons in the way that their crea<strong>to</strong>rs<br />

intended us <strong>to</strong>. I see a <strong>to</strong>mb with a painted tree sheltering a rotund bird which looks rather<br />

like a turkey. In fact this is an olive, signifying eternity and peace, and the bird is a<br />

peacock, symbol of paradise. This is not simply wallpaper for the dead, but a statement of<br />

belief. The icons themselves glow with colors, whose richness I wonder at, but again they<br />

must be read. The gold background symbolizes heaven, and eternity, the red of a cloak is<br />

for love, the green band around the Madonna's head is for hope. The wrinkles painted on<br />

her face are <strong>to</strong> symbolize that beauty comes from within, and is not altered by age. My<br />

guide begins <strong>to</strong> test me, what do I see in front of me? I struggle <strong>to</strong> remember the<br />

unfamiliar visual alphabet that I am being taught, and realize that reliance on the printed<br />

word, and a lack of such meaningful images, has deprived me of a visual vocabulary; that<br />

such iconography is part of another tradition of communication, whose roots, at least in<br />

this part of Greece, are extremely ancient. They have long co-existed with the culture of<br />

the printed word, but have not entirely been supplanted by it.<br />

Portsmouth, England<br />

I am on an advisory board for the Portsmouth record office. Over a period of several<br />

decades they have painstakingly produced nine edited volumes of printed work<br />

cataloguing some of the holdings of their archives. They are arranged thematically,<br />

concerning dockyards, houses in the old <strong>to</strong>wn, legal documents. All are handsome<br />

hardbacks with an individual design. There is, it seems, still a substantial backlog of<br />

volumes in preparation, but printing has become so costly that they have decided <strong>to</strong><br />

publish electronically, which is why I am here. We spend time discussing volumes<br />

awaiting release, and all are relieved <strong>to</strong> find that the next volume should soon be ready<br />

after a period of thirty years in preparation.<br />

There is a certain culture shock on both sides. I am amazed <strong>to</strong> discover how long the<br />

process of editing and publication takes. Thirty years is long, but an average of a decade<br />

seems <strong>to</strong> be quite usual. I reflect that the papers themselves are his<strong>to</strong>ric, and still exist in<br />

the archive, waiting <strong>to</strong> be discovered, even if not calendared. But from the world I am<br />

used <strong>to</strong>, where technology changes so quickly, it is hard <strong>to</strong> return <strong>to</strong> a sense of such<br />

relative lack of change and urgency.<br />

A fellow panel member suggests that what had been thought of as several discrete print<br />

volumes, on large-scale maps, title deeds, city plans, and population data, could easily be<br />

combined, in an electronic environment, with the help of geographic information systems<br />

(GIS) technology. I in turn argue that the idea of separate volumes need not be retained in<br />

an electronic publication. There is no need <strong>to</strong> draw up a publication schedule as has been<br />

done with the print volumes. We could publish data on different themes concurrently, in


several releases, when it is ready, so that the <strong>digital</strong> records will grow at the pace of those<br />

who are editing, and not have <strong>to</strong> suffer delays.<br />

These suggestions are welcomed enthusiastically but the series edi<strong>to</strong>r reminds us that<br />

there are human dimensions <strong>to</strong> this process. Those editing individual volumes may want<br />

<strong>to</strong> see their work identified as a discrete entity, <strong>to</strong> gain credit from funding authorities or<br />

promotion boards. These collections would also, it seems, be incomplete without an<br />

introduction, and that, paradoxically, is usually written last: a major intellectual task and<br />

perhaps the fac<strong>to</strong>r, I speculate privately, which may have delayed publication, since it<br />

involves the synthesis of a vast range of sources and an attempt <strong>to</strong> indicate their<br />

intellectual value <strong>to</strong> a his<strong>to</strong>rian. Would it be possible <strong>to</strong> publish a release of the data<br />

without such an introduction, even if temporarily? We explore different possibilities for<br />

ways that data might appear with different views, <strong>to</strong> accommodate such problems, and<br />

the intellectual adjustments on both sides are almost visible. The his<strong>to</strong>rians and archivists<br />

contemplate the electronic unbinding of the volume as a controlling entity: those<br />

interested in computing are reminded that more traditional elements of the intellectual<br />

culture we are working in cannot be ignored if the project is <strong>to</strong> keep the good will of the<br />

scholars who work on it.<br />

Tools <strong>to</strong> Think With<br />

The two examples above serve as illustration of some of the very complex issues that we<br />

must contend with when considering the relationship between printed and electronic<br />

resources. As Jerome McGann argues, we have grown used <strong>to</strong> books as the primary <strong>to</strong>ols<br />

<strong>to</strong> think with in the <strong>humanities</strong>. We may read one text, but are likely <strong>to</strong> use a variety of<br />

other books as <strong>to</strong>ols, as we attempt <strong>to</strong> interpret texts of all kinds (McGann <strong>2001</strong>, ch. 2). In<br />

the remainder of this chapter I shall demonstrate that what links the two examples that I<br />

have quoted above with McGann's concerns is the question of how we use materials in<br />

<strong>humanities</strong> scholarship. I shall argue that whatever format materials are in, computational<br />

methods must make us reconsider how we read. Since we are so used <strong>to</strong> the idea of<br />

reading as an interpretative strategy we risk taking it for granted, and considering it<br />

mundane when compared <strong>to</strong> exciting new computational methods. But, as the example of<br />

the Macedonian museum shows, reading is a much more broad-ranging process than the<br />

comprehension of a printed text, and this comprehension itself is a complex process<br />

which requires more detailed analysis. It is also arguable that the visual elements of the<br />

graphical user interface (GUI) have also made us rediscover the visual aspects of reading<br />

and comprehension. Both of these processes must be considered in order <strong>to</strong> try <strong>to</strong><br />

understand the complex linkage between <strong>digital</strong> resources and print scholarship, because<br />

if we assume that reading printed text is a simple process, easily replaced by<br />

computational methods of interpreting <strong>digital</strong> resources, we risk underestimating the<br />

richness and complexity of more traditional research in the <strong>humanities</strong>.<br />

When the use of <strong>digital</strong> resources was first becoming widespread, assumptions were<br />

made that such resources could and indeed should replace the culture of interpreting<br />

printed resources by reading them. Enthusiasts championed the use of <strong>digital</strong> resources,<br />

and decried those who did not use them as ill-informed or neo-Luddite. During the 1990s


efforts were made <strong>to</strong> educate academics in the use of <strong>digital</strong> resources. Universities set up<br />

learning media units <strong>to</strong> help with the production of resources, and offered some technical<br />

support <strong>to</strong> academics, though at least in the British system this remains inadequate. The<br />

quality and quantity of <strong>digital</strong> resources available in the <strong>humanities</strong> also increased. And<br />

yet print scholarship is far from dead. Academics in the <strong>humanities</strong> still insisted on<br />

reading books, and writing articles, even if they also used or created <strong>digital</strong> resources. As<br />

I discovered in Portsmouth, cultural fac<strong>to</strong>rs within academia are slow <strong>to</strong> change. The<br />

authority of print publication is still undoubted. Why, after all, is this collection published<br />

in book form? Books are not only convenient, but carry weight with promotion<br />

committees, funding councils, and one's peers. Computational techniques, however,<br />

continue <strong>to</strong> improve and academic culture changes, even if slowly. What is not likely <strong>to</strong><br />

change in the complex dynamics between the two media is the fundamentals of how<br />

<strong>humanities</strong> academics work, and the way that they understand their material. How, then,<br />

should we explain the survival of reading printed texts?<br />

We might begin this process by examining early attempts <strong>to</strong> effect such a culture change.<br />

The 1993 Computers and the Humanities (CHUM) was very much in proselytizing mode.<br />

In the keynote article of a special issue on computers and literary criticism, Olsen argued<br />

that scholars were being wrong-headed. If only they realized what computers really are<br />

useful for, he suggested, there would be nothing <strong>to</strong> s<strong>to</strong>p them using computer<br />

methodology <strong>to</strong> produce important and far-reaching literary research. This is followed by<br />

an interesting collection of articles, making stimulating methodological suggestions. All<br />

of them proceeded from the assumption that critics ought <strong>to</strong> use, and what is more,<br />

should want <strong>to</strong> use <strong>digital</strong> resources in their research. Suggestions include the use of<br />

corpora for studying intertextuality (CHUM 1993, Greco) or cultural and social<br />

phenomena (CHUM 1993, Olsen); scientific or quantitative methodologies (CHUM 1993,<br />

Goldfield) such as those from cognitive science (CHUM 1993, Henry, and Spolsky), and<br />

Artificial Intelligence theory (CHUM 1993, Matsuba). All of these might have proved<br />

fruitful, but no work <strong>to</strong> be found in subsequent mainstream journals suggests that any<br />

literary critics <strong>to</strong>ok note of them.<br />

The reason appears <strong>to</strong> be that the authors of these papers assumed that a lack of<br />

knowledge on the part of their more traditional colleagues must be causing their apparent<br />

conservatism. It appears that they did not countenance the idea that none of these<br />

suggested methods might be fit for what a critic might want <strong>to</strong> do. As Fortier argues in<br />

one of the other articles in the volume, the true core activity of literary study is the study<br />

of the text itself, not theory, nor anything else. He suggests that: "this is not some<br />

reactionary perversity on the part of an entire profession, but a recognition that once<br />

literature studies cease <strong>to</strong> focus on literature, they become something else: sociology,<br />

anthropology, his<strong>to</strong>ry of culture, philosophy speculation, or what have you" (CHUM<br />

1993, Fortier 1993: 376). In other words these writers are offering useful suggestions<br />

about what their literary colleagues might do, instead of listening <strong>to</strong> critics like Fortier<br />

who are quite clear about what it is they want <strong>to</strong> do. Users have been introduced <strong>to</strong> all<br />

sorts of interesting things that can be done with computer analysis or electronic<br />

resources, but very few of them have been asked what it is that they do, and want <strong>to</strong> keep<br />

doing, which is <strong>to</strong> study texts by reading them.


As a result, there is a danger that <strong>humanities</strong> computing enthusiasts may be seen by their<br />

more traditional colleagues as wild-eyed technocrats who play with computers and <strong>digital</strong><br />

resources because they can. We may be seen as playing with technological <strong>to</strong>ys, while<br />

our colleagues perform difficult interpretative tasks by reading texts without the aid of<br />

technology. So if reading is still so highly valued and widely practiced, perhaps in order<br />

<strong>to</strong> be taken seriously as scholars, we as <strong>humanities</strong> computing practitioners should take<br />

the activity of reading seriously as well. As the example of my Portsmouth meeting<br />

shows, both sides of the debate will tend <strong>to</strong> make certain assumptions about scholarly<br />

practice, and it is only when we all understand and value such assumptions that we can<br />

make progress <strong>to</strong>gether. If reading a text is an activity that is not easily abandoned, even<br />

after academics know about and actively use <strong>digital</strong> resources, then it is important for us<br />

<strong>to</strong> ask what then reading might be, and what kind of materials are being read. I shall<br />

consider the second part of the question first, because before we can understand<br />

analytical methods we need <strong>to</strong> look at the material under analysis.<br />

Texts and Tradition<br />

Norman (1999) argues that we must be aware what computer systems are good for and<br />

where they fit in with human expertise. He thinks that computers are of most use when<br />

they complement what humans can do best. He attributes the failure or lack of use of<br />

some apparently good computer systems <strong>to</strong> the problem that they replicate what humans<br />

can do, but do it less efficiently. A clearer understanding of what <strong>humanities</strong> scholars do<br />

in their scholarship is therefore important. Computer analysis is particularly good at<br />

returning quantitative data and has most readily been adopted by scholars in fields where<br />

this kind of analysis is privileged, such as social and economic his<strong>to</strong>ry, or linguistics. A<br />

population dataset or linguistic corpus contains data that is ideal for quantitative analysis,<br />

and other researchers can also use the same data and analysis technique <strong>to</strong> test the<br />

veracity of the results.<br />

However, despite the pioneering work of Corns (1990) and Burrows (1987), literary text<br />

can be particularly badly suited <strong>to</strong> this type of quantitative analysis because of the kind of<br />

questions asked of the data. As Iser (1989) argues, the literary text does not describe<br />

objective reality. Literary data in particular are so complex that they are not well suited <strong>to</strong><br />

quantitative study, polar opposites of being or not being, right and wrong, presence or<br />

absence, but rather the analysis of subtle shades of meaning, of what some people<br />

perceive and others do not. The complexity is demonstrated in the use of figurative<br />

language. As Van Peer (1989: 303) argues this is an intrinsic feature of its "literariness."<br />

Thus we cannot realistically try <strong>to</strong> reduce complex texts <strong>to</strong> any sort of objective and nonambiguous<br />

state for fear of destroying what makes them worth reading and studying.<br />

Computer analysis cannot "recognize" figurative use of language. However, an electronic<br />

text might be marked up in such a way that figurative uses of words are distinguished<br />

from literal uses before any electronic analysis is embarked upon. However, there are two<br />

fundamental problems with this approach. Firstly, it is unlikely that all readers would<br />

agree on what is and is not figurative, nor on absolute numbers of usages. As I. A.<br />

Richards' study (1929) was the first <strong>to</strong> show, readers may produce entirely different


eadings of the same literary text. Secondly, the activity of performing this kind of<br />

markup would be so labor-intensive that a critic might just as well read the text in the<br />

first place. Nor could it be said <strong>to</strong> be any more accurate than manual analysis, because of<br />

the uncertain nature of the data. In many ways we are still in the position that Corns<br />

complained of in 1991 when he remarked that: "Such programmes can produce lists and<br />

tables like a medium producing ec<strong>to</strong>plasm, and what those lists and tables mean is often<br />

as mysterious" (1991: 128). Friendlier user interfaces mean that results are easier <strong>to</strong><br />

interpret, but the point he made is still valid. The program can produce data, but humans<br />

are still vital in its interpretation (CHUM 1993, Lessard and Benard).<br />

Furthermore, interpreting the results of the analysis is a particularly complex activity.<br />

One of the most fundamental ideas in the design of au<strong>to</strong>matic information retrieval<br />

systems is that the researcher must know what he or she is looking for in advance. This<br />

means that they can design the system <strong>to</strong> find this feature and that they know when they<br />

have found it, and how efficient recall is. However, unlike social scientists or linguists,<br />

<strong>humanities</strong> researchers often do not know what they are looking for before they approach<br />

a text, nor may they be immediately certain why it is significant when they find it. If<br />

computer systems are best used <strong>to</strong> find certain features, then this is problematic. They can<br />

only acquire this knowledge by reading that text, and probably many others. Otherwise,<br />

they are likely <strong>to</strong> find it difficult <strong>to</strong> interpret the results of computer analysis, or indeed <strong>to</strong><br />

know what sort of "questions" <strong>to</strong> ask of the text in the first place.<br />

Humanities scholars often do not need <strong>to</strong> analyze huge amounts of text <strong>to</strong> find material of<br />

interest <strong>to</strong> them. They may not need <strong>to</strong> prove a hypothesis as conclusively as possible, or<br />

build up statistical models of the occurrence of features <strong>to</strong> find them of interest, and may<br />

find the exceptional use of language or style as significant as general patterns (S<strong>to</strong>ne<br />

1982). They may therefore see traditional detailed reading of a relatively small amount of<br />

printed text as a more effective method of analysis.<br />

The availability of a large variety of text types may be more important than the amount of<br />

material. His<strong>to</strong>rians require a wide variety of materials such as letters, manuscripts,<br />

archival records, and secondary his<strong>to</strong>rical sources like books and articles (Duff and<br />

Johnson 2002). All of these are <strong>to</strong> be found in print or manuscript sources, some of which<br />

are rare and delicate and often found only in specific archives or libraries. Of course<br />

some materials like these have been digitized, but only a very small proportion. Even<br />

with the largest and most enthusiastic programs of digitization, given constraints of time,<br />

the limited budgets of libraries, and even research councils, it seems likely that this will<br />

remain the case for the foreseeable future. The his<strong>to</strong>rical researcher may also be looking<br />

for an unusual document whose value may not be apparent <strong>to</strong> others: the kind of<br />

document which may be ignored in selective digitization strategies.<br />

Indeed the experience of projects like Portsmouth records society suggests that this has<br />

already been recognized. If the actual material is so unique that a his<strong>to</strong>rian may need <strong>to</strong><br />

see the actual artifact rather than a digitized surrogate, then we may be better advised <strong>to</strong><br />

digitize catalogues, calendars, and finding aids, and allow users <strong>to</strong> make use of them <strong>to</strong><br />

access the material itself. This also returns us <strong>to</strong> the question of the visual aspect of


<strong>humanities</strong> scholarship. Initially the phenomenon of a digitized surrogate increasing the<br />

demand for the original artifact seemed paradoxical <strong>to</strong> librarians and archivists. However,<br />

this acknowledges the need for us <strong>to</strong> appreciate the visual aspects of the interpretation of<br />

<strong>humanities</strong> material. A transcribed manuscript which has been digitized allows us <strong>to</strong><br />

access the information contained in it. It may only be as a result of having seen a <strong>digital</strong><br />

image that the scholar realizes what further potential for interpretation exists, and this, it<br />

appears, may only be satisfied by the artifact itself. This is such a complex process that<br />

we do not yet fully understand its significance.<br />

The nature of the resources that <strong>humanities</strong> scholars use should begin <strong>to</strong> explain why<br />

there will continue <strong>to</strong> be a complex interaction between print and <strong>digital</strong> resources. There<br />

is not always a good fit between the needs of a <strong>humanities</strong> scholar or the tasks that they<br />

might want <strong>to</strong> carry out, and the <strong>digital</strong> resources available. Print still fulfills many<br />

functions and this perhaps encourages scholars <strong>to</strong> produce more, by publishing their<br />

research in printed form. But surely we might argue that computational methods would<br />

allow more powerful and subtle ways of analyzing such material. Reading, after all,<br />

seems such a simple task.<br />

What is Reading?<br />

Whatever assumptions we might be tempted <strong>to</strong> make, the activity of reading even the<br />

simplest text is a highly complex cognitive task, involving what Crowder and Wagner<br />

(1992: 4) describe as "three stupendous achievements": the development of spoken<br />

language, written language, and literacy. Kneepkins and Zwaan (1994: 126) show that:<br />

In processing text, readers perform several basic operations. For example, they decode<br />

letters, assign meaning <strong>to</strong> words, parse the syntactic structure of the sentence, relate<br />

different words and sentences, construct a theme for the text and may infer the objectives<br />

of the author. Readers attempt <strong>to</strong> construct a coherent mental representation of the text. In<br />

this process they use their linguistic knowledge (knowledge of words, grammar), and<br />

their world knowledge (knowledge of what is possible in reality, cultural knowledge,<br />

knowledge of the theme).<br />

These processes are necessary for even the most basic of texts, and therefore the<br />

cognitive effort necessary <strong>to</strong> process a complex text, such as the material commonly used<br />

by <strong>humanities</strong> researchers, must be correspondingly greater. Arguably, the most complex<br />

of all such texts are literary ones, and thus the following section is largely concerned with<br />

such material. Although empirical studies of the reading of literary text are in their<br />

comparative infancy (de Beaugrande 1992) research has ensured that the reading process<br />

is thought of as not simply a matter of recall of knowledge, but as a "complex cognitive<br />

and affective transaction involving text-based, reader-based, and situational fac<strong>to</strong>rs"<br />

(Goetz et al. 1993: 35).


Text-based fac<strong>to</strong>rs<br />

Most <strong>humanities</strong> scholars would agree that their primary task is <strong>to</strong> determine how<br />

meaning can be attributed <strong>to</strong> texts (Dixon et al. 1993). Yet the connection between<br />

meaning and language is an extremely complex problem. As Snelgrove (1990)<br />

concluded, when we read a literary text we understand it not only in terms of the meaning<br />

of specific linguistic features, but also by the creation of large-scale patterns and<br />

structures based on the interrelation of words and ideas <strong>to</strong> one another. This pattern<br />

making means that the relative meaning invested in a word may depend on its position in<br />

a text and the reaction that it may already have evoked in the reader. Dramatic irony, for<br />

example, is effective because we know that a character's speech is loaded with a<br />

significance they do not recognize. Our recognition of this will depend on the mental<br />

patterns and echoes it evokes. Language may become loaded and suffused with meaning<br />

specifically by relations, or its use in certain contexts.<br />

The complexity of the patterns generated by human readers is, however, difficult <strong>to</strong><br />

replicate when computational analysis is used. Text analysis software tends <strong>to</strong> remove the<br />

particular phenomenon under investigation from its immediate context except for the few<br />

words immediately surrounding it. A linguist may collect instances of a particular<br />

phenomenon and present the results of a concordance sorted alphabetically, irrespective<br />

of the order in which the words originally occurred in the text, or the author of them.<br />

However, for a literary critic the patterns created are vital <strong>to</strong> the experience of reading the<br />

text, and <strong>to</strong> the way it becomes meaningful. Thus the fragmented presentation of a<br />

computer analysis program cannot begin <strong>to</strong> approach the kind of understanding of<br />

meaning that we gain by reading a word as part of a narrative structure.<br />

The way we read literary text also depends on the genre of the work. Comprehension<br />

depends on what we already assume about the type of text we recognize it <strong>to</strong> be. Fish<br />

(1980: 326) found that he could persuade students that a list of names was a poem<br />

because of their assumptions about the form that poems usually take. Hanauer (1998) has<br />

found that genre affects the way readers comprehend and recall text, since they are likely<br />

<strong>to</strong> read more slowly and remember different types of information in a poem. Readers also<br />

decode terms and anaphora differently depending on what we know <strong>to</strong> be the genre of a<br />

text; for example, we will expect "cabinet" <strong>to</strong> mean one thing in a newspaper report of a<br />

political speech, and another in a carpentry magazine (Zwaan 1993: 2).<br />

Reader-based fac<strong>to</strong>rs<br />

The way that we extract meaning from a text also depends on many things which are<br />

extrinsic <strong>to</strong> it. The meaning of language changes depending on the associations an<br />

individual reader may make with other features of the text (Miall 1992), and with other<br />

texts and ideas. The creation of such webs of association and causation is central <strong>to</strong> the<br />

his<strong>to</strong>rian's craft. As the eminent his<strong>to</strong>rian G. R. El<strong>to</strong>n put it:


G. M. Young once offered celebrated advice: read in a period until you hear its people<br />

speak.… The truth is that one must read them, study their creations and think about them<br />

until one knows what they are going <strong>to</strong> say next.<br />

(El<strong>to</strong>n 1967: 30)<br />

Interaction between the reader and the text is also affected by fac<strong>to</strong>rs which are particular<br />

<strong>to</strong> the individual reader (Halasz 1992). Iser (1989) argues that narrative texts invite the<br />

interaction of the reader by indeterminacy in the narrative. Where there are gaps of<br />

uncertainty, the reader fills them using their own experience. Readers' experience of a<br />

fictional text will even be affected by the relationship which they build between<br />

themselves and the narra<strong>to</strong>r (Dixon and Ber<strong>to</strong>lussi 1996a).<br />

Situational fac<strong>to</strong>rs<br />

The reader's response <strong>to</strong> a text is likely <strong>to</strong> be affected by situational fac<strong>to</strong>rs, for example<br />

their gender, race, education, social class, and so on. This is also liable <strong>to</strong> change over<br />

time, so that we may experience a text differently at the age of 56 than at 16. As Potter<br />

(1991) argues, these fac<strong>to</strong>rs have yet <strong>to</strong> be taken in<strong>to</strong> account in empirical readership<br />

studies of literary text. Even more subtly than this, however, a reader's appreciation of a<br />

text may be affected by how they feel on a particular day, or if a feature of the text is<br />

reminiscent of their personal experience. Happy readers notice and recall parts of a text<br />

which describe happiness, or evoke the emotion in them, and sad ones the opposite<br />

(Kneepkins and Zwaan 1994: 128). The role of emotional engagement is clearly vital in<br />

literary reading. Yet it is one which is very difficult <strong>to</strong> quantify, or describe, and therefore<br />

is almost impossible for computer analysis <strong>to</strong> simulate.<br />

Reading a text is also affected by the frequency of reading and the expertise of the reader,<br />

as El<strong>to</strong>n's observation suggests (Dixon and Ber<strong>to</strong>lussi 1996b; Dorfman 1996). Dixon and<br />

colleagues (1993) found that the same textual feature might be the cause of varied effects<br />

in different readers and certain effects were only apparent <strong>to</strong> some readers. Although core<br />

effects of the text could usually be discerned on first reading, other, more subtle effects<br />

were only reported on second or subsequent readings, or would only be apparent <strong>to</strong> some<br />

of the readers. They also found that the subtlest of literary effects tended <strong>to</strong> be noticed by<br />

readers who they called "experienced." They concluded that reading is such a complex<br />

procedure that all the effects of the text are unlikely <strong>to</strong> be apparent at once, and that<br />

reading is clearly a skill that needs <strong>to</strong> be learnt and practiced.<br />

Reading the Visual<br />

It should therefore be becoming clear why print resources have continued <strong>to</strong> co-exist with<br />

<strong>digital</strong> ones. The key activity of the <strong>humanities</strong> scholar is <strong>to</strong> read and interpret texts, and<br />

there is little point in using a computational <strong>to</strong>ol <strong>to</strong> replicate what human agency does<br />

best in a much less complex and subtle manner. Reading a printed text is clearly a subtle<br />

and complex analysis technique. It is therefore not surprising that scholars have made the<br />

assumption that <strong>digital</strong> resources and computational techniques that simply replicate the


activity of reading are a pale imitation of an already successful technique. To be of use <strong>to</strong><br />

the <strong>humanities</strong> scholar, it seems that <strong>digital</strong> resources must therefore provide a different<br />

dimension that may change the way that we view our raw materials.<br />

In some ways we can discern a similar movement in <strong>humanities</strong> computing <strong>to</strong> that which<br />

has taken place in computer science. In the 1980s and 1990s Artificial Intelligence<br />

seemed <strong>to</strong> offer the prospect of thinking machines (Jonscher 1999, ch. 5). But the<br />

technology that has captured the imagination of users has not been a computer system<br />

that seeks <strong>to</strong> think for them, but one that provides access <strong>to</strong> material that can provide raw<br />

material for human thought processes, that is, the Internet and World Wide Web. The<br />

popularity of the Web appears <strong>to</strong> have dated from the development of graphical browsers<br />

that gave us access not only <strong>to</strong> textual information, but <strong>to</strong> images. The effect of this and<br />

the rise of the graphical user interface has been <strong>to</strong> re-acquaint us with the power of<br />

images, not only as ways of organizing information, but as way of communicating it. Just<br />

as the images in the museum in Thessaloniki reminded me that there are other ways <strong>to</strong><br />

interpret and communicate ideas, so we have had <strong>to</strong> relearn ways <strong>to</strong> read an image,<br />

whether the frame contains a painted or a pixelated icon.<br />

It is in this area that, I would argue, <strong>digital</strong> resources can make the greatest contribution<br />

<strong>to</strong> <strong>humanities</strong> scholarship. Digitization projects have revolutionized our access <strong>to</strong><br />

resources such as images of manuscripts (Unsworth 2002). The use of three-dimensional<br />

CAD modeling has been extensively used in archaeology <strong>to</strong> help reconstruct the way that<br />

buildings might have looked (Sheffield University 2003). However, the projects that are<br />

most innovative are those that use <strong>digital</strong> resources not for reconstruction or improved<br />

access, though these are of course enormously valuable, but as <strong>to</strong>ols <strong>to</strong> think with. If the<br />

process of reading and interpreting a text is so complex, then it may be that this is best<br />

left <strong>to</strong> our brains as processing devices for at least the medium term. It is, however, in the<br />

realm of the visual that we are seeing some of the most interesting interrelationships of<br />

print scholarship and <strong>digital</strong> resources. We need only look at a small sample of some of<br />

the papers presented at the Association for Literary and Linguistic Computing-<br />

Association for Computers and the Humanities conference (at )<br />

in 2002 <strong>to</strong> see exciting examples of such<br />

developments in progress.<br />

Steve Ramsay argues that we might "remap, reenvision, and re-form" a literary text<br />

(Ramsay 2002). He refers <strong>to</strong> McGann and Drucker's experiments with deforming the way<br />

a text is written on a page (McGann <strong>2001</strong>), but has moved beyond this <strong>to</strong> the use of<br />

Graph View Software. He has used this program, which was originally designed <strong>to</strong> create<br />

graphic representations of numerical data, <strong>to</strong> help in the analysis of dramatic texts. In a<br />

previous interview, he had demonstrated this <strong>to</strong> me, showing a three-dimensional<br />

graphical mapping of Shakespeare's An<strong>to</strong>ny and Cleopatra. This created wonderful<br />

abstract looping patterns which might have been at home in a gallery of modern art. But,<br />

like the Macedonian icons, theses were not simply objects of beauty. Once interpreted,<br />

they show the way that characters move though the play, being drawn inexorably <strong>to</strong>wards<br />

Rome. This visual representation had the effect, not of abolishing the human agency of<br />

the literary critic, but providing, literally, a new vision of the play, perhaps opening up


new vistas <strong>to</strong> the critical view. The significance of such movement, and what it reveals<br />

about the play, is for the critic herself <strong>to</strong> decide, but the program has performed a useful<br />

form of defamiliarization, which would be difficult <strong>to</strong> imagine in a print environment.<br />

We can also see similar types of visual representation of textual information in the<br />

interactive 3-D model of Dante's Inferno. The effect of this is a similar kind of<br />

defamiliarization. A very new view of the information in the text is created, but the effect<br />

of it, at least on this reader, is <strong>to</strong> make her wish <strong>to</strong> return <strong>to</strong> the text itself in printed form,<br />

and <strong>to</strong> read it with new eyes. The <strong>digital</strong> resource has therefore not made reading<br />

redundant, but helped <strong>to</strong> suggest new avenues of interpretation. This project is being<br />

developed at the same research centre, IATH, where McGann is and Ramsay was based.<br />

This is another intriguing connection between the world of <strong>digital</strong> resources and more<br />

traditional forms of scholarship. Computational <strong>to</strong>ols as simple as e-mail have made the<br />

process of scholarly collaboration over large physical distances much easier than before.<br />

Yet it is fascinating <strong>to</strong> note that the physical proximity of scholars such as those at IATH<br />

facilitates the interchange of ideas and makes it possible for methodologies <strong>to</strong> be shared<br />

and for projects and scholars <strong>to</strong> be a creative influence on each other – a process that we<br />

can see at work in the visual dynamics of these IATH projects. This is not an isolated<br />

phenomenon. The fact that Microsoft's campus in Cambridge shares alternate floors of a<br />

building with the university department of computer science shows that such a<br />

technologically advanced organization still values informal creative exchanges as a way<br />

<strong>to</strong> inspire new projects and support existing ones. The Cambridge University Computer<br />

Labora<strong>to</strong>ry's online coffee machine was a star turn of the early Web (Stafford-Fraser<br />

1995). But it was finally switched off in <strong>2001</strong>, perhaps proof that Microsoft's decision <strong>to</strong><br />

privilege meetings over a non-virtual coffee shows their recognition that traditional<br />

methods are still vital in innovation and scholarship.<br />

Two other IATH projects were also discussed at the conference. The Salem witch trials<br />

and Bos<strong>to</strong>n's Back Bay Fens are two projects which both make use of GIS technology <strong>to</strong><br />

integrate textual material and numerical data with spatial information (Pitti et al. 2002).<br />

Once again these projects allow the user <strong>to</strong> visualize the data in different ways.<br />

Connections might be made between people or places in his<strong>to</strong>rical Bos<strong>to</strong>n, whose<br />

physical proximity is much easier for a user <strong>to</strong> establish in the visual environment of a<br />

GUI interface than by the examination of printed data. Where a particular user's<br />

knowledge might be partial, the use of such data lets her literally envision new ones, as a<br />

result of interrogating a large spatial dataset. A new textual narrative can emerge from the<br />

physical linkages.<br />

The case of the Salem witch trials is also intriguing, since the Flash animations actually<br />

allow the user <strong>to</strong> watch as the accusations spread over time and space like a medical,<br />

rather than psychological, epidemic. This idea of mapping this spread is, however, not<br />

new in terms of his<strong>to</strong>riography. In 1974 Boyer and Nissenbaum had pioneered this<br />

approach in a ground-breaking book, Salem Possessed. This contains printed maps of the<br />

area which give us snapshots of the progress of the allegations. We can, therefore, see an<br />

immediate relationship between scholarship in print and a <strong>digital</strong> resource which has<br />

grown out of such theories. What the <strong>digital</strong> resource adds, though, is the immediacy of


eing able <strong>to</strong> watch, run and rerun the sequence in a way that a book, while impressive,<br />

cannot allow us <strong>to</strong> do. Once again, the theories that lie behind the data are, as Pitti et al.<br />

(2002) make clear, very much the product of the scholarship that the his<strong>to</strong>rians who<br />

initiated the projects brought <strong>to</strong> them. Analysis performed on the data will also be done in<br />

the minds of other scholars. But the visual impact of both of these databases supports<br />

human information processing, and may suggest new directions for human analysis <strong>to</strong><br />

take.<br />

As the website for the Valley of the Shadow, one of the two original IATH projects, puts<br />

it, GIS may be used literally "<strong>to</strong> put [his<strong>to</strong>rical] people back in<strong>to</strong> their houses and<br />

businesses" (Ayers et al. <strong>2001</strong>). Valley of the Shadow is itself doing far more than simply<br />

this. The project itself seems <strong>to</strong> be predicated on an understanding of the visual. Even the<br />

basic navigation of the site is organized around the metaphor of a physical archive, where<br />

a user navigates the different materials by visiting, or clicking on, the separate rooms of a<br />

plan of the space. By linking a contemporary map with the data about a particular area,<br />

GIS allows users <strong>to</strong> interpret the data in a new way, giving a concrete meaning <strong>to</strong><br />

statistical data or names of people and places in the registers which are also reproduced<br />

(Thomas 2000). Just as with a literary text, this might cause a his<strong>to</strong>rian <strong>to</strong> return <strong>to</strong> the<br />

numerical or textual data for further analysis, and some of this might use computational<br />

<strong>to</strong>ols <strong>to</strong> aid the process.<br />

But its greatest value is in prompting a fresh approach for consideration. The his<strong>to</strong>rian's<br />

brain is still the <strong>to</strong>ol that determines the significance of the findings. This distinction is<br />

important, since it distinguishes Valley of the Shadow from some earlier computational<br />

projects in similar areas of American his<strong>to</strong>ry. For example, in 1974 Fogle and Engerman<br />

wrote Time on the Cross, in which they sought <strong>to</strong> explain American slavery with the use<br />

of vast amounts of quantitative data on plantation slavery, which was then analyzed<br />

computationally. This, they claimed, would produce a definitive record of the objective<br />

reality of slavery in America. Critics of the book have argued persuasively that the data<br />

were handled much <strong>to</strong>o uncritically. Their findings, it has been claimed, were biased,<br />

because they only used statistical written data, which usually emerged from large<br />

plantations, and ignored the situation of small slave holders who either did not or could<br />

not document details of their holdings, either because the holdings were small or because<br />

the slave holder was illiterate (Ransom and Sutch 1977; Wright 1978). Other his<strong>to</strong>rians<br />

have insisted that anecdotal sources and printed texts must be used <strong>to</strong> complement the<br />

findings, <strong>to</strong> do sufficient justice <strong>to</strong> the complexity of the area. It could be argued that the<br />

problems that they encountered were caused by an over-reliance on computational<br />

analysis of numerical data, and by the implication that this could somehow deliver a<br />

definitive explanation of slavery in a way that would finally put an end <strong>to</strong> controversies<br />

caused by subjective human analysis. A project such as Valley of the Shadow is a<br />

significant progression onwards, not only in computational techniques but also in<br />

scholarly method. It does not rely on one style of data, since it links numerical records <strong>to</strong><br />

textual and spatial data. These resources are then offered as <strong>to</strong>ols <strong>to</strong> aid interpretation,<br />

which takes place in the his<strong>to</strong>rian's brain, rather than in any way seeking <strong>to</strong> supersede<br />

this.


The products of the projects are also varied, ranging from students' projects, which take<br />

the materials and use them <strong>to</strong> create smaller <strong>digital</strong> projects for assessment, <strong>to</strong> more<br />

traditional articles and conference presentations. Most intriguing, perhaps, is the form of<br />

hybrid scholarship that Ed Ayers, the project's founder, and William Thomas have<br />

produced. Ayers had long felt that despite the innovative research that could be<br />

performed with a <strong>digital</strong> resource, there had been very little effect on the nature of<br />

resulting publication. The written article, even if produced in an electronic journal, was<br />

still essentially un<strong>to</strong>uched by the <strong>digital</strong> medium, having the same structure as an article<br />

in a traditional printed journal. Ayers and Thomas (2002) therefore wrote an article which<br />

takes advantage of the electronic medium, by incorporation some of the GIS data, and the<br />

hypertextual navigation system which gives a reader multiple points of entry and of<br />

linkage with other parts of the resource. Readers are given a choice of navigation, a<br />

visual interface with Flash animations or a more traditional text-based interface. The<br />

article might even be printed out, but it is difficult <strong>to</strong> see how it could be fully appreciated<br />

without the use of its <strong>digital</strong> interactive elements. This has appeared in American<br />

His<strong>to</strong>rical Review, a traditional academic journal, announcing its right <strong>to</strong> be considered<br />

part of the mainstream of his<strong>to</strong>rical research. As such it represents a dialogue between the<br />

more traditional world of the academic journal and the possibilities presented by <strong>digital</strong><br />

resources, at once maintaining the continuity of scholarly traditions in his<strong>to</strong>ry, but also<br />

seeking <strong>to</strong> push the boundaries of what is considered <strong>to</strong> be a scholarly publication. The<br />

analysis presented in the paper emerges from human reading and processing of data, but<br />

would not have been possible without the use of the <strong>digital</strong> resource.<br />

It is not only at IATH, however, that work in this area is taking place, despite their<br />

leading position in the field. Thomas Corns et al. (2002), from the University of Bangor<br />

in Wales, has described how one aspect of document visualization can aid human<br />

analysis. Being able <strong>to</strong> digitize a rare manuscript has significantly aided his team in<br />

trying <strong>to</strong> determine whether it was written by Mil<strong>to</strong>n. The simple task of being able <strong>to</strong><br />

cut, paste, and manipulate letter shapes in the digitized text has helped in their<br />

examination of scribal hands. The judgment is that of the scholars, but based on the<br />

ability <strong>to</strong> see a text in a new way, only afforded by digitized resources. This is not a new<br />

technique, and its success is largely dependent on the questions that are being asked of<br />

the data by the human investiga<strong>to</strong>rs. Donaldson (1997) discusses ways in which complex<br />

analysis of <strong>digital</strong> images of seventeenth-century type was used <strong>to</strong> try <strong>to</strong> decide whether<br />

Shakespeare had used the word wife or wise in a couplet from The Tempest, usually<br />

rendered as "So rare a wondered father and a wise / Makes this place paradise" (Act IV,<br />

Scene, I, 11. 122–3). The <strong>digital</strong> research proved inconclusive but might have been<br />

unnecessary, since a Shakespeare scholar might be expected <strong>to</strong> deduce that the rhyme of<br />

wise and paradise is much more likely in the context of the end of a character's speech,<br />

than the word wife, which while tempting for a feminist analysis would not follow the<br />

expected pattern of sound. All of which indicates that the use of <strong>digital</strong> resources can<br />

only be truly meaningful when combined with old-fashioned critical judgment.<br />

Another project being presented at ALLC-ACH, which is very much concerned with<br />

facilitating critical judgment through the realm of the visual, is the Versioning Machine<br />

(Smith 2002). This package, which supports the organization and analysis of text with


multiple variants, is once again a way of helping the user <strong>to</strong> envision a text in a different<br />

way, or even in multiple different ways. The ability <strong>to</strong> display multiple variants<br />

concurrently, <strong>to</strong> color-code comments that are read or unread, selectively <strong>to</strong> show or hide<br />

markup pertaining <strong>to</strong> certain witnesses, gives scholars a different way of perceiving the<br />

text, both in terms of sight and of facilitating the critical process. It is also far less<br />

restrictive than a printed book in the case where the text of a writer might have multiple<br />

variants, none of which the critic can say with certainly is the final version. The case of<br />

Emily Dickinson is a notable one, presented by the MITH team, but it may be that if<br />

freed by <strong>digital</strong> resources like the Versioning Machine of the necessity of having <strong>to</strong><br />

decide on a copy text for an edition, the text of many other writers might be seen as much<br />

more mutable, and less fixed in a final form of texuality.<br />

Print editions, for example of the seventeenth-century poet Richard Crashaw, have forced<br />

edi<strong>to</strong>rs <strong>to</strong> make difficult decisions about whether the author himself made revisions <strong>to</strong><br />

many of the poems. When, as in this case, evidence is contradic<strong>to</strong>ry or inconclusive, it is<br />

surely better <strong>to</strong> be able <strong>to</strong> use <strong>digital</strong> technology such as the Versioning Machine <strong>to</strong> give<br />

users a different way of seeing, and enable them <strong>to</strong> view the variants without edi<strong>to</strong>rial<br />

intervention. The use of the Versioning Machine will not s<strong>to</strong>p the arguments about which<br />

version might be preferred, based as they are on literary judgment and the interpretation<br />

of his<strong>to</strong>rical events, but at least we as readers are not presented with the spurious<br />

certainty that a print edition forces us in<strong>to</strong>. Once again, therefore, such use of computer<br />

technology is not intended <strong>to</strong> provide a substitute for critical analysis, and the vast<br />

processing power of the human brain, rather it gives us a way of reviewing the evidence<br />

of authorial revisions. It makes concrete and real again the metaphors that those terms<br />

have lost in the world of print scholarship.<br />

Conclusion<br />

Even when we look at a small sample of what is taking place in the field, it is clear that<br />

some of the most exciting new developments in the <strong>humanities</strong> computing area seem <strong>to</strong><br />

be looking <strong>to</strong>wards the visual as a way of helping us <strong>to</strong> reinterpret the textual. It appears<br />

that we are moving beyond not printed books and print-based scholarship, but the naive<br />

belief that they can easily be replaced by <strong>digital</strong> resources.<br />

As the example of my visit <strong>to</strong> Portsmouth demonstrated, it is simplistic <strong>to</strong> believe that we<br />

can, or should, rush <strong>to</strong> convince our more traditional colleagues of the inherent value of<br />

<strong>digital</strong> resources, without taking in<strong>to</strong> account the culture of long-established print<br />

scholarship. It is only through negotiations with scholars, and in forging links between<br />

the <strong>digital</strong> and the textual traditions that the most interesting research work is likely <strong>to</strong><br />

emerge.<br />

The materials that <strong>humanities</strong> scholars use in their work are complex, with shifting<br />

shades of meaning that are not easily interpreted. We are only beginning <strong>to</strong> understand<br />

the subtle and complicated processes of interpretation that these require. However, when<br />

we consider that the process of reading a text, which may seem so simple, is in fact so<br />

difficult an operation that computer analysis cannot hope <strong>to</strong> replicate it at present, we can


egin <strong>to</strong> understand why for many scholars the reading of such material in print will<br />

continue <strong>to</strong> form the core activity of their research.<br />

Digital resources can, however, make an important contribution <strong>to</strong> this activity. Far from<br />

attempting <strong>to</strong> replace the scholar's mind as the processing device, computer delivery of<br />

resources can help <strong>to</strong> support the process. The complexity of visual devices as a way of<br />

enshrining memory and communicating knowledge is something that the ancient world<br />

unders<strong>to</strong>od very well, as I learnt when I began <strong>to</strong> read the icons in Thessaloniki. While<br />

much of this knowledge has been lost in the textual obsession of print culture, the<br />

graphical interface of the computer screen has helped us reconnect <strong>to</strong> the world of the<br />

visual and recognize that we can relearn a long-neglected vocabulary of interpretation.<br />

Digital resources can provide us with a new way <strong>to</strong> see, and thus <strong>to</strong> perceive the<br />

complexities in the process of interpreting <strong>humanities</strong> materials. A new way of looking at<br />

a text can lead <strong>to</strong> a way of reading it that is unconstrained by the bindings of the printed<br />

medium, even if it leads us back <strong>to</strong> the pages of a printed book.<br />

Acknowledgments<br />

It is with gratitude that I would like <strong>to</strong> dedicate this chapter <strong>to</strong> the memory of Dr J.<br />

Wilbur Sanders (1936–2002). A man of acute perception and a great teacher, he helped<br />

me <strong>to</strong> see how subtle and complex a process reading might be.<br />

Ayers, E. L. (1999). The Pasts and Futures of Digital His<strong>to</strong>ry. At<br />

http://www.jefferson.village.virginia.edu/vcdh/PastsFutures.html.<br />

Ayers, E. L., A. S. Rubin, and W. G. Thomas (<strong>2001</strong>). The Valley's Electronic Cultural<br />

Atlas Initiative Geographic Information Systems Project. At<br />

http://jefferson.village.virginia/edu/vshadow2/ecai/present/html.<br />

Ayers, E. L. and W. G. Thomas (2002). Two American Communities on the Eve of the<br />

Civil War: An Experiment in Form and Analysis. At<br />

http://jefferson.village.Virginia.edu/vcdh/xml_docs/projects.html.<br />

de Beaugrande, R. (1992). Readers Responding <strong>to</strong> Literature: Coming <strong>to</strong> Grips with<br />

Reality. In Elaine F. Narduccio (ed.), Reader Response <strong>to</strong> Literature: The Empirical<br />

Dimension (pp. 193–210). New York and London: Mou<strong>to</strong>n de Gruyter.<br />

Boyer, P. and S. Nissenbaum (1974). Salem Possessed: The Social Origins of Witchcraft.<br />

Cambridge, MA: Harvard University Press.<br />

Burrows, J. F. (1987). Computation in<strong>to</strong> Criticism: A Study of Jane Austen's Novels and<br />

an Experiment in Method. Oxford: Clarendon Press.<br />

Computers and the Humanities (CHUM), 27 (1993). Special issue on literature and<br />

computing, includes: Thomas, J.-J., Texts On-line (93–104).Henry, C., The Surface of<br />

Language and Humanities Computing (315–22). Spolsky, E., Have It Your Way and


Mine: The Theory of Styles (323–430). Matsuba, S., Finding the Range: Linguistic<br />

Analysis and its Role in Computer-assisted Literary Study (331–40). Greco, L. G., and<br />

Shoemaker, P., Intertextuality and Large Corpora: A Medievalist Approach (349–455).<br />

Bruce, D., Towards the Implementation of Text and Discourse Theory in Computer<br />

Assisted Textual Analysis (357–464). Goldfield, J. D., An Argument for Single-author<br />

and Similar Studies Using Quantitative Methods: Is There Safety in Numbers? (365–<br />

474). Fortier, P. A., Babies, Bathwater and the Study of Literature (375–485). Lessard, G.<br />

and J. Benard, Computerising Celine (387–494). Olsen, M., Signs, Symbols and<br />

Discourses: a New Direction for Computer Aided Literature Studies (309–414). Olsen,<br />

M., Critical Theory and Textual Computing: Comments and Suggestions (395–4400).<br />

Corns, T. N. (1990). Mil<strong>to</strong>n's Language. Oxford: Blackwell.<br />

Corns, T. N. (1991). Computers in the Humanities: Methods and Applications in the<br />

Study of English Literature. Literary and Linguistic Computing 6, 2: 127–31.<br />

Corns, T. N. et al. (2002). Imaging and Amanuensis: Understanding the Manuscript of<br />

De Doctrina Christiana, attributed <strong>to</strong> John Mil<strong>to</strong>n. Proceedings of ALLC/ACH 2002:<br />

New Directions in Humanities Computing (pp. 28–30). The 14th International<br />

Conference. University of Tübingen, July 24–28, 2002.<br />

Crowder, R. G. and R. K. Wagner (1992). The Psychology of Reading: An Introduction,<br />

2nd edn. Oxford: Oxford University Press.<br />

Dixon, P. and M. Ber<strong>to</strong>lussi (1996a). Literary Communication: Effects of Reader-<br />

Narra<strong>to</strong>r Cooperation. Poetics 23: 405–30.<br />

Dixon, P. and M. Ber<strong>to</strong>lussi (1996b). The Effects of Formal Training on Literary<br />

Reception. Poetics 23: 471–87.<br />

Dixon, P. et al. (1993). Literary Processing and Interpretation: Towards Empirical<br />

Foundations. Poetics 22: 5–33.<br />

Donaldson, P. S. (1997). Digital Archive as Expanded Text: Shakespeare and Electronic<br />

Textuality. In K. Sutherland (ed.), Electronic Text: Investigations in Method and Theory<br />

(pp. 173–98). Oxford: Clarendon Press.<br />

Dorfman, M. H. (1996). Evaluating the Interpretive Community: Evidence from Expert<br />

and Novice Readers. Poetics 23: 453–70.<br />

Duff, W. M. and C. A. Johnson (2002). Accidentally Found on Purpose: Informationseeking<br />

Behavior of His<strong>to</strong>rians in Archives. Library Quarterly 72: 472–96.<br />

El<strong>to</strong>n, G. R. (1967). The Practice of His<strong>to</strong>ry. London: Fontana.


Finneran, R. J (1996). The Literary Text in the Digital Age. Ann Arbor: University of<br />

Michigan Press.<br />

Fish, S. (1980). Is There a Text in this Class? The Authority of Interpretive Communities.<br />

Cambridge, MA: Harvard University Press.<br />

Fogel, R. W. and S. L. Engerman (1974). Time on the Cross. London: Wildwood House.<br />

Goetz, et al. (1993). Imagery and Emotional Response. Poetics 22: 35–49.<br />

Halasz, L. (1992). Self-relevant Reading in Literary Understanding. In Elaine F.<br />

Narduccio (ed.), Reader Response <strong>to</strong> Literature: The Empirical Dimension (pp. 229–6).<br />

New York and London: Mou<strong>to</strong>n de Gruyter.<br />

Hanauer, D. (1998). The Genre-specific Hypothesis of Reading: Reading Poetry and<br />

Encyclopaedic Items. Poetics 26: 63–80.<br />

Iser, W. (1989). Prospecting: From Reader Response <strong>to</strong> Literary Anthropology.<br />

Baltimore and London: Johns Hopkins University Press.<br />

Jonscher, C. (1999). The Evolution of Wired Life: From the Alphabet <strong>to</strong> the Soul-Catcher<br />

Chip – How Information Technologies Change Our World. New York: Wiley.<br />

Kneepkins, E. W. E. M. and R. A. Zwaan (1994). Emotions and Literary Text<br />

Comprehension. Poetics 23: 125–38.<br />

McGann, J. (<strong>2001</strong>). Radiant Textuality: Literature after the World Wide Web. New York<br />

and Basings<strong>to</strong>ke: Palgrave.<br />

Miall, D. (1992). Response <strong>to</strong> Poetry: Studies of Language and Structure. In Elaine F.<br />

Narduccio (ed.), Reader Response <strong>to</strong> Literature: The Empirical Dimension (pp. 153–72).<br />

Berlin and New York: Mou<strong>to</strong>n de Gruyter.<br />

Norman, D. A. (1999). The Invisible Computer: Why Good Products Can Fail, the<br />

Personal Computer Is so Complex, and Information Appliances Are the Solution. Bos<strong>to</strong>n:<br />

MIT Press.<br />

Nunberg, G., (ed.) (1996). The Future of the Book (p. 10). Berkeley: University of<br />

California Press.<br />

Pitti, D., C. Jessee, and S. Ramsay (2002). Multiple Architectures and Multiple Media:<br />

the Salem Witch Trials and Bos<strong>to</strong>n's Back Bay Fens. Proceedings of ALLC/ACH 2002:<br />

New Directions in Humanities Computing (pp. 87–91). The 14th International<br />

Conference. University of Tubingen, July 24–28, 2002.


Potter, R. G. (1991). Pragmatic Research on Reader Responses <strong>to</strong> Literature with an<br />

Emphasis on Gender and Reader Responses. Revue Beige de la Philosophie et d'His<strong>to</strong>ire<br />

69, 3: 599–617.<br />

Ramsay, S. (2002). Towards an Algorithmic Criticism. Proceedings of ALLC/ACH 2002:<br />

New Directions in Humanities Computing (p. 99). The 14th International Conference.<br />

University of Tubingen, July 24–28, 2002.<br />

Ransom, R. L. and R. Sutch (1977). One Kind of Freedom: The Economic Consequences<br />

of Emancipation. Cambridge: Cambridge University Press.<br />

Richards, I. A. (1929). Practical Criticism: A Study of Literary Judgements. London:<br />

Kegan Paul.<br />

Sheffield University (2003). The Cistercians in Yorkshire. At<br />

http://cistercians.shef.ac.uk/.<br />

Smith, M. N. (2002). MITH's Lean, Mean Versioning Machine. Proceedings of<br />

ALLC/ACH 2002: New Directions in Humanities Computing (pp. 122–7). The 14th<br />

International Conference. University of Tübingen, July 24–28, 2002.<br />

Snelgrove, T. (1990). A Method for the Analysis of the Structure of Narrative Texts. LLC<br />

5: 221–5.<br />

Stafford-Fraser, Q. (1995). The Trojan Room Coffee Pot: A (Non-technical) Biography.<br />

Cambridge University Computer Labora<strong>to</strong>ry. At<br />

http://www.cl.cam.ac.uk/coffee/qsf/coffee.html.<br />

S<strong>to</strong>ne, S. (1982). Humanities Scholars – Information Needs and Uses. Journal of<br />

Documentation 38, 4: 292–313.<br />

Thomas, W. G. (2000). Preliminary Conclusions from the Valley of the Shadow Project.<br />

Paper delivered at: ECAI/Pacific Neighborhood Consortium Conference, January 11,<br />

2000, University of California at Berkeley. At<br />

http://jefferson.village.virginia.edu/vcdh/ECAI.paper.html.<br />

Unsworth, J. (2002). Using Digital Primary Resources <strong>to</strong> Produce Scholarship in Print.<br />

Paper presented at Literary Studies in Cyberspace: Texts, Contexts, and Criticism.<br />

Modern Language Association Annual Convention, Sunday, December 29, 2002. New<br />

York, USA. At http://jefferson.village.virginia.edu/~jmu2m/cyber-mla.2002/.<br />

Van Peer, W. (1989). Quantitative Studies of Literature, a Critique and an Outlook.<br />

Computers and the Humanities 23: 301–7.<br />

Wright, G. (1978). The Political Economy of the Cot<strong>to</strong>n South'. Households, Markets,<br />

and Wealth in the Nineteenth Century. New York: W. W. Nor<strong>to</strong>n.


Zwaan, R. A. (1993). Aspects of Literary Comprehension: A Cognitive Approach.<br />

Amsterdam: John Benjamins.<br />

26.<br />

Digital Media and the Analysis of Film<br />

Robert Kolker<br />

This his<strong>to</strong>ry of film studies is a short one, dating from the early <strong>to</strong> mid-1960s, and<br />

evolving from a number of his<strong>to</strong>rical events. One was the appearance of the films of the<br />

French New Wave – Jean-Luc Godard, Francois Truffaut, Claude Chabrol, Eric Rohmer,<br />

as well as older figures, such as Alain Resnais. These direc<strong>to</strong>rs, along with Sweden's<br />

Ingmar Bergman and Italy's Federico Fellini and Michelangelo An<strong>to</strong>nioni, among others,<br />

demonstrated that the formal, thematic, and economic givens of American Cinema were<br />

not that at all. Film had a flexible language that could be explored, opened, rethought.<br />

They proved that the conventions of Hollywood s<strong>to</strong>ry telling could be pulled and<br />

stretched, s<strong>to</strong>od on their head. These films, in short, made many people aware that<br />

cinema was a serious form of expression, an art form involved in the production of<br />

thought.<br />

The wonderful paradox here was that many of these European direc<strong>to</strong>rs learned from<br />

watching American film. Not knowing English, they read the visual structure of the films<br />

they saw and discovered a visual energy mostly missed by plot-centric American<br />

filmgoers and film reviewers. They embraced the American style and countered it<br />

simultaneously, all the while refusing the pressures of a studio system that, in America,<br />

saw films as commodities. And, as the writing and the filmmaking of the French New<br />

Wave reached Britain and then the USA, there was an interesting flow-back effect,<br />

turning the eyes of burgeoning American film scholars <strong>to</strong>ward their own cinema, largely<br />

reviled, mostly forgotten.<br />

In their discovery of American film, the French – concentrating on the visual aspects of<br />

what they saw, and understanding that film was essentially about the construction of<br />

images – noticed that when the structures of formal and thematic element cohered, it was<br />

around the film's direc<strong>to</strong>r – no matter what other personnel were involved, including the<br />

screenwriter. As these insights were passed overseas, American film suddenly <strong>to</strong>ok on a<br />

patina of seriousness that only authorship could give it. This, coupled with the excitement<br />

of the new cinema coming from overseas, and the call by students for a broader college<br />

curriculum, led <strong>to</strong> film courses being offered by English professors (like myself), and in<br />

his<strong>to</strong>ry and art his<strong>to</strong>ry departments. Film Studies departments came <strong>to</strong>gether more<br />

slowly. Publishing scholarly film articles and books grew apace.<br />

What was taught and studied in the early days? The auteur theory, in which the film<br />

direc<strong>to</strong>r is seen as the formative creative consciousness of a film, turned out <strong>to</strong> be in<br />

practice not mere idolatry, but a means of analysis. If one could identify a filmmaker by


certain stylistic and thematic traits, these could <strong>to</strong> be unders<strong>to</strong>od and analyzed. Or, taking<br />

a Foucauldian turn, the auteur could be constructed from a group of films, discovered as<br />

auteurs through their work, the analysis of which yields ways of cinematic seeing that<br />

were recognizable from film <strong>to</strong> film.<br />

By the late 1960s and in<strong>to</strong> the 1970s and 1980s, the fundamentals of auteuristn were<br />

enriched by a number of other theoretical and his<strong>to</strong>rical practices. In fact, film studies<br />

was among the earliest disciplines <strong>to</strong> apply feminist and gender theory. Laura Mulvey's<br />

theory of the formal structures of the gendered gaze in her essay "Visual Pleasure and the<br />

Narrative Cinema", published in 1975, remains a <strong>to</strong>uchs<strong>to</strong>ne not only for film studies, but<br />

for art and literary analysis as well. Ideological criticism and cultural analysis, Lacanian<br />

psychoanalytic theory, structuralism, postmodern critique – indeed a variety of theoretical<br />

permutations – have built film (and, finally, television) studies in<strong>to</strong> a rich and productive<br />

discipline.<br />

Throughout this period of growth (if not expansion), one problem remained in both the<br />

teaching and writing of film: the ability <strong>to</strong> prove and demonstrate by means of quotation.<br />

In other words, the literary scholar, the his<strong>to</strong>rian, the art his<strong>to</strong>rian can literally bring the<br />

work under study in<strong>to</strong> her own text <strong>to</strong> prove a point, illustrate an argument, and provide<br />

the text and context for analysis. The film scholar cannot.<br />

In teaching, the ability <strong>to</strong> analyze sequences in front of a class was better served, going<br />

through a long phase of never quite adequate development. In the beginning, there were<br />

only 16 mm copies of films <strong>to</strong> show and analyze. The equivalent of Xerox copies, 16 mm<br />

film was terrible for general viewing, and selecting a sequence <strong>to</strong> study involved putting<br />

pieces of paper in the film reel as it ran and then winding it back <strong>to</strong> find the passage. One<br />

could also put the reel on manual rewinds and use a small viewer <strong>to</strong> find the passage, and<br />

then remount the whole thing back on the projec<strong>to</strong>r. At one point, the department I was<br />

working in purchased an "analyzer" projec<strong>to</strong>r – the kind that was used by football teams<br />

before videotape. One still needed <strong>to</strong> find the passage in the reel, but, theoretically, one<br />

could freeze frames, roll forward or backward, or even step-frame through the sequence.<br />

In reality, the machine usually spat out sprocket holes and <strong>to</strong>re up the film. Videotape<br />

was no better in image quality – worse, in fact, because the manufacturers of video<br />

believed audiences would not put up with the black bars on the <strong>to</strong>p and bot<strong>to</strong>m of the<br />

frame for wide-screen and anamorphic films, and so blew up the image <strong>to</strong> make it fit the<br />

familiar, almost square rectangle, thereby losing about two-thirds of it in a process nicely<br />

named "pan and scan."<br />

Laserdisk and now DVD came close <strong>to</strong> ameliorating the situation. The image resolution<br />

is greatly improved. Good video projection creates a sharp, color correct or accurate<br />

greyscale image in the proper screen ratio. Accessibility <strong>to</strong> the precise place in the film <strong>to</strong><br />

which we want <strong>to</strong> aim our students' attention, or have them find the sequence they wish <strong>to</strong><br />

discuss, is easy. We can now "quote" the passages of the film when teaching them.<br />

But this did not solve the quotation problem for research. The presentation and analysis<br />

of cinematic works, no matter what methodology was used, could not be given the proof


of the thing itself, the moving image upon which the analysis was based. We could only<br />

describe what we saw as a means of bolstering and driving our arguments, and, for<br />

visuals (often on the insistence of publishers), provide studio stills, which are, more often<br />

than not, publicity stills rather than frame enlargements from the film itself. To be sure,<br />

there were some painstaking researchers who, with special lenses and the permission of<br />

archives, made (and still make) their own frame enlargements directly from a 35 mm<br />

print in order <strong>to</strong> get exactly the image they need. But the images were still, and we were<br />

all writing and talking about the images in motion.<br />

The situation began <strong>to</strong> change in the early 1990s. The first indication of a new way <strong>to</strong> do<br />

film studies occurred at a meeting of the Society for Cinema Studies in 1989, when<br />

Stephen Mamber of UCLA hooked up a computer, a laserdisk player, and a television<br />

moni<strong>to</strong>r and controlled the images of the film on the laserdisk through the computer with<br />

a program he had written. This was followed the next year by a colloquium that Mamber<br />

and Bob Rosen set up at UCLA, funded by a MacArthur Grant, that advanced Mamber's<br />

work considerably. He and Rosen had created a database of every shot in Citizen Kane,<br />

Welles's Macbeth, and a film from China, Girl from Hunan (1986). They could access the<br />

shots from the computer, through a multimedia program called Toolbook – an excellent<br />

multimedia package that still exists and is among the best <strong>to</strong> do the kind of work I'm<br />

describing.<br />

The coup came shortly after this, when a colleague at the University of Maryland<br />

demonstrated image overlay hardware that put a moving image from a tape or disk<br />

directly in<strong>to</strong> a window on the computer screen. Quickly following this came inexpensive<br />

video capture hardware and software that allows the capture of small or long sequences<br />

from a videotape or DVD. The captured and compressed images (the initially captured<br />

image creates a huge file which must be compressed with software called "codecs" in<br />

order <strong>to</strong> make them usable) are like any computer file. They can be manipulated, edited,<br />

titled, animated. The film scholar and burgeoning computer user finally discovered that a<br />

fundamental problem faced by the discipline of film studies could now be solved. The<br />

visual text that eluded us in our scholarly work and even our teaching was now at hand.<br />

We could tell, explain, theorize, and demonstrate!<br />

But other problems immediately surfaced. One was obvious: how would this work be<br />

transmitted? On first thought, the web seemed the best and obvious means of distribution<br />

for this new kind of film analysis. It wasn't then; it isn't now. To analyze the moving<br />

image, it has <strong>to</strong> appear promptly, smoothly, with as near perfect resolution as possible,<br />

and at a size no smaller than 320 × 240 pixels or larger (<strong>to</strong> make things confusing, these<br />

numbers refer <strong>to</strong> image resolution, but translate on the screen as image size). This is<br />

impossible on the Web. Even <strong>to</strong>day's streaming video and high-powered computers<br />

cannot offer the image size, resolution, and smoothness of playback that are required,<br />

despite the fact that a few successful Web projects have emerged, mostly using very<br />

small images <strong>to</strong> minimize download time. Until Internet 2 becomes widely available, CD<br />

and DVD are the only suitable media for transmission of moving images. (Recordable<br />

DVD is still in its standardizing phase, with a number of formats competing with each


other.) There is something on the horizon that will allow the Web <strong>to</strong> control a DVD in the<br />

user's machine, and we will get <strong>to</strong> that in a bit.<br />

The other problem involves the programming necessary <strong>to</strong> create a project that combines<br />

text and moving image. There are relatively simple solutions: moving image clips can be<br />

easily embedded in<strong>to</strong> a Power Point presentation, for example. A Word document will<br />

also accept embedded moving images – although, of course, such an essay would have <strong>to</strong><br />

be read on a computer. HTML can be used relatively easily <strong>to</strong> build an offline Web<br />

project which will permit the size and speed of moving images necessary. But more<br />

elaborate presentations that will figuratively or literally open up the image and allow the<br />

user <strong>to</strong> interact with it, that will break up the image in<strong>to</strong> analyzable chunks, or permit<br />

operations on the part of the reader <strong>to</strong>, for example, re-edit a sequence – all require some<br />

programming skills. This is probably not the place <strong>to</strong> address the pains and pleasures of<br />

programming. There is no getting around the fact that one will need <strong>to</strong> know some basics<br />

of programming (and the varieties of online users' groups, who will answer questions);<br />

but ultimately, programming is only half the problem. The other, perhaps most important,<br />

is designing the project, creating an interface, making the screen inviting, as easy or as<br />

complex <strong>to</strong> address as the crea<strong>to</strong>r of it wishes it <strong>to</strong> be. And, preceding the interface, there<br />

must lie a well-thought out concept of what we want a user <strong>to</strong> do, <strong>to</strong> discover, <strong>to</strong> learn. In<br />

other words, in addition <strong>to</strong> the analytic and theoretical skills of a film scholar, there are<br />

added design and usability skills as well.<br />

There are a variety of <strong>to</strong>ols available <strong>to</strong> execute design and usability, each one with its<br />

particular strengths and weaknesses. Macromedia Direc<strong>to</strong>r is good for animations <strong>to</strong><br />

accompany the exposition, but it requires a great deal of programming in a non-intuitive<br />

environment. Toolbook is an excellent, Windows-only program. Its scripting language is<br />

fairly simple (or fairly complex, depending upon your need) and much of it in plain<br />

English. Visual Basic is a powerful <strong>to</strong>ol, requiring a great deal of programming<br />

knowledge. The solution for all of this, of course, is for the film scholar <strong>to</strong> work closely<br />

with a student with programming skills, so that the scholar becomes essentially a concept<br />

and content provider. On the other hand, it is very satisfying <strong>to</strong> learn the necessary<br />

program or multimedia package. More than satisfying, by working out the programming,<br />

one learns how <strong>to</strong> structure and form ideas by and for the computer – in effect<br />

understanding the grammar that will express what you want the viewer <strong>to</strong> see. By<br />

understanding the form and structure of computational possibilities, you can generate the<br />

design, interactivity, the images and analyses that integrate concept, execution, and<br />

reception.<br />

Many scholars have experimented with various modes of computer representations since<br />

the 1990s. I've indicated that the work of Stephen Mamber was responsible for getting me<br />

and others started in using the computer <strong>to</strong> analyze films. Mamber went on <strong>to</strong> do what<br />

still remains the most exciting and complex computer-driven work of cinematic analysis,<br />

his "Digital Hitchcock" project on Hitchcock's The Birds. The project started in part when<br />

the Academy of Motion Picture Arts and Sciences allowed him access <strong>to</strong> the script,<br />

s<strong>to</strong>ryboards, and other material related <strong>to</strong> the film. Programming from scratch, Mamber<br />

created a stunning presentation: beginning with the representation of the first frame of


every shot of the film, all of which he managed <strong>to</strong> put on one screen. In other words, the<br />

whole film is represented by a still of each of its shots and each still is addressable. When<br />

clicked, they bring up the entire shot.<br />

Mamber compares Hitchcock's s<strong>to</strong>ryboard illustrations side by side with moving clips<br />

from the finished screen, <strong>to</strong> demonstrate how closely Hitchcock hewed <strong>to</strong> his original<br />

conceptions. (Hitchcock was fond of saying that, for him, the most boring part of making<br />

a film was actually making the film, because he had completed it before it went <strong>to</strong> the<br />

studio floor.) Mamber's comparison of sketches with shots proves Hitchcock right and<br />

wrong, because it indicates where he deviates from the sketches <strong>to</strong> achieve his greatest<br />

effects. The program allows the clip <strong>to</strong> be played while the successive s<strong>to</strong>ryboard<br />

sketches appear next <strong>to</strong> the sequence.<br />

In this project, Mamber began what has become some of his most exciting work, the<br />

creation of 3-D mock-ups of filmic spaces. Working on the theory of "imaginary spaces",<br />

Mamber shows how the images we look at in a film could not possibly exist. They are<br />

almost the cinematic equivalents of trompe I'oeil, using the two-dimensional surface of<br />

the screen <strong>to</strong> represent a fictional space that the filmmaker deems necessary for us <strong>to</strong> see,<br />

without regard <strong>to</strong> the impossible spaces he or she is creating. By making the entire space<br />

visible by imagining a three-dimensional simulacrum of it (in effect creating a<br />

simulacrum of a simulacrum), Mamber exposes not only the fictional space, but the<br />

processes of representation itself. His spatial representations provide an analytical<br />

understanding of the continuity of movement through space that filmmakers are so keen<br />

on maintaining, despite the optical trickery used <strong>to</strong> create them. His work becomes, in<br />

effect, an exposure of the ideology of the visible. He creates an alternative world of the<br />

alternative world that the film itself creates.<br />

Mamber clarifies a film sequence through 3-D rendering, showing how it is constructed<br />

for our perception. He has made, for example, a painstaking animated reconstruction of<br />

the opening of Max Ophiils's The Earrings of Madame De…, showing the intricate<br />

movements of the camera – by putting the camera in the animation – during an amazingly<br />

long shot. He has done a still rendering of the racetrack and a flythrough of the betting<br />

hall in Kubrick's The Killing in order <strong>to</strong> show the spatial analogues of the complex time<br />

scheme of the film. The line of his investigation is of enormous potential for all narrative<br />

art because he is essentially discovering ways of visualizing narrative. Film is, obviously,<br />

the perfect place <strong>to</strong> start, because its narrative is visual and depends upon spatial<br />

relationships, as well as the temporal additions of editing. But literary narrative also<br />

depends upon the building of imaginary spaces, and the kind of visualizations Mamber is<br />

doing on film could go far <strong>to</strong>ward a mapping of the spaces of the s<strong>to</strong>ry we are asked by<br />

fiction <strong>to</strong> imagine. 1<br />

Other pioneering work in the field of computer-driven film studies includes Marsha<br />

Kinder's 1994 <strong>companion</strong> CD-ROM <strong>to</strong> her book on Spanish film, Blood Cinema. Here,<br />

film clips and narration elaborate the elements of Spanish cinema since the 1950s. Lauren<br />

Rabinowitz's 1995 CD-ROM, The Rebecca Project, is a wonderful example of how CD-<br />

ROMs are able <strong>to</strong> combine images, clips, critical essays, and other documents <strong>to</strong> drill as


deeply and spread as widely information and analysis of Hitchcock's first American film.<br />

Robert Kapsis's Multimedia Hitchcock, displayed at MOMA, also brings <strong>to</strong>gether a<br />

variety of information, images, clips, and commentary on Hitchcock's work. Similarly,<br />

Georgia Tech's Griffith in Context (Ellen Strain, Greg Van Hoosier-Carey, and Patrick<br />

Ledwell), supported by a National Endowment for the Humanities (NEH) grant, takes<br />

representative sequences from Birth of a Nation and makes them available <strong>to</strong> the user in a<br />

variety of ways. The user can view the clips, attempt a re-editing of them, listen <strong>to</strong> a<br />

variety of commentary from film scholars, view documents, and learn about the racial<br />

and cultural context surrounding Griffith's work. MIT's Virtual Screening Room (Henry<br />

Jenkins, Ben Singer, Ellen Draper, and Janet Murray), also the recipient of an NEH grant,<br />

is a huge database of moving images, designed <strong>to</strong> illustrate various elements of film<br />

construction. Adrian Miles's "Singin' in the Rain: A Hypertextual Reading", appearing in<br />

the January, 1998, film issue of Postmodern Culture (at<br />

, is one<br />

of the most ambitious projects involving image and critical analysis, and one of the few<br />

successful web-based projects with moving images. Miles uses small and relatively fastloading<br />

Quicklimes with an intricate hyper-textual analysis – hypervisual might be more<br />

accurate – that not only explores the Kelly-Donen film, but experiments with the<br />

possibilities of a film criticism that follows non-linear, reader-driven paths in which the<br />

images and the text simultaneously elucidate and make themselves more complex. This is<br />

textual and visual criticism turned in<strong>to</strong> Roland Barthes's writerly text.<br />

My own work has followed a number of paths. Early attempts emerged out of an essay I<br />

wrote on Martin Scorsese's Cape Fear (1991). Cape Fear is not great Scorsese, but,<br />

watching it, I was struck by something very curious I could not quite put my finger on: I<br />

could see other films behind it, in it, lurking like ghosts. None of these ghosts were the<br />

original 1961 Cape Fear, an early Psycho imitation. But it was Hitchcock I was seeing,<br />

especially early 1950s Hitchcock, before he had got in<strong>to</strong> his stride; I was seeing<br />

Stagefright (1950), Strangers on a Train (1951), and / Confess (1953). The proof of my<br />

intuition appeared as soon as I looked closely at these three films. And what I had<br />

intuited were not ghosts, but a kind of palimpsest, images and scenes that lay under<br />

Scorsese's film. He was quoting – indeed recreating scenes – from these earlier films.<br />

Writing the essay analyzing these quotes didn't seem sufficient: I wanted a way <strong>to</strong> show<br />

them.<br />

I built a very simple Toolbook program, using a laserdisk and an overlay window, which<br />

merely had a screen of but<strong>to</strong>ns naming the various scenes from Strangers on a Train and<br />

Cape Fear which, when pressed, and with the proper side of the appropriate disk in the<br />

player, displayed the images on the computer screen. The interface was plain, but the first<br />

step had been taken. The laserdisk/computer interface was extremely confining. There<br />

needed <strong>to</strong> be a way <strong>to</strong> capture the images and have them on the computer hard disk,<br />

quickly available and easy <strong>to</strong> insert in<strong>to</strong> a program. Reasonably priced image capture<br />

boards that allowed easy digitizing and compression of short-duration clips, even on the<br />

then low-powered desk-<strong>to</strong>p PCs, were already being developed in the early 1990s. This<br />

was the obvious answer <strong>to</strong> the problem: making image files as accessible and<br />

manipulable as any other <strong>digital</strong> artifact. Here was hardware and software that turned the


moving image in<strong>to</strong> binary code, and once so encoded, almost anything could be done<br />

with it.<br />

With the encouragement and assistance of John Unsworth at the Institute of Advanced<br />

Technology in the Humanities, I wrote I kind of manifes<strong>to</strong>, "The Moving Image<br />

Reclaimed", for a 1994 issue of lATH's online journal Postmodern Culture<br />

(http://muse.jhu.edu/login?uri=/journals/postmodern_culture/v005/5.1kolker.html). The<br />

essay included various moving image files, including the major comparisons of Cape<br />

Fear and Strangers on a Train. It also included an idea of how the computer could be<br />

used literally <strong>to</strong> enter the image and diagram it <strong>to</strong> show how it worked. I <strong>to</strong>ok part of a<br />

very long shot in Citizen Kane, where Mrs Kane sells her son <strong>to</strong> Mr Thatcher for a deed<br />

<strong>to</strong> a copper mine – a sequence in which the movement of the camera and the shifting<br />

positions of the characters tell more interesting things about the s<strong>to</strong>ry than the dialogue<br />

does – and animated it. That is, I plucked each frame from the sequence, and, overlaying<br />

the frame with various colored lines, put <strong>to</strong>gether a variation of a ro<strong>to</strong>scope (an animation<br />

technique where live action is traced over and turned in<strong>to</strong> a car<strong>to</strong>on). The result was a<br />

visualizing – a map – of how eyeline matches (the way a film allows us <strong>to</strong> understand<br />

who is looking at whom, and why) and the changing positions of the various characters<br />

were staged and composed in order <strong>to</strong> represent the spaces of Oedipal conflict.<br />

Unsworth and I decided that MPEG would be the best format for the purpose (this was<br />

some time before the development of streaming video), though we could not transmit the<br />

sound. The essay and images were put online. The images were large, and they <strong>to</strong>ok a<br />

great deal of time <strong>to</strong> download, but the project successfully proved the ability <strong>to</strong> solve the<br />

quotation problem.<br />

While it was a worthwhile experiment, it confirmed that the Web was an imperfect<br />

transmission vehicle for moving images. The next project was the creation of an<br />

interactive program that would provide a visual version of an introduction <strong>to</strong> the basic<br />

elements of film study. Pro<strong>to</strong>typing in Toolbook, I created a series of modules on editing,<br />

point-of-view, mise-en-scene, lighting, camera movement, and other issues that seemed<br />

<strong>to</strong> me most amenable <strong>to</strong> using digitized clips. This was not <strong>to</strong> be a program on how <strong>to</strong><br />

make a film, but, <strong>to</strong> borrow James Monaco's term (who has himself recently published a<br />

DVD full of text and moving images), how <strong>to</strong> read a film. The program uses text and<br />

moving image, both often broken up in<strong>to</strong> successive segments <strong>to</strong> show how a sequence is<br />

developed, and always allowing the user <strong>to</strong> look at the entire clip. It included interactivity<br />

that allowed the user <strong>to</strong>, for example, put <strong>to</strong>gether a montage cell from Eisenstein's<br />

Potemkin or see the results of classical Hollywood three-point lighting by "turning on"<br />

the key, fill, and backlighting on a figure, or step through a sequence in Vertigo <strong>to</strong> show<br />

how camera movement and framing tell a s<strong>to</strong>ry different from the one a character in the<br />

film is telling. It contained a glossary, so that by clicking on any hotword in the text, one<br />

would jump <strong>to</strong> a definition of the term.<br />

The point-of-view (POV) module was among the most challenging. POV is a difficult<br />

concept <strong>to</strong> analyze even in the traditional modes of film theory (though Edward Branigan<br />

and others have made important contributions <strong>to</strong> our understanding). How <strong>to</strong> show how


point-of-view works <strong>to</strong>ward narrative ends – how the film guides our gaze and provides a<br />

"voice" for the narrative – required careful choices and execution. I went <strong>to</strong> a film that<br />

was about point-of-view, Hitchcock's Rear Window (1954). (It is, incidentally, no<br />

accident that so many film/computer projects focus on Hitchcock, one of the most<br />

complex formalists of American film.) By combining stills and moving images, winding<br />

up finally with a 3-D flythrough that indicates the perceptual spaces of the film and the<br />

way they turn at a crucial moment, the user is able <strong>to</strong> gain a visual grip on the process and<br />

understand how filmmakers make us see the way they want us <strong>to</strong> see.<br />

The creation and publication of the project, called Film, Form, and Culture, offers a<br />

useful example of other problems and successes that arise once you have decided <strong>to</strong> go<br />

<strong>digital</strong>. For one thing, most publishers are reluctant <strong>to</strong> publish CDs or DVDs alone. They<br />

are, after all, book publishers, and they still want paper. While continuing <strong>to</strong> pro<strong>to</strong>type<br />

the CD-ROM project, I wrote an introduc<strong>to</strong>ry textbook, which not only introduces terms<br />

and concepts basic <strong>to</strong> film studies, but also examines the his<strong>to</strong>ry of film in the contexts of<br />

the cultures that produced them. The book was not written as a supplement, but as a<br />

stand-alone <strong>companion</strong> <strong>to</strong> the CD. Elements in the text that are covered on the CD are<br />

cited at the end of each chapter. And the book itself is illustrated by <strong>digital</strong> stills – that is,<br />

stills grabbed by the computer directly from the DVD or VHS of a film, a process which<br />

allows creating a sequence of shots which, while still, convey some of the movement<br />

between them, as well as indicating ways in which editing connects various shots.<br />

The publisher of the work, McGraw-Hill, bid out for a professional CD-authoring<br />

company <strong>to</strong> do the distribution CD. An important lesson was learned from this. The film<br />

scholar, used <strong>to</strong> dealing politely with edi<strong>to</strong>rs, who, more often than not, had the right<br />

ideas on how <strong>to</strong> improve the language, organization, and accuracy of a manuscript, is<br />

here confronted with the age-old problem of "crea<strong>to</strong>r" (or, in the discourse of CD or Web<br />

authoring, the "content provider") vs. the producer, who was working within a budget<br />

provided by the publisher. The CD producer converted my pro<strong>to</strong>type in<strong>to</strong> Macromedia<br />

Direc<strong>to</strong>r, in order <strong>to</strong> create a cross-platform product. We worked closely <strong>to</strong> choose a<br />

suitable interface, but some small "creative differences" ensued. The producer wanted a<br />

uniform interface, so that each screen would look the same. This is common<br />

multimedia/web-design practice, though it can limit flexibility, especially when one<br />

wants <strong>to</strong> break the limits of the screen frame, or present alternative ways of laying out the<br />

material. There was an urge <strong>to</strong> cut down on interactivity and slightly diminish an<br />

immersive experience for the user. They insisted on small anomalies, such as not<br />

allowing the cursor <strong>to</strong> change in<strong>to</strong> a hand icon over a link.<br />

These were small, limiting constraints, and in some cases, convincing was needed on my<br />

part, <strong>to</strong> maintain my original conceptions, especially when these involved pedagogical<br />

necessities. But like all such give and take, the constraints were turned <strong>to</strong> advantages,<br />

and, <strong>to</strong> the producers' great credit, the last module, on music, turned out <strong>to</strong> be one of the<br />

best on the CD, and for that module, I provided only concept, assets (images and sounds),<br />

and text; they developed the appearance of the section based upon some preliminary<br />

pro<strong>to</strong>typing on my part – which used Sergei Eisenstein's graphic visualization of how<br />

Prokofiev's score and Eisenstein's images worked in perfect abstract relationship <strong>to</strong> each


other in Alexander Nevsky (1938). In the second edition of the CD, many problems were<br />

solved. The cursor changes <strong>to</strong> a hand over a link. The new material on film noir and on<br />

sound are well executed. The interface remains, and even more interactivity was and can<br />

still be added. As with all <strong>digital</strong> projects, it is a work in progress, always undergoing<br />

improvement with each new edition and the help of very supportive publishers.<br />

But one issue, much more immediately pressing than dealing with producers, hangs like a<br />

cold hand over the project of using <strong>digital</strong> media in the study of film or any other<br />

discipline. This is the issue of copyright and intellectual property law (IP). Simply put, IP<br />

is the legal issue of who owns what, who can use it, and where it can be used. But the<br />

issue is not simple. The US government is trying <strong>to</strong> make copyright laws more and more<br />

inflexible and has made Fair Use (the legally permitted use of small portions of a<br />

copyrighted work for educational purposes) more difficult <strong>to</strong> apply, beginning with the<br />

1990 Digital Millennium Copyright Act. But many of the new laws are undergoing a<br />

challenge. And, with the major exception of Napster, there have been few major test<br />

cases <strong>to</strong> indicate how the various parties – content owner and content user – would fare in<br />

court. No one wants <strong>to</strong> go first!<br />

It might be helpful <strong>to</strong> briefly outline the major copyright laws and then go in<strong>to</strong> some<br />

more detail on the major problems in gaining licenses, rights, and permissions for <strong>digital</strong><br />

media, with the full knowledge that such a complex issue can never be completely<br />

elucidated in a short space.<br />

• The overwhelming majority of feature films are under copyright.<br />

• Copyright has (in its original form) limits:<br />

– Published before 1923, the work is in the Public Domain (PD).<br />

– Works created after January 1, 1978, are copyrighted for the life of the author (in film,<br />

often considered the Studio) plus 70 years.<br />

– Works published from 1923 <strong>to</strong> 1963: copyright holds for 28 years, renewable up <strong>to</strong> 67<br />

years. There are other variations, but these are the most pertinent.<br />

– The Sonny Bono Copyright Term Extension Act extends copyright on many works for<br />

as much as 95 years, and gives foreign works this extension even if the works are not<br />

formally copyrighted. Works published in or before 1922 remain in the Public Domain.<br />

Bono is under litigation in the Supreme Court. The restrictive 1990 Digital Millennium<br />

Copyright Act is also undergoing challenges.<br />

– Legislation has been passed for the "Technology, Education and Copyright<br />

Harmonization Act" (TEACH), that aims <strong>to</strong> modify copyright restrictions for Distance<br />

Education.


– Films in the Public Domain may have "subsidiary rights." That is, while the film may<br />

have come in <strong>to</strong> PD, parts of it, the score, for example, may have been separately<br />

copyrighted. Carol Reed's The Third Man is an example.<br />

– Licensing clips for a CD – even an educational one – takes a strong s<strong>to</strong>mach, a very<br />

busy fax machine, and often a lot of money.<br />

Copyright issues are, perhaps even more than developing one's work, the greatest<br />

hindrance <strong>to</strong> scholars making use of <strong>digital</strong> interventions in film studies. However, they<br />

are not insuperable. Copyright concerns and the energy required for licensing clips<br />

should in fact s<strong>to</strong>p no one, but only cause one <strong>to</strong> act cautiously and judiciously, especially<br />

if the work is aimed at publication. The Library of Congress (the reposi<strong>to</strong>ry for all<br />

copyrighted material) has, for example, a three-volume listing of films in the Public<br />

Domain. The last volume has an appendix that, as of the early 1990s, lists films for which<br />

extensive research on subsidiary rights was made. Many of these films are available on<br />

magnetic or <strong>digital</strong> media. This is the best first source for what is freely available –<br />

especially if the Bono Act is overthrown.<br />

After the research for PD titles, the very best way <strong>to</strong> start the rights and permissions<br />

process, or at least get an entry <strong>to</strong> the studios who own copyrighted material, is <strong>to</strong> contact<br />

the film's direc<strong>to</strong>r – assuming he or she is alive. For Film, Form, and Culture, one<br />

filmmaker gave us complete use of one of his films, and signed a clearance for it.<br />

Another, Oliver S<strong>to</strong>ne, a great filmmaker and an admirer of film studies, wrote <strong>to</strong> each of<br />

his distribu<strong>to</strong>rs asking them <strong>to</strong> help me out. Sometimes even this will not prevail. Disney,<br />

whose Hollywood Films released Nixon, appreciated the personal note, but announced<br />

their policy of never giving permissions for CDs. However, Warner Bros did license<br />

thirty seconds of JFK, and the same length of clip from Citizen Kane, for $3,000 apiece.<br />

For the second edition, I added a module on film noir. I was most interested in using<br />

images from one of the greatest of the late-1940s noir direc<strong>to</strong>rs, Anthony Mann. His<br />

dark, brutal, misanthropic films are currently owned by the famous publisher of children's<br />

literature, Golden Books. The ways of copyright are strange, indeed. Golden Books<br />

charged $1,000 for sixty seconds. In the great scheme of licensing clips, this is not a lot<br />

of money, and publishers who understand the value of having a CD with a book will pay<br />

such a relatively modest fee.<br />

For other clips, the Hitchcock Estate, who, when I started the project, still controlled Rear<br />

Window and Vertigo, worked with Universal pictures, who – unlike the Estate, but at<br />

their urging – reluctantly allowed me <strong>to</strong> use clips from those films. My understanding is<br />

that Universal and other studios have since streamlined both their rights process and their<br />

attitude. Everything else used was Public Domain.<br />

Certainly, none of this was easy (though now it may be at least easier, given such<br />

companies as RightsLine and RightslQ, that can help track who holds various rights <strong>to</strong> a<br />

film). In every successful instance of dealing with a studio, the result will be a long and<br />

intimidating contract, which must be read carefully. The result, however, is worth the


labor. The work of copyright searches and acquisition of licenses allowed Film, Form,<br />

and Culture, for example, <strong>to</strong> present students with access <strong>to</strong> clips from a large variety of<br />

films and uses them in some detail <strong>to</strong> show the precision with which they are made and a<br />

variety of ways in which they can be read and analyzed.<br />

I would only repeat that IP concerns should not be the initial obstacle <strong>to</strong> using moving<br />

images in a pedagogical or scholarly work. There are many ways <strong>to</strong> get what you need, if<br />

not always what you want. And Fair Use should still prevail for in-class pedagogical use,<br />

for <strong>digital</strong> stills in books, and, since the TEACH Act has been passed, for wider<br />

educational distribution. The studios have become somewhat more accommodating, a<br />

response that may have <strong>to</strong> do with DVDs and the unexpected popularity of their<br />

"supplementary material." DVDs also offer something of a new frontier in the<br />

development of digitized media in film studies and, hopefully, a relief for copyright<br />

concerns.<br />

DVDs have proven a boon <strong>to</strong> film scholars and film viewers alike. They have<br />

compensated for the culture's massive error in visual judgment when it chose VHS over<br />

Beta (which had a higher image resolution) in the 1980s. The success of DVD has proven<br />

that image quality does count. They have been of great help <strong>to</strong> film studies: they are<br />

cheap for an academic department <strong>to</strong> buy; they are readily available for students <strong>to</strong> rent<br />

and view at home. Their "supplementary material" (direc<strong>to</strong>rs' and ac<strong>to</strong>rs' commentaries,<br />

demonstrations on how scenes are shot, and how computer graphics are integrated in<strong>to</strong> a<br />

film) have proven enormously popular, even though they are mostly illustrative and<br />

anecdotal. This success, as I noted, came as something of a surprise <strong>to</strong> the studios, and<br />

the outcome is that they may be taking notice of the fact that viewers may, first of all, be<br />

interested in older titles, and, second, do want <strong>to</strong> know more about the film they are<br />

watching. In fact, studios may now be admitting that a film might have educational value!<br />

Herein lies some hope on the licensing issue. The studios (and this is pure fantasy at the<br />

moment) may some day, when a scholar tells them that use of their material will help<br />

further sales of their films by educating new viewers, and without "copying" clips, begin<br />

<strong>to</strong> accept the claim and therefore restrain themselves from deriding the person who makes<br />

such an argument.<br />

Use of DVDs for analysis is being made possible by some new technology that opens<br />

interesting possibilities of using commercially available DVDs, controlling them as one<br />

would a digitized file on the computer, and even creating a set of analytic <strong>to</strong>ols available<br />

on a CD or on a website <strong>to</strong> address the DVD in the individual's own computer. Some of<br />

this work has received funding from the National Endowment for the Humanities.<br />

Again, programming issues are at stake here, and various choices have <strong>to</strong> be made,<br />

depending upon whether the crea<strong>to</strong>r of a DVD-based project wants <strong>to</strong> do a standalone<br />

(that is, a project that exists only on a computer or CD-ROM, along with the DVD) or a<br />

web-based project that contains the controls, a database, and an interface online, which<br />

address the DVD in the user's computer. Indeed, unlike digitized clips, addressing the<br />

DVD from the Web does not result in image degradation, because the image remains on<br />

the user's computer.


The basics for building a set of DVD controls are reasonably straightforward, if you have<br />

some facility at cutting and pasting code. Microsoft offers a complete programmer's<br />

guide <strong>to</strong> DVD control, down <strong>to</strong> the very frame. Full information and very simple code is<br />

given at . Drill through this<br />

page and you will find script that you can cut and paste in<strong>to</strong> an HTML file or a Power<br />

Point slide <strong>to</strong> see how it works. There should be utilities, or plug-ins, available for the<br />

Mac and Mac-based Direc<strong>to</strong>r that would perform similar operations. The Windows<br />

version makes use of the MSWebDVD ActiveX control (the program that contains the<br />

window in which the DVD will play as well as the instructions for writing very simple<br />

code). It is possible that the Javascript version of the code <strong>to</strong> control the DVD will also<br />

function on a web-based project on the Mac, and the latest version of Macromedia<br />

Direc<strong>to</strong>r includes DVD controls, though I have not tested this. The work I have done so<br />

far has been in Toolbook and Visual Basic for standalone use.<br />

Again, the coding issues can be overcome in various ways. But there is yet another issue<br />

involved. While the use of a commercial DVD in no way bypasses copyright issues, it<br />

may (and I must emphasize that I, like anyone else, can only speculate) make licensing<br />

easier <strong>to</strong> obtain or cause even less of a worry for Fair Use in a classroom or other<br />

educational settings. After all, one is not digitizing copyrighted material, but using<br />

something that is already commercially available. The content provider, however, may<br />

still see this as using their material, in ways other than for home viewing.<br />

It may, finally, seem as if we have come full circle <strong>to</strong> the time when a computer<br />

controlled a laserdisk. There are important differences. DVDs are <strong>digital</strong>; laserdisks were<br />

analogue. Laserdisks had <strong>to</strong> be played either on a moni<strong>to</strong>r, or on an overlay window on<br />

the computer screen through special hardware. DVDs play directly through the<br />

computer's graphics display card – as long as there are the proper drivers (which come<br />

au<strong>to</strong>matically with any computer that has a DVD drive). All this makes control of DVDs<br />

easier and more flexible than laserdisk. In a relatively simple Visual Basic program, one<br />

can even capture a still from the DVD directly <strong>to</strong> disk. I have been able <strong>to</strong> program a<br />

DVD so that, with user input, the film is paused, the exact time of the pause is passed <strong>to</strong><br />

the computer, which then uses that time <strong>to</strong> call up other information I have prepared in a<br />

database that is relevant <strong>to</strong> that moment or shot in the film. This can be combined with<br />

other images and analysis that further explains the shot or the sequence it is part of- and<br />

finally the film as a whole. In other words, this combination of database and image, along<br />

with the interactivity that gives the user choice over what part of the film he wishes <strong>to</strong><br />

examine, provides an invaluable method <strong>to</strong> thoroughly analyze a single work.<br />

The database, created and connected <strong>to</strong> the DVD, contains information for every shot in<br />

the film – from dialogue, discussion of the narrative moment, mise-en-scène, editing<br />

patterns, and so on. A database is essentially a static set of cubicles, each containing<br />

various kinds of data, which can be organized and then accessed in a variety of ways. We<br />

use them continuously: whenever you go online <strong>to</strong> use the campus library, <strong>to</strong> shop, or <strong>to</strong><br />

buy an airline ticket, you are manipulating a database. Terabytes of personal information<br />

(more and more being added each day) are s<strong>to</strong>red by government agencies in databases. It<br />

is not a stretch <strong>to</strong> say that the computerized database is the one completely structured


item in an otherwise chaotic world – though this is no guarantee of the veracity of the<br />

data it contains. It must be ordered correctly and filled with well-thought out data, and at<br />

the same time accessible, and manipulable, containing data in the form of text, numbers,<br />

sounds, or images that can be drawn upon in a number of different ways. Using a simple<br />

scripted language called SQL (Standard Query Language, better known as "Sequel"), the<br />

database becomes a malleable, flexible, responsive thing, alive <strong>to</strong> query and incredibly<br />

responsive if you ask the right questions of it. Database tables are themselves<br />

combinable, open <strong>to</strong> cross-queries, with the ability <strong>to</strong> pull <strong>to</strong>gether a large array of<br />

information.<br />

Lev Manovich (<strong>2001</strong>: 226) theorizes that the database is essential <strong>to</strong> computer aesthetics,<br />

part or, perhaps more appropriately, the genesis of the various narratives – including<br />

analytic and theoretical narratives – we can create by computer. "Creating a work in new<br />

media", he writes, "can be unders<strong>to</strong>od as the construction of an interface <strong>to</strong> a database",<br />

the means of accessing, combining, temporalizing, and spatializing static bits of data.<br />

Allow me another example from my own work: before DVDs were available, I had<br />

started work analyzing Robert Altman's 1993 film Short Cuts, using a five-minute<br />

digitized clip. My initial interests were <strong>to</strong> create a search engine so that various key terms<br />

(say, "zoom", a camera lens movement that Altman turns in<strong>to</strong> an aesthetic and narrative<br />

statement) would yield the corresponding shots in which the zoom was used. What I<br />

would like <strong>to</strong> do now, when the DVD becomes available, is <strong>to</strong> create a critical narrative<br />

of the entire film, especially concentrating on its complex narrative structure and the way<br />

that complexity is made perfectly comprehensible <strong>to</strong> the viewer through Altman's editing,<br />

his use of color, the matching of character movement, and, throughout, an abstract,<br />

intuited pattern of movement across the narrative fields of the film, which, indeed, can<br />

itself be thought of as a sort of database of narrative events.<br />

The creation of a database for a three-hour film is no easy task. There are a number of<br />

decisions <strong>to</strong> be made in order <strong>to</strong> make it become dynamic, and all of these must be done<br />

by the imagination and work of the film scholar. There is little au<strong>to</strong>mation here, but an<br />

extraordinary opportunity <strong>to</strong> learn about the film. One must choose the various elements<br />

<strong>to</strong> describe and analyze the shot – for example, narrative elements, mise-en-scene,<br />

editing, dialogue, etc. And, indispensably, though not very creatively, the "in" and "out"<br />

points for each shot – that is, where a shot ends and where the next shot begins must be<br />

found and entered in the database. The in and out points are where the program will first<br />

go <strong>to</strong> find the shot and the relevant information the user wants for that shot. The filling-in<br />

of the database fields is where the major analytic work is done; we have <strong>to</strong> make the<br />

same judgments, the same analytic insights, and apply the same methodologies as we<br />

would writing an article or a book. The difference is a kind of fragmentation and the<br />

adoption of a more aphoristic style than the critical writer is used <strong>to</strong>. At the same time,<br />

the writer has <strong>to</strong> be aware of a thread of analysis running through all the database entries<br />

and make sure that the keywords that the user may want <strong>to</strong> search on appear wherever<br />

appropriate.


The ability <strong>to</strong> search is a major advantage of a database: if the user wants <strong>to</strong> find any<br />

place in the film where a word is used, a color combination is present, a narrative element<br />

can be found, she should be able <strong>to</strong> enter that word, click a but<strong>to</strong>n, and have a list pulled<br />

from the database. Clicking on any entry in the list will bring up the accompanying shot.<br />

The search function opens up an entirely new problem in film analysis. We are, in effect,<br />

giving over some critical work <strong>to</strong> the user of the program. This opens up difficult critical<br />

and functional questions, the most striking of which is: what is the reader going <strong>to</strong> want<br />

<strong>to</strong> search for? Do we second-guess when we create the database, or create a list of<br />

keywords that we know the user may want <strong>to</strong> search on, and then make certain that they<br />

appear in the appropriate fields of the database? Do we provide such a list in the program<br />

itself, perhaps as part of the introduc<strong>to</strong>ry apparatus, thereby suggesting what the user<br />

might want <strong>to</strong> find? When we go <strong>to</strong> a Web search engine, like Google, we know more or<br />

less what we are looking for, even though what we get may not precisely fit our query.<br />

Going <strong>to</strong> the interface of a film database and guessing what we want <strong>to</strong> find, is a problem<br />

of a different order. We have, somehow, <strong>to</strong> present a critical apparatus within the<br />

interface that discusses the program's intent and offers some guidance for the user. As<br />

much interactivity as we provide, we must provide as well a guide <strong>to</strong> the critical narrative<br />

we are striving <strong>to</strong> get the user <strong>to</strong> put <strong>to</strong>gether.<br />

We have, then, discovered a second major advantage of the use of the computer in film<br />

studies. The first was the ability <strong>to</strong> quote from and analyze a given sequence of a film.<br />

The second is the database, as huge as one wants, and as fine-tuned as one needs, full of<br />

ideas, information, analysis, and offering the user the ability <strong>to</strong> connect the data <strong>to</strong> the<br />

precisely relevant images within a complete film.<br />

What we need, finally, <strong>to</strong> make this complete is a program that searches images<br />

themselves. We can tag images in the database and search that way. In other words, we<br />

can describe the image content, color, composition, and narrative relevance, and allow<br />

the user <strong>to</strong> choose these and bring up the related images or shots. A new codec, MPEG-7,<br />

promises the ability <strong>to</strong> tag the image itself. But all these are still text-based. We have <strong>to</strong><br />

write out a description for, or appended <strong>to</strong>, the image and then search for it by entering<br />

the keyword. There is software available <strong>to</strong> search still images by example: that is, by<br />

clicking on one image, other images with similar colors or shapes will be called up.<br />

These are complex applications, not yet easily available <strong>to</strong> a film scholar.<br />

Searching moving images is another matter still. The ability, for example, <strong>to</strong> search a<br />

database of Short Cuts for zoom shots of specific kinds, based merely on clicking on an<br />

example of that kind, would open up new avenues for studying a film's textuality, an<br />

auteur's style, and, most important, begin <strong>to</strong> enable us <strong>to</strong> understand the structure of<br />

cinematic representation itself. Software <strong>to</strong> search moving images is slowly being<br />

developed, although a researcher at Kodak <strong>to</strong>ld me "not in our lifetime." But that was six<br />

years ago.<br />

Note


1 Another aspect of Mamber's investigations, on the theory and ideology of the<br />

surveillance camera, can be found online at .<br />

References for Further Reading<br />

Branigan, Edward (1984). Point of View in the Cinema: A Theory of Narration and<br />

Subjectivity in Classical Film. Berlin and New York: Mou<strong>to</strong>n.<br />

Kolker, Robert (1994). The Moving Image Reclaimed. Postmodern Culture 5.1<br />

(September). At<br />

http://muse.jhu.edu/login?uri=/journals/postmodern_culture/v005/5.1kolker.html.<br />

Kolker, Robert (1998). "Algebraic Figures: Recalculating the Hitchock Formula." In<br />

Play It Again, Sam: Retakes on Remakes, ed. Andrew Hor<strong>to</strong>n and Stuart Y. McDougal.<br />

Berkeley: University of California Press.<br />

Mamber, Stephen (1998). Simultaneity and Overlap in Stanley Kubrick's The Killing.<br />

Postmodern Culture (January). At<br />

http://muse.jhu.edu/login?uri=/journals/postmodern_culture/v008/8.2mamber.html.<br />

Manovich, Lev (<strong>2001</strong>). The Language of New Media. Cambridge, MA: MIT Press.<br />

Miles, Adrian (1998). Singin' in the Rain: A Hypertextual Reading. Postmodern Culture<br />

(January). At<br />

http://muse.jhu.edu/login?uri=/journals/postmodern_culture/v008/8.2miles.html.<br />

Mulvey, Laura (1975). Visual Pleasure and the Narrative Cinema. Screen 16, 3<br />

(Autumn): 6–18. Reprinted at http://www.bbk.ac.uk/hafvm/staff_research/visuall.html.<br />

Information on copyright is from Lolly Gasaway, University of North Carolina<br />

http://www.unc.edu/~unclng/public-d.htm. David Green provided further information on<br />

the Bono Act and treats it and many other IP issues in Mary Case and David Green,<br />

"Rights and Permissions in an Electronic Edition", in Electronic Textual Editing, ed. Lou<br />

Burnard, Katherine O'Brien and John Unsworth (New York: MLA, forthcoming).<br />

Among many sources for information on IP law and policy are Ninch (the National<br />

Initiative for a Networked Cultural Heritage), at http://www.ninch.org, the Digital Future<br />

Coalition, http://www.dfc.org, and the American Library Association,<br />

http://www.ala.org.<br />

27.<br />

Cognitive Stylistics and the Literary Imagination<br />

Ian Lancashire


Cognitive Stylistics analyzes an author's idiolect, his individual language traits. Although<br />

cognitive psychology and neuroscience do not know how the human mind works, they<br />

have detected, through experiments on how people behave (not through personal<br />

testimony about that behavior), features of mental behavior that are consistent with a<br />

standard theory or model. That cautious experimentation uses recall and recognition tests,<br />

and EEG (electroencephalography), PET (positron emission <strong>to</strong>mography), fMRI<br />

(functional magnetic resonance imaging), and other scans. With them, scientists are<br />

painstakingly uncovering steps, and constraints in taking these steps, that characterize<br />

how we mentally create and utter sentences (no word in English covers both oral and<br />

written expressing, but perhaps "uttering" will do). The current standard model describes<br />

two language processes: an unself-conscious creative process, veiled and almost<br />

unknowable, named by older writers the Muse (Jacoby 1971); and a conscious analytical<br />

procedure by which all authors assemble and revise sentences mentally.<br />

These sciences are enhancing our understanding of how authors create both oral and<br />

written texts. New knowledge about language processing in the brain helps us interpret<br />

data from traditional computer text analysis, not because the mind necessarily works<br />

algorithmically, like a computer program, but because quantitative word studies reveal<br />

audi<strong>to</strong>ry networks, and the cognitive model asserts that the brain operates as a massively<br />

distributed group of such networks. Cognitive Stylistics underscores how every utterance<br />

is stamped with signs of its origina<strong>to</strong>r, and with the date of its inception: these traits are<br />

not unique, like fingerprints, but, taken <strong>to</strong>gether, they amount <strong>to</strong> sufficiently distinctive<br />

configurations <strong>to</strong> be useful in authorship attribution or text analysis. Cognitive Stylistics<br />

determines what those traces may be, using concordances and frequency lists of repeated<br />

phenomena and collocations: <strong>to</strong>gether, partially, these conceivably map long-term associational<br />

memories in the author's mind at the time it uttered the text. Such clusters of<br />

repetitions arise from built-in cognitive constraints on, and opportunities in, how we<br />

generate language. The length of fixed phrases, and the complexity of clusters of those<br />

collocations, are important quantitative traits of individuals.<br />

Experimental cognitive psychology res<strong>to</strong>res the human author <strong>to</strong> texts and endows<br />

computer-based stylistics with enhanced explana<strong>to</strong>ry power. The potentially<br />

distinguishable authorship traits <strong>to</strong> which it draws attention include the size of the<br />

author's personal working memory (Just and Carpenter 1992). The cognitive model<br />

shows that, far from being dead, traces of the author do remain in the work, just as<br />

Shakespeare vowed it would in his sonnets, long after his own personal death.<br />

Collocational clusters also change over time as an individual's memory does. As the<br />

implications of cognitive research become unders<strong>to</strong>od, text-analysis systems may change,<br />

and with them stylistics. Texts in languages whose spelling departs widely from its<br />

sounding will likely be routinely translated in<strong>to</strong> a phonological alphabet, before<br />

processing, as researchers recognize the primacy of the audi<strong>to</strong>ry in mental language<br />

processing. Because the standard cognitive model breaks down the old literary distinction<br />

between what is said and how it is said, interpretation and stylistics will come <strong>to</strong>gether.<br />

First Impressions


A paradox underlies stylistic research. Most of us do not know much about how we make<br />

and utter text, simply because we are so expert at uttering. The more we make texts, the<br />

less we have time <strong>to</strong> be self-conscious about doing so. Committing ourselves completely<br />

<strong>to</strong> the action as it unfolds, we no longer attend <strong>to</strong> how it takes place.<br />

Composing for anyone whose style is likely <strong>to</strong> be subject of analysis resembles walking,<br />

cycling or even, sometimes, driving a car on a highway. We can almost entirely devote<br />

our minds <strong>to</strong> other things and yet execute these actions properly. Kosslyn and Koenig say<br />

the more we know something, the harder it is <strong>to</strong> declare how we do it: "when one<br />

becomes an expert in any domain, one often cannot report how one performs the task.<br />

Much, if not most, of the information in memory cannot be directly accessed and<br />

communicated" (1992: 373). Recently one of my students, a professional ac<strong>to</strong>r, described<br />

an unnerving experience she had midway through a theatrical run of Henrik Ibsen's<br />

Enemy of the People. As she s<strong>to</strong>od, waiting <strong>to</strong> go onstage, she could not remember<br />

anything of what she was <strong>to</strong> say or do when performing her part. Yet once she walked<br />

onstage, she executed everything perfectly just in time. She then unders<strong>to</strong>od why many<br />

ac<strong>to</strong>rs repeatedly experience dread and sickness just before going on. So ingrained has<br />

the experience of performing become, over many weeks, that they can no longer summon<br />

the words or even the actions <strong>to</strong> conscious memory. Ironically, even if these ac<strong>to</strong>rs were<br />

<strong>to</strong> succumb <strong>to</strong> amnesia, they still would not lose their roles. These have just become<br />

inaccessible and so are in effect protected against loss. Amnesiacs may forget who they<br />

are, but they can all the same speak and write (Squire 1987: 161, 171; Shimamura et al.<br />

1992). Ironically, the more we do something, the less we can attend <strong>to</strong> how we do it. This<br />

well-observed cognitive effect shows how frightening and even disabling can<br />

neurological processes developed <strong>to</strong> ensure reliable performance of an essential skill be<br />

in practice.<br />

Technology partially conceals this actual neglect from anyone who composes directly<br />

on<strong>to</strong> an artificial memory device. Equipped with a pen, a typewriter, or <strong>digital</strong> editing<br />

<strong>to</strong>ols, authors see their text unfolding from their minds as they manually encode it in<br />

alphanumeric symbols on screen or paper. Making sentences visibly explicit as<br />

composed, writers no longer worry about having <strong>to</strong> s<strong>to</strong>re mentally what they create.<br />

Externalized, the words, phrases, and sentences of utterances can be easily deleted,<br />

rearranged, transformed grammatically, and replaced. We fully attend <strong>to</strong> the analytic task<br />

of doing these things. Because we externalize subvocal or inner speech immediately in<br />

visual form, we feel <strong>to</strong>tally conscious of the mental activity of composing. Yet all we<br />

experience is the s<strong>to</strong>rage and the manipulation of symbols manifested outside the mind.<br />

What happens within the mind remains dark, especially <strong>to</strong> the expert composer, although<br />

novices learning how <strong>to</strong> use a language may assemble, with painful slowness, utterances<br />

in memory, consciously, before they utter them. The inexpert attend completely <strong>to</strong> the<br />

task. They may even be able <strong>to</strong> describe what steps they take consciously, probing their<br />

memory, and editing the results mentally in advance of speaking it. Any native speaker<br />

can follow this method in preparing <strong>to</strong> utter a sentence in their own <strong>to</strong>ngue. It is so slow<br />

in natural conversation and composition as <strong>to</strong> be seldom worth using.


Expert authors in many languages often mention this neglect of attention <strong>to</strong> how they<br />

compose. Although they do not theorize their inability <strong>to</strong> see how they create sentences,<br />

their cumulative testimony is convincing. Being unable <strong>to</strong> see themselves create text, they<br />

characterize the process itself as beyond comprehension. Jean Cocteau explains: "I feel<br />

myself inhabited by a force or being – very little known <strong>to</strong> me. it gives the orders; I<br />

follow" (Plimp<strong>to</strong>n 1989: 106). Elizabeth Hardwick agrees: "I'm not sure I understand the<br />

process of writing. There is, I'm sure, something strange about imaginative concentration.<br />

The brain slowly begins <strong>to</strong> function in a different way, <strong>to</strong> make mysterious connections"<br />

(Plimp<strong>to</strong>n 1989: 113). Fay Weldon candidly admits: "Since I am writing largely out of<br />

my own unconscious much of the time I barely know at the beginning what I am going <strong>to</strong><br />

say" (Winter 1978: 42). Cynthia Ozick imagines language coming out of disembodied<br />

nothingness.<br />

I find when I write I am disembodied. I have no being. Sometimes I'm entranced in the<br />

sense of being in a trance, a condition that speaks, I think, for other writers as well.<br />

Sometimes I discover that I'm actually clawing the air looking for a handhold. The<br />

clawing can be for an idea, for a word. It can be reaching for the use of language, it can<br />

be reaching for the solution <strong>to</strong> something that's happening on the page, wresting out of<br />

nothingness what will happen next. But it's all disembodied…. I fear falling short. I<br />

probably also fear entering that other world; the struggle on the threshold of that<br />

disembodied state is pure terror.<br />

(Wachtel 1993: 16)<br />

John Hersey compares composing <strong>to</strong> "'something like dreaming' … I don't know how <strong>to</strong><br />

draw the line between the conscious management of what you're doing and this state"<br />

(Plimp<strong>to</strong>n 1988: 126). Jonathan Raban thinks of himself as taking down spoken text, as if<br />

in dictation, uttered by an "it" – the pronoun used by Cocteau – rather than emerging<br />

from his own consciousness.<br />

All writers are in some sense secretaries <strong>to</strong> their own books, which emerge by a process<br />

of dictation. You start the thing off and on the first few pages you're in control, but if the<br />

book has any real life of its own, it begins <strong>to</strong> take control, it begins <strong>to</strong> demand certain<br />

things of you, which you may or may not live up <strong>to</strong>, and it imposes shapes and patterns<br />

on you; it calls forth the quality of experience it needs. Or that's what you hope happens. I<br />

don't just sit, making conscious decisions externally about how much of my experience I<br />

am going <strong>to</strong> use.<br />

(Wachtel 1993: 120)<br />

Earlier writers name the voice dictating the text as the Muse, a word for memory. Gore<br />

Vidal describes this unknown as sounding words aloud in the mind and goes further than<br />

others in admitting how ignorant he is of where this voice and its language come from.<br />

I never know what is coming next. The phrase that sounds in the head changes when it<br />

appears on the page. Then I start probing it with a pen, finding new meanings. Sometimes


I burst out laughing at what is happening as I twist and turn sentences. Strange business,<br />

all in all. One never gets <strong>to</strong> the end of it. That's why I go on, I suppose. To see what the<br />

next sentences I write will be.<br />

(Plimp<strong>to</strong>n 1989: 63)<br />

Amy Lowell uses the word "voice" and says, like Ozick, that it comes from something<br />

disembodied, from no body.<br />

I do not hear a voice, but I do hear words pronounced, only the pronouncing is <strong>to</strong>neless.<br />

The words seem <strong>to</strong> be pronounced in my head, but with nobody speaking them.<br />

(Ghiselin 1952: 110)<br />

These very different authors all tried <strong>to</strong> watch how sentences came from their minds and<br />

gave up. They had <strong>to</strong> attribute their composition <strong>to</strong> a disembodied voice or dream that<br />

came out of nothing. In editing on the page or screen, they consciously delete and<br />

rearrange words within passages, adjust syntactic structure, re-sequence parts, and make<br />

word-substitutions, just as we all do. They do not locate the tedious and conscious<br />

sentence-building in short-term memory as the origin of genuine composition. Literary<br />

theorists agree. Colin Martindale characterizes "primary-process cognition" as "freeassociative<br />

… autistic … the thought of dreams and reveries", unlike the mind's problemsolving<br />

capability, "secondary-process cognition" (1990: 56). Mark Turner states that<br />

"All but the smallest fraction of our thought and our labor in performing acts of language<br />

and literature is unconscious and au<strong>to</strong>matic" (1991: 39).<br />

Impressions Sustained<br />

Experiments in cognitive psychology and neuroscience have supported the concept of a<br />

disembodied mental voice that "utters" sentences without exhibiting how they were put<br />

<strong>to</strong>gether.<br />

All verbal utterances, written or oral, are received by the brain, and uttered by it, in an<br />

audi<strong>to</strong>ry-encoded form. Philip Lieberman explains that, during language processing, we<br />

access "words from the brain's dictionary through their sound pattern" (2000: 6, 62).<br />

Further, we internally model whatever we expect <strong>to</strong> hear separately from what we are<br />

hearing. That is, we can only understand heard speech by modeling silently the<br />

articula<strong>to</strong>ry actions necessary <strong>to</strong> produce it. Lieberman explains that<br />

a listener uses a special process, a "speech mode", <strong>to</strong> perceive speech. The incoming<br />

speech signal is hypothetically interpreted by neurally modeling the sequence of<br />

articula<strong>to</strong>ry gestures that produces the best match against the incoming signal. The<br />

internal "articula-<strong>to</strong>ry" representation is the linguistic construct. In other words, we<br />

perceive speech by subvocally modeling speech, without producing any overt articula<strong>to</strong>ry<br />

movements.


(2000: 48)<br />

The "McGurk" effect shows that our internal model for the utterance we are decoding, for<br />

various reasons, may differ significantly from what comes <strong>to</strong> us as audi<strong>to</strong>ry speech.<br />

The effect is apparent when a subject views a motion picture or video of the face of a<br />

person saying the sound [ga] while listening <strong>to</strong> the sound [ba] synchronized <strong>to</strong> start when<br />

the lips of the speaker depicted open. The sound that the listener "hears" is neither [ba]<br />

nor [ga]. The conflicting visually-conveyed labial place-of-articulation cue and the<br />

audi<strong>to</strong>ry velar place of articulation cue yield the percept of the intermediate alveolar [da].<br />

The tape-recorded stimulus is immediately heard as a [ba] when the subject doesn't look<br />

at the visual display.<br />

(McGurk and MacDonald 1976, cited by Lieberman 2000: 57)<br />

If sound alone were responsible for what was heard, the McGurk subject would hear [ba]<br />

at all times. Because it attends <strong>to</strong> visual clues as well, the mind hears something never<br />

sounded, [da]. Only if the subject's mind manufactured speech sounds internally, drawing<br />

on all sensory evidence available, could its model differ from both audi<strong>to</strong>ry and visual<br />

clues individually.<br />

Authors who describe their inner mental activity when creating and uttering sentences are<br />

indeed correct when they characterize the "voice" they hear as bodiless. Even when we<br />

listen <strong>to</strong> the speech of others, it is not the speech of those others that we hear. It is the<br />

brain's own construct of those voices. Obviously, different brains might well perceive<br />

very different sounds, and thus words, from the same speech sounds heard from others.<br />

The mind becomes a reader of things made by a mental process that manifests itself as a<br />

bodiless, at times strange, voice.<br />

Cognitive sciences confirm authors' impressionistic descriptions of the language-making<br />

process as blank and inaccessible. The inner voice utters sentences that appear <strong>to</strong> come<br />

out of nowhere. That we hear a voice suggests we are listening <strong>to</strong> someone else, not<br />

ourselves but someone nameless, unrecognizable, and above all as distant from analysis<br />

as the mind of someone whose words we hear on a radio. We invoke such images for a<br />

form of memory of how <strong>to</strong> do something, creating utterances from natural language,<br />

where we lack the means <strong>to</strong> identify the remembering process with ourselves. That<br />

process exemplifies the expert mind failing <strong>to</strong> attend <strong>to</strong> what it is doing. During<br />

composition, we cannot correct this neglect, as we can when driving a car and suddenly<br />

wake up <strong>to</strong> a recognition that we have been driving on au<strong>to</strong>matic, unattended, for some<br />

miles. We cannot will consciousness of how our minds create utterances. Making them<br />

relies on what is termed procedural or implicit memory, in which what is recalled (that is,<br />

how <strong>to</strong> utter something) is remembered only in the act of doing it. When we try <strong>to</strong><br />

recollect something s<strong>to</strong>red implicitly, we execute the s<strong>to</strong>red procedure. The mind has<br />

never created a readable "manual" of the steps whereby it creates sentences. The only<br />

exceptions are those halting, deliberate activities in our short-term memory in which, as if<br />

on paper, we assemble an utterance, but of course this method <strong>to</strong>o, in the end, relies on


the same mysterious voice or Muse, what cognitive sciences associate with implicit<br />

memory, <strong>to</strong> get it going. That truism, "How can we know what we are going <strong>to</strong> utter until<br />

we have uttered it?", characterizes the uncertainty of waiting for an utterance <strong>to</strong> be set<br />

down on the page or screen, <strong>to</strong> be spoken aloud, and <strong>to</strong> enter working memory. In all<br />

these situations, a silent inner voice precipitates the text out of nowhere.<br />

Twenty-five years ago, Louis Milic distinguished between what writers do unconsciously<br />

in generating language (their stylistic options) and what they do consciously in "scanning,<br />

that is, evaluation of what has been generated" (their rhe<strong>to</strong>rical options; 1971: 85). Milic<br />

anticipated the distinction between implicit or procedural and explicit memory by several<br />

years (Squire 1987: 160). By insisting on the primary role of empirical, rather than of<br />

theoretical or impressionistic, evidence in the study of authorship, Milic directed the<br />

discipline of stylistics <strong>to</strong> the cognitive sciences.<br />

Why cannot we recall how we make a sentence? Why is the mind blocked by implicit<br />

memory in understanding one of the most critical defining features of a human being?<br />

The answer seems <strong>to</strong> lie in what we make memories of. Our long-term memory maker,<br />

located in the hippocampus, can s<strong>to</strong>re language, images, sounds, sensations, ideas, and<br />

feelings, but not neural procedures. Biologically, we appear <strong>to</strong> have no use for recalling,<br />

explicitly, activities by the language-processing centers themselves. Our minds, as they<br />

develop, have no given names for the ac<strong>to</strong>rs and the events at such centers. Such<br />

knowledge is not forbidden. It is likely that it is unnecessary <strong>to</strong> and possibly<br />

counterproductive for our survival.<br />

The Cognitive Model<br />

So far, cognitive sciences have been shown <strong>to</strong> make two critical contributions <strong>to</strong> the<br />

analysis of style: it must be analyzed as audi<strong>to</strong>ry, and it emerges from neural procedures<br />

<strong>to</strong> which we cannot attend. Much else can be learned, however, by reading the scientific<br />

literature, both general studies (Kosslyn and Koenig 1992; Eysenck and Keane 1990; and<br />

Lieberman 2000), and analyses of specific areas like memory (Baddeley 1990; Squire<br />

1987). Recent scientific results, on which these books are based, appear as articles in<br />

journals such as Brain, Brain and Language, Cognitive Psychology, The Journal of<br />

Neuroscience, The Journal of Verbal Learning and Verbal Behavior, Memory and<br />

Cognition, Nature, Neuropsychologia, Psychological Review, and Science. These papers<br />

are accessible, but <strong>to</strong> work with them the intelligent reader must be grounded in the<br />

cognitive model of language processing. Because it is changing now, and will continue <strong>to</strong><br />

do so, work in cognitive stylistics will need steady reassessment. However, the method of<br />

cognitive stylistics, which bases text-stylistics on cognitive effects that experimentation<br />

has found <strong>to</strong> illuminate the mind's style, will remain. Here follows a brief summary of the<br />

emerging model. It breaks down in<strong>to</strong> two parts: memory systems, and neural processes.<br />

Scientists now recognize three basic kinds of human memory: (1) short-term memory,<br />

now known as working memory; and long-term associative memory, which falls in<strong>to</strong> two<br />

kinds, (2) implicit or inaccessible, and (3) explicit or accessible. Implicit long-term


memory includes recall of a procedure in the action, and priming. Explicit long-term<br />

memory includes episodic memory and semantic memory.<br />

Working memory offers short-term s<strong>to</strong>rage of a limited amount of language so that it can<br />

be consciously worked on. This form of memory cannot be separated from processing<br />

activities. Alan Baddeley first proposed twenty years ago a very influential model of<br />

working memory split in<strong>to</strong> three parts: a central executive and two subsystems, a visual<br />

area and a phonological or articula<strong>to</strong>ry loop. The executive, which manages tasks in the<br />

two subsystems, has been localized in the dorsolateral prefrontal cortex (Lieberman 2000:<br />

77). All conscious mental work on language gets done in the articula<strong>to</strong>ry loop. It<br />

encompasses many regions in the brain, including the well-known language centers,<br />

Wernicke's and Broca's areas, which handle, respectively, semantic and syntactic<br />

processing. (When damaged, Wernicke's area leads an individual <strong>to</strong> utter well-formed<br />

nonsense, "word salad", and Broca's area <strong>to</strong> utter a form of agrammatism, characterized<br />

by sentence fragments, understandable but unformed.)<br />

Central <strong>to</strong> the mind's conscious fashioning of language is the subsystem Baddeley calls<br />

the articula<strong>to</strong>ry loop. So called because we must recirculate or rehearse a piece of<br />

language in order <strong>to</strong> keep working on it, this loop has a limited capacity, identified by<br />

Miller in 1956 as "seven, plus or minus two." Experiments have for decades confirmed<br />

these limits and show that individuals do not always reach the maximum potential<br />

capacity. The so-called reading-span test asks individuals <strong>to</strong> remember the final words of<br />

a sequence of unrelated sentences. Test results show a range of from 2 <strong>to</strong> 5.5 final words<br />

(Just and Carpenter 1992). As early as 1975 experiments showed that we could s<strong>to</strong>re in<br />

working memory, for recall, only as many words as we could utter aloud in two seconds<br />

(Baddeley et al. 1975). The number of such words declined as the <strong>to</strong>tal number of<br />

syllables increased, in what was termed "the word-length effect." Other experiments elicit<br />

so-called "effects" in subjects that confirm the audi<strong>to</strong>ry nature of working memory of<br />

language, and its severe capacity limits. The "acoustic similarity effect" shows that the<br />

ability of an individual <strong>to</strong> recollect a sequence of unrelated words suffers if they sound<br />

alike: semantic relations, or lack of them, and dissimilarity in sound have no effect. If<br />

working memory used the images of words, as text, the acoustic character of a word<br />

would not affect manipulation in working memory. The "articula<strong>to</strong>ry suppression" effect<br />

also testifies <strong>to</strong> the audi<strong>to</strong>ry nature of language as consciously worked in memory.<br />

Individuals having <strong>to</strong> repeat aloud, continuously, a single sound or term or number (say, a<br />

function word such as "with") cannot rehearse, subvocally, utterances and so put or keep<br />

them in working memory. Audi<strong>to</strong>ry language immediately, unpreventably, enters it.<br />

Other experiments reveal that syntactically challenging sentences, such as those with<br />

clauses embedded within them centrally, reduce language capacity in working memory.<br />

Recently, Philip Lieberman asserts that we maintain "words in verbal working memory<br />

by means of a rehearsal mechanism (silent speech) in which words are internally modeled<br />

by the neural mechanisms that regulate the production of speech or manual signs" (2000:<br />

6).<br />

What can be learned about the mind's style from the articula<strong>to</strong>ry loop? Its capacity<br />

constraints hamstring conscious mental work on making continuous sentences. It is little


wonder that we seldom mentally assemble or attend <strong>to</strong> editing what we are going <strong>to</strong> say<br />

before we utter it. We have perhaps not had artificial s<strong>to</strong>rage devices (e.g., paper,<br />

computers), where it is very easy <strong>to</strong> edit texts, long enough noticeably <strong>to</strong> atrophy our<br />

already limited working memory. However, we have supplemented that memory, for<br />

language manipulation, with external s<strong>to</strong>rage devices. Increasingly, texts will depart from<br />

the mind's constraints as we assemble sentences that greatly exceed the length and<br />

complexity of ones that can be attended <strong>to</strong> by the unassisted mind. This extension has two<br />

clear effects. First, it produces utterances that the human mind cannot consciously<br />

assimilate in<strong>to</strong> working memory for analysis. This, like the McGurk effect, will cause the<br />

mind <strong>to</strong> work-around the problem and perhaps, in doing so, <strong>to</strong> remodel the ingested<br />

utterance in ways that dis<strong>to</strong>rt it. Second, the very experience of <strong>to</strong>tal control over<br />

utterances that artificial s<strong>to</strong>rage devices give makes all the more unbearable our mental<br />

blindness <strong>to</strong> the generation of utterances. No one can use software <strong>to</strong> create sentences,<br />

outside of playful programs like Joseph Weizenbaum's Eliza and the Postmodernism<br />

Genera<strong>to</strong>r of


ippling effect on all memories linked <strong>to</strong> it. The strength of that activation, or its<br />

"weight", may be proportional <strong>to</strong> the number of times that the linkage between those two<br />

memories has previously been activated.<br />

Long-term associative memory does not s<strong>to</strong>re, in one place, complete utterances. Our<br />

mind's procedural memory of how <strong>to</strong> create or recreate an utterance calls on many<br />

different parts of the brain simultaneously, that is, concurrently, and they operate until the<br />

very instant of utterance or subvocal voicing. The mind's thesaurus (concepts), lexicon<br />

(words), and encyclopedia (images of things) consist of "morphologically decomposed<br />

representations" (Koenig et al. 1992). Different systems responsible for phonemes, lexis,<br />

part-of-speech, syntax, letter-shapes, etc., all s<strong>to</strong>red in different locations, work in<br />

parallel. Current research, for example, locates color words in the ventral temporal lobe,<br />

action words in the left temporal gyrus, names of people in the temporal pole, and words<br />

for animals and <strong>to</strong>ols in, respectively, the anterior and posterior inferotemporal area<br />

(Martin et al. 1995: 102; Lieberman 2000: 63, 65; Ojemann 1991). The concept of a<br />

typical noun resembles an address list that itemizes the locations of separately s<strong>to</strong>red<br />

traits or features. Words associated with concepts are kept separately and matched<br />

<strong>to</strong>gether by a mediating area of the brain (Damasio and Damasio 1992) termed a<br />

"convergence zone." The "combina<strong>to</strong>rial arrangements that build features in<strong>to</strong> entities,<br />

and entities in<strong>to</strong> events, i.e. their spatial and temporal coincidences, are recorded in<br />

separate neural ensembles, called convergence zones … [found in] association cortexes,<br />

limbic cortexes, and nonlimbic subcortical nuclei such as the basal ganglia… [where they<br />

form] hierarchical and heterarchical networks" (Damasio et al. 1990: 105). Convergence<br />

zones are keys <strong>to</strong> neural networks.<br />

One type of long-term associative memory is priming. It is a wildcard in the formation of<br />

our memory s<strong>to</strong>re. Sensory experience lays down primes in the mind; we are never aware<br />

of and we do not attend <strong>to</strong> them. Kosslyn and Koenig describe how researchers read<br />

word-pairs <strong>to</strong> patients who had been anaesthetized for surgery. Later, when these patients<br />

were asked for the second, associated member of each word-pair, they replied with the<br />

formerly primed words more than with equally likely associated words (1992: 376). This<br />

effect is often termed repetition priming. A "prior exposure <strong>to</strong> a stimulus facilitates later<br />

processing of that stimulus" (1992: 374). Primes create sometimes unexpected links<br />

between an experience or an idea that we regard as common, and other things that would<br />

not ordinarily be associated with it. Even if everyone shared the same fuzzy definition of<br />

a simple concept, individual experiences impacting on us in the force of primes would<br />

subtly alter that already fuzzy definition. When we search long-term memory, we are<br />

intentionally, consciously, launching a prime-like probe. This type of prime always<br />

places semantic restrictions on retrieval. For instance, priming with the word "present" in<br />

the hope of raising memories related <strong>to</strong> the meaning "gift" will not elicit anything related<br />

<strong>to</strong> the meaning "now." When "primes are unattended, words related <strong>to</strong> either meaning<br />

appear <strong>to</strong> be facilitated" (Posner and Raichle 1997: 148–51). That is, when someone or<br />

some experience triggers long-term memory, what surfaces in equally unattended shape<br />

has all strings attached.


How does the mind use these associative networks <strong>to</strong> provide an utterance? That process<br />

remains elusive. Kosslyn and Koening say that the brain forces output of an utterance<br />

au<strong>to</strong>matically "via a process of constraint satisfaction" (1992: 48, 268–69) in which what<br />

might be termed the best fit survives. This fit is <strong>to</strong> some pragmatic goal that meets the<br />

person's needs, however they may be said <strong>to</strong> exist. Emotions, desires, and purposes<br />

inform those needs. If cognition activates many brain sites in parallel, and if our vaguely<br />

sensed intentions determine what associative networks are selected <strong>to</strong> supply the<br />

semantic gist of what we will say, it is little wonder that we cannot describe how the<br />

Muse works. Working memory – the only mental place where we can consciously attend<br />

<strong>to</strong> language – is not big enough <strong>to</strong> hold this complex cascade of mental events and is also<br />

inherently unsuited <strong>to</strong> doing so. Mental processes are not images or sounds.<br />

So-called experiments "in nature" (that is, patients with brain damage) and devices that<br />

image brain activity have at least identified some essential components of this sentencemaking<br />

process. In the classical model of language brain function, Lichtheim (1885) and<br />

Geschwind (1970) proposed that two regions of the neocortex were responsible: the<br />

posterior Wernicke's area did semantic processing and sent language data <strong>to</strong> the frontal<br />

Broca's area, which clothed it with syntactic form and passed it on <strong>to</strong> the mo<strong>to</strong>r cortex for<br />

speaking. This model relied on ample medical evidence that patients with damage in<br />

Wernicke's area displayed faulty or nonsensical semantics and comprehension, and that<br />

those with damage in Broca's area revealed stacca<strong>to</strong>, fragmented speech with agrammatism.<br />

No one disputes this evidence from brain damage, but localizing function so simply<br />

is now impossible. Neural activity during linguistic processing turns out <strong>to</strong> be massively<br />

parallel and distributed. Language does not follow one path but many. Also, after damage<br />

<strong>to</strong> Broca's and Wernicke's areas, the brain can enlist "alternate neuroana<strong>to</strong>mical<br />

structures" for language use (Lieberman 2000: 5) and recover functionality. Lieberman<br />

and his colleagues have also recently shown that subcortical basal ganglia structures, in<br />

one of the most evolutionally ancient (reptilian) parts of the brain, help regulate language<br />

processing. As far as the brain is concerned, the watchword is indeed in the plural,<br />

location, location, location.<br />

The Mind's Style<br />

Louis Milic some decades ago argued that stylistics must abandon impressionism for<br />

quantitative measures. Since then, researchers who compiled numeric data about style<br />

and made such measures have been puzzled <strong>to</strong> explain how they illuminate literary works<br />

or the authors who made them. Cognitive stylistics asserts that literary texts do not have<br />

style; individual minds do, in uttering. It contends that individual styles are profoundly<br />

affected by the neural constraints surrounding mental language processes. Because minds<br />

can only be indirectly analyzed, stylistics as a discipline must do research at the interface<br />

of cognitive sciences and corpus linguistics. Cognitive psychology and neuroscience tell<br />

us what <strong>to</strong> expect. Corpus linguistics extracts quantitative features of texts that can be<br />

analyzed in terms of how they match what human sciences predict will be found.<br />

How, then, do these sciences characterize language uttered by the disembodied sub-vocal<br />

voice long named the Muse? Keeping in mind that scientists base their models of how the


mind works on experimentally discovered effects, and that cognitive stylistics is at an<br />

early stage, a clear profile of the mind's style is beginning <strong>to</strong> emerge. It is:<br />

• audi<strong>to</strong>ry. Language utterances as s<strong>to</strong>red, processed, and retrieved are phonological, not<br />

sequences of visual symbols, not alphabetic.<br />

• lexico-syntactic. Grammar and vocabulary cannot be separated: that is, syntactic<br />

structure imposed by Broca's area, and semantic fields by Wernicke's area,<br />

simultaneously participate in a unified, parallel, non-sequential process.<br />

• combina<strong>to</strong>ry. The building blocks of any utterance are networks, what Damasio's<br />

convergence zones wire <strong>to</strong>gether. These blocks are not discrete words. The mind knows<br />

word-image-concept-sound combinations, not dictionary headwords and their<br />

explanations.<br />

• built from two-second-long units. These combinations appear in repeated phrases or<br />

unfixed collocations that are not more than 5–9 units in length. This follows if, as<br />

scientists suspect, working memory constraints are associated with a deeper limitation<br />

existing at the level of neural networks. Indeed, one use of computer text analysis is <strong>to</strong><br />

help determine the size and complexity of long-term-memory networks, how many things<br />

can converge on a convergence zone.<br />

• biased <strong>to</strong> parataxis. Working memory is slowed when it must manipulate centrally<br />

embedded clauses. The mind, working <strong>to</strong>wards a best fit in generating a sentence, may<br />

also well opt for simpler syntactical structures, unnested syntactic constructions, that is,<br />

paratactic sentences that take the form of a list of clauses linked by conjunctions.<br />

• semantically indeterminate. No conventional thesaurus, encyclopedia, or dictionary can<br />

adequately document the individual mind's convergence zones, what may underlie<br />

semantic fields and associative clusters, simply because every individual's long-term<br />

s<strong>to</strong>re is affected by fortui<strong>to</strong>us associations, primes. The traits associated with concepts<br />

and words alter subtly as the weights that measure their binding strength change over<br />

time, affected by everything we directly experience through the senses. This partly<br />

explains the language effect known as lexical indeterminacy (Pilch 1988). Many words<br />

cannot be defined precisely enough <strong>to</strong> avoid misunderstandings. Individuals use words<br />

differently and only partially overlap with others.<br />

• time-sensitive. As memory changes (or deteriorates), so do the characteristic traits of its<br />

utterances. Style is tied always <strong>to</strong> the health and age of the author's brain.<br />

Other features of the mind's default style may be known and certainly will be discovered.<br />

These inbuilt traits are enough <strong>to</strong> initiate research.<br />

Tools


Text-analysis algorithms and <strong>to</strong>ols are available <strong>to</strong>day <strong>to</strong> undertake rudimentary<br />

computer-based cognitive Stylistics research.<br />

Analysis is complicated by the need for enriched texts. We cannot use orthographically<br />

spelled text alone because the brain uniformly recognizes, s<strong>to</strong>res, retrieves, and operates<br />

in working memory on language as audi<strong>to</strong>ry data, "silent speech." Each word must be<br />

available in orthographic and phonemic forms, at least, and optimally be segmented in<strong>to</strong><br />

syllables, and tagged with morphological information. Organizations such as the Speech<br />

Assessment Methods Phonetic Alphabet (SAMPA,<br />

) offer rigorous modern British and<br />

American alphabets. SIL International () has useful databases and<br />

software for this enrichment. Some au<strong>to</strong>matic phonetic conversion <strong>to</strong>ols recommended by<br />

the Simplified Spelling Society () use less rigorous<br />

alphabets. Phonetic conversion in general can only be done well by researchers who<br />

understand the sounds of a language. In early periods, before printing, spelling may<br />

represent word-sounds adequately for analysis.<br />

Researchers have been able <strong>to</strong> locate repeated words and fixed phrases since the<br />

invention of the KWIC concordancer in the late 1950s. Software by computational<br />

linguistics exists <strong>to</strong> generate the repeating fixed phrases of texts, termed "n-grams"<br />

(Fletcher 2002). Techniques for generating unfixed repeating phrases, that is,<br />

collocations, has been slower <strong>to</strong> develop. Collgen, a free TACT program I developed in<br />

the early 1990s, works only with word-pairs, word-word collocations, in small corpora.<br />

Xtract (1993), by Smadja and McKeown, is unavailable, but Word Sketch by Kilgarriff<br />

and Tugwell looks promising. Every researcher faces the not inconsiderable task of<br />

defining collocation itself and selecting a statistical measure that ranks repeating<br />

collocations by their significance. Collins Wordbanks Online<br />

() offers interactive searching for wordcombinations<br />

extracted from a 56-million-word Bank of English with two significance<br />

scores: "mutual information" and t-test. Berry-Rogghe (1973), Choueka et al. (1983), and<br />

Church and Hands (1990) offer various measures. Budanitsky and Hirst (<strong>2001</strong>) evaluate<br />

them. Significance rating conceivably recovers information about the strength of<br />

associativity among items s<strong>to</strong>red in long-term memory. By processing concordance data,<br />

representations of such clusters can be built.<br />

I am not aware of any software that generates repeating clusters of words, fixed phrases,<br />

and unfixed phrases (collocations), say, <strong>to</strong> a limit of seven units, plus or minus two, but<br />

computational algorithms written for large industrial applications in data mining might<br />

permit repurposing.<br />

Applications<br />

Some trial literary analyses tested whether traits in works by two English poets, Geoffrey<br />

Chaucer and Shakespeare, could be accounted for within the standard cognitive model of<br />

language processing. My studies used texts with normalized orthography, untranscribed<br />

phonetically because Middle and Early Modern English pronunciation is not yet


sufficiently well unders<strong>to</strong>od. The software, Collgen, was limited <strong>to</strong> repeated fixed phrases<br />

and node-collocate pairs. Unfixed groups of three and more collocates were uncollected.<br />

These limitations aside, published results of the analyses tended <strong>to</strong> affirm that the styles<br />

of both authors might be cognitively based, and partly recoverable.<br />

I studied two passages from Shakespeare's works, Hamlet, III. 1 (the so-called "nunnery<br />

scene"), and Troilus and Cressida, 1.3.1–29 (Agamemnon's first speech), and two parts<br />

of Chaucer's Canterbury Tales, the General Prologue and the Manciple's prologue and<br />

tale, both in the context of the complete Canterbury Tales. The principal repeated<br />

vocabulary unit of both authors was the word-combination. In The Canterbury Tales,<br />

Chaucer used 12,000 word-forms but 22,000 repeating fixed phrases. Over two periods,<br />

1589–94 and 1597–1604, Shakespeare's different fixed phrases at least doubled his word<br />

types. The vocabulary of both poets consisted, not of single words, but of little networks,<br />

a fact consistent with associative long-term memory. The sizes of these networks were<br />

well within what working memory could accommodate. The 464 phrasal repetends<br />

appearing in both Chaucer's General Prologue and the rest of his Canterbury Tales<br />

averaged 2.45 words. They fell in<strong>to</strong> 177 networks. Repeating fixed phrases in<br />

Shakespeare's texts in both periods averaged 2.5 words. Chaucer's largest repeating<br />

combination in the General Prologue (853 lines) had nine words. Shakespeare's largest in<br />

Hamlet III.l, under 200 lines long, had five words. A second Shakespeare analysis, of<br />

Agamemnon's speech in Troilus and Cressida. 1.3.1–29, found 107 phrasal repetends<br />

(repeating elsewhere in Shakespeare's works) in a passage that has only 159 different<br />

word-forms. Most combinations are two words in length, and the maximum has four. It is<br />

possible that the constraints of working memory affected the quantitative profile of the<br />

verbal networks employed by both men.<br />

For both authors, I used text-graphs of these phrasal repetends <strong>to</strong> depict how they<br />

intersected. In the Chaucer study, three overlapping "say"-"tell"-"speak" graphs drew<br />

attention <strong>to</strong> Chaucer's unnoticed marking of the three verbs: they distinguished "between<br />

speaking (words), telling tales, and saying truths or sooths" (Lancashire 1992a: 349).<br />

Gary Shawver's doc<strong>to</strong>ral thesis at the University of Toron<strong>to</strong>, "A Chaucerian Narra<strong>to</strong>logy:<br />

'S<strong>to</strong>rie' and 'Tale' in Chaucer's Narrative Practice" (1999), developed a finely detailed<br />

theory of Chaucer's narrative by taking this passing observation seriously. Two<br />

intersecting phrasal repetend graphs on the various word-forms for "true" and "love" in<br />

Hamlet also revealed a small network including the term "prove."<br />

Other findings illustrate the variety of useful applications of cognitive Stylistics.<br />

Chaucer's phrasal repetends in the General Prologue that repeated somewhere in the rest<br />

of the tales were graphed against those tales. After taking in<strong>to</strong> account their different<br />

sizes, a distribution showed that the General Prologue shared more repetends with a quite<br />

unrelated tale, by the Manciple, found always just preceding the last tale, by the Parson.<br />

One possible interpretation of these results is that Chaucer wrote the two works in the<br />

same year. The 107 phrasal repetends in Agamemnon's speech in Shakespeare's Troilus<br />

and Cressida served a different purpose. They explained why Shakespeare used an odd<br />

sequence of images in lines that critics thought ill-conceived. Passages from earlier<br />

poems and plays here documented associative linkages that are private.


Conclusions<br />

Annotating texts for their repeating combinations conceivably finds traces of an author's<br />

long-term associative memory that make up his idiolect. Combina<strong>to</strong>rial analysis also<br />

assists in close reading. It puts in<strong>to</strong> sharp relief the unrepeating words in those passages<br />

that may mark recent mental acquisitions. (Very long phrasal repetends enter the author's<br />

text manually, copied from other sources.) Analysis exposes some repetitions, notably<br />

syntactic strings, that may show that an author primes his mind by looking at the<br />

unfolding composition on page or screen and so moving the language in<strong>to</strong> the artificial<br />

memory. The image of just-composed text can lead <strong>to</strong> re-use of the grammatical<br />

structures, that is, function words. Entering working memory as an image, then converted<br />

subvocally for the articula<strong>to</strong>ry loop, writing stimulates variations on itself. Yet cognitive<br />

Stylistics is a young field. It must undertake hundreds of case studies <strong>to</strong> buttress these<br />

hypotheses. There are three special challenges.<br />

We still have no firm model of repeating word-combinations. Fixed-phrase repetends (ngrams)<br />

and collocations (order-free collocate groups) are known <strong>to</strong> vary according <strong>to</strong> the<br />

span or the window, measured in words, within which they are measured. The tighter the<br />

window (e.g., five words), the smaller the length of the repeating repetends. If a window<br />

becomes as large as a page, it can contain two or more instances of the same fixed phrase.<br />

In that case, should the repetend be redefined as a cluster of that repeating phrase? We do<br />

not know how large such repeating clusters are allowed <strong>to</strong> get. If we increase the size of<br />

the window <strong>to</strong> the complete text, we then have only one complex cluster that turns out<br />

not <strong>to</strong> be a repetend after all because it never repeats. The window itself must be<br />

theorized. It could be set at the length of working memory, but mental associational<br />

networks may be larger. If eight words on either side of one member of a collocate pair<br />

were set as the span (i.e., the maximum capacity of the articula<strong>to</strong>ry loop in working<br />

memory), what would we find if, after collecting all repeating collocates for a node word,<br />

we then treated those collocates themselves as nodes and went on <strong>to</strong> collect their<br />

collocates, and so on? At what point does the original node no longer reasonably<br />

associate with a distant collocate of one of its node-collocates?<br />

We now have a reasonable vocabulary for describing the repeating fixed phrase and the<br />

cluster of a node and its collocates. Word-combinations that have reached their greatest<br />

phrasal length or collocate number are termed "maximals" (Altenberg and Eeg-Olofsson<br />

1990: 8). Shorter, more frequent substrings or sub-collocate groups, which appear <strong>to</strong><br />

function as kernels or attrac<strong>to</strong>rs for other units, are called "subordinates"; and substrings<br />

of those subordinates, substrings that do not occur more frequently than the subordinate<br />

in which they appear, are termed "fragments" (Lancashire, 1992b). Kjellmer (1991: 112)<br />

uses the phrase "right-and-left predictive" for nodes that accumulate collocates <strong>to</strong> the<br />

right or the left. Sinclair (1991: 121) characterizes the frequency behavior of a node <strong>to</strong> be<br />

"upward" if its collocates occur more frequently than it, and "downward" if less<br />

frequently. (For example, nouns collocate upward with function words, and function<br />

words collocate downward with nouns.) Repetend features such as these affect our sense<br />

of the strength of association – often termed semantic distance – between parts of a fixed<br />

phrase or node-collocates cluster. So far, we calculate strength on the basis of frequencies


(expected and actual) and mutual distance (measured in words), but size and other traits<br />

must be taken in<strong>to</strong> account. For example, consider two node-collocate pairs, both<br />

separated by five words, and both sharing the same frequency profiles. (That is, both<br />

pairs share an actual frequency of co-occurrence that exceeds the expected frequency by<br />

the same amount.) Do they have different strengths of association if one of the pairs<br />

consists of two single words, and the other of two four-word fixed phrases? Or consider a<br />

repetend consisting of a single open-class word (e.g., noun, adjective, non-auxiliary verb,<br />

etc.), grammatically marked by a function word (e.g., article, preposition, etc.). Can we<br />

reasonably compare the strength of association there with that governing two open-class<br />

words that collocate freely and at some distance from one another? (In other words, can<br />

we treat grammatical and lexical collocations identically?) And how should we locate,<br />

name, and characterize words that have no significant repeating collocates and partner in<br />

no repeated phrases? These are words with a null "constructional tendency" (Kjellmer,<br />

1994: ix). The more we know about combina<strong>to</strong>ry repetends, the more we can bring <strong>to</strong><br />

understanding the mind's style and even maybe our long-term memory. Texts follow<br />

from an individual's brain functions, but experimental sciences do not know how <strong>to</strong> use<br />

texts as evidence. Cognitive stylistics can assist.<br />

The most common application of stylistics is authorship attribution. It locates markertraits<br />

in an unattributed work that match selected traits of only one of the candidates. This<br />

method systematically neglects the overall stylistic profile of any one author. We need<br />

many more text-based studies of how a single author uses repeating word-combinations<br />

over time. Humanities researchers must also go beyond texts for their evidence. We can<br />

enhance our knowledge about composition, and its relationship <strong>to</strong> brain function, by<br />

undertaking experimental studies common in psychology and neuroscience. If the<br />

<strong>humanities</strong> expect the sciences <strong>to</strong> attend <strong>to</strong> text-based analysis, they must begin <strong>to</strong><br />

undertake controlled experiments <strong>to</strong> analyze the linguistic behavior of living writers as<br />

they compose. Interviews and tests before, during, and after periods of composition can<br />

be combined with detailed capture of the author's keystrokes as the creative process takes<br />

place. The necessary research <strong>to</strong>ols for these experiments are now widely applied <strong>to</strong><br />

human-user interaction research.<br />

In cognitive Stylistics, the <strong>humanities</strong> can take a leading role in profoundly important<br />

research straddling the medical sciences, corpus and computational linguistics, literary<br />

studies, and a community of living authors.<br />

References for Further Reading<br />

Altenberg, B. and M. Eeg-Olofsson (1990). Phraseology in Spoken English: Presentation<br />

of a Project. In J. AartsW. Meijs (eds.), Theory and Practice in Corpus Linguistics (pp.<br />

1–26). Amsterdam: Rodopi.<br />

Baddeley, A. (1986). Working Memory. Oxford: Clarendon Press.<br />

Baddeley, A. (1990). Human Memory: Theory and Practice. Hove and London: Erlbaum.


Baddeley, A. (1992). Is Working Memory Working? The Fifteenth Bartlett Lecture.<br />

Quarterly Journal of Experimental Psychology 44A, 1: 1–31.<br />

Baddeley, A. (1993). Working Memory. Science 255: 556–9.<br />

Baddeley, A. (1998). Recent Developments in Working Memory. Current Opinion in<br />

Neurobiology 8, 2(April): 234–8.<br />

Baddeley, A., S. E. Gathercole, and C. Papagno (1998). The Phonological Loop as a<br />

Language Learning Device. Psychological Review. 158–73.<br />

Baddeley, A., et al. (1975). Word Length and the Structure of Short-term Memory.<br />

Journal of Verbal Learning and Verbal Behavior 14, 6: 575–89.<br />

Berry-Rogghe, G. L. M. (1973). The Computation of Collocations and Their Relevance<br />

in Lexical Studies. In A. J. Aitken, R. W. Bailey, and N. Hamil<strong>to</strong>n-Smith (eds.), The<br />

Computer and Literary Studies (pp. 103–12). Edinburgh: Edinburgh University Press.<br />

Budanitsky, Alexander and Graeme Hirst (<strong>2001</strong>). Semantic Distance in WordNet: An<br />

Experimental, Application-oriented Evaluation of Five Measures. Workshop on WordNet<br />

and Other Lexical Resources, Second meeting of the North American Chapter of the<br />

Association for Computational Linguistics, Pittsburgh.<br />

Caplan, D. and G. S. Waters (1990). Short-term Memory and Language Comprehension:<br />

a Cortical Review of the Neuropsychological Literature. In G. Villar and T. Shallice<br />

(eds.), Neuropsychological Impairments of S.T.M. Cambridge: Cambridge University<br />

Press.<br />

Chang, T. M. (1986). Semantic Memory: Facts and Models. Psychological Bulletin 99, 2:<br />

199–220.<br />

Choueka, Y., T. Klein, and E. Neuwitz (1983). Au<strong>to</strong>matical Retrieval of Frequent<br />

Idiomatic and Collocational Expressions in a Large Corpus. Journal for Literary and<br />

Linguistic Computing 4: 34–8.<br />

Church, Kenneth Ward and Patrick Hands (1990). Word Association Norms, Mutual<br />

Information, and Lexicography. Computational Linguistics 16, 1: 22–220.<br />

Collins, A. M. and E. F Loftus (1975). A Spreading Activation Theory of Semantic<br />

Processing. Psychological Review 82: 407–28.<br />

Courtney, S. M., L. Petit, M. M. Jose, L. G. Ungerleider, and J. V. Haxby (1998). An<br />

Area Specialized for Spatial Working Memory in Human Frontal Cortex. Science 279:<br />

1347–51.


Damasio, An<strong>to</strong>nio R. (1995). Descartes' Error: Emotion, Reason, and the Human Brain.<br />

New York: Avon.<br />

Damasio, An<strong>to</strong>nio R. and Hanna Damasio (1992). Brain and Language. Scientific<br />

American 267, 3: 88–95.<br />

Damasio, An<strong>to</strong>nio R., Daniel Tranel, and Hanna Damasio (1990). Face Agnosia and the<br />

Neural Substrates of Memory. Annual Review of Nearoscience 13: 89–109.<br />

D'Esposi<strong>to</strong>, M., J. A. Detre, D. C. Alsop, R. K. Shin, S. Atlas, and M. Grossman (1995).<br />

The Neural Basis of the Central Executive System of Working Memory. Science 378:<br />

279–81.<br />

Eysenck, Michael W. and Mark, T. Keane (1990). Cognitive Psychology: A Student's<br />

Handbook. Hove, East Sussex: Erlbaum.<br />

Fletcher, William H. (2002). kfNgram Information & Help. URL: At<br />

http://www.chesapeake.net/~fletcher/kfNgramHelp.html.<br />

Gathercole, S. E. and A. D. Baddeley (1993). Working Memory and Language. Hillside,<br />

PA: Lawrence Erlbaum.<br />

Geschwind, Norman (1970). The Organization of Language and the Brain. Science 170:<br />

940–4.<br />

Geschwind, Norman (1979). Specializations of the Human Brain. In The Brain (pp. 108–<br />

17). A Scientific American Book. San Francisco: W. H. Freeman.<br />

Ghiselin, Brewster, (ed.) (1952). The Creative Process: A Symposium. New York:<br />

Men<strong>to</strong>r.<br />

Grasby, P. M., C. D. Frith, K. J. Fris<strong>to</strong>n, C. Bench, R. S. J. Frackowiak, and R. J. Dolan<br />

(1993). Functional Mapping of Brain Areas Implicated in Audi<strong>to</strong>ry–Verbal Memory<br />

Function. Brain 116: 1–20.<br />

Hagoort, Peter (1993). Impairments of Lexical–Semantic Processing in Aphasia:<br />

Evidence from the Processing of Lexical Ambiguities. Brain and Language 45: 189–232.<br />

Hoey, M. (1991). Patterns of Lexis in Text. Oxford: Oxford University Press.<br />

Jacoby, Mario (1971). The Muse as a Symbol of Literary Creativity. In Joseph P. Strelka<br />

(ed.), Anagogic Qualities of Literature (pp. 36–50). University Park and London:<br />

Pennsylvania State University Press.<br />

Just, Marcel A. and Patricia A. Carpenter (1992). A Capacity Theory of Comprehension:<br />

Individual Differences in Working Memory. Psychological Review 99: 122–49.


Kertesz, Andrew (1983). Localization of Lesions in Wernicke's Aphasia. In Andrew<br />

Kertesz (ed.), Localization in Neuropsychology (pp. 208–30). New York: Academic<br />

Press.<br />

Kilgarriff, Adam and David Tugwell (<strong>2001</strong>). WORD SKETCH: Extraction and Display of<br />

Significant Collocations for Lexicography. COLLOCATION: Computational Extraction,<br />

Analysis and Exploitation (pp. 32–8). 39th ACL and 10th EACL. Toulouse (July).<br />

Kjellmer, G. (1991). A Mint of Phrases. In K. Aijmer and B. Altenberg (eds.), Corpus<br />

Linguistics: Studies in Honour of Jan Svartvik (pp. 111–27). London: Longman.<br />

Kjellmer, Goran (1994). A Dictionary of English Collocations. Oxford: Clarendon Press.<br />

Koenig, Olivier, Corinne Wetzel, and Alfonso Caramazza (1992). Evidence for Different<br />

Types of Lexical Representations in the Cerebral Hemispheres. Cognitive<br />

Neuropsychology 9, 1: 33–220.<br />

Kosslyn, Stephen M. and Olivier Koenig (1992). Wet Mind: The New Cognitive<br />

Neuroscience. New York: Free Press.<br />

Lancashire, Ian (1992a). Chaucer's Repetends from The General Prologue of the<br />

Canterbury Tales. In R. A. Taylor, James F. Burke, Patricia J. Eberle, Ian Lancashire,<br />

and Brian Merrilees (eds.), The Centre and its Compass: Studies in Medieval Literature<br />

in Honor of Professor John Leyerle (pp. 315–65). Kalamazoo, MI: Western Michigan<br />

University Press.<br />

Lancashire, Ian (1992b). Phrasal Repetends in Literary Stylistics: Shakespeare's Hamlet<br />

III.l. Research in Humanities Computing 4 (pp. 34–68). Selected Papers from the<br />

ALLC/ACH Conference, Christ Church, Oxford, April, 1992. Oxford: Clarendon Press,<br />

1996.<br />

Lancashire, Ian (1993a). Chaucer's Phrasal Repetends and The Manciple's Prologue and<br />

Tale. In Ian Lancashire (ed.), Computer-Based Chaucer Studies (pp. 99–122). CCHWP 3.<br />

Toron<strong>to</strong>: Centre for Computing in the Humanities.<br />

Lancashire, Ian (1993b). Computer-assisted Critical Analysis: A Case Study of Margaret<br />

Atwood's The Handmaid's Tale. In George Landow and Paul Delany (eds.), The Digital<br />

Word (pp. 293–318). Cambridge, MA: MIT Press.<br />

Lancashire, Ian (1993c). Uttering and Editing: Computational Text Analysis and<br />

Cognitive Studies in Authorship. Texte: Revue de Critique et de Theorie Litteraire 13/14:<br />

173–218.<br />

Lancashire, Ian (1999). Probing Shakespeare's Idiolect in Troilus and Cressida 1.3.1–29.<br />

University of Toron<strong>to</strong> Quarterly 68, 3: 728–67.


Lancashire, Ian, in collaboration with John Bradley, Willard McCarty, Michael Stairs,<br />

and T. R. Wooldridge (1996). Using TACT with Electronic Texts: A Guide <strong>to</strong> Text-<br />

Analysis Computing Tools, Version 2.1 for MS-DOS and PC DOS. New York: Modern<br />

Language Association of America.<br />

Lavric, A., S. Forstmeier, and G. Rippon (2000). Differences in Working Memory<br />

Involvement in Analytical and Creative Tasks: An ERP Study. Neuroreport 11, 8: 1613–<br />

220.<br />

Levine, David N. and Eric Sweet (1983). Localization of Lesions in Broca's Mo<strong>to</strong>r<br />

Aphasia. In Andrew Kertesz (ed.), Localization in Neuropsychology (pp. 185–208). New<br />

York: Academic Press.<br />

Lichtheim, L. (1885). On Aphasia. Brain 7: 433–84.<br />

Lieberman, Philip (2000). Human Language and Our Reptilian Brain: The Subcortical<br />

Bases of Speech, Syntax, and Thought. Cambridge, MA: Harvard University Press.<br />

Longoni, A. M., J. T. E. Richardson, and A. Aiello (1993). Articula<strong>to</strong>ry Rehearsal and<br />

Phonological S<strong>to</strong>rage in Working Memory. Memory and Cognition 21: 11–22.<br />

MacDonald, Maryellen, Marcel A. Just, and Patricia A. Carpenter (1992). Working<br />

Memory Constraints on the Processing of Syntactic Ambiguity. Cognitive Psychology 24:<br />

56–98.<br />

Martin, A., J. V. Haxby, F. M. Lalonde, C. L. Wiggs, and L. G. Ungerleider (1995).<br />

Discrete Cortical Regions Associated with Knowledge of Color and Knowledge of Action.<br />

Science 270: 102–5.<br />

Martindale, Colin (1990). The Clockwork Muse: The Predictability of Artistic Change.<br />

New York: Basic Books.<br />

Miles, C., D. M. Jones, and C. A. Madden (1991). Locus of the Irrelevant Speech Effect<br />

in Short-term Memory. Journal of Experimental Psychology: Learning, Memory, and<br />

Cognition 17: 578–84.<br />

Milic, Louis T. (1971). Rhe<strong>to</strong>rical Choice and Stylistic Option: The Conscious and<br />

Unconscious Poles. In Seymour Chatman (ed. and tr.), Literary Style: A Symposium.<br />

London and New York: Oxford University Press.<br />

Miller, G. A. (1956). The Magical Number Seven, Plus or Minus Two: Some Limits on<br />

our Capacity for Processing Information. Psychological Review 63: 89–97.<br />

Miyake, Akira and Priti Shah (1998). Models of Working Memory: Mechanisms of Active<br />

Maintenance and Executive Control. New York: Cambridge University Press.


Ojemann, George A. (1991). Cortical Organization of Language. Journal of<br />

Neuroscience 11, 8 (August): 2281–7.<br />

Ojemann, George A., F. Ojemann, E. Lattich, and M. Berger (1989). Cortical Language<br />

Localization in Left Dominant Hemisphere: an Electrical Stimulation Mapping<br />

Investigation in 117 Patients. Journal of Neurosurgery 71: 316–26.<br />

Paulesu, E., C. Firth, and R. Frackowiak (1993). The Neural Correlates of the Verbal<br />

Component of Working Memory. Nature 362: 342–45.<br />

Perani, D., S. Bressi, S. F. Cappa, G. Vallar, M. Alberoni, F. Grassi, C. Caltagirone, L.<br />

Cipolotti, M. Franceschi, G. L. Lenzi, and F. Fazio (1993). Evidence of Multiple Memory<br />

Systems in the Human Brain. Brain 116: 903–19.<br />

Petersen, S. E. and J. A. Fiez (1993). The Processing of Single Words Studied with<br />

Positron Emission Tomography. Annual Review of Neuroscience 16: 509–30.<br />

Pilch, Herbert (1988). Lexical Indeterminacy. In E. G. Stanley and T. F. Hoad (eds.),<br />

Words for Robert Burchfield's Sixty-fifth Anniversary (pp. 133–41). Cambridge: D. S.<br />

Brewer.<br />

Plimp<strong>to</strong>n, George, (ed.) (1988). Writers at Work: The Paris Review Interviews. London:<br />

Penguin.<br />

Plimp<strong>to</strong>n, George, (ed.) (1989). The Writer's Chapbook: A Compendium of Fact,<br />

Opinion, Wit, and Advice from the 20th Century's Preeminent Writers, rev. edn. London:<br />

Penguin Books.<br />

Posner, Michael I. and Marcus E. Raichle (1997). Images of Mind (pp. 117–18, 123–9).<br />

New York: Scientific American Library.<br />

Schacter, Daniel L., C.-Y Peter Chiu, and Kevin N. Ochsner (1993). Implicit Memory: A<br />

Selective Review. Annual Review of Neuroscience 16, 159–82.<br />

Schweickert, R. and B. Boriff (1986). Short-term Memory Capacity: Magic Number or<br />

Magic Spell? Journal of Experimental Psychology: Learning, Memory, and Cognition 12,<br />

3: 419–25.<br />

Scott, Mike and Geoff Thompson, (eds.) (<strong>2001</strong>). Patterns of Text in Honour of Michael<br />

Hoey. Amsterdam: John Benjamins.<br />

Shallice, T. and B. Butterworth (1977). Short-term Memory and Spontaneous Speech.<br />

Neuropsychologia 15:729–35.


Shawver, Gary (1999). A Chaucerian Narra<strong>to</strong>logy: "S<strong>to</strong>rie" and "Tale" in Chaucer's<br />

Narrative Practice. PhD thesis, University of Toron<strong>to</strong>. Cf.<br />

homepages.nyu.edu/~gs74/nonComIndex.html#4.<br />

Sheridan, Jenny and Glyn W. Humphreys (1993). A Verbal-semantic Category-specific<br />

Recognition Impairment. Cognitive Neuropsychology 10, 2: 143–220.<br />

Shimamura, Arthur P., Felicia B. Gershberg, Paul J. Jurica, Jennifer A. Mangels, and<br />

Robert T. Knight (1992). Intact Implicit Memory in Patients with Frontal Lobe Lesions.<br />

Neuropsychologia 30, 10: 931–220.<br />

Sinclair, John (1991). Corpus Concordance, Collocation. Oxford: Oxford University<br />

Press.<br />

Smadja, Frank (1994). Retrieving Collocations from Text: Xtract. Computational<br />

Linguistics 19, 1: 143–77.<br />

Squire, Larry R. (1987). Memory and Brain. New York: Oxford University Press.<br />

Tulving, E. (1983). Elements of Episodic Memory. Oxford: Oxford University Press.<br />

Turner, Mark (1991). Reading Minds: The Study of English in the Age of Cognitive<br />

Science. Prince<strong>to</strong>n, NJ: Prince<strong>to</strong>n University Press.<br />

Vallar, G., A. M. D. Betta, and M. C. Silveri (1997). The Phonological Short-term S<strong>to</strong>rerehearsal<br />

System. Neuropsychologia 35: 795–812.<br />

Wachtel, Eleanor, (ed.) (1993). Writers and Company. Toron<strong>to</strong>: Alfred A. Knopf Canada.<br />

Winter, Nina (1978). Interview with the Muse: Remarkable Women Speak on Creativity<br />

and Power. Berkeley, CA: Moon.<br />

Zola-Morgan, S. and L. R. Squire (1993). Neuroana<strong>to</strong>my of Memory. Annual Review of<br />

Neuroscience 16: 547–63.<br />

28.<br />

Multivariant Narratives<br />

Marie-Laure Ryan<br />

Media theorists divide the his<strong>to</strong>ry of writing in<strong>to</strong> four periods delimited by technological<br />

innovations: the oral age; the chirographic age (manuscript writing); the print age; and<br />

the <strong>digital</strong> age. The material support of language passed from unique <strong>to</strong> freely copiable;<br />

from restricted <strong>to</strong> a live audience <strong>to</strong> widely distributed; and from evanescent <strong>to</strong> durable,


only <strong>to</strong> return <strong>to</strong> a strange combination of evanescence and durability: though <strong>digital</strong> texts<br />

can be s<strong>to</strong>red in a wide variety of memory devices, computer systems become rapidly<br />

obsolete and their archives unreadable.<br />

We know that the invention of writing and of the printing press had major consequences<br />

for textuality and narrativity. In oral cultures, as Walter Ong (1982) has shown, narrative<br />

was used as a mnemonic device for the transmission of knowledge; its memorization was<br />

facilitated by prosodic features – meter, rhyme, alliteration – as well as by fixed formulae<br />

and standardized images; and the limitations of memory were compensated by a<br />

relatively free episodic structure which allowed, within limits, permutation of its units.<br />

The invention of writing made it possible <strong>to</strong> shape the flat line of epic plots in<strong>to</strong> the curve<br />

of dramatic form, a much more condensed narrative structure that allows a tighter<br />

management of emotional responses. Writing also froze the free order of plots in<strong>to</strong> a<br />

fixed sequence. The printing press increased the length of narrative, revived the episodic<br />

pattern of epic poetry <strong>to</strong> fill the generous frame of the book, rendered mnemonic devices<br />

obsolete, and led <strong>to</strong> the birth of the novel, a relatively unconstrained narrative form that<br />

<strong>to</strong>ok plot <strong>to</strong> unprecedented levels of complexity: framing, embedding, branching,<br />

digressions, disruptions of temporal sequence, and multiple plot lines. Thanks <strong>to</strong> the<br />

spatiality of the page, images found their way in<strong>to</strong> texts, and visual presentation was<br />

eventually recognized and exploited as an expressive device. After these three stages, is<br />

there anything left for <strong>digital</strong> media <strong>to</strong> develop in the narrative terri<strong>to</strong>ry?<br />

This question would not be worth asking if the computer merely served as a medium of<br />

transmission for print texts (as it does when a digitized version of a Stephen King novel is<br />

sold online) or as an instrument of production for texts <strong>to</strong> be experienced in print: most<br />

novels, after all, come out of word processors. A truly <strong>digital</strong> text, or narrative, is one that<br />

cannot be transferred in<strong>to</strong> the print medium without significant loss. It depends on the<br />

computer as a sustaining environment, and it uses the screen (or any other display device)<br />

as a stage for performance.<br />

What, then, are the properties of <strong>digital</strong> media, and, by extension, of <strong>digital</strong> texts, that<br />

bear upon the development of narrative? Several new media theorists (Murray 1997;<br />

Manovich <strong>2001</strong>) have offered their own lists of distinctive features or essential properties<br />

of <strong>digital</strong> systems. The list proposed below is a distillation of the features I regard as the<br />

most relevant <strong>to</strong> the issues of textuality and narrativity:<br />

• Algorithm-driven operation. Computers are machines that can run a variety of<br />

programs, through which they can perform a variety of tasks. The behavior of <strong>digital</strong><br />

objects, such as texts, images, and sound, is therefore regulated by an invisible code, the<br />

machine-language instructions of the supporting software.<br />

• Reactive and interactive nature. This property is a direct consequence of the preceding<br />

one. Computer code is based on conditional statements (if… then) that execute different<br />

instructions, depending on the state of the system or on external input. I call a system<br />

reactive when it responds <strong>to</strong> changes in the environment or <strong>to</strong> non- intentional user<br />

actions; it is interactive when the input originates in a deliberate user action. (Interactivity


does not necessarily mean, however, that the system will act in the way intended by the<br />

user.)<br />

• Performantial aspect. Another consequence of the first property. A <strong>digital</strong> text is like a<br />

music score or theater script: its written inscription is meant <strong>to</strong> be executed, either by the<br />

underlying code alone, or through a feedback loop that leads from the user <strong>to</strong> the<br />

underlying code <strong>to</strong> the display, and back <strong>to</strong> the user. Digital texts thus present the same<br />

contrast as the classic performing arts between the invariability of the script and the<br />

variability of its execution.<br />

• Multiple sensory and semiotic channels, or what we may call multimedia capabilities, if<br />

we are not afraid of the apparent paradox of talking about multimedia media. Digital<br />

environments can combine text, sound, still pictures, and animations.<br />

• Networking capabilities. Digital media connect machines and people across space and<br />

bring them <strong>to</strong>gether in virtual environments. This opens the possibility of multi-user<br />

systems and live (real-time) as well as delayed (asynchronous) communication.<br />

• Volatile signs. Computer memory is made of bits whose value can switch back and<br />

forth between positive and negative. The value of these bits determines the display. This<br />

means that unlike books or paintings, <strong>digital</strong> texts can be refreshed and rewritten, without<br />

having <strong>to</strong> throw away the material support. This property explains the unparalleled<br />

fluidity of <strong>digital</strong> images.<br />

• Modularity. Because the computer makes it so easy <strong>to</strong> reproduce data, <strong>digital</strong> works<br />

tend <strong>to</strong> be composed of many au<strong>to</strong>nomous objects. These objects can be used in many<br />

different contexts and combinations, and undergo various transformations, during the run<br />

of the work.<br />

The output of a hidden program, <strong>digital</strong> narrative is shaped not only by the general<br />

properties of its "material" medium (i.e., silicon chips), but also by the specific<br />

affordances of the system through which it is created and executed. An Infocom<br />

interactive fiction of the 1980s or a S<strong>to</strong>ryspace hypertext narrative of the early 1990s<br />

differs significantly from a Flash game or Direc<strong>to</strong>r "movie" produced in this new century.<br />

The code of authoring programs is a second-order means of expression, and the various<br />

software supports should, therefore, be considered the submedia of <strong>digital</strong>ity – just as<br />

clay tablets, papyrus scrolls, and codices are the submedia of manuscript writing.<br />

To complete the preliminaries of my discussion, let me briefly define what I mean by<br />

narrative. I endorse a medium-free, semantically based definition, according <strong>to</strong> which<br />

narrative is a type of meaning, or mental image generated in response <strong>to</strong> certain stimuli.<br />

A narrative text is an artifact designed <strong>to</strong> bring this meaning <strong>to</strong> mind. But the cognitive<br />

construct specific <strong>to</strong> narrativity can also be formed in response <strong>to</strong> stimuli not expressly<br />

designed for this purpose, for instance as an interpretation of life itself. This does not<br />

make life in<strong>to</strong> "a" narrative, but it means that life may possess narrative potential – what<br />

we may call "narrativity." A narrative script (as I will call the relevant cognitive


construct) pictures a world situated in time and populated by intelligent agents. The time<br />

span framed by the representation encompasses a series of different states mediated by<br />

accidental happenings and deliberate actions. To understand the sequence as narrative<br />

means <strong>to</strong> be capable of reconstructing the motivations of the agents and the causal<br />

connections between events and states. As a mental representation of a temporal<br />

sequence of events, narrative is not only linear – or multilinear, when it follows several<br />

parallel or interwoven destinies – but vec<strong>to</strong>rial: a plot must be followed in a specific<br />

direction, from birth <strong>to</strong> death, beginning <strong>to</strong> end.<br />

In contrast <strong>to</strong> some of the narrower definitions endorsed by narra<strong>to</strong>logists, the present<br />

approach does not limit narration <strong>to</strong> the re-presentation of past events by a narra<strong>to</strong>r, but<br />

accepts various modalities: narrative scripts can be <strong>to</strong>ld (diegetic mode), they can be<br />

shown (mimetic mode), or they can be enacted, not for the benefit of an audience, as is<br />

the case in drama, but as a self-rewarding activity. In this last mode, which prevails in<br />

computer games and participa<strong>to</strong>ry environments, the interac<strong>to</strong>r is also the beneficiary of<br />

the text. Rather than being necessarily pre-encoded in a semiotic body, moreover,<br />

narrative scripts can be dynamically generated during the performance of a text through a<br />

simulation of actions and events. In contrast <strong>to</strong> scholars such as Espen Aarseth, who want<br />

<strong>to</strong> maintain a strict distinction between computer games and narrative (a concept<br />

implicitly narrowed down <strong>to</strong> its literary manifestations), my definition regards games and<br />

literary texts (or rather, those games and those literary texts that satisfy the proper<br />

semantic conditions) as two different narrative modalities. In the discussion below, I will<br />

therefore consider texts conceived as literature as well as texts regarded as games, but I<br />

will exclude the rich field of electronic poetry, as well as some innovative <strong>digital</strong> texts<br />

that rely <strong>to</strong>o much on alea<strong>to</strong>ry principles <strong>to</strong> fulfill the strict requirements of narrative<br />

coherence.<br />

The development of <strong>digital</strong> textuality hit the literary scene at a time when "serious"<br />

literature – in contrast <strong>to</strong> popular culture – was in the grip of an epistemological and<br />

aesthetic crisis that challenged the closure, stability, and vec<strong>to</strong>riality of narrative form.<br />

The epistemological issue was a critique of the alleged blindness of classical narrative <strong>to</strong><br />

the complexity of the problem of truth: facts are asserted by authoritative narra<strong>to</strong>rs as a<br />

matter of absolute knowledge, and the reader is asked <strong>to</strong> take the text as the canonical<br />

version of the world. For the his<strong>to</strong>rian Hayden White (1987), narrative is not a reliable<br />

way <strong>to</strong> gain knowledge about the past, because it always involves a fabrication. Reality,<br />

he claims, does not offer itself <strong>to</strong> perception in the shape of a s<strong>to</strong>ry. And for the discourse<br />

analysts Elinor Ochs and Lisa Capps, a truly "authentic" narrative of personal experience<br />

would not be a stable reconstruction of the past but "the airing and evaluating of<br />

alternative possible understandings of past events" (<strong>2001</strong>: 17). The aesthetic issue arose<br />

from a similar desire <strong>to</strong> open up the text <strong>to</strong> multiple variants. Throughout the twentieth<br />

century, as Umber<strong>to</strong> Eco has shown in The Open Work (1989), artists in many media<br />

have been obsessed with the idea of making the text endlessly self-renewable, of<br />

capturing infinity in its necessarily bounded body, of turning the work of art from a static<br />

self-identical object in<strong>to</strong> a matrix of virtualities. Modern science has recently come up<br />

with some models and names for this aesthetic ideal: emergence, complexity, distributed<br />

intelligence.


Through its permanent inscription and linear reading pro<strong>to</strong>col, its bound spine and<br />

bounded content, the book was perceived as an obstacle <strong>to</strong> the dream of the text that<br />

keeps giving. If the problem came from the limitations of print, the solution might come<br />

from the affordances of <strong>digital</strong> technology. Thanks <strong>to</strong> the properties of reactivity,<br />

interactivity, volatility, and modularity, every run of a <strong>digital</strong> text can be turned in<strong>to</strong> a<br />

performance of different virtualities. Out of a limited number of elements, the computer<br />

allows the creation of a vast number of versions. Janet Murray (1997: 155–62) calls this<br />

property the "kaleidoscopic" nature of the medium.<br />

Twentieth-century writers did not await the advent of the <strong>digital</strong> age <strong>to</strong> develop a<br />

kaleidoscopic form of textuality. They did so through the technique of physically<br />

"chunking" and rearranging the text. We find the principle at work in the combina<strong>to</strong>rial<br />

algorithms developed by the members of the Oulipo literary movement: novels written on<br />

decks of cards which could be shuffled <strong>to</strong> produce different texts (Marc Saporta's<br />

Composition No 1, 1961); or sonnets obtained by selecting each verse from a matrix of<br />

poems (Raymond Queneau's Cent mille milliards de poemes). In these last two examples<br />

the combina<strong>to</strong>rial principle is alea<strong>to</strong>ry, and it does not guarantee the production of<br />

meaning, especially not of narrative meaning. We can make a picture out of any<br />

arrangement of pic<strong>to</strong>rial fragments; and perhaps even a "text" in the same way, since nonsense<br />

can have a certain poetic force, but certainly not a s<strong>to</strong>ry. Consider for instance what<br />

would happen if we reshuffled the functions of Vladimir Propp's Morphology of the<br />

Folktale (1962 [1928]):<br />

The villain is defeated. One member of a family either lacks something or desires <strong>to</strong> have<br />

something. The hero is recognized. The hero is married and ascends <strong>to</strong> the throne. The<br />

hero and the villain join in direct combat. A false hero presents unfounded claims. An<br />

interdiction is addressed <strong>to</strong> the hero.<br />

An alternative <strong>to</strong> random sequencing, also practiced by print authors, is a combination of<br />

chunking and directed linking. After a given fragment of text the reader is given the<br />

choice between two or more alternatives. In "A S<strong>to</strong>ry As You Like It" by the French<br />

Oulipo member Raymond Queneau, the reader is asked: (a) "Do you wish <strong>to</strong> hear the<br />

s<strong>to</strong>ry of the three big skinny beanpoles?" (yes – go <strong>to</strong> b; no – go <strong>to</strong> c); (b) "Do you want<br />

<strong>to</strong> hear the s<strong>to</strong>ry of the three meddling mediocre bushes?" (yes – go <strong>to</strong> d; no – go <strong>to</strong> e).<br />

The expansion of each node creates a decision tree. Because there is only one way <strong>to</strong><br />

reach a given node, whether terminal or medial, the author has absolute control over the<br />

information available <strong>to</strong> readers at every moment of their itinerary. This prevents the<br />

nonsense of the Propp example. But the ratio of number of paths vs. number of units in a<br />

tree-based system is not efficient; most works based on this algorithm offer a single-digit<br />

number of variants (two, for instance, in Julio Cortazar's Hopscotch (1966); about eight,<br />

in a typical Choose Your Own Adventures children's book).<br />

The <strong>digital</strong> medium offers a much more powerful engine for the creation of multi-variant<br />

narrative than either alea<strong>to</strong>ry combination or unconditional branching because the<br />

passage from one chunk of text <strong>to</strong> the next can be meaningfully controlled by code. In a<br />

hypertextual system, for instance, the operation of clicking on but<strong>to</strong>ns activates the


execution of some machine-language instructions, usually a "go<strong>to</strong>" <strong>to</strong> a certain memory<br />

address and an instruction <strong>to</strong> display the group of data beginning at this address. But the<br />

"go<strong>to</strong>" can be embedded in an "if… then… else" statement that imposes conditions on the<br />

branching. From the node "hero arrives at castle", for instance, the computer could be<br />

instructed <strong>to</strong> display the nodes "marriage <strong>to</strong> princess" or "king gives hero mission <strong>to</strong><br />

rescue princess", depending on whether or not the reader has already visited "hero defeats<br />

villain." In a computer game, similarly, the player's actions function as links between<br />

segments in the sense that they trigger the execution of code, which leads <strong>to</strong> changes in<br />

the display and in the global state of the system. What players can do at a given moment<br />

depends on what they did in the past. The placement of links or the timing of user action<br />

in a well-designed narrative or informational system cannot be a random act, unless the<br />

point is <strong>to</strong> signify randomness itself. The creation of multivariant narratives depends on<br />

the existence of pro<strong>to</strong>cols that maintain linear coherence on the cognitive level – for, as<br />

the Propp example suggests, you cannot freely permute events in your mental<br />

representation of a s<strong>to</strong>ry without consequences for the construction of causal and<br />

temporal relations. Here I would like <strong>to</strong> explore some of the system configurations,<br />

linking strategies, and modes of user participation that enable <strong>digital</strong> texts <strong>to</strong> achieve the<br />

difficult task of combining variability with narrativity. My investigation will focus on<br />

three aspects of narrative: discourse, which is the presentation of the narrative script<br />

through a particular medium, point of view, and plot – the narrative script itself.<br />

Variable Discourse<br />

Most of the <strong>digital</strong> texts that implement this type of variability are hypertexts, a form of<br />

organization described as follows in the documentation for S<strong>to</strong>ryspace, the authoring<br />

software published by Eastgate:<br />

The word "hypertext" refers <strong>to</strong> a specific kind of writing. While cus<strong>to</strong>mary text appears in<br />

a single sequence (e.g. page two always follows page one), hypertext presents the reader<br />

with multiple pathways through a document. These pathways are created by using<br />

hypertext links, a special kind of hypertext object that permits readers <strong>to</strong> actuate<br />

sequences, just like turning a page…. Simply put, hypertexts are branching textual<br />

objects that allow the reader <strong>to</strong> decide where <strong>to</strong> go next.<br />

(S<strong>to</strong>ryspace FAQ; available from <br />

By the admission of its developers, S<strong>to</strong>ryspace was designed <strong>to</strong> handle "large texts"<br />

(novels and databases rather than poems and short s<strong>to</strong>ries) with heavily interlinked nodes.<br />

One of the convenient features of S<strong>to</strong>ryspace is the au<strong>to</strong>matic building of a graph that<br />

keeps track of the system of links. This diagram gives the writer an overview of the<br />

developing network; but it can also serve as a navigational <strong>to</strong>ol for the reader.<br />

Figure 28.1 shows the organizational map of Stuart Moulthrop's novel Vic<strong>to</strong>ry Garden<br />

(1991). This map (inserted as art, rather than generated by the system) represents only the<br />

upper layer of the textual architecture: each named site on the map stands for a region of


the text with its own finely grained system of links, and it would be necessary <strong>to</strong> zoom in<br />

<strong>to</strong> a larger-scale image <strong>to</strong> get a full view of the textual architecture. The configuration of<br />

the map of Vic<strong>to</strong>ry Garden is what theorists would call a network, or unrestricted graph.<br />

The presence of circuits – the formal trademark of a network – means that there may be<br />

many different ways <strong>to</strong> get <strong>to</strong> the same node. The system designer can control the reader's<br />

itinerary on the local level (where <strong>to</strong> go from a given node) but not on the global level.<br />

This feature discourages what I call a "narrative" interpretation of the sequence viewed<br />

by the reader: an interpretation that narrowly associates the order of appearance of lexia<br />

with a chronological and causal chain of events in the reference world. This type of<br />

reading would frequently lead <strong>to</strong> nonsense. For instance, if we first visited a node where a<br />

character is dead, and then a node where the character is alive, we would have <strong>to</strong> imagine<br />

a miraculous resurrection <strong>to</strong> make a s<strong>to</strong>ry out of the sequence. It would take superhuman<br />

intelligence, or a map with very few decision points, for the designer <strong>to</strong> guarantee that<br />

every path will form a logically coherent plot. The map of Vic<strong>to</strong>ry Garden is not a<br />

projection of branching possible parallel times in<strong>to</strong> a spatial image but just what it claims<br />

<strong>to</strong> be: the map of a garden, or labyrinth, or graveyard – in any case a purely spatial<br />

construct. Despite its allusions <strong>to</strong> J. L. Borges's Garden of forking Paths, where paths<br />

represent possible futures rather than routes in space, Moulthrop's Vic<strong>to</strong>ry Garden does<br />

not outline multiple destinies for its heroes, but rather traces many pathways in<strong>to</strong> a<br />

reasonably solid and chronologically organized narrative core. There is no necessary<br />

global parallelism between the progression of narrative time and the progression of the<br />

reader on the paths of the garden, though there can be partial relations: for instance a<br />

stretch of path without branches or with a default continuation that captures a sequence of<br />

events in chronological order. (Vic<strong>to</strong>ry Garden is particularly rich in such linear paths:<br />

they are what enables the reader <strong>to</strong> get a general idea of the plot.)<br />

28.1 Figure The map of Vic<strong>to</strong>ry Garden


A text based on a map as complex as figure 28.1 does not tell a different s<strong>to</strong>ry for every<br />

reader, or with every reading session, it rather tells a s<strong>to</strong>ry in many different ways,<br />

varying discourse instead of plot. Approaching the text like a jigsaw puzzle, the reader<br />

rearranges lexia mentally, so that a fragment encountered at Tl in the reading sequence<br />

may be assigned time slot T22 in the reader's final reconstruction of the plot. Similarly,<br />

the reader of an epic text that begins in medias res will reconstrue a chronological order<br />

of events that differs from the order of their presentation.<br />

With its implication of a stable, determinate, complete, and coherent image <strong>to</strong> be<br />

recovered in its entirety, the image of the jigsaw puzzle admittedly fails <strong>to</strong> capture<br />

important aspects of the hypertextual reading experience. How does the phenomenon<br />

resist the metaphor? The reconstruction of narrative meaning is hardly ever complete<br />

because readers rarely visit all the lexia. Narrative is often only one of the threads in the<br />

textual web; following the postmodern aesthetics of the collage, most "classic" hypertexts<br />

interweave narration with metatextual comments, philosophical reflections, and intertextual<br />

imports. There may be not just one but many s<strong>to</strong>ries in the textual network, just as<br />

a print text may tell a complex s<strong>to</strong>ry with many subplots, or the text may present different<br />

versions of the same events without designating one of them as true. (Michael Joyce's<br />

afternoon (1987) circles, for instance, among several versions of an accident witnessed<br />

by the narra<strong>to</strong>r.) Even in its print form, narrative discourse rarely allows the linearization<br />

of all its information. Islands of free-floating events (mostly of the mental type) usually<br />

co-exist with streams of temporally ordered states of affairs. Whether print or <strong>digital</strong>,<br />

literary texts may actively militate against narrativity and its linear organization, though<br />

many of the texts that sound their demand for a post-narrative mode of signification do so<br />

from a platform of narrative fragments that maintain the reader's interest.<br />

The reconstruction of a reasonably consistent narrative world from a scrambled discourse<br />

would quickly become a tiresome activity if the reader's role were exhausted by the<br />

metaphor of the jigsaw puzzle. Any picture can be cut up, boxed, and sold as a puzzle. A<br />

narrow mapping of hypertext on<strong>to</strong> puzzles would therefore mean that the significance of<br />

the reader's involvement is independent of the narrative content of the text. From a<br />

literary point of view, the best hypertexts are those that manage <strong>to</strong> present the reader's<br />

activity of moving through the network and reassembling the narrative as a symbolic<br />

gesture endowed with a meaning specific <strong>to</strong> the text, a meaning which cannot be<br />

predicted by reading the medium as a built-in message. The hypertextual mechanism<br />

does not make a text au<strong>to</strong>matically innovative and significant; it is up <strong>to</strong> the author <strong>to</strong> put<br />

it in the service of a unique textual idea, of a metaphor that gives meaning <strong>to</strong> the reader's<br />

activity. In Vic<strong>to</strong>ry Garden, this activity is framed as an exploration of space. With its<br />

2,804 links connecting 993 lexia through a variety of paths planned over long stretches<br />

rather than from node <strong>to</strong> node (so that the narrative retains considerable linear<br />

coherence), Vic<strong>to</strong>ry Garden takes Daedalian architecture <strong>to</strong> a level of complexity that<br />

probably no reader has fully appreciated. Moulthrop's metaphor of the Garden of Forking<br />

Paths was so suggestive of a radically new mode of reading, and described the structure<br />

favored by S<strong>to</strong>ryspace so well, that it has become something of a theoretical cliche. In


Michael Joyce's afternoon, the often frustrated search of the reader for explanations and<br />

narrative continuity emulates the frantic calls of the narra<strong>to</strong>r <strong>to</strong> find out whether or not<br />

the accident he witnessed involved, and perhaps killed his ex-wife and son (Bolter 1991:<br />

126). And in Shelley Jackson's Patchwork Girl, the reader is made <strong>to</strong> stitch <strong>to</strong>gether a<br />

text out of heterogeneous fragments, some recycled from other texts, just as the narra<strong>to</strong>rcharacter<br />

Mary Shelley assembles a monster (allegory of a multiple, decentered<br />

subjectivity) by sewing <strong>to</strong>gether body parts collected from different women, and just as<br />

Shelley Jackson constructs a narrative identity for the monster from the s<strong>to</strong>ries of these<br />

women.<br />

Variable Point of View<br />

When the idea of interactive television was first introduced <strong>to</strong> the public, the advantage<br />

of the new technology was presented as an opportunity for specta<strong>to</strong>rs <strong>to</strong> view live<br />

broadcasts from a variety of perspectives. The capture of different cameras would be<br />

shown in small windows on the side of the screen, and by clicking on one of these<br />

windows the viewer would expand the image <strong>to</strong> the entire screen. Interactive movies, a<br />

genre that never really got off the ground, also played with the idea of variable point of<br />

view. In the exposition of I'm Your Man (1998), a DVD interactive movie playable on<br />

personal computers, we are given the choice between three options, each bearing a<br />

character's name. If we choose Leslie, the heroine, we will hear about Leslie's mission <strong>to</strong><br />

give some important computer disk files <strong>to</strong> an FBI agent at a party. If we choose Richard,<br />

the villain, we will find out about his sinister plan <strong>to</strong> pose as a fake agent in order <strong>to</strong> get<br />

the documents from Leslie and then <strong>to</strong> kill her. (This is a very cheesy plot.) If we choose<br />

Jack, the fool-turned-hero-in-spite-of-himself, we will simply accompany him on his way<br />

<strong>to</strong> the party, looking alternatively at Jack and with Jack at the derrieres and decolletes of<br />

female passers-by. At every decision point we can switch <strong>to</strong> the corresponding moment<br />

on one of the other two narrative tracks. The switches have no impact on the plot, or on<br />

the order of its presentation. Time moves inexorably forward, and by selecting a point of<br />

view we miss information that will only become available when we s<strong>to</strong>p the movie and<br />

start from the beginning again. Because of the temporal structure of film, it takes several<br />

passes through the movie <strong>to</strong> tie the three strands <strong>to</strong>gether in<strong>to</strong> one coherent plot.<br />

Purely textual environments can vary point of view without running in<strong>to</strong> this problem<br />

because a written text normally exists simultaneously in all of its parts. (Digital texts may<br />

change this, by destroying some of the links after the reader has visited them.) In the<br />

hypertext short s<strong>to</strong>ry "A Long Wild Smile" by Jeff Parker (found at<br />

), the reader can move back and forth<br />

between two narrative strands: one narrated by a woman's fiance and another by her<br />

lover. Each node has links <strong>to</strong> the partner narrative, enabling the reader <strong>to</strong> follow both<br />

versions in parallel. Some links make very quick incursions in<strong>to</strong> the other perspective,<br />

highlighting micro-level conflicts. In the following sequence, as Parker explains, the<br />

underlined words (which also function as links <strong>to</strong> the other passage) represent the<br />

interpretation by two different characters of the same string of sound:


(Node 1: Fiance's perspective) When I awake, she's joined him in the kitchen. She stares<br />

at the fridge while he stares at her. She says, "Did you ever ride ponies?" He says, "No."<br />

(Node 2: Lover's perspective) "Do you ever write rhyming poetry?" she said, taking a big<br />

sip of water and passing the cup <strong>to</strong> me. I drank and passed it back <strong>to</strong> her.<br />

This is a subtle authorial way <strong>to</strong> suggest misunderstanding (most likely the fiance's)<br />

behind the back of the two narra<strong>to</strong>rs. Locked in their own perspectives, the lover and the<br />

fiance quote what they hear, and they are not aware that the other interpreted the same<br />

words differently. But this kind of effect could have been achieved in print by<br />

juxtaposing the two passages and distinguishing them with indentation or with different<br />

fonts.<br />

Variations in point of view have been so successfully achieved in "old media" (think of<br />

William Faulkner's novel The Sound and the Fury, or of Akira Kurosawa's film<br />

Rashomon) that they seldom carry <strong>digital</strong> texts all by themselves. The online soap operas<br />

The Spot and The Lurker Files, which consisted of interwoven first-person narratives that<br />

represent the perspective of different characters, have now disappeared from the Web.<br />

Interactive film and TV will have <strong>to</strong> find other selling points <strong>to</strong> succeed commercially –<br />

such as the possibility for the latter <strong>to</strong> play any document from an archive. The linking<br />

design of Parker's text pursues a wider variety of literary effects than merely switching<br />

back and forth between two perspectives. Even a work as crude as I'm Your Man<br />

combines variable point of view with other types of choices, such as deciding what the<br />

characters should do next. It is perhaps computer games that show the most efficient use<br />

of variable point of view in a <strong>digital</strong> environment. Many games enable players <strong>to</strong> switch<br />

from a god's eye perspective, third-person display, through which they see their character<br />

as a moving object on a map of the playing field, <strong>to</strong> a first-person, horizontal perspective<br />

that shows the game world <strong>to</strong> the players through the eyes of their avatar. One view is for<br />

the planning of strategy in a suspended time, and the other for the execution of moves in<br />

the heat of the game-action.<br />

Variable Plot<br />

The simplest way <strong>to</strong> implement variability on the level of plot is <strong>to</strong> arrange multiple<br />

narrative developments along the branches of a decision tree. We see this principle at<br />

work in Choose Your Own Adventures children's s<strong>to</strong>ries, in the Queneau text quoted<br />

above, and in postmodern narratives with two or more endings or with different reading<br />

paths (Hopscotch). All these examples come from print texts, and all are fairly static. The<br />

various narrative lines are pre-inscribed within the database, limited in number (it would<br />

take God-like vision <strong>to</strong> control the combina<strong>to</strong>rial explosion of an arborescent structure<br />

with more than three levels), and the reader activates them through a purely selective type<br />

of interactivity. As long as this scheme is maintained, the <strong>digital</strong> medium does not really<br />

facilitate the creation of multiple plots. It makes travel along the branches a little bit<br />

easier, since clicking brings text <strong>to</strong> the reader, while print makes the reader go <strong>to</strong> the text,<br />

but the author still has <strong>to</strong> write the individual s<strong>to</strong>ries. But <strong>digital</strong> media present one<br />

significant advantage over print. They allow the user <strong>to</strong> interact with the text not just


selectively but also productively. It is only when the user contributes elements <strong>to</strong> a<br />

developing s<strong>to</strong>ry, allowing plots <strong>to</strong> be dynamically generated at run-time, that a system's<br />

narrative productivity can be raised above the level reachable by print media. In this type<br />

of system the player's actions perform an individualized narrative by responding <strong>to</strong> the<br />

affordances of the textual world. Since no two players will take the same moves, no two<br />

runs of the program will produce the same narrative trace. This productive type of<br />

interactivity is found mostly in computer games, although it also occurs in chatrooms,<br />

MOOs (Multi-User Dungeons, Object-Oriented), dialogue systems (the famous ELIZA<br />

program), virtual reality installations, and projects in Interactive Drama (Ryan 200la).<br />

Here I will examine its effect on plot in three types of games: interactive fiction, firstperson<br />

shooters, and simulation games.<br />

The lowest degree of plot-level variability occurs when the player's actions fill in the<br />

blank spots in a pre-determined narrative script. This situation occurs in interactive<br />

fiction (IF), a text-only <strong>digital</strong> genre that flourished in the early 1980s, when computer<br />

graphics were <strong>to</strong>o primitive <strong>to</strong> create an immersive environment. The genre nearly<br />

disappeared in the 1990s, displaced from the game market by fancy graphic interfaces<br />

and exiled from the front stage of <strong>digital</strong> literature by hypertext, a genre which quickly<br />

became the darling of theorists. But IF survives in small Internet niches, practiced<br />

without commercial and theoretical pressures by enthusiastic writers and programmers<br />

who freely share their work with a devoted community.<br />

In an interactive fiction, we impersonate a character within the s<strong>to</strong>ry world. We specify<br />

the actions of this character by typing instructions <strong>to</strong> the system in a simplified version of<br />

a natural language on an old-fashioned command-line interface. A parser analyzes the<br />

input, and the system responds with appropriate, or sometimes comically inappropriate<br />

actions. Since most IPs are designed as adventure games, the user's character has a task <strong>to</strong><br />

perform, and the plot only reveals itself fully <strong>to</strong> the user who is able <strong>to</strong> complete the task<br />

by solving all the problems.<br />

The downside of a system that grants free speech <strong>to</strong> the user is the impossibility of<br />

incorporating every verbal action in<strong>to</strong> a coherent dialogue. IF systems understand only a<br />

limited vocabulary – typically 200 verbs, and as many nouns as there are different objects<br />

in the game-world – and they reject any input that cannot be parsed by their limited<br />

syntactic and semantic competence. This means that while the user can technically<br />

perform an unlimited number of actions, only some of these actions yield events in the<br />

developing s<strong>to</strong>ry. The following examples, all taken from Spider and Web by Andrew<br />

Plotkin (Plotkin, website), illustrate various relations between input and plot. In the first<br />

example, player input counts as a turn in a conversation between the player's character<br />

and another member of the s<strong>to</strong>ry world, and it reproduces exactly a verbal event within<br />

this world (> represents the output of the system):<br />

Interrogation chamber [You are imprisoned in a chair]<br />

"Don't be absurd", [the interroga<strong>to</strong>r] says. "… if you'd had enough sense <strong>to</strong> walk away<br />

from that door, you wouldn't be here. You don't and you didn't and are; we caught you.


And you're going <strong>to</strong> start by telling me how you got through that door. Do you<br />

understand me? Player: Yes.<br />

>The man nods briefly – a man satisfied with the least important detail.<br />

In the next example, the player and the system takes turn at producing the s<strong>to</strong>ry, though<br />

the player's input is not an act of narration but a command <strong>to</strong> the system. The system<br />

responds by updating its knowledge base, by narrating the result of the action, and<br />

occasionally (though this is not exemplified in the passage below) by retelling the action<br />

<strong>to</strong> the player in more elaborate language:<br />

>You are carrying nothing of importance, except a lockpick.<br />

Player: Put lockpick on plate.<br />

>The pick locks itself rigidly <strong>to</strong> the black plate by the door. Its status dot flashes <strong>to</strong><br />

yellow.<br />

Player: Wait.<br />

>Time passes.<br />

The door slides neatly open.<br />

In the last example, the player's actions cannot be executed because they cannot be parsed<br />

by the system. Their event-status within the narrative is uncertain: do they count as futile<br />

attempts <strong>to</strong> solve problems, or are they al<strong>to</strong>gether external <strong>to</strong> the s<strong>to</strong>ry world? The<br />

response of the system is clearly external, and since it is situated on the same on<strong>to</strong>logical<br />

level as the player's attempted action, the whole dialogue may be regarded as taking place<br />

outside narrative space and time:<br />

>A plain metal door faces you <strong>to</strong> the east, near the alley's end. It's firmly shut.<br />

Player: Open door.<br />

>You don't see how.<br />

Player: Scream in despair.<br />

>That's not a verb I recognize.<br />

Interactive fiction thus keeps the plot on the projected track by expelling the actions that<br />

do not fit within the script <strong>to</strong> a sphere that narra<strong>to</strong>logists would call the extra-diegetic or<br />

para-textual.


Most interactive fictions trace a narrowly scripted trajec<strong>to</strong>ry. Players must visit a number<br />

of locations, either in free or fixed order, and they must solve a problem <strong>to</strong> get <strong>to</strong> the next<br />

location. In Spider and Web, for instance, player A may try the blaster and the lockpick <strong>to</strong><br />

enter a space before succeeding with the scan scrambler; player B may never solve the<br />

problem and be caught by the guards in the hallway (a fate which can be averted by<br />

consulting the solutions posted on the Web); while player C may succeed on the first try.<br />

The only room for variation resides in the player's unsuccessful attempts. Spider and Web<br />

creates an original variation on this pattern by presenting the user's failed actions not as<br />

actual events but as lies <strong>to</strong>ld by the player's character <strong>to</strong> an interroga<strong>to</strong>r who wants <strong>to</strong><br />

know how the character has managed <strong>to</strong> infiltrate a secret war labora<strong>to</strong>ry. Some<br />

interactive fictions have two or three endings, but in general, the variety of the player's<br />

input does not translate in<strong>to</strong> an equal variety on the level of plot. Though they can<br />

develop very imaginative scripts that spice up the solving of problems with narrative<br />

interest, IF texts offer little incentive <strong>to</strong> re-enter their world once the game has been<br />

beaten.<br />

But endless replayability is not a reliable sign of high narrative variability. Consider the<br />

case of the so-called "First Person Shooter" or FPS (Wolfenstein, Doom, Quake, Unreal<br />

Tournament, etc.), a type of game that is played over and over again by its fans. Here also<br />

the user is cast as a character situated in both the time and space of the fictional world.<br />

The actions of the user determine the fate of the puppet character, and by extension, the<br />

fate of the fictional world. The script of the game simulates a live encounter with<br />

opponents, who can be manipulated either by the computer itself, or, if the game is<br />

played in a networked environment, by other human players. Whereas IPs allow users <strong>to</strong><br />

type whatever they want, but filter out a narrow range of acceptable inputs, FPS games<br />

accept all of the user's actions as part of the game, but they limit these actions <strong>to</strong> moving<br />

around the game world, gathering ammunition, selecting weapons, and aiming and firing<br />

them. If we regard the player's moves as the writing of a "life s<strong>to</strong>ry" for the character,<br />

every run of the system produces a new life, and consequently a new narrative trace. This<br />

narrative is created dramatically by being enacted, rather than diegetically by being<br />

narrated. Because the speed of the processor and the number of variables involved in the<br />

execution of the game script make the system's reaction partially unpredictable <strong>to</strong> the<br />

player, FPS games are inexhaustible matrices of different lives for the player's character.<br />

The s<strong>to</strong>ries of these lives remain in a virtual state until they are mentally replayed. When<br />

players tell about their game experience, they do so by producing a standard "diegetic"<br />

narrative. For instance: "I was out of ammo, and Tyrus and Jason were surrounding me,<br />

and I thought I was dead, but I made this cool jump, and landed on the other side of the<br />

wall, and I found a cache of new ammo. I got back <strong>to</strong> the dungeon and killed them both,<br />

and we won the game." But while FPS games are played over and over again, players<br />

rarely "replay" in their minds the s<strong>to</strong>ry of a game, because these s<strong>to</strong>ries only differ from<br />

each other in who shot whom and who died when. This mono<strong>to</strong>nous diversity is only<br />

compounded by the fact that all FPSs implement the same narrative archetype – the s<strong>to</strong>ry<br />

of the quest – through the same motif: a physical confrontation between hero and villain.<br />

Though FPSs have a narrative base, like almost all recent computer games, the s<strong>to</strong>ries<br />

they generate are worth experiencing in the first person but rarely worth telling <strong>to</strong> a third<br />

party.


For a combination of replayability and narrative diversity, no formula can presently rival<br />

the type of game known as simulation or as "god game" – the latter name due <strong>to</strong> the fact<br />

that the player manipulates the objects of the game world from an external position of<br />

power. Simulation games create a dynamic model of a complex entity, such as a city<br />

(Simcity), an empire (Caesar), or a family (The Sims). An emergent process, the global<br />

evolution of this entity is the product of an algorithm that computes the interrelated<br />

consequences of the actions of many individual agents and continually feeds this output<br />

back in<strong>to</strong> its own engine. In The Sims, for instance, a game whose master plot is a pursuit<br />

of happiness that requires a steady climb up the economic and social ladder, the wellbeing<br />

of the family of Bob and Betty Newbie is a balancing act through which this couple<br />

of suburbanites must manage the resources of time <strong>to</strong> satisfy bodily demands (eating,<br />

sleeping, using the bathroom), emotional demands (their relation <strong>to</strong> each other),<br />

entertainment demands (satisfied by buying the proper commodities), social demands<br />

(getting along with neighbors), and economic pressures (they must earn money <strong>to</strong> fill<br />

their house with pleasure-giving commodities). The game has been criticized for its<br />

capitalist philosophy, but the comic texts that ridicule some of the objects available for<br />

purchase can just as easily be read as a satire of consumerism.<br />

In keeping with its conception of life as a s<strong>to</strong>ry that is constantly being written by the<br />

interaction between individuals and their environment, The Sims presents each object<br />

within the game world as an opportunity for action. A computer, for instance, affords the<br />

actions of playing games (good <strong>to</strong> cheer up depressed characters) or of finding a job<br />

(good <strong>to</strong> expand buying power). Characters are like objects: they <strong>to</strong>o offer opportunities<br />

for actions by other characters. By mousing over Betty Newbie, the user playing Bob will<br />

for instance discover that Bob can kiss her, hug her, tickle her, or talk <strong>to</strong> her. The<br />

possibilities of action evolve during the run of the program, and since affordances are<br />

determined by the global state of the system, as well as by the nature of the objects, the<br />

user's choices will always produce a coherent narrative development.<br />

The object of the game is not <strong>to</strong> win, since life never runs out of problems <strong>to</strong> be solved,<br />

but <strong>to</strong> manage the life of the characters as successfully as possible according <strong>to</strong> cultural<br />

standards (make them rich), or <strong>to</strong> the player's personal idea of dramatic development (for<br />

instance, create interpersonal conflicts and situations leading <strong>to</strong> catastrophic events).<br />

Some players (the novelists) want <strong>to</strong> produce specific scenarios; others (the virtual<br />

his<strong>to</strong>rians) are more interested in discovering how certain actions will affect the fate of<br />

the family. It is the same curiosity that makes us wonder "what would my life be like now<br />

if I hadn't met this stranger in a bar" or that inspires practitioners of virtual his<strong>to</strong>ry <strong>to</strong><br />

write essays about "what would have happened in US foreign policy if JFK had not been<br />

murdered." The motivation of both types of players is much more narrative than that<br />

which drives the players of FPS. In contrast <strong>to</strong> the lives of FPS players, the narratives<br />

produced by The Sims can be enjoyed retrospectively as well as during the run of the<br />

program. The system allows users <strong>to</strong> retell the s<strong>to</strong>ry of their game, or <strong>to</strong> invent their own<br />

Sim s<strong>to</strong>ries, by taking snapshots of the screen and by complementing their pictures with<br />

text. Many of these car<strong>to</strong>on-form s<strong>to</strong>ries are posted on the game's website. Whether or<br />

not they make interesting reading, and whether or not they chronicle actual games, the


urge of players <strong>to</strong> share them with other players provides ample evidence that computer<br />

games can produce a genuine narrative interest in their fictional worlds.<br />

The ludic pleasure of deciphering the logic of the system – what game designers call<br />

reverse engineering- cannot be separated from the narrative pleasure of watching the<br />

s<strong>to</strong>ry unfold. Without playing skills, the player would be unable <strong>to</strong> create interesting<br />

s<strong>to</strong>ries. On my first attempt <strong>to</strong> play The Sims, for instance, the cooking range caught fire,<br />

and because I hadn't bought a phone I could not call the fire department. I watched<br />

helplessly my whole family die, and my only option was <strong>to</strong> let the last survivor mourn for<br />

his loved ones before being himself engulfed by the flames. At this point there was<br />

nothing <strong>to</strong> do but start again from scratch. It is only by learning how the system "thinks"<br />

that players can increase their authorial power over their family. But since the system<br />

throws in unexpected events, the player's control over the fate of the characters is never<br />

absolute. For narrative <strong>to</strong> be pleasurable, apprecia<strong>to</strong>rs must be able <strong>to</strong> anticipate <strong>to</strong> some<br />

extent the development of the s<strong>to</strong>ry, but the s<strong>to</strong>ry must be able <strong>to</strong> fulfill expectations in a<br />

surprising way. This time-tested formula holds no less for simulation games than for<br />

novels and drama.<br />

For all its narrative productivity, however, the formula of The Sims does not entirely<br />

satisfy the ambition of Will Wright, the designer of the game. He envisions a networked<br />

system that detects the most innovative subplots produced by users, updates the program<br />

with devices that facilitate their creation, and sends the new version <strong>to</strong> thousands of other<br />

users <strong>to</strong> find out if they, <strong>to</strong>o, will attempt these subplots. "So in fact you [will] have the<br />

players… cross-pollinating their creativity", but the exchange will be mediated by the<br />

computer (Pearce 2002: 6). What Wright has in mind is a narrative variability of a higher<br />

power – a variability that affects the s<strong>to</strong>rytelling engine.<br />

Beyond Narrative<br />

In all the previous examples, variability remained within the bounds of narrative logic.<br />

Let me conclude my survey with an example that starts well within these bounds, but<br />

transgresses them during the run of the program, so that the text performs before the<br />

reader's eyes the dissolution of narrative. The text in question is The Impermanence<br />

Agent, by Noah Wardrip-Fruin and Brion Moss (2002). It consists of two windows on the<br />

screen. One of them contains the input narrative: a s<strong>to</strong>ry inspired by the death of the<br />

author's grandmother, Nana, illustrated with family pho<strong>to</strong>s. The other window contains<br />

texts from various authors and memorial imagery from multiple cultures. The content of<br />

both windows scrolls down slowly by itself, then returns <strong>to</strong> the <strong>to</strong>p, in an infinite loop. A<br />

reactive, rather than interactive text which takes full advantage of the networking<br />

capability of the medium, The Impermanence Agent modifies the content of each window<br />

by gradually integrating materials culled from the user's "scrapbook" – the file in which<br />

the Internet browser of the user's system s<strong>to</strong>res text and images from the websites most<br />

recently visited. Whereas the original s<strong>to</strong>ry reads "As a girl, at recess, Joan played<br />

punching games with the boys" it may become, "As a girl, at recess, Joan <strong>to</strong> follow in the<br />

Bowlers' footsteps played TeacberSource: Health and Fitness with the boys" (Wardrip-<br />

Fruin and Moss 2002: 14). A "lightweight intelligence model" selects "interesting" or


frequent words from the user's scrapbook, and makes sure that they fit syntactically in<strong>to</strong><br />

the text, but the model does not check the output for semantic coherence. The visual<br />

material undergoes similar blending with pictures from the scrapbook. The authors claim<br />

that this algorithm cus<strong>to</strong>mizes narrative discourse <strong>to</strong> the reader's interests, but their<br />

definition of narrative is so loose that it accepts any grammatical sequence of words.<br />

After several loops through the program, all that is left of the original s<strong>to</strong>ry is bits and<br />

pieces of an alea<strong>to</strong>ry collage, as the text is invaded by fragments of other texts that may<br />

point <strong>to</strong>wards, but never fully tell their own s<strong>to</strong>ries. My point in presenting this example<br />

is not <strong>to</strong> deny artistic potential <strong>to</strong> this kind of project, but rather <strong>to</strong> locate the point where<br />

multivariant textuality leaves narrativity behind and becomes conceptual art.<br />

The textual phenomena described in this chapter represent two extremes on the cultural<br />

spectrum. While computer games have taken popular culture by s<strong>to</strong>rm, generating a<br />

billion-dollar industry that rivals Hollywood and Disneyland, hypertext is an arcane<br />

academic genre read mostly by theorists and prospective authors. What remains <strong>to</strong> be<br />

conquered for <strong>digital</strong> textuality is the terri<strong>to</strong>ry that lies between the stereotyped narrative<br />

scripts of popular culture and the militant anti-narrativity of so many experimental texts:<br />

a terri<strong>to</strong>ry where narrative form is neither frozen nor ostracized, but recognized as an<br />

endlessly productive source of knowledge and aesthetic experiences. In the early 1990s,<br />

when theorists embraced hypertext as the genre that would carry the future of <strong>digital</strong><br />

literature, the concepts of non-linearity and spatiality s<strong>to</strong>od at the <strong>to</strong>p of their list of<br />

aesthetic preferences. But narrative, as we have seen, is a fundamentally temporal, and<br />

consequently linear form of meaning. It was the great achievement of the twentiethcentury<br />

novel <strong>to</strong> have created complex networks of internal relations, through which the<br />

present word, passage, motif, or episode activated the energies of other textual elements<br />

<strong>to</strong> form patterns of signification that transcended the linear development of the plot. This<br />

process of irradiation produced what critics have called a spatial form. But the greatest<br />

spatial novels (for instance Proust's A la recherche du temps perdu) can still be read for<br />

the plot, because their spatial organization complements, rather than destroys, the<br />

temporality of their narrative layer. Computer games, especially the much-maligned FPS,<br />

also present a very efficient use of time and space, if by space one understands the<br />

concrete geography of the game world rather than an abstract formal pattern. The<br />

pleasure of FPS consists in equal parts of moving through a three-dimensional<br />

environment that constantly updates itself <strong>to</strong> reflect the player's perspective, and of taking<br />

the right action at the right moment in a world relentlessly subjected <strong>to</strong> the ticking of the<br />

clock. For <strong>digital</strong> texts <strong>to</strong> establish themselves within the cultural middle ground – the<br />

narratives of the educated but not professional public – they must do the opposite of what<br />

the twentieth-century novel achieved, and perhaps learn a lesson from computer games,<br />

without succumbing <strong>to</strong> their propensity for repetitive themes and stereotyped s<strong>to</strong>rylines:<br />

naturally spatial, these texts must reconquer the narrative temporality that fuels the<br />

reader's desire.<br />

References for Further Reading<br />

Aarseth, E. (1997). Cybertext. Perspectives on Ergodic Literature. Baltimore: Johns<br />

Hopkins University Press.


Aarseth, E. (1999). Aporia and Epiphany in Doom and The Speaking Clock: The<br />

Temporality of Ergodic Art. In Cyberspace Textuality (pp. 31–41), ed. M.-L. Ryan.<br />

Blooming<strong>to</strong>n: Indiana University Press.<br />

Aarseth, E. (forthcoming). Repurposing the Novel - Narrative Literature in the Turning<br />

Universe. In The Novel, ed. Franco Moretti. Prince<strong>to</strong>n, NJ: Prince<strong>to</strong>n University Press.<br />

Bolter, J. (1991). Writing Space: The Computer, Hypertext, and the His<strong>to</strong>ry of Writing.<br />

Hillsdale, NJ: Lawrence Erlbaum.<br />

Cortázar, Julio (1966). Hopscotch (Rayuela), tr. G. Rabassa. New York: Pantheon.<br />

Douglas, J. Y. (2000). The End of Books - or Books Without End? Reading Interactive<br />

Narratives. Ann Arbor: University of Michigan Press.<br />

Eco, U. (1989 [1962]). The Open Work. Cambridge, MA: Harvard University Press.<br />

Hayles, N. K. (2000a). Flickering Connectivities in Shelley Jackson's Patchwork Girl'.<br />

The Importance of Media-specific Analysis. Postmodern Culture 10, 2. At<br />

http://www.iath.virginia.edu/pmc/text-only/issue.lOO/10.2hayles.txt.<br />

Hayles, N. K. (2000b). The Transformation of Narrative and the Materiality of<br />

Hypertext. Narrative 9, 1: 21–220.<br />

I'm Your Man (1998). Dir. B. Bejean. A Choice Point Film. Presented by Planet Theory<br />

in association with DVD international. DVD edition produced by B. Franzblau.<br />

Jackson, S. (1995). Patchwork Girl. Hypertext software. Cambridge, MA: Eastgate<br />

Systems.<br />

Joyce, M. (1987). afternoon, a s<strong>to</strong>ry. Hypertext software. Cambridge, MA: Eastgate<br />

Systems.<br />

Joyce, M. (1995). Of Two Minds: Hypertext, Pedagogy and Poetics. Ann Arbor:<br />

University of Michigan Press.<br />

Koskimaa, R. (2000). Digital Literature: From Text <strong>to</strong> Hypertext and Beyond.<br />

Unpublished PhD dissertation. University of Jyväskylä Available at:<br />

http://www.cc.jyu.fi/~koskimaa/thesis/.<br />

Landow, G. (1997). Hypertext 2.0: The Convergence of Contemporary Critical Theory<br />

and Technology. Baltimore: Johns Hopkins University Press.<br />

Manovich, L. (<strong>2001</strong>). The Language of New Media. Cambridge, MA: MIT Press.


Montfort, N. (2003). Twisty Little Passages: An Approach <strong>to</strong> Interactive Fiction.<br />

Cambridge, MA: MIT Press.<br />

Moulthrop, S. (1991). Vic<strong>to</strong>ry Garden. Hypertext software. Cambridge, MA: Eastgate<br />

Systems.<br />

Murray, J. (1997). Hamlet on the Holodeck: The Future of Narrative in Cyberspace. New<br />

York: Free Press.<br />

Ochs, E. and L. Capps (<strong>2001</strong>). Living Narrative: Creating Lives in Everyday S<strong>to</strong>rytelling.<br />

Cambridge, MA: Harvard University Press.<br />

Ong, W. (1982). Orality and Literacy. The Technologizing of the Word. London:<br />

Methuen.<br />

Parker, J. A Poetics of the Link. Electronic Book Review 12. Accessed April 21, 2004. At<br />

http://altx.com/ebr12/park/park.htm.<br />

Pearce, C. (2002). Sims, Battlebots, Cellular Au<strong>to</strong>mata, God and Go. A Conversation<br />

with Will Wright. Gamestudies 2. At http://gamestudies.org/0102/pearce.<br />

Plotkin, A. Spider and Web. Interactive Fiction. Accessed April 21, 2004. At<br />

http://www.wurb.com/if/game/207. Solution <strong>to</strong> the game (by PJG, modified by Nils<br />

Barth) at http://www.ifarchive.org/if-archive/solutions/tangle.sol.<br />

Propp, Vladimir (1968 [1928]). Morphology of the Folktale, tr. L. Scott, rev. L. Wagner.<br />

Austin: University of Texas Press.<br />

Queneau, R. A S<strong>to</strong>ry As You Like It: Interactive Version. Accessed April 21, 2004. At<br />

http://www.thing.de/projekte/7:9%23/queneau_l.html.<br />

Ryan, M.-L. (<strong>2001</strong>a). Narrative as Virtual Reality: Immersion and Interactivity in<br />

Literature and Digital Media. Baltimore: Johns Hopkins University Press.<br />

Ryan, M.-L. (<strong>2001</strong>b). Beyond Myth and Metaphor: Narrative in Digital Media.<br />

Gamestudies 1. At http://www.gamestudies.org/0101/ryan/.<br />

Saporta, Marc (1961). Composition No 1. Paris: Seuil.<br />

Sloane, S. (2000). Digital Fictions: S<strong>to</strong>rytelling in a Material World. Stamford, CT:<br />

Ablex.<br />

The Sims (2000). Computer game. Designer: Will Wright. Maxis/Electronic Arts. At<br />

http://www.thesims.com.


Wardrip-Fruin, Noah and Brion Moss (with A. C. Chapman and Duane Whitehurst)<br />

(2002). The Impermanence Agent. Project and context. CyberText Yearbook, ed. Markku<br />

Eskelinen and Raine Koskimaa (pp. 13–58). University of Jyvaskyla: Publications of the<br />

Research Centre for Contemporary Culture.<br />

White, H. (1987). The Content of the Form: Narrative Discourse and His<strong>to</strong>rical<br />

Representation. Baltimore: Johns Hopkins University Press.<br />

29.<br />

Speculative Computing: Aesthetic Provocations in<br />

Humanities Computing<br />

Johanna Drucker (and Bethany Nowviskie)<br />

With roots in computational linguistics, stylometrics, and other quantitative statistical<br />

methods for analyzing features of textual documents, <strong>humanities</strong> computing has had very<br />

little use for analytic <strong>to</strong>ols with foundations in visual epistemology. In this respect<br />

<strong>humanities</strong> computing follows the text-based (dare I still say – logocentric?) approach<br />

typical of traditional <strong>humanities</strong>. "Digital" <strong>humanities</strong> are distinguished by the use of<br />

computational methods, of course, but they also make frequent use of visual means of<br />

information display (tables, graphs, and other forms of data presentation) that have<br />

become common in desk<strong>to</strong>p and Web environments. Two significant challenges arise as a<br />

result of these developments. The first is <strong>to</strong> meet requirements that humanistic thought<br />

conform <strong>to</strong> the logical systematicity required by computational methods. The second is <strong>to</strong><br />

overcome humanists' long-standing resistance (ranging from passively ignorant <strong>to</strong><br />

actively hostile) <strong>to</strong> visual forms of knowledge production. Either by itself would raise a<br />

host of issues. But the addition of a third challenge – <strong>to</strong> engage computing <strong>to</strong> produce<br />

useful aesthetic provocations – pushes mathesis and graphesis in<strong>to</strong> even more unfamiliar<br />

collaborations. Speculative approaches <strong>to</strong> <strong>digital</strong> <strong>humanities</strong> engage subjective and<br />

intuitive <strong>to</strong>ols, including visual interfaces, as primary means of interpretation in<br />

computational environments. Most importantly, the speculative approach is premised on<br />

the idea that a work is constituted in an interpretation enacted by an interpreter. The<br />

computational processes that serve speculative inquiry must be dynamic and constitutive<br />

in their operation, not merely procedural and mechanistic.<br />

To some extent these notions are a radical departure from established practices in <strong>digital</strong><br />

<strong>humanities</strong>. As the articles in this volume attest, many of the practices in <strong>digital</strong><br />

<strong>humanities</strong> are becoming standardized. Technical and practical environments have<br />

become more stable. So have procedures for developing meta-data, for content modeling,<br />

for document classification and organization, search instruments, and the various<br />

pro<strong>to</strong>cols of ordering, sorting, and accessing information in <strong>digital</strong> formats. In its<br />

broadest conceptualization, speculative computing is premised on the conviction that<br />

logical, systematic knowledge representation, adequate though it may be for many fields


of inquiry, including many aspects of the <strong>humanities</strong>, is not sufficient for the<br />

interpretation of imaginative artifacts.<br />

Intellectual debates, collateral but substantive, have arisen as <strong>digital</strong> humanists engage<br />

long-standing critical discussions of the "textual condition" in its material, graphical,<br />

bibliographical, semantic, and social dimensions. No task of information management is<br />

without its theoretical subtext, just as no act of instrumental application is without its<br />

ideological aspects. We know that the "technical" tasks we perform are themselves acts<br />

of interpretation. Intellectual decisions that enable even such fundamental activities as<br />

keyword searching are fraught with interpretative baggage. We know this – just as surely<br />

as we understand that the front page of any search engine (the world according <strong>to</strong><br />

"Yahoo!") is as idiosyncratic as the Chinese Emperor's Encyclopedia famously conjured<br />

by Borges and commented upon by Foucault. Any "order of things" is always an<br />

expression of human foibles and quirks, however naturalized it appears at a particular<br />

cultural moment. We often pretend otherwise in order <strong>to</strong> enact the necessary day-<strong>to</strong>-day<br />

"job" in front of us, bracketing out the (sometimes egregious) assumptions that allow<br />

computational methods (such as markup or data models) <strong>to</strong> operate effectively.<br />

Still, we didn't arrive at <strong>digital</strong> <strong>humanities</strong> naively. Though work in <strong>digital</strong> <strong>humanities</strong> has<br />

turned some relativists in<strong>to</strong> pragmatists under pressure of technical exigencies, it has also<br />

reinvigorated our collective attention <strong>to</strong> the heart of our intellectual undertaking. As the<br />

applied knowledge of <strong>digital</strong> <strong>humanities</strong> becomes integrated in<strong>to</strong> libraries and archives,<br />

providing the foundation for collection management and delivery systems, the ecological<br />

niche occupied by theory is called on <strong>to</strong> foster new self-reflective activity. We are not<br />

only able <strong>to</strong> use <strong>digital</strong> instruments <strong>to</strong> extend <strong>humanities</strong> research, but <strong>to</strong> reflect on the<br />

methods and premises that shape our approach <strong>to</strong> knowledge and our understanding of<br />

how interpretation is framed. Digital <strong>humanities</strong> projects are not simply mechanistic<br />

applications of technical knowledge, but occasions for critical self-consciousness.<br />

Such assertions beg for substantiation. Can we demonstrate that <strong>humanities</strong> computing<br />

isn't "just" or "merely" a technical innovation, but a critical watershed as important as<br />

deconstruction, cultural studies, feminist thinking? To do so, we have <strong>to</strong> show that <strong>digital</strong><br />

approaches don't simply provide objects of study in new formats, but shift the critical<br />

ground on which we conceptualize our activity. The challenge is <strong>to</strong> structure instruments<br />

that engage and enable these investigations, not only those that allow theoretically<br />

glossed discussion of them. From a distance, even a middle distance of practical<br />

engagement, much of what is currently done in <strong>digital</strong> <strong>humanities</strong> has the look of<br />

au<strong>to</strong>mation. Distinguished from augmentation by Douglas Engelbart, one of the<br />

pioneering figures of graphical interface design, au<strong>to</strong>mation suggests mechanistic<br />

application of technical knowledge according <strong>to</strong> invariant principles. Once put in<strong>to</strong><br />

motion, an au<strong>to</strong>matic system operates, and its success or benefit depends on the original<br />

design. By contrast, Engelbart suggested that augmentation extends our intellectual and<br />

cognitive – even imaginative – capabilities through prosthetic means, enhancing the very<br />

capabilities according <strong>to</strong> which the operations we program in<strong>to</strong> a computer can be<br />

conceived. Creating programs that have emergent properties, or that bootstrap their<br />

capabilities through feedback loops or other recursive structures, is one stream of


esearch work. Creating <strong>digital</strong> environments that engage human capacities for subjective<br />

interpretation, interpellating the subjective in<strong>to</strong> computational activity, is another.<br />

Prevailing approaches <strong>to</strong> <strong>humanities</strong> computing tend <strong>to</strong> lock users in<strong>to</strong> procedural<br />

strictures. Once determined, a data structure or content model becomes a template<br />

restricting interpretation. Not in your tag set? Not subject <strong>to</strong> hierarchical ordering? Too<br />

bad. Current methods don't allow much flexibility – a little like learning <strong>to</strong> dance by<br />

fitting your feet <strong>to</strong> footsteps molded in<strong>to</strong> concrete. Speculative computing suggests that<br />

the concrete be replaced by plasticine that remains malleable, receptive <strong>to</strong> the trace of<br />

interpretative moves. Computational management of <strong>humanities</strong> documents requires that<br />

"content" has be subjected <strong>to</strong> analysis and then put in<strong>to</strong> conformity with formal<br />

principles.<br />

Much of the intellectual charge of <strong>digital</strong> <strong>humanities</strong> has come from the confrontation<br />

between the seemingly ambiguous nature of imaginative artifacts and the requirements<br />

for formal dis-ambiguation essential for data structures and schema. The requirement that<br />

a work of fiction or poetry be unders<strong>to</strong>od as an "ordered hierarchy of content objects"<br />

(Allen Renear's oft-cited phrase of principles underlying the Text Encoding Initiative)<br />

raises issues, as Jerome McGann has pointed out. Productive as these exchanges have<br />

been, they haven't made the shrug of resignation that accompanies acceptance of such<br />

procedures and presses them in<strong>to</strong> practice in<strong>to</strong> anything more than a futile protest against<br />

the Leviathan of standardization. Alternatives are clearly needed, not merely objections.<br />

The problems are not just with ordered hierarchies, but with the assumption that an<br />

artifact is a stable, constant object for empirical observation, rather than a work produced<br />

through interpretation.<br />

Speculative computing is an experiment designed <strong>to</strong> explore alternative approaches. On a<br />

technical level, the challenge is <strong>to</strong> change the sequence of events through which the<br />

process of "dis-ambiguation" occurs. Interpretation of subjective activity can be<br />

formalized concurrent with its production – at least, that is the design principle we have<br />

used as the basis of Temporal Modeling.<br />

By creating a constrained visual interface, Temporal Modeling puts subjective<br />

interpretation within the system, rather than outside it. The subjective, intuitive<br />

interpretation is captured and then formalized in<strong>to</strong> a structured data scheme, rather than<br />

the other way around. The interface gives rise <strong>to</strong> XML exported in a form that can be<br />

used <strong>to</strong> design a document type definition (DTD) or <strong>to</strong> be transformed through use of<br />

Extensible Stylesheet Language Transformation (XSLT) or other manipulations. A<br />

description of the technical parameters that realize these conceptual premises is described<br />

in the case study below. The project is grounded on convictions that subjective<br />

approaches <strong>to</strong> knowledge representation can function with an intellectual rigor<br />

comparable <strong>to</strong> that usually claimed by more overtly formal systems of thought. This<br />

experimental approach has potential <strong>to</strong> expand <strong>humanities</strong> computing in theoretical scope<br />

and practical application.


Our path in<strong>to</strong> the "speculative" has been charted by means of aesthetic exploration,<br />

emphasizing visual means of interpretation. These are informed by the his<strong>to</strong>ry of<br />

aesthetics in descriptive and generative approaches, as well as by the anomalous<br />

principles of 'pataphysics, that invention of the late nineteenth-century French poetphilosopher<br />

Alfred Jarry. An outline of these aesthetic, literary, and critical traditions,<br />

and their role in the ongoing development of <strong>digital</strong> <strong>humanities</strong>, forms the first part of<br />

this chapter. This is followed by a discussion of the project that demonstrates the working<br />

viability of the precepts of speculative computing, Temporal Modeling. I invited Bethany<br />

Nowviskie <strong>to</strong> author this final section, since the conceptual and technical development of<br />

this research project has proceeded largely under her intellectual guidance.<br />

Aesthetics and Digital Humanities<br />

The his<strong>to</strong>ry of aesthetics is populated chiefly by descriptive approaches. These are<br />

concerned with truth value, the specificity of individual media and activity "proper" <strong>to</strong><br />

their form, the development of taste and knowledge, and the capacity of aesthetics <strong>to</strong><br />

contribute <strong>to</strong> moral improvement – and, of course, notions of beauty and the aesthetic<br />

experience. These concerns are useful in assessing the aesthetic capabilities of <strong>digital</strong><br />

media – as well as visual forms of knowledge production – even if only because of the<br />

peculiar prejudices such traditional approaches have instilled in our common<br />

understanding. For instance, long-standing tensions between images and text-based forms<br />

of knowledge production still plague humanist inquiry. A disposition against visual<br />

epistemology is deeply rooted in conceptions of image and word within their morally and<br />

theoretically charged his<strong>to</strong>ry in Western philosophy. A schematic review of such key<br />

traditional issues provides a useful preface <strong>to</strong> understanding current concerns, particularly<br />

as visual <strong>to</strong>ols become integrated in<strong>to</strong> <strong>digital</strong> contexts as primary instruments of<br />

<strong>humanities</strong> activity.<br />

Fundamental distinctions differentiate descriptive modes from the intellectual traditions<br />

that inform our project: generative aesthetics, 'pataphysics, speculative thought, and<br />

quantum poetics. Generative approaches are concerned with the creation of form, rather<br />

than its assessment on grounds of truth, purity, epistemological, cognitive, or formal<br />

value. Speculative aesthetics is a rubric hatched for our specific purposes, and<br />

incorporates emergent and dynamic principles in<strong>to</strong> interface design while also making a<br />

place for subjectivity within the computational environment. 'Pataphysics inverts the<br />

scientific method, proceeding from and sustaining exceptions and unique cases, while<br />

quantum methods insist on conditions of indeterminacy as that which is intervened in any<br />

interpretative act. Dynamic and productive with respect <strong>to</strong> the subject-object dialectic of<br />

perception and cognition, the quantum extensions of speculative aesthetics have<br />

implications for applied and theoretical dimensions of computational <strong>humanities</strong>.<br />

Before plunging in<strong>to</strong> the vertiginous world of speculative and ';pataphysical endeavors,<br />

some frameworks of traditional aesthetics provide useful points of departure for<br />

understanding the difficulties of introducing visual means of knowledge representation<br />

in<strong>to</strong> <strong>digital</strong> <strong>humanities</strong> contexts. To reiterate, the themes of descriptive aesthetics that are<br />

most potently brought <strong>to</strong> bear on <strong>digital</strong> images are: truth value, "purity" or capabilities of


a medium, the cognitive values of aesthetics, and the moral improvement aesthetic<br />

experience supposedly fosters. Debates about beauty I shall leave aside, except in so far<br />

as they <strong>to</strong>uch on questions of utility, and the commonplace distinction between applied<br />

and artistic activity.<br />

The emphasis placed on the distinction between truth-value and imitation in classical<br />

philosophy persists in contemporary suspicions of <strong>digital</strong> images. The simulacral<br />

representations that circulate in cyberspace (including <strong>digital</strong> displays of information in<br />

visual form) are so many removes from "truth" that they would be charged with multiple<br />

counts of aesthetic violation in any Socratic court. Pla<strong>to</strong>nic hierarchies, and their negative<br />

stigmatization of images as mere imitations of illusions, are famously entrenched in<br />

Western thought. Whether we consider the virtual image <strong>to</strong> be a thing-in-itself, with<br />

on<strong>to</strong>logical status as a first-order imitation, or as a mimetic form and thus yet another<br />

remove from those Ideas whose truth we attempt <strong>to</strong> ascertain, hardly matters. The fixed<br />

hierarchy assesses aesthetic judgment against a well-marked scale of authenticity.<br />

From a theological perspective, images are subject <strong>to</strong> negative judgment except when<br />

they serve as instruments of meditation, as material forms whose properties function as a<br />

first rung on the long ladder <strong>to</strong>wards enlightenment. Such attitudes are characterized by a<br />

disregard for embodied intelligence and of the positive capacities of sensory perception.<br />

Denigrating the epistemological capacity of visualization, they assume that art and<br />

artifice are debased from the outset – as deceptive, indulgent acts of hubris – or worse,<br />

temptations <strong>to</strong> sinful sensuality. But if images are necessarily inferior <strong>to</strong> some "Idea"<br />

whose pale shadow they represent, <strong>digital</strong> images are redeemed only when they bring the<br />

ideal form of data in<strong>to</strong> presentation. The difficulty of such reasoning, however, is that it<br />

collapses in<strong>to</strong> questions about what form data has in a disembodied condition.<br />

Aris<strong>to</strong>tle's concern with bow things are made, not just how "truthful" they are, suggested<br />

that it was necessary <strong>to</strong> pay attention <strong>to</strong> the properties particular <strong>to</strong> each medium. The<br />

idea of a "proper" character for poetry was opposed <strong>to</strong> – or at least distinct from – that of<br />

visual forms. Likewise, sculpture was distinguished from painting, and so on, in an<br />

approach dependent on the specificity of media and their identifying properties. This<br />

notion of "propriety" led <strong>to</strong> differentiation among aesthetic forms on the basis of media,<br />

providing a philosophical foundation for distinctions that resonate throughout literary and<br />

visual studies. Investigation of the distinct properties of media was formulated most<br />

famously in the modern era by Gotthold Lessing (Laocoon, 1766). The value judgments<br />

that lurk in his distinctions continue <strong>to</strong> surface in disciplinary and critical activity <strong>to</strong> the<br />

present day. Such boundaries are well policed. But "new media" challenge these<br />

distinctions through the use of meta-technologies and inter-media sensibility. In addition,<br />

artistic practices forged from conceptual, procedural, and computational realms can't be<br />

well accommodated by aesthetic structures with "purist" underpinnings. If a data file can<br />

be input from a typewriter keyboard and output as musical notes then the idea of the<br />

"purity" of media seems irrelevant.<br />

Critics trained in or focused on the modern tradition (in its twentieth-century form and<br />

reaching back in<strong>to</strong> eighteenth-century aesthetics) have difficulty letting go of the


longstanding distinction between textual and visual forms of representation – as well as<br />

of the hierarchy that places text above image. The disjunct between literary and visual<br />

modernism, the very premise of an au<strong>to</strong>nomous visuality freed of literary and textual<br />

referents, continues <strong>to</strong> position these approaches <strong>to</strong> knowledge representation within<br />

separate domains. The consequences are profound. Intellectual training in the <strong>humanities</strong><br />

only rarely includes skills in interpretation of images or media in any but the most<br />

thematic or iconographic terms. The idea that visual representation has the capacity <strong>to</strong><br />

serve as a primary <strong>to</strong>ol of knowledge production is an almost foreign notion <strong>to</strong> most<br />

humanists. Add <strong>to</strong> this that many latter-day formalists conceive of <strong>digital</strong> objects as<br />

"immaterial" and the complicated legacy of hierarchical aesthetics becomes a very real<br />

obstacle <strong>to</strong> be overcome. Naivety aside (<strong>digital</strong> artifacts are highly, complexly material),<br />

these habits of thought work against conceptualizing a visual approach <strong>to</strong> <strong>digital</strong><br />

<strong>humanities</strong>. Nonetheless, visualization <strong>to</strong>ols have long been a part of the analysis of<br />

statistical methods in <strong>digital</strong> <strong>humanities</strong>. So long as these are kept in their place, a<br />

secondary and subservient "display" of information, their dubious character is at least<br />

held in check. Other conceptual difficulties arise when visual interfaces are used <strong>to</strong><br />

create, rather than merely present, information.<br />

In a ideologically grounded aesthetics, forms of creative expression are unders<strong>to</strong>od <strong>to</strong><br />

participate in spiritual evolution, moral improvement – or its opposite.<br />

Whether staged as cultural or individual improvements in character through exposure <strong>to</strong><br />

the "best that has been thought" embodied in the artifacts of high, fine art, the idea<br />

lingers: the arts, visual, musical, or poetical, somehow contribute <strong>to</strong> moral improvement.<br />

Samuel Taylor Coleridge, Matthew Arnold, and Walter Pater all reinforced this sermon<br />

on moral uplift in the nineteenth century. Even now the <strong>humanities</strong> and fine arts often<br />

find themselves justified on these grounds. The links among ideas of progress and the<br />

application of "<strong>digital</strong>" technology <strong>to</strong> <strong>humanities</strong> continue <strong>to</strong> be plagued by pernicious<br />

notions of improvement.<br />

Hegel wrote a fine script for the progressive his<strong>to</strong>ry of aesthetic forms. The cultural<br />

authority of technology is insidiously bound <strong>to</strong> such teleologies – especially when it<br />

becomes interlinked with the place granted <strong>to</strong> instrumental rationality in modern culture.<br />

The speculative approach, which interjects a long-repressed streak of subjective<br />

ambiguity, threatens the idea that <strong>digital</strong> representations present a perfect match of idea<br />

and form.<br />

But Hegel's dialectic provides a benefit. It reorients our understanding of aesthetic form,<br />

pivoting it away from the classical conception of static, fixed ideal. The interaction of<br />

thesis and antithesis in Hegelian principles provides a dynamic basis for thinking about<br />

transformation and change – but within a structure of progress <strong>to</strong>wards an Absolute.<br />

Hegel believed that art was concerned "with the liberation of the mind and spirit from the<br />

content and forms of finitude" (Hegel 1975) that would compensate for the "bitter labour<br />

of knowledge" (ibid.). Aesthetic experience, presumably, follows this visionary path. If<br />

an aesthetic mode could manage <strong>to</strong> manifest ideal thought, presumably in the form of<br />

"pure data" – and give it perfect form through technological means, then descriptive


aesthetics would have in its sights a sense of teleological completion. Mind, reason,<br />

aesthetic expression – all would align. But evidence that humankind has reached a<br />

pinnacle of spiritual perfection in this barely new millennium is in short supply. In this<br />

era following post-structuralism, and influenced by a generation of deconstruction and<br />

post-colonial theory, we can't really still imagine we are making "progress." Still, the idea<br />

that <strong>digital</strong> technology provides a high point of human intelligence, or other<br />

characterizations of its capabilities in superlative terms, persists.<br />

Eighteenth-century aestheticians placed attention on the nature of subjective experience,<br />

rather than remaining focused on standards of harmonious perfection for objectified<br />

forms. In discussions of taste, subjective opinion comes <strong>to</strong> the fore. Well suited <strong>to</strong> an era<br />

of careful cultivation of elite sensibility, this discussion of taste and refinement<br />

emphasizes the idea of expertise. Connoisseurship is the epi<strong>to</strong>me of knowledge created<br />

through systematic refining of sensation. Alexander Baumgarten sought in aesthetic<br />

experience the perfection proper <strong>to</strong> thought. He conceived that the object of aesthetics<br />

was "<strong>to</strong> analyse the faculty of knowledge" (Baumgarten 1735, sections 115–16;<br />

Beardsley 1966: 157) or "<strong>to</strong> investigate the kind of perfection proper <strong>to</strong> perception which<br />

is a lower level of cognition but au<strong>to</strong>nomous and possessed of its own laws" (ibid.). The<br />

final phrase resonates profoundly, granting aesthetics a substantive, rather than trivial,<br />

role. But aesthetic sensibilities – and objects – were distinguished from those of tecbne or<br />

utility. The class divide of laborer and intellectual aesthete is reinforced in this<br />

distinction. The legacy of this attitude persists most perniciously, and the idea of the<br />

aesthetic function of utilitarian objects is as bracketed in <strong>digital</strong> environments as it is in<br />

the well-marked domains of applied and "pure" arts.<br />

In what is arguably the most influential work in modern aesthetics, Immanuel Kant<br />

elevated the role of aesthetics – but at a price. The Critique of Judgment (1790), known<br />

as the "third" critique – since it bridged the first and second critiques of Pure Reason<br />

(knowledge) and Practical Reason (sensation) – contained an outline of aesthetics as the<br />

understanding of design, order, and form. But this understanding was meant as the<br />

apperception of "Purposiveness without purpose." In other words, appreciation of design<br />

outside of utility was the goal of aesthetics. Knowledge seeking must be "free",<br />

disinterested, without end, aim, he asserted. In his system of three modes of<br />

consciousness – knowledge, desire, and feeling (Pure Reason, Practical Reason, and<br />

Judgment) – Kant positioned aesthetics between knowledge and desire, between pure and<br />

practical reasons. Aesthetic judgment served as a bridge between mind and sense. But<br />

what about the function of emergent and participa<strong>to</strong>ry subjectivity? Subjectivity that<br />

affects the system of judgment? These are alien notions. For the Enlightenment thinker,<br />

the objects under observation and the mind of the observer interact from au<strong>to</strong>nomous<br />

realms of stability. We cannot look <strong>to</strong> Kant for anticipation of a "quantum" aesthetics in<br />

which conditions exist in indeterminacy until intervened by a participa<strong>to</strong>ry sentience.<br />

In summary, we can see that traditional aesthetics bequeaths intellectual parameters on<br />

which we can distinguish degrees of truth, imitation, refinement, taste, and even the<br />

"special powers of each medium" that are contributing strains <strong>to</strong> understanding the


knowledge-production aspects of visual aesthetics in <strong>digital</strong> media. But none provide a<br />

foundation for a generative approach, let alone a quantum and/or speculative one.<br />

Why?<br />

For all their differences these approaches share a common characteristic. They are all<br />

descriptive systems. They assume that form pre-exists the act of apprehension, that<br />

aesthetic objects are independent of subjective perception – and vice versa. They assume<br />

stable, static relations between knowledge and its representation – even if epistemes<br />

change (e.g., Hegel's dialectical forms evolve, but they do not depend on contingent<br />

circumstances of apperception in order <strong>to</strong> come in<strong>to</strong> being). The very foundations of<br />

<strong>digital</strong> media, however, are procedural, generative, and iterative in ways that bring these<br />

issues <strong>to</strong> the fore. We can transfer the insights gleaned from our understanding of <strong>digital</strong><br />

artifacts on<strong>to</strong> traditional documents – and we should – just as similar insights could have<br />

arisen from non-<strong>digital</strong> practices. The speculative approach is not specific <strong>to</strong> <strong>digital</strong><br />

practices – nor are generative methods. Both, however, are premised very differently<br />

from that of formal, rational, empirical, or classical aesthetics.<br />

Generative Aesthetics<br />

The Jewish ecstatic traditions of gematria (a method of interpretation of letter patterns in<br />

sacred texts) and Kabala (with its inducement of trance conditions through repetitive<br />

cornbina<strong>to</strong>ric meditation) provide precedents for enacting and understanding generative<br />

practices. Secular literary and artistic traditions have also drawn on permutational,<br />

cornbina<strong>to</strong>ric, and other programmed means of production. Alea<strong>to</strong>ry procedures<br />

(seemingly at odds with formal constraints of a "program" but incorporated in<strong>to</strong><br />

instructions and procedures) have been used <strong>to</strong> generate imaginative and aesthetic works<br />

for more than a century, in accord with the enigmatic cautions uttered by Stephane<br />

Mallarme that the "throw of the dice would never abolish chance." In each of the domains<br />

just cited, aesthetic production engages with non-rational systems of thought – whether<br />

mystical, heretical, secular, or irreverent. Among these, twentieth-century developments<br />

in generative aesthetics have a specific place and relevance for <strong>digital</strong> <strong>humanities</strong>.<br />

"Generative aesthetics" is the phrase used by the German mathematician and visual poet<br />

Max Sense <strong>to</strong> designate works created using algorithmic processes and computational<br />

means for their production. "The system of generative aesthetics aims at a numerical and<br />

operational description of characteristics of aesthetic structures", Sense wrote in his<br />

prescription for a generative aesthetics in the early 1960s (Sense 1971: 57). Algorithmically<br />

generated computations would give rise <strong>to</strong> datasets in turn expressed in visual or<br />

other output displays. Sense's formalist bias is evident. He focused on the description of<br />

formal properties of visual images, looking for a match between their appearance and the<br />

instructions that bring them in<strong>to</strong> being. They were evidence of the elegance and formal<br />

beauty of algorithms. They also demonstrated the ability of a "machine" <strong>to</strong> produce<br />

aesthetically harmonious images. His rational, mathematical disposition succeeded in<br />

further distancing subjectivity from art, suggesting that form exists independent of any<br />

viewer or artist. Sense's systematic means preclude subjectivity. But his essay marks an


important miles<strong>to</strong>ne in the his<strong>to</strong>ry of aesthetics, articulating as it does a procedural<br />

approach <strong>to</strong> form giving that is compatible with computational methods.<br />

Whether such work has any capacity <strong>to</strong> become emergent at a level beyond the<br />

programmed processes of its original conception is another question. Procedural<br />

approaches are limited because they focus on calculation (manipulation of quantifiable<br />

parameters) rather than symbolic properties of computing (manipulation at the level of<br />

represented information), thus remaining mechanistic in conception and execution.<br />

Reconceptualizing the mathematical premises of cornbina<strong>to</strong>ric and permutational<br />

processes so they work at the level of symbolic, even semantic and expressive levels, is<br />

crucial <strong>to</strong> the extension of generative aesthetics in<strong>to</strong> speculative, pataphysical, or<br />

quantum approaches.<br />

Generative aesthetics has a different lineage than that of traditional aesthetics. Here the<br />

key points of reference would not be Baumgarten and Kant, Hegel and Walter Pater,<br />

Roger Fry, Clive Bell, or Theodor Adorno – but the generative morphology of the fifthcentury<br />

bc Sanskrit grammarian Panini, the rational calculus of Leibniz, the visionary<br />

work of Charles Babbage, George Boole, Alan Turing, Herbert Simon, and Marvin<br />

Minsky. Other important contributions come from the traditions of self-consciously<br />

procedural poetics and art such as that of Lautreamont, Duchamp, Cage, Lewitt, Maciunas,<br />

S<strong>to</strong>ckhausen, and so on. The keyword vocabulary in this approach would not consist<br />

of beauty, truth, mimesis, taste, and form – but of emergent, au<strong>to</strong>poietic, generative,<br />

iterative, algorithmic, speculative, and so on.<br />

The intellectual tradition of generative aesthetics inspired artists working in conceptual<br />

and procedural approaches throughout the twentieth century. Earlier precedents can be<br />

found, but Dada strategies of composition made chance operations a crucial element of<br />

poetic and visual art. The working methods of Marcel Duchamp provide ample testimony<br />

<strong>to</strong> the success of this experiment. Duchamp's exemplary "unfinished" piece, The Large<br />

Glass, records a sequence of representations created through actions put in<strong>to</strong> play <strong>to</strong><br />

create tangible traces of abstract thought. Duchamp precipitated form from such activity<br />

in<strong>to</strong> material residue, rather than addressing the formal parameters of artistic form-giving<br />

according <strong>to</strong> traditional notions of beauty (proportion, harmony, or truth, for instance).<br />

He marked a radical departure from even the innovative visual means of earlier avantgarde<br />

visual traditions (Post-Impressionism, Cubism, Futurism, and so forth). For all their<br />

conceptual invention, these were still bound up with visual styles.<br />

In the 1960s and 1970s, many works characteristic of these approaches were instructionbased.<br />

Composer John Cage made extensive use of chance operations, establishing his<br />

visual scores as points of departure for improvisational response, rather than as<br />

prescriptive guidelines for replication of ideal forms of musical works. Fluxus artists such<br />

as George Brecht, George Maciunas, Robert Whitman, or Alison Knowles drew on some<br />

of the conceptual parameters invoked by Dada artists a generation earlier.<br />

The decades of the 1950s and 1960s are peopled with individuals prone <strong>to</strong> such inspired<br />

imaginings: Herbert Franke and Melvin Pruitt, Jascia Reichardt, and the heterogeneous


esearch teams at Bell Labs such as Kenneth Knowl<strong>to</strong>n, Leon Harmon, and dozens of<br />

other artists, worked in robotics, electronics, video, visual and audio signal processing, or<br />

the use of new technology that engaged combina<strong>to</strong>ric or permutational methods for<br />

production of poetry, prose, music, or other works. The legacy of this work remains<br />

active. Digital art-making exists in all disciplines and genres, often hybridized with<br />

traditional approaches in ways that integrate procedural methods and material production.<br />

One of the most sustained and significant projects in this spirit is Harold Cohen's Aaron<br />

project. As a demonstration of artificial aesthetics, an attempt <strong>to</strong> encode artistic creativity<br />

in several levels of instructions, Aaron is a highly developed instance of generative work.<br />

Aaron was first conceived in 1973, and not surprisingly its first iteration corresponded <strong>to</strong><br />

artificial vision research at the time. The conviction that perceptual processes, if<br />

sufficiently unders<strong>to</strong>od, would provide a basis for computational models predominated in<br />

research done by such pioneers as David Marr in the 1970s. Only as this work progressed<br />

did researchers realize that perceptual processing of visual information had <strong>to</strong> be<br />

accompanied by higher-order cognitive representations. Merely understanding<br />

"perception" was inadequate. Cognitive schemata possessed of the capacity for emerging<br />

complexity must also be fac<strong>to</strong>red in<strong>to</strong> the explanation of the way vision worked.<br />

Aaron reached a temporary impasse when it became clear that the methods of generating<br />

shape and form within its programs had <strong>to</strong> be informed by such world-based knowledge<br />

as the fact that tree trunks were thicker at the bot<strong>to</strong>m than at the <strong>to</strong>p. Vision, cognition,<br />

and representation were all engaged in a dialogue of percepts and concepts. Programming<br />

these in<strong>to</strong> Aaron's operation pushed the project <strong>to</strong>wards increasingly sophisticated AI<br />

research. Aaron did not simulate sensory perceptual processing (with its own complex<br />

mechanisms of sorting, classifying, actively seeking stimuli as well as responding <strong>to</strong><br />

them), but the cognitive representations of "intellectualized" knowledge about visual<br />

forms and their production developed for Aaron made a dramatic demonstration of<br />

generative aesthetics. Aaron was designed <strong>to</strong> create original expressive artifacts – new<br />

works of art. Because such projects have come in<strong>to</strong> being as generative machines before<br />

our very eyes, through well-recorded stages, they have shown us more and more<br />

precisely just how that constitutive activity of cognitive function can be conceived.<br />

'Pataphysical Sensibilities and Quantum Methods<br />

Before returning <strong>to</strong> speculative computing, and <strong>to</strong> the case study of this chapter, a note<br />

about 'pataphysics is in order. I introduced 'pataphysics almost in passing in the<br />

introduction above, not <strong>to</strong> diminish the impact of slipping this peculiar gorilla in<strong>to</strong> the<br />

chorus, but because I want <strong>to</strong> suggest that it offers an imaginative fillip <strong>to</strong> speculative<br />

computing, rather than the other way around.<br />

An invention of the late nineteenth-century French physicist poet Alfred Jarry,<br />

'pataphysics is a science of unique solutions, of exceptions. 'Pataphysics celebrates the<br />

idiosyncratic and particular within the world of phenomena, thus providing a framework<br />

for an aesthetics of specificity within generative practice. (This contrasts with Sense's


generative approach, which appears content with generalities of conception and formal<br />

execution.)<br />

The original founder of 'pataphysics, Alfred Jarry, declared the principles for the new<br />

science in the fantastic pages of his novel Dr Faustroll, 'Patapbysicien: "Faustroll<br />

defined the universe as that which is the exception <strong>to</strong> oneself." In his introduction <strong>to</strong> Dr<br />

Faustroll Roger Shattuck described the three basic principles of Jarry's belief system:<br />

clinamen, syzygy, and ethernity. Shattuck wrote: "Clinamen, an infinitesimal and<br />

fortui<strong>to</strong>us swerve in the motion of an a<strong>to</strong>m, formed the basis of Lucretius's theory of<br />

matter and was invoked by Lord Kelvin when he proposed his 'kinetic theory of matter.'<br />

To Jarry in 1898 it signified the very principle of creation, of reality as an exception<br />

rather than the rule." Just as Jarry was proposing this suggestive reconceptualization of<br />

physics, his contemporary Stephane Mallarme was calling the bluff on the end game <strong>to</strong><br />

metaphysics. Peter Burger suggests that Mallarme's conception of "the absolute"<br />

coincides with a conception of aesthetic pleasure conceived of as a technological game,<br />

driven by a non-existent mechanism. The substantive manifestation in poetic form shows<br />

the workings of the mechanism as it enacts, unfolds. Generative and speculative<br />

aesthetics are anticipated in the conceptualization of Mallarme's approach.<br />

What has any of this <strong>to</strong> do with computing?<br />

Without 'pataphysical and speculative capabilities, instrumental reason locks computing<br />

in<strong>to</strong> engineering problem-solving logical sensibility, programs that only work within the<br />

already defined parameters. The binarism between reason and its opposite, endemic <strong>to</strong><br />

Western thought, founds scientific inquiry in<strong>to</strong> truth on an empirical method. Pledged <strong>to</strong><br />

rational systematic consistency, this binarism finds an unambiguous articulation in Book<br />

X of Pla<strong>to</strong>'s Republic. "The better part of the soul is that which trusts <strong>to</strong> measure and<br />

calculation." The poet and visual artist "implant an evil constitution" – indulging the<br />

"irrational nature" which is "very far removed from the true." Ancient words, they<br />

prejudice the current condition in which the cultural authority of the computer derives<br />

from its relation <strong>to</strong> symbolic logic at the expense of those inventions and sensibilities that<br />

characterize imaginative thought. By contrast, speculative approaches seek <strong>to</strong> create<br />

parameter-shifting, open-ended, inventive capabilities – humanistic and imaginative by<br />

nature and disposition. Quantum methods extend these principles. Simply stated,<br />

quantum interpretation notes that all situations are in a condition of indeterminacy<br />

distributed across a range of probability until they are intervened by observation. The<br />

goal of 'pataphysical and speculative computing is <strong>to</strong> keep <strong>digital</strong> <strong>humanities</strong> from falling<br />

in<strong>to</strong> mere technical application of standard practices (either administered/info<br />

management or engineering/statistical calculations). To do so requires finding ways <strong>to</strong><br />

implement imaginative operations.<br />

Speculative Computing and the Use of Aesthetic<br />

Provocation


Visual or graphic design has played almost no part in <strong>humanities</strong> computing, except for<br />

the organized display of already structured information. Why should this be necessary?<br />

Or continue <strong>to</strong> be true? What are the possibilities of integrating subjective perspectives<br />

in<strong>to</strong> the process of <strong>digital</strong> <strong>humanities</strong>. And though emergent systems for dynamic<br />

interface are not realizable, they are certainly conceivable. Such perspectives differentiate<br />

speculative approaches from generative ones.<br />

The attitude that pervades information design as a field is almost entirely subsumed by<br />

notions that data pre-exist display, and that the task of visual form-giving is merely <strong>to</strong><br />

turn a cognitive exercise in<strong>to</strong> a perceptual one. While the value of intelligent information<br />

design in the interpretation of statistical data can't be overestimated, and dismissing the<br />

importance of this activity would be ridiculous, the limits of this approach also have <strong>to</strong> be<br />

pointed out. Why? Because they circumscribe the condition of knowledge in their<br />

apparent suggestion that information exists independently of visual presentation and just<br />

waits for the "best" form in which it can be represented. Many of the <strong>digital</strong> humanists<br />

I've encountered treat graphic design as a kind of accessorizing exercise, a dressing-up of<br />

information for public presentation after the real work of analysis has been put in<strong>to</strong> the<br />

content model, data structure, or processing algorithm. Arguing against this attitude<br />

requires rethinking of the way embodiment gives rise <strong>to</strong> information in a primary sense.<br />

It also requires recognition that embodiment is not a static or objective process, but one<br />

that is dynamic and subjective.<br />

Speculative computing is a technical term, fully compatible with the mechanistic reason<br />

of technological operations. It refers <strong>to</strong> the anticipation of probable outcomes along<br />

possible forward branches in the processing of data. Speculation is used <strong>to</strong> maximize<br />

efficient performance. By calculating the most likely next steps, it speeds up processing.<br />

Unused paths are discarded as new possibilities are calculated. Speculation doesn't<br />

eliminate options, but, as in any instance of gambling, the process weights the likelihood<br />

of one path over another in advance of its occurrence. Speculation is a mathematical<br />

operation unrelated <strong>to</strong> metaphysics or narrative theory, grounded in probability and<br />

statistical assessments. Logic-based, and quantitative, the process is pure tecbne, applied<br />

knowledge, highly crafted, and utterly remote from any notion of poiesis or aesthetic<br />

expression. Metaphorically, speculation invokes notions of possible worlds spiraling<br />

outward from every node in the processing chain, vivid as the rings of radio signals in the<br />

old RKO studios film logo. To a narra<strong>to</strong>logist, the process suggests the garden of forking<br />

paths, a way <strong>to</strong> read computing as a tale structured by nodes and branches.<br />

The phrase "speculative computing" resonates with suggestive possibilities, conjuring<br />

images of unlikely outcomes and surprise events, imaginative leaps across the circuits<br />

that comprise the electronic synapses of <strong>digital</strong> technology. The question that hangs in<br />

that splendid interval is a fundamental one for many areas of computing application: can<br />

the logic-based procedures of computational method be used <strong>to</strong> produce an aesthetic<br />

provocation? We know, of course, that the logic of computing methods does not in any<br />

way preclude their being used for illogical ends – or for the processing of information<br />

that is unsystematic, silly, trivial, or in any other way outside the bounds of logical<br />

function. Very few fully logical or formally systematic forms of knowledge exist in


human thought beyond those few branches of mathematics or calculation grounded in<br />

unambiguous procedures. Can speculation engage these formalized models of human<br />

imagination at the level of computational processing? To include an intuitive site for<br />

processing subjective interpretation in<strong>to</strong> formal means rather than circumscribing it from<br />

the outset? If so, what might those outcomes look like and suggest <strong>to</strong> the <strong>humanities</strong><br />

scholar engaged with the use of <strong>digital</strong> <strong>to</strong>ols? Does the computer have the capacity <strong>to</strong><br />

generate a provocative aesthetic artifact?<br />

Speculative computing extends the recognition that interpretation takes place from inside<br />

a system, rather than from outside. Speculative approaches make it possible for subjective<br />

interpretation <strong>to</strong> have a role in shaping the processes, not just the structures, of <strong>digital</strong><br />

<strong>humanities</strong>. When this occurs, outcomes go beyond descriptive, generative, or predictive<br />

approaches <strong>to</strong> become speculative. New knowledge can be created.<br />

These are big claims. Can they be substantiated?<br />

Temporal Modeling 1<br />

Temporal Modeling is a time machine for <strong>humanities</strong> computing. Not only does it take<br />

time and the temporal relations inherent in <strong>humanities</strong> data as its computational and<br />

aesthetic domain, enabling the production and manipulation of elaborate, subjectively<br />

inflected timelines, but it also allows its users <strong>to</strong> intervene in and alter the conventional<br />

interpretative sequence of visual thinking in <strong>digital</strong> <strong>humanities</strong>.<br />

The Temporal Modeling environment, under ongoing development at SpecLab<br />

(University of Virginia), embodies a reversal of the increasingly familiar practice of<br />

generating visualizations algorithmically from marked or structured data, data that have<br />

already been modeled and made <strong>to</strong> conform <strong>to</strong> a logical system. The aesthetic<br />

provocations Johanna Drucker describes are most typically unders<strong>to</strong>od <strong>to</strong> exist at the<br />

edges or termini of <strong>humanities</strong> computing projects. These are the graphs and charts we<br />

generate from large bodies of data according <strong>to</strong> strict, pre-defined procedures for<br />

knowledge representation, and which often enchant us with their ability <strong>to</strong> reveal hidden<br />

patterns and augment our understanding of encoded material. They are, however,<br />

fundamentally static and (as they depend on structured data and defined constraints)<br />

predictable, and we are hard-pressed <strong>to</strong> argue that they instantiate any truly new<br />

perspective on the data they reflect. Why, given the fresh possibilities for graphesis the<br />

computer affords, should we be content with an after-the-fact analysis of algorithmically<br />

produced representations alone? Temporal Modeling suggests a new ordering of aesthetic<br />

provocation, algorithmic process, and hermen-eutic understanding in the work of <strong>digital</strong><br />

<strong>humanities</strong>, a methodological reversal which makes visualization a procedure rather than<br />

a product and integrates interpretation in<strong>to</strong> digitization in a concrete way.<br />

How concrete? The Temporal Modeling <strong>to</strong>ols incorporate an intuitive kind of sketching –<br />

within a loosely constrained but highly defined visual environment – in<strong>to</strong> the earliest<br />

phases of content modeling, thereby letting visualization drive the intellectual work of<br />

data organization and interpretation in the context of temporal relations. Aesthetic


provocation becomes dynamic, part of a complex dialogue in which the user is required<br />

<strong>to</strong> respond <strong>to</strong> visualizations in kind. Response in kind, that is, in the visual language of<br />

the Temporal Modeling <strong>to</strong>olset, opens up new ways of thinking about <strong>digital</strong> objects,<br />

about the relation of image <strong>to</strong> information, and about the subjective position of any<br />

interpreter within a seemingly logical or analytic system. Our chief innovation is the<br />

translation of user gestures and image-orderings that arise from this iterative dialogue<br />

in<strong>to</strong> an accurate and expressive XML schema, which can be exported <strong>to</strong> other systems,<br />

transformed using XSLT, and even employed as a document type definition (DTD) in<br />

conventional data-markup practices. The sketching or composition environment in which<br />

this rich data capture takes place (the Temporal Modeling PlaySpace) is closely wedded<br />

<strong>to</strong> a sister-environment, the DisplaySpace. There, we provide a set of filters and<br />

interactive <strong>to</strong>ols for the manipulation and display of more familiar, algorithmically<br />

generated visualizations, derivative from PlaySpace schemata or the already-encoded<br />

data structures of established <strong>humanities</strong> computing projects. Like the PlaySpace, though,<br />

the Temporal Modeling DisplaySpace emphasizes the flux and subjectivity common <strong>to</strong><br />

both our human perception of time and our facility for interpretation in the <strong>humanities</strong>.<br />

We have not rejected display in favor of the playful engagement our composition<br />

environment fosters; instead, we hope <strong>to</strong> show that a new, procedural understanding of<br />

graphic knowledge enhances and even transfigures visualization in the older modes.<br />

Our work in building the PlaySpace, with which we began the project in the Summer of<br />

<strong>2001</strong> and which now nears completion, has required a constant articulation of its<br />

distinction from the DisplaySpace – the implementation of which forms the next phase of<br />

Temporal Modeling. What quality of appearance or use distinguishes a display <strong>to</strong>ol from<br />

an editing <strong>to</strong>ol? At their heart, the mechanisms and processes of the PlaySpace are bound<br />

up in: the positioning of temporal objects (such as events, intervals, and points in time)<br />

on the axis of a timeline; the labeling of those objects using text, color, size, and quality;<br />

the relation of objects <strong>to</strong> specific temporal granularities (the standards by which we mark<br />

hours, seasons, aeons); and, in complex interaction, the relation of objects <strong>to</strong> each other.<br />

Each of these interpretative actions – the specification of objects and orderings, their<br />

explication and interrelation – additionally involves a practice we designate inflection.<br />

Inflection is the graphic manifestation of subjective and interpretative positioning <strong>to</strong>ward<br />

a temporal object or (in a sometimes startling display of warping and adjustment) <strong>to</strong> a<br />

region of time. This positioning can be on the part of the prime interpreter, the user of the<br />

PlaySpace, or inflections can be employed <strong>to</strong> represent and theorize external<br />

subjectivities: the implied interpretative standpoint of a character in a work of fiction, for<br />

instance, or of an his<strong>to</strong>rical figure, movement, or Zeitgeist. The energies of the PlaySpace<br />

are all bound up in enabling understanding through iterative visual construction in an<br />

editing environment that implies infinite visual breadth and depth. In contrast, the<br />

DisplaySpace channels energy in<strong>to</strong> iterative visual reflection by providing a responsive,<br />

richly-layered surface in which subjectivity and inflection in temporal relations are not<br />

fashioned but may be reconfigured.<br />

I want <strong>to</strong> focus here on some specific qualities and <strong>to</strong>ols of Temporal Modeling,<br />

especially as they relate <strong>to</strong> the embeddedness of subjectivity, uncertainty, and<br />

interpretation in every act of representation, which we take as a special province of


speculative computing. Our very process of design self-consciously embodies this<br />

orientation <strong>to</strong>ward information and software engineering. We make every effort <strong>to</strong> work<br />

from imagery as much as from on<strong>to</strong>logy, coupling our research efforts in the philosophy<br />

and data-driven classification of temporal relations with the intuitive and experimental<br />

work of artists of whom we asked questions such as: "What does a slow day look like?"<br />

or "How might you paint anticipation or regret?" As our underlying architecture became<br />

more stable and we began <strong>to</strong> assemble a preliminary notation system for temporal objects<br />

and inflections, we made a practice of asking of each sketch we floated, "What does this<br />

imply?" and "What relationships might it express?" No visual impulse was dismissed out<br />

of hand; instead, we retained each evocative image, frequently finding use for it later,<br />

when our iterative process of development had revealed more about its implications in<br />

context.<br />

In this way, the necessity of a special feature of the Temporal Modeling Project was<br />

impressed on us: a capacity for expansion and adjustment. The objects, actions, and<br />

relations defined by our schemata and programming are not married inextricably with<br />

certain graphics and on-screen animations or display modes. Just as we have provided<br />

<strong>to</strong>ols for captioning and coloring (and the ability <strong>to</strong> regularize cus<strong>to</strong>m-made systems with<br />

legends and labels), we have also made possible the upload and substitution of user-made<br />

standard vec<strong>to</strong>r graphics (SVG) for the generic notation systems we've devised. This is<br />

more than mere window-dressing. Our intense methodological emphasis on the<br />

importance of visual understanding allows the substitution of a single set of graphics<br />

(representing inflections for, say, mood or foreshadowing) <strong>to</strong> alter radically the<br />

statements made possible by Temporal Modeling's loose grammar. Users are invited <strong>to</strong><br />

intervene in the interpretative processes enabled by our <strong>to</strong>ol almost at its root level.<br />

A similar sensibility governs the output of a session in the PlaySpace environment.<br />

PlaySpace visualizations consist of objects and inflections in relation <strong>to</strong> each other and<br />

(optionally) <strong>to</strong> the metric of one or more temporal axes. The editing process involves the<br />

placement and manipulation of these graphics on a series of user-generated, transparent<br />

layers, which enable groupings and operations on groups (such as zooms, granularity<br />

adjustments, panning and positioning, simple overlays, and changes in intensity or alphavalue)<br />

meant <strong>to</strong> enhance experimentation and iterative information design inside the<br />

visual field. When the user is satisfied that a particular on-screen configuration represents<br />

an understanding of his data worth preserving, he may elect <strong>to</strong> save his work as a model.<br />

This means that the PlaySpace will remember both the positioning of graphic notations<br />

on-screen and the underlying data model (in the form of an XML schema) that these<br />

positions express. This data model can then be exported and used elsewhere or even<br />

edited outside the PlaySpace and uploaded again for visual application. Most interesting<br />

is the way in which transparent editing layers function in the definition of PlaySpace<br />

models. The process of saving a model requires that the user identify those layers<br />

belonging <strong>to</strong> a particular, nameable interpretation of his material. This means that a single<br />

PlaySpace session (which can support the creation of as many layers as hardware<br />

limitations make feasible) might embed dozens of different interpretative models: some<br />

of which are radical departures from a norm; some of which differ from each other by a<br />

small, yet significant, margin; and some of which are old friends, imported in<strong>to</strong> the


PlaySpace from past sessions, from collections of instructional models representing<br />

conventional understandings of his<strong>to</strong>ry or fiction, or from the efforts of colleagues<br />

working in collaboration on research problems in time and temporal relations. A model is<br />

an interpretative expression of a particular dataset. More importantly, it is what the<br />

interpreter says it is at any given point in time. We find the flexibility inherent in this<br />

mode of operation akin <strong>to</strong> the intuitive and analytical work of the traditional <strong>humanities</strong><br />

at its best.<br />

Our policies of welcoming (and anticipating) the upload of user-designed graphic<br />

notation and of enforcing the formalization of interpretative models in the loosest terms<br />

possible are examples of Temporal Modeling's encouragement of hermeneutic practice in<br />

computational contexts. In some sense, this practice is still external <strong>to</strong> the visual<br />

environment we have built, even as it forms an integral part of the methodology<br />

Temporal Modeling is designed <strong>to</strong> reinforce. I wish <strong>to</strong> close here with a description of a<br />

new and exciting <strong>to</strong>ol for encoding interpretation and subjectivity within the designed<br />

Temporal Modeling environment: a mechanism we call the nowslider.<br />

"Nowsliding" is a neologism for a practice all of us do constantly – on which, in fact, our<br />

understanding of ourselves and our lives depends. Nowsliding is the subjective<br />

positioning of the self along a temporal axis and in relation <strong>to</strong> the points, events,<br />

intervals, and inflections through which we classify experience and make time<br />

meaningful. You nowslide when you picture your world at the present moment and, some<br />

ticks of the clock later, again at another ever-evolving present. You nowslide, <strong>to</strong>o, when<br />

you imagine and project the future or interpret and recall the past. Our <strong>to</strong>olset allows a<br />

graphic literaliza-tion of this subjective positioning and temporal imagining, in the shape<br />

of configurable, evolving timelines whose content and form at any given "moment" are<br />

dependent on the position of a sliding icon, representative of the subjective viewpoint.<br />

Multiple independent or interdependent points of view are possible within the context of<br />

a single set of data, and the visual quality of nowsliding may be specified in the<br />

construction of a particular model.<br />

At present, two display modes for the nowslider are in development. The first is a<br />

catastrophic mode, in which new axial iterations (or imagined past- and future-lines)<br />

spring in a tree structure from well-defined instances on a primary temporal axis. In this<br />

way, PlaySpace users can express the human tendency <strong>to</strong> re-evaluate the past or make<br />

predictions about the future in the face of sudden, perspective-altering events. New<br />

subjective positions on the primary axis of time (and new happenings) can provoke more<br />

iterations, which do not supplant past imaginings or interpretations, but rather co-exist<br />

with them, attached as they are <strong>to</strong> a different temporal locus. In this way, timelines are<br />

made <strong>to</strong> bristle with possibility, while still preserving a distinct chronology and single<br />

path. Our nowsliders also function in a continuous mode – distinct from catastrophism –<br />

in which past and future iterations fade in and out, change in position or quality, appear<br />

or disappear, all within the primary axis of the subjective viewpoint. No new lines are<br />

spawned; instead, this mode presents time as a continuum of interpretation, in which past<br />

and present are in constant flux and their shape and very content are dependent on the<br />

interpretative pressure of the now.


Our taking of temporal subjectivity and the shaping force of interpretation as the content<br />

and overarching theme of the PlaySpace and DisplaySpace environments we have built is<br />

meant <strong>to</strong> reinforce the goal of the Temporal Modeling <strong>to</strong>ols and, by extension, of<br />

speculative computing. Our goal is <strong>to</strong> place the hermeneut inside a visual and algorithmic<br />

system, where his or her very presence alters an otherwise mechanistic process at the<br />

quantum level. Humanists are already skilled at the abstract classification and encoding<br />

that data modeling requires. We understand algorithmic work and can appreciate the<br />

transformative and revela<strong>to</strong>ry power of visual and structural deformance. We at least<br />

think we know what <strong>to</strong> do with a picture or a graph. What we haven't yet tried in a<br />

rigorous and systematic way is the injection of the subjective positioning any act of<br />

interpretation both requires and embodies in<strong>to</strong> a computational, self-consciously visual<br />

environment. If speculative computing has a contribution <strong>to</strong> make <strong>to</strong> the methods and<br />

outcomes of <strong>digital</strong> <strong>humanities</strong>, this is it.<br />

See also Chapter 16: Marking Texts of Many Dimensions; Chapter 13: How the<br />

Computer Works<br />

Note<br />

1 This section was co-authored with Bethany Nowviskie.<br />

References for Further Reading<br />

Amelunxen, H. V., S. Iglhaut, and F. Rötzer, (eds.) (1996). Pho<strong>to</strong>graphy after<br />

Pho<strong>to</strong>graphy. Munich: G&B Arts.<br />

Baumgarten, A. (1735). Reflections on Poetry. London.<br />

Baumgarten, A. (1750). Aesthetica. London.<br />

Beardsley, M. (1966). Aesthetics. Tuscaloosa: University of Alabama Press.<br />

Bense, M. (1971). The Projects of Generative Aesthetics. In J. Reichardt (ed.),<br />

Cybernetics, Art and Ideas (pp. 57–60). New York: Graphics Society.<br />

Bök, C. (2002). Pataphysics: The Poetics of an Imaginary Science. Evans<strong>to</strong>n, IL:<br />

Northwestern University Press.<br />

Bürger, P. (1998). Mallarmé. In M. Kelly (ed.), Encyclopedia of Aesthetics (pp. 177–8).<br />

Oxford: Oxford University Press.<br />

Dunn, D., (ed.) (1992). Pioneers of Electronic Art. Santa Fe: Ars Electronica and The<br />

Vasulkas.


Engelbart, D. (1963). A Conceptual Framework for the Augmentation of Man's Intellect.<br />

In P. W. Hower<strong>to</strong>n and D. C. Weeks (eds.), The Augmentation of Man's Intellect by<br />

Machine, Vision in Information Handling, vol. 1. Washing<strong>to</strong>n, DC: Spartan Books.<br />

Franke, H. and H. S. Helbig (1993). Generative Mathematics: Mathematically Described<br />

and Calculated Visual Art. In M. Emmer (ed.), The Visual Mind. Cambridge, MA: MIT<br />

Press.<br />

Glazier, L. P. (2002). Digital Poetics: The Making of E-Poetics. Tuscaloosa, AL, and<br />

London: University of Alabama Press.<br />

Hegel, G. W. F. (1975). Aesthetics: Lectures on Fine Arts. Oxford: Oxford University<br />

Press. (Translated by T. M. Moore. Citation is from http://www.hegel.net/spirit/abs/art.).<br />

Hockey, S. (2000). Electronic Texts in the Humanities. Oxford and New York: Oxford<br />

University Press.<br />

Holtzman, S. (1994). Digital Mantras. Cambridge, MA: MIT Press.<br />

Jarry, A. (1996). Exploits and Opinions of Dr. Faustroll, Pataphysician. Bos<strong>to</strong>n: Exact<br />

Change.<br />

Lister, M., (ed.) (1995). The Pho<strong>to</strong>graphic Image in Digital Culture. London and New<br />

York: Routledge.<br />

Marr, D. (1982). Vision. New York: W. H. Freeman.<br />

McGann, J. J. (<strong>2001</strong>). Radiant Textuality. Basings<strong>to</strong>ke and New York: Palgrave.<br />

Prueitt, M. (1984). Art and the Computer. New York: McGraw-Hill.<br />

Renear, A. (1997). Out of Praxis: Three (Meta) Theories of Textuality. In K. Sutherland<br />

(ed.), Electronic Text. Oxford: Clarendon Press.<br />

Ritchin, F. (1990). In Our Own Image. New York: Aperture.<br />

Shanken, E. (n.d.). The House that Jack Built. At http://www.duke.edu/~giftwrap/,<br />

http:mitpres.mit.edu/e-journals/LEA/ARTICLES/jack.html.<br />

Sutherland, K., (ed.) (1997). Electronic Text. Oxford: Clarendon Press.<br />

Zhang, Y., L. Rauchwerger, and J. Torrellas (1999). Hardware for Speculative<br />

Reduction, Parallelization and Optimization in DSM Multiprocessors. Proceedings of the<br />

1st Workshop on Parallel Computing for Irregular Applications, 26. At<br />

http://citeseer.nj.nec.com/article/zhang99hardware.html.


Temporal Modeling is freely available at http://www.speculativecomputing.org, and is<br />

delivered <strong>to</strong> the Web using XML-enabled Macromedia Flash MX and Zope, an objec<strong>to</strong>riented<br />

open-source application server.<br />

30.<br />

Robotic Poetics<br />

William Winder<br />

Robots<br />

There is a fundamental link between language and robots. Whatever material constitution<br />

they may have – arms and legs, cogs and wheels, and engines – the crucial ingredients,<br />

the ones that separate robots from wrenches, are their instructions. Instructions are<br />

perhaps inscribed in some concrete medium, but they are quintessentially abstract,<br />

existing more in the netherworld of code than in concrete form. That code netherworld is<br />

where humans and robots inevitably meet. It is where humans are most robotic and where<br />

robots are most human.<br />

Such was Turing's reasoning that led him <strong>to</strong> choose the test of dialogue <strong>to</strong> detect human<br />

intelligence. It is no coincidence that a computer scientist of Turing's stature might come<br />

<strong>to</strong> the same conclusion as such foundational thinkers as C. S. Peirce and F. de Saussure.<br />

Each in his own way <strong>to</strong>ok the axiomatic position that there is no thought without signs,<br />

and Peirce, at least, comes <strong>to</strong> the conclusion that a person is in fact fundamentally a sign<br />

– and so <strong>to</strong>o, though inversely, for robots, which are best described as personified signs.<br />

Robots are instructions that take human form.<br />

Mechanized writing has been at the center of philosophical debate as far back as the<br />

invention of writing itself, but the humanists' version of the Turing test is relatively<br />

recent. It is, of course, not the conversational game between man and machine Turing<br />

described, but rather a literary version of the Turing test, where the goal is <strong>to</strong> simulate the<br />

author (see Kurzweil 2002, for a practical example). There is no doubt that mechanical<br />

processes underlie literary texts (Zholkovsky 1984: 53) and they have been used<br />

sporadically throughout his<strong>to</strong>ry as an explicit device of creation (Swift's literary engine,<br />

the Surrealists' cadavre exquis, Bryon Gyson and Burroughs's cut-ups). Yet only recently<br />

have mechanical processes become the principal focus of a literary movement, in the<br />

form of the (largely French) group OULIPO (ouvoir de littérature potentielle – Workshop<br />

for Potential Literature), which by a plagiat par anticipation (anticipa<strong>to</strong>ry plagiarism)<br />

<strong>to</strong>ok computational constraints as the source of their creations.<br />

Let us call robotic poetics (RP hereafter) the study of creative writing done by robots:<br />

Definition 1: RP is the study of robotic authors and the au<strong>to</strong>matic generation of creative<br />

texts.


RP has as many facets as au<strong>to</strong>matic writing has dimensions (see Fournel 1986).<br />

Generated text has a robotic author, itself created by a human programmer. There is a<br />

poetics of creating that author (here creativity lies in the writing of instructions); a poetics<br />

of generating a specific text (how those instructions play out <strong>to</strong> make a given text); and a<br />

poetics of reading generated literature (how the reader will read a text knowing, or not,<br />

that it is au<strong>to</strong>matically generated). We simply do not read a generated text the same way<br />

we read a one-off hand-crafted text. We read in the shadow of textual variants, those<br />

sibling texts that could have been generated just as easily: generated texts come in<br />

swarms.<br />

We will consider here RP as the field that covers all these facets of robotic creation,<br />

without narrowing the field <strong>to</strong> a single aspect. We will talk about robotic authors (the<br />

program that generates), robotic texts (generated texts), text engineers (those who<br />

program), and so on. We will perhaps focus narrowly here on creative natural language<br />

texts, but that should not be taken as RP's natural limit. What is central <strong>to</strong> RP is not text<br />

(which in any case should be taken as multimedia text), but rather printing, the central<br />

metaphor for any action a robot might take. Literally any medium can be printed on or<br />

imprinted. Thus CAD (computer-aided design) and CNC (computer numeric control) are<br />

used in fac<strong>to</strong>ries <strong>to</strong> "print" machine parts. What is central is the way the texts of<br />

instructions are au<strong>to</strong>matically transmuted in<strong>to</strong> something other, sometimes other texts,<br />

sometimes things.<br />

Textual generation systems have many applications in the <strong>humanities</strong>. Lessard and<br />

Levison offer a good example of the range of applications. Their generation system, Vinci<br />

(see below), has been used for foreign-language exercise generation, simulation of<br />

oulipian constraints, teaching linguistics, punning and humor, word and s<strong>to</strong>ry generation,<br />

among others. (See their site for documentation, download, and an extensive publication<br />

list.)<br />

Two red threads that RP follows are human creativity and combina<strong>to</strong>ry practices. These<br />

two dimensions are oddly opposed, yet intertwined. A combina<strong>to</strong>ry can be systematic or<br />

random, but its value is inevitably placed second <strong>to</strong> what is creative. A roll of the dice is<br />

not a creative act, since the outcome is within a foreseeable range of events. It is a<br />

distinctly human ability <strong>to</strong> creatively define the possible, <strong>to</strong> envisage possibilities outside<br />

the foreseeable. And yet we <strong>to</strong>o are combina<strong>to</strong>ry creatures. As Saussure and Peirce<br />

underline, because we think and live through signs, we must be signs (argues Peirce).<br />

What are signs? Nothing if not combina<strong>to</strong>ry conspira<strong>to</strong>rialness.<br />

Let us formulate a second possible definition of the field that covers the combina<strong>to</strong>rial<br />

dimension of robots:<br />

Definition 2:, RP is <strong>humanities</strong> combina<strong>to</strong>rial studies.<br />

RP is an interdisciplinary, interstitial field – a field that takes on a certain consistency in<br />

the overlap of more clearly defined disciplines such as cognitive science and<br />

computational linguistics (Anis 1995). That interdiscplinarity comes naturally through


computers, which are by nature interdisciplinary machines: anything can be digitized;<br />

everything – so it would seem – can be virtualized (see chapter 28, this volume) and thus<br />

converge in the computer.<br />

Potential and Real Texts<br />

Poetics is traditionally conceived as the science of message construction, a systematic<br />

study of texts that describes why and how they make sense or have a particular effect.<br />

Even in this traditional vein, poetics describes text generation, the ergonomics of texts<br />

and the aesthetics of combinations. RP pushes the notion one step further in that it studies<br />

what creates the mechanics of creativity. What is central in RP is the virtualized (or<br />

potentia-lized) dimension of a work of art; understanding virtual texts tells us much about<br />

real texts and their creative core. A robotic author is precisely the site of a corpus of<br />

virtual texts.<br />

Pierre Lévy (1998: 141–2) contrasts the virtual with the potential, the actual, and the real<br />

(see Figure 30.1). The north/south planes oppose what is inert (substances) and what is<br />

dynamic (events); the east/west, what is unrealized (potential, virtual) and realized (real,<br />

actual). The movement from north <strong>to</strong> south is described as "jectification" – transforming<br />

what is fixed in<strong>to</strong> a dynamic interaction between subjectified objects (money and<br />

language, for example, exist only through our collective subjectivity) and objectified<br />

subjects (the self is constructed by concrete choices of technological solution <strong>to</strong> problems<br />

– the subject can choose <strong>to</strong> be a blond or a brunette). The movement in the opposite<br />

direction, south <strong>to</strong> north, is <strong>to</strong>wards the fixation of predefined structures.<br />

A real text is a text instance that derives its primary value from its simple existence. It has<br />

a secondary value in that it can also be "reused" through potentialization <strong>to</strong> enhance a<br />

potential text, i.e. a range of possible texts. Transducer bots (my neologism; see below),<br />

which scramble real texts <strong>to</strong> form new texts, are a clear example of potentialization of<br />

real texts.<br />

A potential text represents a set of real texts, each of which is created through the reader's<br />

choice. Queneau's Cent mille milliards de poemes (One hundred trillion poems) is the<br />

classic example of this kind of text. For each of the 14 lines of a sonnet, Queneau offers<br />

10 substitute or variant lines. Since each of the 14 lines has 10 possible realizations and<br />

any combination of these choices is "poetically grammatical", the number of possible<br />

poems is 10 14 , or one hundred trillion.<br />

This (14 by 10) combina<strong>to</strong>rial poem is printed in book form. The book looks as if it has<br />

10 pages with 1 poem per page. However, the pages of this book are cut horizontally<br />

between the verses so that each verse can be turned separately. By turning partial pages in<br />

this way, verses on the pages below can be successively revealed and each distinct<br />

combina<strong>to</strong>ry sonnet can be brought <strong>to</strong> the fore (see Figure 30.2).


According <strong>to</strong> Queneau's calculations one would have <strong>to</strong> read 24 hours a day for 200<br />

millennia <strong>to</strong> read all the combinations; this is clearly a work that no one could ever<br />

possibly read.<br />

Queneau's text shows us how, in Lévy's terms, the potential is transformed in<strong>to</strong> the real.<br />

On the other hand a transducer bot represents the potentialization of a real text.<br />

Transducers take as input any text and produce another text, using various transcoding<br />

procedures. One example is the Shannonizer, which more or less randomly rearranges<br />

input text and intersperses it with canned expressions. The Shannonizer will transform<br />

this text (edited by a robotic Dr Seuss):<br />

Windows XP delivers a host of great features, including improved reliability and security,<br />

along with great experiences in real-time communications, mobile support, home<br />

networking, <strong>digital</strong> pho<strong>to</strong>s, <strong>digital</strong> video, and <strong>digital</strong> music. As a result of these higher<br />

standards, you may see some compatibility issues with your computer systems and<br />

software programs. In most cases, however, you can expect those problems <strong>to</strong> be at an<br />

absolute minimum. Here's why.<br />

(Windows XP "Compatibility", Internet advertisement)<br />

in<strong>to</strong> this text:<br />

Windows XP delivers as edited by Dr Seuss<br />

Windows XP delivers a host of great features, sir. Thank goodness I'm a result of sour<br />

gooseberry jam! Just look! No, <strong>digital</strong> pho<strong>to</strong>s, home networking, you may see some<br />

compatibility issues with great experiences in their paddles in real-time communications,<br />

we serve hot dogs, along with your computer systems and the dark. Take a star, funny<br />

things are everywhere. In most cases, however, along with great experiences in real-time<br />

communications, not one little bed all night.<br />

(The Shannonizer transforms a Windows XP advertisement)<br />

The Shannonizer potentializes in that it recreates new texts from old. A more<br />

sophisticated potentializer is Kurzweil's poetry genera<strong>to</strong>r ("Cyberart Technologies"),<br />

which "recycles" the poetry of any given corpus. (As Balpe remarks, poetry is the easiest<br />

genre <strong>to</strong> simulate – See Balpe and Magné 1991: 129.)<br />

Games, Dialogue, and Texts<br />

A potential text represents a field of possibilities defined by the text engineer. A given<br />

realization of the text is simply a selection, a roll of the dice, by the user. Potential texts<br />

are textual kaleidoscopes and the basis of such ancient textual games as Tarot, I Ching, or<br />

horoscope predictions.


There is a great range of potential texts, and many possible classifications (for example,<br />

Aarseth 1997: 58ff). The following represents a very simple classification with some<br />

pro<strong>to</strong>typical examples:<br />

1 Combina<strong>to</strong>ry texts or textual kaleidoscopes, such as hypertext, which generate<br />

variants or allow the reader <strong>to</strong> select variants: mail merge in word processors or<br />

Queneau's sonnets.<br />

2 MUDs (multi-user domains), which generate and manage a virtual universe for the<br />

interaction of several users. (See chapter 28, this volume.)<br />

3 Transducer bots, which rearrange text: the Shannonizer or Kurzweil's poetry engine.<br />

4 Genera<strong>to</strong>rs, such as Vinci (Lessard and Levison 2002) or KPML (Reiter 2002).<br />

5 Chatbots or Eliza variants, such as ALICE, which have Turing-style dialogues with<br />

the user.<br />

Such texts are at the same time text, game, and dialogue. Queneau's text requires that the<br />

reader select a poem by turning <strong>to</strong> the appropriate verses. The choice is not determined in<br />

advance, though the field of possibilities is. Games are in general dialogues between two<br />

or more players, though in some games, such as solitaire, the second player is reduced <strong>to</strong><br />

the mechanical combination that resists – the shuffle that gives the distribution of cards<br />

that the solitaire player must "beat."<br />

"Playtexts" (Motte 1995) have peculiar properties, though all can be represented as a<br />

sequence of signs, just as chess games have a notation for each move of a game. Lewis<br />

Carroll will thus reproduce the chess problem that informed Alice in Wonderland in the<br />

preface <strong>to</strong> the text itself. Chess offers the example of a now highly computerized game in<br />

which the robotic player routinely outstrips its designers (Aarseth 1997: 27). Whether<br />

authorship will be au<strong>to</strong>mated <strong>to</strong> the same degree remains a question. It is nevertheless<br />

useful at this point <strong>to</strong> consider more closely the general structure of games.<br />

A chess game is a combina<strong>to</strong>ry, but its potential nature is distributed in many layers of<br />

increasing complexity. A chess piece has no pertinent parts, but a board with its<br />

distribution of pieces does. Even so, we can just as easily consider each possible board an<br />

unanalyzed "piece" of a larger construct, the sequence of board configurations that make<br />

up a given game. We will recognize three very broad, relative layers of play: pre-play,<br />

play, and post-play.<br />

Pre-play (syntactic level) concerns the mastering of the fundamental rules of the game<br />

which define the playing field. A player who could in some way visualize all the possible<br />

board configurations would completely master this level. Even so, he or she would not<br />

necessarily win a single game, even against someone who could not visualize the entire<br />

field of possibilities.


Play (semantic level) concerns the relative value of a move and its fundamental effect or<br />

consequences. The players' vying is a kind of dialogue at a higher combina<strong>to</strong>rial level<br />

than pre-play, though it presupposes pre-play vying in the sense that each player is more<br />

or less expert at seeing potential moves. Play adds a layer of meaning since the choice of<br />

a complete game must be shared with the opponent. At the same time, play includes an<br />

unlimited number of possible sub-games. For example, in chess, the opening play is of<br />

itself a sub-game, as is the end game. Yet all the sub-games blend with the general goal<br />

of winning.<br />

Post-play (pragmatic level) concerns aesthetic sensibility; it is not the simple winning or<br />

losing of the game, but rather the way one wins or loses that becomes the game. Post-play<br />

can make the game mean just about anything: for example, out of deference, one player<br />

might seek <strong>to</strong> lose, perhaps <strong>to</strong> encourage the other player. At the post-play level the game<br />

opens on <strong>to</strong> any number of further games.<br />

Abstractly, the pre-play level in chess is always the same. However, a real player must<br />

devise a set of rules for seeing pre-play and play potential, and those heuristic rules can<br />

change and indeed should change as the player improves. At the moment, it would seem<br />

that programmers of chess machines, which do great things through brute search<br />

algorithms, have not yet addressed this further level of chess prowess, that rule-changing<br />

creativity that allows players <strong>to</strong> improve or, while playing chess, <strong>to</strong> play more than just<br />

chess.<br />

Reading a playtext involves vying at all these levels. Potential texts are like solitaire: they<br />

are only worthy adversaries at the pre-play level, or perhaps at the play level. The game<br />

s<strong>to</strong>ps with a win or loss for the computer. In human interaction, the game can be<br />

transformed in<strong>to</strong> just about any game the players wish; it becomes mostly post-play, or<br />

higher. That facile movement <strong>to</strong> higher levels of play is the ultimate focus of RP, both for<br />

understanding what robotic texts mean and how <strong>to</strong> build them.<br />

Virtual and Actualized Texts<br />

Many features not clearly associated with traditional print literature come <strong>to</strong> the forefront<br />

in the robotic text. All texts are minimally interactive, but a robotic text has concrete<br />

structures that are designed <strong>to</strong> maintain a more human-like feedback loop with the reader.<br />

Many special functions, while possible with a paper text, become central <strong>to</strong> the ethos of a<br />

robotic text. Clearly, time is one parameter that can be used in a novel way. Thus Fournel<br />

contrasts the computer implementation of Queneau's sonnets with the printed book:<br />

The printed collection [of Queneau's poems] is prettily conceived, but the manipulation<br />

of the strips on which each verse is printed is sometimes tedious. The computer, though,<br />

makes a selection in the corpus in function of the length of the "reader's" name and the<br />

time which he takes <strong>to</strong> type it in<strong>to</strong> the terminal, then prints the sonnet, which bears the<br />

double signature of Queneau and his reader.<br />

(Fournel 1986: 140)


This subjectification, as Lévy calls it, characterizes the move <strong>to</strong>wards the virtual text, (see<br />

Figure 30.1). Yet moving from the potential <strong>to</strong> the virtual is not simply a question of<br />

degree. Both virtual and potential texts are alike in that both are equally unreadable (as<br />

Queneau remarks above). Virtual texts have however a kind of open-ended creativity that<br />

would ultimately define success under the Turing test. A computerized version of<br />

Queneau's text simply does not have that degree of intelligence. Today's systems only<br />

approach asymp<strong>to</strong>tically the virtual text, which remains a challenge. Even so, many<br />

potential texts seem <strong>to</strong> "simulate" virtual texts (!).<br />

The virtual text remains the ultimate object of RP. How does one instill a robot with<br />

creative subjectivity? There are certainly many tantalizing steps that seem <strong>to</strong> inch the<br />

potential text <strong>to</strong>wards the virtual. The single most common step is <strong>to</strong> borrow the reader's<br />

subjectivity by implicating him or her in the construction of the text. This transfer of<br />

subjectivity is no doubt the source of the Eliza effect: readers fill in what the computer<br />

leaves blank. Though such sleights of hand concerning subjectivity hybridize (reader and<br />

text) rather than virtualize, they do offer a tantalizing mirage of what RP seeks:<br />

Definition 3:, RP is the poetics of virtualization and virtual text.<br />

At present, the only obvious path <strong>to</strong> virtualization is by degrees, through a dogged<br />

expansion of the computer's generative potential using ever more powerful techniques -<br />

such was the evolution of computer chess. It is useful therefore <strong>to</strong> look at some concrete<br />

examples of text generation and follow how the generative power of applications can be<br />

expanded through progressive generalization.<br />

Generative Applications: Prolog and the Tigerbil<br />

There are many technical dimensions associated with a generation project, and a<br />

confusing diversity of applications. Very little can be called standard in this area. Anyone<br />

wishing <strong>to</strong> become a text engineer and actually generate text will face a number of<br />

practical choices (operating systems, programming environment, grammar formalism,<br />

lexicon format, etc.), many of which have considerable theoretical implications.<br />

One of the major stumbling blocks in the field is the short life cycle of any software.<br />

Publications in computer-generated literature included, as late as 1986, pages of poorly<br />

printed code in DOS BASIC and were fairly unreadable. Transcribing code from books<br />

was extremely error-prone, especially in the unfriendly DOS BASIC environment. It<br />

would seem that pseudo code and abstract structures are the best way <strong>to</strong> approach any<br />

presentation of this rapidly evolving field. Danlos's use of pseudo-code in her Linguistic<br />

Basis of Text Generation (Danlos 1987) makes it more accessible <strong>to</strong>day. More recent<br />

texts on text generation (Reiter and Dale 2000), grammar formalism (Butt et al. 1989), or<br />

computational linguistics (Gazdar and Mellish 1989; Jurafsky and Martin 2000) use<br />

higher-level programming languages and grammar formalisms. (See SIGGEN 2002 and<br />

Calder 2002 for recent publications on generation.)


And yet the heart of generation lies in the code, because it is in the code that virtual<br />

structures are condensed in<strong>to</strong> single-step operations – the code becomes the text<br />

engineer's principal language. A simple example, au<strong>to</strong>mating Lewis Carroll's<br />

portmanteau words, should suffice <strong>to</strong> illustrate, in a very minimal way, both the problems<br />

and powers associated with a given programming environment.<br />

Let us program a system that will create portmanteau words by combining words whose<br />

spelling overlaps. That is, one word will end with the same letters with which the other<br />

word begins. To investigate that relation over a given set of words (the size of the set<br />

being almost immaterial for the computer), one must exhaustively compare every<br />

possible pair of words in the set. Each word must be decomposed in<strong>to</strong> all its possible<br />

front sequences and corresponding end sequences. For example, "tiger" and "gerbil" must<br />

be split in<strong>to</strong>: t iger, ti ger, tig er, tige r, tiger / g erbil, ge rbil, ger bil, gerb il, gerbi l,<br />

gerbil. Then each front of one must be compared <strong>to</strong> the ends of the other and vice versa <strong>to</strong><br />

check for (the longest) match: tigerbil.<br />

In a high-level language like Prolog, built-in programming constructs allow the<br />

specification of that splitting process <strong>to</strong> be very detailed and yet concise. The following<br />

code (Coelho et al. 1982: 18) uses a small set of French names of animals. It prints out:<br />

alliga<strong>to</strong>rtue, caribours, chevalliga<strong>to</strong>r, chevalapin, vacheval. (Some overlaps in English are<br />

beelephant, birdog, birdonkey, birdove, birduck, camelephant, camelion, catiger,<br />

caturkey, chickite, cowl, cowolf, crowl, crowolf, cuckoowl, dogecko, dogibbon, dogoat,<br />

dogoose, dovelephant, duckite, elephantiger, elephanturkey, frogecko, frogibbon, frogoat,<br />

frogoose, geckowl, goatiger, goaturkey, gooselephant, horselephant, jackalion,<br />

kitelephant, mouselephant, owlion, pigecko, pigibbon, pigoat, pigoose, roosterooster,<br />

sheepig, sheepigeon, snakelephant, tigerooster, wolfox, wolfrog. Of these, only<br />

camelephant and crowl overlap by more than one letter.) The actual core of the algorithm<br />

is only in the append and mutation routines; the others just deal with data representation<br />

and display. On most Prolog systems, append would be a built-in function, and so the<br />

programmer would only design mutation.


We can explain the mutation algorithm as shown in Table 30.1. The append code will try<br />

all the possible ways <strong>to</strong> split a list, just as the animal code will form all possible pairs of<br />

animals' names. The search for possible combinations (through the backtracking<br />

function) is a built-in dimension of any procedure and gives Prolog its specific character<br />

as a programming language. In fact, the programming issue in Prolog is really how <strong>to</strong><br />

contain the search for combinations. Thus, in the list for English overlaps, we find<br />

"roosterooster"; since we did not specify explicitly that Y and Z should not be the same,<br />

that combination was produced along with the others.<br />

AI programming languages such as Lisp and Prolog come closest <strong>to</strong> a pseudo-code for<br />

generation. Unfortunately, considerable time and energy is required <strong>to</strong> master the power<br />

and concision of a high-level programming language. It is clear nonetheless that if these<br />

languages are preferred for AI applications, it is precisely because they represent an<br />

important step <strong>to</strong>wards the virtual. Algorithms must ultimately stand on each other's<br />

shoulders, as mutation's use of append shows.<br />

Many generation systems are end-user applications which require no programming at all,<br />

but they do not allow any extension of virtualization (Kurzweil's poetry genera<strong>to</strong>r or the<br />

Shannonizer). Others are quite sophisticated specifications, such as ALICE and A/ML (AI<br />

XML), but do not deal in any direct way with crucial linguistic dimensions. Grammar


management packages are by definition still closer <strong>to</strong> questions of language generation<br />

(see the NLP Software Registry); some, such as KPML (Reiter 2002; Reiter and Dale<br />

2000) or Vinci (Lessard and Levison 2002) are explicitly conceived for generation. In all<br />

cases, the real question is how <strong>to</strong> make the programming language increasingly highlevel<br />

and expressive; grammar specification and programming are increasingly fused. For<br />

the moment, no metalanguage would allow the programmer <strong>to</strong> code a portmanteau<br />

genera<strong>to</strong>r without specifying in a very detailed manner character-string manipulations.<br />

30.1 Table The mutation algorithm<br />

mutation(X):- Try <strong>to</strong> make an X (a word) with the following properties:<br />

animal(Y),<br />

animal(Z),<br />

append(Y1,Y2,Y),<br />

Y1 =[],<br />

append(Y2,Z2,Z),<br />

Z2/= [],<br />

append(Y1,Z,X).<br />

To find that X, you will first have find two animals (whose names<br />

have been split in<strong>to</strong> lists of characters), Y and Z. (Finding two<br />

candidate animals will be the job of code defined elsewhere, in the<br />

animal definition.)<br />

Divide animal Y in<strong>to</strong> a front part (Y1) and an end part (Y2).<br />

Make sure that the end part (Y2) is not the whole word, which<br />

would be the case should Y1 be the empty list.<br />

Divide the other animal, Z, in<strong>to</strong> two parts such that the front is Y2,<br />

i.e., the same end found in the Y animal.<br />

Again, make sure that there is some overlap;<br />

Y2 should not be empty.<br />

Splice <strong>to</strong>gether Y1 (front of Y), with whole of Z <strong>to</strong> give X, the<br />

portmanteau we seek.<br />

Generalizing Generation: Grammar Levels<br />

A genera<strong>to</strong>r is fundamentally a mail merge function, as is found in most word processors.<br />

A template for a merge has blank spots that are filled in using a database of content items.<br />

For example, from the sentence "Jack finds Jill irresistible", we can construct the<br />

template: " finds ".


With that data a combina<strong>to</strong>rial generation will produce (6 * 6 * 6 =) 216 sentences having<br />

the same general form as the original. As can be seen, the combinations depend on the<br />

distinction between tag classes: the set is used twice, the set once.<br />

We might wish <strong>to</strong> refine our set of categories <strong>to</strong> distinguish between different aspects of<br />

the content items.


The template " finds " restricts the<br />

combina<strong>to</strong>ry <strong>to</strong> (3 * 3 * 2 =) 18 combinations. Other restrictions might be of the logical<br />

sort, for example excluding the (schizophrenic!) cases where the first and second person<br />

is the same person: " finds ". Variable A will<br />

be instantiated randomly and ♭A will be chosen randomly as well, but excluding<br />

whoever was chosen for A.<br />

Sentence grammars: context-free grammars<br />

Advanced genera<strong>to</strong>rs differ from mail merge only in generality, i.e., advanced systems<br />

construct the most general descriptions possible of all dimensions of the text.<br />

Unfortunately, as structures become more general, the metalanguage and the control of<br />

generation become increasingly complex. The grammar of that seemingly simple<br />

sentence is not simple if considered as part of English grammar in general. A more<br />

general description of the sentence requires a framework for grammar description that is<br />

ultimately similar <strong>to</strong> a programming language. There is often a direct translation between<br />

programming constructs and grammar metalanguage. Prolog has in fact a built-in<br />

grammar description language called definite clause grammars (DCGs), which are<br />

essentially context-free grammars (CFGs) (see Jurafsky and Martin 2000 for this and<br />

other formalisms). A CFG of the template " <br />

" is:<br />

Each rule represents a step in the description of a possible sentence. What is <strong>to</strong> the left of<br />

the arrow is rewritten in the form <strong>to</strong> the right. When generating a sentence, the<br />

substitution process continues until all the categories (left-hand side of the arrow) have


een rewritten in<strong>to</strong> words (terminal entries, which are not on the left-hand side of any<br />

rewrite rule). Diagrammatically a given sentence is a network of rewrites (see Figure<br />

30.3). This "stacking up" of grammatical categories in a tree form is called a phrasestructure<br />

tree diagram.<br />

Distinguishes (attributes)<br />

If our generative needs were limited <strong>to</strong> this one type of sentence, this simple CFG would<br />

be sufficient: from the database of possible nouns, adjectives, and verbs, it could generate<br />

all possible combinations. However, as it is, this grammar is <strong>to</strong>o specific <strong>to</strong> be reused as a<br />

component of a larger grammar; it only describes the very specific behavior of attributive<br />

verbs (verbs whose adjective complement is attributed <strong>to</strong> the direct object complement).<br />

Other sentences, such as the simple transitive sentence "Jack sees Jill", do not fit this<br />

framework: "Jack sees Jill irresistible" is not grammatical.<br />

We can generalize our grammar fairly simply by adding subclassifications <strong>to</strong> the main<br />

categories. We want <strong>to</strong> limit the VP <strong>to</strong> a particular subset of verbs, those that are<br />

attributive, and expand the grammar <strong>to</strong> allow for simple transitive constructions:<br />

30.3 Figure A phrase-structure tree diagram<br />

This is a more general grammar since it subcategorizes V so that we can distinguish<br />

between attributive and simple transitive verbs. (In Vinci, described below, distinguishers<br />

are called attributes.) Other distinguishers could be added <strong>to</strong> control various kinds of<br />

agreement between categories, such as number agreement (subject-verb). We will add the<br />

case of a plural subject:


The first rule states that the value of the number distinguisher, NUM, must be the same in<br />

both the VP and the NP. That value will be assigned in the rewrite rules that follow it:<br />

when the NP subject is finally chosen, the value of the number distinguisher will be<br />

transferred, through NUM, <strong>to</strong> VP and will constrain the choices from that node down the<br />

tree <strong>to</strong> the selection of the verb.<br />

Word grammars: morphology<br />

The generalizations we have made so far are generalizing extensions <strong>to</strong> the grammar;<br />

they allow for other possible sentences, or variants of the original. Another generalization<br />

concerns what was originally taken as the data items of the merge: the words themselves<br />

have their own grammar, their morphology. Most English verbs have a simple<br />

morphology, just a choice between "s" and "es" in the third person:<br />

Our morphological rules are more complex than in the previous CFG, but now they are<br />

more general: adding new verbs <strong>to</strong> the database requires adding only the infinitives <strong>to</strong> one<br />

of the four classes, rather than adding two inflected forms in different rules.


Text grammars: metagrammars<br />

All grammar might in principle be integrated in<strong>to</strong> a CF sentence grammar. Word<br />

morphology becomes a module of a sentence grammar, and a text of whatever length can<br />

be transformed in<strong>to</strong> a single sentence. Yet grammars traditionally describe separately the<br />

word, sentence, and textual levels. Text grammars bring with them a number of difficult<br />

issues. Minimally, a text is two or more sentences that cohere. That coherence expresses<br />

itself in many ways.<br />

The sentence genera<strong>to</strong>r we have described so far could easily generate two sentences <strong>to</strong><br />

make a text: "Jack finds Jill irresistible. Jack finds Jill irrepressible." To add a textual<br />

level <strong>to</strong> the grammar, it is sufficient <strong>to</strong> add the rules:<br />

This describes minimally a text as two sentences, but it does not generate a very natural<br />

one. We would prefer our first text <strong>to</strong> be rendered as: "Jack finds Jill irresistible and<br />

irrepressible." Furthermore, if we generated "Jack finds Jill irresistible. Jack finds Jill<br />

irritating", we would prefer <strong>to</strong> have "Jack finds Jill irresistible yet irritating."<br />

This process of removing repetitive structures, expressing the underlying logical<br />

relations, and making a natural synthesis of the information is called aggregation. In the<br />

first case, it is simply a question of suppressing the start of the second sentence and<br />

adding "and", but <strong>to</strong> accomplish that in a general fashion, we would need <strong>to</strong> understand<br />

how the two sentences compare in their structure. Not all sentences that end in adjectives<br />

can be conflated in this same way. (We cannot conflate "Jack finds Jill irresistible" and<br />

"Jack is irrepressible.") Suppression of the first part of the sentence and coordination of<br />

the adjectives is possible because each adjective is in exactly the same syntactic<br />

environment.<br />

Textual grammars are meta or control grammars; they direct the generation of sentences.<br />

Descriptions at this level are obviously quite complex since they are twice removed from<br />

an actual sentence. A control grammar would not normally attempt <strong>to</strong> reorganize what is<br />

already generated. Comparing and manipulating two parse trees would be a very complex<br />

process. Rather, typically it is more straightforward <strong>to</strong> allow the control grammar <strong>to</strong> build<br />

a plan of the text and then <strong>to</strong> select the grammatical resources needed <strong>to</strong> fulfill the plan.<br />

We might represent the plan or semantic skele<strong>to</strong>n of our text using predicate calculus. In<br />

this small grammar, each verb would have a general semantic description. One class of<br />

attributive verbs expresses the fact that the subject, X, believes that the object<br />

complement, Y, possesses a certain quality, Z: belief(X, is(Y, Z)). The predicate "belief"<br />

reflects the modal part of the fundamental meaning behind "finds", "considers", or


"believes"; "is", the attributive part. If our text is defined by two facts of this type, the<br />

rules of predicate logic will allow us <strong>to</strong> understand how they can be combined.<br />

From belief(X, is(Y,Z)) and belief(X, is(Y,W)) we can deduce (using logical<br />

transformations) that belief(X,&(is(Y,Z),is(Y,W))). From that expression, we simplify <strong>to</strong><br />

belief(X, is(Y,&(Z,W))). That is the general plan for our text, derived by transformation of<br />

formulas that preserve meaning. Our grammar would then have <strong>to</strong> take that meaning and<br />

translate it in<strong>to</strong> the appropriate grammatical structures and finally in<strong>to</strong> words.<br />

Vinci 1<br />

To understand how control grammars work, we will consider in more detail a small<br />

fragment of a s<strong>to</strong>ry genera<strong>to</strong>r coded in Vinci, the generation system designed by Lessard<br />

and Levison.<br />

Vinci supports all the grammar coding structures we have discussed so far: CFGs,<br />

inheritance of distinguishers (attributes), sophisticated morphological generation, and<br />

much more. The feature that will concern us here is their PRESELECT, which plays the<br />

role of a sentence planner. This system does not (yet) have a text planner as such, but the<br />

sentence planner does give us insight in<strong>to</strong> how a full-blown text planner would work.<br />

(KPML is a more complete implementation on this point, but considerably more complex.<br />

See Reiter and Dale 2000.)<br />

The following is a Vinci-generated s<strong>to</strong>ry designed by Lessard. It is presented here divided<br />

according <strong>to</strong> the Proppian narrative functions that served as its generative framework (an<br />

approach that is common since Klein et al. 1977):<br />

A Vinci-generated fairy tale<br />

(present king) _Il était une fois un roi (Once upon a time there was a king)<br />

(name king) qui s'appelait Pierre, (named Pierre.)<br />

_Il était beau, intelligent, bon, et il avait des cheveux roux. (He was<br />

(describe king)<br />

handsome, intelligent, good, and he had red hair.)<br />

(present<br />

_Il avait une enfant, la princesse. (He had a child, the princess.)<br />

victim)<br />

(describe<br />

victim)<br />

(make<br />

interdiction)<br />

_Elle était belle, intelligente, forte, bonne, et elle avait des cheveux<br />

blonds. (She was beautiful, intelligent, strong, good, and she had blond<br />

hair.)<br />

_Le roi ordonna a la princesse de rester au chateau. (The king ordered the<br />

princess <strong>to</strong> stay in the castle.)<br />

(disobey) _Elle désobéit quand meme. (She disobeyed nevertheless.)<br />

(leave) _Elle alia dans la forêt. (She went in<strong>to</strong> the forest.)


(encounter) _où elle rencontra un sorcier. (where she met a sorcerer.)<br />

(name villain) _Il s'appelait Merloc. (His name was Merloc.)<br />

(describe _Il était laid, intelligent, fort, mechant, et il avait des cheveux blonds. (He<br />

villain) was ugly, intelligent, strong, wicked, and he had blond hair.)<br />

(kidnap) _Il l'enleva. (He kidnapped her.)<br />

(ask)<br />

_Pour la sauver, le roi demanda l'aide d'un prince. (To save her, the king<br />

asked a prince for help.)<br />

_Heureusement, le prince rencontra un lutin qui lui fournit une grenouille<br />

(give) magique. (Luckily, the prince met an elf who provided him with a magic<br />

frog.)<br />

(kill)<br />

_Le prince l'utilisa pour tuer le sorcier. (The prince used it <strong>to</strong> kill the<br />

sorcerer.)<br />

(marry)<br />

_Il épousa la princesse et ils eurent beaucoup d'enfants. (He married the<br />

princess and they had many children.)<br />

This s<strong>to</strong>ry requires a number of support files and a separate grammar for each narrative<br />

function. The main files that concern us here are:<br />

1 The lexicon, where words are s<strong>to</strong>red with their attributes and indications for<br />

generating morphological variants. The lexicon, and the grammar in the user file, is<br />

expandable.<br />

2 The attribute file, which defines the attribute categories and the values of each<br />

attribute (which we called distinguishers in the previous sections). For example, in the<br />

Genre category (gender) there will be two attribute values, masc and fém.<br />

3 The user file, which is the main grammar file where the CFGs are found.<br />

Without any preselection, the system uses whatever grammar has been defined in the user<br />

file and generates random output. The PRESELECT, along with the SELECT function,<br />

controls how the grammar will generate. For example, we will want certain verbs <strong>to</strong> have<br />

certain nouns as subjects, and not others (e.g., "It is raining" is acceptable; "Paul is<br />

raining" is not). In this fairy tale, each narrative function concerns the sphere of action of<br />

a given subject and that subject must be the focus throughout the function. Attributes, the<br />

SELECT, and the PRESELECT functions of Vinci allow us <strong>to</strong> describe such constraints.<br />

Random generation will happen within the boundaries defined by these structures.<br />

Table 30.2 contains a detailed commentary of the user file for the second function, the<br />

"Name the king" function, produces the output "qui s'appelait X" (who is called X).<br />

Grammar control is achieved in Vinci principally through controlling the flow of<br />

attributes in the phrase structure trees and through loading and unloading grammar<br />

modules in user files. This is a simulation of grammar control, but it is clear that in a full<br />

implementation of a metagrammar, the grammar rules themselves become the objects of<br />

manipulation and take on the same status, but at a higher level, that words have in a


sentence grammar. To correctly generate sentences in a general manner, we had <strong>to</strong><br />

develop a complex system of grammatical categories and attributes. A metagrammar<br />

must in the same way develop the descriptive language, not <strong>to</strong> describe sentences, but <strong>to</strong><br />

describe what describes sentences: the CFG rules.<br />

In short, the most logical way <strong>to</strong> generate text under Vinci would be <strong>to</strong> use two Vincis:<br />

one <strong>to</strong> (randomly) generate a constrained grammar and the second <strong>to</strong> use that grammar <strong>to</strong><br />

randomly generate the final text. The PRESELECT and the SELECT rules represent a<br />

foreshadowing of that ultimate metagrammar.<br />

30.2 Table The "name the king" function<br />

PRESELECT = roi<br />

N[sing]/"roi"<br />

ROOT = NOMMER[roi]<br />

NOMMER = INHERIT Ps:<br />

Personnage; SELECT Ge:<br />

Genre in Ps, No: Nombre in<br />

Ps, Mo: Moralite in Ps;<br />

PRON[pronrel, suj, clit]<br />

PRON[pronref, p3, No]<br />

According <strong>to</strong> this PRESELECT instruction, the system will<br />

first choose a singular N (noun) whose lexical entry is "roi."<br />

"Roi" is defined in the lexicon as: "roi"|N|masc, Nombre,<br />

Determ, humain, male, roux, bon ("king"|N|masc, Number,<br />

Determiner, human, male, red-haired, good).<br />

This preselected lexical item will be used in the SELECT<br />

rule below. (Note that the "|" is a field delimiter in the<br />

dictionary entries.)<br />

The highest node is always ROOT. For this generation, its<br />

only child node is NOMMER with the semantic attribute<br />

roi. For generating "Name the villain", NOMMER is given<br />

attribute scelerat (villain) instead.<br />

In creating the daughter nodes for NOMMER as described<br />

below, we first inherit its Personnage attribute, in this case<br />

roi, giving this value <strong>to</strong> the variable Ps. Through Ps, we<br />

then access the lexical item preselected for roi (i.e., "roi") <strong>to</strong><br />

set up three variables (Ge, No, Mo). Notice that "roi" is<br />

masc, and has the attribute bon (good) taken from the<br />

Moralite category. The preselection has also been<br />

designated sing.<br />

To summarize the transfer of attributes: the lexical entry<br />

"roi" gives its attributes <strong>to</strong> the preselection labelled roi, one<br />

of the Personnage attributes defined in the attributes file.<br />

The NOMMER node has roi as its Personnage attribute,<br />

which is inherited under the name of Ps. Through Ps, the<br />

variables Ge, No, and Mo obtain values which are used <strong>to</strong><br />

select appropriate PRON and N words.<br />

"qui": the attributes select the subject relative pronoun in a<br />

clitic construction<br />

"se": a reflexive pronoun, 3rd person, with number (sing)<br />

defined in the PRESELECT. Morphological rules of this


entry deal with the elision of "e."<br />

V[vpron, imparf, p3, "appelait": the imperfect tense, 3rd person singular, of the<br />

No]/"appeler"<br />

verb "appeler."<br />

"Pierre" is a possible N[prp] (prp = proper noun) choice.<br />

His entry in the lexicon is: "Pierre"|N|prp, masc, sing,<br />

humain, Cheveux, Obeissance, Apparence, Force, Moralite,<br />

Intelligence." Of course, we do not know whether Pierre is<br />

actually bon – he is not defined as such in the lexicon. He<br />

N[prp, Mo, Ge, No] does however possess the attribute category Moralite, and<br />

that is sufficient for him <strong>to</strong> qualify as a potential roi. On the<br />

other hand, "Merloc", another N[prp], will not be chosen<br />

from the lexicon since his Mo does not match bon (he is<br />

méchant!: "Merloc"|N|prp, masc, sing, humain, Cheveux,<br />

Apparence, Force, mechant, Intelligence.<br />

PONCT[pt] Then a period.<br />

Definition 4, RP is the study of abstraction and meta processes.<br />

Art and Virtuality<br />

We have outlined a seemingly new area of study called robotic poetics. We have not tried<br />

<strong>to</strong> define all its dimensions or <strong>to</strong> subdivide it in any way. Rather we have focused on a<br />

number of interwoven problematics that seem <strong>to</strong> concern any form RP might take in the<br />

future: mechanization, combina<strong>to</strong>ry, virtualization, and abstraction <strong>to</strong>wards<br />

metastructures. There are many formalist poetics that have dealt with some of these<br />

issues in the past, and some are now being adapted <strong>to</strong> the new computational context<br />

(enterprises such as Rastier's interpretative semantics or Zholkovsky's poetics of<br />

expressiveness). However, RP's most pivotal shift in emphasis lies in the confluence of<br />

traditionally separate fields of interest: generation merges with analysis, theory with<br />

practice, creativity with criticism.<br />

Whatever the techniques that ultimately come <strong>to</strong> inform RP, the field is inevitably<br />

structured by the romantic image of a creative artificial intelligence. Winograd (1990)<br />

discusses that image in a practical manner. In his Count Zero (1986) William Gibson<br />

presents perhaps a more powerful literary description of RP's idealized object. One strand<br />

of the intrigue focuses on Marly Krushkhova, an art dealer who has been hired by Joseph<br />

Virek <strong>to</strong> track down the author of beautiful, shoebox-sized found-object sculptures. Her<br />

employer is astronomically wealthy, but now an immense blob of cancerous cells<br />

maintained in a vat at a S<strong>to</strong>ckholm labora<strong>to</strong>ry. Virek wants <strong>to</strong> transfer his mind out of the<br />

vat and in<strong>to</strong> an AI construct. He believes that the artist Marly seeks – an AI construct –<br />

holds the key <strong>to</strong> his transmogrification. Near the end of the novel Marly travels <strong>to</strong> an<br />

abandoned space station <strong>to</strong> meet the artist. She arrives in the weightless domed room<br />

where the robot is housed. Its arms stir the floating space debris:


There were dozens of the arms, manipula<strong>to</strong>rs, tipped with pliers, hexdrivers, knives, a<br />

subminiature circular saw, a dentist's drill…. They bristled from the alloy thorax of what<br />

must once have been a construction remote, the sort of unmanned, semiau<strong>to</strong>nomous<br />

device she knew from childhood videos of the high frontier. But this one was welded in<strong>to</strong><br />

the apex of the dome, its sides fused with the fabric of the Place, and hundreds of cables<br />

and optic lines snaked across the geodesies <strong>to</strong> enter it. Two of the arms, tipped with<br />

delicate force-feedback devices, were extended; the soft pads cradled an unfinished box.<br />

(Gibson 1986: 217)<br />

The debris, from which the mechanical arms extract components and fashion its box<br />

sculptures, is composed of personal belongings abandoned by a family of now decadent<br />

AI pioneers:<br />

A yellowing kid glove, the faceted crystal s<strong>to</strong>pper from some vial of vanished perfume,<br />

an armless doll with a face of French porcelain, a fat gold-fitted black fountain pen,<br />

rectangular segments of perf board, the crimpled red and green snake of a silk cravat….<br />

Endless, the slow swarm, the spinning things.<br />

(1986: 217)<br />

Virek, a pleasant, human-like "generated image" (1986: 227), appears on the terminal<br />

screen in the AI's workshop; Marly is caught floating between the fake glowing image<br />

that masks Virek's deformed and decadent humanity and the very concrete arms of the<br />

superhuman robotic artist. The AI explains that the Virek tera<strong>to</strong>ma might at best become<br />

"the least of my broken selves" (ibid.) – Marly recaps <strong>to</strong> the AI:<br />

You are someone else's collage. Your maker is the true artist. Was it the mad daughter [of<br />

the AI pioneer family]? It doesn't matter. Someone brought the machine here, welded it<br />

<strong>to</strong> the dome, and wired it <strong>to</strong> the traces of memory. And spilled, somehow, all the worn<br />

sad evidence of a family's humanity, and left it <strong>to</strong> be stirred, <strong>to</strong> be sorted by a poet. To be<br />

sealed away in boxes. I know of no more extraordinary work than this. No more complex<br />

gesture.<br />

(1986: 227)<br />

Why this fascination with the robotic poet? Why should the robotic author be so<br />

"extraordinary", the most "complex gesture"? There are many reasons why artists might<br />

be drawn <strong>to</strong> robotics, but one profound reason might lie in the nature of art itself.<br />

Art, like religion, is a window on<strong>to</strong> something beyond, though we know not what. The<br />

work of art entices us beyond who we are and what we understand, and we are left<br />

floating, like Marly, in a space defined only by the fact that it is new; one step beyond<br />

who we were, somewhere we thought we could not possibly be. The pleasure and the<br />

importance of art is perhaps that move up the meta organization of our understanding, <strong>to</strong><br />

a more general space that gives us a different, if not a better insight in<strong>to</strong> the world.


There is nothing new about this confluence of problematics. Art as combina<strong>to</strong>ry, love,<br />

divine madness, knowledge, and writing are at the heart of Pla<strong>to</strong>'s Phaedrus. Socrates<br />

cites Midas's epitaph, spoken by a bronze Eliza, a kind of rigid chatbot:<br />

Bronze maiden am I and on Midas' mound I lie.<br />

As long as water flows and tall trees bloom,<br />

Right here fixed fast on the tearful <strong>to</strong>mb,<br />

I shall announce <strong>to</strong> all who pass near: Midas is dead and buried here.<br />

(264e; cited in Carson 1986: 134)<br />

Socrates's point seems <strong>to</strong> be that combina<strong>to</strong>ry, like Midas's gold, is without meaning:<br />

I suppose you notice that it makes no difference which line is read first or which is read<br />

last.<br />

(264e; cited in Carson 1986: 135)<br />

Art lies elsewhere than in combina<strong>to</strong>ry, and yet art is fundamentally combination, as its<br />

etymology suggests. It lies somewhere in the vast domain of post-play. Socrates contrasts<br />

uncreative potential and creative virtuality; true knowledge and sophistry; love and<br />

convention. Artful discourse must have a beginning and an end. It must go somewhere,<br />

and that where is up. Like eros, it is a winged thing (Carson 1986).<br />

Pla<strong>to</strong>'s multifaceted Phaedrus gives us a very ancient and general definition of RP:<br />

Definition 5, RP is the study of art through the foil of mechanical art.<br />

Note<br />

1 My thanks go <strong>to</strong> Greg Lessard and Michael Levison for technical assistance with<br />

Vinci as well as their explanations of its formalism.<br />

References for Further Reading<br />

ACL (Association for Computational Linguistics) (2002). Natural Language Software<br />

Registry. At http://registry.dfki.de/.<br />

A.L.I.C.E AI Foundation (2002). A.L.I.C.E. AI Foundation. At http://alicebot.org/.<br />

Aarseth, Epsen J. (1997). Cybertext: Perspectives in Ergodic Literature. Baltimore: Johns<br />

Hopkins University Press.<br />

Anis, J. (1995). La Génération de textes littéraires: cas particulier de la génération de<br />

textes ou discipline à part? [Generation of literary texts: a particular example of text


generation or a separate field?]. In Littérature et infarmatique, ed. A. Vuillemin and M.<br />

Lenoble (pp. 33–48). Arras: Ar<strong>to</strong>is Presses Université.<br />

Balpe, Jean-Pierre and Bernard Magné, (eds.) (1991). L'Imagination infarmatique de la<br />

littérature [Literature's computerized imagination]. Saint-Denis: Presses Universitaires de<br />

Vincennes.<br />

Butt, Miriam, Tracey Holloway King, María-Eugenia Niño, and Fréderique Segond<br />

(1989). A Grammar Writer's Cookbook. Stanford, CA: CSLI Publications.<br />

Calder, Jo (2002). The Scottish Natural Language Generation Homepage. At<br />

http://www.cogsci.ed.ac.uk/~jo/Gen/snlg/nlg.html.<br />

Carson, Anne (1986). Eros the Bittersweet. Prince<strong>to</strong>n, NJ: Prince<strong>to</strong>n University Press.<br />

Coelho, Helder, José Carlos Cotta, and Luís Moniz Pereira (1982). How <strong>to</strong> Solve It with<br />

Prolog, 3rd edn. Lisbon: Ministerio Da habiacao e Obras Públicas.<br />

Danlos, Laurence (1987). The Linguistic Basis of Text Generation, tr. Dominique Debize<br />

and Colin Henderson. Cambridge: Cambridge University Press. (Original work published<br />

1985.).<br />

Fournel, Paul (1986). Computer and the Writer: The Centre Pompidou Experiment. In<br />

Oulipo: A Primer of Potential Literature, ed. and tr. W. F. Motte (pp. 140–2). Lincoln<br />

and London: University of Nebraska Press.<br />

Gazdar, Gerald and Chris Mellish (1989). Natural Language Processing in PROLOG.<br />

New York: Addison-Wesley Publishing.<br />

Gibson, William (1986). Count Zero. New York: Ace Books.<br />

Jurafsky, Daniel and James H. Martin (2000). Speech and Language Processing. Upper<br />

Saddle River, NJ: Prentice Hall.<br />

Klein, Sheldon, J. F. Aeschlimann, M. A. Appelbaum, D. F. Balsiger, E. J. Curtis, M.<br />

Foster, S. D. Kalish, S. J. Kamin, Y.-D. Lee, L. A. Price, and D. F. Salsieder (1977).<br />

Modelling Propp and Lévi-Strauss in a Meta-symbolic Simulation System. In Patterns in<br />

Oral Literature, ed. H. Jason and D. Segal (pp. 141–222). The Hague: Mou<strong>to</strong>n.<br />

Kurzweil, Ray (2002). The Age of Intelligent Machines: "A (Kind of) Turing Test." At<br />

http://www.kurzweilcyberart.com/poetry/rkcp_akindofturingtest.php3.<br />

Lessard, Greg and Michael Levison (2002). Vinci Labora<strong>to</strong>ry. At<br />

http://www.cs.queensu.ca/CompLing/.


Lévy, Pierre (1998). Qu'est-ce que le virtuel? [What is the virtual?]. Paris: La Découverte<br />

and Syros.<br />

Motte, Warren (1995). Playtexts: Ludics in Contemporary Literature. Lincoln: University<br />

of Nebraska Press.<br />

Queneau, Raymond (1973). Cent mille milliards dépoémes [One hundred trillion poems].<br />

Paris: Gallimard.<br />

Rastier, François, Marc Cavazza, and Anne Abeillé (1994). Sémantique pour l'analyse<br />

[Semantic for analysis]. Paris: Masson.<br />

Reiter, Ehud (2002). KPML. At http://www.fb10.unibremen.de/anglistik/langpro/kpml/README.html.<br />

Reiter, Ehud, and Robert Dale (2000). Building Natural Language Generation Systems.<br />

Cambridge: Cambridge University Press.<br />

Shannon_Team (2002). The Shannonizer. At http://www.nightgarden.com/shannon.htm.<br />

SIGGEN (2002). ACL Special Interest Group on Generation. At http://www.siggen.org/.<br />

Winder, Bill (1994). Le Robo-poète: litterature et critique à l'ére électronique [The robopoet:<br />

literature and criticism in the electronic era]. In Littérature, Informatique, Lecture,<br />

ed. A. Vuillemin and M. Lenoble (pp. 187–213). Limoges: Presses Universitaires de<br />

Limoges.<br />

Winograd, Terry (1990). Thinking Machines: Can There Be? Are We? In The<br />

Foundations of Artificial Intelligence, ed. D. Patridge and Y. Wilks (pp. 167–89).<br />

Cambridge: Cambridge University Press.<br />

Zholkovsky, Alexander (1984). Themes and Texts. Ithaca and London: Cornell<br />

University Press.<br />

30.1 Figure The potential, the real, the actual, and the virtual


30.2 Figure Raymond Queneau, Cent mille milliards de poèmes<br />

31.<br />

Designing Sustainable Projects and Publications<br />

Daniel V. Pitti<br />

Introduction<br />

Designing complex, sustainable <strong>digital</strong> <strong>humanities</strong> projects and publications requires<br />

familiarity with both the research subject and available technologies. It is essential that<br />

the scholar have a clear and explicit understanding of his intellectual objectives and the<br />

resources <strong>to</strong> be used in achieving them. At the same time, the scholar must have<br />

sufficient comprehension of available technologies, in order <strong>to</strong> judge their suitability for<br />

representing and exploiting the resources in a manner that best serves his intellectual<br />

objectives. Given the dual expertise required, scholars frequently find it necessary <strong>to</strong><br />

collaborate with technologists in the design and implementation processes, who bring<br />

different understandings, experience, and expertise <strong>to</strong> the work. Collaboration in and of<br />

itself may present challenges, since most humanists generally work alone and share the<br />

research's results, not the process.<br />

Collaborative design involves iterative analysis and definition, frequently accompanied<br />

by pro<strong>to</strong>type implementations that test the accuracy and validity of the analysis. While<br />

computers enable sophisticated data processing and manipulation <strong>to</strong> produce desired<br />

results, the scholar must analyze, define, and represent the data in detail. Each data<br />

component and each relation between data components needs <strong>to</strong> be identified, and, when<br />

represented in the computer, named and circumscribed. Data must not simply be entered<br />

in<strong>to</strong> a computer, but be accompanied by additional data that identify the data<br />

components. This decomposition is necessary <strong>to</strong> instruct the computer, through programs<br />

or applications, <strong>to</strong> process the data <strong>to</strong> achieve the objectives, a process of locating,<br />

identifying, and manipulating the data.


The data need not be endlessly analyzed and decomposed. It is only necessary <strong>to</strong> identify<br />

the data components and relations deemed essential <strong>to</strong> achieving desired results. To<br />

determine what is essential in reaching the objectives, it is necessary also <strong>to</strong> understand<br />

the objectives in detail and subject them <strong>to</strong> a rigorous analysis, defining and making<br />

explicit each important function. Analysis of the objectives informs analysis of the data,<br />

determining what data and features of the data will be required <strong>to</strong> meet the objectives,<br />

and thus what features will need <strong>to</strong> be identified and explicitly represented. Conversely,<br />

analysis of the data will inform analysis of the objectives.<br />

Analyzing objectives relies on identifying and defining the intended user community (or<br />

communities) and the uses. Are the intended users peers, sharing the same or similar<br />

research interests? Are they students, or perhaps the interested public? If the users are<br />

students, then at what level: K-12, K-16, 9–12, or 13–16? Perhaps they are a very focused<br />

group, such as high school literature or his<strong>to</strong>ry classes? For the intended user community,<br />

is the project primarily intended <strong>to</strong> facilitate research, scholarly communication of<br />

research, reference, pedagogy, or perhaps a combination of one or more of these? The<br />

answers <strong>to</strong> these questions will assist in determining the functions that will be most likely<br />

<strong>to</strong> serve the perceived needs and at the appropriate level or levels.<br />

As important as users are, those responsible for creation and maintenance of the project<br />

must also be considered. In fact, they may be considered another class of users, as distinct<br />

from the so-called end users. Before a project begins <strong>to</strong> produce and publish data content,<br />

the project must have an infrastructure that will support its production and publishing.<br />

Someone – project manager, researcher, lead design person – needs <strong>to</strong> specify the steps or<br />

workflow and methods involved in the complex tasks of identifying, documenting, and<br />

digitizing artifacts; creating original <strong>digital</strong> resources; and the control and management of<br />

both the assets created and the methods used. The design must incorporate the crea<strong>to</strong>r's<br />

requirements. Control, management, and production methods are easily overlooked if the<br />

designer focuses exclusively on the publication and end user. Asset management does not<br />

present a particular problem in the early stages of a project, when there are only a handful<br />

of objects <strong>to</strong> track. As the number of files increases, though, it becomes increasingly<br />

difficult <strong>to</strong> keep track of file names, addresses and versions.<br />

Focusing on achieving particular short-term results may also lead <strong>to</strong> neglecting<br />

consideration of the durability or persistence of the content and its publication. Ongoing<br />

changes in computer technology present long-term challenges. Changes in both<br />

applications and the data notations with which they work will require replacing one<br />

version of an application with another, or with an entirely different application, and will<br />

eventually require migrating the data from one notation <strong>to</strong> another. Generally these<br />

changes do not have a major impact in the first four or five years of a project, and thus<br />

may not seem significant <strong>to</strong> projects that are intended <strong>to</strong> be completed within three or<br />

four years. Most researchers and crea<strong>to</strong>rs, though, will want their work <strong>to</strong> be available for<br />

at least a few years after completion. If a project has or will have a publisher, the<br />

publisher will certainly want it <strong>to</strong> remain viable as long as it has a market. The cultural<br />

heritage community, important participants in scholarly communication, may judge a<br />

publication <strong>to</strong> have cultural value for future generations. Archivists, librarians, and


museum cura<strong>to</strong>rs, in collaboration with technologists, have been struggling seriously<br />

since at least the 1980s <strong>to</strong> develop methods for ensuring long-term access <strong>to</strong> <strong>digital</strong><br />

information. Though many critical challenges remain, there is broad consensus that<br />

consistent and well-documented use of computer and intellectual standards is essential.<br />

Economy must also be considered. There are the obvious up-front costs of computer<br />

hardware and software, plus maintenance costs as these break or become obsolete.<br />

Human resources are also an important economic fac<strong>to</strong>r. Human resources must be<br />

devoted <strong>to</strong> the design, pro<strong>to</strong>typing, and, ultimately, implementation. Once implemented,<br />

more resources are needed <strong>to</strong> maintain a project's infrastructure. Data production and<br />

maintenance also require human resources. Data needs <strong>to</strong> be identified, collected, and<br />

analyzed; created, and maintained in machine-readable form; and reviewed for accuracy<br />

and consistency. Realizing complex, detailed objectives requires complex, detailed<br />

implementations, production methods, and, frequently, a high degree of data<br />

differentiation. As a general rule, the more detailed and complex a project becomes, the<br />

more time it will take <strong>to</strong> design, implement, populate with data, and maintain. Thus every<br />

objective comes with costs, and the costs frequently have <strong>to</strong> be weighed against the<br />

importance of the objective. Some objectives, while worthwhile, may simply be <strong>to</strong>o<br />

expensive.<br />

Design is a complex exercise: given this complexity, the design process is iterative, and<br />

each iteration leads progressively <strong>to</strong> a coherent, integrated system. It is quite impossible<br />

<strong>to</strong> simultaneously consider all fac<strong>to</strong>rs. Both the whole and the parts must be considered in<br />

turn, and each from several perspectives. Consideration of each in turn leads <strong>to</strong><br />

reconsideration of the others, with ongoing modifications and adjustments. Pro<strong>to</strong>typing of<br />

various parts <strong>to</strong> test economic, intellectual, and technological feasibility is important. It is<br />

quite common <strong>to</strong> find an apparent solution <strong>to</strong> a particular problem while considering one<br />

or two fac<strong>to</strong>rs but <strong>to</strong> have it fail when other fac<strong>to</strong>rs are added, requiring it <strong>to</strong> be modified<br />

or even abandoned. Good design, then, requires a great deal of persistence and patience.<br />

Disciplinary Perspective and Objects of Interest<br />

Humanists study human artifacts, attempting <strong>to</strong> understand, from a particular perspective<br />

and within a specific context, what it means <strong>to</strong> be human. The term artifact is used here in<br />

the broadest sense: an artifact is any object created by humans. Artifacts may be primary<br />

objects of interest or serve as evidence of human activities, events, and intentions.<br />

Archivists make a useful distinction between an object intentionally created, as an end in<br />

itself, and an object that is a by-product of human activities, the object typically<br />

functioning instrumentally as a means or <strong>to</strong>ol. Novels, poems, textbooks, paintings, films,<br />

sculptures, and symphonies are created as ends in themselves. Court records, birth<br />

certificates, memoranda, sales receipts, deeds, and correspondence facilitate or function<br />

as a means <strong>to</strong> other objectives, or as evidence or a record of particular actions taken.<br />

After serving their intended purpose, instrumental artifacts primarily function as<br />

evidence. Artifacts that are ends in themselves may of course also function as evidence.<br />

A novel, for example, may be viewed as a literary object, with particular aesthetic and<br />

intellectual qualities, or a source of evidence for language usage or for identifying


important cultural values in a particular his<strong>to</strong>rical context. Many disciplines share an<br />

interest in the same or similar objects or have similar assumptions and methods. Linguists<br />

and students of literature may share an interest in written texts, perhaps even literary<br />

texts, but differ substantially in what characteristics of texts are of interest or considered<br />

significant. Socio-cultural his<strong>to</strong>rical studies, while differing widely in the phenomena<br />

studied, will nevertheless employ similar methods, with the shared objective of a<br />

his<strong>to</strong>rical understanding.<br />

In order <strong>to</strong> apply computer technology <strong>to</strong> <strong>humanities</strong> research, it is necessary <strong>to</strong> represent<br />

in machine-readable form the artifacts or objects of primary or evidentiary interest in the<br />

research, as well as secondary information used in the description, analysis, and<br />

interpretation of the objects. The representations must reflect the disciplinary perspective<br />

and methods of the researcher, and facilitate the analytic and communication objectives<br />

of the project. In addition <strong>to</strong> perspective and intended uses, the method of representation<br />

will depend upon the nature of the artifact <strong>to</strong> be represented. As modes of human<br />

communication, pic<strong>to</strong>rial materials, printed texts, mathematical formulae, tabular data,<br />

sound, film, sculptures, buildings, and maps, <strong>to</strong> name a few possible artifact types, all<br />

have distinct observable characteristics or features, and representations will generally<br />

isolate and emphasize those deemed essential <strong>to</strong> the research.<br />

All information in computers exists in a coded form that enables it <strong>to</strong> be mechanically<br />

read and processed by the computer. Such codes are ultimately based on a binary system.<br />

Combinations of fixed sequences of 0s and 1s are used <strong>to</strong> create more complex<br />

representations, such as picture elements or pixels, sound samples, and alphabetic and<br />

numeric characters. Each of these is then combined <strong>to</strong> create more complex<br />

representations. For example, an array of pixels is used <strong>to</strong> form a picture, a sequence of<br />

alphabetic and numeric characters a text, and a sequence of sound samples a song. The<br />

codes so developed are used <strong>to</strong> represent both data and the programs that instruct the<br />

computer in reading and processing the data. There are a variety of ways of creating<br />

machine-readable texts. Text can be entered or transcribed from a keyboard. Some<br />

printed textual materials can be scanned and then rendered in<strong>to</strong> machine-readable form<br />

using optical character recognition (OCR) software, although many texts, particularly in<br />

manuscript form, exceed the capability of current OCR technology. Either way, creating<br />

a sequence of machine-readable characters does not take advantage of major<br />

technological advances made in the representation and manipulation of written language<br />

and mathematics.<br />

Written language and mathematics are composed of character data, marks, or graphs.<br />

Characters include ideographs (such as Chinese characters) and phonographs (as in the<br />

Latin alphabet), as well as Arabic numbers and other logograms. The discrete codes used<br />

in writing are mapped or reduced <strong>to</strong> machine-readable codes when represented in<br />

computers. Unicode (defined in ISO/IEC 10646) represents the most extensive attempt <strong>to</strong><br />

map the known reper<strong>to</strong>ire of characters used by humans <strong>to</strong> machine-readable codes.<br />

While written language (text) and mathematics are not the only types of information<br />

represented in computers, text constitutes an essential component of all <strong>humanities</strong><br />

projects. Many projects will use one or more his<strong>to</strong>rical texts, the literary works of an


author or group of related authors, a national constitution, or a philosophical work as the<br />

principal subject. For other projects, particular texts and records may not be the object of<br />

study, but texts may still be used for information about individuals, families, and<br />

organizations; intellectual, artistic, political, or other cultural movements; events; and<br />

other socio-cultural phenomena. Projects employ text <strong>to</strong> describe and document resources<br />

and <strong>to</strong> communicate the results of analysis and research. Further, text is used <strong>to</strong> identify,<br />

describe, control, and provide access <strong>to</strong> the <strong>digital</strong> assets collected in any project, and is<br />

essential in describing, effecting, and controlling interrelations between entities and the<br />

<strong>digital</strong> assets used <strong>to</strong> represent them. Comparing, identifying, describing, and<br />

documenting such interrelations are essential components of all <strong>humanities</strong> research.<br />

The power and complexity of computer processing of textual data require complex<br />

representation of the text. A simple sequence of characters is of limited utility. In order <strong>to</strong><br />

take full advantage of computing, textual data must be decomposed in<strong>to</strong> their logical<br />

components, with each component identified or named, and its boundaries explicitly or<br />

implicitly marked. This decomposition and identification is necessary because<br />

sophisticated, accurate, and reliable processing of text is only possible when the computer<br />

can be unambiguously instructed <strong>to</strong> identify and isolate characters and strings of<br />

characters and <strong>to</strong> perform specified operations on them. There are a variety of specialized<br />

applications available for recording and performing mathematical operations on<br />

numerical data. There are also a variety of applications available for recording and<br />

processing textual data, though most do not have characteristics and features that make<br />

them appropriate for sustainable, intellectually complex <strong>humanities</strong> research.<br />

Database and markup technologies represent the predominant technologies available for<br />

textual information. Though they differ significantly, they have several essential<br />

characteristics and features in common. Both exist in widely supported standard forms.<br />

Open, community-based and widely embraced standards offer reasonable assurance that<br />

the represented information will be durable and portable. Database and markup<br />

technologies are either based on or will accept standard character encodings (ASCII or,<br />

increasingly, Unicode). Both technologies enable an explicit separation of the logical<br />

components of information from the operations that are applied <strong>to</strong> them. This separation<br />

is effected by means of two essential features. First, each requires user-assigned names<br />

for designating and delimiting the logical components of textual objects. While there are<br />

some constraints, given the available character reper<strong>to</strong>ire, the number of possible names<br />

is unlimited. Second, each supports the logical interrelating and structuring of the data<br />

components.<br />

Though each of these technologies involves naming and structurally interrelating textual<br />

entities and components, each does so using distinctly different methods, and exploits the<br />

data in different ways. Deciding which technology <strong>to</strong> use will require analyzing the<br />

nature of the information <strong>to</strong> be represented and identifying those operations you would<br />

like <strong>to</strong> perform on it. With both technologies, the naming and structuring requires<br />

anticipating the operations. Each of these technologies is optimized <strong>to</strong> perform a set of<br />

well-defined operations. Though there is some overlap in functionality, the two<br />

technologies are best described as complementary rather than competitive. Information


est represented in databases is characterized as "data-centric", and information best<br />

represented in markup technologies is characterized as "document-centric." Steve<br />

DeRose, in a paper delivered in 1995 at a conference devoted <strong>to</strong> discussing the encoding<br />

of archival finding aids, made much the same distinction, though he used the terms<br />

"document" and "form" <strong>to</strong> distinguish the two information types (see Steve DeRose,<br />

"Navigation, Access, and Control Using Structured Information", in The American<br />

Archivist (Chicago: Society of American Archivists), 60,3, Summer 1997, pp. 298–309).<br />

Information collected on many kinds of forms is data-centric. Job applications are a<br />

familiar example. Name, birth date, street address, city, country or state, postal codes,<br />

education, skills, previous employment, date application completed, name or description<br />

of a position sought, references, and so on, are easily mapped in<strong>to</strong> a database and easily<br />

retrieved in a variety of ways. Medical and student records, and census data are other<br />

common examples of data-centric information. These familiar examples have the<br />

following characteristics in common:<br />

• Regular number of components or fields in each discrete information unit.<br />

• Order of the components or fields is generally not significant.<br />

• Each information component is restricted <strong>to</strong> data. That is, it has no embedded<br />

delimiters, other than the formal constraints of data typing (for example, a date may be<br />

constrained <strong>to</strong> a sequence of eight Arabic numbers, in the order year-month-day).<br />

• Highly regularized structure, possibly with a fixed, though shallow, hierarchy.<br />

• Relations between discrete information units have a fixed number of types (though the<br />

number of occurrences of each type may or may not be constrained).<br />

• Processing of data-centric information (such as accurate recall and relevance retrieval,<br />

sorting, value comparison, and mathematical computation) is highly dependent on<br />

controlled values and thus highly dependent on machine-enforced data typing, authority<br />

files, and a high degree of formality, accuracy, and consistency in data creation and<br />

maintenance.<br />

Common types of databases are hierarchical, network, relational, and object-oriented.<br />

Today, relational databases are by far the most prevalent. Structured Query Language<br />

(SQL), an ISO standard first codified in 1986 and substantially revised in 1992 and again<br />

in 1999, is the major standard for relational databases. While compliance with the<br />

standard is somewhat irregular, most relational databases comply sufficiently <strong>to</strong> ensure<br />

the portability and durability of data across applications. There are only a few viable<br />

implementations of object-oriented databases, though the technology has contributed<br />

conceptual and functional models that influenced the 1999 revision of SQL. SQL is both<br />

a Data Definition Language (DDL) and a Data Manipulation Language (DML). As a<br />

DDL, it allows users <strong>to</strong> define and name tables of data, and <strong>to</strong> interrelate the tables. Such<br />

definitions are frequently called database schemas. Tables have rows and columns.


Within the context of databases, rows are commonly called records, and columns called<br />

fields in the records. Each column or field is assigned a name, typically a descriptive term<br />

indicating the type of data <strong>to</strong> be contained in it. The DML facilitates updating and<br />

maintaining the data, as well as sophisticated querying and manipulating of data and data<br />

relations.<br />

Most his<strong>to</strong>rical or traditional documents and records are <strong>to</strong>o irregular for direct<br />

representation in databases. Data in databases are rigorously structured and systematic<br />

and most his<strong>to</strong>rical documents and records simply are not. Nevertheless, some documents<br />

and records will be reducible <strong>to</strong> database structures, having characteristics approximating<br />

those listed above. Many traditional government, education, and business records may<br />

lend themselves <strong>to</strong> representation in databases, as will many ledgers and accounting<br />

books. But a doc<strong>to</strong>ral dissertation, while a record, and generally quite regularized in<br />

structure, lacks other features of data that it would have <strong>to</strong> have <strong>to</strong> fit comfortably in a<br />

database architecture – for example, the order of the text components matters very much,<br />

and if they are rearranged arbitrarily, the argument will be rendered incoherent.<br />

While database technology may be inappropriate for representing most his<strong>to</strong>rical<br />

documents and records, it is very appropriate technology for recording analytic<br />

descriptions of artifacts and in systematically describing abstract and concrete<br />

phenomena based on analysis of evidence found in artifacts. Analytic descriptive<br />

surrogates will be useful in a wide variety of projects. Archaeologists, for example, may<br />

be working with thousands of objects. Cataloguing these objects involves systematically<br />

recording a vast array of details, frequently including highly articulated classification<br />

schemes and controlled vocabularies. Database technology will almost always be the<br />

most appropriate and effective <strong>to</strong>ol for collecting, classifying, comparing, and evaluating<br />

artifacts in one or many media.<br />

Database technology is generally an appropriate and effective <strong>to</strong>ol for documenting and<br />

interrelating records concerning people, events, places, movements, artifacts, or themes.<br />

Socio-cultural his<strong>to</strong>rical projects in particular may find databases useful in documenting<br />

and interrelating social and cultural phenomena. While we can digitize most if not all<br />

artifacts, many concrete and abstract phenomena are not susceptible <strong>to</strong> direct digitization.<br />

For example, any project that involves identifying a large number of people,<br />

organizations, or families will need <strong>to</strong> represent them using descriptive surrogates. A<br />

picture of a person, when available, might be very important, but textual identification<br />

and description will be necessary. In turn, if an essential part of the research involves<br />

describing and relating people <strong>to</strong> one another, <strong>to</strong> artifacts created by them or that provide<br />

documentary evidence of them, <strong>to</strong> events, <strong>to</strong> intellectual movements, and so on, then<br />

database technology will frequently be the most effective <strong>to</strong>ol. Individuals, artifacts,<br />

places, events, intellectual movements, can each be identified and described (including<br />

chronological data where appropriate). Each recorded description can be associated with<br />

related descriptions. When populated, such databases enable users <strong>to</strong> locate individual<br />

entities, abstract as well as concrete, and <strong>to</strong> see and navigate relations between entities. If<br />

the database is designed and configured with sufficient descriptive and classifica<strong>to</strong>ry<br />

detail, many complex analytic queries will be possible, queries that reveal intellectual


elations between unique entities, but also between categories or classes of entities based<br />

on shared characteristics.<br />

Textbooks, novels, poems, collections of poetry, newspapers, journals, and journal<br />

articles are all examples of document-centric data. These familiar objects have in<br />

common the following characteristics:<br />

• Irregular number of parts or pieces. Documents, even documents of a particular type, do<br />

not all have the same number of textual divisions (parts, chapters, sections, and so on),<br />

paragraphs, tables, lists, and so on.<br />

• Serial order is significant. It matters whether this chapter follows that chapter, and<br />

whether this paragraph follows that paragraph. If the order is not maintained, then<br />

intelligibility and sense break down.<br />

• Semi-regular structure and unbounded hierarchy.<br />

• Arbitrary intermixing of text and markup, or what is technically called mixed content.<br />

• Arbitrary number of interrelations (or references) within and among documents and<br />

other information types, and generally the types of relations are unconstrained or only<br />

loosely constrained.<br />

Extensible Markup Language (XML), first codified in 1998 by the World Wide Web<br />

Consortium (W3C), is the predominant markup technologies standard. XML is a direct<br />

descendant of Standard Generalized Markup Language (SGML), first codified in 1986 by<br />

the International Standards Organization (ISO). While the relationship between XML and<br />

SGML is quite complex, XML may be viewed as "normalized" SGML, retaining the<br />

essential functionality of SGML, while eliminating those features of SGML that were<br />

problematic for computer programmers.<br />

XML is a descriptive or declarative approach <strong>to</strong> encoding textual information in<br />

computers. XML does not provide an off-the-shelf tagset that one can simply take home<br />

and apply <strong>to</strong> a letter, novel, article, or poem. Instead, it is a standard grammar for<br />

expressing a set of rules according <strong>to</strong> which a class of documents will be marked up.<br />

XML provides conventions for naming the logical components of documents, and a<br />

syntax and metalanguage for defining and expressing the logical structure and relations<br />

among the components. Using these conventions, individuals or members of a<br />

community sharing objectives with respect <strong>to</strong> a particular type of document can work<br />

<strong>to</strong>gether <strong>to</strong> encode those documents.<br />

One means of expressing analytic models written in compliance with formal XML<br />

requirements is the document type definition, or DTD. For example, the Association of<br />

American Publishers has developed four DTDs for books, journals, journal articles, and<br />

mathematical formulae. After thorough revision, this standard has been released as an<br />

ANSI/NISO/ISO standard, 12083- A consortium of software developers and producers


has developed a DTD for computer manuals and documentation called DocBook. The<br />

Text Encoding Initiative (TEI) has developed a complex suite of DTDs for the<br />

representation of literary and linguistic materials. Archivists have developed a DTD for<br />

archival description or finding aids called encoded archival description (EAD). A large<br />

number of government, education and research, business, industry, and other institutions<br />

and professions are currently developing DTDs for shared document types. DTDs shared<br />

and followed by a community can themselves become standards. The ANSI/NISO/ISO<br />

12083, DocBook, TEI, and EAD DTDs are all examples of standards.<br />

The standard DTDs listed above have two common features. First, they are developed<br />

and promulgated by broad communities. For example, TEI was developed by humanists<br />

representing a wide variety of disciplines and interests. The development process (which<br />

is ongoing) requires negotiation and consensus-building. Second, each DTD is<br />

authoritatively documented, with semantics and structure defined and described in detail.<br />

Such documentation helps make the language public rather than private, open rather than<br />

closed, standard rather than proprietary, and thereby promotes communication. These two<br />

features – consensus and documentation – are essential characteristics of standards, but<br />

they are not sufficient in and of themselves <strong>to</strong> make standards successful (i.e., make them<br />

widely unders<strong>to</strong>od and applied within an acceptable range of uniformity).<br />

The distinction between document-centric and data-centric information, while useful, is<br />

also frequently problematic. It is a useful distinction because it highlights the strengths of<br />

markup languages and databases, although it does not reflect information reality, as such.<br />

Frequently any given instance of information is not purely one or the other, but a mixture<br />

of both. Predominantly data-centric information may have some components or features<br />

that are document-centric, and vice-versa. The distinction is of more than theoretical<br />

interest, as each information type is best represented using a different technology, and<br />

each technology is optimized <strong>to</strong> perform a specific set of well-defined functions and<br />

either does not perform other functions, or performs them less than optimally. When<br />

deciding the best method, it is unfortunately still necessary <strong>to</strong> weigh what is more and<br />

less important in your information and <strong>to</strong> make trade-offs.<br />

In addition <strong>to</strong> text representation. XML is increasingly used in realizing other functional<br />

objectives. Since the advent of XML in 1998, many database developers have embraced<br />

XML syntax as a data-transport syntax, that is, for passing information in machine-<strong>to</strong>machine<br />

and machine-<strong>to</strong>-human communication, XML is also increasingly being used <strong>to</strong><br />

develop standards for declarative processing of information. Rendering and querying are<br />

probably the two primary operations performed on XML-encoded textual documents. We<br />

want <strong>to</strong> be able <strong>to</strong> transform encoded documents in<strong>to</strong> human-readable form, on computer<br />

screens, and on paper. We also want <strong>to</strong> be able <strong>to</strong> query individual documents and<br />

collections of documents. There are currently several related standards for rendering and<br />

querying XML documents that have been approved or are under development: Extensible<br />

Stylesheet Language – Transformation (XSLT), Extensible Stylesheet Language –<br />

Formatting Objects (XSLFO), and XQuery. These declarative standards are significant<br />

because they ensure not only the durability of encoded texts, but also the durability of the<br />

experienced presentation.


Users and Uses<br />

Developing XML and database encoding schemes, as we have seen, involves<br />

consideration of the scholarly perspective, the nature and content of the information <strong>to</strong> be<br />

represented, and its intended use. For example, if we have a collection of texts by several<br />

authors, and we want users <strong>to</strong> be able <strong>to</strong> locate and identify all texts by a given author,<br />

then we must explicitly identify and delimit the author of each text. In the same way, all<br />

intended uses must be identified, and appropriate names and structures must be<br />

incorporated in<strong>to</strong> the encoding schemes and in<strong>to</strong> the encoded data. In a sense, then, XML<br />

and database encoding schemes represent the perspective and understanding of their<br />

designers, and each DTD or schema is an implied argument about the nature of the<br />

material under examination and the ways in which we can use it.<br />

Specifying an intended use requires identifying and defining users and user communities.<br />

The researcher is an obvious user. In fact, the researcher might be the only user, if the<br />

intention is <strong>to</strong> facilitate collection and analysis of data, with analytic results and<br />

interpretation published separately. Such an approach, though, deprives readers of access<br />

<strong>to</strong> the resources and methods used, and there is an emerging consensus that one of the<br />

great benefits of computers and networks is that they allow us <strong>to</strong> expose our evidence and<br />

our methods <strong>to</strong> evaluation and use by others. If that is considered desirable, design must<br />

then take other users in<strong>to</strong> consideration. As a general rule, the narrower and more specific<br />

the intended audience, the easier it will be <strong>to</strong> identify and define the uses of the data. If<br />

the intended audience is more broadly defined, it is useful <strong>to</strong> classify the users and uses<br />

according <strong>to</strong> educational and intellectual levels and abilities. While there may be overlap<br />

in functional requirements at a broad level, there will generally be significant differences<br />

in the complexity and detail of the apparatus made available <strong>to</strong> users.<br />

We can divide user functions or operations in<strong>to</strong> three types: querying, rendering, and<br />

navigation. This division is simply an analytic convenience. The categories are<br />

interrelated and interdependent, and some functions may involve combinations of other<br />

functions. Querying, simply characterized, involves users submitting words, phrases,<br />

dates, and the like, <strong>to</strong> be matched against data, with matched data or links <strong>to</strong> matched<br />

data returned, typically in an ordered list, <strong>to</strong> the user. Browsing is a type of querying,<br />

though the author of the research collection predetermines the query and thus the list of<br />

results returned. Rendering is the process by which machine-readable information is<br />

transformed in<strong>to</strong> human-readable information. All information represented in computers,<br />

whether text, graphic, pic<strong>to</strong>rial, or sound, must be rendered. Some rendering is<br />

straightforward, with a direct relation between the machine-readable data and its humanreadable<br />

representation. Navigational apparatus are roughly analogous <strong>to</strong> the title page,<br />

table of contents, and other information provided in the preliminaries of a book. They<br />

inform the user of the title of the research collection, its intellectual content, scope, and<br />

organization, and provide paths or a means <strong>to</strong> access and traverse sub-collections and<br />

individual texts, graphics, pictures, videos, or sound files. The navigational apparatus will<br />

frequently employ both text and graphics, with each selectively linked <strong>to</strong> other<br />

navigational apparatus or directly <strong>to</strong> available items in the collection. Much like the<br />

design of the representation of <strong>digital</strong> artifacts, and necessarily related <strong>to</strong> it, designing the


navigational apparatus requires a detailed analysis of each navigational function and steps<br />

<strong>to</strong> ensure that the underlying representation of the data will support it.<br />

Determining and defining the querying, rendering, and navigational functions <strong>to</strong> be<br />

provided <strong>to</strong> users should not be an exclusively abstract or imaginative exercise. Many<br />

designers of publicly accessible information employ user studies <strong>to</strong> inform and guide the<br />

design process. Aesthetics play an important role in communication. Much of the<br />

traditional role of publishers involves the aesthetic design of publications. The "look and<br />

feel" of <strong>digital</strong> publications is no less important. While content is undoubtedly more<br />

important than form, the effective communication of content is undermined when<br />

aesthetics are not taken sufficiently in<strong>to</strong> consideration. Researchers may wish <strong>to</strong> enlist the<br />

assistance of professional designers. While formal, professionally designed studies may<br />

be beyond the means of most researchers, some attempt at gathering information from<br />

users should be made. Interface design should not be deferred <strong>to</strong> the end of the design<br />

process. Creating pro<strong>to</strong>types and mockups of the interface, including designing the visual<br />

and textual apparatus <strong>to</strong> be used in querying, rendering, and navigating the collection and<br />

its parts, will inform the analysis of the artifacts and related objects <strong>to</strong> be <strong>digital</strong>ly<br />

collected, and the encoding systems used in representing them.<br />

Creating and Maintaining Collections<br />

Creating, maintaining, managing, and publishing a <strong>digital</strong> research collection requires an<br />

infrastructure <strong>to</strong> ensure that the ongoing process is efficient, reliable, and controlled.<br />

Building a <strong>digital</strong> collection involves a large number of interrelated activities. Using a<br />

variety of traditional and <strong>digital</strong> finding aids, researchers must discover and locate<br />

primary and secondary resources in private collections and in public and private archives,<br />

museums, and libraries. Intellectual property rights must be negotiated. The researcher<br />

must visit the reposi<strong>to</strong>ry and <strong>digital</strong>ly capture the resource, or arrange for the resource <strong>to</strong><br />

be <strong>digital</strong>ly captured by the reposi<strong>to</strong>ry, or, when possible, arrange <strong>to</strong> borrow the resource<br />

for <strong>digital</strong> capture. After each resource is <strong>digital</strong>ly captured, the capture file or files may<br />

be edited <strong>to</strong> improve quality and fidelity. One or more derivative files may also be<br />

created. Digital derivatives may be simply alternative versions. For example, smaller<br />

image files may be derived from larger image files through sampling or compression <strong>to</strong><br />

facilitate efficient Internet transmission or <strong>to</strong> serve different uses. Derivatives also may<br />

involve alternative representational technologies that enhance the utility or functionality<br />

of the information, for example, an XML-encoded text rather than bitmap images of<br />

original print materials. Creation of enhanced derivatives may constitute the primary<br />

research activity, and typically will involve extended analysis and editing of successive<br />

versions of files. Finally, complex relations within and between files and file components<br />

may be identified and explicitly recorded or encoded, and must be managed and<br />

controlled <strong>to</strong> ensure their persistence. All of these activities will require documentation<br />

and management that in turn will require recording detailed information.<br />

The detailed information needed <strong>to</strong> intellectually and physically manage individual files<br />

and collections of interrelated files can be divided in<strong>to</strong> four principal and interrelated<br />

categories: descriptive, administrative, file, and relations or structural data.


These categories represent the prevailing library model for recording management and<br />

control data, commonly called metadata, and they have been formalized in the Metadata<br />

Encoding and Transmission Standard (METS). Each of these four areas addresses<br />

specific kinds of management and control data. Each is also dependent on the others.<br />

Only when they are related or linked <strong>to</strong> one another will they provide an effective means<br />

of intellectual, legal, and physical control over collections of <strong>digital</strong> materials. These<br />

sorts of data are primarily data-centric, and thus best represented using database<br />

technology.<br />

Descriptive data function as an intellectual surrogate for an object. Descriptive<br />

information is needed <strong>to</strong> identify the intellectual objects used in building as well as<br />

objects <strong>digital</strong>ly represented in a research collection. Documenting and keeping track of<br />

the sources and evidence used in creating a research project is as important in <strong>digital</strong><br />

research as it is in traditional research. For example, if particular sources are used <strong>to</strong><br />

create a database that describes individuals and organizations, those sources need <strong>to</strong> be<br />

documented even though they may not be directly accessible in the collection. Typically,<br />

descriptive information includes author, title, and publisher information and may also<br />

contain subject information. When existing traditional media are digitized, it is necessary<br />

<strong>to</strong> describe both the original and its <strong>digital</strong> representation. While there may be<br />

overlapping information, the two objects are distinct manifestations of the same work.<br />

Traditional media will generally be held in both public reposi<strong>to</strong>ries and private<br />

collections. This is particularly important for unique materials, such as manuscripts and<br />

archival records, but also important for copies, such as published books, as copies are in<br />

fact unique, even if only in subtle ways. The reposi<strong>to</strong>ries and collections, even the private<br />

collection of the researcher, that hold the resources need <strong>to</strong> be recorded and related <strong>to</strong> the<br />

description of each resource used or digitized. When interrelated with administrative, file,<br />

and relations data, descriptive data serves the dual role of documenting the intellectual<br />

content, and attesting <strong>to</strong> the provenance and thus the authenticity of the sources and their<br />

<strong>digital</strong> derivatives.<br />

When linked <strong>to</strong> related descriptive and file data, administrative data enable several<br />

essential activities:<br />

• managing the agents, methods, and technology involved in creating <strong>digital</strong> files;<br />

• ensuring the integrity of those <strong>digital</strong> files;<br />

• tracking their intellectual relation <strong>to</strong> the sources from which they are derived;<br />

• managing the negotiation of rights for use;<br />

• controlling access based on negotiated agreements.<br />

A wide variety of hardware and software exists for capturing traditional media, and the<br />

environmental conditions and context under which capture takes place varies. Digital<br />

capture always results in some loss of information, even if it enhances access and analytic


uses. An image of a page from a book, for example, will not capture the tactile features of<br />

the page. Hardware and software used in capture differ in quality. Each may also, in time,<br />

suffer degradation of performance. Hardware and software also continue <strong>to</strong> develop and<br />

improve. The environmental conditions that affect capture may also vary. For example,<br />

differences in lighting will influence the quality of pho<strong>to</strong>graphs, ambient noise will<br />

influence sound recording. The knowledge and skill of the technician will also affect the<br />

results. Direct capture devices for image, sound, and video have a variety of options<br />

available that will affect both the quality and depth of information captured.<br />

Following capture, versions may be directly derived from the capture file or files, or<br />

alternative forms may be created using different representational methods. Versions of<br />

files are directly derived from the capture file, and represent the same kind of information<br />

(such as pixels or sound samples) but employ different underlying representations that<br />

facilitate different uses of the information. Tag Image File Format (TIFF), for example, is<br />

currently considered the most lossless capture format for pic<strong>to</strong>rial or image data, but they<br />

are generally quite large. JPEG (Joint Pho<strong>to</strong>graphic Experts Group), a compression<br />

technology, is generally used for creating smaller files that are more efficiently<br />

transmitted over the Internet. Such compression results in a loss of information, and thus<br />

impacts the image's intellectual content and its relation <strong>to</strong> the capture file and the original<br />

image. Other kinds of capture involve human judgment and intervention, as the<br />

underlying representations are fundamentally different. A bitmap page image differs<br />

fundamentally from an XML encoding of the text on the page, since the former is a pixel<br />

representation and the latter a character-based encoding. While OCR software can be<br />

used <strong>to</strong> convert pixel representations of characters <strong>to</strong> character representations, manual<br />

correction of recognition errors is still necessary. Accurate encoding of a text's<br />

components requires judgment and manual intervention, even if some encoding is<br />

susceptible <strong>to</strong> machine-assisted processing. The agents, methods, and technology used in<br />

deriving versions and alternative forms of data need <strong>to</strong> be recorded and related <strong>to</strong> each<br />

file in order <strong>to</strong> provide the information necessary <strong>to</strong> evaluate the authenticity and<br />

integrity of the derived data.<br />

Intellectual property law or contract law will protect many of the resources collected and<br />

published in a <strong>digital</strong> <strong>humanities</strong> research project. Use of these materials will require<br />

negotiating with the property owners, and subsequently controlling and managing access<br />

based on agreements. Currently there are several competing initiatives <strong>to</strong> develop<br />

standards for <strong>digital</strong> rights management, and it is <strong>to</strong>o early <strong>to</strong> determine which if any will<br />

satisfy international laws and be widely embraced by users and developers. A successful<br />

standard will lead <strong>to</strong> the availability of software that will enable and moni<strong>to</strong>r access,<br />

including fee-based access and fair use. It is essential for both legal and moral reasons<br />

that researchers providing access <strong>to</strong> protected resources negotiate use of such materials,<br />

record in detail the substance of agreements, and enforce, <strong>to</strong> the extent possible, any<br />

agreed-upon restrictions. It is equally important, however, that researchers assert fair use<br />

when it is appropriate <strong>to</strong> do so: fair use, like a right of way, is a right established in<br />

practice and excessive self-policing will eventually undermine that right. The Stanford<br />

University Library's site on copyright and fair use is a good place <strong>to</strong> look for guidance on


when it is appropriate <strong>to</strong> assert this right (see for more<br />

information).<br />

File data enable management and control of the files used <strong>to</strong> s<strong>to</strong>re <strong>digital</strong> data. When<br />

linked <strong>to</strong> related descriptive data, the intellectual content of files can be identified, and<br />

when linked <strong>to</strong> administrative data, the authenticity, integrity, and quality of the files can<br />

be evaluated, and access <strong>to</strong> copyright-protected materials can be controlled. It is quite<br />

easy <strong>to</strong> overlook the need <strong>to</strong> describe and control files early on in projects. It is easy <strong>to</strong><br />

simply create ad hoc file names and direc<strong>to</strong>ry structures <strong>to</strong> address their management and<br />

control. As projects develop, however, the number of files increases and the use of ad hoc<br />

file names and direc<strong>to</strong>ry structures begins <strong>to</strong> break down. It becomes difficult <strong>to</strong><br />

remember file names, what exactly each file name designates, and where in the direc<strong>to</strong>ry<br />

structure files will be located.<br />

On the other hand, it is quite common <strong>to</strong> create complex, semantically overburdened file<br />

names when using file names and direc<strong>to</strong>ries <strong>to</strong> manage files. Such file names will<br />

typically attempt, in abbreviated form, <strong>to</strong> designate two or more of the following: source<br />

reposi<strong>to</strong>ry, identity of the object represented in a descriptive word, crea<strong>to</strong>r (of the<br />

original, the <strong>digital</strong> file, or both), version, date, and <strong>digital</strong> format. If the naming rules are<br />

not carefully documented, such faceted naming schemes generally collapse quite quickly,<br />

and even when well documented, are rarely effective as the number of files increases.<br />

Such information should be recorded in descriptive and administrative data linked <strong>to</strong> the<br />

address of the s<strong>to</strong>rage location of the file. A complete address will include the address of<br />

the server, the direc<strong>to</strong>ry, and the file name. The name and structure of the direc<strong>to</strong>ry and<br />

file name should be simple, and easily extensible as a collection grows.<br />

There are a wide variety of possible approaches <strong>to</strong> naming and addressing files.<br />

Completely arbitrary naming and addressing schemes are infinitely extensible, though it<br />

is possible <strong>to</strong> design simple, extensible schemes that provide some basic and useful clues<br />

for those managing the files. For example, most projects will collect representations of<br />

many artifacts from many reposi<strong>to</strong>ries. Each artifact may be captured in one or more<br />

<strong>digital</strong> files, with one or more derivative files created from each capture file. In such<br />

scenarios, it is important <strong>to</strong> create records for each reposi<strong>to</strong>ry or collection source. Linked<br />

<strong>to</strong> each of these reposi<strong>to</strong>ry records will be records describing each source artifact. In turn,<br />

each artifact record will be linked <strong>to</strong> one or more records describing capture files, with<br />

each capture record potentially linked <strong>to</strong> one or more records describing derivative files.<br />

Given this scenario, a reasonable file-naming scheme might include an abbreviated<br />

identifier for the reposi<strong>to</strong>ry, followed by three numbers separated by periods identifying<br />

the artifact, the capture, and the derivative. Following file-naming convention, a suffix<br />

designating the notation or format of the file is appended <strong>to</strong> the end. Ideally, the numbers<br />

reflect the order of acquisition or capture. For example, the first artifact <strong>digital</strong>ly acquired<br />

from the Library of Congress might have the base name "loc.OOOl", with "loc" being an<br />

abbreviated identifier for "Library of Congress." The second <strong>digital</strong> file capture of this<br />

artifact would have the base name "loc.0001.02" and the fourth derivative of the original<br />

capture "loc.0001.02.04." A suffix would then be added <strong>to</strong> the base <strong>to</strong> complete the file<br />

name. For example, if the capture file were in the TIFF notation and the derivative in the


JPEG notation, then the file names would be "loc.0001.02.tif" and "loc.0001.02.04.jpg."<br />

Full identification of any file can be achieved by querying the database for the file name.<br />

The direc<strong>to</strong>ry structure might simply consist of a series of folders for each reposi<strong>to</strong>ry.<br />

The file-naming scheme in this example is relatively simple, sustainable, and despite the<br />

weak semantics, very useful, even without querying the database. All files derived from<br />

artifacts in the same reposi<strong>to</strong>ry will be collocated when sorted, and all related versions of<br />

a file will also be collocated. In addition, the order in which artifacts are identified and<br />

collected will be reflected in the sorting order of each reposi<strong>to</strong>ry. The example provided<br />

here is not intended <strong>to</strong> be normative. Varying patterns of collecting, the nature of the<br />

artifacts collected, and other fac<strong>to</strong>rs might lead <strong>to</strong> alternative and more appropriate<br />

schemes. The important lesson, though, is that database technology is ideally suited <strong>to</strong><br />

management and control of intellectual and <strong>digital</strong> assets, and clever naming schemes and<br />

direc<strong>to</strong>ry structures are not. Naming schemes should be semantically and structurally<br />

simple, with linked database records carrying the burden of detailed descriptive and<br />

administrative data.<br />

Interrelating information is important as both an intellectual and a management activity.<br />

Interrelating intellectual objects within and between files requires creating explicit<br />

machine-readable data that allow au<strong>to</strong>mated correlation or collocation of related<br />

resources. For our purposes, there are two types of relations: intrinsic and extrinsic.<br />

Intrinsic relation data support is a feature of the character-based database and markup<br />

technologies. Database technology enables the creation, maintenance, and use of relations<br />

between records or tables within a particular database. XML technology enables the<br />

creation, maintenance, and use of relations between components within a particular<br />

encoded document. Care must be taken in ensuring that the feature is properly employed<br />

<strong>to</strong> ensure the integrity of the relations.<br />

Extrinsic relations are external <strong>to</strong> databases and XML-encoded documents. Relations<br />

between one database and another, between a database and XML documents, between<br />

XML documents, between XML documents and files in other notations or encodings, and<br />

relations between image or sound files and other image or sound files are all types of<br />

extrinsic relations. Extrinsic relations are not currently well supported by standards and<br />

standard-based software. There are a large number of loosely connected or independent<br />

initiatives <strong>to</strong> address the standardization of extrinsic relations. Extrinsic relations thus<br />

present particular challenges in the design of <strong>digital</strong> research collections. Especially<br />

difficult are relations between resources in the collection under the control of the<br />

researcher, and resources under the control of others.<br />

In the absence of standard methods for recording and managing extrinsic relations, it is<br />

necessary <strong>to</strong> design ad hoc methods. Given the complexity of managing relations, it is<br />

difficult <strong>to</strong> provide specific guidance. Both database and markup technologies provide<br />

some support for extrinsic relations, but this support must be augmented <strong>to</strong> provide<br />

reasonable control of relations. Some databases support direct s<strong>to</strong>rage of Binary Large<br />

Objects (BLOBs). Despite the term, BLOBs may include not only binary data, such as<br />

image and sound files, but also complex character data such as XML documents, vec<strong>to</strong>r


graphics, and computer programs. While not standardized, XML Catalog, a standard for<br />

mapping the extrinsic relations of XML documents, provides some support. XML<br />

catalogues, though, primarily represent a method for communicating data about relations<br />

and offer no direct support for managing them effectively. Database technology is<br />

reasonably effective in managing the relations data communicated in XML catalogues.<br />

XQuery offers a standardized but indirect way of articulating extrinsic relations, though it<br />

is not intended as a means of controlling them or ensuring referential integrity. According<br />

<strong>to</strong> the World Wide Web Consortium, XQuery is "designed <strong>to</strong> be a language in which<br />

queries are concise and easily unders<strong>to</strong>od. It is also flexible enough <strong>to</strong> query a broad<br />

spectrum of XML information sources, including both databases and documents" (see<br />

). However, at this point none of these<br />

approaches will be completely reliable and effective without care and vigilance on the<br />

part of the human edi<strong>to</strong>rs and data managers.<br />

Conclusion: Collaboration<br />

In the late twentieth and early twenty-first centuries, the most significant impact of<br />

information technology may be increased collaboration. Collaboration, when successful,<br />

offers many intellectual, professional, and social benefits. A group of scholars working<br />

<strong>to</strong>gether can create research collections more intellectually complex and comprehensive<br />

than is possible for an individual working alone. Complementary and even competing<br />

disciplinary perspectives and specialties will support broader and richer analysis and<br />

understanding. Collaboration will enable greater productivity. Collaboration between<br />

humanists and technologists may lead <strong>to</strong> more profound understandings and more<br />

incisive <strong>to</strong>ols than either would develop by working alone. The explicitness and logical<br />

rigor required <strong>to</strong> represent and use <strong>digital</strong> information exposes <strong>to</strong> criticism many of the<br />

traditional disciplinary assumptions, and often leads <strong>to</strong> a deeper understanding of our<br />

methods and the subjects <strong>to</strong> which they are applied.<br />

Despite these benefits, collaboration, whether with fellow humanists or technologists,<br />

also presents unfamiliar challenges, which require careful attention and time – often more<br />

time than is anticipated at the beginning. The most fundamental challenge of<br />

collaboration is balancing individual interests and shared objectives. Collabora<strong>to</strong>rs need<br />

<strong>to</strong> discuss, negotiate, and clearly document the research and publication objectives, a<br />

process that will require cooperation and compromise. Collabora<strong>to</strong>rs also need <strong>to</strong><br />

recognize individual differences in aptitude for <strong>digital</strong> technology, and differences in<br />

expertise, methodology, and emphasis. Individual responsibilities, obligations, and<br />

production goals need <strong>to</strong> be negotiated and documented, and should reflect expertise and<br />

aptitude. Intellectual and social compatibility, trust, and mutual respect are essential<br />

characteristics of successful collaboration. Individual professional and intellectual needs<br />

and goals must also be recognized. The nature of the academic culture of recognition and<br />

rewards requires that individual scholars be productive and that the scholarship be of high<br />

quality. Individual contributions need <strong>to</strong> be clearly documented, in order <strong>to</strong> provide<br />

reliable evidence for reviewers. Depending upon the nature of the collaboration,<br />

individual contributions may be intermingled, and without careful project design it may<br />

be difficult or impossible <strong>to</strong> reliably indicate who is responsible for what. Such


documentation is essential, for example, in the statements of responsibility incorporated<br />

in<strong>to</strong> the descriptive data correlated with <strong>digital</strong> objects, and in bylines of <strong>digital</strong> objects,<br />

where this is possible.<br />

In collaborative projects, standards and guidelines or manuals will need <strong>to</strong> be augmented<br />

with training. Left alone, different people will interpret and apply guidelines differently.<br />

With respect <strong>to</strong> character-based technologies, in particular text encoding and database<br />

entry, but including geographic information systems (GIS), computer-assisted design<br />

(CAD), and many graphic systems, data entry and encoding will also involve analysis<br />

and interpretation of source materials. Because human judgment is involved, it will be<br />

impossible <strong>to</strong> entirely eliminate differences and <strong>to</strong> achieve absolute, complete intellectual<br />

and technological consistency. Happily, absolute consistency is rarely necessary – it is<br />

generally only necessary <strong>to</strong> have absolute consistency in those data components that are<br />

directly involved in au<strong>to</strong>mated processing. An acceptable range of consistency can<br />

generally be achieved with training. Typically such training, as in all education, will<br />

require both instruction and an initial period of intensive, thorough review, evaluation,<br />

and revision. After intellectual and technological quality reaches an acceptable level,<br />

periodic review will ensure that the level is sustained.<br />

Given the amount of communication involved, it is generally easier <strong>to</strong> collaborate when<br />

all or most of the participants work in the same location. One of the most attractive<br />

benefits of the Internet, though, is the ability <strong>to</strong> communicate with anyone else on the<br />

Internet, regardless of where they are connected, and <strong>to</strong> work remotely in the creation and<br />

maintenance of shared resources. As with all other aspects of designing a research<br />

collection, care must be taken <strong>to</strong> provide an infrastructure <strong>to</strong> support communication and<br />

shared editing of the database. It is advisable <strong>to</strong> set up an e-mail list for project<br />

communication (using software such as Mailman, Majordomo, or Listserv), and <strong>to</strong><br />

archive the correspondence it distributes (software such as Hypermail can au<strong>to</strong>matically<br />

produce a web-accessible version of this archive). An e-mail list will facilitate<br />

broadcasting messages <strong>to</strong> a group. Archiving the messages will serve the purpose of<br />

preserving discussions, negotiations, decisions, and other transactions that document the<br />

creation and maintenance of the collection. Among other things, such an archive will<br />

serve the important function of providing documentary evidence for reviewers of<br />

individual participation and contributions. Shared editing also needs <strong>to</strong> be managed,<br />

typically by a database, <strong>to</strong> control and manage workflow, tracking additions, changes,<br />

and deletions. As attractive and useful as remote collaboration can be, it frequently is still<br />

necessary <strong>to</strong> meet periodically in person for discussions, negotiations, and training.<br />

Collaboration with technologists may also present its own challenges. Although<br />

technologists who elect <strong>to</strong> participate in <strong>digital</strong> <strong>humanities</strong> projects may themselves have<br />

some background in the <strong>humanities</strong>, it will more often be the case that technologists have<br />

little training in <strong>humanities</strong> disciplines. Humanities methods, perspectives, and values<br />

may be strange <strong>to</strong> them, and unfamiliar terminology may hinder communication.<br />

Carefully negotiated and apparently shared understandings will frequently be illusory,<br />

requiring further discussion and renegotiation. Close and successful collaboration will


equire goodwill and persistence, and will rely on developing a shared language and an<br />

enduring mutual understanding.<br />

Depending upon the size and constitution of a collaborative group, it may be necessary <strong>to</strong><br />

formally address issues of administration and supervision. A small number of peers who<br />

are comfortable working with one another may rely only on informal discussion and<br />

consensus. A project driven by one scholar who enlists graduate assistants will implicitly<br />

be led and managed by that scholar. Larger groups, involving scholars of various ranks<br />

and prestige and perhaps also graduate assistants, may require formal designation of a<br />

leader. Even when a particular individual is in charge, a project manager or<br />

administrative assistant may be beneficial, <strong>to</strong> tend <strong>to</strong> coordination and moni<strong>to</strong>ring<br />

deadlines and the like.<br />

Finally, collaboration depends on sustainable project design. It is relatively easy for one<br />

person <strong>to</strong> work, in the short run, in an idiosyncratic framework of his or her own design:<br />

it is more difficult for two people <strong>to</strong> do this, and it becomes increasingly difficult the<br />

more people are involved, and the longer the project continues. If a project is<br />

collaborative, and if it is <strong>to</strong> succeed, it will require the attributes of standardization,<br />

documentation, and thoughtful, iterative design that have been recommended throughout<br />

this chapter, in different contexts.<br />

32.<br />

Conversion of Primary Sources<br />

Marilyn Deegan and Simon Tanner<br />

Material Types<br />

The primary source materials for humanists come in many data forms, and the <strong>digital</strong><br />

capture of that data needs <strong>to</strong> be carefully considered in relation <strong>to</strong> its formats and the<br />

materials that form the substrates upon which it is carried. It is difficult <strong>to</strong> list these<br />

comprehensively, but it is true <strong>to</strong> say that anything that can be pho<strong>to</strong>graphed can be<br />

digitized, and that some materials can be pho<strong>to</strong>graphed or digitized with more fidelity<br />

than others. This is an interesting point, as many digitization projects aim <strong>to</strong> present some<br />

kind of "true" representation of the original, and some materials can get closer <strong>to</strong> this goal<br />

than others.<br />

The materials of interest <strong>to</strong> humanists are likely <strong>to</strong> include the following, and there are<br />

probably many more examples that could be given. However, this list covers a<br />

sufficiently broad area as <strong>to</strong> encompass most of the problems that might arise in<br />

digitization.


Documents<br />

This category covers a huge variety of materials, from all epochs of his<strong>to</strong>ry, and includes<br />

both manuscript and printed artifacts. Manuscripts can be from all periods, in any<br />

language, and written on a huge variety of surfaces: paper, parchment, birch bark,<br />

papyrus, lead tablets, wood, s<strong>to</strong>ne, etc. They will almost certainly have script or character<br />

set issues, and so may require special software for display or analysis. They may be<br />

music manuscripts, which will have issues of notation and representation of a whole<br />

range of special characters; or a series of personal letters, which being loose-leafed will<br />

pose their own organizational problems. Manuscripts may be composite in that they will<br />

have a number of different (often unrelated) texts in them, and there may be difficulties<br />

with bindings. They may be very large or very small – which will have implications for<br />

scanning mechanisms and resolutions. With handwritten texts, there can be no au<strong>to</strong>mated<br />

recognition of characters (although some interesting work is now being done in this area<br />

by the developers of palm computers and tablet PC computers – we need <strong>to</strong> wait <strong>to</strong> see if<br />

the work being done on handwriting recognition for data input feeds in<strong>to</strong> the recognition<br />

of scripts on manuscript materials of the past), and they may also have visual images or<br />

illuminated letters. The documents that humanists might want <strong>to</strong> digitize also include<br />

printed works from the last 500 years, which come in a huge variety: books, journals,<br />

newspapers, posters, letters, typescript, gray literature, musical scores, ephemera,<br />

advertisements, and many other printed sources. Earlier materials and incunabula will<br />

have many of the same problems as manuscripts, and they may be written on paper or<br />

parchment. With printed materials, there can be a wide range of font and typesetting<br />

issues, and there is often illustrative content <strong>to</strong> deal with as well. Printed materials will<br />

also come in a huge range of sizes, from posters <strong>to</strong> postage stamps.<br />

Visual materials<br />

These may be on many different kinds of substrates: canvas, paper, glass, film, fabric,<br />

etc., and for humanists are likely <strong>to</strong> include manuscript images, paintings, drawings,<br />

many different types of pho<strong>to</strong>graphs (film negatives, glass plate negatives, slides, prints),<br />

stained glass, fabrics, maps, architectural drawings, etc.<br />

Three-dimensional objects and artifacts<br />

With present technologies, it is not possible <strong>to</strong> create a "true" facsimile of such objects,<br />

but there are now three-dimensional modeling techniques that can give good<br />

representations of three-dimensional materials. These are likely <strong>to</strong> include the whole<br />

range of museum objects, sculpture, architecture, buildings, archaeological artifacts.<br />

Time-based media<br />

There is increasing interest in the digitization of film, video, and sound, and indeed there<br />

have been large advances in the techniques available <strong>to</strong> produce time-based media in<br />

born-<strong>digital</strong> form in the commercial world, and also <strong>to</strong> digitize analogue originals. The<br />

Star Wars movies are a good example of the his<strong>to</strong>rical transition from analogue <strong>to</strong> <strong>digital</strong>.


The first three movies were filmed <strong>to</strong>tally in analogue and then later <strong>digital</strong>ly remastered;<br />

then all the films were made available on DVD. Finally, the last film <strong>to</strong> be made, Attack<br />

of the Clones, was <strong>to</strong>tally filmed in the <strong>digital</strong> medium.<br />

The conversion of three-dimensional objects and of time-based media is more<br />

problematic than that of text or still images. It draws upon newer technologies, the<br />

standards are not as well supported, and the file sizes produced are very large. The<br />

hardware and software <strong>to</strong> manipulate such materials is also generally more costly.<br />

The Nature of Digital Data<br />

If the nature of <strong>humanities</strong> data is very complex, the nature of <strong>digital</strong> data in their<br />

underlying form is seemingly very simple: all <strong>digital</strong> data, from whatever original they<br />

derive, have the same underlying structure, that of the "bit" or the fonary digit. A bit is an<br />

electronic impulse that can be represented by two states, "on" or "off", also written as "1"<br />

or "0." A "byte" consists of 8 bits, and 1 byte represents 1 alphanumeric character. A 10letter<br />

word, for example, would be 10 bytes. Bits and bytes are linked <strong>to</strong>gether in chains<br />

of millions of electronic impulses; this is known as the "bit stream." A "kilobyte" is 1,024<br />

bytes, and a "megabyte" 1,024 kilobytes. Digital images are represented by "pixels" or<br />

picture elements – dots on the computer screen or printed on paper. Pixels can carry a<br />

range of values, but at the simplest level, one pixel equals one bit, and is represented in<br />

binary form as "black" (off) or "white" (on). Images captured at this level are "bi-<strong>to</strong>nal" –<br />

pure black and white. Images can also be represented as 8-bit images, which have 256<br />

shades of either gray or color and 24-bit images, which have millions of colors – more<br />

than the eye can distinguish. The number of bits chosen <strong>to</strong> represent each pixel is known<br />

as the "bit depth", and devices capable of displaying and printing images of higher bit<br />

depths than this (36 or 48 bits) are now emerging.<br />

Bit depth is the number of bits per pixel, "resolution" is the number of pixels (or printed<br />

dots) per inch, known as ppi or dpi. The higher the resolution, the higher the density of a<br />

<strong>digital</strong> image. The resolution of most computer screens is generally in the range of 75 <strong>to</strong><br />

150 pixels per inch. This is adequate for display purposes (unless the image needs <strong>to</strong> be<br />

enlarged on-screen <strong>to</strong> show fine detail), but visual art or pho<strong>to</strong>graphic content displayed<br />

at this resolution is inadequate for printing (especially in color), though images of black<br />

and white printed text or line art are often acceptable. High-density images of originals<br />

(manuscripts, pho<strong>to</strong>graphs, etc.) need <strong>to</strong> be captured in the range 300–600 ppi for print<br />

quality output. Note that this will depend on the size of the original materials: 35 mm<br />

slides or microfilm originals will need <strong>to</strong> be captured at much higher resolutions, and<br />

scanners are now available offering resolutions of up <strong>to</strong> 4,000 dpi for such materials.<br />

These issues are discussed further below.<br />

Almost any kind of information can be represented in these seemingly simple structures,<br />

as patterns of the most intricate complexity can be built up. Most primary sources as<br />

outlined above are capable of <strong>digital</strong> representation, and when <strong>digital</strong> they are susceptible<br />

<strong>to</strong> manipulation, interrogation, transmission, and cross-linking in ways that are beyond<br />

the capacity of analogue media. Creating an electronic pho<strong>to</strong>copy of a plain page of text


is not a complex technical process with modern equipment, but being able <strong>to</strong> then<br />

au<strong>to</strong>matically recognize all the alphanumeric characters it contains, plus the structural<br />

layout and metadata elements, is a highly sophisticated operation. Alphanumeric symbols<br />

are the easiest objects <strong>to</strong> represent in <strong>digital</strong> form, and <strong>digital</strong> text has been around for as<br />

long as there have been computers.<br />

As will be clear from what is outlined above, there are diverse source materials in the<br />

<strong>humanities</strong> and, using a variety of techniques, they are all amenable <strong>to</strong> <strong>digital</strong> capture.<br />

Nowadays there are many different approaches that can be taken <strong>to</strong> capture such artifacts,<br />

dependent upon (a) the materials themselves; (b) the reasons for capturing them; (c) the<br />

technical and financial resources available <strong>to</strong> the project; and (d) the potential uses and<br />

users. One point that should be emphasized is that <strong>digital</strong> capture, especially capture of<br />

materials with significant image content, is a skilled process that is best entrusted <strong>to</strong><br />

professionals if high-quality archive images are required. However, for the production of<br />

medium- or lower-quality materials for Web delivery or for use in teaching, and for<br />

learning about the issues and processes attendant upon <strong>digital</strong> capture, scholars and<br />

students may find it valuable <strong>to</strong> experiment with the range of medium-cost, good-quality<br />

capture devices now available commercially.<br />

There is rarely only one method for the capture of original source materials, and so<br />

careful planning and assessment of all the cost, quality, conservation, usage, and access<br />

needs for the resultant product needs <strong>to</strong> be done in order <strong>to</strong> decide which of several<br />

options should be chosen. It is vital when making these decisions that provision is made<br />

for the long-term survivability of the materials as well as for immediate project needs: the<br />

costs of sustaining a <strong>digital</strong> resource are usually greater than those of creating it (and are<br />

very difficult <strong>to</strong> estimate) so good planning is essential if the investments made in <strong>digital</strong><br />

conversion are <strong>to</strong> be profitable. (See chapters 31 and 37, this volume, for more<br />

information on project design and long-term preservation of <strong>digital</strong> materials.) However,<br />

when working with sources held by cultural institutions, the range of digitization options<br />

may be limited by what that institution is willing <strong>to</strong> provide: rarely will an institution<br />

allow outsiders <strong>to</strong> work with its holdings <strong>to</strong> create <strong>digital</strong> surrogates, especially of rare,<br />

unique, or fragile materials. Many now offer digitization services as adjuncts or<br />

alternatives <strong>to</strong> their pho<strong>to</strong>graphic services, at varying costs, and projects will need <strong>to</strong><br />

order <strong>digital</strong> files as they would formerly have ordered pho<strong>to</strong>graphs. See Tanner and<br />

Deegan (2002) for a detailed survey of such services in the UK and Europe.<br />

The Advantages and Disadvantages of Digital<br />

Conversion<br />

Digital conversion, done properly, is a difficult, time-consuming and costly business, and<br />

some of the properties of <strong>digital</strong> objects outlined above can be prove <strong>to</strong> be<br />

disadvantageous <strong>to</strong> the presentation and survivability of cultural materials – the resultant<br />

<strong>digital</strong> object, for instance, is evanescent and mutable in a way that the analogue original<br />

isn't. It can disappear in a flash if a hard drive crashes or a CD is corrupt; it can be<br />

changed without trace with an ease that forgers can only dream about. However, the


digitization of resources opens up new modes of use for humanists, enables a much wider<br />

potential audience, and gives a renewed means of viewing our cultural heritage. These<br />

advantages may outweigh the difficulties and disadvantages, provided the project is well<br />

thought out and well managed – and this applies however large or small the project might<br />

be. The advantages of digitization for humanists include:<br />

• the ability <strong>to</strong> republish out-of-print materials<br />

• rapid access <strong>to</strong> materials held remotely<br />

• potential <strong>to</strong> display materials which are in inaccessible formats, for instance, large<br />

volumes or maps<br />

• "virtual reunification" – allowing dispersed collections <strong>to</strong> be brought <strong>to</strong>gether<br />

• the ability <strong>to</strong> enhance <strong>digital</strong> images in terms of size, sharpness, color contrast, noise<br />

reduction, etc.<br />

• the potential for integration in<strong>to</strong> teaching materials<br />

• enhanced searchability, including full text<br />

• integration of different media (images, sounds, video etc.)<br />

• the potential for presenting a critical mass of materials for analysis or comparison.<br />

Any individual, group or institution considering digitization of primary sources will need<br />

<strong>to</strong> evaluate potential digitization projects using criteria such as these. They will also need<br />

<strong>to</strong> assess the actual and potential user base, and consider whether this will change when<br />

materials are made available in <strong>digital</strong> form. Fragile originals which are kept under very<br />

restricted access conditions may have huge appeal <strong>to</strong> a wide audience when made<br />

available in a form which does not damage the originals. A good example of this is the<br />

suffrage banners collection at the Women's Library (formerly known as the Fawcett<br />

Library). This is a unique collection of large-sized women's suffrage banners, many of<br />

which are woven from a variety of materials – cot<strong>to</strong>n and velvet, for instance, often with<br />

applique lettering – which are now in a fragile state. The sheer size of the banners and<br />

their fragility means that viewing the original is heavily restricted, but digitization of the<br />

banners has opened up the potential for much wider access for scholarly use, adding a<br />

vital dimension <strong>to</strong> what is known about suffrage marches in the UK at the beginning of<br />

the twentieth century (see ).<br />

An important question that should be asked at the beginning of any digitization project,<br />

concerns exactly what it is that the digitization is aiming <strong>to</strong> capture. Is the aim <strong>to</strong> produce<br />

a full facsimile of the original that when printed out could stand in for the original? Some<br />

projects have started with that aim, and then found that a huge ancillary benefit was<br />

gained by also having the <strong>digital</strong> file for online access and manipulation. The Digital


Image Archive of Medieval Music (DIAMM) project, for instance, had as its original<br />

goal the capture of a specific corpus of fifteenth-century British polyphony fragments for<br />

printed facsimile publication in volumes such as the Early English Church Music<br />

(EECM) series. However, early studies showed that there was much <strong>to</strong> be gained from<br />

obtaining high-resolution <strong>digital</strong> images in preference <strong>to</strong> slides or prints. This was not<br />

only because of the evidence for growing exploitation of <strong>digital</strong> resources at that time<br />

(1997), but also because many of these fragments were badly damaged and <strong>digital</strong><br />

res<strong>to</strong>ration offered opportunities not possible with conventional pho<strong>to</strong>graphy. The project<br />

therefore decided <strong>to</strong> capture manuscript images in the best quality possible using highend<br />

<strong>digital</strong> imaging equipment, set up according <strong>to</strong> the most rigorous professional<br />

standards; <strong>to</strong> archive the images in an uncompressed form; <strong>to</strong> enhance and reprocess the<br />

images in order <strong>to</strong> wring every possible piece of information from them; <strong>to</strong> preserve all<br />

these images – archive and derivative – for the long term. That has proved <strong>to</strong> be an<br />

excellent strategy for the project, especially as the image enhancement techniques have<br />

revealed hither<strong>to</strong> unknown pieces of music on fragments that had been scraped down and<br />

overwritten with text: digitization has not just enhanced existing <strong>humanities</strong> sources, it<br />

has allowed the discovery of new ones (see ).<br />

Digital techniques can also allow a user experience of unique textual materials that is<br />

simply not possible with the objects themselves. Good-quality images can be integrated<br />

with other media for a more complete user experience. In <strong>2001</strong>, the British Library<br />

produced a high-quality <strong>digital</strong> facsimile of the fifteenth-century Sherborne Missal that is<br />

on display in its galleries on the largest <strong>to</strong>uch screen in the UK. The unique feature of this<br />

resource is that the pages can be turned by hand, and it is possible <strong>to</strong> zoom in at any point<br />

on the page at the <strong>to</strong>uch of a finger on the screen. High-quality sound reproduction<br />

accompanies the images, allowing users <strong>to</strong> hear the religious offices which make up the<br />

text being sung by a monastic choir. This is now available on CD-ROM, and such<br />

facsimiles of other unique objects are now being produced by the British Library. See<br />

.<br />

The Cornell Brittle Books project also started with the aim of recreating analogue<br />

originals: books printed on acid paper were crumbling away, and the goal was <strong>to</strong> replace<br />

these with print surrogates on non-acid paper. Much experimentation with microfilm and<br />

digitization techniques in the course of this project has produced results which have<br />

helped <strong>to</strong> set the benchmarks and standards for the conversion of print materials <strong>to</strong> <strong>digital</strong><br />

formats all around the world. See Chapman et al. (1999) for further details.<br />

Another aim of <strong>digital</strong> conversion might be <strong>to</strong> capture the content of a source without<br />

necessarily capturing its form. So an edition of the work of a literary author might be<br />

rekeyed and re-edited in electronic form without particular reference <strong>to</strong> the visual<br />

characteristics of an existing print or manuscript version. Or, if the aim is <strong>to</strong> add<br />

searchability <strong>to</strong> a written source while preserving the visual form, text might be converted<br />

<strong>to</strong> electronic form and then attached <strong>to</strong> the image.<br />

With a visual source such as a fine art object or early manuscript, what level of<br />

information is needed? The intellectual content or the physical detail of brushstrokes,


canvas grain, the pores of the skin of the animal used <strong>to</strong> make the parchment, scratched<br />

glosses? Is some kind of analysis or reconstruction the aim? The physics department at<br />

the University of Bologna has developed a <strong>digital</strong> x-ray system for the analysis of<br />

paintings (Rossi et al. 2000) and the Beowulf Project, a collaboration between the British<br />

Library and the University of Kentucky, has used advanced imaging techniques for the<br />

recovery of letters and fragments obscured by clumsy repair techniques, and<br />

demonstrated the use of computer imaging <strong>to</strong> res<strong>to</strong>re virtually the hidden letters <strong>to</strong> their<br />

place in the manuscript. See .<br />

With three-dimensional objects, no true representation of three-dimensional space can be<br />

achieved within the two-dimensional confines of a computer screen, but some excellent<br />

results are being achieved using three-dimensional modeling and virtual reality<br />

techniques. The Virtual Harlem project, for instance, has produced a reconstruction of<br />

Harlem, New York, during the time of the Harlem Renaissance in the 1920s, using<br />

modeling and VR immersion techniques. (See <br />

and Carter 1999.) The Cistercians in Yorkshire project is creating imaginative<br />

reconstructions of Cistercian abbeys as they might have been in the Middle Ages, using<br />

three-dimensional modeling and detailed his<strong>to</strong>rical research. (See<br />

.)<br />

Methods of Digital Capture<br />

Text<br />

Humanists have been capturing, analyzing, and presenting textual data in <strong>digital</strong> form for<br />

as long as there have been computers capable of processing alphanumeric symbols.<br />

Thankfully, long gone are the days when pioneers such as Busa and Kenny laboriously<br />

entered text on punchcards for processing on mainframe machines – now it is an<br />

everyday matter for scholars and students <strong>to</strong> sit in the library transcribing text straight<br />

in<strong>to</strong> a lap<strong>to</strong>p or palm<strong>to</strong>p computer, and the advent of tablet PC computers could make this<br />

even easier and more straightforward. Gone, <strong>to</strong>o, are the days when every individual or<br />

project invented codes, systems, or symbols of their own <strong>to</strong> identify special features, and<br />

when any character that could not be represented in ASCII had <strong>to</strong> be receded in some<br />

arcane form. Now the work of standards bodies such as the Text Encoding Initiative<br />

(TEI, ), the World Wide Web Consortium (W3C, ) and<br />

the Dublin Core Metadata Initiative (DCMI, ) has provided<br />

standards and sche-mas for the markup of textual features and for the addition of<br />

metadata <strong>to</strong> non-textual materials that renders them reusable, interoperable, and<br />

exchangable. The work done <strong>to</strong> develop the Unicode standard, <strong>to</strong>o, means that there is a<br />

standard way of encoding characters for most of the languages of the world so that they<br />

<strong>to</strong>o are interoperable and exchangable<br />

().<br />

In considering the different features of electronic text, and the methods of their capture, it<br />

is important <strong>to</strong> make the distinction between machine-readable electronic text and<br />

machine-viewable electronic text. A bitmap of a page, for instance, is machine-


displayable, and can show all the features of the original of which it is a representation.<br />

But it is not susceptible <strong>to</strong> any processing or editing, so though it is human-readable, it is<br />

not machine-readable. In machine-readable text, every individual entity, as well as the<br />

formatting instructions and other codes, is represented separately and is therefore<br />

amenable <strong>to</strong> manipulation. Most electronic texts are closer <strong>to</strong> one of these two forms than<br />

the other, although there are some systems emerging which process text in such a way as<br />

<strong>to</strong> have the elements (and advantages) of both. These are discussed further below.<br />

For the scholar working intensively on an individual text, there is probably no substitute<br />

for transcribing the text directly on <strong>to</strong> the computer and adding appropriate tagging and<br />

metadata according <strong>to</strong> a standardized framework. The great benefit of the work done by<br />

the standards bodies discussed above is that the schemas proposed by them are<br />

extensible: they can be adapted <strong>to</strong> the needs of individual scholars, sources and projects.<br />

Such rekeying can be done using almost any standard word processor or text edi<strong>to</strong>r,<br />

though there are specialist packages like XMetaL which allow the easy production and<br />

validation of text marked up in XML (Extensible Markup Language).<br />

Many scholars and students, empowered by improvements in the technology, are<br />

engaging in larger collaborative projects which require more substantial amounts of text<br />

<strong>to</strong> be captured. There are now some very large projects such as the Early English Books<br />

Online Text Creation Partnership (EEBO TCP, ),<br />

which is creating accurately keyboarded and tagged editions of up <strong>to</strong> 25,000 volumes<br />

from the EEBO corpus of 125,000 volumes of English texts from between 1473 and<br />

1800, which has been microfilmed and digitized by ProQuest. The tagged editions will be<br />

linked <strong>to</strong> the image files. For projects like these, specialist bureaux provide rekeying and<br />

tagging services at acceptable costs with assured levels of accuracy (up <strong>to</strong> 99–995<br />

percent). This accuracy is achieved through double or triple rekeying: two or three<br />

opera<strong>to</strong>rs key the same text, then software is used <strong>to</strong> compare them with each other. Any<br />

differences are highlighted and then they can be corrected manually. This is much faster,<br />

less prone <strong>to</strong> subjective or linguistic errors, and therefore cheaper than proofreading and<br />

correction. There will always be a need for quality control of text produced by any<br />

method, but most specialist bureaux give an excellent, accurate, reliable service: the<br />

competition is fierce in this business, which has driven costs down and quality up. For an<br />

example of a project that effectively used the double rekeying method see the Old Bailey<br />

Court Proceedings, 1674 <strong>to</strong> 1834 ().<br />

Optical character<br />

The capture of text by rekeying either by an individual or by a specialist bureau is<br />

undertaken either because very high levels of accuracy are needed for textual analysis or<br />

for publication, or because the originals are not susceptible <strong>to</strong> any au<strong>to</strong>mated processes,<br />

which is often the case with older materials such as the EEBO corpus discussed above.<br />

While this level of accuracy is sometimes desirable, it comes at a high cost. Where<br />

modern, high-quality printed originals exist, it may be possible <strong>to</strong> capture text using<br />

optical character recognition (OCR) methods, which can give a relatively accurate result.<br />

Accuracy can then be improved using a variety of au<strong>to</strong>mated and manual methods:


passing the text through spellcheckers with specialist dictionaries and thesauri, and<br />

manual proofing. Research is currently being done <strong>to</strong> improve OCR packages and <strong>to</strong><br />

enable them <strong>to</strong> recognize page structures and even add tagging: the European Unionfunded<br />

METAe Project (the Metadata Engine Project) is developing au<strong>to</strong>matic processes<br />

for the recognition of complex textual structures, including text divisions such as<br />

chapters, sub-chapters, page numbers, headlines, footnotes, graphs, caption lines, etc. See<br />

. Olive Software's Active Paper Archive can also recognize<br />

complex page structures from the most difficult of texts: newspapers. This is described<br />

more fully below. OCR techniques were originally developed <strong>to</strong> provide texts for the<br />

blind. OCR engines can operate on a wide range of character sets and fonts, though they<br />

have problems with non-alphabetic character sets because of the large number of<br />

symbols, and also with cursive scripts such as Arabic. Software can be "trained" on new<br />

texts and unfamiliar characters so that accuracy improves over time and across larger<br />

volumes of data. Though this requires human intervention in the early stages, the<br />

improvement in accuracy over large volumes of data is worth the initial effort.<br />

Though OCR can give excellent results if (a) the originals are modern and in good<br />

condition and (b) there is good quality control, projects must consider carefully the costs<br />

and benefits of deciding between a rekeying approach and OCR. Human time is always<br />

the most costly part of any operation, and it can prove <strong>to</strong> be more time-consuming and<br />

costly <strong>to</strong> correct OCR (even when it seems relatively accurate) than <strong>to</strong> go for bureau<br />

rekeying with guaranteed accuracy. It is worth bearing in mind that what seem like<br />

accurate results (between 95 and 99 percent, for instance) would mean that there would<br />

be between 1 and 5 incorrect characters per 100 characters. Assuming there are on<br />

average 5 characters per word then a 1 percent character error rate equates <strong>to</strong> a word error<br />

rate of 1 in 20 or higher.<br />

OCR with fuzzy matching<br />

OCR, as suggested above, is an imperfect method of text capture which can require a<br />

great deal of post-processing if accurate text is <strong>to</strong> be produced. In an ideal world, one<br />

would always aim <strong>to</strong> produce electronic text <strong>to</strong> the highest possible standards of<br />

accuracy; indeed, for some projects and purposes, accurate text <strong>to</strong> the highest level<br />

attainable is essential, and worth what it can cost in terms of time and financial outlay.<br />

However, for other purposes, speed of capture and volume are more important than<br />

quality and so some means has <strong>to</strong> be found <strong>to</strong> overcome the problems of inaccurate OCR.<br />

What needs <strong>to</strong> be taken in<strong>to</strong> account is the reason a text is <strong>to</strong> be captured <strong>digital</strong>ly and<br />

made available. If the text has important structural features which need <strong>to</strong> be encoded in<br />

the <strong>digital</strong> version, and these cannot be captured au<strong>to</strong>matically, or if a definitive edition is<br />

<strong>to</strong> be produced in either print or electronic form from the captured text, then high levels<br />

of accuracy are paramount. If, however, retrieval of the information contained within<br />

large volumes of text is the desired result, then it may be possible <strong>to</strong> work with the raw<br />

OCR output from scanners, without post-processing. A number of text retrieval products<br />

are now available which allow searches <strong>to</strong> be performed on inaccurate text using "fuzzy<br />

matching" techniques. However, at the moment, fuzzy searching will only work with<br />

suitable linguistic and contextual dictionaries, and therefore is pretty much limited <strong>to</strong> the


English language. Other languages with small populations of speakers are poorly<br />

represented, as are pre-1900 linguistic features.<br />

Hybrid solutions: page images with underlying searchable text<br />

An increasing number of projects and institutions are choosing <strong>to</strong> deliver textual content<br />

in <strong>digital</strong> form through a range of hybrid solutions. The user is presented with a facsimile<br />

image of the original for printing and viewing, and attached <strong>to</strong> each page of the work is a<br />

searchable text file. This text file can be produced by rekeying, as in the EEBO project<br />

described above, or by OCR with or without correction. Decisions about the method of<br />

production of the underlying text will depend on the condition of the originals and the<br />

level of accuracy of retrieval required. With the EEBO project, the materials date from<br />

the fifteenth <strong>to</strong> the eighteenth centuries, and they have been scanned from microfilm, so<br />

OCR is not a viable option. The anticipated uses, <strong>to</strong>o, envisage detailed linguistic<br />

research, which means that retrieval of individual words or phrases will be needed by<br />

users. The highest possible level of accuracy is therefore a requirement for this project.<br />

The Forced Migration Online (FMO) project based at the Refugee Studies Centre,<br />

University of Oxford, is taking a different approach. FMO is a portal <strong>to</strong> a whole range of<br />

materials and organizations concerned with the study of the phenomenon of forced<br />

migration worldwide, with content contributed by an international group of partners. One<br />

key component of FMO is a <strong>digital</strong> library of gray literature and of journals in the field.<br />

The <strong>digital</strong> library is produced by attaching text files of uncorrected OCR <strong>to</strong> page images:<br />

the OCR text is used for searching and is hidden from the user; the page images are for<br />

viewing and printing. What is important <strong>to</strong> users of FMO is documents or parts of<br />

documents dealing with key <strong>to</strong>pics, rather than that they can retrieve individual instances<br />

of words or phrases. This type of solution can deliver very large volumes of material at<br />

significantly lower cost than rekeying, but the trade-offs in some loss of accuracy have <strong>to</strong><br />

be unders<strong>to</strong>od and accepted. Some of the OCR inaccuracies can be mitigated by using<br />

fuzzy search algorithms, but this can give rise <strong>to</strong> problems of over-retrieval. FMO<br />

() uses Olive Software's Active Paper Archive, a product<br />

which offers au<strong>to</strong>matic zoning and characterization of complex documents, as well as<br />

OCR and complex search and retrieval using fuzzy matching. See Deegan (2002) and<br />

.<br />

One document type that responds well <strong>to</strong> hybrid solutions is newspapers, which are highvolume,<br />

low-value (generally), mixed media, and usually large in size. Newspaper<br />

digitization is being undertaken by a number of companies, adopting different capture<br />

strategies and business models. ProQuest are using a rekeying solution <strong>to</strong> capture full text<br />

from a number of his<strong>to</strong>ric newspapers, and are selling access <strong>to</strong> these through their portal,<br />

His<strong>to</strong>ry Online (). These newspapers include The<br />

Times, the New York Times and the Wall Street Journal. A number of Canadian<br />

newspapers are digitizing page images and offering searchability using the Cold North<br />

Wind software program (); and the TIDEN project, a<br />

large collaboration of Scandinavian libraries, is capturing content using advanced OCR<br />

and delivering page images with searchability at the page level using a powerful


commercial search engine, Convera's RetrievalWare (). OCLC (the<br />

Online Computer Library Center) has an his<strong>to</strong>ric newspaper digitization service which<br />

provides an end-<strong>to</strong>-end solution: the input is microfilm or TIFF images, the output a fully<br />

searchable archive with each individual item (article, image, advertisement) separated<br />

and marked up in XML. Searches may be carried out on articles, elements within articles<br />

(title, byline, text), and image captions. OCLC also uses Olive Software's Active Paper<br />

Archive.<br />

Images<br />

As we suggest above, humanists deal with a very large range of image-based materials,<br />

and there will be different strategies for capture that could be employed, according <strong>to</strong><br />

potential usage and cost fac<strong>to</strong>rs. Many humanists, <strong>to</strong>o, may be considering <strong>digital</strong> images<br />

as primary source materials rather than as secondary surrogates: increasingly<br />

pho<strong>to</strong>graphers are turning from film <strong>to</strong> <strong>digital</strong>, and artists are creating <strong>digital</strong> art works<br />

from scratch. Many <strong>digital</strong> images needed by humanists are taken from items outside<br />

their control: objects that are held in cultural institutions. These institutions have their<br />

own facilities for creating images which scholars and students will need <strong>to</strong> use. If they<br />

don't have such facilities, analogue surrogates (usually pho<strong>to</strong>graphic) can be ordered and<br />

digitization done from the surrogate. The costs charged by institutions vary a great deal<br />

(see Tanner and Deegan 2002). Sometimes this depends on whether a reproduction fee is<br />

being charged. Institutions may offer a bewildering choice of <strong>digital</strong> and analogue<br />

surrogate formats with a concomitant complexity of costs. A thorough understanding of<br />

the materials, the digitization technologies and the implications of any technical choices,<br />

the metadata that need <strong>to</strong> be added <strong>to</strong> render the images useful and usable (which is<br />

something that will almost certainly be added by the scholar or student), and what the<br />

potential cost fac<strong>to</strong>rs might be are all essential for humanists embarking on <strong>digital</strong><br />

projects, whether they will be doing their own capture or not. Understanding of the<br />

materials is taken as a given here, so this section will concentrate upon technical issues.<br />

Technical issues in image capture<br />

Most image materials <strong>to</strong> be captured by humanists will need a high level of fidelity <strong>to</strong> the<br />

original. This means that capture should be at an appropriate resolution, relative <strong>to</strong> the<br />

format and size of the original, and at an appropriate bit depth. As outlined above,<br />

resolution is the measure of information density in an electronic image and is usually<br />

measured in dots per inch (dpi) or pixels per inch (ppi). The more dots per inch there are,<br />

the denser the image information provided. This can lead <strong>to</strong> high levels of detail being<br />

captured at higher resolutions. The definition of "high" resolution is based upon fac<strong>to</strong>rs<br />

such as original media size, the nature of the information, and the eventual use. Therefore<br />

600 dpi would be considered high-resolution for a pho<strong>to</strong>graphic print, but would be<br />

considered low-resolution for a 35 mm slide. Careful consideration must be given <strong>to</strong> the<br />

amount of information content required over the eventual file size. It must be<br />

remembered that resolution is always a fac<strong>to</strong>r of two things: (1) the size of the original<br />

and (2) the number of dots or pixels. Resolution is expressed in different ways according<br />

<strong>to</strong> what particular part of the <strong>digital</strong> process is being discussed: hardware capabilities,


absolute value of the <strong>digital</strong> image, capture of analogue original, or printed output.<br />

Hardware capabilities are referred <strong>to</strong> differently as well: resolution calculated for a<br />

flatbed scanner, which has a fixed relationship with originals (because they are placed on<br />

a fixed platen and the scanner passes over them at a fixed distance) is expressed in dpi.<br />

With <strong>digital</strong> cameras, which have a variable dpi in relation <strong>to</strong> the originals, given that<br />

they can be moved closer or further away, resolution is expressed in absolute terms,<br />

either by their x and y dimensions (12,000 × 12,000, say, for the highest-quality<br />

professional <strong>digital</strong> cameras) or by the <strong>to</strong>tal number of pixels (4 million, for instance, for<br />

a good-quality, compact camera). The <strong>digital</strong> image itself is best expressed in absolute<br />

terms: if expressed in dpi, the size of the original always needs <strong>to</strong> be known <strong>to</strong> be<br />

meaningful.<br />

The current choices of hardware for <strong>digital</strong> capture include flatbed scanners, which are<br />

used for reflective and transmissive materials. These can currently deliver up <strong>to</strong> 5,000<br />

dpi, but can cost tens of thousands of dollars, so most projects can realistically only<br />

afford scanners in the high-end range of 2,400 <strong>to</strong> 3,000 dpi. Bespoke 35 mm film<br />

scanners, which are used for transmissive materials such as slides and film negatives, can<br />

deliver up <strong>to</strong> 4,000 dpi. Drum scanners may also be considered as they can deliver much<br />

higher relative resolutions and quality, but they are generally not used in this context as<br />

the process is destructive <strong>to</strong> the original pho<strong>to</strong>graphic transparency and the unit cost of<br />

creation is higher. Digital cameras can be used for any kind of material, but are generally<br />

recommended for those materials not suitable for scanning with flatbed or film scanners:<br />

tightly bound books or manuscripts, art images, three-dimensional objects such as<br />

sculpture or architecture. Digital cameras are becoming popular as replacements for<br />

conventional film cameras in the domestic and professional markets, and so there is now<br />

a huge choice. High-end cameras for purchase by image studios cost tens of thousands of<br />

dollars, but such have been the recent advances in the technologies that superb results can<br />

be gained from cameras costing much less than this – when capturing images from<br />

smaller originals, even some of the compact cameras can deliver archive-quality scans.<br />

However, they need <strong>to</strong> be set up professionally, and professional stands and lighting must<br />

be used.<br />

For color scanning, the current recommendation for bit depth is that high-quality<br />

originals be captured at 24 bit, which renders more than 16 million colors – more than the<br />

human eye can distinguish, and enough <strong>to</strong> give pho<strong>to</strong>realistic output when printed. For<br />

black and white materials with <strong>to</strong>ne, 8 bits per pixel is recommended, which gives 256<br />

levels of gray, enough <strong>to</strong> give pho<strong>to</strong>realistic printed output.<br />

The kinds of images humanists will need for most purposes are likely <strong>to</strong> be of the highest<br />

possible quality for two reasons. First, humanists will generally need fine levels of detail<br />

in the images, and, secondly, the images will in many cases have been taken from rare or<br />

unique originals, which might also be very fragile. Digital capture, wherever possible,<br />

should be done once only, and a <strong>digital</strong> surrogate captured that will satisfy all anticipated<br />

present and future uses. This surrogate is known as the "<strong>digital</strong> master" and should be<br />

kept under preservation conditions. (See chapter 37, this volume.) Any manipulations or<br />

post-processing should be carried out on copies of this master image. The <strong>digital</strong> master


will probably be a very large file: the highest-quality <strong>digital</strong> cameras (12,000 × 12,000<br />

pixels) produce files of up <strong>to</strong> 350Mb, which means that it is not possible <strong>to</strong> s<strong>to</strong>re more<br />

than one on a regular CD-ROM disk. A 35 mm color transparency captured at 2,700 dpi<br />

(the norm for most slide scanners) in 24-bit color would give a file size of 25 Mb, which<br />

means that around 22 images could be s<strong>to</strong>red on one CD-ROM. The file format generally<br />

used for <strong>digital</strong> masters is the TIFF (Tagged Image File Format), a de fac<strong>to</strong> standard for<br />

<strong>digital</strong> imaging. There are many other file formats available, but TIFF can be<br />

recommended as the safest choice for the long term. The "Tagged" in the title means that<br />

various types of information can be s<strong>to</strong>red in a file header of the TIFF files.<br />

The actual methods of operation of scanning equipment vary, and it is not the intention of<br />

this chapter <strong>to</strong> give details of such methodologies. There are a number of books which<br />

give general advice, and hardware manufacturers have training manuals and online help.<br />

Anyone who wishes <strong>to</strong> learn <strong>digital</strong> capture techniques in detail is advised <strong>to</strong> seek out<br />

professional training through their own institution, or from one of the specialist<br />

organizations offering courses: some of these are listed in the bibliography at the end of<br />

this chapter.<br />

Compression and derivatives<br />

It is possible <strong>to</strong> reduce file sizes of <strong>digital</strong> images using compression techniques, though<br />

this is often not recommended for the <strong>digital</strong> masters. Compression comes in two forms:<br />

"lossless", meaning that there is no loss of data through the process, and "lossy", meaning<br />

that data is lost, and can never be recovered. There are two lossless compression<br />

techniques that are often used for TIFF master files, and which can be recommended here<br />

– the LZW compression algorithm for materials with color content, and the CCITT<br />

Group 4 format for materials with 1-bit, black and white content.<br />

Derivative images from <strong>digital</strong> masters are usually created using lossy compression<br />

methods, which can give much greater reduction in file sizes than lossless compression<br />

for color and greyscale images. Lossy compression is acceptable for many uses of the<br />

images, especially for Web or CD-ROM delivery purposes. However, excessive<br />

compression can cause problems in the viewable images, creating artifacts such as<br />

pixelation, dotted or stepped lines, regularly repeated patterns, moire, halos, etc. For the<br />

scholar seeking the highest level of fidelity <strong>to</strong> the originals, this is likely <strong>to</strong> be<br />

unacceptable, and so experimentation will be needed <strong>to</strong> give the best compromise<br />

between file size and visual quality. The main format for derivative images for delivery<br />

<strong>to</strong> the web or on CD-ROM is currently JPEG. This is a lossy compression format that can<br />

offer considerable reduction in file sizes if the highest levels of compression are used, but<br />

this comes at the cost of some compromise of quality. However, it can give color<br />

thumbnail images of only around 7 KB and screen resolution images of around 60 KB – a<br />

considerable benefit if bandwidth is an issue.


Audio and video capture<br />

As technology has advanced since the 1990s, more and more humanists are becoming<br />

interested in the capture of time-based media. Media studies is an important and growing<br />

area, and his<strong>to</strong>rians of the modern period <strong>to</strong>o derive great benefit from having <strong>digital</strong><br />

access <strong>to</strong> time-based primary sources such as news reports, film, etc. Literary scholars<br />

also benefit greatly from access <strong>to</strong> plays, and <strong>to</strong> filmed versions of literary works.<br />

In principle the conversion of audio or video from an analogue medium <strong>to</strong> a <strong>digital</strong> data<br />

file is simple, it is in the detail that complexity occurs. The biggest problem caused in the<br />

<strong>digital</strong> capture of video and audio is the resultant file sizes:<br />

There is no other way <strong>to</strong> say it. Video takes up a lot of everything; bandwidth, s<strong>to</strong>rage<br />

space, time, and money. We might like <strong>to</strong> think that all things <strong>digital</strong> are preferable <strong>to</strong> all<br />

things analog but the brutal truth is that while analog formats like video might not be<br />

exact or precise they are remarkably efficient when it comes <strong>to</strong> s<strong>to</strong>ring and transmitting<br />

vast amounts of information.<br />

(Wright <strong>2001</strong>a)<br />

Raw, uncompressed video is about 21 MB/sec.<br />

(Wright <strong>2001</strong>b)<br />

There are suppliers that can convert audio and video, and many scholars will want <strong>to</strong><br />

outsource such work. However, an understanding of the processes is of great importance.<br />

The first stage in the capture process is <strong>to</strong> have the original video or audio in a format that<br />

is convertible <strong>to</strong> <strong>digital</strong> formats. There are over thirty types of film and video s<strong>to</strong>ck and<br />

many types of audio formats that the original content may be recorded upon. Most<br />

humanists, however, should only need <strong>to</strong> concern themselves with a limited number of<br />

formats, including:<br />

• VHS or Betamax video<br />

• mini-DV for <strong>digital</strong> video cameras<br />

• DAT tape for audio<br />

• cassette tape for audio<br />

• CD-ROM for audio<br />

The equipment needed <strong>to</strong> capture video and audio includes:<br />

• video player capable of high-quality playback of VHS and mini-DV tapes


• hi-fi audio player capable of high-quality playback of DAT and cassette tapes<br />

• small TV for review and location of video clips<br />

• high-quality speakers and/or headphones for review and location of audio clips<br />

Once the suitable video or audio clip has been identified on the original via a playback<br />

device, this has <strong>to</strong> be connected <strong>to</strong> the capture device for rendering in a <strong>digital</strong> video<br />

(DV) or <strong>digital</strong> audio (DA) format upon a computer system. This is normally achieved by<br />

connecting input/output leads from the playback device <strong>to</strong> an integrated <strong>digital</strong> capture<br />

card in the computer or via connection <strong>to</strong> a device known as a "breakout box." The<br />

breakout box merely allows for more types of input/output leads <strong>to</strong> be used in connecting<br />

<strong>to</strong> and from the playback device. Whatever the method of connection used, there must be<br />

a capture card resident in the machine <strong>to</strong> allow for <strong>digital</strong> data capture from the playback<br />

device. Usually, a separate capture card is recommended for video and for audio:<br />

although all video capture cards will accept audio input, the quality required for pure<br />

audio capture is such that a bespoke capture card for this purpose is recommended;<br />

indeed, it is probably best <strong>to</strong> plan the capture of audio and video on separate PCs – each<br />

card requires its own Input/Output range, Interrupt Request number and its own software<br />

<strong>to</strong> control the capture. These are temperamental and tend <strong>to</strong> want <strong>to</strong> occupy the same I/O<br />

range, interrupts and memory spaces, leading <strong>to</strong> system conflicts and reliability problems.<br />

This of course adds <strong>to</strong> cost if a project is planning <strong>to</strong> capture both audio and video.<br />

The benefit of having a capture card is that it introduces render-free, real-time <strong>digital</strong><br />

video editing, DV and analogue input/output, and this is usually augmented with direct<br />

output for DVD, and it also provides Internet streaming capabilities, along with a suite of<br />

<strong>digital</strong> video production <strong>to</strong>ols. Render-free real-time capture and editing is important<br />

because capture and editing can then happen without the long waits (which can be many<br />

minutes) associated with software packages for rendering and editing. The capture from<br />

video or audio will be <strong>to</strong> DV or DA, usually at a compression rate of around 5:1 which<br />

gives 2.5–3.5 MB/sec. This capture is <strong>to</strong> a high standard even though there is<br />

compression built in<strong>to</strong> the process. It is also <strong>to</strong>o large for display and use on the Internet<br />

and so further compression and file format changes are required once edited in<strong>to</strong> a<br />

suitable form for delivery.<br />

Editing of Captured Content<br />

The editing of the content captured is done via software <strong>to</strong>ols known as non-linear editing<br />

suites. These allow the content <strong>to</strong> be manipulated, edited, spliced, and otherwise changed<br />

<strong>to</strong> facilitate the production of suitable content for the prospective end user. The ability <strong>to</strong><br />

do this in real time is essential <strong>to</strong> the speed and accuracy of the eventual output. Also, the<br />

editing suite should have suitable compressors for output <strong>to</strong> Web formats and Internet<br />

streaming.


Compressors for the Web<br />

When viewing video or listening <strong>to</strong> audio on the Web the content has <strong>to</strong> go through a<br />

process of compression and de-compression (CODEC). There is initial compression at<br />

the production end using a suitable CODEC (e.g., Sorenson, Qdesign) <strong>to</strong> gain the desired<br />

file size, frames per second, transmission rate, and quality. This content may then be<br />

saved in a file format suitable for the architecture of the delivery network and expected<br />

user environment – possibly QuickTime, Windows Media or Real. The user then uses<br />

viewer software <strong>to</strong> decompress the content and view/hear the content.<br />

Compression is a difficult balancing act between gaining the smallest file size and<br />

retaining suitable quality. In video, compression works by taking a reference frame that<br />

contains all the information available and then subsequent frames are represented as<br />

changes from the reference frame. The number of frames between the reference frames is<br />

a defining fac<strong>to</strong>r in the compression and quality – the fewer reference frames the more<br />

compressed the file but the lower quality the images are likely <strong>to</strong> be. However, if the<br />

number of reference frames is increased <strong>to</strong> improve visual quality then the file size will<br />

also rise. As the nature of the content in video can differ radically, this compression<br />

process has <strong>to</strong> be done on a case by case basis. A video of a "talking head" type interview<br />

will require fewer reference frames than, say, sports content because the amount of<br />

change from frame <strong>to</strong> frame is less in the "talking head" example. Thus higher<br />

compression is possible for some content than for others if aiming for the same visual<br />

quality – there is no "one size fits all" equation in audio and video compression.<br />

It is generally assumed with images that the main cost will be in the initial capture and<br />

that creating surrogates for end-user viewing on the Internet will be quick and cheap. In<br />

the video sphere this paradigm is turned on its head, with the initial capture usually<br />

cheaper than the creation of the compressed surrogate for end-user viewing as a direct<br />

consequence of the increased amount of human intervention needed <strong>to</strong> make a goodquality<br />

Internet version at low bandwidth.<br />

Metadata<br />

Given that humanists will almost certainly be capturing primary source data at a high<br />

quality and with long-term survival as a key goal, it is important that this data be<br />

documented properly so that cura<strong>to</strong>rs and users of the future understand what it is that<br />

they are dealing with. Metadata is one of the critical components of <strong>digital</strong> resource<br />

conversion and use, and is needed at all stages in the creation and management of the<br />

resource. Any crea<strong>to</strong>r of <strong>digital</strong> objects should take as much care in the creation of the<br />

metadata as they do in the creation of the data itself – time and effort expended at the<br />

creation stage recording good-quality metadata is likely <strong>to</strong> save users much grief, and <strong>to</strong><br />

result in a well-formed <strong>digital</strong> object which will survive for the long term.<br />

It is vitally important that projects and individuals give a great deal of thought <strong>to</strong> this<br />

documentation of data right from the start. Having archive-quality <strong>digital</strong> master files is<br />

useless if the filenames mean nothing <strong>to</strong> anyone but the crea<strong>to</strong>r, and there is no indication


of date of creation, file format, type of compression, etc. Such information is known as<br />

technical or administrative metadata. Descriptive metadata refers <strong>to</strong> the attributes of the<br />

object being described and can be extensive: attributes such as: "title", "crea<strong>to</strong>r",<br />

"subject", "date", "keywords", "abstract", etc. In fact, many of the things that would be<br />

catalogued in a traditional cataloguing system. If materials are being captured by<br />

professional studios or bureaux, then some administrative and technical metadata will<br />

probably be added at source. Or it may be possible <strong>to</strong> request or supply project-specific<br />

metadata. Descriptive metadata can only be added by experts who understand the nature<br />

of the source materials, and it is an intellectually challenging task in itself <strong>to</strong> produce<br />

good descriptive metadata.<br />

Publications<br />

Understanding the capture processes for primary source materials is essential for<br />

humanists intending <strong>to</strong> engage in <strong>digital</strong> projects, even if they are never going <strong>to</strong> carry out<br />

conversion activities directly. Knowing the implications of the various decisions that<br />

have <strong>to</strong> be taken in any project is of vital importance for short- and long-term costs as<br />

well as for the long-term survivability of the materials <strong>to</strong> which time, care, and funds<br />

have been devoted. The references below cover most aspects of conversion for most<br />

types of <strong>humanities</strong> sources, and there are also many websites with bibliographies and<br />

reports on <strong>digital</strong> conversion methods. The technologies and methods change constantly,<br />

but the underlying principles outlined here should endure.<br />

References for Further Reading<br />

Baca, M., (ed.) (1998). Introduction <strong>to</strong> Metadata, Pathways <strong>to</strong> Digital Information. Los<br />

Angeles, CA: Getty Research Institute.<br />

Carter, B. (1999). From Imagination <strong>to</strong> Reality: Using Immersion Technology in an<br />

African American Literature Course. Literary and Linguistic Computing 14: 55–65.<br />

Chapman, S., P. Conway, and A. R. Kenney (1999). Digital Imaging and Preservation<br />

Microfilm: The Future of the Hybrid Approach for the Preservation of Brittle Books.<br />

RLG DigiNews. Available at http://www.rlg.org/preserv/diginews/diginews3-1.html.<br />

Davies, A. and P. Fennessey (<strong>2001</strong>). Digital Imaging for Pho<strong>to</strong>graphers. London: Focal<br />

Press.<br />

Deegan, M., E. Steinvel, and E. King (2002). Digitizing His<strong>to</strong>ric Newspapers: Progress<br />

and Prospects. RLG DigiNews. Available at<br />

http://www.rlg.org/preserv/diginews/diginews6-4.html#feature2.<br />

Deegan, M. and S. Tanner (2002). Digital Futures: Strategies for the Information Age.<br />

London: Library Association Publishing.


Feeney, M., (ed.) (1999). Digital Culture: Maximising the Nation's Investment. London:<br />

National Preservation Office.<br />

Getz, M. (1997). Evaluating Digital Strategies for S<strong>to</strong>ring and Retrieving Scholarly<br />

Information. In S. H. Lee (ed.), Economics of Digital Information: Collection, S<strong>to</strong>rage<br />

and Delivery. Bingham<strong>to</strong>n, NY: Haworth Press.<br />

Gould, S. and R. Ebdon (1999). IFLA/UNESCO Survey on Digitisation and Preservation.<br />

IFLA Offices for UAP and International Lending Available at<br />

http://www.unesco.org/webworld/mdm/survey_index_en.html.<br />

Hazen, D., J. Horrell, and J. Merrill-Oldham (1998). Selecting Research Collections for<br />

Digitization CLIR. Available at http://www.clir.org/pubs/reports/hazen/pub74.html.<br />

Kenney, A. R. and O. Y. Rieger, (eds.) (2000). Moving Theory in<strong>to</strong> Practice: Digital<br />

Imaging for Libraries and Archives. Mountain View, CA: Research Libraries Group.<br />

Klijn, E. and Y. de Lusenet (2000). In the Picture: Preservation and Digitisation of<br />

European Pho<strong>to</strong>graphic Collections. Amsterdam: Koninklijke Bibliotheek.<br />

Lacey, J. (2002). The Complete Guide <strong>to</strong> Digital Imaging. London: Thames and Hudson.<br />

Lagoze, C. and S. Payette (2000). Metadata: Principles, Practices and Challenges. In A.<br />

R. Kenney and O. Y. Rieger (eds.), Moving Theory in<strong>to</strong> Practice: Digital Imaging for<br />

Libraries and Archives (pp. 84–100). Mountain View, CA: Research Libraries Group.<br />

Lawrence, G. W., et al. (2000). Risk Management of Digital Information: A File Format<br />

Investigation. Council on Library and Information Resources. Available at<br />

http://www.clir.org.pubs/reports/reports.html.<br />

Parry, D. (1998). Virtually New: Creating the Digital Collection. London: Library and<br />

Information Commission.<br />

Robinson, P. (1993). The Digitization of Primary Textual Sources. London: Office for<br />

Humanities Communication Publications, 3, King's College, London.<br />

Rossi, M. F. Casali, A. Bacchilega, and D. Romani (2000). An Experimental X-ray<br />

Digital Detec<strong>to</strong>r for Investigation of Paintings. 15th World Conference on<br />

NonDestructive Testing, Rome, 15–21 Oc<strong>to</strong>ber.<br />

Smith, A. (1999). Why Digitize? CLIR. Available at<br />

http://www.clir.org/pubs/abstract/pub80.html.<br />

Tanner, S. and M. Deegan (2002). Exploring Charging Models for Digital Cultural<br />

Heritage. Available at http://heds.herts.ac.uk.


Watkinson, J. (<strong>2001</strong>). Introduction <strong>to</strong> Digital Video. London: Focal Press.<br />

Wright, G. (<strong>2001</strong>a). Building a Digital Video Capture System. Part I, Tom's Hardware<br />

Guide. Available at http://www.<strong>to</strong>mshardware.com/video/<strong>2001</strong>0524/.<br />

Wright, G. (<strong>2001</strong>b). Building a Digital Video Capture System. Part II, Tom's Hardware<br />

Guide. Available at http://www.<strong>to</strong>mshardware.com/video/<strong>2001</strong>0801/.<br />

Courses<br />

School for Scanning, North East Documentation Center, .<br />

Cornell University Library, Department of Preservation and Conservation,<br />

.<br />

SEPIA (Safeguarding European Pho<strong>to</strong>graphic Images for Access), project training<br />

courses, .<br />

33.<br />

Text Tools<br />

John Bradley<br />

One should not be surprised that a central interest of computing humanists is <strong>to</strong>ols <strong>to</strong><br />

manipulate text. For this community, the purpose of digitizing text is <strong>to</strong> allow the text <strong>to</strong><br />

be manipulated for scholarly purposes, and text <strong>to</strong>ols provide the mechanisms <strong>to</strong> support<br />

this. Some of the scholarly potential of digitizing a text has been recognized from the<br />

earliest days of computing. Father Rober<strong>to</strong> Busa's work on the Index Thomisticus (which<br />

began in the 1940s) involved the indexing of the writings of Thomas Aquinas, and arose<br />

out of work for his doc<strong>to</strong>ral thesis when he discovered that he needed <strong>to</strong> systematically<br />

examine the uses of the preposition in. Busa reports that because of this he imagined<br />

using "some sort of machinery" that would make it possible. Later, but still relatively<br />

early in the his<strong>to</strong>ry of <strong>humanities</strong> computing, the writings of John B. Smith recognized<br />

some of the special significance of text in digitized form (see a relatively late piece<br />

entitled "Computer Criticism", Smith 1989). Indeed, even though computing <strong>to</strong>ols for<br />

manipulating text have been with us for many years, we are still in the early stages of<br />

understanding the full significance of working with texts electronically.<br />

At one time an important text <strong>to</strong>ol that one would have described in an article such as this<br />

would have been the word processor. Today, however, the word processor has become<br />

ubiqui<strong>to</strong>us both inside and outside the academy and it is no longer necessary <strong>to</strong> describe<br />

its benefits or categorize its applications. Today many people believe that the World<br />

Wide Web can be at the center of computing for humanists, and much discussion has<br />

focused on the significance of presenting materials on the WWW for a worldwide


audience. Although some of the <strong>to</strong>ols I will describe here could greatly facilitate the<br />

preparation of materials for the WWW, I will not be writing very much about HTML or<br />

the significance of using it. Instead, this article will focus on <strong>to</strong>ols that can manipulate<br />

texts in ways that might usefully support the development of scholarship on those texts<br />

rather than merely prepare and present them.<br />

Some of the <strong>to</strong>ols described here have been developed specifically <strong>to</strong> support <strong>humanities</strong><br />

scholarship. The intellectual origins of many in this group flow directly from the wordoriented<br />

work of Father Busa and other early researchers, and these ones tend <strong>to</strong> be<br />

mainly designed <strong>to</strong> support a certain kind of word- or wordform-oriented inquiry.<br />

Another important group of <strong>to</strong>ols described later were not developed by humanists and<br />

have a broad range of applications in general computing, although they have proven<br />

themselves <strong>to</strong> be powerful <strong>to</strong>ols when applied <strong>to</strong> scholarly tasks. The chapter finishes<br />

with a brief description of TuStep – a system developed by Wilhelm Ott that is<br />

potentially useful in a range of applications within and outside the <strong>humanities</strong>, but was<br />

developed with a <strong>humanities</strong> focus.<br />

Not surprisingly, given their very different origins, the two groups present very different<br />

challenges <strong>to</strong> the <strong>humanities</strong> researcher who is beginning <strong>to</strong> use them. The first group of<br />

<strong>to</strong>ols, generally designed <strong>to</strong> support text analysis, are relatively straightforward <strong>to</strong><br />

understand once one is operating within a tightly defined problem domain. The generalpurpose<br />

<strong>to</strong>ols, however, are not so easily categorized, and can perform a very broad range<br />

of tasks on a text. Because they are not focused on a specific problem domain, as the<br />

"Text Analysis" <strong>to</strong>ols are, they contain very abstract components, and the difficulty for<br />

the beginner is <strong>to</strong> understand whether, and then how, these abstract elements can be<br />

applied <strong>to</strong> any particular problem.<br />

Those who begin with the <strong>to</strong>ols that are specially developed for textual analysis or critical<br />

work often find that they wish <strong>to</strong> do something that goes beyond what their software does<br />

– for example, apply a statistical program <strong>to</strong> some materials that the software has given<br />

them. Often the form of the data generated by their first program is different from that<br />

needed by the second program. As we will see, one of the important tasks of the generalpurpose<br />

<strong>to</strong>ols described later in this chapter is <strong>to</strong> provide ways <strong>to</strong> bridge these gaps.<br />

Tools <strong>to</strong> Support Text Analysis<br />

Paradoxically, the task of writing general-purpose software is highly specialized; it is also<br />

both costly and difficult. Perhaps for this reason most <strong>humanities</strong> computing software<br />

available "off the shelf" (with the exception of TuStep) performs a relatively simple set of<br />

functions. Of those pieces of software that do exist, many are designed <strong>to</strong> assist with text<br />

analysis.<br />

Wordform-oriented <strong>to</strong>ols<br />

Most of this section describes software that supports the analysis of texts through wordforms,<br />

and relies on the KWIC (Key Word In Context) and related displays. These <strong>to</strong>ols


are the most accessible for new users, and often are taught <strong>to</strong> students. See, for example,<br />

Hawthorne's (1994) article for examples of student uses of TACT. An ances<strong>to</strong>r of many<br />

of the currently available wordform-oriented programs is the Oxford Concordance<br />

Program (OCP), with which they share a number of concepts.<br />

All the "wordform" programs described in this section provide a relatively limited set of<br />

operations. Nonetheless, there are reports of them being useful in a broad range of textresearch<br />

activities, including:<br />

• language teaching and learning;<br />

• literary research;<br />

• linguistic research, including corpus linguistics, translation and language engineering,<br />

supported at least at a relatively basic level (although see the next paragraph);<br />

• lexicography; and<br />

• content analysis in many disciplines including those from both the <strong>humanities</strong> and the<br />

social sciences.<br />

As the category suggests, they all operate on wordforms and <strong>to</strong> work they must be able <strong>to</strong><br />

au<strong>to</strong>matically collect the wordforms from the provided texts. Most of the programs<br />

described here are thus oriented <strong>to</strong>wards alphabetic languages, and allow one <strong>to</strong> identify<br />

both the characters that are acting as letters in the target language, and <strong>to</strong> identify<br />

characters that may appear in a word but are not letters (such as a hyphen).<br />

At the center of most of these programs is the word list – a list of all wordforms that<br />

occur in the processed text, generally shown with a count of how many times that<br />

wordform occurred and usually sorted in alphabetical order or by frequency, with the<br />

most frequently occurring forms listed first. In most programs, users can ask <strong>to</strong> view<br />

information about a wordform by selecting it from this list, e.g., "show me all<br />

occurrences of the word 'avuncular.'" In some programs one types in a query <strong>to</strong> select the<br />

words one wants <strong>to</strong> view. Wildcard schemes in queries can support word stemming, e.g.,<br />

"show me all occurrences of the word beginning 'system,'" and some allow for the<br />

wildcard <strong>to</strong> appear in other places than the end, thereby allowing for queries such as<br />

"show me all occurrences of words ending in 'ing.'" Some pieces of software allow for<br />

selection based around the occurrence close <strong>to</strong>gether of one or more wordforms. An<br />

example of selection by phrase would be "show me where the phrase 'the* moon' occurs"<br />

(where "*" means any word). A collocation selection query would allow one <strong>to</strong> ask<br />

"show me where the words 'moon' and 'gold' appear close <strong>to</strong>gether."<br />

After you have selected wordforms of interest the programs are ready <strong>to</strong> show results. All<br />

the <strong>to</strong>ols offer a KWIC display format – similar <strong>to</strong> that shown below – where we see<br />

occurrences of the word "sceptic" in David Hume's Dialogues Concerning Natural<br />

Religion.


Sceptic (11)<br />

[1,47] abstractions. In vain would the sceptic make a distinction<br />

[1,48] <strong>to</strong> science, even no speculative sceptic, pretends <strong>to</strong> entertain<br />

[1,49] and philosophy, that Atheist and Sceptic are almost synonymous.<br />

[1,49] by which the most determined sceptic must allow himself <strong>to</strong><br />

[2,60] of my faculties? You might cry out sceptic and railer, as much as<br />

[3,65] profession of every reasonable sceptic is only <strong>to</strong> reject<br />

[8,97] prepare a compleat triumph for the Sceptic; who tells them, that<br />

[11,121] <strong>to</strong> judge on such a subject. I am Sceptic enough <strong>to</strong> allow, that<br />

[12,130] absolutely insolvable. No Sceptic denies that we lie<br />

[12,130] merit that name, is, that the Sceptic, from habit, caprice,<br />

[12,139] To be a philosophical Sceptic is, in a man of<br />

A viewer uses the KWIC display <strong>to</strong> see what the context of the word says about how it is<br />

used in the particular text.<br />

The KWIC display shown above also indicates where in the text the word occurs: the first<br />

of the two numbers in square brackets is the number of the dialogue, and the second is a<br />

page number in a standard edition. This information comes from textual markup. Markup<br />

can often provide useful information about the text's structure, and some software allows<br />

the markup <strong>to</strong> be identified and used in several ways:<br />

(a) As shown in the KWIC display above, it can be used <strong>to</strong> indicate where a particular<br />

word occurrence occurs in the text – e.g., "paragraph 6 of chapter 2", or "spoken by<br />

Hamlet."<br />

(b) It can be used <strong>to</strong> select wordform occurrences that might be of interest – e.g., "tell me<br />

all the places where Hamlet speaks of a color."<br />

(c) It can be used <strong>to</strong> define a collocation range – e.g., "show me all the words that occur<br />

in the same sentence as a color-word."<br />

(d) It can be used as the basis for a distribution – "show me how a word is used by each<br />

speaker."<br />

(e) It can be used <strong>to</strong> delimit text <strong>to</strong> be ignored for concording – e.g., "do not include text<br />

found in the prologue in your index."<br />

Although working with markup is clearly useful, support for XML/SGML markup is<br />

poor in current wordform programs. None of the wordform-oriented software described


here can make good use of the full expressive power of TEI (the Text Encoding<br />

Initiative) XML markup, and some cannot even properly recognize the appearance of<br />

markup in a text document so that they can at least ignore it. A couple of the programs<br />

recognize a rather more simple form of markup – called COCOA markup after the piece<br />

of concordance software (COCOA, for count and concordance generation on the Atlas)<br />

for which it was first designed. A typical COCOA tag might look like this: "." The angle brackets announce the presence of a COCOA tag, and the tag's<br />

content is divided in<strong>to</strong> two parts. The first part (here, an "S") announces the kind of thing<br />

being identified – "S" presumably stands for "speaker." The second part announces that<br />

starting with the spot in the text where the tag has appeared, the Speaker is "Hamlet."<br />

Although perhaps at first glance the COCOA tagging scheme might remind one a little of<br />

XML or SGML markup, it is technically quite different, and is structurally largely<br />

incompatible with texts marked up in XML/SGML.<br />

Some wordform software offers other kinds of displays for results. A distribution graph,<br />

for example, could show graphically how the word use was distributed through the text,<br />

and might allow the user <strong>to</strong> see areas in the text where usage was concentrated. Some<br />

programs provide displays that focus on words that appear near <strong>to</strong> a selected word.<br />

TACT's collocation display, for example, is a list of all wordforms that occur near <strong>to</strong><br />

selected words, ordered so that those with proportionally the most of their occurrences<br />

near the selected words (and therefore perhaps associated with the selected word itself)<br />

appear near the <strong>to</strong>p of the list.<br />

Some of the programs allow for some degree of manual lemmatization of wordforms –<br />

the grouping <strong>to</strong>gether of different forms that constitute a single word in<strong>to</strong> a single group,<br />

although none of the wordform-oriented <strong>to</strong>ols include software <strong>to</strong> do this task<br />

au<strong>to</strong>matically. Stemming with wildcards can help <strong>to</strong> lemmatize many words in languages<br />

where the initial word stems can be used <strong>to</strong> select possible related wordforms. Several<br />

programs, including Concordance, WordSmith Tools, and TACT, allow one <strong>to</strong> s<strong>to</strong>re a list<br />

of wordforms that belong in a single lemma in a separate file and au<strong>to</strong>matically apply<br />

these against a text. A group of words for the verb "<strong>to</strong> be" in English would contain the<br />

various wordforms that make it up: "am, are, is, be, was, were", etc. By applying this list<br />

the software would group these various wordforms under a single heading: "<strong>to</strong> be." Of<br />

course, the task of doing a proper lemmatization is often more complex than this. For one<br />

thing, one must separate homographs (a single wordform that might have several<br />

different meanings, such as "lead" in English). To allow the machine <strong>to</strong> do this task<br />

au<strong>to</strong>matically would require it <strong>to</strong> understand the use of each word occurrence in its<br />

particular context -something outside the capabilities of any of the software reviewed<br />

here. TACT, however, allows the user <strong>to</strong> review the groups created by the simple<br />

wordform selection process and individually exclude wordform occurrences – thereby<br />

allowing the separation of homographs <strong>to</strong> be carried out manually, although at the cost of<br />

a significant amount of work.<br />

The following paragraphs outline some of the important characteristics of the pieces of<br />

wordform-oriented software that are available <strong>to</strong>day. (Only software readily available at<br />

the time of writing will be discussed – thus software such as OCP or WordCruncher,


important and influential as they were in their day, are not described.) There are many<br />

differences in detail that might make one <strong>to</strong>ol better suited <strong>to</strong> a task at hand than another.<br />

If you choose and use only one, you may need <strong>to</strong> direct your investigations <strong>to</strong> suit the<br />

program's strengths – a case of the <strong>to</strong>ol directing the work rather than the other way<br />

around.<br />

Concordance 3.0 is a Windows-based program written by R. J. S. Watt, who is Senior<br />

Lecturer in English at the University of Dundee. An evaluation copy (30 days) of the<br />

software is available at , with information about<br />

current prices. As the name suggests, the program focuses on providing a good KWIC<br />

Concordance generation and display engine. The user loads a text, selects words from the<br />

resultant word list (displayed on the left) and immediately sees a KWIC display on the<br />

right. The program supports some degree of lemmatization, and makes limited use of<br />

COCOA-like tags. I found its Windows interface quite easy <strong>to</strong> use. Concordance has<br />

been used with languages written in a range of alphabets. Furthermore, Professor Marjoie<br />

Chann, Ohio State University, reports success in using Concordance 3–0 even with<br />

Chinese, Japanese, and Korean texts on Windows 2000/XP. See <br />

for details.<br />

Concordance 3–0 has the ability <strong>to</strong> transform its concordance in<strong>to</strong> what becomes a large<br />

number of interconnected HTML pages that <strong>to</strong>gether could be made available on the<br />

WWW – a "Web Concordance." The resulting web pages present the full KWIC<br />

concordance, with links between the headwords in one frame, and the KWIC contexts in<br />

another. In the sample text that I tried (0.127 MB in size, containing about 23,300 words)<br />

the resulting concordance website was 7.86 MB, and contained 204 interlinked HTML<br />

documents.<br />

MonoConc and ParaConc is software developed by Michael Barlow at the Department of<br />

Linguistics, Rice University. See , and A<br />

Guide <strong>to</strong> MonoConc written by Barlow at .<br />

MonoConc is sold by the Athelstan company (Hous<strong>to</strong>n, Texas). With MonoConc you<br />

enter a search string <strong>to</strong> select words and it then generates a KWIC concordance. The<br />

search query language supports various kinds of wildcards, so that, for example, "speak*"<br />

will select all words beginning with "speak." There is also a query specification language<br />

(similar <strong>to</strong> regular expressions) that provides a more sophisticated pattern-matching<br />

capability. MonoConc also provides a way of sorting the KWIC display by preceding or<br />

following words – an option that causes similar phrases involving the selected word <strong>to</strong> be<br />

more evident. Like other programs described here, MonoConc can also generate some<br />

wordform frequency lists. ParaConc provides similar functions when one is working with<br />

parallel texts in more than one language.<br />

TACT is a suite of programs developed first by John Bradley and Lidio Presutti, and then<br />

somewhat further developed by Ian Lancashire and Mike Stairs, at the University of<br />

Toron<strong>to</strong>. The software was developed in the late 1980s and early 1990s and presents a<br />

DOS, rather than Windows, interface. It is still used, in spite of this limitation, by a<br />

significant user community. The software is available free from


– however, there is no manual for it<br />

available from that site. Instead, a user's guide – containing both the suite of programs<br />

and a selection of electronic texts on CD-ROM – can be purchased from the Modern<br />

Language Association (MLA).<br />

Although TACT's DOS-based interface is difficult <strong>to</strong> learn for users familiar with<br />

Windows, TACT remains popular with some users because it presents a relatively broad<br />

range of functions within the general wordform-oriented context described above,<br />

including the most sophisticated support for textual markup and word collocation of any<br />

of the programs described here. TACT works immediately with texts from the standard<br />

western European languages, and can be extended <strong>to</strong> work with other languages based on<br />

the Roman alphabet, and with classical Greek, although how <strong>to</strong> do this is unfortunately<br />

poorly documented. Compared <strong>to</strong> other programs described here, TACT is limited in the<br />

size of texts it can handle. Although users have reported working successfully with texts<br />

as large as the Bible, that is roughly TACT's upper limit. New users will find the system<br />

difficult at the beginning because of the DOS interface, the need <strong>to</strong> purchase the User's<br />

Guide <strong>to</strong> receive any written help, and the lack of technical support. TACTweb<br />

(http://tactweb.<strong>humanities</strong>.mcmaster.ca/) allows TACT's textbases <strong>to</strong> be searched by<br />

queries over the WWW.<br />

WordSmith Tools Version 3–0 is a suite of programs developed by Mike Scott<br />

(Department of English Language and Literature, University of Liverpool) and now<br />

distributed by the Oxford University Press. A version may be downloaded for evaluation<br />

from or the author's website at<br />

. Word-Smith Tools consists of six programs, one of<br />

which is a Controller component – used <strong>to</strong> invoke the others – and several others provide<br />

facilities for the basic text editing of large text files, and will not be described further<br />

here. Most users will begin by pointing the WordList component at one or more text<br />

documents <strong>to</strong> generate a word list. From there the user can select some words <strong>to</strong> form the<br />

basis for KWIC concordance, displayed in the Concord component. An unusual feature<br />

of WordSmith Tools is its KeyWords component. This allows the word list generated by<br />

your text <strong>to</strong> be compared with a control word list (for example, a word list generated<br />

from the British National Corpus is available over the WWW for this purpose), and the<br />

words that occur proportionally more frequently in your text than in the control text are<br />

emphasized in the resulting display.<br />

Although WordSmith Tools has only limited facilities <strong>to</strong> handle text with markup, it<br />

provides a richer set of functions <strong>to</strong> support statistical analysis than the other programs<br />

mentioned here. WordSmith Tools are Windows-based, but the modular nature of the<br />

program, where the word list, concordance, and keyword functions are packaged in<br />

different, albeit linked, components, made the software slightly more difficult for me <strong>to</strong><br />

master than some of the other programs mentioned here.<br />

Readers of this chapter may have noted that all the software mentioned so far runs on the<br />

IBM PC and/or Windows. Users of the Apple Macin<strong>to</strong>sh have a smaller number of<br />

programs <strong>to</strong> choose from. The AnyText search engine, marketed by Linguist's Software, is


described as "a HyperCard -based Full Proximity Boolean Search Engine and Index<br />

Genera<strong>to</strong>r that allows you <strong>to</strong> create concordances and do FAST word searches on<br />

ordinary text files in English, Greek and Russian languages." The software was originally<br />

designed <strong>to</strong> work with Biblical text, and it can be purchased by itself, or with one or more<br />

versions of the Bible included (at a higher cost). More information is available at<br />

– but it should be noted that HyperCard<br />

itself is no longer supported in the current generation of Macin<strong>to</strong>sh operating systems<br />

(OS X). D. W Rand's Concorder Version 3–1 is available free of charge, although a<br />

printed manual is sold by Les Publications CRM of the Universite de Montreal, Canada.<br />

More information about Concorder can be found at<br />

.<br />

Several sets of <strong>to</strong>ols developed for linguistics-oriented research support more<br />

sophisticated word-oriented research. See, for example, the GATE system developed at<br />

the University of Sheffield's Natural Language Processing Group at ,<br />

or TIPSTER, developed with the support of various agencies in the USA<br />

(http://www.itl.nist.gov/iaui/894.02/related_projects/tipster/). As powerful as these<br />

environments generally are, they have been developed in a technical culture that is<br />

largely foreign <strong>to</strong> the <strong>humanities</strong>, and they require both extensive knowledge of<br />

linguistics and computing environments such as Unix <strong>to</strong> operate. They are not discussed<br />

further in this chapter.<br />

Qualitative analysis<br />

Software developed for qualitative analysis within the social sciences provides a very<br />

different view of how computers can assist with the analysis of texts. Here the focus is on<br />

software that allows the user <strong>to</strong> explicitly label and organize thematic categories that they<br />

discover while reading the text. Three widely used pieces of software that work in this<br />

way are Nud*ist, NVivo (both from QSR International: ) and<br />

Atlas.ti from Scientific Software Development, Berlin<br />

(). An interesting review of text analysis software that<br />

comes from this social sciences perspective is provided in the book A Review of Software<br />

for Text Analysis (Alexa and Zuell 1999). An article describing the use of these <strong>to</strong>ols for<br />

research can be found in Kelle (1997).<br />

Text collation<br />

The use of computers <strong>to</strong> support the collation of textual variants is almost as old as the<br />

wordform work described earlier in this chapter. Although the early belief that collation<br />

was largely a mechanical task has been shown <strong>to</strong> be false, much useful work has been<br />

done in this area, nonetheless. Perhaps the best known <strong>to</strong>ol available <strong>to</strong> support collation<br />

is Peter Robinson's Collate program for the Macin<strong>to</strong>sh. You can read more about it at<br />

Getting Started with Collate 2, from<br />

.


Wilhelm Ott's TuStep system has components in it that are specifically designed <strong>to</strong><br />

support collation and the broader task of the preparation of a critical edition. See the<br />

further discussion about TuStep later in this chapter.<br />

A Text Tool for XML: XSLT<br />

Up <strong>to</strong> now we have focused primarily on <strong>to</strong>ols specifically developed <strong>to</strong> support certain<br />

kinds of research in the <strong>humanities</strong>. If your work is focused on these particular tasks, then<br />

you may find many of your needs met by these <strong>to</strong>ols. However, for many of us, the work<br />

we do with a text is not so tightly focused, and these <strong>to</strong>ols will rarely be sufficient <strong>to</strong><br />

support all the work we wish <strong>to</strong> do.<br />

In this section we will describe a set of <strong>to</strong>ols built on XSLT that are available <strong>to</strong> work<br />

with XML markup. The Text Encoding Initiative (TEI) has made it evident that many of<br />

the things in a text that are of interest <strong>to</strong> textual scholars can be usefully identified with<br />

SGML or XML markup. However, once these things are marked up this is rarely the end<br />

of the s<strong>to</strong>ry – after all, one puts the markup in so that one can do further work with it<br />

later. These days one often wants <strong>to</strong> publish such materials on the WWW and, of course,<br />

the resulting website would contain the base text in HTML. Often, however, the markup<br />

can provide the source for various supporting index documents that provide a set of<br />

alternative entry points in<strong>to</strong> the text itself. The au<strong>to</strong>matic generation of such indices and<br />

of the original text in HTML is well supported by XSLT.<br />

XSLT stands for "Extensible Stylesheet Language: Transformations" (compare with<br />

XML: "Extensible Markup Language"). It is one of a pair of standards that are named<br />

XSL (Extensible Stylesheet Language). The second part of XSL, XSLFO – "Extensible<br />

Stylesheet Language: Formatting Objects" – provides a language describing how a text<br />

should be laid out for display – usually for the printed page. The XSLT language is<br />

developed by the World Wide Web Consortium (W3C: ), a group<br />

of individuals and corporations who have taken on the task of maintaining and<br />

developing standards for the World Wide Web.<br />

The emphasis in XSLT is on transformation, and <strong>to</strong>day the principal use of XSLT is the<br />

transformation of an XML document in<strong>to</strong> HTML for presentation on the WWW. The<br />

virtues of using a form of XML more complex and specifically adapted than HTML as<br />

the initial markup scheme for a text have been extensively described elsewhere in this<br />

book, and I will not repeat them here. However, if one is <strong>to</strong> present material on the<br />

WWW then (at least <strong>to</strong>day) one needs <strong>to</strong> present the material in HTML, and an approach<br />

that consistently transforms any well-formed XML in<strong>to</strong> HTML is very valuable. Often a<br />

single tag in XML needs <strong>to</strong> be transformed in<strong>to</strong> a set, sometimes a complex set, of HTML<br />

tags. In XSLT one describes this transformation once, in what is called in XSLT a<br />

template. An XSLT processor can then apply the specification given in the template <strong>to</strong> all<br />

places in the original text where it fits. Normally the transformation process requires the<br />

specification of many different templates. Thus, one groups the set of templates one<br />

needs in a single document, called an XSLT stylesheet.


Here is a simple example that will give a little of the flavor of XSLT. Suppose the<br />

original text is marked up in TEI (the example, slightly modified <strong>to</strong> be in XML rather<br />

than SGML, is from section 6.3–2.2. "Emphatic Words and Phrases" of the TEI P3<br />

guidelines): For presentation in<br />

HTML, the q element needs <strong>to</strong> be transformed so that the contained text is surrounded by<br />

quote marks. The empb element, when the rend attribute is set <strong>to</strong> "italic", needs <strong>to</strong> be<br />

transformed in<strong>to</strong> HTML's / element. The following two templates, which would, of<br />

course, be included with many others in a complete XSLT stylesheet, would accomplish<br />

this: XSLT is itself written in XML. In<br />

this excerpt we see two XSLT template elements. Each has a match attribute which<br />

specifies where the particular transformation contained therein should be applied. The<br />

match attribute for the first template element simply asserts that this template is<br />

applicable for all q elements in the input text. The match attribute in the second template<br />

asserts that it is applicable <strong>to</strong> all emph elements which have their rend attribute set <strong>to</strong><br />

italic. The content of each template element describes what should be generated as output<br />

from the transformation process. In both template elements you can see the applytemplates<br />

element which, in effect, says "place the matched element's content here while<br />

also applying their templates as appropriate." In the first template, then, the content is<br />

simply surrounded by quote marks. In the second template, the content is surrounded by<br />

HTML's / element. If these transformation templates were applied <strong>to</strong> the text above, then,<br />

the resulting output would be: Although this<br />

example gives a little bit of the flavor of what XSLT can do, by itself it barely hints at the<br />

kinds of transformations that can be expressed in XSLT. With an XSLT stylesheet, for<br />

example, it is possible <strong>to</strong> specify a transformation that reorganizes and duplicates the<br />

content of some elements so that, say, each section heading in an input document appears<br />

in two places in the output – once as heading for its respective section, but also at the<br />

front of the document as an element in a table of contents. Material can also be sorted and<br />

grouped in different ways (although grouping is rather complex <strong>to</strong> express in the XSLT<br />

version 1). It is possible, for example, <strong>to</strong> write a transformation stylesheet that scans an<br />

XML document looking for references <strong>to</strong> people that have been tagged using TEI's<br />

"name" tag: (e.g., from TEI P3 guidelines section 6.4.1, slightly modified):<br />

XSLT can be used <strong>to</strong> select all material in name tags and generate an index from them.<br />

All the references <strong>to</strong> "de la Mare, Walter" and <strong>to</strong> the places "Charl<strong>to</strong>n" and "Kent" could<br />

be grouped <strong>to</strong>gether under their respective headwords, and links could be generated


elow each heading <strong>to</strong> take the user from the index <strong>to</strong> the spots in the text where the<br />

references were found.<br />

Examples so far have emphasized the use of XSLT <strong>to</strong> generate output in HTML for<br />

making material available on the WWW. However, XSLT is designed <strong>to</strong> produce output<br />

in other XML formats as well as HTML. As XML becomes the scheme of choice for<br />

feeding data <strong>to</strong> many programs, it is likely that XSLT will be used more and more <strong>to</strong><br />

transform one type of XML markup in<strong>to</strong> another. For example, suppose you had a text<br />

with all the words tagged, and you wished <strong>to</strong> perform a statistical calculation based, say,<br />

on the number of occurrences of the wordforms in different chapters, one could write a<br />

(complex) stylesheet that would locate the words, count up the number of occurrences of<br />

each wordform in each chapter, and generate a document containing this information. If<br />

the statistical analysis program accepted data of this kind in XML then your stylesheet<br />

could write out its results in the XML format your statistical program wanted.<br />

Learning XSLT is difficult, particularly if you wish <strong>to</strong> take advantage of its more<br />

advanced features. Jeni Tennison's Book Beginning XSLT (Tennison 2002) provides a<br />

thorough introduction <strong>to</strong> many features of XSLT and is understandable <strong>to</strong> the nonprogrammer,<br />

although one does have <strong>to</strong> put up with examples drawn from the processing<br />

of television program listings! Michael Kay's XSLT: Programmer's Reference (Kay <strong>2001</strong>)<br />

is, as the name suggests, a complete, in-depth review of XSLT and the software that<br />

supports it, and is aimed at a rather technical audience. A search on Google for "XSLT<br />

Tu<strong>to</strong>rials" will find many online tu<strong>to</strong>rials. A number of learning resources are also listed<br />

at .<br />

There are several XSLT transformation <strong>to</strong>ols available, and all the ones described here<br />

are free.<br />

Microsoft's MSXML (XML Core Services) contains an XSLT processor for Windows and<br />

is incorporated inside the company's Internet Explorer (IE) browser. If you send IE an<br />

XML document with an XSLT stylesheet, it will transform the XML within the browser<br />

and present the display as if the resulting HTML document had been sent instead.<br />

Microsoft was much involved in the development of XSLT, and until Internet Explorer 6,<br />

the processor provided was not compatible with the now-standard version of XSLT.<br />

Hence, be sure that you are using the latest version. Information about MSXML, and a<br />

download package <strong>to</strong> set it up on your Windows machine can be found at<br />

.<br />

There are several XSLT processors available that run separately from a browser. To use<br />

one of them you would give your XML document and XSLT stylesheet <strong>to</strong> it, and it would<br />

save results of the transformation in<strong>to</strong> a file on your hard disk. HTML files generated this<br />

way would then be made available using any web server, and by transforming the<br />

documents in HTML before serving them, you allow any web browser <strong>to</strong> view them. The<br />

two most widely used XSLT processors are SAXON and XALAN. Both these XSLT<br />

processors are most readily run on a Windows machine from the DOS prompt. You will


need <strong>to</strong> be familiar with the basics of a DOS window in order <strong>to</strong> be able <strong>to</strong> use these<br />

processors effectively.<br />

Michael Kay's SAXON processor () is considered <strong>to</strong> be<br />

very highly conformant with the XSLT standard. Saxon versions up <strong>to</strong> 6.5.2 have been<br />

packaged in an "Instant Saxon" format that runs readily on many Windows machines, and<br />

is relatively easy <strong>to</strong> set up and use. It is packaged in a ZIP file on the Saxon website, and<br />

setting it up involves only unzipping the two files it contains in<strong>to</strong> an appropriate direc<strong>to</strong>ry<br />

on your computer. Once I had unzipped the Saxon program from the ZIP file I was<br />

immediately able <strong>to</strong> use it with a Hamlet text and stylesheet by starting a DOS/<br />

Command window and typing the command:<br />

(which asked SAXON <strong>to</strong> read an XML<br />

document in hamlet, xml, and an XSLT stylesheet in hamlet.xsl and <strong>to</strong> generate as output<br />

a file hamlet.htm). The ZIP file contains an HTML document that describes how <strong>to</strong><br />

invoke the Saxon program in some detail. A version of SAXON is also available which<br />

operates with the support of Sun's Java system. If you are familiar with Java, you might<br />

find it as easy <strong>to</strong> use this version instead.<br />

XALAN, from the Apache Software Foundation, is another widely used XSLT processor.<br />

It comes in two versions. One version, implemented in Java, can be found at<br />

, and requires you <strong>to</strong> also set up Sun's Java<br />

system on your computer. The second version (at )<br />

does not require Java, but seems <strong>to</strong> be a less complete implementation of<br />

XSLT.<br />

Macin<strong>to</strong>sh OS X users who are able <strong>to</strong> set up Java on their machines in the background<br />

Unix environment that underpins OS X will be able <strong>to</strong> use XALAN or SAXON on their<br />

machines.<br />

Perl and TuStep: General Purpose Text Manipulation<br />

In the first section of this chapter we described <strong>to</strong>ols designed <strong>to</strong> support particular kinds<br />

of text-oriented tasks. The section on XML introduced <strong>to</strong>ols that are potentially much<br />

more general-purpose. In this section we come <strong>to</strong> the <strong>to</strong>ols that could be applied the most<br />

broadly, and we will be describing two "<strong>to</strong>olkit environments" that at first may seem<br />

quite different: Perl and TuStep. By the end of the chapter we will have shown some of<br />

the ways in which they are also similar.<br />

Perl<br />

Perl was created by Larry Wall in 1987 <strong>to</strong> respond <strong>to</strong> a need <strong>to</strong> be able <strong>to</strong> perform<br />

complex administrative-oriented text processing and report generation of data on Unix<br />

systems. Perl was so successful that it soon became the <strong>to</strong>ol of choice for getting all sorts<br />

of things done in the Unix environment, and it eventually spread from the Unix world <strong>to</strong><br />

many other computing environments. It has become (in Andrew Johnson's words) "the<br />

<strong>to</strong>ol <strong>to</strong> use for filling gaps everywhere." Perl is good at reading, processing, transforming


and writing plain text. If you have ASCII data that come out of one program in one<br />

format, and you want <strong>to</strong> feed them in<strong>to</strong> another program that requires a slightly different<br />

format, then Perl might allow you <strong>to</strong> quickly develop the transformation you need in a<br />

matter of minutes. We will introduce Perl here, but other languages such as Python can<br />

do similar work.<br />

Perl is a programming language, and as such shares some of its characteristics with other<br />

programming languages – including a syntactical structure that is broadly similar <strong>to</strong><br />

several important ones. Thus, those who are familiar with other widely used languages<br />

such as Java or C will start off already knowing something of how <strong>to</strong> write Perl<br />

programs. This means that for many computing professionals, Perl is "easy <strong>to</strong> learn."<br />

However, if you do not have a programming background, you will find that you will not<br />

be able <strong>to</strong> draw in the years of training and experience that allow Perl <strong>to</strong> be "easy <strong>to</strong><br />

learn" for this community. Instead, initially you are likely <strong>to</strong> find learning <strong>to</strong> program in<br />

Perl slow-going. Make friends with a programmer!<br />

Filling of gaps – Perl's strength – often involves the creation of small pieces of<br />

throwaway software specifically tailored for a task at hand. When one knows how <strong>to</strong><br />

write Perl, one can often throw a little program <strong>to</strong>gether <strong>to</strong> do something simple in a<br />

matter of minutes. The price for this informal nature of the language is sometimes paid in<br />

performance – the program might take longer <strong>to</strong> run than a carefully written program <strong>to</strong><br />

do the same task in C would take – however, the C program <strong>to</strong> do the same job might<br />

take a day <strong>to</strong> write.<br />

To help introduce Perl, I will describe a bit of the Perl programming I did for Willard<br />

McCarty's Onomasticm of Ovid's Metamorphoses. In this project, McCarty has marked<br />

up some aspects of personification in the Metamorphoses by a tagging scheme of his own<br />

invention. Here is an example from his text:<br />

These seven lines represent<br />

one line from book 13 of the Metamorphoses. The first line, above, is the actual Latin<br />

text of Ovid's poem. The remaining six lines represent the tagging that McCarty has<br />

added. Most of the lines of poetry, like this one, have a number of tags attached <strong>to</strong> them.<br />

Each tag is surrounded by brace brackets "[" and "]", and the contents are divided in<strong>to</strong><br />

three parts. The first part (in the first tag "on") categorizes the kind of personification –<br />

here the individual is personified by being given a person-like attribute. Tags beginning<br />

with "v" represent personification through a verb. The second part (here "Achilles")<br />

identifies the person/god being talked about, and the third part (here "[arma] optima")<br />

links the tag <strong>to</strong> words in the text. Tags can be more complex, but this is sufficient for our<br />

purposes here.


One obvious way <strong>to</strong> transform this text and make the tagging very useful would be <strong>to</strong> turn<br />

it inside out – producing indices based around the tags, producing something that looks<br />

like this (here showing a fragment from the index of Attributes):<br />

You can see the reference <strong>to</strong> Achilles arma optima from line 13–40 in the second line of<br />

the Arma/Achilles entry. The indices were represented as a set of linked HTML pages,<br />

with references <strong>to</strong> the poem acting as links. One can use the indices <strong>to</strong> first get an<br />

overview of the kind of attributes that appear across the entire text, and can then use the<br />

index <strong>to</strong> review all the occurrences in the poem.<br />

Writing the Perl code <strong>to</strong> transform this tagged text in<strong>to</strong> these indices was, of course,<br />

significantly more than a 10-minute job, and it is beyond the space we have here <strong>to</strong> begin<br />

<strong>to</strong> analyze the processing in any detail. However, a small example of Perl coding will<br />

give you the flavor of what working with Perl is like.<br />

An important component of Perl is its regular expression processor. Regular expressions<br />

provide a way <strong>to</strong> express relatively complex patterns <strong>to</strong> select and pull apart text for<br />

further processing. The expression: uses the regular expression opera<strong>to</strong>r<br />

sequence ".*?" and ".*" <strong>to</strong> specify "any sequence of characters." This pattern matches the<br />

form of all the tags in McCarty's text: a "{" followed by any sequence of characters,<br />

followed by a "/", followed by any sequence of characters, followed by a ":", followed by<br />

any sequence of characters and ending with a "}".<br />

If we put parentheses around the "any sequence of characters" in this expression, then<br />

Perl will make available for further processing the component of each specific text string<br />

that matched the pattern. Thus, the sequence: if applied against the<br />

sequence "[on/ Achilles: [arma] optima]" would have Perl identify "on" as the first part,<br />

"Achilles" as the second part, and "[arma] optima" as the third part, and make those<br />

subcomponents available <strong>to</strong> the rest of the program for further processing. Here is the<br />

same regular expression as it might be used in a bit of Perl code:<br />

The example has been slightly simplified from


what was actually used. Even then one can see the formal nature of the Perl language, and<br />

perhaps get a sense of the challenge it presents for non-programmers <strong>to</strong> learn. The first<br />

line says "fetch a new line of input from the input file and do the following chunk of Perl<br />

code (enclosed in '[]' s) for each line." The second line says "assign the line of input you<br />

read in the first line <strong>to</strong> $txt, and chop off any ending newline character that you find."<br />

The third line says "if the string of characters in $txt matches the given pattern (you can<br />

see the pattern we described above embedded in this line), do the lines enclosed below<br />

within '[]' s." The fourth line asks Perl <strong>to</strong> save the first bit of the matched regular<br />

expression in $code, the second bit in $name, and the third bit in $text. The code would<br />

continue with instructions that made Perl do something useful with the chunks of the tag<br />

in $code, $name, and $text.<br />

During my analysis of the task of transforming the tagged text in<strong>to</strong> a set of linked indices,<br />

it became evident that the process <strong>to</strong> generate the separate indices was largely the same<br />

for each index, albeit using different data by selecting different tags from the tagged text.<br />

Thus, it became clear that the entire processing could be written as a small set of separate<br />

modular programs that could be directed <strong>to</strong> each appropriate tagset in turn. We chose,<br />

then, <strong>to</strong> apply Perl <strong>to</strong> the task of transforming the marked-up text in<strong>to</strong> indices (expressed<br />

as collections of HTML pages) by applying a modular approach – breaking the<br />

transformation tasks in<strong>to</strong> a number of smaller steps, writing Perl code for each one of the<br />

bits. Then, by combining them <strong>to</strong>gether in the correct order, the entire task would be<br />

performed.<br />

For example, a small program we called text2htm.pl transformed the text and tags in<strong>to</strong> a<br />

set of HTML documents. A small module was also written (called tagproc.pl) that<br />

extracted data from the tags and s<strong>to</strong>red them in a large table-like structure which would<br />

be used, and added <strong>to</strong>, by the other processing Perl programs that worked on the material.<br />

These two programs only needed <strong>to</strong> be run once. A set of three Perl scripts then did the<br />

bulk of the work of transforming the tabular data derived from the tags in<strong>to</strong> a particular<br />

index presentation (divided in<strong>to</strong> a number of separate HTML pages) you see above. Two<br />

further scripts were also used during the generation of each index <strong>to</strong> create two kinds of<br />

navigational aids for the index. Thus, after the first two programs were run, the other five<br />

could be used in sequence for each of the three indices that were <strong>to</strong> be generated.<br />

When the Perl scripts were ready, it was a simple matter <strong>to</strong> record the sequence in which<br />

they should run and what data they should process in a small DOS batch file. When this<br />

has been done the process of transforming the entire text with its thousands of tags in<strong>to</strong><br />

the integrated collection of several thousand HTML pages could then be performed by<br />

typing only one command. When the batch file was invoked, it was then a matter of<br />

simply waiting for the couple of minutes while the computer worked its way through the<br />

batch file and Perl scripts <strong>to</strong> generate the completed version.<br />

The writing of the Perl scripts <strong>to</strong> work over the material was not a trivial matter.<br />

However, because they could be packaged up and invoked in such a simple manner, they<br />

allowed McCarty <strong>to</strong> quickly generate a new version of the reference website for his<br />

materials as often as he liked.


Getting started with Perl<br />

Perl is free software. It was originally developed for Unix systems and is nowadays<br />

provided as a built-in <strong>to</strong>ol in systems like Linux. The ActiveState company<br />

(http://www.activestate.com) has produced the most popular (free) version of Perl for<br />

Windows machines. Versions of Perl have been available on the Macin<strong>to</strong>sh for several<br />

years (see, for example, MacPerl), although until OS X (which is Unix-based), they were<br />

generally incomplete implementations.<br />

Development of Perl is, these days, a worldwide phenomenon. There is not only<br />

continued work <strong>to</strong> enhance and further develop the basic Perl language, but there is a<br />

substantial number of developers who develop Perl modules that allow the Perl<br />

programmer <strong>to</strong> do tasks other than simply process text files. The DBI module, for<br />

example, provides a standard way <strong>to</strong> allow a Perl script <strong>to</strong> connect <strong>to</strong> and extract data<br />

from a relational database. There is a set of XML modules that allow XML files <strong>to</strong> be<br />

manipulated in Perl programs. Indeed, among the thousands of modules now available it<br />

is possible <strong>to</strong> find one that will help you apply Perl <strong>to</strong> almost any kind of computer data<br />

now available, including material available over the Internet. The CPAN website<br />

() is the usual place <strong>to</strong> go for information about all the modules<br />

that people have written. For Windows Perl users, the ActiveState company maintains a<br />

set of these modules specially packaged up for use on Windows, and allows you <strong>to</strong><br />

download any of them and set them up for free, simply by invoking their "ppm" program<br />

and asking for them by name.<br />

Perl, like any full-function programming language, is difficult for non-programmers <strong>to</strong><br />

learn, and most books and online tu<strong>to</strong>rials that purport <strong>to</strong> provide an "easy" introduction<br />

<strong>to</strong> Perl are, in fact, written for the programmer audience. Andrew L. Johnson's book<br />

Elements of Programming with Perl (Johnson 2000), however, provides an introduction<br />

<strong>to</strong> Perl programming for non-programmers, and is recommended for this purpose.<br />

TuStep<br />

Willhelm Ott's TuStep (Tubingen System of Text Processing Programs) system has a<br />

long his<strong>to</strong>ry. The software that developed in<strong>to</strong> TuStep was begun in 1966, and the first<br />

version of TuStep that was made available under that name appeared in 1978. It is still in<br />

use <strong>to</strong>day – extensively at Tubingen and in other centers in Germany, and in a number of<br />

other places in the world – for the manipulation of texts for all sorts of scholarly<br />

purposes. TuStep consists of a set of 45 commands that operate in an environment<br />

containing a text edi<strong>to</strong>r and a macro program. Ott describes the task domain for his<br />

system as "textdata processing", and describes what he means by this in the introduction<br />

<strong>to</strong> the English version of the TuStep manual:<br />

These basic operations of textdata processing (and the corresponding TUSTEP programs)<br />

encompass functions that can be generally characterized as follows: Collation of different<br />

text versions; correcting not only in the interactive Edi<strong>to</strong>r mode, but also with the use of<br />

prepared (or au<strong>to</strong>matically generated) correction instructions; breakdown of text in<strong>to</strong>


user-defined elements (z. B. semantic patterns); sorting text elements or lengthy units of<br />

text according <strong>to</strong> a number of different alphabets and other criteria; Generating an index<br />

by compiling sorted text-elements; Processing textdata according <strong>to</strong> user-defined<br />

selective criteria, replacing, moving, enhancing, concluding and comparing text,<br />

calculating numerical values contained in the text (e.g. calender dates) or those which can<br />

be determined from the text (e.g. the number of words in a sentence), and output in<br />

various formats, including those required by other operating systems (e.g. SPSS for<br />

statistical evaluation).<br />

(1993: 8)<br />

TuStep, then, consists of a large number of rather abstract processes that can be instructed<br />

<strong>to</strong> perform a large number of individual operations on a set of text files. The processes<br />

are designed <strong>to</strong> be combined <strong>to</strong>gether in a sequence <strong>to</strong> perform a large number of useful<br />

tasks. The sequence of operations can then be s<strong>to</strong>red in a command file and subsequently<br />

reapplied at any time. One of TuStep's components is a high-quality typesetting program<br />

which is controlled by a command language that can be generated by other elements of<br />

the system, and which is capable of producing letterpress-quality printing – indeed a<br />

significant use of TuStep is in the preparation of printed critical editions. In the same way<br />

that a set of woodworking <strong>to</strong>ols can, in the hands of a skilled craftsperson, produce<br />

beautiful results, TuStep's <strong>to</strong>ols can be applied <strong>to</strong> text <strong>to</strong> produce beautiful printed<br />

editions. Of course, it is only after years of practice that the woodworker can produce a<br />

beautiful bookcase using the basic carpentry <strong>to</strong>ols such as a saw, plane, chisel, etc. In the<br />

same way, one has <strong>to</strong> learn how <strong>to</strong> take the similarly abstract <strong>to</strong>ols provided by TuStep<br />

and combine them <strong>to</strong> generate the best product that they are capable of producing.<br />

We mentioned earlier that TuStep modules are meant <strong>to</strong> be combined in<strong>to</strong> a sequence of<br />

operations. To generate an index of McCarty's Onomasticm material described earlier<br />

using TuStep, one would first invoke its PINDEX program <strong>to</strong> extract the materials that<br />

are meant <strong>to</strong> make up the index – putting the result in<strong>to</strong> an intermediate file. Then one<br />

would sort the resulting intermediate file and use the result of that <strong>to</strong> feed in<strong>to</strong> another<br />

program called GINDEX <strong>to</strong> transform the properly ordered file in<strong>to</strong> a hierarchical indexlike<br />

structure. Next one could take the hierarchically structured entries and write<br />

instructions that would transform them in<strong>to</strong> a series of typesetting codes that could drive<br />

the typesetting program.<br />

TuStep modules often incorporate a large number of separate functions within a single<br />

program. Parameters are provided <strong>to</strong> tell the PINDEX program how <strong>to</strong> recognize material<br />

in the input text file that belongs in the index, and how <strong>to</strong> assign the pieces that are found<br />

in<strong>to</strong> a multilevel index. Instructions <strong>to</strong> TuStep programs are given in a notational scheme<br />

that is especially developed for the program. For example, the following lines would be<br />

s<strong>to</strong>red in a file and passed <strong>to</strong> the PINDEX program <strong>to</strong> tell it how <strong>to</strong> extract the tags from


McCarty's Onomasticon texts: The letters at the front of each line specify<br />

parameters that the PINDEX program is <strong>to</strong> use. The "ea" line specifies that an index entry<br />

from the text file begins after a ":." The "ee" line indiciates that the text for an index entry<br />

ends before either a "]" or a "_." The "ena" and "ene" lines identify characters that delimit<br />

supplementary text that belong with the collected index entry. The "enk" line specifies<br />

that the text after the index entry should be in bold.<br />

TuStep is a very versatile system, but learning how <strong>to</strong> use the software components <strong>to</strong> do<br />

the things you want is not easy. TuStep can be inexpensively purchased from the<br />

University of Tubingen. Most of the documentation for TuStep is in German. There is a<br />

version of the TuStep reference manual in English, but it cannot really be used <strong>to</strong> learn<br />

the program in the first place. There is a tu<strong>to</strong>rial on TuStep, in German, called Lernbucb<br />

TUSTEP, and there are training courses each March and September in Tubingen that will<br />

help you get started. There is a brief introduction <strong>to</strong> TuStep, its concepts, and its facilities<br />

in Ott (2000). There is also further information on the TuStep website at<br />

.<br />

TuStep and Perl compared<br />

Both TuStep and Perl can be used <strong>to</strong> do almost any kind of manipulation of a text in<br />

electronic form. Perl is a general-purpose programming language and can also perform<br />

any kind of transformation that you can ask a computer <strong>to</strong> do on a text file, although the<br />

task of learning how <strong>to</strong> use it is large – especially for non-programmers. TuStep is<br />

especially developed for performing tasks of interest <strong>to</strong> scholars working with texts,<br />

although it will still take real work <strong>to</strong> learn how <strong>to</strong> use it. TuStep also comes with a highquality<br />

typesetting program which is capable of professional-quality typesetting –<br />

something Perl does not have (although, of course, Perl could be used <strong>to</strong> generate texts <strong>to</strong><br />

feed in<strong>to</strong> other typesetting systems such as TeX). For non-programmers, then, both<br />

TuStep and Perl have steep and long learning curves. Since TuStep is specifically<br />

designed for scholarly work, it does contain some helpful features – such as strategies<br />

that are built in for the proper sorting of text in many languages – that make specifying<br />

processing easier. However, on the other hand, there is a great deal of material available<br />

about how <strong>to</strong> program in Perl, and it is possible <strong>to</strong> hire a programmer who already knows<br />

Perl <strong>to</strong> do work for you.<br />

Conclusion<br />

The reader will, by this point, have hopefully seen that the range of text <strong>to</strong>ols described<br />

here can support a very wide range of scholarly activities on texts, and in the limited<br />

space available here it is possible <strong>to</strong> only hint at the full potential of these <strong>to</strong>ols. Susan<br />

Hockey's book Electronic Texts in the Humanities (Hockey 2000), while not explicitly


about text <strong>to</strong>ols, does describe many different projects and suggests the range of things<br />

that a set of <strong>to</strong>ols like the ones described here can accomplish.<br />

Certainly, XSLT, Perl, and TuStep are complex and powerful systems. Certain kinds of<br />

useful results can be achieved with a modest amount of work, but learning <strong>to</strong> use any of<br />

them both efficiently and <strong>to</strong> their full potential will commit the learner <strong>to</strong> a process that<br />

could take years <strong>to</strong> complete. Often reviews of software for the <strong>humanities</strong> talk about<br />

"user friendliness." Are these <strong>to</strong>ols "user-friendly?" If you mean, "are they immediately<br />

usable", then the answer is "no." However, are a good set of carpentry <strong>to</strong>ols userfriendly?<br />

They are if you are skilled in their use. In a similar way, these <strong>to</strong>ols can do<br />

significant work on texts, but all of them require a significant amount of skill <strong>to</strong> make use<br />

of them.<br />

Bibliography<br />

ActivePerl (2002). ActiveState Corp. At<br />

http://www.activestate.com/Products/ActivePerl/.<br />

Alexa, Melina and Cornelia Zuell (1999). A Review of Software for Text Analysis.<br />

Mannheim: ZUMA.<br />

Barlow, Michael (1995). A Guide <strong>to</strong> MonoConc. At<br />

http://www.ruf.rice.edu/~barlow/mc.html.<br />

Clark, James, (ed.) (1999). XSL Transformations (XSLT): Version 1.0. W3C. Accessed<br />

November 16, 1999. At http://www.w3c.org/TR/xslt.<br />

Concordance. Accessed November 10, 2002. At http://www.rjcw.freeserve.co.uk/.<br />

Hawthorne, Mark (1994). The Computer in Literary Analysis: Using TACT with Students.<br />

Computers and the Humanities 28: 19–27.<br />

Hockey, Susan (2000). Electronic Texts in the Humanities. Oxford: Oxford University<br />

Press.<br />

Johnson, Andrew L. (2000). Elements of Programming with Perl. Greenwich, CT:<br />

Manning Publications.<br />

Kay, Michael (<strong>2001</strong>). XSLT: Programmer's Reference. Birmingham, UK: WROX Press.<br />

Kay, Michael (2002). SAXON: The XSLT Processor. Accessed August 28, 2002. At<br />

http://saxon.sourceforge.net/.<br />

Kelle, U. (1997). Theory-building in Qualitative Research and Computer Programs for<br />

the Management of Textual Data. Sociological Research Online 2, 2. At<br />

http://www.socresonline.org.Uk/socresonline/2/2/l.html.


Lancashire, Ian, et al. (1996). Using TACT with Electronic Texts. New York: Modern<br />

Languages Association of America.<br />

Microsoft MSDN: XML. Microsoft. At http://msdn.microsoft.com/xml.<br />

Ott, W. (2000). Strategies and Tools for Textual Scholarship: the Tübingen System of<br />

Text Processing Programs (TUSTEP). Literary and Linguistic Computing 15, 1: 93–108.<br />

Robinson, Peter (2000). Collate Software: Resources. Accessed March 2, 2000. At<br />

http://www.cta.dmu.ac.uk/projects/collate/res.html.<br />

Scott, M. (2002). Mike Scott's Web: WordSmith Tools. Accessed July 16, 2002. At<br />

http://www.liv.ac.uk/~ms2928/.<br />

Smith, J. B. (1989). Computer Criticism. In Rosanne G. Potter (ed.), Literary Computing<br />

and Literary Criticism: Theoretical and Practical Essays on Theme and Rhe<strong>to</strong>ric (pp.<br />

13–44). Philadelphia: University of Pennsylvania Press.<br />

Tennison, Jeni (2002). Beginning XSLT. Birmingham, UK: WROX Press.<br />

TUebingen System of TExt processing Programs: TuStep. Accessed Oc<strong>to</strong>ber 22, 2002.<br />

At http://www.uni-tuebingen.de/zdv/tustep/tustep_eng.html.<br />

Wall, Larry, et al. (1996). Programming Perl. Sebas<strong>to</strong>pol, CA: O'Reilly and Associates.<br />

Wooldridge, T. Russon, (ed.) (1991). A TACT Exemplar. Toron<strong>to</strong>: CCH, University of<br />

Toron<strong>to</strong>.<br />

34.<br />

"So the Colors Cover the Wires": Interface, Aesthetics,<br />

and Usability<br />

Matthew G. Kirschenbaum<br />

Introduction: A Glass Darkly<br />

The idea of interface, and related concepts such as design and usability, are some of the<br />

most vexed in contemporary computing. Definitions of interface typically invoke the<br />

image of a "surface" or a "boundary" where two or more "systems", "devices", or<br />

"entities" come in<strong>to</strong> "contact" or "interact." Though these terms encourage spatial<br />

interpretation, most interfaces also embody temporal, haptic, and cognitive elements. The<br />

steering wheel of a car, the control panel of a VCR, and the handle of a door are all<br />

examples of everyday interfaces that exhibit these dimensions. In the context of


computers and computing, the word "interface" is often used interchangeably with<br />

"graphical user interface", or GUI, most frequently encountered as a desk<strong>to</strong>p windows<br />

environment. The command line prompt is perhaps the best-known alternative <strong>to</strong> the<br />

GUI, but there are a plethora of others, including screen readers, motion trackers, tangible<br />

user interfaces (TUIs, breathtakingly rendered in the 2002 film Minority Report), and<br />

immersive or augmented computing environments. In the <strong>humanities</strong>, meanwhile, it is<br />

increasingly common <strong>to</strong> encounter the idea that a book or a page is a kind of interface, a<br />

response <strong>to</strong> the understanding that the conventions of manuscript and print culture are no<br />

less technologically determined than those of the <strong>digital</strong> world. At least one observer,<br />

Steven Johnson, has defined our present his<strong>to</strong>rical moment as an "interface culture", a<br />

term he wields <strong>to</strong> embrace not only the ubiquity of computers and electronic devices but<br />

also the way in which interface has come <strong>to</strong> function as a kind of trope or cultural<br />

organizing principle – what the industrial novel was <strong>to</strong> the nineteenth century or<br />

television <strong>to</strong> the suburban American 1950s are his examples.<br />

As much as it is talked about, however, interface can at times seem little loved. Usability<br />

guru Donald A. Norman writes: "The real problem with interface is that it is an interface.<br />

Interfaces get in the way. I don't want <strong>to</strong> focus my energies on interface. I want <strong>to</strong> focus<br />

on the job" (2002: 210). Nicholas Negroponte holds that the "secret" of interface design<br />

is <strong>to</strong> "make it go away" (1995: 93). To further complicate matters, interface is often, in<br />

practice, a highly recursive phenomenon. Take the experience of a user sitting at her<br />

computer and browsing the Web, perhaps accessing content at a <strong>digital</strong> <strong>humanities</strong> site.<br />

The site's internal design imposes one layer of interface between the user and the content,<br />

and the web browser – its but<strong>to</strong>ns and menus and frames – immediately imposes another.<br />

The user's desk<strong>to</strong>p environment and operating system then impose a third layer of<br />

interface. The ergonomics of the situation (we'll assume our user is working with a<br />

keyboard and mouse, looking at a screen positioned the recommended 18 inches away)<br />

create still another layer of interface, a layer which becomes apparent when one considers<br />

alternatives such as accessing the same content with a PDA or a wearable device, or in a<br />

room-based virtual reality setting like a CAVE. Importantly, each of these "layers", as I<br />

have been calling them, exhibits the potential for interaction with one another as well as<br />

with the user. The desk<strong>to</strong>p environment governs the behavior of the browser software,<br />

whose features and functions in turn directly affect many aspects of the user's interaction<br />

with the site's internal design and content.<br />

While everything I have just been rehearsing is familiar enough in computer science<br />

circles, particularly the domain known as human-computer interaction (HCI, also<br />

sometimes identified as human-computer interface), aspects of this narrative may seem<br />

problematic <strong>to</strong> readers trained in <strong>humanities</strong> disciplines. It would not be hard <strong>to</strong> find<br />

someone willing <strong>to</strong> argue that my entire scenario is the product of yet another<br />

unacknowledged interface, a kind of common cultural gateway whose socially<br />

constructed ideologies govern our expectations with regard <strong>to</strong> technology, representation,<br />

and access <strong>to</strong> information. Moreover, in the scenario sketched above, my distinction<br />

between different layers of interface and something I casually called "content" is one that<br />

runs counter <strong>to</strong> decades of work in literary and cultural criticism, where form and content<br />

are almost instinctively unders<strong>to</strong>od as inextricable from one another. Thus, the weight of


established wisdom in a field like interface design rests on a fundamental disconnect with<br />

the prevailing intellectual assumptions of most humanists – that an "interface", whether<br />

the windows and icons of a website or the placement of a poem on a page, can somehow<br />

be on<strong>to</strong>logically decoupled from whatever "content" it happens <strong>to</strong> embody. This is<br />

precisely the point at which Brenda Laurel begins her dissent in Computers as Theatre,<br />

her influential critique of mainstream interface theory:<br />

Usually we think about interactive computing in terms of two things: an application and<br />

an interface. In the reigning view, these two things are conceptually distinct: An<br />

application provides specific functionality for specific goals, and an interface represents<br />

that functionality <strong>to</strong> people. The interface is the thing that we communicate with – the<br />

thing we "talk" <strong>to</strong> – the thing that mediates between us and the inner workings of the<br />

machine. The interface is typically designed last, after the application is thoroughly<br />

conceived and perhaps even implemented; it is attached <strong>to</strong> a preexisting bundle of<br />

"functionality" <strong>to</strong> serve as its contact surface.<br />

(Laurel 1991: xv)<br />

Laurel is writing <strong>to</strong> challenge this prevailing viewpoint. Yet the truth is that, from a<br />

developer's perspective, the interface is often not only conceptually distinct, but also<br />

computationally distinct. As John M. Carroll points out, it wasn't until the comparatively<br />

recent advent of languages like Visual Basic that it even became practical <strong>to</strong> program<br />

both a user interface and an application's underlying functionality with the same code<br />

(Carroll 2002: xxxii). Like the his<strong>to</strong>ry of hand-press printing, which teaches us that<br />

paper-making, typesetting, etching or engraving, and bookbinding came <strong>to</strong> encompass<br />

very different domains of labor and technical expertise (rarely housed under the same<br />

roof), resource development in the <strong>digital</strong> world is typically highly segmented and<br />

compartmentalized.<br />

Interfaces are mostly discussed in mundane and utilitarian terms, but computing lore<br />

contains its share of instances where poor interface design has had lethal and catastrophic<br />

results. One no<strong>to</strong>rious episode involved the Therac-25, a machine employed for cancer<br />

radiation therapy in the mid-1980s, whose cumbersome software interface contributed <strong>to</strong><br />

several patient deaths from overexposure. Likewise, the small-plane accident that killed<br />

singer-songwriter John Denver has been attributed <strong>to</strong> a poorly designed cockpit interface,<br />

specifically the placement of a fuel switch. While the stakes in <strong>digital</strong> <strong>humanities</strong><br />

research are (happily) other than life and death, interface is nonetheless an indispensable<br />

component of any project.<br />

Indeed, interface presents a number of interesting and unique problems for the <strong>digital</strong><br />

humanist. Understandably driven by pragmatic and utilitarian needs, the interface is also<br />

where representation and its attendant ideologies are most conspicuous <strong>to</strong> our critical<br />

eyes. Ostensibly wedded <strong>to</strong> the ideal of user-friendliness, the interface is also where we<br />

deploy our most creative features and imaginative flourishes. Too often put <strong>to</strong>gether as<br />

the final phase of a project under a tight deadline and an even tighter budget, the interface<br />

becomes the first and in most respects the exclusive experience of the project for its end


users. Seemingly the most creative or intuitive stage of the development process, the<br />

interface is also potentially the most empirical, subject <strong>to</strong> the rigorous quantitative<br />

usability testing pioneered by HCI. This chapter makes no attempt <strong>to</strong> offer a<br />

comprehensive survey of the vast professional literature on interface and usability, nor<br />

does it seek <strong>to</strong> serve as a design primer or guide <strong>to</strong> best practices. (Readers interested in<br />

those <strong>to</strong>pics should consult the chapter's suggestions for further reading.) Instead, in the<br />

pages that follow I seek <strong>to</strong> develop some broader frameworks for thinking about the<br />

major challenges that interface poses for both theorists and developers in the <strong>digital</strong><br />

<strong>humanities</strong>.<br />

Ways of Seeing<br />

Computers compute, of course, but computers <strong>to</strong>day, from most users' points of view, are<br />

not so much engines of computation as venues for representation. The first use of CRT<br />

displays as output devices in the 1950s irretrievably situated computers within a complex<br />

cultural genealogy of screens, a genealogy which also includes video, television, cinema,<br />

pho<strong>to</strong>graphy, and indeed, as Lev Manovich and others have argued, the full lineage of<br />

visual aesthetics in the West since the advent of perspective. This context is important<br />

because it allows interface design <strong>to</strong> take its place alongside the other representational<br />

forms that have inhabited our many varieties of screens, frames, and windows. Moreover,<br />

although a graphical user interface built around the familiar desk<strong>to</strong>p and windows<br />

metaphor currently constitutes the normative interface experience for the vast majority of<br />

users, it is worth remembering that this particular graphical environment, and the<br />

representational practices it encourages, is his<strong>to</strong>rically quite specific.<br />

From one perspective, our interface and display technologies have remained remarkably<br />

stable and consistent over the past thirty years. Though the Al<strong>to</strong>, built at Xerox PARC in<br />

1973, is widely regarded as the first computer <strong>to</strong> implement a functional graphical user<br />

interface, the more important date is probably 1968, when Stanford's Douglas Engelbart<br />

demonstrated an operational GUI <strong>to</strong> a standing-room-only audience in San Francisco.<br />

Steven Johnson articulates the importance of Engelbart's presentation, which included a<br />

feature recognizable as "windows", this way:<br />

[H]is<strong>to</strong>rians a hundred years from now will probably accord it the same weight and<br />

significance we now bes<strong>to</strong>w on Ben Franklin's kite-flying experiments or Alexander<br />

Graham Bell's accidental phone conversation with Watson. Engelbart's thirty-minute<br />

demo was our first public glimpse of information-space, and we are still living in its<br />

shadow.<br />

(Johnson 1997: 11)<br />

By "information-space", Johnson means the abrupt transformation of the screen from a<br />

simple and subordinate output device <strong>to</strong> a bounded representational system possessed of<br />

its own on<strong>to</strong>logical integrity and legitimacy, a transformation that depends partly on the<br />

heightened visual acuity a graphical interface demands, but ultimately on the combined<br />

concepts of interactivity and direct manipulation. Engelbart's demo included the first


public use of a mouse, and the sight of its pointer sweeping across the screen instantly<br />

collapsed the stark input/output rhythms of batch-process and command-line computing<br />

in<strong>to</strong> a single, continuous sweep of user activity. Just as important, however, was the<br />

spectacle of Engelbart dividing his screen in<strong>to</strong> distinct regions, heterogeneous in content<br />

but spatially adjoining, a feat demonstrated most dramatically by a window with a live<br />

audio and video feed <strong>to</strong> a colleague in Menlo Park, California. The computer – or more<br />

specifically, the screen – had clearly become a much more complex representational<br />

space, an information space whose surface owed as much <strong>to</strong> modernist collage as it did <strong>to</strong><br />

brute force calculation. A crucial refinement came several years later when a team at<br />

Xerox PARC, led by Alan Kay (and including many of Engelbart's former Stanford<br />

colleagues), arrived at the realization that windows could actually overlap, thereby<br />

immediately imbuing the screen with the third dimension we take for granted <strong>to</strong>day. As<br />

Johnson suggests, "The whole idea of imagining a computer as an environment, a virtual<br />

world, comes out of this seemingly modest innovation" (1997: 47).<br />

While Engelbart's 1968 demo is the most venerable <strong>to</strong>uchs<strong>to</strong>ne for contemporary humancomputer<br />

interaction, there are other origin s<strong>to</strong>ries that bear repeating. Ivan Sutherland,<br />

for example, working on a PhD thesis at MIT in 1963, introduced Sketchpad, a system<br />

that allowed users <strong>to</strong> draw lines on a screen in real time with what we would <strong>to</strong>day<br />

recognize as a light pen. Sketchpad is significant because it reminds us that the current<br />

hegemony of mouse and keyboard was not always in place, and indeed, there are<br />

indications that alternative input devices like light pens – which fundamentally alter the<br />

nature of one's bodily interaction with a computer – may once again displace the mouse<br />

(see this chapter's "Coda" section, below). Sketchpad is also significant in another<br />

context, the his<strong>to</strong>ry of computer graphics (without which there would be no graphical<br />

user interfaces). Nicholas Negroponte comments:<br />

The achievement was of such magnitude and breadth that it <strong>to</strong>ok some of us a decade <strong>to</strong><br />

understand and appreciate all of its contributions. Sketchpad introduced many new<br />

concepts: dynamic graphics, visual simulation, constraint reduction, pen tracking, and a<br />

virtually infinite coordinate system, just <strong>to</strong> name a few. Sketchpad was the big bang of<br />

computer graphics.<br />

(1995: 103)<br />

Sketchpad was a vec<strong>to</strong>r system, meaning that the lines and shapes drawn by the user were<br />

mathematical formulas (vec<strong>to</strong>rs) that could be reproduced at will on-screen. Vec<strong>to</strong>r<br />

images provide another important context for understanding the significance of<br />

Englebart's work because the latter led directly <strong>to</strong> Xerox PARC's refinement of<br />

bitmapping as an alternative <strong>to</strong> Sketchpad and the era's prevailing vec<strong>to</strong>r displays. A<br />

"bitmap", as many readers will know, is a grid or matrix of pixels ("picture elements"),<br />

which, not unlike a Seurat painting or a pho<strong>to</strong>graphic half<strong>to</strong>ne, yield a coherent visual<br />

image through the optical interpretation of the aggregate composition. Bitmapped<br />

displays are what permit the gorgeous, high-quality facsimile images that we see in many<br />

of <strong>to</strong>day's <strong>digital</strong> <strong>humanities</strong> projects, but their significance is much greater. (Note that<br />

bitmap displays, which are also known as raster displays, can refer <strong>to</strong> individual image


files in formats such as JPEG, TIFF, or GIF, as well as <strong>to</strong> an entire screen display; there<br />

is also a proprietary image format known as "Bitmap", or BMP, which is not <strong>to</strong> be<br />

confused with bitmapping as a general concept.) If vec<strong>to</strong>r images were the graphical<br />

inheritance of the computer's mathematical roots, bitmapped images were the visual<br />

realization of Turing's ideal of the universal machine: bitmaps enabled the computer<br />

screen <strong>to</strong> function as a representational surface capable of emulating other<br />

representational surfaces. Through bitmapping, the computer screen was transformed in<strong>to</strong><br />

a second-order or "meta" representational venue. This transformation quickly gave rise <strong>to</strong><br />

intensive research on pho<strong>to</strong>realistic rendering techniques in computer graphics as well as<br />

(eventually) the advent of hardware devices like scanners and <strong>digital</strong> cameras – which<br />

enable the computer screen <strong>to</strong> operate in the pho<strong>to</strong>graphic tradition. (JPEG compression<br />

algorithms, it is worth noting, were introduced precisely <strong>to</strong> provide an image format that<br />

lent itself <strong>to</strong> reproducing pho<strong>to</strong>graphic images.)<br />

William M. Ivins, in Prints and Visual Communication (1953), his landmark survey of<br />

print-making technologies in the West, argues eloquently for the importance of<br />

pho<strong>to</strong>graphy and what he terms "exactly repeatable visual statements" in enabling the<br />

dissemination of scientific knowledge. Bitmapping, I would argue, endows the computer<br />

screen with much those same qualities and capabilities, and although Manovich is right <strong>to</strong><br />

point <strong>to</strong> the origins of computer graphics in the vec<strong>to</strong>r images of Cold War radar displays,<br />

the visual authority of computers as we know it <strong>to</strong>day clearly owes more <strong>to</strong> the advent of<br />

bitmapping. Taken <strong>to</strong>gether, however, Sutherland and Englelbart laid the foundations for<br />

contemporary computer graphics and <strong>to</strong>day's graphical user interface through their<br />

competing paradigms of vec<strong>to</strong>r and bitmap displays; competing not in a strict commercial<br />

sense, but in offering alternative visions of the computer as a representational medium<br />

and as an information space.<br />

Unlike bitmap or raster graphics, vec<strong>to</strong>r graphics are not well suited <strong>to</strong> representing<br />

continuous <strong>to</strong>ne (especially pho<strong>to</strong>graphic) images. Consequently, they may seem of little<br />

use in the <strong>digital</strong> <strong>humanities</strong>, where much of our work consists in providing high-quality<br />

facsimile renderings of documents, artwork, and other artifacts of cultural heritage.<br />

However, because vec<strong>to</strong>r graphics exist as a kind of mathematical abstraction, with no<br />

one-<strong>to</strong>-one mapping <strong>to</strong> an external referent, they are scalable and modular in ways that<br />

raster graphics are not. Today the most popular vehicle for vec<strong>to</strong>r graphics is the<br />

animation <strong>to</strong>ol Flash, which, characterized by its colorful, dynamic displays, is rapidly<br />

colonizing large segments of the Web; indeed, there are those who believe that at the<br />

level of interface design the Web itself will eventually be made over as an animated<br />

Flash-based environment, with static HTML (more likely, XML) documents existing as<br />

subsidiary, special-purpose content. Interestingly for our purposes, Flash is also capable<br />

of supporting embedded bitmapped images, suggesting that the representational field I<br />

described earlier has receded by one full order of magnitude and that vec<strong>to</strong>r graphics are<br />

now the true heir <strong>to</strong> Turing's universalism. (On the other hand, it remains true that all<br />

general-purpose screen displays, whether LCD or CRT, are rendered as bitmaps.)<br />

To build a Flash animation, the designer creates a so-called "movie" consisting of a series<br />

of timed, sequenced, or triggered events. For some this may suggest that the Web is


evolving in<strong>to</strong> a medium that owes more <strong>to</strong> television than <strong>to</strong> the now-familiar (and<br />

comfortably humanistic) conceit of the "page." I would cast the argument differently. If,<br />

as Jerome McGann has repeatedly said, computers lend themselves <strong>to</strong> representing books<br />

because they exist on a different material plane, then vec<strong>to</strong>r images, whose mathematical<br />

construction allows them <strong>to</strong> operate in registers unavailable <strong>to</strong> raster images, may offer<br />

the opportunity for a similar re-conception of our current information spaces. Whereas<br />

Alan Kay's overlapping windows added a third dimension <strong>to</strong> our interfaces, event-based<br />

vec<strong>to</strong>r graphics may, I believe, give rise <strong>to</strong> more playful and pliable information places:<br />

interfaces that occupy a material middle ground between the bitmapped data objects they<br />

enfold and the bitmapped surface of the screen.<br />

If all of that sounds hopelessly speculative and abstract, there are nonetheless some<br />

tantalizing glimpses <strong>to</strong> be had from the <strong>digital</strong> art world. Tomoko Takahashi's<br />

WordPerbect [sic], for instance, a Flash-based parody of a typical word processing<br />

interface: users encounter what appears <strong>to</strong> be a hand-drawn car<strong>to</strong>on GUI, roughly limned<br />

in black ink on a white background. Typing generates crude, seemingly handwritten, lessthan-perfect<br />

characters on the screen. Takahashi's word processor is fully functional, but<br />

the interface yields an inversion of the typical user-friendly experience, one that serves <strong>to</strong><br />

underscore the distinctive materialities of both print and electronic textuality. Clicking<br />

the Mail icon produces the following set of instructions, which appear as a scrap of<br />

notepaper "taped" <strong>to</strong> the screen: "print the document, put in<strong>to</strong> an envelope or ssomething<br />

similair [sic] which can contain the document. Go <strong>to</strong> post office and weigh it and buy<br />

stamps…" and so on, for another hundred words, including further typos and blemishes.<br />

Jason Nelson, another Flash artist, in a piece entitled the last machine with moving parts,<br />

deploys the color picker interface familiar <strong>to</strong> users of image processing and paint<br />

programs <strong>to</strong> control the sequence of events in an animated word poem. As the reader<br />

moves the picker over the color palette, arrangements of words are pulled in and out of<br />

the visual field. "So the colors / cover the wires", reads this text at one point – and so the<br />

brilliant disguises of Nelson and Takahashi's localized parodies of interface culture cover<br />

(and discover) the hardwired his<strong>to</strong>ries of our information spaces.<br />

The Artist of the Beautiful<br />

Mary Shelley's 1818 novel Frankenstein is often invoked in discussions of computing for<br />

its meditation on the dualistic nature of science and technology, and has long been read as<br />

a parable of the promise and the peril of the Promethean flame. Yet Frankenstein is also,<br />

importantly, a novel of aesthetics. The anonymous creature at the center of the text<br />

("Frankenstein's monster") is described repeatedly as a "wretch", "a thing such as even<br />

Dante could not have conceived" (Shelley 1992: 57). In fact, the creature's visage is so<br />

hideous that apart from Vic<strong>to</strong>r (his crea<strong>to</strong>r), the only person who can stand <strong>to</strong> conduct an<br />

extended conversation with him is a blind man.<br />

It might be tempting <strong>to</strong> adopt Frankenstein for my own ends and make Shelley's creature<br />

in<strong>to</strong> a figure for the graphical user interfaces of the current day and age; wretched and<br />

hideous are qualifiers we can debate, but there is no doubt that most of our desk<strong>to</strong>p views<br />

are oddly unlovely, dull and listless information spaces that, as has been pointed out


many times, hew <strong>to</strong> the conspicuously corporate metaphor of the office. But I would like<br />

<strong>to</strong> turn instead <strong>to</strong> another, less well-known text in the Frankenstein tradition: Nathaniel<br />

Hawthorne's (1844) short s<strong>to</strong>ry, "The Artist of the Beautiful." It is the tale of one Owen<br />

Warland, a watchmaker with an exquisitely refined aesthetic sensibility who, bored with<br />

crafting mechanisms in the service of "old blind Father Time", eventually succeeds in his<br />

experiments <strong>to</strong> "spiritualize machinery", imbuing a tiny, mechanical butterfly with the<br />

living breath of his artist's imagination:<br />

He produced, as he spoke, what seemed a jewel box…. This case of ebony the artist<br />

opened, and bade Annie place her fingers on its edge. She did so, but almost screamed as<br />

a butterfly fluttered forth, and, alighting on her finger's tip, sat waving the ample<br />

magnificence of its purple and gold-speckled wings, as if in prelude <strong>to</strong> a flight. It is<br />

impossible <strong>to</strong> express by words the glory, the splendor, the delicate gorgeousness which<br />

were softened in<strong>to</strong> the beauty of this object. Nature's ideal butterfly was here realized in<br />

all its perfection; not in the pattern of such faded insects as flit among earthly flowers, but<br />

of those which hover across the meads of paradise for child-angels and the spirits of<br />

departed infants <strong>to</strong> disport themselves with. The rich down was visible upon its wings;<br />

the lustre of its eyes seemed instinct with spirit. The firelight glimmered around this<br />

wonder – the candles gleamed upon it; but it glistened apparently by its own radiance,<br />

and illuminated the finger and outstretched hand on which it rested with a white gleam<br />

like that of precious s<strong>to</strong>nes.<br />

(Hawthorne 1948: 235–6)<br />

Unlike Shelley's Frankenstein, however, there is some ambiguity as <strong>to</strong> the true nature of<br />

Warland's creation – whether it is indeed a living creature or a delicate clockwork<br />

au<strong>to</strong>ma<strong>to</strong>n. "But is it alive?" the onlookers in the s<strong>to</strong>ry ask repeatedly. Warland refuses <strong>to</strong><br />

answer definitively: "'Wherefore ask who created it, so it be beautiful?' … 'it may well be<br />

said <strong>to</strong> possess life, for it has absorbed my own being in<strong>to</strong> itself; and in the secret of that<br />

butterfly, and in its beauty, – which is not merely outward, but deep as its whole system,<br />

– is represented the intellect, the imagination, the sensibility, the soul of an Artist of the<br />

Beautiful!'" (1948: 237).<br />

As in Frankenstein, "The Artist of the Beautiful" is structured by the classic binaries of<br />

nature and culture, animate spirit and technological artifice; but it is also a s<strong>to</strong>ry about<br />

form versus function, art for art's sake colliding with honest industry and labor. Warland's<br />

chief rival and antagonist is Robert Danforth, the village blacksmith, whose response <strong>to</strong><br />

the butterfly is <strong>to</strong> exclaim, "There is more real use in one downright blow of my sledge<br />

hammer than in the whole five years' labor that our friend Owen has wasted on this<br />

butterfly" (1948: 238). These are the same binaries that have structured debate in<br />

interface and design since Apple introduced the Macin<strong>to</strong>sh; it is not hard <strong>to</strong> hear the echo<br />

of Danforth's sledgehammer in the firm finger striking the Return key at the end of the<br />

command line, or <strong>to</strong> see Warland's labored handicraft reflected in the incessant mouse<br />

manipulations of the GUI. What role, then, should aesthetics have in interface design?<br />

How do we balance the competing demands of truth and beauty? For while most software<br />

and websites have pragmatic or functional ends, an interface such as Apple's OS X – with


its zooming windows, gossamer transparencies, and luscious drop shadows – encourages<br />

us <strong>to</strong> cultivate an aesthetic sensibility even in the most mundane corners of the desk<strong>to</strong>p.<br />

Yale computer scientist David Gelernter has written at length on the subject of aesthetics<br />

in software design:<br />

Most computer technologists don't like <strong>to</strong> discuss it, but the importance of beauty is a<br />

consistent (if sometimes inconspicuous) thread in the software literature. Beauty is more<br />

important in computing than anywhere else in technology. … Beauty is important in<br />

engineering terms because software is so complicated…. Beauty is our most reliable<br />

guide <strong>to</strong> achieving software's ultimate goal: <strong>to</strong> break free of the computer, <strong>to</strong> break free<br />

conceptually…. But as we throw off the limits, what guides us? How do we know where<br />

<strong>to</strong> head? Beauty is the best guide we have.<br />

(Gelernter 1998: 22–3)<br />

Gelernter, who often comes across as an unreconstructed Pla<strong>to</strong>nist in his writing, goes on<br />

<strong>to</strong> speak of "deep beauty", his term for an idealized integration of form and function that<br />

bears a striking resemblance <strong>to</strong> Owen Warland's statement (above) that the beauty he<br />

apprehends is "not merely outward, but deep as its whole system." Gelernter also quotes<br />

approvingly Ted Nelson's dictum that "the integration of software cannot be achieved by<br />

committee…. It must be controlled by dicta<strong>to</strong>rial artists" (1998: 22). His<strong>to</strong>rically,<br />

however, human-computer interaction has its origins not in the poet's eye in a fine frenzy<br />

rolling, but rather in the quantitative usability testing of the sort pioneered by Ben<br />

Shneiderman and his colleagues in the 1970s. (Shneiderman himself, interestingly, has<br />

just recently put forward the Renaissance artist/technologist Leonardo da Vinci as the<br />

inspirational muse for his vision of a user-oriented "new computing.") I propose <strong>to</strong> begin<br />

addressing the aesthetics of interface by narrowing the field <strong>to</strong> look closely at two<br />

projects, both from the realm of text analysis, and each of which, it seems <strong>to</strong> me, exhibits<br />

a "beautiful" user interface.<br />

The first is a pro<strong>to</strong>type <strong>to</strong>ol called Eye-ConTact. Conceived and implemented by<br />

Geoffrey Rockwell and John Bradley in explicit relation <strong>to</strong> its predecessor TACT, Eye-<br />

ConTact is perhaps not beautiful in the superficial sense of being (especially) pleasing <strong>to</strong><br />

look at; nonetheless, I believe the program well instantiates Gelernter's "deep beauty" in<br />

its stated approach <strong>to</strong> textual analysis. Rockwell describes its operations this way:<br />

Eye-Contact deals with the problem of recording the logic of an exploration by<br />

encouraging the user <strong>to</strong> lay out the fundamental steps in a visual environment. The user<br />

creates the logic by dragging out icons and "connecting the dots." This has the advantage<br />

of acting as both a record of the flow of choices made and a synoptic description of that<br />

flow, which should make the research easier <strong>to</strong> grasp.<br />

(Rockwell, website)


Note then that Eye-ConTact requires the user <strong>to</strong> make explicit procedural choices which<br />

are then recorded and represented in an evolving graphical construction of each specific<br />

analytical operation. The software, in other words, serves <strong>to</strong> model a series of ongoing<br />

hermeneutical events (by contrast, TACT lacks the means <strong>to</strong> log a record of a user's<br />

operations with a text). Eye-ConTact is also a realization of what Bruce Tognazzini, who<br />

has had a distinguished career at Apple, celebrates as a visible (as opposed <strong>to</strong> a merely<br />

graphical) interface: "A visible interface is a complete environment in which users can<br />

work comfortably, always aware of where they are, where they are going, and what<br />

objects are available <strong>to</strong> them along the way" (Tognazzini 1992: xiii). Still another way of<br />

thinking about Eye-ConTact is as an example of what happens when an application<br />

adopts as central that which is peripheral in most other <strong>to</strong>ols, in this case the His<strong>to</strong>ry<br />

feature – an approach that offers the basis for some interesting experiments in software<br />

design (see also in this regard Matthew Fuller's essay on his Web Stalker browser). Eye-<br />

ConTact clearly illustrates the way in which a particular <strong>to</strong>ol, rather than seeking <strong>to</strong> hide<br />

its user interface, can instead use that interface as an active – indeed essential –<br />

component of the intellectual activity it aims <strong>to</strong> support: "not merely outward, but deep as<br />

its whole system."<br />

My second example is even more contemporary. In April 2002, information designer and<br />

<strong>digital</strong> artist W. Bradford Paley debuted a web-based <strong>to</strong>ol called TextArc. Drawing upon<br />

publicly available electronic texts in the online collections of Project Gutenberg, TextArc<br />

produces intricate visualizations of novels and other literary works. Every word of the<br />

original text is rendered on-screen, both in a one-pixel font that reprints the entire work<br />

line by line clockwise around the perimeter of the display, and then again in a larger font,<br />

a cloud of words with those appearing most frequently clustered brightly in the center.<br />

Paley describes the result as "[s]ome funny combination of an index, concordance, and<br />

summary." Clicking on a word highlights all of its appearances within the visualized text<br />

as well as generating rays or spokes connecting them so that a user can study whatever<br />

patterns may emerge. The visualization is also dynamically linked <strong>to</strong> a clear reading text<br />

of the work and a keyword-in-context concordance, and <strong>to</strong>gether these <strong>to</strong>ols offer a<br />

powerful package for textual investigation. What is most striking about TextArc,<br />

however, is not its analytical engine, but rather its gorgeous, luminescent fields of display<br />

that seem <strong>to</strong> subordinate traditional hermeneutics <strong>to</strong> more s<strong>to</strong>chastic modes of knowledge<br />

representation. The visualizations have a marked aesthetic dimension, asserting their<br />

integrity on a purely visual register independent of any functional use <strong>to</strong> which they<br />

might be put. Paley unders<strong>to</strong>od this from the start, and indeed supports TextArc (which is<br />

free) in part by selling hard copies of its output, each offset-printed and ready for framing<br />

on high-quality paper.<br />

TextArc has received a great deal of positive press, including mainstream coverage in the<br />

New York Times; yet the basic functionality it provides – word frequency counts,<br />

distribution patterns, and keyword-in-context displays – has long been available with<br />

other <strong>to</strong>ols. The platform-independent Oxford Concordance Program, which was capable<br />

of generating word lists, indices, and concordances for text analysis, first appeared in<br />

1981, followed by desk<strong>to</strong>p software such as WordCruncher (1985) and TACT (1988).<br />

Yet all of these programs, as well as more recent packages like WordSmith Tools, deploy


stark, ascetic interfaces very different from the illuminated constellations of a TextArc<br />

visualization.<br />

Undoubtedly that contrast has much <strong>to</strong> do with the <strong>to</strong>ols and development environments<br />

available <strong>to</strong> the authors of those earlier packages, as well as limited time and limited<br />

resources. These fac<strong>to</strong>rs cannot be overstated: in one software project with which I was<br />

involved we spent eleven months of a one-year development period on the underlying<br />

architecture – which performed flawlessly – but mere days on the final user interface (all<br />

but unusable). Then the money ran out. Such problems are endemic <strong>to</strong> any institutional<br />

setting. But the <strong>digital</strong> <strong>humanities</strong> have also not yet begun (or else only just begun – see<br />

chapter 29, this volume) <strong>to</strong> initiate a serious conversation about its relationship <strong>to</strong> visual<br />

design, aesthetics, and, yes, even beauty. And just as Owen Warland's gilded creation fed<br />

the skepticism of Robert Danforth and the other artisans of his village, so <strong>to</strong>o do<br />

conspicuous graphical displays often engender mistrust in contemporary academic<br />

settings – as though, like the traditional library carrel, our electronic surroundings have <strong>to</strong><br />

be sensually impoverished in order <strong>to</strong> be intellectually viable. Visually suggestive<br />

interfaces are often derided as "slick", or "eye candy", or gratui<strong>to</strong>usly "cool", or else – in<br />

an interesting bit of synaesthesia – <strong>to</strong>o cluttered with "bells and whistles."<br />

To understand what is at issue here we might return <strong>to</strong> Donald A. Norman's contention,<br />

quoted in the opening section: "The real problem with interface is that it is an interface.<br />

Interfaces get in the way. I don't want <strong>to</strong> focus my energies on interface. I want <strong>to</strong> focus<br />

on the job" (2002: 210). Interestingly, despite this fierce disavowal of interface, in their<br />

original typographic presentation Norman's words are printed in boldface, for emphasis.<br />

Pausing for a moment, we can enumerate many typographical conventions for cueing a<br />

reader: not just the weight, but also the size and face of the type, its justification, margins,<br />

and so forth. Not <strong>to</strong> mention more diffuse bibliographical features including paper,<br />

binding, illustrations, and even the size, shape, and heft of the codex. As scholars such as<br />

Johanna Drucker and Jerome McGann (among others) have long argued, these extralinguistic<br />

elements cannot be written off as merely expendable or extraneous <strong>to</strong> "the text<br />

itself." They are, indeed, the working vocabulary of a particular graphical user interface<br />

that has become transparent <strong>to</strong> us only through long familiarity. All of us know how <strong>to</strong><br />

read a modern newspaper or magazine in terms of its visual and typographic layout as<br />

well as its journalistic content. The debate over transparency in interface design that<br />

Donald Norman and Brenda Laurel (and more recently Matthew Fuller, Jay David Bolter<br />

and Diane Gromala) have all participated in thus mirrors the debate in the literary and<br />

textual studies community over the nature of a book's "contact surface", those physical<br />

and material features widely seen as incidental <strong>to</strong> the production of textual meaning. In<br />

both communities, the ideal of transparency is now being called in<strong>to</strong> question and<br />

replaced with a broader awareness of how the visual (and aural, or tactile and olfac<strong>to</strong>ry)<br />

elements on page or screen function as integral aspects of the information experience,<br />

rather than as afterthoughts <strong>to</strong> some "pre-existing bundle of functionality."<br />

Some may object that the language I have been using ("integral aspects of the<br />

information experience", "deep beauty", "deep as the whole system"), in addition <strong>to</strong><br />

fostering a curious kind of New Critical organicism – for Cleanth Brooks, a poem was,


famously, a well-wrought urn – is also militated against by actual information design<br />

practices, which frequently make a virtue of the deliberate and explicit segregation of<br />

form and function. We see this in markup languages such as SGML and XML, when data<br />

tagged with a descriptive schema are then rendered with external stylesheets. Or if not<br />

stylesheets then so-called "skins", which allow different visual themes <strong>to</strong> be swapped in<br />

and out of a web page or desk<strong>to</strong>p application. With the appropriate skins a generic MP3<br />

application can be dressed <strong>to</strong> look like a jukebox or a crystal radio or in a thousand other<br />

guises. Likewise, a site called the CSS Zen Garden offers a striking demonstration of the<br />

power of stylesheets <strong>to</strong> represent the "same" underlying content in a dazzling array of<br />

different configurations. Moreover, as noted earlier, interface is recursive: thus we can<br />

use stylesheets <strong>to</strong> skin our data, skins <strong>to</strong> thematize our browsers and applications, and<br />

desk<strong>to</strong>p themes <strong>to</strong> stylize our information spaces. Do we not then see interface and<br />

application cleaving apart in precisely the ways Laurel and Gelernter speak against?<br />

The truth is that the variability introduced by such effects is literally and deliberately only<br />

skin-deep. Standards compliance and validated code are essential. At the Zen Garden site,<br />

valid XHTML and CSS "cleave" in the other sense of that Janus-word: they work<br />

<strong>to</strong>gether as an integrated whole fused by the centripetal force of an open, communitybased<br />

standards environment. Littering this "road <strong>to</strong> enlightenment", notes the site's<br />

designer, are the "dark and dreary" relics of the past: "browser-specific tags, incompatible<br />

[Document Object Models], and broken CSS support." The visual effects that play across<br />

the shimmering surface of a site such as the Zen Garden are now practical precisely<br />

because the Web design community has come <strong>to</strong> an understanding (as expressed by its<br />

standards and <strong>to</strong>ols) that clearly regards a well-wrought information space as a deeply<br />

integrated system, not just a series of on-screen effects. Compare this <strong>to</strong> the state of<br />

affairs captured in David Siegel's old essay "Severe Tire Damage on the Information<br />

Superhighway", advertised as "an open letter <strong>to</strong> Netscape Communications and the<br />

WWW community", which details the baroque array of spoofs, tricks, workarounds,<br />

fixes, and kludges that characterized professional Web design circa 1996.<br />

If there is a lesson here for the <strong>digital</strong> <strong>humanities</strong> it is simply this: just as interface cannot<br />

– finally – be decoupled from functionality, neither can aesthetics be decoupled from<br />

interface. Nor does talk of beauty always have <strong>to</strong> resort <strong>to</strong> mystification. I ultimately<br />

prefer the poet and critic Lisa Samuels <strong>to</strong> either Gelernter's neo-Pla<strong>to</strong>nism or Hawthorne's<br />

transcendentalism: "[W]e think that those parts of beauty which resist the translation back<br />

<strong>to</strong> knowledge are uselessly private and uncommunicative. In fact, they are what beauty<br />

'knows': that knowledge is also – perhaps most importantly – what we do not yet know….<br />

Beauty is therefore endlessly talk-inspiring, predictive rather than descriptive, dynamic<br />

rather than settled, infinitely serious and useful." The important new wave of software<br />

studies (led by Manovich, Fuller, Bolter and Gromala, and others), which cultivates<br />

granular, material readings of the inevitable cultural and ideological biases encoded by<br />

particular applications and interfaces, offers one way <strong>to</strong> offset the potential risks of<br />

unvarnished aestheticism, and this literature should be watched – many of its contribu<strong>to</strong>rs<br />

are also practicing <strong>digital</strong> artists. In the meantime, in the <strong>digital</strong> <strong>humanities</strong>, where we<br />

deal with the rich and varied legacy of cultural heritage, we ought <strong>to</strong> think about what it


means <strong>to</strong> be artists of the beautiful ourselves. As we will see in the next section, however,<br />

being an artist isn't always easy.<br />

The Blake Archive for Humans<br />

"The Blake Archive for Humans" is the title of a page that collects William Blake<br />

resources on the Web. It is authored and maintained by an amateur enthusiast who links<br />

<strong>to</strong>gether a wide range of material. As visi<strong>to</strong>rs quickly learn, however, the page defines<br />

itself in part through explicit contrast <strong>to</strong> the Blake Archive for "non-humans": that is, the<br />

William Blake Archive (WB A), an extensive scholarly text- and image-encoding project<br />

that has been freely available online since 1995, with funding and support from the Getty<br />

Grant Program, the National Endowment for the Humanities, the Library of Congress,<br />

and the Institute for Advanced Technology in the Humanities at the University of<br />

Virginia. As of this writing the Archive contains SGML-encoded electronic facsimile<br />

editions of some 49 separate copies of all 19 of Blake's illuminated books, as well as a<br />

growing selection of paintings, drawings, separate plates, engravings, and manuscript<br />

materials. The Blake Archive has also won the Modern Language Association's 2002–3<br />

prize for a Distinguished Scholarly Edition, the first time this award has been given <strong>to</strong> an<br />

electronic project.<br />

What could possibly be wrong with any of that? The author of the "Blake Archive for<br />

Humans" site expresses his objections this way:<br />

Their site may be complete and accurate, but it is not particularly easy <strong>to</strong> use, and it's<br />

chock-full of all the ridiculous trappings of the scholarly profession. While such things<br />

certainly have their place, they can also interfere with the appreciation of the actual work<br />

itself. On the other hand, this site is the only place you're going <strong>to</strong> find this many of the<br />

illuminated books in forms you can actually read…. And it, <strong>to</strong>o is a work in progress, so<br />

it will only be getting better; especially when we all have broadband connections.<br />

I will say at the outset that I am not a disinterested party, having been affiliated with the<br />

William Blake Archive since 1997, first as its project manager and now as a consultant. If,<br />

however, as Blake claimed, "opposition is true friendship", then the opinions expressed<br />

above offer a friendlier-than-usual occasion for examining aspects of the WB As<br />

interface – and the development and design process that produced it – in some detail as<br />

this chapter's central case study in usability. For the phrase "the Blake Archive for<br />

humans" can't help but raise the damning specter of divided loyalties, the suggestion that<br />

the Blake Archive, with its admittedly formidable layouts of menus and but<strong>to</strong>n panels and<br />

search tables, may ultimately be more suited <strong>to</strong> a machine's vision than the human eye.<br />

And there is a certain truth <strong>to</strong> that, since the evolution of the Archive has been marked by<br />

constant trade-offs between the (perceived) needs of our (human) users and certain nonnegotiable<br />

demands of our adopted technologies. This divide between humans and<br />

machines is the crux of applied human-computer interaction, and is nowhere more visible<br />

than at the level of the user interface.


The notion that the WBA is "not particularly easy <strong>to</strong> use" and "chock-full" of "scholarly<br />

trappings" in fact typifies a certain range of responses we have received over the years.<br />

On the one hand, while we are happy <strong>to</strong> have users from many different constituencies,<br />

the site's primary mission has always been expressly conceived as scholarly research. To<br />

a certain point we make no apologies for that, since we emphatically believe there is<br />

place for serious <strong>humanities</strong> scholarship – presented in all its cus<strong>to</strong>mary "completeness<br />

and accuracy" – on the Web. On the other hand, however, the comments I quoted above<br />

clearly speak <strong>to</strong> a real, felt frustration that the most comprehensive online resource on its<br />

subject should sometimes prove awkward for the non-specialist <strong>to</strong> navigate and use.<br />

Don't we understand (I hear these users saying) that what really matters here is Blake, and<br />

access <strong>to</strong> "the actual work itself?" Sometimes a particular feature of the site, which may<br />

appear intrusive <strong>to</strong> lay users (one of those aforementioned bells and whistles), is indeed<br />

there for more specialized audiences; for example, the ImageSizer applet, which allows<br />

scholars <strong>to</strong> display Blake's work on-screen at its true physical dimensions. This slippage<br />

between scholarly complexity and bad design is suggestive, however, and deserving of<br />

further study – many years ago, in the pages of the New York Times Review of Books,<br />

Lewis Mumford and Edmund Wilson famously railed against the user interfaces they<br />

found in the edited texts of the Center for Editions of American Authors, with their<br />

"barbed wire" thickets of scholarly apparatus. I suspect something of that same dynamic<br />

is at work here, and the challenge would seem <strong>to</strong> lie in distinguishing legitimate<br />

intellectual complexity from what is merely a poorly executed design.<br />

But that is perhaps easier said than done. The interface users encounter when they come<br />

<strong>to</strong> the Archive on the Web <strong>to</strong>day is known <strong>to</strong> us behind the scenes as "WBA 2.0." In<br />

2000–1 we under<strong>to</strong>ok a major overhaul that introduced several new features in<strong>to</strong> the<br />

Archive and also remedied certain design flaws that had been noted by users and<br />

reviewers.<br />

As Stuart Curran wrote in his review of the 1.0 version of the site (Figure 34.1), the thencurrent<br />

design was "luckily malleable and [could] be altered <strong>to</strong> accommodate altering<br />

circumstances." The revised design we see in WBA 2.0 (Figure 34.2) clearly has been<br />

altered in a number of ways, and there is not space here <strong>to</strong> discuss all of them. A modest<br />

interface element, such as the arrow icons that are visible just beneath the pulldown<br />

menus <strong>to</strong>ward the bot<strong>to</strong>m of the screen, serves well <strong>to</strong> illustrate the nature of the revision<br />

process. The problem was that users who wished <strong>to</strong> view the images of a work in<br />

sequence were forced <strong>to</strong> scroll down the page <strong>to</strong> examine the image and perhaps work<br />

with the transcription, enlargement, or other associated features, and then scroll back up<br />

again <strong>to</strong> where the "Previous" and "Next" links were located. Moreover, the long vertical<br />

profile of the page meant that its <strong>to</strong>p and bot<strong>to</strong>m halves could not both be kept on-screen<br />

at the same time. This turned out <strong>to</strong> be particularly awkward in classroom settings, where<br />

the instruc<strong>to</strong>r was forced <strong>to</strong> do much scrolling upwards and downwards in order present<br />

students with a sequential series of images. The arrows situated in the lower portion of<br />

the screen now alleviate that extraneous scrolling, a simple design solution <strong>to</strong> a more or<br />

less straightforward usability problem.


The generous screen-space around the facsimile image of Blake's work (which<br />

contributes significantly <strong>to</strong> the page's vertical dimensions) is partly the result of the<br />

display area for our ImageSizer applet, but it is also a product of the site's aesthetic<br />

values. From the outset, the WBA was conceived as an image-based resource, with the<br />

"actual work itself" (<strong>to</strong> borrow our friendly critic's phrase) visibly central <strong>to</strong> the user's<br />

active eye, and but<strong>to</strong>ns, links, and apparatus positioned either above or below. While the<br />

earlier version of the site relied on a tables layout for its presentation of this apparatus,<br />

WBA 2.0 employs pulldown menus. The pull-downs offer an efficient way <strong>to</strong> condense a<br />

number of different user options, but there was considerable internal debate among the<br />

project team as <strong>to</strong> whether or not they were the right solution. I reproduce below an<br />

abridged version of a long and thoughtful piece of e-mail from Morris Eaves, one of the<br />

WBA's three general edi<strong>to</strong>rs (Eaves, e-mail <strong>to</strong> blake-proj Discussion List, "WBA 2.0."<br />

July 11, 2000), not only because it captures well the flavor of those internal discussions,<br />

but also because it illustrates the way in which <strong>humanities</strong> scholars are nowadays thrust<br />

in<strong>to</strong> the role of interface designer – an impromptu role that some are quite comfortable in<br />

assuming, while others find it al<strong>to</strong>gether unnatural. Eaves is clearly in the former<br />

category:


34.1 Figure WBA 1.0: The William Blake Archive (1997–<strong>2001</strong>)


34.2 Figure WBA 2.0: William Blake Archive (<strong>2001</strong> <strong>to</strong> the present)<br />

I'm pretty sure I've read the whole discussion of the WBA 2.0 interface from start <strong>to</strong><br />

finish. I know that time is short – but one of the nicest things about these <strong>digital</strong><br />

confections is that they can be changed…. Here are a few personal reactions <strong>to</strong> the<br />

options that have been presented and discussed:…<br />

For me at least, there's a small chasm between form and function here. We all want the<br />

page <strong>to</strong> look good, and we all want it <strong>to</strong> respond readily and efficiently <strong>to</strong> the user's<br />

needs…. The way <strong>to</strong> make it look the best, probably, is <strong>to</strong> put everything on the page<br />

except the image behind a <strong>digital</strong> curtain and hide it from users until they want <strong>to</strong> do<br />

something – then they can open the curtain, rummage through those ugly old technical<br />

contents (previous/next, comparisons, enlargements, indexes, etc.) <strong>to</strong> find what they


want, and close the curtain again <strong>to</strong> recreate the pretty image on a nice, paperlike<br />

background. …<br />

On the other hand, the way <strong>to</strong> make the page function the best is probably <strong>to</strong> pull down<br />

the curtain, arrange all the <strong>to</strong>ols neatly, and label them clearly, so that the user always has<br />

everything needful right at hand. Commercial kitchens usually have open shelves and<br />

overhead racks; home kitchens usually have doors on the cabinets and <strong>to</strong>ols in drawers<br />

that close.<br />

One of our problems is that our user's <strong>to</strong>olbox is so large – much larger than the <strong>to</strong>olbox<br />

of a book-reader. So we have a lot more <strong>to</strong> label, organize, and (maybe) hide until<br />

needed…. It's worth remembering how hard it is for us <strong>to</strong> look at this stuff without<br />

insider's prejudice – after all, we've been using Imagesizer and Inote and several of the<br />

other functions available in those lower areas of the page for a very long time now, so we<br />

constitute the least confusable set of users on earth. But I think it's very hard <strong>to</strong> imagine<br />

other users – and I don't mean dumb users, I mean very smart, alert, expert Blakean users<br />

– who wouldn't welcome some visual help in organizing that formidable display of <strong>to</strong>ols<br />

and information below the image.<br />

PS: Final confession about my kitchen: Georgia and I have two big commercial overhead<br />

racks full of hanging stuff, but then, at her insistence, we also have doors on our cabinets<br />

and drawers that close. I may be showing my hand when I say that I've never been<br />

completely happy about those doors on the cabinets – we built open shelves in our<br />

Albuquerque kitchen and I always loved them. The Rochester kitchen compromise must<br />

have something <strong>to</strong> do with living <strong>to</strong>gether for 100 years.<br />

As it happens, we arrived at a similar compromise (in much less time) for WBA 2.0, the<br />

doors variously put on and left off different interface elements. And as both Curran and<br />

Eaves note, "<strong>digital</strong> confections" like the WBA can always be changed – at least in<br />

theory. But in practice, as I suggested in my introduction, the interface tends <strong>to</strong> come<br />

very late in a project's development cycle (Eaves's recognition above that "time is short"<br />

is a reference <strong>to</strong> the fact that our lead programmer was about <strong>to</strong> be reassigned <strong>to</strong> another<br />

project). Most <strong>digital</strong> <strong>humanities</strong> scholarship is produced incrementally, in layers; at least<br />

for large-scale projects like the WBA, housed within particular institutional ecosystems,<br />

those layers tend <strong>to</strong> be more sedentary than we like <strong>to</strong> admit. The truth is that WBA 2.0 is<br />

unlikely <strong>to</strong> change much in its interface until the site's next major incarnation as WBA<br />

3.0.<br />

There is also another consideration, beyond the exigencies of project management. The<br />

conventional wisdom, reinforced by commercial marketing campaigns like Microsoft's<br />

"Where Do You Want <strong>to</strong> Go Today?" and AT&T's "You Will", is that computers can do<br />

just about anything we want them <strong>to</strong> do. As programmers and developers know, however,<br />

there are always limitations that come with particular hardware and software<br />

environments, and those limitations render the computer an eminently material medium,<br />

this materiality not so different in its way from the characteristic marks and traces left by<br />

Blake's brushes and burins. Both versions of the WBA, for example, rely upon Inso's


DynaWeb software <strong>to</strong> provide the stylesheets needed <strong>to</strong> display our SGML-encoded<br />

electronic editions in a conventional web browser. We have, in the process, cus<strong>to</strong>mized<br />

the appearance and behavior of the DynaWeb environment <strong>to</strong> such an extent that it may<br />

be unrecognizable from its out-of-the-box implementation. But while DynaWeb has been<br />

an enormous boon <strong>to</strong> the Archive, making possible (among much else) powerful search<br />

functions, its idiosyncratic architecture clearly imposes constraints that manifest<br />

themselves at the level of the site's interface. In the case of the navigation arrows that I<br />

discussed above, for example, some readers may have wondered why we opted for icons<br />

on the bot<strong>to</strong>m half of the page but retained the textual "Previous/Next" links up <strong>to</strong>p. The<br />

answer is that (for reasons <strong>to</strong>o esoteric <strong>to</strong> detail) the DynaWeb environment could not be<br />

made <strong>to</strong> accommodate the graphical image files of the icons in the <strong>to</strong>p portion of the<br />

screen. Curran, likewise, notes the strong hierarchical nature of the WBA: "the user must<br />

descend four levels <strong>to</strong> get <strong>to</strong> the texts of the individual illuminated works…. This notion<br />

of penetrating <strong>to</strong> an inner sanctum is, of course, antithetical <strong>to</strong> Blake." We have attempted<br />

<strong>to</strong> rectify this in WBA 2.0 through the addition of Comparison and Naviga<strong>to</strong>r features,<br />

but the broader point is that this particular order of things is largely an artifact of the<br />

DynaWeb architecture, which was originally intended <strong>to</strong> support not Blake's illuminated<br />

visions but, rather, large volumes of text organized in <strong>to</strong>p-down hierarchical structures.<br />

The technologies we work with at the WBA thus constantly make their presence felt,<br />

visibly and palpably pushing back against the interface we attempt <strong>to</strong> enfold around them.<br />

This situation is particularly acute in the <strong>digital</strong> <strong>humanities</strong>, where necessity often<br />

dictates that we adopt and adapt <strong>to</strong>ols and technologies that were originally developed for<br />

other needs and audiences. If, as I suggested earlier, interface design is a dialectic<br />

between the competing demands of human and machine, then the art and science of<br />

usability lies in striking a balance between the two.<br />

The William Blake Archive is – emphatically – for humans. But humans are not<br />

homogeneous, and different users will have different needs. While it will be impossible<br />

<strong>to</strong> please everybody all of the time, a design team must at least ensure that it is meeting<br />

the needs of its most important user communities most of the time. We believe we can<br />

now make this claim. Following the example of the large portals, more and more sites on<br />

the Web are also trending <strong>to</strong>wards user-cus<strong>to</strong>mizable interfaces, and this is a potential<br />

long-term solution for a project like the WBA. But it is a solution that, like all others, will<br />

have <strong>to</strong> be tempered by the non-virtual realities of staffing and development time,<br />

software and data standards, and project funding.<br />

Coda: Magic Carpet Ride<br />

The basic conventions of the desk<strong>to</strong>p windows GUI have not evolved much since their<br />

popular inception with the Apple Macin<strong>to</strong>sh in 1984. Until recently, however, our display<br />

hardware had not changed very greatly either (despite higher resolutions and the<br />

universal shift from monochrome <strong>to</strong> color). But if, as I insisted earlier, the addition of a<br />

glass pane in the 1950s irrevocably situated the computer within the sphere of visual<br />

representation, then everywhere <strong>to</strong>day there are signs that computers may be on the verge<br />

of another broad-based shift in the tangible construction of their information spaces. The<br />

rise of featherweight lap<strong>to</strong>ps, tablet computers, PDAs, and wearable devices, on the one


hand, and wall-sized or room-based projection and display systems, on the other, is even<br />

now wrenching apart the Procrustean setup of the desk<strong>to</strong>p workstation, which has forced<br />

users <strong>to</strong> accept what hindsight will reveal <strong>to</strong> be an almost unbearably constricted and<br />

con<strong>to</strong>rted relationship with our machines (while the ongoing pandemic of carpal tunnel<br />

syndrome and other repetitive strain injuries offer more immediate and irrefutable bodily<br />

evidence). In this same climate, HCI research has also convincingly demonstrated that the<br />

conventions of the desk<strong>to</strong>p GUI do not scale well <strong>to</strong> either larger displays, such as one<br />

might find with wall-sized projections, or <strong>to</strong> smaller displays, such as one now finds on a<br />

PDA.<br />

What, then, does the future hold? Of one thing I am sure: the typewriter and the television<br />

set will not enjoy their conceptual monopoly over our computing machinery for much<br />

longer. My own ideal system might look and feel something like this. I think of it as a<br />

magic carpet: a rectangle of thin, flexible, waterproof plastic, perhaps 3x4 feet, which I<br />

carry about rolled up under my arm (or folded in a bag). I can lay it out on any table<strong>to</strong>p or<br />

flat surface, or else unfold only a corner of it, like newspaper readers on a train. The<br />

plastic sheet is actually an LCD screen, with an embedded wireless uplink <strong>to</strong> the Web.<br />

Applications, both local and remote, appear on its surface like the tiles of a mosaic. I<br />

move them about physically, dragging, shrinking, or enlarging them with my hands,<br />

pushing and pulling them through the information space. Text entry is primarily by voice<br />

recognition. The keyboard, when needed, is a holographic projection coupled <strong>to</strong> a motion<br />

tracker. Data are s<strong>to</strong>red on a solid state memory stick I keep on my keychain, or else<br />

uploaded directly <strong>to</strong> secure network servers.<br />

All of this may sound like over-indulgent science fiction – "Hamlet on the holodeck", <strong>to</strong><br />

borrow a phrase from Janet Murray. But in fact, most of the elements listed here –<br />

wireless networking, voice recognition, keychain data s<strong>to</strong>rage, <strong>to</strong>uch-screen and tangible<br />

user interfaces – are already in common use. And the rest – the motion trackers, paperthin<br />

LCDs, and holographic input devices – are all already at or past the initial<br />

development stage. The "magic" carpet is actually just an extrapolation of real-world<br />

research that is ongoing at places like the Tangible Media Group (and elsewhere) in the<br />

Media Lab at MIT, the Interactivity Lab at Stanford, the Metaverse Lab at the University<br />

of Kentucky, the GVU Center at Georgia Tech, and the Human-Computer Interaction<br />

Lab at the University of Maryland. While the fortunes of any one individual technology<br />

will invariably prove volatile, even a quick scan of these sites leaves no question that the<br />

next decade of research in interface, usability, and HCI will take us for quite a ride. One<br />

of the major challenges for the <strong>digital</strong> <strong>humanities</strong> in the coming decade will therefore be<br />

designing for interfaces (and designing interfaces themselves) outside of the 13- <strong>to</strong> 21inch<br />

comfort zone of the desk<strong>to</strong>p box.<br />

Works Cited or Suggestions for Further Reading<br />

The Blake Archive for Humans. Accessed November 15, 2002. At<br />

http://www.squibix.net/blake/.


Bolter, J. D. and D. Gromala (2003). Windows and Mirrors: Design, Digital Art, and the<br />

Myth of Transparency. Cambridge, MA: MIT Press.<br />

Carroll, J. M. (2002). Human-Computer Interaction in the New Millennium. Bos<strong>to</strong>n:<br />

Addison-Wesley.<br />

Cooper, A. (1995). About Face: The Essentials of User Interface Design. Hoboken, NJ:<br />

John Wiley.<br />

CSS Zen Garden: The Beauty of CSS Design. Accessed December 30, 2003. At<br />

http://www.csszengarden.com/.<br />

Curran, S. (1999). Review of the William Blake Archive. TEXT 12: 216–19.<br />

Drucker, J. (1994). The Visible Word: Experimental Typography and Modern Art, 1909–<br />

1923. Chicago: University of Chicago Press.<br />

Fuller, M. (2003). Behind the Blip: Essays on the Culture of Software. Brooklyn, NY:<br />

Au<strong>to</strong>nomedia.<br />

Gelernter, D. (1998). Machine Beauty: Elegance and the Heart of Technology. New<br />

York: Basic Books.<br />

GVU Center, Georgia Tech. Accessed May 16, 2003. At http://www.cc.gatech.edu/gvu/.<br />

Hawthorne, N. (1948). The Portable Hawthorne, ed. M. Cowley. NewYork: Penguin<br />

Books.<br />

Human-Computer Interaction Lab, University of Maryland. Accessed May 16, 2003. At<br />

http://www.cs.umd.edu/hcil.<br />

Ivins, W. M. (1953). Prints and Visual Communication. Cambridge, MA: MIT Press.<br />

Johnson, S. (1997). Interface Culture: How New Technology Transforms the Way We<br />

Create and Communicate. San Francisco: HarperEdge.<br />

Laurel, B. (1991). Computers as Theatre. Reading, MA: Addison-Wesley.<br />

Manovich, L. (<strong>2001</strong>). The Language of New Media. Cambridge, MA: MIT Press.<br />

McGann, J. (1993). Black Riders: The Visible Language of Modernism. Prince<strong>to</strong>n, NJ:<br />

Prince<strong>to</strong>n University Press.<br />

Metaverse Lab, University of Kentucky. Accessed November 15, 2002. At<br />

http://www.metaverselab.org/.


Negroponte, N. (1995). Being Digital. New York: Knopf.<br />

Nelson, J. (2000). The Last Machine with Moving Parts. Accessed May 14, 2003. At<br />

http://www.heliozoa.com/ending3.html.<br />

Nielsen, J. (2000). Designing Web Usability: The Practice of Simplicity. Indianapolis:<br />

New Riders Publishing.<br />

Norman, D. A. (1990). Why Interfaces Don't Work. In B. Laurel (ed.), The Art of Human-<br />

Computer Interface Design (pp. 209–19). Reading, MA: Addison Wesley.<br />

Norman, D. A. (2002). The Design of Everyday Things. New York: Basic Books.<br />

Paley, B. Text Arc. Accessed November 15, 2002. At http://www.textarc.org.<br />

Raskin, J. (2000). The Humane Interface: New Directions for Designing Interactive<br />

Systems. Bos<strong>to</strong>n: Addison-Wesley.<br />

Rockwell, G. Eye-ConTact: Towards a New Design for Text-Analysis Tools. Computers<br />

and the Humanities Working Papers. November 15, 2002. At<br />

http://www.chass.u<strong>to</strong>ron<strong>to</strong>.ca/epc/chwp/rockwell/index.html.<br />

Samuels, L. (1997). Poetry and the Problem of Beauty. Modern Language Studies 27, 2.<br />

At http://epc.buffalo.edu/authors/samuels/beauty.html.<br />

Shelley, M. (1992). Frankenstein, ed. M. Hindle. London: Penguin Classics.<br />

Shneiderman, B. (1998). Designing the User Interface: Strategies for Effective Human-<br />

Computer Interaction. Reading, MA: Addison-Wesley.<br />

Shneiderman, B. (2002). Leonardo's Lap<strong>to</strong>p: Human Needs and the New Computing<br />

Technologies. Cambridge, MA: MIT Press.<br />

Siegel, D. (1995–6). Severe Tire Damage on the Information Superhighway. At<br />

http://www.dsiegel.com/damage/index.html.<br />

Stephenson, N. (1999). In the Beginning Was the Command Line. New York: Avon.<br />

Takahashi, T. Word Perhect. Accessed May 14, 2003. At http://www.e-<br />

2.org/commissions_wordperhect.html.<br />

Tangible Media Group, MIT Media Lab. Accessed May 16, 2003. At<br />

http://tangible.media.mit.edu/.<br />

Tognazzini, B. (1992). Tog on Interface. Reading, MA: Addison-Wesley.


William Blake Archive. Eds. M. Eaves, R. N. Essick, and J. Viscomi. Accessed<br />

November 15, 2002. At http://www.blakearchive.org.<br />

35.<br />

Intermediation and its Malcontents: Validating<br />

Professionalism in the Age of Raw Dissemination<br />

Michael Jensen<br />

Over the last few years it has become clear that <strong>digital</strong> publishing is far more than just<br />

making a <strong>digital</strong> version of paper available online. Rather, we are (finally) beginning <strong>to</strong><br />

view networked e-content as a new form of publication in its own right, with its own<br />

strengths and weaknesses, its own imperatives, its own necessary logics. What I hope <strong>to</strong><br />

do in the next few thousands of words is <strong>to</strong> explore the issues of networked publishing<br />

from a nonprofit publisher's perspective, having been involved in computers and <strong>digital</strong><br />

publishing within the nonprofit publishing world since the late 1980s.<br />

I'll make a case, I hope, for encouraging a broader engagement between the many sec<strong>to</strong>rs<br />

within scholarly communication (publishers, scholars, libraries, technologists, edi<strong>to</strong>rs,<br />

collabora<strong>to</strong>rs) by dint of common fundamental mission, audiences, and funding sources.<br />

Further, I hope <strong>to</strong> survey some of the themes that are important <strong>to</strong> consider as networked<br />

publishing matures from its infancy, matters of significance <strong>to</strong> anyone considering the<br />

issues raised by <strong>digital</strong> publications in the new century. Finally, I hope <strong>to</strong> address the<br />

value of professionalism in the <strong>digital</strong> publishing arena.<br />

The Nonprofit's Metamission<br />

At an Online Computer Library Center (OCLC) conference in the late 1990s, the new<br />

President of the University of Arizona, Peter. W. Likins, made an intellectually<br />

compelling distinction that I've tried <strong>to</strong> promote ever since:<br />

A for-profit's mission is <strong>to</strong> create as much value for its s<strong>to</strong>ckholders as possible, within<br />

the constraints of society. The non-profit's mission is <strong>to</strong> create as much value for society<br />

as possible, within the constraints of its money.<br />

The academic and educational communication sec<strong>to</strong>rs have much in common, which<br />

ought <strong>to</strong> be driving many of the trends in online communication. How do we take<br />

advantage of this moment of unparalleled interconnectivity <strong>to</strong> "create as much value for<br />

society as possible", given the cultural constraints on our money? Do we maintain the<br />

same structures as before? Do the new capabilities militate for a revolution?<br />

Networked publishing is going through its <strong>to</strong>ddler phase, it seems <strong>to</strong> me – stumbly, never<br />

quite balanced, uncertain of its size and abilities, and amazed and delighted by itself. One


key <strong>to</strong> <strong>to</strong>ddlerhood is an utter ignorance of one's own ability <strong>to</strong> be ignorant. To counter<br />

that, let me first outline what I think are some of the constraints and capabilities of the<br />

current system – from the perspective of a nonprofit publisher.<br />

The Current System<br />

University presses, scholarly societies, academic publishers, and other nonprofit<br />

publishers currently perform a set of societal and scholarly roles, as well as performing a<br />

set of practical roles.<br />

The societal and scholarly roles include identifying promising (and often nascent)<br />

scholarly research, encouraging the development of scholarly work, and enhancing the<br />

accessibility, readability, and comprehensibility of that work. In the context of scholarly<br />

communications in general, these publishers validate scholarly material, integrate and<br />

coordinate scholarship in discipline-specific collections, and (not least) provide authority<br />

for <strong>humanities</strong> and social sciences scholarship for tenure and promotion. They are also<br />

frequently publishers of what are called "midlist" publications – works which may sell<br />

between 500 and 5,000 copies, indicating a small (if committed) audience. It also can be<br />

the scholarly publisher's job both <strong>to</strong> provide a promotional context for a work, and <strong>to</strong> put<br />

the works in<strong>to</strong> the largest context within the discipline or the field.<br />

Everything Old is New Again<br />

For an author, scholar, or scholarly publisher in the twenty-first century, the seemingly<br />

obvious approach <strong>to</strong> the new capabilities – and the seeming imperative – is <strong>to</strong> perform all<br />

these roles in the "e-arena." It's becoming ever more clear that this seeming simplicity is<br />

anything but simple. Let me lay out some of the complexities before moving on.<br />

In the world of print publishing, there is a truism: that "journals publishing can't be done<br />

small" – that developing the skill-set and software for maintaining and applying<br />

individual subscription mechanisms is so costly that you need at least six or eight journals<br />

<strong>to</strong> share the costs. Otherwise, the publishing requirements and concomitant costs cannot<br />

be amortized across enough projects <strong>to</strong> make it affordable. Because of this principle,<br />

among the 120 university presses, only a handful have journals programs. "Stay with<br />

your strengths", says conventional wisdom.<br />

That truism applies <strong>to</strong> technology as well. When the implications of the revolutionary<br />

changes of the last few years are taken in<strong>to</strong> account, it becomes evident that entirely new<br />

systems have <strong>to</strong> be created <strong>to</strong> handle the realities of the New Publishing. It's not as if the<br />

set of problems a publisher must face is only transformed in<strong>to</strong> a new set of problems<br />

when you fac<strong>to</strong>r in networked publishing. Rather, the diversity of problems – and the<br />

diversity of skills necessary <strong>to</strong> address them – blossoms. Entirely new problems,<br />

unconnected <strong>to</strong> anything from the print world, must be confronted.


A print publisher must deal with physical returns of books from books<strong>to</strong>res, while<br />

strategizing about how much content <strong>to</strong> give away on the Press website. A print publisher<br />

must determine print run sizes (because most books are still generally more economic <strong>to</strong><br />

print in quantities of 600 at a time than they are <strong>to</strong> "print on demand" – at least as of<br />

January 2004), while deciding whether <strong>to</strong> include a CD in the book and adding $10 <strong>to</strong> the<br />

retail price. A print and <strong>digital</strong> publisher must not only produce product, record-keep, and<br />

generate monthly reports on the order fulfillment of paper books (an inven<strong>to</strong>ry-based<br />

system), but may also need <strong>to</strong> produce, record-keep, and generate monthly reports on<br />

inven<strong>to</strong>ryless material (e-books), and may need <strong>to</strong> do so on potential-inven<strong>to</strong>ry material<br />

(print-on-demand), and may even need <strong>to</strong> produce, record-keep, and the rest on<br />

subscription-based access methodologies. This is not a decision like "paper or plastic."<br />

It's rather "cash or coins or check card or credit card or check or voucher or foodstamps<br />

or green-stamps or goldstamps or barter." It's not only "transformed", it's<br />

metamorphosed.<br />

The electronic enterprises themselves hold danger: it is painfully easy <strong>to</strong> choose the<br />

wrong preferred technology for an audience or discipline by misunderstanding their<br />

readiness for new technology or misreading their desire for alternative presentation<br />

modes. Or choose the wrong technology for material by force-fitting new <strong>to</strong>ols on old<br />

content, or locking older accommodations <strong>to</strong> print restrictions in<strong>to</strong> new presentations. Or<br />

choose the wrong technology for a changing environment and by so doing lock the<br />

content in<strong>to</strong> a non-adaptive format. Or spend the publication budget on cutting-edge culde-sac<br />

software, or confuse "cool" with "valuable", or conflate readers with purchasers,<br />

or audience with cus<strong>to</strong>mer, such as by presuming that by developing a site aimed at<br />

students <strong>to</strong> promote one's texts, one can entice the libraries serving those markets <strong>to</strong><br />

purchase what their readers are acquiring for free. Or depend overly much on the novelty<br />

value of e-publishing, and believe that simply by building it, you will attract business.<br />

Any single one of these threatens <strong>to</strong> either beggar the publisher or compromise the<br />

project, and shouts "Danger!" <strong>to</strong> most publishers.<br />

Those who have deep and broad knowledge of the processes of publication – publishing<br />

professionals – are vital in many ways. Who better than a professional publisher <strong>to</strong> help<br />

navigate these dangerous waters?<br />

Appropriate Presentational Models<br />

Every living reader <strong>to</strong>day has grown up in a world in which paper dominated the<br />

informationscape in terms of authority, validity, and value. The primary models of<br />

argumentation and presentation in scientific and scholarly discourse are still<br />

predominantly linear, partly because of the nature of exposition and presentation in a<br />

linear-publishing, paper-based world.<br />

This makes it likely that a large proportion of the output of scholars and specialists can be<br />

expected, whether consciously or unconsciously, <strong>to</strong> mimic the modalities of linear<br />

expression. Does that mean that most publications are served best by paper? Perhaps -<br />

though the answer <strong>to</strong> that question depends on many fac<strong>to</strong>rs. Some material is ideal for


online publication (reference works, huge resource bases, highly interconnected and<br />

changing material, multimedia collections), though there may be no simple cost-recovery<br />

system that can be implemented effectively for such publications. Other material may not<br />

be well suited <strong>to</strong> <strong>digital</strong> publication – novels, traditional monographs, linear<br />

argumentation of thesis/promotion/conclusion, poetry, art (still), typographically<br />

challenging material, etc.<br />

Different media serve different content types more or less effectively; it's important <strong>to</strong><br />

recognize what a medium is best suited for. A million JPGs couldn't tell the s<strong>to</strong>ry of<br />

Moby Dick like the words do; a million words can't thrill like the Ninth Symphony. We<br />

are just discovering the potential of networked presentation, but shouldn't be seduced in<strong>to</strong><br />

thinking that Javascript or Flash is the inevitable best (or only eventual) means of<br />

communication.<br />

Appropriate Cost-recovery Models<br />

Just as there are media appropriate for particular communications, there are economic<br />

models appropriate for particular media, content, and audience.<br />

"Cost recovery is destiny", and an organization's mission should dictate cost-recovery<br />

models. For nonprofit publishers with whom dissemination is the priority, new models<br />

are more easily implementable than with nonprofit publishers for whom fiscal stability is<br />

the prime directive. For associations, cost-recovery models are different than for corporately<br />

owned publishers. Fiction publishers have different uses of the medium than<br />

reference publishers do.<br />

Distinguishing between the possible publishing choices is complicated. To make the best<br />

choices, one must understand appropriate (even available) technologies, appropriate cost<br />

recovery and maintenance plans, and what the appropriate media (and cost recovery<br />

mechanisms) are for the particular audience.<br />

What model will work best with your project? Institutional subscription? Print-ready PDF<br />

for a fee? Print-on-demand? Time-based access? Free old material, charged new? Free<br />

new material, charged archive? Making these choices ain't easy.<br />

"Free" Publishing<br />

I once realized that a project I was directing was spending more on developing accessrestriction<br />

mechanisms (for subscription-only access) than it was on developing deeperexploration<br />

mechanisms. Over 60 percent of our development dollars were being spent<br />

on keeping people from reading our material. That seemed silly, and helped convince me<br />

that I should be a bit more open <strong>to</strong> alternative models for scholarly and academic<br />

publishing.


At the National Academies Press, we make more than 3,000 reports of the National<br />

Academy of Sciences, the Institute of Medicine, and the National Academy of<br />

Engineering freely available <strong>to</strong> any browser, hoping <strong>to</strong> encourage browsers <strong>to</strong> purchase<br />

what we consider <strong>to</strong> be an "optimal version" – the print book, or multi-page PDF. We<br />

recognize that nearly all of our publications are designed for print reading (being<br />

predominantly linear exposition of how the best science may inform issues of public<br />

policy), and so see the online versions as only a minor "threat" <strong>to</strong> traditional print sales.<br />

That's one model of "free" publishing – using the Web <strong>to</strong> promote the content, while<br />

selling the container.<br />

In the new networked environment, there are many mechanisms of nearly "outlay-free<br />

publishing" – though it's worth discriminating outlay-free publishing from the mythical<br />

"cost-free" publishing. With enough volunteer labor, or hierarchical control, or<br />

browbeating, or passion, <strong>to</strong>day virtually anything can be made available "for free" via<br />

online presentation. The work may not be best served by this model, but it is a viable<br />

means of presentation.<br />

The Web makes such raw dissemination possible, and it's an exciting prospect. Today,<br />

we're still in the wonderful volunteerism phase, where individual labors of love and<br />

institutional experimentation in<strong>to</strong> electronic publishing are going on. My worry is that<br />

volunteerism is reaching its limits as the true complexities of online publishing begin <strong>to</strong><br />

demand cost-recovery mechanisms for sustenance.<br />

Like many other volunteer efforts, e-projects have proven hard <strong>to</strong> sustain. Countless<br />

worthwhile volunteer projects have died when a central figure moves on, or when the<br />

project reaches that first grand plateau (after which the real work begins), or when the<br />

grant funding for the "e-project" is two years gone.<br />

Promoting a Sustainable Scholarly Information<br />

Infrastructure<br />

Especially in an online world where disintermediation is at play, it's my contention that<br />

there is still a need for some kinds of intermediation – for publishers <strong>to</strong> be involved in the<br />

"making public the fruits of scholarly research" (as Daniel Quoit Oilman, first direc<strong>to</strong>r of<br />

the oldest University Press in the country, Johns Hopkins, so famously said). I'd further<br />

contend that it generally makes more sense for our culture <strong>to</strong> support nonprofit,<br />

networked publishing through its institutional mechanisms (university presses,<br />

association publishing, NGO publishing, and the like) than <strong>to</strong> either relinquish it <strong>to</strong> the<br />

commercial sec<strong>to</strong>r, or <strong>to</strong> presume that volunteerism will enable educated discourse and<br />

debate based on a rich, robust set of resources disseminated online.<br />

To accomplish that, it's important <strong>to</strong> have cultural cost-recovery systems that support the<br />

right processes – that promote experimentation, expand dissemination and access, yet<br />

prevent disruption. Without an appropriate means of supporting the institutional


acquisition, digitization, and presentation of significant resources, real dangers accrue,<br />

the likes of which I've personally witnessed.<br />

The "Publishing Revolution" in Eastern Europe<br />

I saw the results of those dangers first-hand, in another place and time. During the period<br />

1990 <strong>to</strong> 1995, during and after the fall of the Soviet Union, I was involved in a sequence<br />

of educational endeavors helping publishers in Es<strong>to</strong>nia, Latvia, Lithuania, Poland,<br />

Hungary, and Czechoslovakia adapt <strong>to</strong> new technologies, understand how nonprofit<br />

publishing worked in a free-market capitalist system, and helping university presses adapt<br />

<strong>to</strong> the new realities of technology.<br />

By far the deepest experience I had was in Czechoslovakia, before it split in<strong>to</strong> two<br />

countries. Throughout this period I worked with publishers and scholars, frequently<br />

visiting Prague and other <strong>to</strong>wns, and watching as the <strong>to</strong>talitarian socialist model – which<br />

had strengths as well as significant weaknesses – was replaced by systems constructed by<br />

amateur capitalists who had at best a naive understanding of the nature of capitalism.<br />

This created a publishing system with significant weaknesses, as well as (perhaps) a few<br />

strengths, especially in arenas like public-value, scholarly, and educational publishing.<br />

The Soviet model of scholarly publishing was tremendously inefficient. Every university<br />

published its own material for its own students: introduc<strong>to</strong>ry biology coursebooks,<br />

collections of essays, lecture notes, monographs, specific research, all were printed and<br />

bound at the professor's behest. There was heavy subvention from the universities, which<br />

were (in turn) heavily subvened by the state. There was virtually no economic feedback<br />

in this system of scholarly publishing, because there were virtually no cost-recovery<br />

mechanisms at all, apart from a <strong>to</strong>ken fee of the equivalent of about 25 cents or so per<br />

"scripta", as the class publications were called. Hardbacks cost the equivalent of a pack of<br />

Czech cigarettes.<br />

Edi<strong>to</strong>rial selection hardly entered in<strong>to</strong> the matter at the "university publishing" level.<br />

Every year, a few works from university publishing were designated as worthy of being<br />

published in hardback, usually in order <strong>to</strong> give their universities a medium of exchange<br />

for publications from the outside, non-Soviet world (librarians still remember the<br />

awkward near-barter system between Soviet institutions and Western ones).<br />

Often these hardbacks were lavishly produced, but there was rarely any relationship<br />

between "list price" and publication costs. The system separated all indirect costs <strong>to</strong> the<br />

point of immeasurability, by massive bureaucracy, and in fact discouraged any costcontainment<br />

systems based on merit or audience. Instead, decisions were based mostly on<br />

"old-boy" status. An important professor, for example, could insist on having 50,000<br />

copies of his book on the ana<strong>to</strong>my and aerodynamics of bat wings printed (in Czech,<br />

without translation); this was <strong>to</strong> his advantage because his personal royalty was based (by<br />

the Soviet diktat) on number of pages printed, rather than number of copies sold.


Other state-run publishing houses also published scholarly work in philosophy, science,<br />

metaphysics, etc.; they had more freedom of choice of what <strong>to</strong> publish than university<br />

presses, but their work was also heavily subsidized, and prices and print runs were often<br />

at the whim of "important" people. Nonetheless, publishing was considered a social<br />

value, and publishers operated in that way.<br />

When the Soviet Union collapsed, there were countless warehouses with hundreds of<br />

thousands of copies of the writings of Stalin which nobody would buy, even at the<br />

minuscule prices charged in the Soviet days, much less in the post-Soviet economic<br />

crises. And those warehouses also contained about 49,800 copies of that book on the<br />

aerodynamics of the bat wing. In Czech.<br />

But that same socially supported, exceedingly expensive publishing industry produced a<br />

broad variety of valuable, diverse, and very inexpensive books, which deeply affected the<br />

habits of the Czech culture. New books came out every Wednesday, and book sales<br />

bloomed like flowers in a field – every square had books<strong>to</strong>res, every tram s<strong>to</strong>p had a<br />

cardtable with someone selling books.<br />

When I first spent time in Prague, just after the November 1989 revolution, I saw<br />

everyone – and I mean the butcher and the hardhat and the professor alike – reading<br />

books. On the trams, the metro, the streetcorners, and lining up <strong>to</strong> pay a few crowns for<br />

new titles, the citizenry read books – not True Romance, or escapist potboilers but, rather,<br />

philosophy, his<strong>to</strong>ry, science, metaphysics. The inefficient system of subvention had<br />

created a highly literate, exceedingly well-read populace, who read for fun, and read in<br />

depth, all the time.<br />

When the revolution set in fully, suddenly universities, whose subventions were being<br />

completely reconsidered by new governments, were telling their "presses" that they had<br />

<strong>to</strong> become self-sufficient in two years, and many were <strong>to</strong>ld they had <strong>to</strong> start giving money<br />

back <strong>to</strong> their universities – that is, make a profit – in the third year (much like several<br />

major US university presses have been asked recently <strong>to</strong> do). For most of the Czech<br />

university policy-makers, their recipe for capitalism was: add a pinch of slogan-level<br />

ideas picked up from reruns of Dallas, blend with the flour of the Voice of America, add<br />

a dash of Hayek, and finally spice with an understanding gleaned from dinner-table<br />

conversations.<br />

The resulting policies gave little consideration <strong>to</strong> the realities of publishing costs and cost<br />

recovery; had no understanding of the infrastructure needs of the industry (like<br />

distribution, and warehousing, not <strong>to</strong> mention computers, databases, edi<strong>to</strong>rial experience,<br />

knowledge); had no understanding of the place of scholarly publishing in the educational<br />

system; and no recognition that in a revolutionary economy, nobody would have spare<br />

money <strong>to</strong> make discretionary purchases.<br />

Three years after the revolution, the prices for books were often ten <strong>to</strong> fifty times what<br />

they were in 1990. The publishers who were managing <strong>to</strong> survive did so by subsidizing<br />

their continuing translations of Derrida and Roethke with pornography. During that


period, books<strong>to</strong>res closed down everywhere. Publishers closed down across the country.<br />

Citizens s<strong>to</strong>pped reading every day. By 1995, nobody was reading metaphysics on the<br />

tram. A quarter of the university presses I knew of were closed, well over half of the<br />

books<strong>to</strong>res I knew of in Prague were closed, and the scholars I'd befriended were telling<br />

me that they couldn't get anything published any more – there were fewer outlets than<br />

ever.<br />

This transformation was a sad thing <strong>to</strong> watch, especially when I'd been so delighted by<br />

the intense literacy I'd seen initially. The culture turned <strong>to</strong> free or cheap content<br />

(television, newspapers, advertising, radio, Walkmen) rather than <strong>to</strong> the new – and more<br />

complex – ideas contained in books. Promotion and branding became the watchwords of<br />

the day, rather than subtlety of analysis or clarity of presentation.<br />

Let me be clear that I think that neither model was fully right: the absurd redundancies<br />

and inefficiencies of the Soviet system were far <strong>to</strong>o costly for me <strong>to</strong> support. However,<br />

the result was frequently a marvelously high level of intellectual discourse. Regardless,<br />

the follow-on quasi-capitalist system was far <strong>to</strong>o brutal, and had consequences that they<br />

are still feeling in the Czech Republic <strong>to</strong> this day: far fewer high-level publications in<br />

their own language, far fewer high-quality scholarly publications in general (a significant<br />

problem in a small language group), and cultural costs that are hard <strong>to</strong> quantify but easy<br />

<strong>to</strong> identify as causing a kind of intellectual hunger.<br />

What this has <strong>to</strong> do with the current revolution of <strong>digital</strong> presentation should be selfevident,<br />

though the parallels are somewhat indirect. We must recognize that we are in a<br />

revolutionary period, and we must be careful not <strong>to</strong> damage the valuable qualities of the<br />

current system based on inexperienced premises – that is, based on a naive understanding<br />

of the coming state of scholarly communication, which is based on a few visionaries'<br />

description of what is "inevitable." As users and crea<strong>to</strong>rs of scholarly communication, we<br />

must carefully assess what qualities we want <strong>to</strong> maintain from the current system, and be<br />

sure that we create evolutionary pressures that encourage a "scholarly communication<br />

biosystem" which serves scholarship well.<br />

My fear is that without some care, we may undercut (if not destroy) the valuable<br />

characteristics of our current nonprofit publishing system by revolutionizing it without<br />

understanding it. That is, if we revolutionize ourselves out of the ability <strong>to</strong> effectively<br />

support a nonprofit industry disseminating public value, we will have done more harm<br />

than good.<br />

The Value of Publishers<br />

Currently, scholarly publishers – specialists in the selection, preparation, presentation,<br />

and dissemination of scholarly material – are a valuable and necessary part of scholarly<br />

communication. There are some who are still under the false impression that in the new<br />

environment, publishers aren't necessary, just as there are others under the false<br />

impression that libraries are soon <strong>to</strong> be moot, or that universities are outmoded<br />

institutions.


In our "disintermediated world" there is some truth <strong>to</strong> all of those points. Every<br />

intermediary is potentially erasable by the direct producer-<strong>to</strong>-end-user contact made<br />

possible by the existence of the Internet.<br />

It's true, for example, that some professors could teach students directly, without a<br />

university's intervention, or with only one big university's accreditation (and conceivably<br />

only a handful of professors); similarly, scholars could spend the time producing the<br />

complicated presentational structures required for effective online publication, without<br />

the need for a publisher. Universities could conceivably contract with mega-information<br />

providers for just-in-time provision of scholarly content using contractual agents,<br />

replacing the need for an active library.<br />

But I maintain that having specialists do what they do best is the most efficient model in<br />

general, and that intermediaries, though not required, will tailor the intermediation for<br />

particular audiences, and will identify and produce material preferable <strong>to</strong> non-selected,<br />

non-edited, non-vetted, non-enhanced material. None of the intermediaries I mentioned –<br />

the universities, the scholarly community, the scholarly publishers – will become moot in<br />

our lifetime, if for no other reason than the inertia of the credentialing culture combined<br />

with the inertia of presentation systems.<br />

Enough Naive <strong>to</strong> Go Around<br />

For well over a decade I've watched members of the computer scientist community and<br />

the librarian community naively discuss publishing as either a networked-database<br />

problem or as a distribution and classification problem, and conclude that they each can<br />

do the job of publishing better than the existing publishing institutions.<br />

And while some very interesting and valuable projects have come out of both<br />

communities (SPARC, the physics preprint server, UVAs Etext Center, etc.), many of the<br />

rest were diversions of resources that I believe might have been more usefully applied <strong>to</strong><br />

supporting and expanding the scholarly world's existing strengths in the scholarly/<br />

academic publishing arena. The shortcomings in the perspectives of the various interest<br />

groups are worth noting:<br />

• Librarians <strong>to</strong>o often think of scholarly content as primarily something <strong>to</strong> purchase,<br />

categorize, metatag, and archive (and provide <strong>to</strong> patrons), independent of the content's<br />

quality or utility.<br />

• Technologists <strong>to</strong>o often see the world of scholarly content as a superlarge reference set,<br />

crying out for full-text indexing and au<strong>to</strong>mated interconnections. "It's really just a bunch<br />

of XML content types", you can almost hear them muttering.<br />

• Publishers – even nonprofit publishers – <strong>to</strong>o often see the world of scholarly content as<br />

primarily composed of products, each requiring a self-sustaining business model <strong>to</strong> be<br />

deemed "successful." Unless a publication is self-supporting, it's suspect.


• And status-conferring enterprises (like tenure committees) seem <strong>to</strong> see innovative work<br />

engaging <strong>digital</strong> <strong>to</strong>ols in the service of scholarship as being "interesting exercises" akin<br />

<strong>to</strong>, say, generating palindromic verse, or writing a novel without an "o" in it.<br />

The <strong>humanities</strong>/social sciences scholar generally sees his/her work as an abstract and<br />

tremendously personal effort <strong>to</strong> illuminate something previously hidden. S/he sees<br />

technology simultaneously as a hindrance <strong>to</strong>, a boon <strong>to</strong>, an obvious precondition for, and<br />

utterly inconsequential <strong>to</strong>, his/her work. Scholarly communication is all of these, and<br />

even more. There are other players in the game of scholarly publishing: intermediaries,<br />

agents, public information officers, sycophants, supporting scholars and scholarship,<br />

student engagement, pedagogical application, external validation, etc. Each has its own<br />

interests and agendas, and sees the world through its own interest-lens.<br />

The question scholarly publishing should be answering is "how can we most<br />

appropriately support the creation and presentation of intellectually interesting material,<br />

maximize its communicative and pedagogical effectiveness, ensure its stability and<br />

continual engagement with the growing information universe, and enhance the<br />

reputations and careers of its crea<strong>to</strong>rs and sustainers?" If that question is asked, we are<br />

likely <strong>to</strong> answer it.<br />

Without a mission driven by that question, the evolutionary result will be driven by the<br />

interests of the enterprises initiating the publication: library, scholar, nonprofit press,<br />

academic departments, associations – or, if the publication is initiated in the commercial<br />

sec<strong>to</strong>r, by the potential for profit.<br />

Mission-driven Publishing<br />

The projects with sustainable futures, it seems <strong>to</strong> me, are those which are joint<br />

partnerships between stakeholders – such as scholar/publisher enterprises,<br />

publisher/library, scholar/technologist, association/publisher,<br />

association/scholar/publisher, etc. Hybrid vigor is a wonderful thing, and makes for<br />

hearty crops.<br />

The His<strong>to</strong>ry Cooperative (), for example, is a joint<br />

project between the two major his<strong>to</strong>rical associations (the American His<strong>to</strong>rical<br />

Association and the Organization of American His<strong>to</strong>rians) and the University of Illinois<br />

Press, with technology participation from the National Academies Press. These groups<br />

form the executive core of a project enabling the integrated networked publication of<br />

His<strong>to</strong>ry journals, on terms established by the executive body. Currently consisting of nine<br />

journals and growing, the two-year-old project is increasing its specialized <strong>to</strong>ols, its<br />

interdisciplinary value, and its pedagogical utility. More importantly, it has helped these<br />

groups understand each other in a collegial environment of shared goals.<br />

The National Academies Press is another example () – a publisher<br />

"run by its authors" in the sense that it is a dual-mission-driven publisher, expected <strong>to</strong><br />

maximize dissemination of the 180 reports annually produced by the Academies, and


sustain itself through print sales. This expert/publisher/technologist project has resulted in<br />

3,000+ reports being browsed by 7 million visi<strong>to</strong>rs for free in 2002. Every page is<br />

accessible for free, yet tens of thousands of book orders annually support the enterprise.<br />

Project Muse (), a joint publisher/library/journal project, uses<br />

institutional subscription <strong>to</strong> make itself available online. Within subscribed organizations,<br />

free unlimited browsing of over 100 journals is enabled.<br />

CIAO, the Columbia International Affairs Online (), is a joint<br />

publisher/technologist/library project, providing a resource pertaining <strong>to</strong> International<br />

Affairs that combines formal, informal, primary, secondary, and "gray" literature <strong>to</strong>gether<br />

in a coherent, institutional-subscription-based context.<br />

The University of Virginia Press's Electronic Imprint () is<br />

leveraging the strengths of its University's electronic-publishing capabilities (IATH, the<br />

E-text Center, the Press), <strong>to</strong> craft sustainability models for significant electronic<br />

publishing projects. The Press's institutional skill (and ability) in processing income<br />

enables that craftsmanship, which may end up maintaining and growing important works<br />

of scholarship.<br />

These and other examples of multiple institutions recognizing their common goals give<br />

strength <strong>to</strong> the hope that coordination is possible among the various stakeholders in the<br />

scholarly publishing arena.<br />

The Big C<br />

If the nonprofit stakeholders can coordinate, even slowly and gently, <strong>to</strong> achieve that<br />

mission <strong>to</strong> "support the creation and presentation of intellectually interesting material,<br />

maximize its communicative and pedagogical effectiveness, ensure its stability and<br />

continual engagement with the growing information universe, and enhance the<br />

reputations and careers of its crea<strong>to</strong>rs and sustainers", then perhaps their shared interests<br />

will become clearer in confronting the Big C of educational and scholarly publishing:<br />

Copyright.<br />

"The Big C" usually means Cancer, and in this context is intended <strong>to</strong>, since the principles<br />

of copyright are currently metastasizing in ways unhealthy for our culture. The principles<br />

of author's rights have been shanghaied for the benefit of the vested interests of<br />

copyright-holders: publishers, museums, rights-holders, and other potential parasites.<br />

I'm generally disgusted by the intellectual and academic boat-anchor represented by<br />

intellectual property law in the newly networked world. So far, copyright law has done<br />

more <strong>to</strong> preclude full intellectual analysis of our <strong>digital</strong> condition than anything else in<br />

society. As an author, I certainly am glad <strong>to</strong> assert rights <strong>to</strong> my own work, but the blade<br />

honed <strong>to</strong> cut paper doesn't necessarily cut wood well, nor styrofoam, nor tin. That is,<br />

intellectual property law differentiates quite poorly between use types, which means that<br />

I can't glancingly allude (even transclusively) <strong>to</strong> someone else's work without risking


lawsuits; I can't sample Snoop Dogg's samples without that risk; without risking legal<br />

action, I can't quote my own words (that is, if I've contractually signed away rights <strong>to</strong><br />

those words). I certainly cannot harvest the video footage of CNN or the articles of the<br />

New York Times and the Washing<strong>to</strong>n Post, and then use that footage in a multimedia<br />

monograph, analyzing Bush's strategies for forcing a war attitude <strong>to</strong>ward Iraq in the two<br />

pre-election months of 2002 – even if it does not threaten CNN's or the Times's or the<br />

Post's cost-recovery mechanisms.<br />

Intellectual property restrictions, compounded by the attitudes of publishers,<br />

technologists, librarians, and academic bureaucracies, have left many writers and scholars<br />

dispirited. To acquire tenure, the scholar must have work published in print. To do that,<br />

great labor is required on many counts, but developing rich Webliographies is not one of<br />

them. Integrating intellectual-property-controlled material in<strong>to</strong> one's own work is<br />

dangerous, unless it's done in the "dead" form of print.<br />

In light of these barriers <strong>to</strong> full intellectual analysis of our own cultural and social<br />

condition, it's clear <strong>to</strong> me that something needs <strong>to</strong> change. Since the 1990s, nonprofit<br />

publishers have been kneecapping themselves, by routinely siding with the for-profit<br />

publishers on intellectual property and copyright issues, instead of working <strong>to</strong> develop<br />

more culture-friendly standards and practices by engaging more fully and flexibly with<br />

the authors, scholars, societies, associations, libraries, and academic institutions upon<br />

which nonprofit publishing depends.<br />

By making these choices (often passively, by making no choice), the nonprofit publishers<br />

have allowed themselves <strong>to</strong> be tarred with the same brush that is painting the commercial<br />

publishers, who are seen as parasites feeding off the authors, rather than as participants in<br />

the process of scholarly communication. In this way the nonprofit publishing sec<strong>to</strong>r is<br />

impeding its own ability <strong>to</strong> be recognized as a resource; instead, it has allowed the meme<br />

that "all publishers are the same" <strong>to</strong> prevail. From my perspective, nonprofit publishers –<br />

academic, nonprofit, scholarly publishers – are very different from commercial<br />

publishers.<br />

The Needful Intermediary<br />

The naive belief that publishers are a needless intermediary permeates a great deal of the<br />

intellectual culture, mostly because so many publishers can be seen <strong>to</strong> be parasitic freeriders.<br />

Most truly aren't, I believe – we may be parasites, but it's hard work doing it. The<br />

value of the dissemination a good publisher can provide is rarely unders<strong>to</strong>od, and the<br />

ways in which nonprofit, academic publishers can assist the pursuit of knowledge have<br />

been poorly presented.<br />

Remember the Likins quotation from earlier – "a for-profit's mission is <strong>to</strong> create as much<br />

value for its s<strong>to</strong>ckholders as possible, within the constraints of society." Commercial<br />

publishers are by definition trying <strong>to</strong> make money. But the nonprofit publishing world is<br />

by definition about something different, even if it's forced <strong>to</strong> confront the constraints of<br />

budgets and risk. In Likins's words, the goal is <strong>to</strong> "create as much value for society as


possible, within the constraints of its money." That's what the National Academies Press<br />

tries <strong>to</strong> do for its institution – which itself is trying <strong>to</strong> create as much value for society as<br />

possible. We are part service, part business, part entrepreneur, part inven<strong>to</strong>r, and an<br />

interesting model <strong>to</strong> explore.<br />

As the publisher for the National Academies (the National Academy of Sciences, the<br />

National Academy of Engineering, the Institute of Medicine, and the National Research<br />

Council), we produce an actual physical entity – a book. Print is far from dead. In fact, it<br />

is our belief that for at least the next decade, without such an artifact, a National<br />

Academies publication would be perceived as something akin <strong>to</strong> a long, very erudite<br />

memo. We produce something that is a pleasure <strong>to</strong> read, thereby providing the prestige of<br />

a real book, an item that can be purchased and possessed, seen in books<strong>to</strong>res, promoted in<br />

the world, and read in an office, a plane, a bathroom, or a tram.<br />

The Press makes every page of its publications browsable online, and is almost certainly<br />

the most open book publisher in the world. We had almost 9 million visi<strong>to</strong>rs in 2003;<br />

reading almost 53 million report pages. That dissemination allows the recommendations<br />

of the reports <strong>to</strong> be communicated throughout the world <strong>to</strong> anyone wanting <strong>to</strong> read them.<br />

The Press also disseminates and promulgates the works of the National Academies<br />

through traditional means. A colleague recently sent an e-mail <strong>to</strong> me and said, "I ran in<strong>to</strong><br />

an NAP title at Powell's yesterday. It was in the Agriculture section and had a picture of a<br />

field on the cover." That small impact – of being found serendipi<strong>to</strong>usly while performing<br />

concrete <strong>to</strong>pical searches – is easy <strong>to</strong> underestimate, when the goal is <strong>to</strong> provide value for<br />

society. In contrast, that impact is meaningless – in fact not desired – for a commercial<br />

publisher.<br />

The Press's business model provides an economic incentive at the institutional level for<br />

"push" marketing – catalogues, direct mail, e-mail marketing, targeted contacts of all<br />

kinds. Without a business underlying the enterprise, it's hard <strong>to</strong> justify the expense of, for<br />

example, space ads in a particular journal.<br />

The Press also provides an economic brake for political influence from "big swaggerers."<br />

Without a business underpinning, there is little <strong>to</strong> prevent the influence of a loud<br />

institutional voice from directing limited resources <strong>to</strong> his/her own pet project – like<br />

translating a Czech bat wing book. Similar waste is just as easy <strong>to</strong> generate via other "bad<br />

publishing decisions": bad online Webwork, insistence on "multimedia" or interactivity<br />

or animation for no communicative reason, poor design of a print component, insistence<br />

on four-color everywhere, dismal titling, and on and on.<br />

We provide a business reason for good judgment, which can save huge amounts of<br />

money. A "print-on-demand" (POD) book may have a fixed unit cost – say $6 per unit –<br />

regardless of the print run. A unit cost for an offset print run can be as little as $1.50 per<br />

unit. The advantages of POD are certain, within certain constraints; the financial benefit<br />

of good judgment (predicting demand sufficiently <strong>to</strong> decide between traditional offset


and POD) is not encouraged by a POD-only model. Without a business-based motiva<strong>to</strong>r,<br />

routine internal expedience will be the intrinsically rewarded model.<br />

We also provide a business reason <strong>to</strong> generate "pull" – <strong>to</strong> enhance the search value and<br />

search visibility of National Academies texts within the context of the world's search<br />

engines. Without professional attention at that level, there would at best be spotty "search<br />

engine optimization" by those within the institution with the funds and staff <strong>to</strong> do so.<br />

There would be no systematic, institution-wide strategy <strong>to</strong> continuously strengthen the<br />

online visibility of all Academies reports.<br />

Finally, we provide a business motiva<strong>to</strong>r that can justify underwriting the development of<br />

publishing enhancements. Since 1999, the NAP has underwritten the development of<br />

"special collections" software, the development of novel search and information<br />

discovery projects, and much more. These sorts of <strong>to</strong>ols are justified because, on our<br />

website, they do a better job of engaging the value of our unique resources <strong>to</strong> readers. We<br />

want <strong>to</strong> promote that value, because any researcher might be the one visi<strong>to</strong>r in two<br />

hundred who might decide <strong>to</strong> buy the book.<br />

The Press is far from perfect, but it's a successful model of alternative modes of<br />

publishing, responding <strong>to</strong> the unique needs of its parent institution and authors, and <strong>to</strong> the<br />

media requirements of its primary audiences.<br />

Conclusions<br />

If there are four things I hope the reader takes away from this article, they are the<br />

following. First, the depiction of the collapse of a vibrant state of public literacy in<br />

Prague between 1990 and 1995, in which <strong>to</strong>o-sudden changes in the economics of<br />

publishing resulted in the loss of cultural and societal habits that encouraged educational<br />

engagement. Naive revolutionaries are usually the most dangerous. To edi<strong>to</strong>rialize, it's<br />

important <strong>to</strong> recognize that revolutions can <strong>to</strong>o often end up in dicta<strong>to</strong>rships of one kind<br />

or another. In this <strong>digital</strong> revolution, we in the USA are running the risk of ceding (by<br />

passive acceptance) the rights <strong>to</strong> the intellectual property of the public, nonprofit, and<br />

educational sec<strong>to</strong>rs <strong>to</strong> commercial interests, whose metric of value is commercial, not<br />

societal.<br />

Next, I hope I made clear the need <strong>to</strong> acknowledge the limits of volunteerism, and <strong>to</strong> see<br />

that some projects require a foundation beyond the passionate involvement of a few<br />

people; that cost-recovery models can dictate the possibilities. A cost-recovery model can<br />

enable or inhibit more development, may encourage or discourage dissemination and<br />

access, or may allow gymnastics on the high wire, rather than simply forcing a static<br />

balance on the tightrope of survival.<br />

I hope I made the point for a broader engagement between the many sec<strong>to</strong>rs within<br />

scholarly communication (publishers, scholars, libraries, technologists, edi<strong>to</strong>rs,<br />

collabora<strong>to</strong>rs), and made a convincing argument that because of our common


fundamental mission, audiences, and funding sources, we should be collaborating more<br />

with each other, and making accommodations for each other more often.<br />

Finally, I hope that the underlying theme of publishing as a set of choices, with which<br />

professional help is useful, was clear. There are ever more potential traps that every<br />

publishing project of scope and depth must now recognize and elude. Choosing<br />

appropriate partners – outside institutions, intermediaries, dissemina<strong>to</strong>rs – can make the<br />

difference between a project of scholarly/academic/educational significance and a<br />

publishing novelty.<br />

Networked publishing will begin <strong>to</strong> mature in the next decade. I hope that we can work <strong>to</strong><br />

ensure that it matures within a rich community of support and engagement, and that the<br />

public sec<strong>to</strong>r finds ways <strong>to</strong> add value <strong>to</strong> society through a robust culture of publishing.<br />

36.<br />

The Past, Present, and Future of Digital Libraries<br />

Howard Besser<br />

Digital libraries will be critical <strong>to</strong> future <strong>humanities</strong> scholarship. Not only will they<br />

provide access <strong>to</strong> a host of source materials that humanists need in order <strong>to</strong> do their work,<br />

but these libraries will also enable new forms of research that were difficult or impossible<br />

<strong>to</strong> undertake before. This chapter gives a his<strong>to</strong>ry of <strong>digital</strong> libraries. It pays particular<br />

attention <strong>to</strong> how they have thus far failed <strong>to</strong> incorporate several key elements of<br />

conventional libraries, and discusses current and future <strong>digital</strong> library developments that<br />

are likely <strong>to</strong> provide these missing elements. Developments of concern <strong>to</strong> humanists<br />

(such as preservation, the linking of collections <strong>to</strong> one another, and standards) are<br />

discussed in detail.<br />

Why are Digital Libraries Important <strong>to</strong> Humanists?<br />

His<strong>to</strong>rically, libraries have been critical <strong>to</strong> <strong>humanities</strong> scholarship. Libraries provide<br />

access <strong>to</strong> original works, <strong>to</strong> correspondence and other commentary that helps<br />

contextualize those works, and <strong>to</strong> previous <strong>humanities</strong> commentaries.<br />

As we enter the twenty-first century, <strong>digital</strong> libraries appear <strong>to</strong> be as critical <strong>to</strong> <strong>humanities</strong><br />

scholarship as brick-and-mortar libraries were <strong>to</strong> scholarship in previous centuries. Not<br />

only do <strong>digital</strong> libraries provide access <strong>to</strong> original source material, contextualization, and<br />

commentaries, but they also provide a set of additional resources and service (many of<br />

them closely matching <strong>humanities</strong> trends that emerged in the late twentieth century).<br />

Digital libraries allow scholars <strong>to</strong> engage in a host of activities that were difficult or<br />

impossible <strong>to</strong> do before. Libraries, archives, and museums have been gathering <strong>to</strong>gether<br />

high-quality <strong>digital</strong> surrogates of original source material from many different


eposi<strong>to</strong>ries so that they appear <strong>to</strong> be a single reposi<strong>to</strong>ry <strong>to</strong> users (for example, see the<br />

Online Archive of California ) or Arts<strong>to</strong>r<br />

(). Researchers can now consult online facsimiles of rare works<br />

residing in a host of different institutions without having <strong>to</strong> visit each one. Students now<br />

have the opportunity <strong>to</strong> explore facsimiles of rare works and correspondence. Researchers<br />

who engage in lexical analysis now have the opportunity <strong>to</strong> count word/phrase<br />

occurrences or do syntactical analysis not just on a single work, but across a whole body<br />

of works. Digital libraries permit instruc<strong>to</strong>rs and reposi<strong>to</strong>ry managers <strong>to</strong> reflect multiple<br />

interpretations of works, authors, or ideas alongside each other – a realization of<br />

decentered critical authority, one of the tenets of postmodernism.<br />

But it would be a mistake <strong>to</strong> see <strong>digital</strong> libraries as primarily providing ways <strong>to</strong> access<br />

material more quickly or more easily, without having <strong>to</strong> visit a reposi<strong>to</strong>ry across the<br />

country. Though the promise of <strong>digital</strong> technology in almost any field has been <strong>to</strong> let one<br />

do the same things one did before but better and faster, the more fundamental result has<br />

often been the capability of doing entirely new things. It is very possible that <strong>digital</strong><br />

libraries will enable future <strong>humanities</strong> scholars <strong>to</strong> engage in new activities that we haven't<br />

yet envisioned.<br />

What is a Library?<br />

Traditionally, libraries have been more than just collections. They have components<br />

(including service <strong>to</strong> a clientele, stewardship over a collection, sustainability, and the<br />

ability <strong>to</strong> find material that exists outside that collection) and they uphold ethical<br />

traditions (including free speech, privacy, and equal access). In the last decade of the<br />

twentieth century, both technological changes and reduced funding for the public sec<strong>to</strong>r<br />

led <strong>to</strong> significant changes in libraries. For the first time in his<strong>to</strong>ry it became possible <strong>to</strong><br />

divorce the physical aspects of a library from the (<strong>digital</strong>) access and services that a<br />

library provides. This has led <strong>to</strong> much discussion of the past and possible future role of<br />

libraries, and some speculation as <strong>to</strong> whether they have a future. In her discussion of why<br />

conventional libraries will not disappear simply because we develop online collections,<br />

Christine Borgman states that the conventional library's role is <strong>to</strong> "select, collect,<br />

organize, preserve, conserve, and provide access <strong>to</strong> information in many media, <strong>to</strong> many<br />

communities of users." I have argued, elsewhere, that the four core characteristics of a<br />

public library are: that it is a physical place, that is, a focus spot for continuous<br />

educational development, that it has a mission <strong>to</strong> serve the underserved, and that it is a<br />

guaran<strong>to</strong>r of public access <strong>to</strong> information (Besser 1998).<br />

Almost all conventional libraries have a strong service component. All but the smallest<br />

libraries tend <strong>to</strong> have a substantial "public service" unit. Library schools teach about<br />

service (from "public service" courses <strong>to</strong> "reference interviews"). And the public in<br />

general regard librarians as helpful people who can meet their information needs. Many<br />

libraries also deliver information <strong>to</strong> multiple clienteles. They are very good at using the<br />

same collection <strong>to</strong> serve many different groups of users, each group incorporating<br />

different modalities of learning and interacting, different levels of knowledge of a certain<br />

subject, etc. Public libraries serve people of all ages and professions, from those barely


able <strong>to</strong> read, <strong>to</strong> high school students, <strong>to</strong> college students, <strong>to</strong> professors, <strong>to</strong> blue-collar<br />

workers. Academic libraries serve undergraduates who may know very little in a<br />

particular field, faculty who may be specialists in that field, and non-native English<br />

speakers who may understand detailed concepts in a particular domain, but have<br />

difficulty grasping the language in which those concepts are expressed.<br />

Most libraries also incorporate the component of stewardship over a collection. For some<br />

libraries, this is primarily a matter of reshelving and circulation control. But for most<br />

libraries, this includes a serious preservation function over at least a portion of their<br />

collection. For research libraries and special collections, preservation is a significant<br />

portion of their core responsibilities, but even school, public, and special libraries are<br />

usually responsible for maintaining a core collection of local records and works over long<br />

periods of time.<br />

Libraries are organizations that last for long periods of time. Though occasionally a<br />

library does "go out of business", in general, libraries are social entities that have a great<br />

deal of stability. Though services may occasionally change slightly, people rely on their<br />

libraries <strong>to</strong> provide a sustainable set of services. And when services do change, there is<br />

usually a lengthy period when input is solicited from those who might be affected by<br />

those changes.<br />

Another key component of libraries is that each library offers the service of providing<br />

information that is not physically housed within that library. Libraries see themselves as<br />

part of a networked world of libraries that work <strong>to</strong>gether <strong>to</strong> deliver information <strong>to</strong> an<br />

individual (who may deal directly only with his or her own library). Tools such as union<br />

catalogues and services such as inter-library loan have produced a sort of interoperable<br />

library network that was used <strong>to</strong> search for and deliver material from afar long before the<br />

advent of the World Wide Web.<br />

Libraries also have strong ethical traditions. These include fervent protection of readers'<br />

privacy, equal access <strong>to</strong> information, diversity of information, serving the underserved,<br />

etc. (see resolutions of the American Library Association – American Library<br />

Association 1995). Librarians also serve as public guardians over information, advocating<br />

for these ethical values.<br />

The library tradition of privacy protection is very strong. Librarians have risked serving<br />

jail time rather than turn over whole sets of patron borrowing records. Libraries in the<br />

USA have even designed their circulation systems <strong>to</strong> only save aggregate borrowing<br />

statistics; they do not save individual statistics that could subsequently be data-mined <strong>to</strong><br />

determine what an individual had borrowed.<br />

Librarians believe strongly in equal access <strong>to</strong> information. Librarians traditionally see<br />

themselves as providing information <strong>to</strong> those who cannot afford <strong>to</strong> pay for that<br />

information on the open market. And the American Library Association even mounted a<br />

court challenge <strong>to</strong> the Communications Decency Act because it prevented library users<br />

from accessing information that they could access from venues outside the library.


Librarians have been in the forefront of the struggle against the privatizing of US<br />

government information on the grounds that those steps would limit the access of people<br />

who could not afford <strong>to</strong> pay for it.<br />

Librarians also work <strong>to</strong> ensure diversity of information. Libraries purposely collect<br />

material from a wide variety of perspectives. Collection development policies often stress<br />

collection diversity. And librarians pride themselves on being able <strong>to</strong> offer patrons a rich<br />

and diverse set of information.<br />

Librarians are key public advocates for these ethical values. As guardians of information,<br />

they try <strong>to</strong> make sure that its richness, context, and value do not get lost.<br />

As more and more information is available in <strong>digital</strong> form, a common misperception is<br />

that a large body of online materials constitutes a "<strong>digital</strong> library." But a library is much<br />

more than an online collection of materials. Libraries (either <strong>digital</strong> or brick-and-mortar)<br />

have both services and ethical traditions that are a critical part of the functions they serve.<br />

The <strong>digital</strong> collections we build will not truly be <strong>digital</strong> libraries until they incorporate a<br />

significant number of these services and ethical traditions.<br />

Brief Digital Library His<strong>to</strong>ry<br />

The first major acknowledgment of the importance of <strong>digital</strong> libraries came in a 1994<br />

announcement that $24.4 million of US federal funds would be dispersed among six<br />

universities for "<strong>digital</strong> library" research (NSF 1994). This funding came through a joint<br />

initiative of the National Science Foundation (NSF), the Department of Defense<br />

Advanced Research Projects Agency (ARPA), and the National Aeronautics and Space<br />

Administration (NASA). The projects were at Carnegie Mellon University, the University<br />

of California-Berkeley, the University of Michigan, the University of Illinois, the<br />

University of California-Santa Barbara, and Stanford University.<br />

These six well-funded projects helped set in motion the popular definition of a "<strong>digital</strong><br />

library." These projects were computer science experiments, primarily in the areas of<br />

architecture and information retrieval. According <strong>to</strong> an edi<strong>to</strong>rial in D-Lib Magazine,<br />

"Rightly or wrongly, the DLI-1 grants were frequently criticized as exercises in pure<br />

research, with few practical applications" (Hirtle 1999).<br />

Though these projects were exciting attempts <strong>to</strong> experiment with <strong>digital</strong> collections, in no<br />

sense of the word did they resemble libraries. They had little or no service components,<br />

no cus<strong>to</strong>dianship over collections, no sustainability, no base of users, and no ethical<br />

traditions. We will call this the "experimental" stage of <strong>digital</strong> library development (see<br />

Table 36.1). Because efforts during this experimental stage were the first <strong>to</strong> receive such<br />

widespread acknowledgment under the term "<strong>digital</strong> library", they established a popular<br />

understanding of that term that has persisted for many years.


36.1 Table Stages of <strong>digital</strong> library development<br />

Stage Date Sponsor What<br />

I<br />

Experimental<br />

1994 NSF/ARPA/NASA<br />

II Developing 1998/99 NSF/ARPA/NASA,<br />

DLF/CLIR<br />

III Mature ?<br />

Funded through normal<br />

channels?<br />

Experiments on collections of <strong>digital</strong><br />

materials<br />

Begin <strong>to</strong> consider cus<strong>to</strong>dianship,<br />

sustainability, user communities<br />

Real sustainable interoperable <strong>digital</strong><br />

libraries<br />

By 1996, social scientists who had previously worked with conventional libraries began<br />

trying <strong>to</strong> broaden the term "<strong>digital</strong> libraries" (Bishop and Star 1996; Borgman et al.<br />

1996). But the real breakthrough came in late 1998 when the US federal government<br />

issued their highly funded DL-2 awards (Griffin 1999) <strong>to</strong> projects that contained some<br />

elements of traditional library service, such as cus<strong>to</strong>dianship, sustainability, and<br />

relationship <strong>to</strong> a community of users. Also around that time, administra<strong>to</strong>rs of<br />

conventional libraries began building serious <strong>digital</strong> components.<br />

As librarians and social scientists became more involved in these <strong>digital</strong> projects, efforts<br />

moved away from computer science experiments in<strong>to</strong> projects that were more<br />

operational. We shall call this the "developing" stage of <strong>digital</strong> libraries. By the late<br />

1990s, particularly under the influence of the US Digital Library Federation, projects<br />

began <strong>to</strong> address traditional library components such as stewardship over a collection and<br />

interoperability between collections. But even though <strong>digital</strong> library developers have<br />

made great progress on issues such as real interoperability and <strong>digital</strong> preservation, these<br />

are far from being solved in a robust operational environment. In order <strong>to</strong> enter the<br />

"mature" stage where we can really call these new entities "<strong>digital</strong> libraries", they will<br />

need <strong>to</strong> make much more progress in moving conventional library components, such as<br />

sustainability and interoperability, in<strong>to</strong> the <strong>digital</strong> realm. And developers need <strong>to</strong> begin <strong>to</strong><br />

seriously address how they can move library ethical traditions (such as free speech,<br />

privacy, and equal access) in<strong>to</strong> the <strong>digital</strong> realm as well. The remainder of this chapter<br />

examines important efforts <strong>to</strong> move us in those directions.<br />

Moving <strong>to</strong> a More User-centered Architecture<br />

Both the early computer science experiments in <strong>digital</strong> libraries and the earlier initial<br />

efforts <strong>to</strong> build online public access catalogues (OPACs) followed a model similar <strong>to</strong> that<br />

in Figure 36.1. Under this model, a user needed <strong>to</strong> interact with each <strong>digital</strong> reposi<strong>to</strong>ry<br />

independently, <strong>to</strong> learn the syntax supported by each <strong>digital</strong> reposi<strong>to</strong>ry, and <strong>to</strong> have<br />

installed on his or her own computer the applications software needed <strong>to</strong> view the types<br />

of <strong>digital</strong> objects supported by each <strong>digital</strong> reposi<strong>to</strong>ry.<br />

So, in order for a user <strong>to</strong> search Reposi<strong>to</strong>ry A, s/he would need <strong>to</strong> first adjust <strong>to</strong><br />

Reposi<strong>to</strong>ry As specialized user interface, then learn the search syntax supported by this


eposi<strong>to</strong>ry. (For example, NOTIS-based OPACs required search syntax like A=Besser,<br />

Howard, while Inovative-based OPACs required search syntax like FIND PN Besser,<br />

Howard.) Once the search was completed, s/he could retrieve the appropriate <strong>digital</strong><br />

objects, but would not necessarily be able <strong>to</strong> view them. Each reposi<strong>to</strong>ry would only<br />

support a limited number of encoding formats, and would require that the user have<br />

specific software installed on their personal computer (such as viewers for Microsoft<br />

Word 98, SGML, Adobe Acrobat, TIFF, PNG, JPEG, or specialized software distributed<br />

by that reposi<strong>to</strong>ry) in order <strong>to</strong> view the <strong>digital</strong> object. Thus users might search and find<br />

relevant works, but not be able <strong>to</strong> view them.<br />

The user would then have <strong>to</strong> repeat this process with Reposi<strong>to</strong>ry B, C, D, etc., and each of<br />

these reposi<strong>to</strong>ries might have required a different syntax and different set of viewers.<br />

Once the user had searched several different reposi<strong>to</strong>ries, he or she still could not<br />

examine all retrieved objects <strong>to</strong>gether. There was no way of merging sets. And because<br />

different reposi<strong>to</strong>ries supported different viewing software, any attempt <strong>to</strong> examine<br />

objects from several reposi<strong>to</strong>ries would likely require going back and forth between<br />

several different applications software used for display.<br />

Obviously the model in Figure 36.1 was not very user-friendly. Users don't want <strong>to</strong> learn<br />

several search syntaxes, they don't want <strong>to</strong> install a variety of viewing applications, and<br />

they want <strong>to</strong> make a single query that accesses a variety of different reposi<strong>to</strong>ries. Users<br />

want <strong>to</strong> access an interoperable information world, where a set of separate reposi<strong>to</strong>ries<br />

looks <strong>to</strong> them like a single information portal. A more user-friendly model is outlined in<br />

Figure 36.2. Under this model, a user makes a single query that propagates across<br />

multiple reposi<strong>to</strong>ries. The user must only learn one search syntax.<br />

The user doesn't need <strong>to</strong> have a large number of software applications installed for<br />

viewing, and retrieved sets of <strong>digital</strong> objects may be looked at <strong>to</strong>gether on the user's<br />

workstation. The model in Figure 36.2 envisions a world of interoperable <strong>digital</strong><br />

reposi<strong>to</strong>ries, and is a model we need <strong>to</strong> strive for.<br />

Over the years, developers have made some significant progress <strong>to</strong>wards the Figure 36.2<br />

model, particularly in the area of OPACs. Web browsers have provided a common "lookand-feel"<br />

between different reposi<strong>to</strong>ry user interfaces. The Z39–50 pro<strong>to</strong>cols have<br />

allowed users <strong>to</strong> employ a single, familiar search syntax, even when the reposi<strong>to</strong>ry's<br />

native search syntax appears foreign. Z39–50 has also promised <strong>to</strong> let user queries<br />

propagate <strong>to</strong> different reposi<strong>to</strong>ries. But when one leaves the world of OPACs and enters<br />

the world of <strong>digital</strong> reposi<strong>to</strong>ries, much work still needs <strong>to</strong> be done <strong>to</strong> achieve real<br />

interoperability. Most of this work involves creation and adoption of a wide variety of<br />

standards: from standards for the various types of metadata (administrative, structural,<br />

identification, longevity), <strong>to</strong> ways of making those metadata visible <strong>to</strong> external systems<br />

(harvesting), <strong>to</strong> common architectures that will support interoperability (open archives).


General processes and stages of technological development<br />

The au<strong>to</strong>mation of any type of conventional process often follows a series of pragmatic<br />

steps as well as a series of conceptual stages.<br />

Pragmatic implementation steps usually begin by using technology <strong>to</strong> experiment with<br />

new methods of performing some function, followed by building operational systems,<br />

followed by building interoperable operational systems. And at the later stages of this,<br />

developers begin trying <strong>to</strong> make these systems useful for users. We have seen this pattern<br />

(experimental systems <strong>to</strong> operational systems <strong>to</strong> interoperable systems <strong>to</strong> useful systems)<br />

repeated in the development of OPACs, Indexing and Abstracting services, and image<br />

retrieval. The au<strong>to</strong>mation of each of these has begun with experiments, followed by<br />

implementations that envisioned closed operational systems (with known bodies of users<br />

who need <strong>to</strong> learn particular user interfaces and syntaxes <strong>to</strong> interact with the system),<br />

followed by implementations that allowed the user <strong>to</strong> more easily interact with multiple<br />

systems (and sometimes <strong>to</strong> even search across various systems). Today's "<strong>digital</strong><br />

libraries" are not much beyond the early experimental stage, and need a lot more work <strong>to</strong><br />

make them truly interoperable and user-centered.<br />

The conceptual steps typically include first trying <strong>to</strong> replicate core activities that<br />

functioned in the analogue environment, then attempting <strong>to</strong> replicate some (but not all) of<br />

the non-core analogue functions, then (after using the systems for some time) discovering<br />

and implementing new functions that did not exist within the previous analogue<br />

environment. Only with this final step do we actually realize the new functional<br />

environment enabled by the new technology. So, for example, word processors were<br />

initially built as typewriters with s<strong>to</strong>rage mechanisms, but over time grew <strong>to</strong> incorporate<br />

functions such as spell-checking and revision-tracking, and eventually enabled very<br />

different functions (such as desk<strong>to</strong>p publishing). Our early efforts at creating MARC<br />

records began as ways <strong>to</strong> au<strong>to</strong>mate the production of catalogue cards, then moved <strong>to</strong> the<br />

creation of bibliographic utilities and their union catalogues, then <strong>to</strong> OPACs.<br />

Functionally, our OPACs began as mere replicas of card catalogues, then added Boolean<br />

searching, then title-word searching capabilities, and now are poised <strong>to</strong> allow users <strong>to</strong><br />

propagate distributed searches across a series of OPACs. Today's <strong>digital</strong> collections are<br />

not much past the initial stage where we are replicating the collections of content and<br />

cataloguing that existed in analogue form, and just beginning <strong>to</strong> add minor functions. In<br />

the future we can expect our <strong>digital</strong> libraries <strong>to</strong> incorporate a variety of functions that<br />

employ the new technological environments in ways we can hardly imagine <strong>to</strong>day.<br />

The Importance of Standards<br />

In moving from dispersed <strong>digital</strong> collections <strong>to</strong> interoperable <strong>digital</strong> libraries, the most<br />

important activity developers need <strong>to</strong> focus on is standards. This includes standards and<br />

pro<strong>to</strong>cols for open archives and metadata harvesting. But most important is the wide<br />

variety of metadata standards needed. The most extensive metadata activities have<br />

focused on discovery metadata (such as the Dublin Core), but metadata also include a<br />

wide variety of other functions: structural metadata are used for turning the pages of a


<strong>digital</strong> book, administrative metadata are used <strong>to</strong> ensure that all the individual pages of a<br />

<strong>digital</strong> book are kept <strong>to</strong>gether over time, computer-based image retrieval systems employ<br />

metadata <strong>to</strong> help users search for similar colors, shapes, and textures, etc. Developers<br />

need <strong>to</strong> widely employ descriptive metadata for consistent description, discovery<br />

metadata for finding works, administrative metadata for viewing and maintaining works,<br />

structural metadata for navigation through an individual work, identification metadata <strong>to</strong><br />

determine that one has accessed the proper version of a work, and terms and conditions<br />

metadata for compliance with use constraints.<br />

Having consensus over metadata and other standards is important for a variety of reasons.<br />

Administrative and longevity metadata are needed <strong>to</strong> manage <strong>digital</strong> files over time, <strong>to</strong><br />

make sure all the necessary files are kept <strong>to</strong>gether, and <strong>to</strong> help view these files when<br />

<strong>to</strong>day's application software becomes unusable. Because of the mutability of <strong>digital</strong><br />

works, developers need standards <strong>to</strong> ensure the veracity of a work, and <strong>to</strong> help assure<br />

users that a particular work has not been altered, and is indeed the version of the work<br />

that it purports <strong>to</strong> be. And developers need a variety of types of metadata and standards <strong>to</strong><br />

allow various <strong>digital</strong> collections <strong>to</strong> interoperate, and <strong>to</strong> help users feel that they can<br />

search across groups of collections. One side benefit of reaching consensus over metadata<br />

that will be recorded in a consistent manner is that vendors will have an economic<br />

incentive <strong>to</strong> re-<strong>to</strong>ol applications <strong>to</strong> incorporate this metadata (because they can spread<br />

their costs over a wide variety of institutions who will want <strong>to</strong> employ these standards).<br />

The various metadata types<br />

Libraries have had agreements on metadata standards for many decades. The Anglo-<br />

American Cataloguing Rules defined a set of descriptive metadata for bibliographic (and<br />

later, other) works, and the MARC format gave us a syntax for transporting those<br />

bibliographic records. Likewise, Library of Congress Subject Headings and Sears Subject<br />

Headings have for many years provided us with discovery metadata <strong>to</strong> help users find<br />

relevant material. In the last quarter of the twentieth century, other types of discovery<br />

metadata emerged <strong>to</strong> serve specialized fields, including the Art and Architecture<br />

Thesaurus (AAT) and the Medical Subject Headings (MeSH). Though both AAT and<br />

MeSH envisioned use in an online environment, both were developed in an era when<br />

indexing and cataloguing records might sit on a computer, but the works they referred <strong>to</strong><br />

would not. And both were developed at a point in time when the merging of records from<br />

these specialized fields with records for more general works was unlikely <strong>to</strong> take place on<br />

a widespread basis.<br />

The rapid acceptance of the World Wide Web led a number of us <strong>to</strong> consider how one<br />

might allow users <strong>to</strong> search across a variety of online records and resources, particularly<br />

when some of those resources received extensive cataloguing, while others received little<br />

or none. This led <strong>to</strong> the March 1995 meeting that defined the Dublin Core as a type of<br />

discovery metadata that would allow users <strong>to</strong> search across a wide variety of resources<br />

including both highly catalogued (often legacy) material, and material (much of it new<br />

and in electronic form) that was assigned only a minimal amount of metadata. We<br />

envisioned the Dublin Core as a kind of unifying set of metadata that would permit


discovery across all types of <strong>digital</strong> records and resources. Library cataloguing records or<br />

museum collection management records could be "dumbed down" <strong>to</strong> look like Dublin<br />

Core (DC) records, while DC records for resources like an individual's research paper<br />

might either be easy enough for the individual <strong>to</strong> create, or might even be au<strong>to</strong>matically<br />

generated by the individual's word processor. The not-yet-realized promise of the Dublin<br />

Core (see next section on Harvesting) was <strong>to</strong> provide discovery-level interoperability<br />

across all types of online indexes and resources, from the highly-catalogued OPACs <strong>to</strong><br />

the websites and webpages of individuals and organizations. And because the Dublin<br />

Core has been in existence for approximately seven years (and has recently been<br />

designated as NISO Standard Z39–85), it is more developed and better known than any<br />

of the other types of metadata created for electronic resources.<br />

Though the Dublin Core was developed as a form of <strong>digital</strong> metadata <strong>to</strong> be applied <strong>to</strong><br />

works in both <strong>digital</strong> and non-<strong>digital</strong> form, a variety of other metadata types have more<br />

recently been developed specifically for collections of works in <strong>digital</strong> form. Below we<br />

will briefly discuss efforts <strong>to</strong> define structural metadata, administrative metadata,<br />

identification metadata (particularly for images), and longevity metadata. All these<br />

metadata types are critical for moving from a set of independent <strong>digital</strong> collections <strong>to</strong> real<br />

interoperable <strong>digital</strong> libraries. Hence they all incorporate functions likely <strong>to</strong> lead either <strong>to</strong><br />

increased interoperability, or <strong>to</strong> the fuller and more robust services that characterize a<br />

library rather than a collection.<br />

Structural metadata recognize that, for many works in <strong>digital</strong> form, it is not enough<br />

merely <strong>to</strong> display the work; users may need <strong>to</strong> navigate through the work. Structural<br />

metadata recognize that users expect certain "behaviors" from a work. For example,<br />

imagine a book that is composed of hundreds of <strong>digital</strong> files, each one the scan of a single<br />

book page. Structural metadata are needed for users <strong>to</strong> perform the normal behaviors they<br />

might expect from a book. Users will expect <strong>to</strong> be able <strong>to</strong> view the table of contents, then<br />

jump <strong>to</strong> a particular chapter. As they read through that chapter, they will expect <strong>to</strong> turn<br />

the page, and occasionally go back <strong>to</strong> re-read the previous page. When they come <strong>to</strong> a<br />

citation, they will want <strong>to</strong> jump <strong>to</strong> the bibliography <strong>to</strong> read the citation, then jump back.<br />

And when they come <strong>to</strong> a footnote marker, they may want <strong>to</strong> jump <strong>to</strong> where they can read<br />

the footnote contents, then jump back. These are all just normal behaviors we expect<br />

from any type of book, but these behaviors all require structural metadata. Without them,<br />

the book would just be a series of individual scanned pages, and users would have a great<br />

deal of difficulty trying <strong>to</strong> even put the pages in the correct order, let alone read the book.<br />

Structural metadata have a role in any kind of material that would benefit from internal<br />

navigation (including diaries and journals).<br />

Administrative metadata maintain the information necessary in order <strong>to</strong> keep a <strong>digital</strong><br />

work accessible over time. In the case of a digitized book, the administrative metadata<br />

would note all the individual files needed <strong>to</strong> assemble the book, where the files were<br />

located, and what file formats and applications software would be necessary in order <strong>to</strong><br />

view the book or its individual pages. Administrative metadata become particularly<br />

important when moving files <strong>to</strong> a new server, or engaging in activities related <strong>to</strong> <strong>digital</strong><br />

longevity such as refreshing or migration.


Instead of employing open standards for structural and administrative metadata, many<br />

individuals and organizations choose <strong>to</strong> encode their documents within commercial<br />

products such as Adobe Acrobat. Although this is highly convenient (particularly given<br />

the proliferation of Acrobat readers), it could be a dangerous practice for libraries and<br />

similar reposi<strong>to</strong>ries. Commercial products are proprietary, and focus on immediate<br />

convenience rather than long-term access. Hence, there is no guarantee of continued<br />

compatibility or future access <strong>to</strong> works encoded in earlier versions of commercial<br />

software. In order <strong>to</strong> cope with long-term preservation and access issues, as well as <strong>to</strong><br />

provide a higher level of structural functionality, in 1997 a group of US libraries began<br />

the Making of America II Project <strong>to</strong> define structural and administrative metadata<br />

standards for library special collection material (Hurley et al. 1999). These standards<br />

were further refined within the Technology Architecture and Standards Committee of the<br />

California Digital Library (CDL 200la), and have since been renamed the Metadata<br />

Encoding and Transmission Standards (METS). METS was adopted by the US Digital<br />

Library Federation and is now maintained by the US Library of Congress<br />

(). The Research Libraries Group recently<br />

announced plans <strong>to</strong> lead the METS development effort.<br />

Identification metadata attempt <strong>to</strong> address the proliferation of different versions and<br />

editions of <strong>digital</strong> works. In the print world, the publishing cycle usually both enforced an<br />

edi<strong>to</strong>rial process and created time lags between the issuance of variant works. But for a<br />

highly mutable <strong>digital</strong> work, those processes and time lags are often eliminated, and<br />

variants of the work are created quickly and with very little thought about their impact on<br />

what we used <strong>to</strong> call bibliographic control. In addition, the networked <strong>digital</strong> environment<br />

itself leads information distribu<strong>to</strong>rs <strong>to</strong> provide a variety of different forms of a given<br />

work (HTML, Postscript, Acrobat, XML, and Microsoft Word forms of documents <strong>to</strong><br />

support different user capabilities and needs; thumbnail, medium-sized, and large images<br />

<strong>to</strong> support image browsing, viewing, and study).<br />

To illustrate the identification metadata problem, let us turn <strong>to</strong> Figure 36.3, an illustration<br />

of variant forms of images. The original object (a sheep) is shown in the upper left<br />

corner. There are four pho<strong>to</strong>graphs of the original object, taken from three different<br />

angles. Two of the pho<strong>to</strong>graphs (A and D) are taken from the same angle, but pho<strong>to</strong>graph<br />

D has captured a fly on the side of the sheep. The images <strong>to</strong> the right of pho<strong>to</strong>graph D are<br />

variant forms of that pho<strong>to</strong>graph after image processing was done <strong>to</strong> remove the fly from<br />

the side of the sheep, while the images below pho<strong>to</strong>graph D show variant forms that<br />

include the fly. Spread throughout the figure are variant forms including different-sized<br />

images (thumbnail <strong>to</strong> high-resolution), different compression ratios (including both lossy<br />

and lossless), and different file encoding formats (PICT, TIFF, JFIF). Certain users may<br />

only want particular resolutions or compression ratios (e.g., a serious researcher may<br />

require an uncompressed high-resolution image). All the images illustrated in this figure<br />

share a base set of metadata that refer <strong>to</strong> the initial object of origin (the sheep), and<br />

adapting Leazer's idea of Bibliographic Families (Leazer and Smiraglia 1999), we can<br />

say that all the images in this illustration form an "Image Family" which shares a<br />

common set of metadata. Each image instantiation in the family also inherits metadata<br />

from its parents, and knowledge about that inheritance can often be critical <strong>to</strong> someone


viewing a particular instantiation (for example, someone using one of the lower-right<br />

corner instantiations <strong>to</strong> study wool characteristics should be able <strong>to</strong> ascertain that image<br />

processing was done on a parent or grandparent of this image [<strong>to</strong> remove the fly], and<br />

that this image processing might affect the matting of the wool). Thus, it is critical that<br />

any image inherits important metadata from its lineage, or that systems at least provide<br />

ways that a researcher can trace upwards in lineage <strong>to</strong> discover metadata that might affect<br />

their use of the work. A first step in this direction is in the US National Information<br />

Standards Organization efforts at creating a Technical Imaging Metadata Standard that<br />

incorporates "change his<strong>to</strong>ry" and "source data" (NISO 2002). But our community has<br />

much more work <strong>to</strong> do in order <strong>to</strong> provide the type of identification of variant forms that<br />

users (particularly researchers) have come <strong>to</strong> expect from libraries of analogue materials.<br />

And we still need <strong>to</strong> come <strong>to</strong> grips with the problem of how <strong>to</strong> preserve dynamic<br />

documents – documents that are essentially alive, and changing on a daily basis.<br />

As <strong>digital</strong> library developers construct large collections of material in <strong>digital</strong> form, we<br />

need <strong>to</strong> consider how <strong>digital</strong> works will provoke changes in long-standing practices that<br />

have grown up around analogue works. Elsewhere this author has outlined how electronic<br />

art will likely provoke changes in conservation and preservation practices (Besser 200la)<br />

and how the growing body of moving image material in <strong>digital</strong> form is beginning <strong>to</strong><br />

reshape the role of film archives and archivists (Besser 200Ib). But at a very pragmatic<br />

level, all reposi<strong>to</strong>ries of <strong>digital</strong> works need <strong>to</strong> worry about the persistence of those works<br />

over time. Longevity metadata are necessary in order <strong>to</strong> keep <strong>digital</strong> material over long<br />

periods of time. While saving bits may be fairly straightforward, saving <strong>digital</strong> works is<br />

not. Digital works are very fragile, and pro-active steps need <strong>to</strong> be taken in order <strong>to</strong> make<br />

sure that they persist over time (for more on this subject, see the Digital Longevity<br />

website maintained by this author at ).<br />

Elsewhere, I have outlined five key fac<strong>to</strong>rs that pose <strong>digital</strong> longevity challenges (the<br />

Viewing Problem, the Scrambling Problem, the Interrelation Problem, the Cus<strong>to</strong>dial<br />

Problem, and the Translation Problem), and have suggested that community consensus<br />

over metadata can be a key fac<strong>to</strong>r in helping <strong>digital</strong> works persist over time (Besser<br />

2000). Recently, the major US bibliographic utilities have begun serious efforts <strong>to</strong> reach<br />

consensus on preservation metadata for <strong>digital</strong> works (OCLC/RLG <strong>2001</strong>a, <strong>2001</strong>b).<br />

Widespread adoption of this type of standard will make the challenge of <strong>digital</strong><br />

persistence much more tractable. And late in <strong>2001</strong> the Library of Congress, with the help<br />

of the Council on Library and Information Resources, began a planning process for a<br />

National Digital Information Infrastructure and Preservation Program (see<br />

). Widespread support for the emerging<br />

metadata standards mentioned here will greatly improve interoperability between<br />

collections and the sustainability of those collections over time. This will help us move<br />

away from isolated experiments with <strong>digital</strong> collections <strong>to</strong>wards sustainable <strong>digital</strong><br />

libraries.


36.3 Figure Image families<br />

Metadata philosophies and harvesting: Warwick vs. MARC<br />

Though agreement on a variety of metadata standards is a necessary prerequisite for<br />

interoperable <strong>digital</strong> collections, implementation of interoperability also requires a set of<br />

architectures and a common approach <strong>to</strong> making that metadata available <strong>to</strong> other<br />

collections, middleware, and end users. In this section we will discuss two philosophical<br />

approaches <strong>to</strong> metadata, as well as methods for sharing metadata with applications and<br />

individuals outside the designer's home collection.<br />

Libraries have traditionally employed the MARC/AACR2 philosophical approach <strong>to</strong><br />

metadata. This approach employs a single overarching schema <strong>to</strong> cover all types of works


and all groups of users. As new types of works arise, new fields are added <strong>to</strong> the MARC/<br />

AACR2 framework, or rules for existing fields are changed <strong>to</strong> accommodate these new<br />

works. And as communities emerge with new metadata needs, these are also incorporated<br />

in<strong>to</strong> the existing schema. The MARC/AACR2 philosophy maintains that one big schema<br />

should serve all user needs for all types of works. Critics of this approach point out that<br />

the schema has become so overly complex that only highly trained specialists (library<br />

cataloguers) are able <strong>to</strong> assign metadata using it, and that the system is <strong>to</strong>o slow <strong>to</strong> adapt<br />

<strong>to</strong> emerging types of works. They also claim that groups of users often have sets of<br />

metadata needs that the controllers of MARC/AACR2 are unwilling <strong>to</strong> accommodate.<br />

In recent years, a rival philosophy has emerged from within the Dublin Core community.<br />

This philosophy, based upon the Warwick Framework, relies upon interlocking<br />

containers and packages of metadata, each maintained by a particular community.<br />

According <strong>to</strong> this philosophy, each community can support the packages of metadata it<br />

needs for its own particular uses, while still interoperating with the metadata packages<br />

from other communities. Under this philosophy, the Dublin Core serves as a unifying set<br />

of metadata <strong>to</strong> allow discovery across all communities. And even within the Dublin Core<br />

(DC), certain communities can employ qualifiers that meet their own detailed needs,<br />

while still providing useful metadata <strong>to</strong> other communities. (For example, the library<br />

community could use qualifiers <strong>to</strong> reflect the nuances of differences between main title,<br />

alternate title, transliterated title, and translated title, while other communities could find<br />

any of these as part of a search under unqualified title.) This philosophy supports<br />

metadata packages that are modular, overlapping, extensible, and community-based.<br />

Advocates believe that they will aid commonality between communities while still<br />

providing full functionality within each community. This approach is designed for a<br />

networked set of communities <strong>to</strong> interrelate <strong>to</strong> one another.<br />

No matter which philosophical approach one follows, any collection faces the pragmatic<br />

issue of how <strong>to</strong> make their metadata available <strong>to</strong> other collections and external searching<br />

software. The traditional library model was <strong>to</strong> export MARC records <strong>to</strong> a bibliographic<br />

utility (like OCLC or RLIN) and <strong>to</strong> have all external users search through that utility.<br />

While this works fine for MARC-based records, increasingly users want <strong>to</strong> search across<br />

a much wider base of information from a world not circumscribed by bibliographic<br />

records (such as web pages and sites, PDF documents, images, databases, etc.).<br />

Therefore, most <strong>digital</strong> collections are beginning <strong>to</strong> consider how <strong>to</strong> export reduced<br />

records in<strong>to</strong> a space where they can be picked up by Internet search engines. For records<br />

following the MARC/AACR2 approach, this means extracting simple records (probably<br />

in DC format) from complex MARC records, and exporting these. For both the Warwick<br />

and the MARC/AACR2 approaches, this means developing methods for metadata<br />

harvesting that allow the appropriate exported records <strong>to</strong> be found by Internet search<br />

engines. A number of projects are currently under way <strong>to</strong> test metadata harvesting.<br />

Best practices<br />

Along with standards and architectures, community agreement on best practices is<br />

another important ingredient in helping make collections of <strong>digital</strong> materials more


interoperable and sustainable. Best practices ensure that content and metadata from<br />

different collections will meet minimum standards for preservation purposes, and that<br />

users can expect a baseline quality level.<br />

A key best practice principle is that any <strong>digital</strong> project needs <strong>to</strong> consider users, potential<br />

users, uses, and actual characteristics of the collections (Besser and Trant 1995). This<br />

means that decision-making both on <strong>digital</strong> conversion of analogue material and on<br />

metadata assignment needs <strong>to</strong> be carefully planned at the start of a <strong>digital</strong> project. The<br />

pioneering best practices for <strong>digital</strong> conversion developed by the Technology<br />

Architecture and Standards Committee of the California Digital Library (CDL <strong>2001</strong>b)<br />

introduced important concepts designed <strong>to</strong> aid in the longevity and sustainability of<br />

<strong>digital</strong> collections. These concepts included differentiating between masters and<br />

derivatives, inclusion of greyscale targets and rulers in the scan, using objective<br />

measurements <strong>to</strong> determine scanner settings (rather than matching the image on a nearby<br />

moni<strong>to</strong>r <strong>to</strong> the original object), s<strong>to</strong>ring in common formats, and avoiding compression<br />

(particularly lossy compression). This document also suggested that collections strive <strong>to</strong><br />

capture as much metadata as is reasonably possible (including metadata about the<br />

scanning process itself). The relationship of best practices for scanning <strong>to</strong> the longevity of<br />

<strong>digital</strong> works is more fully explained in the Digital Library Federation's draft benchmarks<br />

for <strong>digital</strong> reproductions (DLF <strong>2001</strong>).<br />

The Making of America II Project (Hurley et al. 1999) introduced the idea that metadata<br />

could begin in fairly raw form, and, over time, move <strong>to</strong>ward being seared and eventually<br />

cooked. This notion of incrementally upgrading metadata appears <strong>to</strong> have assuaged the<br />

fears of some groups that metadata schemes like METS were <strong>to</strong>o overblown and<br />

complicated for them <strong>to</strong> undertake. In effect, this notion appears <strong>to</strong> have increased the<br />

adoption level of more complicated metadata schemes.<br />

Other standards issues<br />

A number of other standards issues need <strong>to</strong> be addressed in order <strong>to</strong> bring interoperability<br />

and other conventional library services <strong>to</strong> our emerging <strong>digital</strong> libraries. These include<br />

open archives, metadata harvesting, persistent identification, helping users find the<br />

appropriate copy, and user authentication.<br />

Metadata s<strong>to</strong>red in local systems are often not viewable by external applications or users<br />

that may be trying <strong>to</strong> discover local resources. This seriously inhibits interoperability, and<br />

the ability of a user <strong>to</strong> search multiple collections. The Open Archives Initiative<br />

() is tackling this problem by developing and testing<br />

interoperability pro<strong>to</strong>cols that will allow applications <strong>to</strong> harvest metadata, even those<br />

residing in deep archives. The success of a project like this is critical <strong>to</strong> providing users<br />

with the type of access outlined in Figure 36.2. In addition, this project's focus on e-print<br />

archives should provide users with a diverse body of content free of onerous constraints.<br />

Persistent naming is still an important issue for building real <strong>digital</strong> libraries. Though the<br />

World Wide Web has brought us increased access <strong>to</strong> works, Web architecture has


violated traditional library practices of providing relative location information for a work<br />

by instead providing a precise location address. Failures in the Web's precise locationaddressing<br />

system (the URL) produce the most common error message that Web users<br />

encounter: 404 – File Not Found. Most of these error messages result from normal<br />

maintenance of a website (renaming higher-order folders or direc<strong>to</strong>ries, re-organizing file<br />

locations). Librarians would never consider telling a user that <strong>to</strong> find the book they're<br />

seeking they must go <strong>to</strong> the third tier of the stacks, in the eighth row, the fifth bookcase,<br />

the third shelf, and grab the seventh book from the left; they know that once someone<br />

removes the third book from the left, the entire system of locating will break down. Yet,<br />

this is the type of system that URLs are based upon. In recent years there has been much<br />

work done on indirect naming (in the form of PURLS, URNs, and handles). But <strong>to</strong><br />

replicate the power that libraries have developed, we need truly persistent naming. This<br />

means more than just the indication of a location for a particular work. Sophisticated<br />

persistent naming would include the ability <strong>to</strong> designate a work by its name, and <strong>to</strong><br />

distinguish between various instantiations of that work and their physical locations. Just<br />

as conventional libraries are able <strong>to</strong> handle versions and editions and direct users <strong>to</strong><br />

particular copies of these, our <strong>digital</strong> libraries will need <strong>to</strong> use identification metadata <strong>to</strong><br />

direct users <strong>to</strong> an appropriate instantiation of the work they are seeking.<br />

Ever since the advent of indexing and abstracting services, conventional libraries have<br />

had <strong>to</strong> face the problem of answering a user's query with a list of works, some of which<br />

may not be readily available. Conventional libraries have striven <strong>to</strong> educate users that<br />

some sources are not physically present in the library, and they have also developed both<br />

interlibrary loan and document delivery services <strong>to</strong> help get material <strong>to</strong> users in a timely<br />

fashion. But the recent proliferation of licensed full-text electronic resources has greatly<br />

complicated this problem. For a variety of reasons, it may be very difficult <strong>to</strong> match a<br />

user with an appropriate document they are licensed <strong>to</strong> use: certain users may be covered<br />

by a given license while others are not; the same document may be provided by several<br />

aggrega<strong>to</strong>rs under different licenses; much of the online content is physically s<strong>to</strong>red by<br />

the licensor (content provider) rather than by the licensee (library). Recent<br />

standardization efforts have begun <strong>to</strong> address part of this problem; the US National<br />

Information Standards Organization has formed the OpenURL Standard Committee AX<br />

() <strong>to</strong> allow a query <strong>to</strong> carry context-sensitive<br />

information. This will help a library authenticate their licensees <strong>to</strong> a remote content site.<br />

But there are still many more problems <strong>to</strong> solve in getting users appropriate copies<br />

(particularly when content licenses are complex and overlapping).<br />

Still another critically important standards and pro<strong>to</strong>cols area that is under development is<br />

that of authentication of users. With more and more licensed content being physically<br />

s<strong>to</strong>red by content providers, those providers want assurances that users accessing the<br />

content are indeed covered by valid licenses <strong>to</strong> do so. Yet conventional methods of user<br />

authentication (such as password or IP addressing) would allow content providers <strong>to</strong> track<br />

what an individual reads, and develop complex profiles of user habits. Legal scholars<br />

have warned of the dangers this poses (Cohen 1996), and it flies in the face of the<br />

important library ethical tradition of privacy. Work has begun on a project that lets an<br />

institution authenticate users <strong>to</strong> a resource provider without revealing individual


identities. It still remains <strong>to</strong> be seen whether the Shibboleth project<br />

() will be acceptable <strong>to</strong> resource providers,<br />

yet still provide the privacy and anonymity that libraries have traditionally assured users.<br />

Success becomes more questionable in the wake of the September 11, <strong>2001</strong>, US building<br />

destructions, as the US federal government has increased the pressure <strong>to</strong> eliminate<br />

anonymous library access (ALA et al. <strong>2001</strong>).<br />

The Next Stage: Moving from Isolated Digital<br />

Collections <strong>to</strong> Interoperable Digital Libraries<br />

Conventional libraries have both functional components and ethical traditions. The <strong>digital</strong><br />

collections currently under construction will not truly be "<strong>digital</strong> libraries" until they<br />

incorporate a significant number of the components of conventional libraries, and adhere<br />

<strong>to</strong> many of the important ethical traditions and values of libraries. And though our <strong>digital</strong><br />

collections have made significant progress in these areas in the past seven years, they still<br />

have a long way <strong>to</strong> go.<br />

For the component of interoperability, projects such as open archives, metadata<br />

harvesting, and structural and administrative metadata hold great promise. For the<br />

component of stewardship over collections, <strong>digital</strong> preservation projects have finally<br />

begun, but it will be some time before we will be able <strong>to</strong> say with confidence that we can<br />

preserve the portion of our culture that is in <strong>digital</strong> form. Additionally, developers have<br />

only recently begun <strong>to</strong> grapple with the issue of economic sustainability for <strong>digital</strong><br />

libraries (CLIR <strong>2001</strong>). For other components such as service <strong>to</strong> clientele, they have<br />

barely scratched the surface in one important area where conventional libraries perform<br />

very well – delivering information <strong>to</strong> different groups of users (by age level, knowledge<br />

base, particular need, etc.) in ways appropriate <strong>to</strong> each group. The California Digital<br />

Library and the UCLA/ Pacific Bell Initiative for 21st Century Literacies have recently<br />

completed a project <strong>to</strong> explore this problem<br />

().<br />

Those constructing <strong>digital</strong> collections have spent less energy trying <strong>to</strong> build library<br />

ethical traditions in<strong>to</strong> our systems, and in many cases have relied on those outside the<br />

<strong>digital</strong> library community (such as the American Library Association filing lawsuits on<br />

privacy and free speech, or the Internet Engineering Task Force developing pro<strong>to</strong>cols <strong>to</strong><br />

preserve privacy) <strong>to</strong> work on upholding library ethical traditions such as free speech,<br />

privacy, and equal access. But as Lawrence Lessig has made clear, the choices we make<br />

in the architecture and design of our systems will limit the social choices we can make<br />

around use of those systems in the future (Lessig 1999). For example, some of our online<br />

public-access circulation systems purposely saved only aggregate user data so that no one<br />

in the future could attempt <strong>to</strong> track individual reading habits. While recent projects such<br />

as Shibboleth are trying <strong>to</strong> design library ethical traditions in<strong>to</strong> the technological<br />

infrastructure, the <strong>digital</strong> libraries we're building still have not addressed many of our<br />

important library ethical traditions.


As designers build <strong>digital</strong> collections they will also need <strong>to</strong> uphold library ethical<br />

traditions of equal access and diversity of information. Both of these are threatened by<br />

the commercialization of intellectual property. As we see the increased commodification<br />

of information and consolidation of the content industry in<strong>to</strong> fewer and fewer hands, less<br />

and less creative work enters the public domain and more requires payment <strong>to</strong> view<br />

(Besser 2002a). Commodification and consolidation also bring with them a concentration<br />

on "best-seller" works and a limiting of diversity of works (Besser 1995, 1998). The<br />

builders of <strong>digital</strong> collections need <strong>to</strong> go beyond content that is popular and become<br />

aggressive about collecting content that reflects wide diversity; instead of using the nowfamiliar<br />

opportunistic approach <strong>to</strong> converting content <strong>to</strong> <strong>digital</strong> form, they will need <strong>to</strong><br />

develop carefully planned <strong>digital</strong> collection development policies. Developers will find<br />

that they need <strong>to</strong> closely collaborate with others both <strong>to</strong> leverage resources, and <strong>to</strong> ensure<br />

that the efforts of different organizations <strong>to</strong>gether look like virtual <strong>digital</strong> libraries <strong>to</strong> our<br />

users. Developers and scholars also need <strong>to</strong> be involved in these efforts, <strong>to</strong> ensure that a<br />

broad range of content eventually leaves the marketplace and enters the public domain.<br />

We need <strong>to</strong> involve ourselves in struggles such as the American Library Association's<br />

current effort <strong>to</strong> build a coalition <strong>to</strong> protect the "information commons" in cyberspace.<br />

Digital library developers also will find that they need <strong>to</strong> maintain the role of the library<br />

as guardian over individuals' rights <strong>to</strong> access a rich variety of information, and <strong>to</strong> see that<br />

information within its context. They will need <strong>to</strong> continue <strong>to</strong> be vigilant about making<br />

sure that other forms of "equal access <strong>to</strong> information" extend <strong>to</strong> the new <strong>digital</strong> world.<br />

They would also do well <strong>to</strong> extend the "library bill of rights" in<strong>to</strong> cyberspace, and they<br />

will find themselves having <strong>to</strong> struggle <strong>to</strong> keep the <strong>digital</strong> world from increasing the<br />

distance between "haves" and "have-nots."<br />

Finally, in the move <strong>to</strong>wards constructing <strong>digital</strong> libraries, we need <strong>to</strong> remember that<br />

libraries are not merely collections of works. They have both services and ethical<br />

traditions and values that are a critical part of their functions. Libraries interoperate with<br />

each other <strong>to</strong> serve the information needs of a variety of different user groups <strong>to</strong>day, and<br />

expect <strong>to</strong> sustain themselves and their collections so that they can serve users 100 years<br />

from now. They defend their users' rights <strong>to</strong> access content, and <strong>to</strong> do so with some<br />

degree of privacy or anonymity. The <strong>digital</strong> collections being built will not truly be<br />

<strong>digital</strong> libraries until they incorporate a significant number of these services and ethical<br />

traditions.<br />

Acknowledgments<br />

Extensive portions of this chapter were previously published in the online journal First<br />

Monday as "The Next Stage: Moving from Isolated Digital Collections <strong>to</strong> Interoperable<br />

Digital Libraries" (Besser 2002c) and in the Vic<strong>to</strong>rian Association for Library<br />

Au<strong>to</strong>mation's E-Volving Information Futures 2002 conference proceedings (Besser<br />

2002a). This chapter also involves a synthesis of a number of different talks and<br />

workshops the author delivered between 1996 and 2002, and he wishes <strong>to</strong> thank the many<br />

individuals who offered critical comments or encouragement after those presentations.


References<br />

American Library Association (1995). American Library Association Code of Ethics<br />

adopted by the ALA Council on June 28, 1995. At<br />

http://www.ala.org/alaorg/oif/ethics.html.<br />

American Library Association et al. (ALA) (<strong>2001</strong>). Library Community Statement on<br />

Proposed Anti-Terrorism Measures (Oc<strong>to</strong>ber 2). At http://www.ala.org/washoff/.<br />

Besser, Howard (1995). From Internet <strong>to</strong> Information Superhighway. In James Brook and<br />

Iain A. Boal (eds.), Resisting the Virtual Life: The Culture and Politics of Information<br />

(pp. 59–70). San Francisco: City Lights.<br />

Besser, Howard (1997). The Changing Role of Pho<strong>to</strong>graphic Collections with the Advent<br />

of Digitization. In Katherine Jones-Garmil (ed.), The Wired Museum (pp. 115–27).<br />

Washing<strong>to</strong>n: American Association of Museums.<br />

Besser, Howard (1998). The Shape of the 21st Century Library. In Mil<strong>to</strong>n Wolf et al.<br />

(eds.), Information Imagineering: Meeting at the Interface (pp. 133–46). Chicago:<br />

American Library Association.<br />

Besser, Howard (2000). Digital Longevity. In Maxine K. Sitts (ed.), Handbook for Digital<br />

Projects: A Management Tool for Preservation and Access (pp. 155–66). Andover, MA:<br />

Northeast Document Conservation Center.<br />

Besser, Howard (<strong>2001</strong>a). Longevity of Electronic Art. In David Bearman and Franca<br />

Garzot<strong>to</strong> (eds.), ICHIM 01 International Cultural Heritage Informatics Meeting: Cultural<br />

Heritage and Technologies in the Third Millennium, vol. 1: Full Papers. Proceedings of<br />

the September 3–7, <strong>2001</strong>, Milan Meeting (pp. 263–75). Milan: Politecnico di Milano.<br />

Besser, Howard (<strong>2001</strong>b). Digital Preservation of Moving Image Material. The Moving<br />

Image (Fall): 39–55.<br />

Besser, Howard (2002a). Moving from Isolated Digital Collections <strong>to</strong> Interoperable<br />

Digital Libraries. In Vic<strong>to</strong>rian Association for Library Au<strong>to</strong>mation's E-Volving<br />

Information Futures. Proceedings of the 11 Biennial Conference, 6–8 February, 2002,<br />

Melbourne, vol. 2 (pp. 707–25).<br />

Besser, Howard (2002b). Commodification of Culture Harms Crea<strong>to</strong>rs. Information<br />

Commons website by Ameri can Library Association. At http://www.infocommons.org.<br />

http:///gseis.ucla.edu/~howard/Copyright/ala-commons.html.<br />

Besser, Howard (2002c). The Next Stage: Moving from Isolated Digital Collections <strong>to</strong><br />

Interoperable Digital Libraries. First Monday 7, 6 (June). At<br />

http://www.firstmonday.org/issues/issue7_6/besser/.


Besser, Howard (website). Digital Longevity. At http://sunsite.berkeley.edu/Longevity/.<br />

Besser, Howard and Jennifer Trant (1995). Introduction <strong>to</strong> Imaging: Issues in<br />

Constructing an Image Database. Santa Monica: Getty Art His<strong>to</strong>ry Information Program.<br />

Bishop, Ann and Susan Leigh Star (1996). Social Informatics for Digital Library Use and<br />

Infrastructure. In Martha Williams (ed.), Annual Review of Information Science and<br />

Technology 31 (pp. 301–401). Medford, NJ: Information Today.<br />

Borgman, Christine (1997). Now that We Have Digital Collections, Why Do We Need<br />

Libraries? In Candy Schwartz and Mark Rorvig (eds.), ASIS '97: Proceedings of the 60th<br />

ASIS Annual Meeting, vol. 34. Medford, NJ: Information Today.<br />

Borgman, Christine, et al. (1996). Social Aspects of Digital Libraries, Final Report <strong>to</strong> the<br />

National Science Foundation. Digital Libraries Initiative. At<br />

http://dli.grainger.uiuc.edu/national.htm.<br />

California Digital Library (CDL) (<strong>2001</strong>a). Technology Architecture and Standards<br />

Committee (May 18), California Digital Library Digital Object Standard: Metadata,<br />

Content and Encoding. California Digital Library. At<br />

http://www.cdlib.org/about/publications/.<br />

California Digital Library (<strong>2001</strong>b). Best Practices for Image Capture. California Digital<br />

Library (February). At http://www.cdlib.org/about/publications/.<br />

CLIR (<strong>2001</strong>). Building and Sustaining Digital Collections: Models for Libraries and<br />

Museums (August). Council on Library and Information Resources (CLIR). At<br />

http://www.clir.org/pubs/reports/publ00/publ00.<strong>pdf</strong>.<br />

Cohen, Julie (1996). A Right <strong>to</strong> Read Anonymously: A Closer Look at "Copyright<br />

Management" in Cyberspace. Connecticut Law Review 28: 981–1039.<br />

Digital Library Federation (DLF) (<strong>2001</strong>). Draft Benchmark for Digital Reproductions of<br />

Printed Books and Serial Publications (July). At<br />

http://www.diglib.org/standards/draftbmark.htm.<br />

Digital Preservation and Archive Committee (DPAC) (<strong>2001</strong>). Draft Final Report<br />

(Oc<strong>to</strong>ber 8).<br />

Griffin, Stephen M. (1999). Digital Libraries Initiative – Phase 2: Fiscal Year 1999<br />

Awards, D-Lib Magazine 5, 7–8 (July/August). At<br />

http://www.dlib.org/dlib/july99/07griffin.html.<br />

Hirtle, Peter (1999). A New Generation of Digital Library Research. Edi<strong>to</strong>rial. D-Lib<br />

Magazine 5, 7–8 (July/August). At http://www.dlib.org/dlib/july99/07edi<strong>to</strong>rial.html.


Hurley, Bernard, John Price-Wilkin, Merrillee Proffitt, and Howard Besser (1999). The<br />

Making of America II Testbed Project: A Digital Library Service Model. Council on<br />

Library and Information Resources. Digital Library Federation (December). At<br />

http://sunsite.berkeley.edu/moa2/wp-v2.html.<br />

Leazer, Gregory and Richard Smiraglia (1999). Bibliographic Families in the Library<br />

Catalogue: A Qualitative Analysis and Grounded Theory. Library Resources and<br />

Technical Services 43, 4 (Oc<strong>to</strong>ber).<br />

Lessig, Lawrence (1999). Code and Other Laws of Cyberspace. New York: Basic Books.<br />

McClure, Charles et al. (1987). Planning and Role Setting for Public Libraries (figure<br />

11). Chicago: American Library Association.<br />

National Information Standards Organization (NISO) (2002). Standards Committee AU.<br />

Data Dictionary – Technical Metadata for Digital Still Images (June 1). At<br />

http://www.niso.org/standards/resources/Z39_37_trial_use.<strong>pdf</strong>.<br />

National Science Foundation (NSF) (1994). NSF Announces Awards for Digital Libraries<br />

Research; $24.4 Million <strong>to</strong> Fund Advanced Research and Technology Development by<br />

Teams from Universities, Industries and Other Institutions (September). At<br />

http://elib.cs.berkeley.edu:80/admin/proposal/nsf-press-release.html.<br />

OCLC/RLG Working Group on Preservation Metadata (<strong>2001</strong>a). Preservation Metadata<br />

for Digital Objects: A Review of the State of the Art (January 31). At<br />

http://www.oclc.org/<strong>digital</strong>preservation/presmeta_wp.<strong>pdf</strong>.<br />

OCLC/RLG Working Group on Preservation Metadata (<strong>2001</strong>b). A Recommendation for<br />

Content Information (Oc<strong>to</strong>ber). At<br />

http://www.oclc.org/research/pmwg/contentinformation.<strong>pdf</strong>.<br />

36.1 Figure "Traditional" <strong>digital</strong> collection model


36.2 Figure<br />

37.<br />

Preservation<br />

Abby Smith<br />

What is Preservation and Why Does it Matter?<br />

The purpose of preserving cultural and intellectual resources is <strong>to</strong> make their use possible<br />

at some unknown future time. A complete and reliable record of the past is important for<br />

many reasons, not the least of which is <strong>to</strong> provide an audit trail of the actions, thoughts,<br />

deeds, and misdeeds of those who have gone before us. For the <strong>humanities</strong> – a field of<br />

open-ended inquiry in<strong>to</strong> the nature of humankind and especially of the culture it creates –<br />

access <strong>to</strong> the recorded information and knowledge of the past is absolutely crucial, as<br />

both its many subjects of inquiry and its methodologies rely heavily on retrospective as<br />

well as on current resources. Preservation is a uniquely important public good that<br />

underpins the health and well-being of humanistic research and teaching.<br />

Preservation is also vitally important for the growth of an intellectual field and the<br />

professional development of its practitioners. Advances in a field require that there be<br />

ease of communication between its practitioners and that the barriers <strong>to</strong> research and<br />

publishing be as low as possible within a system that values peer review and widespread<br />

sharing and vetting of ideas. Digital technologies have radically lowered the barriers of<br />

communication between colleagues, and between teachers and students. Before e-mail, it<br />

was not easy <strong>to</strong> keep current with one's colleagues in distant locations; and before<br />

listservs and search engines it was hard <strong>to</strong> learn about the work others were doing or <strong>to</strong><br />

hunt down interesting leads in fields related <strong>to</strong> one's own. Those who aspire <strong>to</strong> make a<br />

mark in the <strong>humanities</strong> may be attracted <strong>to</strong> new technologies <strong>to</strong> advance their research<br />

agenda, but those who also aspire <strong>to</strong> make a career in the <strong>humanities</strong> now feel hampered<br />

by the barriers <strong>to</strong> electronic publishing, peer review, and reward for work in advanced<br />

<strong>humanities</strong> computing. The common perception that <strong>digital</strong> creations are not permanent is


among the chief obstacles <strong>to</strong> the widespread adoption of <strong>digital</strong> publishing, and few<br />

scholars are rewarded and promoted for their work in this area.<br />

In research-oriented institutions such as libraries, archives, and his<strong>to</strong>rical societies,<br />

primary and secondary sources should be maintained in a state that allows – if not<br />

encourages – use, and therefore the concept of "fitness for use" is the primary principle<br />

that guides preservation decisions, actions, and investments. (This is in contrast <strong>to</strong><br />

museums, which seldom loan objects <strong>to</strong> patrons or make them available for people <strong>to</strong><br />

<strong>to</strong>uch and manipulate.) Of course, fitness for use also entails describing or cataloguing an<br />

item, ensuring that it is easily found in s<strong>to</strong>rage and retrieved for use, and securing the<br />

object from mishandling, accidental damage, or theft (Price and Smith 2000). Imagine a<br />

run of journals that contains information a researcher wants <strong>to</strong> consult: the researcher<br />

must be able <strong>to</strong> know what the title is (it is often found in a catalogue record in a<br />

database) and where it is held. He or she must be able <strong>to</strong> call the journals up from their<br />

location, whether on site, in remote s<strong>to</strong>rage, or through inter-library loan or document<br />

delivery. Finally, the researcher must find the journals <strong>to</strong> be the actual title and dates<br />

requested, with no pages missing in each volume and no volume missing from the run,<br />

for them <strong>to</strong> be of use.<br />

In the <strong>digital</strong> realm, the ability <strong>to</strong> know about, locate and retrieve, and then verify (or<br />

reasonably assume) that a <strong>digital</strong> object is authentic, complete, and undis<strong>to</strong>rted is as<br />

crucial <strong>to</strong> "fitness for use" or preservation as it is for analogue objects – the manuscripts<br />

and maps, posters and prints, books and journals, or other genres of information that are<br />

captured in continuous waves, as opposed <strong>to</strong> discrete bits, and then recorded on<strong>to</strong><br />

physical media for access (see chapter 32, this volume).<br />

The general approach <strong>to</strong> preserving analogue and <strong>digital</strong> information is exactly the same<br />

– <strong>to</strong> reduce risk of information loss <strong>to</strong> an acceptable level – but the strategies used <strong>to</strong><br />

insure against loss are quite different. In the analogue realm, information is recorded on<br />

<strong>to</strong> and retrieved from a physical medium, such as paper, cassette tapes, parchment, film,<br />

and so forth. But as paper turns brittle, cassette tapes break, or film fades, information is<br />

lost. Therefore, the most common strategy for preserving the information recorded on<strong>to</strong><br />

these media is <strong>to</strong> ensure the physical integrity of the medium, or carrier. The primary<br />

technical challenges <strong>to</strong> analogue preservation involve stabilizing, conserving, or<br />

protecting the material integrity of recorded information. Physical objects, such as books<br />

and magnetic tapes, inevitably age and degrade, and both environmental stresses such as<br />

excess heat or humidity, and the stresses of use tend <strong>to</strong> accelerate that loss. Reducing<br />

stress <strong>to</strong> the object, either by providing optimal s<strong>to</strong>rage conditions or by restricting use of<br />

the object in one way or another (including providing a copy or surrogate <strong>to</strong> the<br />

researcher rather than the original object), are the most common means <strong>to</strong> preserve fragile<br />

materials. Inevitably, preservation involves a trade-off of benefits between current and<br />

future users, because every use of the object risks some loss of information, some<br />

deterioration of the physical artifact, some compromise <strong>to</strong> authenticity, or some risk <strong>to</strong><br />

the data integrity.


In the <strong>digital</strong> realm, there are significant trade-offs between preservation and access as<br />

well, but for entirely different reasons. In this realm information is immaterial, and the bit<br />

stream is not fixed on <strong>to</strong> a stable physical object but must be created ("instantiated" or<br />

"rendered") each time it is used. The trade-offs made between long-term preservation and<br />

current ease of access stem not from so-called data dependencies on physical media per<br />

se, but rather from data dependencies on hardware, software, and, <strong>to</strong> a lesser degree, on<br />

the physical carrier as well.<br />

What is archiving?<br />

The concept of <strong>digital</strong> preservation is widely discussed among professional communities,<br />

including librarians, archivists, computer scientists, and engineers, but for most people,<br />

preservation is not a common or commonly unders<strong>to</strong>od term. "Archiving", though, is<br />

widely used by computer users. To non-professionals, including many scholars and<br />

researchers unfamiliar with the technical aspects of librarianship and archival theory,<br />

<strong>digital</strong> archiving means s<strong>to</strong>ring non-current materials some place "offline" so that they<br />

can be used again. But the terms "archiving", "preservation", and "s<strong>to</strong>rage" have<br />

meaningful technical distinctions – as meaningful as the difference between "brain" and<br />

"mind" <strong>to</strong> a neuroscientist. To avoid confusion, and <strong>to</strong> articulate the special technical<br />

needs of managing <strong>digital</strong> information as compared with analogue, many professionals<br />

are now using the term "persistence" <strong>to</strong> mean long-term access – preservation by another<br />

name. The terms "preservation", "long-term access", and "persistence" will be used<br />

interchangeably here.<br />

Technical Challenges <strong>to</strong> Digital Preservation and Why<br />

They Matter<br />

The goal of <strong>digital</strong> preservation is <strong>to</strong> ensure that <strong>digital</strong> information – be it textual,<br />

numeric, audio, visual, or geospatial – be accessible <strong>to</strong> a future user in an authentic and<br />

complete form. Digital objects are made of bit streams of 0s and 1s arranged in a logical<br />

order that can be rendered on<strong>to</strong> an interface (usually a screen) through computer<br />

hardware and software. The persistence of both the bit stream and the logical order for<br />

rendering is essential for long-term access <strong>to</strong> <strong>digital</strong> objects.<br />

As described by computer scientists and engineers, the two salient challenges <strong>to</strong> <strong>digital</strong><br />

preservation are:<br />

• physical preservation: how <strong>to</strong> maintain the integrity of the bits, the Os and Is that reside<br />

on a s<strong>to</strong>rage medium such as a CD or hard drive; and<br />

• logical preservation: how <strong>to</strong> maintain the integrity of the logical ordering of the object,<br />

that code that makes the bits "renderable" in<strong>to</strong> <strong>digital</strong> objects.<br />

In the broader preservation community outside the sphere of computer science, these<br />

challenges are more often spoken of as:


• media degradation: how <strong>to</strong> ensure that the bits survive intact and that the magnetic tape<br />

or disk or drive on which they are s<strong>to</strong>red do not degrade, demagnetize, or otherwise result<br />

in data loss (this type of loss is also referred <strong>to</strong> as "bit rot"); and<br />

• hardware/software dependencies: how <strong>to</strong> ensure that data can be rendered or read in the<br />

future when the software they were written in and/or the hardware on which they were<br />

designed <strong>to</strong> run are obsolete and no longer supported at the point of use.<br />

Because of these technical dependencies, <strong>digital</strong> objects are by nature very fragile, often<br />

more at risk of data loss and even sudden death than information recorded on brittle paper<br />

or nitrate film s<strong>to</strong>ck.<br />

And from these overarching technical dependencies devolve nearly all other fac<strong>to</strong>rs that<br />

put <strong>digital</strong> data at high risk of corruption, degradation, and loss – the legal, social,<br />

intellectual, and financial fac<strong>to</strong>rs that will determine whether or not we are able <strong>to</strong> build<br />

an infrastructure that will support preservation of these valuable but fragile cultural and<br />

intellectual resources in<strong>to</strong> the future. It may well be that <strong>humanities</strong> scholars, teachers,<br />

and students are most immediately affected by the copyright restrictions, economic<br />

barriers, and intellectual challenges <strong>to</strong> working with <strong>digital</strong> information, but it will be<br />

difficult <strong>to</strong> gain leverage over any of these problems without a basic understanding of the<br />

ultimate technical problems from which all these proximate problems arise. Without<br />

appreciating the technical barriers <strong>to</strong> preservation and all the crucial dependencies they<br />

entail, humanists will not be able <strong>to</strong> create or use <strong>digital</strong> objects that are authentic,<br />

reliable, and of value in<strong>to</strong> the future. So a bit more detail is in order.<br />

Media degradation: Magnetic tape, a primary s<strong>to</strong>rage medium for <strong>digital</strong> as well as<br />

analogue information, is very vulnerable <strong>to</strong> physical deterioration, usually in the form of<br />

separation of the signal (the encoded information itself) from the substrate (the tape on<br />

which the thin layer of bits reside). Tapes need <strong>to</strong> be "exercised" (wound and rewound) <strong>to</strong><br />

maintain even tension and <strong>to</strong> ensure that the signal is not separating. Tapes also need <strong>to</strong><br />

be reformatted from time <strong>to</strong> time, though rates of deterioration are surprisingly variable<br />

(as little as five years in some cases) and so the only sure way <strong>to</strong> know tapes are sound is<br />

by frequent and labor-intensive examinations. CDs are also known <strong>to</strong> suffer from<br />

physical degradation, also in annoyingly unpredictable ways and time frames.<br />

(Preservationists have <strong>to</strong> rely on predictable rates of loss if they are <strong>to</strong> develop<br />

preservation strategies that go beyond hand-crafted solutions for single-item treatments.<br />

Given the scale of <strong>digital</strong> information that deserves preservation, all preservation<br />

strategies will ultimately need <strong>to</strong> be au<strong>to</strong>mated in whole or in part <strong>to</strong> be effective.)<br />

Information s<strong>to</strong>red on hard drives is generally less prone <strong>to</strong> media degradation. But these<br />

media have not been in use long enough for there <strong>to</strong> be meaningful data about how they<br />

have performed over the decades. Finally, a s<strong>to</strong>rage medium may itself be physically<br />

intact and still carry information, or signal, that has suffered degradation – tape that has<br />

been demagnetized is such an example.<br />

Hardware/software obsolescence: Data can be perfectly intact physically on a s<strong>to</strong>rage<br />

medium and yet be unreadable because the hardware and software – the playback


machine and the code in which the data are written – are obsolete. We know a good deal<br />

about hardware obsolescence and its perils already from the numerous defunct playback<br />

machines that "old" audio and visual resources required, such as Beta video equipment,<br />

16 mm home movie projec<strong>to</strong>rs, and numerous proprietary dictation machines. The<br />

problem of software obsolescence may be newer, but it is chiefly the proliferation of<br />

software codes, and their rapid supersession by the next release, that makes this computer<br />

software concern intractable. In the case of software, which comprises the operating<br />

system, the application, and the format, there are multiple layers that each require<br />

attending <strong>to</strong> when preservation strategies, such as those listed below, are developed.<br />

There are currently four strategies under various phases of research, development, and<br />

deployment for addressing the problems of media degradation and hardware/software<br />

obsolescence (Greenstein and Smith 2002).<br />

• Migration. Digital information is transferred, or rewritten, from one hardware/software<br />

configuration <strong>to</strong> a more current one over time as the old formats are superseded by new<br />

ones. Often, a <strong>digital</strong> reposi<strong>to</strong>ry where data are s<strong>to</strong>red will reformat or "normalize" data<br />

going in<strong>to</strong> the reposi<strong>to</strong>ry; that is, the reposi<strong>to</strong>ry will put the data in<strong>to</strong> a standard format<br />

that can be reliably managed over time. As necessary and cost-effective as this process<br />

may be in the long run, it can be expensive and time-consuming. In addition, <strong>digital</strong> files<br />

translated in<strong>to</strong> another format will lose some information with each successive<br />

reformatting (loss similar <strong>to</strong> that of translations from one language <strong>to</strong> another), ranging<br />

from formatting or presentation information <strong>to</strong> potentially more serious forms of loss.<br />

Migration works best for simple data formats and does not work well at all for<br />

multimedia objects. It is the technique most commonly deployed <strong>to</strong>day and it shows<br />

considerable reliability with ASCII text and some numeric databases of the sort that<br />

financial institutions use.<br />

• Emulation. Emulation aims <strong>to</strong> preserve the look and feel of a <strong>digital</strong> object, that is, <strong>to</strong><br />

preserve the functionality of the software as well as the information content of the object.<br />

It requires that information about the encoding and hardware environments be fully<br />

documented and s<strong>to</strong>red with the object itself so that it can be emulated or essentially<br />

recreated on successive generations of hardware/software (though, of course, if that<br />

information is itself <strong>digital</strong>, further problems of accessibility <strong>to</strong> that information must also<br />

be anticipated). Emulation for preservation is currently only in the research phase. People<br />

are able <strong>to</strong> emulate retrospectively recently deceased genres of <strong>digital</strong> objects – such as<br />

certain computer games – but not prospectively for objects <strong>to</strong> be read on unknown<br />

machines and software programs in the distant future. Many in the field doubt that<br />

proprietary software makers will ever allow their software code <strong>to</strong> accompany objects, as<br />

stipulated by emulation, and indeed some in the software industry say that documentation<br />

is never complete enough <strong>to</strong> allow for the kind of emulation 100 or 200 years out that<br />

would satisfy a preservation demand (Rothenberg 1999; Bearman 1999; Holdsworth and<br />

Wheatley 2000). Perhaps more <strong>to</strong> the point, software programs that are several decades or<br />

centuries old may well be no more accessible <strong>to</strong> contemporary users than medieval<br />

manuscripts are accessible <strong>to</strong> present-day readers who have not been trained <strong>to</strong> read


medieval Latin in a variety of idiosyncratic hands. Nonetheless, emulation remains a<br />

tantalizing notion that continues <strong>to</strong> attract research dollars.<br />

• Persistent object preservation. A relatively new approach being tested by the National<br />

Archives and Records Administration for electronic records such as e-mails, persistent<br />

object preservation (POP) "entails explicitly declaring the properties (e.g., content,<br />

structure, context, presentation) of the original <strong>digital</strong> information that ensure its<br />

persistence" (Greenstein and Smith 2002). It envisions wrapping a <strong>digital</strong> object with the<br />

information necessary <strong>to</strong> recreate it on current software (not the original software<br />

envisioned by emulation). This strategy has been successfully tested in its research phase<br />

and the Archives is now developing an implementation program for it. At present it<br />

seems most promising for <strong>digital</strong> information objects such as official records and other<br />

highly structured genres that do not require extensive normalization, or changing it <strong>to</strong><br />

bring it in<strong>to</strong> a common preservation norm upon deposit in<strong>to</strong> the reposi<strong>to</strong>ry. Ironically,<br />

this approach is conceptually related <strong>to</strong> the efforts of some <strong>digital</strong> artists, creating very<br />

idiosyncratic and "un-normalizable" <strong>digital</strong> creations, who are making declarations at the<br />

time of creation on how <strong>to</strong> recreate the art at some time in the future. They do so by<br />

specifying which features of the hardware and software environment are intrinsic and<br />

authentic, and which are fungible and need not be preserved (such things as screen<br />

resolution, processing speed, and so forth, that can affect the look and feel of the <strong>digital</strong><br />

art work) (Thibodeau 2002; Mayfield 2002).<br />

• Technology preservation. This strategy addresses future problems of obsolescence by<br />

preserving the <strong>digital</strong> object <strong>to</strong>gether with the hardware, operating system, and program<br />

of the original. While many will agree that, for all sorts of reasons, someone somewhere<br />

should be collecting and preserving all generations of hardware and software in <strong>digital</strong><br />

information technology, it is hard <strong>to</strong> imagine this approach as much more than a<br />

technology museum attempting production-level work, doomed <strong>to</strong> an uncertain future. It<br />

is unlikely <strong>to</strong> be scalable as an everyday solution for accessing information on orphaned<br />

platforms, but it is highly likely that something like a museum of old technology, with<br />

plentiful documentation about the original hardware and software, will be important for<br />

future <strong>digital</strong> archaeology and data mining. Currently, <strong>digital</strong> archaeologists are able,<br />

with often considerable effort, <strong>to</strong> rescue data from degraded tapes and corrupted (or<br />

erased) hard drives. With some attention now <strong>to</strong> capturing extensive information about<br />

successive generations of hardware and software, future computer engineers should be<br />

able <strong>to</strong> get some information off the old machines.<br />

The preservation and access trade-offs for <strong>digital</strong> information are similar <strong>to</strong> those for<br />

analogue. To make an informed decision about how <strong>to</strong> preserve an embrittled book<br />

whose pages are crumbling or spine is broken, for example, one must weigh the relative<br />

merits of aggressive and expensive conservation treatment <strong>to</strong> preserve the book as an<br />

artifact, versus the less expensive option of reformatting the book's content on<strong>to</strong><br />

microfilm or scanning it, losing most of the information integral <strong>to</strong> the artifact as physical<br />

object in the process. One always needs <strong>to</strong> identify the chief values of a given resource <strong>to</strong><br />

decide between a number of preservation options. In this case, the question would be<br />

whether one values the artifactual or the informational content more highly. This


consideration would apply equally in the <strong>digital</strong> realm: some of the technologies<br />

described above will be cheaper, more easily au<strong>to</strong>mated, and in that sense more scalable<br />

over time than others. Migration appears <strong>to</strong> be suitable for simpler formats in which the<br />

look and feel matters less and in which loss of information at the margins is an acceptable<br />

risk. The other approaches have not been tested yet in large scale over several decades,<br />

but the more options we have for ensuring persistence, the likelier we are <strong>to</strong> make<br />

informed decisions about what we save and how. There is no silver bullet for preserving<br />

<strong>digital</strong> information, and that may turn out <strong>to</strong> be good news in the long run.<br />

Crucial Dependencies and Their Implications for the<br />

Humanities<br />

Preservation by benign neglect has proven an amazingly robust strategy over time, at<br />

least for print-on-paper. One can passively manage a large portion of library collections<br />

fairly cheaply. One can put a well-catalogued book on a shelf in good s<strong>to</strong>rage conditions<br />

and expect <strong>to</strong> be able <strong>to</strong> retrieve it in 100 years in fine shape for use if no one has called it<br />

from the shelf. But neglect in the <strong>digital</strong> realm is never benign. Neglect of <strong>digital</strong> data is a<br />

death sentence. A <strong>digital</strong> object needs <strong>to</strong> be optimized for preservation at the time of its<br />

creation (and often again at the time of its deposit in<strong>to</strong> a reposi<strong>to</strong>ry), and then it must be<br />

conscientiously managed over time if it is <strong>to</strong> stand a chance of being used in the future.<br />

The need for standard file formats and metadata and the role of data<br />

crea<strong>to</strong>rs<br />

All the technical strategies outlined above are crucially dependent on standard file<br />

formats and metadata schemas for the creation and persistence of <strong>digital</strong> objects. File<br />

formats that are proprietary are often identified as being especially at risk, because they<br />

are in principle dependent on support from an enterprise that may go out of business.<br />

Even a format so widely used that it is a de fac<strong>to</strong> standard, such as Adobe Systems, Inc.'s<br />

portable document format (PDF), is treated with great caution by those responsible for<br />

persistence. The owner of a such a de fac<strong>to</strong> standard has no legal obligation <strong>to</strong> release its<br />

source code or any other proprietary information in the event that it goes bankrupt or<br />

decides <strong>to</strong> s<strong>to</strong>p supporting the format for one reason or another (such as creating a better<br />

and more lucrative file format).<br />

Commercial interests are not always in conflict with preservation interests, but when they<br />

are, commercial interests must prevail if the commerce is <strong>to</strong> survive. For that reason, the<br />

effort <strong>to</strong> develop and promote adoption of non-proprietary software, especially so-called<br />

open source code, is very strong among preservationists. (Open source, as opposed <strong>to</strong><br />

proprietary, can be supported by non-commercial as well as commercial users.) But <strong>to</strong> the<br />

extent that commercial services are often in a better position <strong>to</strong> support innovation and<br />

development efforts, the preservation community must embrace both commercial and<br />

non-commercial formats. While preservationists can declare which standards and formats<br />

they would like <strong>to</strong> see used, dictating <strong>to</strong> the marketplace, or ignoring it al<strong>to</strong>gether, is not a<br />

promising solution <strong>to</strong> this problem. One way <strong>to</strong> ensure that important but potentially


vulnerable proprietary file formats are protected if they are orphaned is for leading<br />

institutions with a preservation mandate – national libraries, large research institutions, or<br />

government archives – <strong>to</strong> develop so-called fail-safe agreements with software makers<br />

that allow the code for the format <strong>to</strong> go in<strong>to</strong> receivership or be deeded over <strong>to</strong> a trusted<br />

third party. (For more on file formats, see chapter 32, this volume.)<br />

Metadata schemas – approaches <strong>to</strong> describing information assets for access, retrieval,<br />

preservation, or internal management – is another area in which there is a delicate<br />

balance between what is required for ease of access (and of creation) and what is required<br />

<strong>to</strong> ensure persistence. Extensive efforts have been made by librarians, archivists, and<br />

scholars <strong>to</strong> develop sophisticated markup schemes that are preservation-friendly, such as<br />

the Text Encoding Initiative (TEI) Guidelines, which were first expressed in SGML, or<br />

Encoded Archival Description, which, like the current instantiation of the TEI, is written<br />

in XML, and open for all communities, commercial and non-commercial alike. The<br />

barriers <strong>to</strong> using these schemas can be high, however, and many authors and crea<strong>to</strong>rs who<br />

understand the importance of creating good metadata nevertheless find these schema <strong>to</strong>o<br />

complicated or time-consuming <strong>to</strong> use consistently. It can be frustrating <strong>to</strong> find out that<br />

there are best practices for the creation of preservable <strong>digital</strong> objects, but that those<br />

practices are prohibitively labor-intensive for most practitioners.<br />

The more normalized and standard a <strong>digital</strong> object is, the easier it is for a <strong>digital</strong><br />

reposi<strong>to</strong>ry <strong>to</strong> take it in (a process curiously called "ingest"), <strong>to</strong> manage it over time, and <strong>to</strong><br />

provide the objects back <strong>to</strong> users in their original form. Indeed, most reposi<strong>to</strong>ries under<br />

development in libraries and archives declare that they will assume responsibility for<br />

persistence only if the objects they receive are in certain file formats accompanied by<br />

certain metadata. This is in sharp contrast <strong>to</strong> the more straightforward world of books,<br />

pho<strong>to</strong>graphs, or maps, where one can preserve the artifact without having <strong>to</strong> catalogue it<br />

first. Boxes of unsorted and undescribed sources can languish for years before being<br />

discovered, and once described or catalogued, they can have a productive life as a<br />

resource. Although fully searchable text could, in theory, be retrieved without much<br />

metadata in the future, it is hard <strong>to</strong> imagine how a complex or multimedia <strong>digital</strong> object<br />

that goes in<strong>to</strong> s<strong>to</strong>rage of any kind could ever survive, let alone be discovered and used, if<br />

it were not accompanied by good metadata. This creates very large up-front costs for<br />

<strong>digital</strong> preservation, both in time and money, and it is not yet clear who is obligated <strong>to</strong><br />

assume those costs.<br />

For non-standard or unsupported formats and metadata schemas, <strong>digital</strong> reposi<strong>to</strong>ries<br />

might simply promise that they will deliver back the bits as they were received (that is,<br />

provide physical preservation) but will make no such promises about the legibility of the<br />

files (that is, not guarantee logical preservation). Though developed in good faith, these<br />

policies can be frustrating <strong>to</strong> crea<strong>to</strong>rs of complex <strong>digital</strong> objects, or <strong>to</strong> those who are not<br />

used <strong>to</strong> or interested in investing their own time in preparing their work for permanent<br />

retention. This is what publishers and libraries have traditionally done, after all. Some ask<br />

why should it be different now.


There is no question that a <strong>digital</strong> object's file format and metadata schema greatly affect<br />

its persistence and how it will be made available in the future. This crucial dependency of<br />

<strong>digital</strong> information on format and markup begs the question of who should pay for file<br />

preparation and what economic model will support this expensive enterprise. Who are the<br />

stakeholders in <strong>digital</strong> preservation, and what are their roles in this new information<br />

landscape? There is now an interesting negotiation under way between data crea<strong>to</strong>rs and<br />

distribu<strong>to</strong>rs on the one hand, and libraries on the other, about who will bear the costs of<br />

ingest. Various institutions of higher learning that are stepping up <strong>to</strong> the challenge of<br />

<strong>digital</strong> preservation are working out a variety of local models that bear watching closely<br />

(see below, the section on solutions and current activities).<br />

Need for early preservation action and the role of copyright<br />

Regardless of the outcome, it seems clear that those who create intellectual property in<br />

<strong>digital</strong> form need <strong>to</strong> be more informed about what is at risk if they ignore longevity issues<br />

at the time of creation. This means that scholars should be attending <strong>to</strong> the information<br />

resources crucial <strong>to</strong> their fields by developing and adopting the document standards vital<br />

for their research and teaching, with the advice of preservationists and computer<br />

scientists where appropriate. The examples of computing-intensive sciences such as<br />

genomics that have developed professional tracks in informatics might prove fruitful for<br />

humanists as more and more computing power is applied <strong>to</strong> humanistic inquiry and<br />

pedagogy. Such a function would not be entirely new <strong>to</strong> the <strong>humanities</strong>; in the nineteenth<br />

century, a large number of eminent scholars became heads of libraries and archives in an<br />

age when a scholar was the information specialist par excellence.<br />

One of the crucial differences between the information needs of scientists and those of<br />

humanists is that the latter tend <strong>to</strong> use a great variety of sources that are created outside<br />

the academy and that are largely protected by copyright. Indeed, there is probably no type<br />

of information created or recorded by human beings that could not be of value for<br />

<strong>humanities</strong> research at some time, and little of it may be under the direct control of the<br />

researchers who most value it for its research potential. The chief concern about<br />

copyright that impinges directly on preservation is the length of copyright protection that<br />

current legislation extends <strong>to</strong> the rights holders – essentially, the life of the crea<strong>to</strong>r plus<br />

70 years (or more) (Copyright Office, website). Why that matters <strong>to</strong> preservation goes<br />

back <strong>to</strong> the legal regime that allows libraries and archives <strong>to</strong> preserve materials that are<br />

protected by copyright. Institutions with a preservation mission receive or buy<br />

information – books, journals, manuscripts, maps – <strong>to</strong> which they may have no<br />

intellectual rights. But rights over the physical objects themselves do transfer, and the law<br />

allows those institutions <strong>to</strong> copy the information in those artifacts for the purposes of<br />

preservation.<br />

This transfer of property rights <strong>to</strong> collecting institutions breaks down with the new<br />

market in <strong>digital</strong> information. Publishers and distribu<strong>to</strong>rs of <strong>digital</strong> information very<br />

seldom sell their wares. They license them. This means that libraries no longer own the<br />

journals, databases, and other <strong>digital</strong> intellectual property <strong>to</strong> which they provide access,<br />

and they have no incentive <strong>to</strong> preserve information that they essentially rent. Because


publishers are not in the business of securing and preserving information "in perpetuity",<br />

as the phrase goes, there is potentially a wealth of valuable <strong>digital</strong> resources that no<br />

institution is claiming <strong>to</strong> preserve in this new information landscape. Some libraries,<br />

concerned about the potentially catastrophic loss of primary sources and scholarly<br />

literature, have successfully negotiated "preservation clauses" in their licensing<br />

agreements, stipulating that publishers give them physical copies (usually CDs) of the<br />

<strong>digital</strong> data if they cease licensing it, so that they can have perpetual access <strong>to</strong> what they<br />

paid for. While CDs are good for current access needs, few libraries consider them <strong>to</strong> be<br />

archival media. Some commercial and non-commercial publishers of academic literature<br />

have forged experimental agreements with libraries, ensuring that in the event of a<br />

business failure, the <strong>digital</strong> files of the publisher will go <strong>to</strong> the library.<br />

It is natural that a bibliocentric culture such as the academy has moved first on the issue<br />

of what scholars themselves publish. The greater threat <strong>to</strong> the his<strong>to</strong>rical record, however,<br />

is not <strong>to</strong> the secondary literature on which publishers and libraries are chiefly focused.<br />

The exponential growth of visual resources and sound recordings in the past 150 years<br />

has produced a wealth of primary source materials in audiovisual formats <strong>to</strong> which<br />

humanists will demand access in the future. It is likely that most of these resources, from<br />

performing arts <strong>to</strong> moving image, pho<strong>to</strong>graphs, music, radio and television broadcasting,<br />

geospatial objects, and more, are created for the marketplace and are under copyright<br />

protection (Lyman and Varian 2000).<br />

Efforts have barely begun <strong>to</strong> negotiate with the major media companies and their trade<br />

associations <strong>to</strong> make film and television studios, recording companies, news pho<strong>to</strong><br />

services, <strong>digital</strong> car<strong>to</strong>graphers, and others aware of their implied mandate <strong>to</strong> preserve<br />

their corporate <strong>digital</strong> assets for the greater good of a common cultural heritage. As long<br />

as those cultural and intellectual resources are under the control of enterprises that do not<br />

know about and take up their preservation mandate, there is a serious risk of major losses<br />

for the future, analogous <strong>to</strong> the fate of films in the first 50 years of their existence. More<br />

than 80 percent of silent films made in the United States and 50 percent made before<br />

1950 are lost, presumably for ever. Viewing film as "commercial product", the studios<br />

had no interest in retaining them after their productive life, and libraries and archives had<br />

no interest in acquiring them on behalf of researchers. It is critical that humanists start<br />

<strong>to</strong>day <strong>to</strong> identify the <strong>digital</strong> resources that may be of great value now or in the future so<br />

that they can be captured and preserved before their date of expiration arrives.<br />

Need <strong>to</strong> define the values of the <strong>digital</strong> object and the role of the<br />

humanist<br />

Among the greatest values of <strong>digital</strong> information technologies for scholars and students is<br />

the ability of <strong>digital</strong> information <strong>to</strong> transform the very nature of inquiry. Not bound <strong>to</strong><br />

discrete physical artifacts, <strong>digital</strong> information is available anywhere at any time<br />

(dependent on connectivity). Through computing applications that most computer users<br />

will never understand in fine grain, <strong>digital</strong> objects can be easily manipulated, combined,<br />

erased, and cloned, all without leaving the physical traces of tape, erasure marks and<br />

whiteouts, the tell-tale shadow of the pho<strong>to</strong>copied version, and other subtle physical clues


that apprise us of the authenticity and provenance, or origin, of the pho<strong>to</strong>graph or map we<br />

hold in our hands.<br />

Inquiring minds who need <strong>to</strong> rely on the sources they use <strong>to</strong> be authentic – for something<br />

<strong>to</strong> be what it purports <strong>to</strong> be – are faced with special concerns in the <strong>digital</strong> realm, and<br />

considerable work is being done in this area <strong>to</strong> advance our trust in <strong>digital</strong> sources and <strong>to</strong><br />

develop <strong>digital</strong> information literacy among users (Bearman and Trant 1998; CLIR 2000).<br />

Where this issue of the malleability of <strong>digital</strong> information most affects humanistic<br />

research and preservation, beyond the crucial issue of authenticity, is that of fixity and<br />

stability. While the genius of <strong>digital</strong> objects is their ability <strong>to</strong> be modified for different<br />

purposes, there are many reasons why information must be fixed and stable at some point<br />

<strong>to</strong> be reliable in the context of research and interpretation. For example, <strong>to</strong> the extent that<br />

research and interpretation builds on previous works, both primary and secondary, it is<br />

important for the underlying sources of an interpretation or scientific experiment or<br />

observation <strong>to</strong> be accessible <strong>to</strong> users in the form in which it was cited by the crea<strong>to</strong>r. It<br />

would be useless <strong>to</strong> have an article proposing a new interpretation of the Salem witch<br />

trials rely on diary sources that are not accessible <strong>to</strong> others <strong>to</strong> investigate and verify. But<br />

when writers cite as their primary sources web-based materials and the reader can find<br />

only a dead link, that is, in effect, the same thing.<br />

To the extent that the pursuit of knowledge in any field builds upon the work of others,<br />

the chain of reference and ease of linking <strong>to</strong> reference sources are crucial. Whose<br />

responsibility is it <strong>to</strong> maintain the persistence of links in an author's article or a student's<br />

essay? Somehow, we expect the question of persistence <strong>to</strong> be taken care of by some vital<br />

but invisible infrastructure, not unlike the water that comes out of the tap when we turn<br />

the knob. Clearly, that infrastructure does not yet exist. But even if it did, there are still<br />

nagging issues about persistence that scholars and researchers need <strong>to</strong> resolve, such as the<br />

one known as "versioning", or deciding which iteration of a dynamic and changing<br />

resource should be captured and curated for preservation. This is a familiar problem <strong>to</strong><br />

those who work in broadcast media, and <strong>digital</strong> humanists can profit greatly from the<br />

sophisticated thinking that has gone on in audiovisual archives for generations.<br />

The problem of persistent linking <strong>to</strong> sources has larger implications for the growth of the<br />

<strong>humanities</strong> as a part of academic life, and for the support of emerging trends in<br />

scholarship and teaching. Until publishing a journal article, a computer model, or a<br />

musical analysis in <strong>digital</strong> form is seem as persistent and therefore a potentially longlasting<br />

contribution <strong>to</strong> the chain of knowledge creation and use, few people will be<br />

attracted <strong>to</strong> work for reward and tenure in these media, no matter how superior the media<br />

may be for the research in<strong>to</strong> and expression of an idea.<br />

Solutions and Current Activities<br />

There has been considerable activity in both the basic research communities (chiefly<br />

among computer scientists and information scientists) and at individual institutions<br />

(chiefly libraries and federal agencies) <strong>to</strong> address many of the critical technical issues of<br />

building and sustaining <strong>digital</strong> reposi<strong>to</strong>ries for long-term management and persistence.


The private sec<strong>to</strong>r, while clearly a leading innova<strong>to</strong>r in information technologies, both in<br />

the development of hardware and software and in the management of <strong>digital</strong> assets such<br />

as television, film, and recorded sound, has not played a leading public role in the<br />

development of <strong>digital</strong> preservation systems. That is primarily because the time horizons<br />

of the preservation community and of the commercial sec<strong>to</strong>rs are radically different.<br />

Most data s<strong>to</strong>rage systems in the private sec<strong>to</strong>r aim for retention of data for no more than<br />

five <strong>to</strong> ten years (the latter not being a number that a data s<strong>to</strong>rage business will commit<br />

<strong>to</strong>). The time horizon of preservation for libraries, archives, and research institutions must<br />

include many generations of inquiring humans, not just the next two generations of<br />

hardware or software upgrades.<br />

Because <strong>digital</strong> preservation is so complex and expensive, it is unlikely that <strong>digital</strong><br />

reposi<strong>to</strong>ries will spring up in the thousands of institutions that have traditionally served as<br />

preservation centers for books. Nor should they. In a networked environment in which<br />

one does not need access <strong>to</strong> a physical object <strong>to</strong> have access <strong>to</strong> information, the<br />

relationship between ownership (and physical cus<strong>to</strong>dy) of information and access <strong>to</strong> it<br />

will be transformed. Within the research and nonprofit communities, it is likely that the<br />

system of <strong>digital</strong> preservation reposi<strong>to</strong>ries, or <strong>digital</strong> archives, will be distributed among a<br />

few major ac<strong>to</strong>rs that work on behalf of a large universe of users. They will be, in other<br />

words, part of the so-called public goods information economy that research and teaching<br />

have traditionally relied upon for core services such as preservation and collection<br />

building.<br />

Among the major ac<strong>to</strong>rs in <strong>digital</strong> archives will be academic disciplines whose <strong>digital</strong><br />

information assets are crucial <strong>to</strong> the field as a whole. Examples include organizations<br />

such as the Inter-university Consortium for Political and Social Research (ICPSR), which<br />

manages social science datasets, and the Human Genome Data Bank, which preserves<br />

genetic data. Both are supported directly by the disciplines themselves and through<br />

federal grants. The data in these archives are not necessarily complex, but they are highly<br />

structured and the deposi<strong>to</strong>rs are responsible for preparing the data for deposit. JSTOR, a<br />

non-commercial service that preserves and provides access <strong>to</strong> <strong>digital</strong> versions of key<br />

scholarly journals, is another model of a preservation enterprise designed <strong>to</strong> meet the<br />

needs of researchers. It is run on behalf of researchers and financially supported by<br />

libraries through subscriptions. (Its start-up costs were provided by a private foundation.)<br />

Some large research university libraries, such as the University of California, Harvard<br />

University, Massachusetts Institute of Technology, and Stanford University, are<br />

beginning <strong>to</strong> develop and deploy <strong>digital</strong> reposi<strong>to</strong>ries that will be responsible for some<br />

circumscribed portion of the <strong>digital</strong> output of their faculties. Another <strong>digital</strong> library<br />

leader, Cornell, has recently taken under its wing the disciplinary pre-print archive<br />

developed <strong>to</strong> serve the high-energy physics community and its need for rapid<br />

dissemination of information among the small but geographically far-flung members of<br />

that field. In the sciences, there are several interesting models of <strong>digital</strong> information being<br />

created, curated, and preserved by members of the discipline, among which arXiv.org is<br />

surely the best known. This model appears difficult <strong>to</strong> emulate in the <strong>humanities</strong> because<br />

arts and <strong>humanities</strong> disciplines do not create shared information resources that are then


used by many different research teams. Nevertheless, where redundant collections, such<br />

as journals and slide libraries, exist across many campuses, subscription-based services<br />

are being developed <strong>to</strong> provide access <strong>to</strong> and preserve those resources through <strong>digital</strong><br />

surrogates. Examples include JSTOR, AMICO, and ARTs<strong>to</strong>r, now under development.<br />

The economies of scale can be achieved only if the libraries and museums that subscribe<br />

<strong>to</strong> these services believe that the provider will persistently manage the <strong>digital</strong> surrogates<br />

and that they are therefore able <strong>to</strong> dispose of extra copies of journals or slides.<br />

Learned societies may be a logical locus of <strong>digital</strong> archives, as they are trusted third<br />

parties within a field and widely supported by members of the communities they serve.<br />

But they are not, as a rule, well positioned <strong>to</strong> undertake the serious capital expenditures<br />

that reposi<strong>to</strong>ry services require. That seems <strong>to</strong> be the reason why these subscription-based<br />

preservation and access services have appeared in the marketplace.<br />

The Library of Congress (LC), which is the seat of the Copyright Office and receives for<br />

inclusion in its collections one or more copies of all works deposited for copyright<br />

protection, is beginning <strong>to</strong> grapple with the implications of <strong>digital</strong> deposits and what they<br />

mean for the growth of the Library's collections. (At present, with close <strong>to</strong> 120 million<br />

items in its collections, it is several times larger than other <strong>humanities</strong> collections in the<br />

United States.) LC is developing a strategy <strong>to</strong> build a national infrastructure for the<br />

preservation of <strong>digital</strong> heritage that would leverage the existing and future preservation<br />

activities across the nation (and around the globe) <strong>to</strong> ensure that the greatest number of<br />

people can have persistent rights-protected access <strong>to</strong> that heritage. The National Archives<br />

is also working <strong>to</strong> acquire and preserve the <strong>digital</strong> output of the federal government,<br />

though this will entail an expenditure of public resources for preservation that is<br />

unprecedented in a nation that prides itself on its accountability <strong>to</strong> its people.<br />

The final word about major ac<strong>to</strong>rs in preservation belongs <strong>to</strong> the small group of visionary<br />

private collec<strong>to</strong>rs that have fueled the growth of great <strong>humanities</strong> collections for<br />

centuries. The outstanding exemplar of the <strong>digital</strong> collec<strong>to</strong>r is Brewster Kahle, who has<br />

designed and built the Internet Archive, which captures and preserves a large number of<br />

publicly available sites. While the visible and publicly available Web that the Internet<br />

Archive harvests is a small portion of the <strong>to</strong>tal Web (Lyman 2002), the Archive has<br />

massive amounts of culturally rich material. Kahle maintains the Archive as a<br />

preservation reposi<strong>to</strong>ry – that is its explicit mission – and in that sense his enterprise is a<br />

beguiling peek in<strong>to</strong> the future of collecting in the <strong>digital</strong> realm.<br />

Selection for Preservation<br />

While much work remains <strong>to</strong> ensure that <strong>digital</strong> objects of high cultural and research<br />

value persist in<strong>to</strong> the future, many experts are cautiously optimistic that, with enough<br />

funding and will, technical issues will be addressed and acceptable solutions will be<br />

found. The issue that continues <strong>to</strong> daunt the most thoughtful among those engaged in<br />

preservation strategies is selection: how <strong>to</strong> determine what, of the massive amount of<br />

information available, should be captured and s<strong>to</strong>red and managed over time.


In theory, there is nothing created by the hands of mankind that is not of potential<br />

research value for <strong>humanities</strong> scholars, even the humblest scrap of data – tax records,<br />

laundry lists, porn sites, personal websites, weblogs, and so forth. Technology optimists<br />

who believe that it will be possible <strong>to</strong> "save everything" through au<strong>to</strong>mated procedures<br />

have advocated doing so. Some teenager is out there <strong>to</strong>day, they point out, who will be<br />

president of the United States in 30 years, and she no doubt already has her own website.<br />

If we save all websites we are bound <strong>to</strong> save hers. If completeness of the his<strong>to</strong>rical record<br />

is a value that society should support, then saving everything that can be saved appears <strong>to</strong><br />

be the safest policy <strong>to</strong> minimize the risk of information loss.<br />

There may well be compelling social reasons <strong>to</strong> capture everything from the Web and<br />

save it forever – if it were possible and if there were no legal and privacy issues. (Most of<br />

the Web, the so-called Deep Web, is not publicly available, and many copyright experts<br />

tend <strong>to</strong> think that all non-federal sites are copyright-protected.) Certainly, everything<br />

created by public officials in the course of doing their business belongs in the public<br />

record, in complete and undis<strong>to</strong>rted form. Further, there are the huge and expensive s<strong>to</strong>res<br />

of data about our world – from census data <strong>to</strong> the petabytes of data sent back <strong>to</strong> Earth<br />

from orbiting satellites – that may prove invaluable in future scientific problem solving.<br />

On the other hand, humanists have traditionally valued the enduring quality of an<br />

information object as much as the quantity of raw data it may yield. There is reason <strong>to</strong><br />

think that humanists in the future will be equally interested in the depth of information in<br />

a source as they are in the sheer quantity of it. Indeed, some his<strong>to</strong>rians working in the<br />

modern period already deplore the promiscuity of paper-based record making and<br />

keeping in contemporary life.<br />

But the <strong>digital</strong> revolution has scarcely begun, and the <strong>humanities</strong> have been slower <strong>to</strong><br />

adopt the technology and test its potential for transforming the nature of inquiry than<br />

have other disciplines that rely more heavily on quantitative information. Some fields –<br />

his<strong>to</strong>ry is a good example – have gone through periods when quantitative analysis has<br />

been widely used, but this was before the advent of computing power on <strong>to</strong>day's scale. If<br />

and when humanists discover the ways that computers can truly change the work of<br />

research and teaching, then we can expect <strong>to</strong> see the growth of large and commonly used<br />

databases that will demand heavy investments in time and money and that will, therefore,<br />

beg the question of persistence.<br />

We will not be able in the future <strong>to</strong> rely on traditional assessments of value for<br />

determining what deserves preservation. In the <strong>digital</strong> realm, there will be no uniqueness,<br />

no scarcity, no category of "rare." There will remain only the signal categories of<br />

evidential value, aesthetic value, and associational value, the very criteria that are, by<br />

their subjectivity, best assessed by scholar experts. The role of humanists in building and<br />

preserving collections of high research value will become as important as it was in the<br />

Renaissance or the nineteenth century. Unlike those eras, however, when scholars could<br />

understand the value of sources as they have revealed themselves over time, there is no<br />

distinction between collecting "just in case" something proves later <strong>to</strong> be valuable, and<br />

"just in time" for someone <strong>to</strong> use now. Scholars cannot leave it <strong>to</strong> later generations <strong>to</strong><br />

collect materials created <strong>to</strong>day. They must assume a more active role in the stewardship


of research collections than they have played since the nineteenth century or, indeed,<br />

ever.<br />

Digital Preservation as a Strategy for Preserving Non<strong>digital</strong><br />

Collections<br />

There seems little doubt that materials that are "born <strong>digital</strong>" need <strong>to</strong> be preserved in<br />

<strong>digital</strong> form. Yet another reason why ensuring the persistence of <strong>digital</strong> information is<br />

crucial <strong>to</strong> the future of <strong>humanities</strong> scholarship and teaching is that a huge body of very<br />

fragile analogue materials demands <strong>digital</strong> reformatting for preservation. Most moving<br />

image and recorded sound sources exist on media that are fragile, such as nitrate film,<br />

audio tapes, or lacquer disks, and they all require reformatting on <strong>to</strong> fresher media <strong>to</strong><br />

remain accessible for use. Copying analogue information <strong>to</strong> another analogue format<br />

results in significant loss of signal within a generation or two (imagine copying a video of<br />

a television program over and over), so most experts believe that reformatting on<strong>to</strong> <strong>digital</strong><br />

media is the safest strategy. Digital reformatting does not result in significant (in most<br />

cases in any) loss of information.<br />

For those who prize access <strong>to</strong> the original source materials for research and teaching, the<br />

original artifact is irreplaceable. For a large number of other uses, an excellent surrogate<br />

or access copy is fine. Digital technology can enhance the preservation of artifacts by<br />

providing superlative surrogates of original sources while at the same time protecting the<br />

artifact from overuse. (See chapter 32, this volume.)<br />

All this means that, as crea<strong>to</strong>rs and consumers of <strong>digital</strong> information, humanists are<br />

vitally interested in <strong>digital</strong> preservation, both for <strong>digital</strong> resources and for the abundance<br />

of valuable but physically fragile analogue collections that must rely on digitization.<br />

Given the extraordinary evanescence of <strong>digital</strong> information, it is crucial now for them <strong>to</strong><br />

engage the copyright issues, <strong>to</strong> develop the economic models for building and sustaining<br />

the core infrastructure that will support persistent access, and, most proximate <strong>to</strong> their<br />

daily lives, <strong>to</strong> ensure that students and practitioners of the <strong>humanities</strong> are appropriately<br />

trained and equipped <strong>to</strong> conduct research, write up results for dissemination, and enable a<br />

new generation of students <strong>to</strong> engage in the open-ended inquiry in<strong>to</strong> culture that is at the<br />

core of the humanistic enterprise.<br />

Acknowledgments<br />

The author thanks Amy Friedlander, Kathlin Smith, and David Rumsey for their<br />

invaluable help.<br />

Bibliography<br />

Bearman, David (1999). Reality and Chimeras in the Preservation of Electronic Records.<br />

D-Lib Magazine 5, 4. At http://www.dlib.org/dlib/april99/bearman/04bearman.html.


Bearman, David and Jennifer Trant (1998). Authenticity of Digital Resources: Towards a<br />

Statement of Requirements in the Research Process. D-Lib Magazine (June). At<br />

http://www.dlib.org/dlib/june98/06bearman.html.<br />

Copyright Office. Circulars 15 and 15a. Accessed April 27, 2004. At<br />

http://www.loc.gov/copyright/circs/circ!5.<strong>pdf</strong> (CLIR)<br />

http://www.loc.gov/copyright/circs/circl5a.<strong>pdf</strong>.<br />

Council on Library and Information Resources; (2000). Authenticity in a Digital<br />

Environment. Washing<strong>to</strong>n, DC: Council on Library and Information Resources. At<br />

http://www.clir.org/pubs/reports/pub92/contents.html.<br />

Greenstein, Daniel and Abby Smith (2002). Digital Preservation in the United States:<br />

Survey of Current Research, Practice, and Common Understandings. At<br />

http://www.<strong>digital</strong>preservation.gov.<br />

Holdsworth, David and Paul Wheatley (2000). Emulation, Preservation and Abstraction.<br />

CAMiLEON Project, University of Leeds. At<br />

http://129.11.152.25/CAMiLEON/dh/ep5.html.<br />

Lyman, Peter (2002). Archiving the World Wide Web. In Building a National Strategy for<br />

Digital Preservation: Issues in Digital Media Archiving. Washing<strong>to</strong>n, DC: Council on<br />

Library and Information Resources and the Library of Congress. At<br />

http://www.clir.org/pubs/reports/publ06/con-tents.html; and<br />

http://www.<strong>digital</strong>preservation.gov/ndiipp/repor/repor_back_web.html.<br />

Lyman, Peter, and Hal R. Varian (2000). How Much Information? At<br />

http://www.sims.berkeley.edu/how-much-info.<br />

Mayfield, Kedra (2002). How <strong>to</strong> Preserve Digital Art. Wired (July 23). At<br />

http://www.wired.com/news/culture/0,1284,53712,00.html.<br />

Price, Laura and Abby Smith (2000). Managing Cultural Assets from a Business<br />

Perspective. Washing<strong>to</strong>n, DC: Council on Library and Information Resources. At<br />

http://www.clir.org/pubs/reports/pub90/contents.html.<br />

Rothenberg, Jeff (1999). Avoiding Technological Quicksand: Finding a Viable Technical<br />

Foundation for Digital Preservation. Washing<strong>to</strong>n, DC: Council on Library and<br />

Information Resources. At http://www.clir.org/pubs/reports/pub77/contents.html.<br />

Thibodeau, Kenneth (2002). Overview of Technological Approaches <strong>to</strong> Digital<br />

Preservation and Challenges in Coming Years. In The State of Digital Preservation: An<br />

International Perspective. Conference Proceedings. Documentation Abstracts, Inc.,<br />

Institutes for Information Science Washing<strong>to</strong>n, DC, April 24–25, 2002. Washing<strong>to</strong>n, DC:<br />

Council on Library and Information Resources. At<br />

http://www.clir.org/pubs/reports/publ07/contents.html.


Websites of Organizations and Projects Noted<br />

Art Museum Image Consortium (AMICO). http://www.amico.org.<br />

ARTs<strong>to</strong>r. http://www.mellon.org/programs/otheractivities/ARTs<strong>to</strong>r/ARTs<strong>to</strong>r.htm.<br />

ArXiv.org. http://www.arXiv.org.<br />

Internet Archive, http://www.archive.org.<br />

Inter-university Consortium for Political and Social Research (ICPSR).<br />

http://www.icpsr.umich.edu.<br />

JSTOR. http://www.js<strong>to</strong>r.org.<br />

Library of Congress. The National Digital Information and Infrastructure Preservation<br />

Program of the. Library of Congress is available at http://www.<strong>digital</strong>preservation.gov.<br />

National Archives and Record Administration (NARA). Electronic records Archives,<br />

http://www.archives.gov/electronic_records_archives.<br />

National Archives and Record Administration (NARA). Electronic records Archives<br />

http://www.archives.gov/electronic_records_archives.<br />

For Further Reading<br />

On the technical, social, organizational, and legal issues related <strong>to</strong> <strong>digital</strong> preservation,<br />

the best source continues <strong>to</strong> be Preserving Digital Information: Report of the Task Force<br />

on Archiving of Digital Information (Washing<strong>to</strong>n, DC, and Mountain View, CA:<br />

Commission of Preservation and Access and the Research Libraries Group, Inc.).<br />

Available at http://www.rlg.org/ArchTF.<br />

Digital preservation is undergoing rapid development, and few publications remain<br />

current for long. To keep abreast of developments in <strong>digital</strong> preservation, see the<br />

following list, which themselves regularly digest or cite the latest information on the<br />

subject and present the leading research of interest <strong>to</strong> humanists on the subject:.<br />

D-Lib Magazine. http://www.dlib.org.<br />

Council on Library and Information Resources (CLIR). http://www.clir.org.<br />

Digital Library Federation (DLF). http://www.diglib.org.<br />

National Science Foundation (NSF), Digital Libraries Initiative (DLI).<br />

http://www.dli2.nsf.gov.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!