Category Archives: library

library of congress to archive electronic literature (suggest a link)

The Electronic Literature Organization seeks your assistance in selecting “works of imaginative writing that take advantage of the capabilities of the standalone or networked computer” for preservation by the LOC and Internet Archive:

The Library of Congress has asked the Electronic Literature Organization to collect a sample of 300 web sites related to the field and to contribute that sample to the Internet Archive’s Archive-It project. The sites selected will be crawled and archived to the extent that the Archive-It technology allows. The result will be full-text searchable collections of the spidered HTML files in the Internet Archive’s Wayback Machine. The ELO will enter metadata including a short description and keywords for each URL entered into the database. The ELO Board of Directors, Literary Advisory Board, membership, and community are encouraged to suggest sites here for three sets of links.
-? Electronic Literature: Collections of Works: Sites that aggregate works of electronic literature by multiple authors, such as online journals and anthologies.
-? Electronic Literature: Individual Works: Individual works of electronic literature and collections of works by a single author, as opposed to collections of works by multiple authors.
-? Electronic Literature: Context: Sites related to the critical, theoretical, and institutional contexts of electronic literature.

More info on how to suggest links at the ELO wiki.

googlesoft gets the brush-off

This is welcome. Several leading American research libraries including the Boston Public and the Smithsonian have said no thanks to Google and Microsoft book digitization deals, opting instead for the more costly but less restrictive Open Content Alliance/Internet Archive program. The NY Times reports, and explains how private foundations like Sloan are funding some of the OCA partnerships.

the really modern library

This is a request for comments. We’re in the very early stages of devising, in partnership with Peter Brantley and the Digital Library Federation, what could become a major initiative around the question of mass digitization. It’s called “The Really Modern Library.”
Over the course of this month, starting Thursday in Los Angeles, we’re holding a series of three invited brainstorm sessions (the second in London, the third in New York) with an eclectic assortment of creative thinkers from the arts, publishing, media, design, academic and library worlds to better wrap our minds around the problems and sketch out some concrete ideas for intervention. Below I’ve reproduced some text we’ve been sending around describing the basic idea for the project, followed by a rough agenda for our first meeting. The latter consists mainly of questions, most of which, if not all, could probably use some fine tuning. Please feel encouraged to post responses, both to the individual questions and to the project concept as a whole. Also please add your own queries, observations or advice.
The Really Modern Library (basically)
The goal of this project is to shed light on the big questions about future accessibility and usability of analog culture in a digital, networked world.
We are in the midst of a historic “upload,” a frenetic rush to transfer the vast wealth of analog culture to the digital domain. Mass digitization of print, images, sound and film/video proceeds apace through the efforts of actors public and private, and yet it is still barely understood how the media of the past ought to be preserved, presented and interconnected for the future. How might we bring the records of our culture with us in ways that respect the originals but also take advantage of new media technologies to enhance and reinvent them?
Our aim with the Really Modern Library project is not to build a physical or even a virtual library, but to stimulate new thinking about mass digitization and, through the generation of inspiring new designs, interfaces and conceptual models, to spur innovation in publishing, media, libraries, academia and the arts.
The meeting in October will have two purposes. The first is to deepen and extend our understanding of the goals of the project and how they might best be achieved. The second is to begin outlining plans for a major international design competition calling for proposals, sketches, and prototypes for a hypothetical “really modern library.” This competition will seek entries ranging from the highly particular (for e.g., designs for digital editions of analog works, or new tools and interfaces for handling pre-digital media) to the broadly conceptual (ideas of how to visualize, browse and make use of large networked collections).
This project is animated by a strong belief that it is the network, more than the simple conversion of atoms to bits, that constitutes the real paradigm shift inherent in digital communication. Therefore, a central question of the Really Modern Library project and competition will be: how does the digital network change our relationship with analog objects? What does it mean for readers/researchers/learners to be in direct communication in and around pieces of media? What should be the *social* architecture of a really modern library?
The call for entries will go out to as broad a community as possible, including designers, artists, programmers, hackers, librarians, archivists, activists, educators, students and creative amateurs. Our present intent is to raise a large sum of money to administer the competition and to have a pool for prizes that is sufficiently large and meaningful that it can compel significant attention from the sort of minds we want working on these problems.
Meeting Agenda
Although we have tended to divide the Really Modern Library Project into two stages – the first addressing the question of how we might best take analog culture with us into the digitally networked future and the second, how the digitally networked library of the future might best be conceived and organized – these questions are joined at the hip and not easily or productively isolated from each other.
Realistically, any substantive answer to the question of how to re-present artifacts of analog culture in the digital network immediately raises issues ranging from new forms of browsing (in a social network) to new forms of reading (in a social network) which have everything to do with the broader infrastructure of the library itself.
We’re going to divide the day roughly in half, spending the morning confronting the broader conceptual issues and the afternoon discussing what kind of concrete intervention might make sense.
Questions to think about in preparation for the morning discussion:
* if it’s assumed that form and content are inextricably linked, what happens when we take a book and render it on a dynamic electronic screen rather than bound paper? same question for movies which move from the large theatrical presentation to the intimacy of the personal screen. interestingly the “old” analog forms aren’t as singular as they might seem. books are read silently alone or out loud in public; music is played live and listened to on recordings. a recording of a Beethoven symphony on ten 78rpm discs presents quite a different experience than listening to it on an iPod with random access. from this perspective how do we define the essence of a work which needs to be respected and protected in the act of re-presentation?
* twenty years ago we added audio commentary tracks to movies and textual commentary to music. given the spectacular advances in computing power, what are the most compelling enhancements we might imagine. (in preparation for this, you may find it useful to look at a series of exchanges that took place on the if:book blog regarding an “ideal presentation of Ulysses” (here and here).
* what are the affordances of locating a work in the shared social space of a digital network. what is the value of putting readers, viewers, and listeners of specific works in touch with each other. what can we imagine about the range of interactions that are possible and worthwhile. be expansive here, extrapolating as far out as possible from current technical possibilities.
* it seems to us that visualization tools will be crucial in the digital future both for opening up analog works in new ways and for browsing and making sense of massive media archives. if everything is theoretically connected to everything else, how do we make those connections visible in a way that illuminates rather than overwhelms? and how do we visualize the additional and sometimes transformative connections that people make individually and communally around works? how do we visualize the conversation that emerges?
* in the digital environment, all media break down into ones and zeros. all media can be experienced on a single device: a computer. what are the implications of this? what are the challenges in keeping historical differences between media forms in perspective as digitization melts everything together?
* what happens when computers can start reading all the records of human civilization? in other words, when all analog media are digitized, what kind of advanced data crunching can we do and what sorts of things might it reveal?
* most analog works were intended to be experienced with all of one’s attention, but the way we read/watch/listen/look is changing. even when engaging with non-networked media -? a paper book, a print newspaper, a compact disc, a DVD, a collection of photos -? we increasingly find ourselves Googling alongside. Al Pacino paces outside the bank in ‘dog day afternoon’ firing up the crowded street with “Attica! Attica!” I flip to Wikipedia and do quick read on the Attica prison riots. reading “song of myself” in “leaves of grass,” i find my way to the online Whitman archive, which allows me to compare every iteration of Whitman’s evolutionary work. or reading “ulysses” i open up Google Earth and retrace Bloom’s steps by satellite. while leafing through a book of caravaggio’s paintings, a quick google video search leads me to a related episode in simon schama’s “power of art” documentary series and a series of online essays. as radiohead’s new album plays, i browse fan sites and blogs for backstory, b-sides and touring info. the immediacy and proximity of such supplementary resources changes our relationship to the primary ones. the ratio of text to context is shifting. how should this influence the structure and design of future digital editions?
Afternoon questions:
* if we do decide to mount a competition (we’re still far from decided on whether this is the right approach), how exactly should it work? first off, what are we judging? what are we hoping to reward? what is the structure of this contest? what are the motivators? a big concern is that the top-down model -? panel of prestigious judges, serious prize money etc. -? feels very old-fashioned and ignores the way in which much of the recent innovation in digital media has taken place: an emergent, grassroots ferment… open source culture, web2.0, or what have you. how can we combine the heft and focused energy of the former with the looseness and dynamism of the latter? is there a way to achieve some sort of top-down orchestration of emergent creativity? is “competition” maybe the wrong word? and how do we create a meaningful public forum that can raise consciousness of these issues more generally? an accompanying website? some other kind of publication? public events? a conference?
* where are the leverage points are for an intervention in this area? what are the key consituencies, national vs. international?
* for reasons both practical and political, we’ve considered restricting this contest to the public domain. practical in that the public domain provides an unencumbered test bed of creative content for contributors to work with (no copyright hassles). political in that we wish to draw attention to the threat posed to the public domain by commercially driven digitization projects ( i.e. the recent spate of deals between Google and libraries, the National Archives’ deal with Footnote.com and Amazon, the Smithsonian with Showtime etc.). emphasizing the public domain could also exert pressure on the media industries, who to date have been more concerned with preserving old architectures of revenue than with adapting creatively to the digital age. making the public domain more attractive, more dynamic and more *usable* than the private domain could serve as a wake-up call to the big media incumbents, and more importantly, to contemporary artists and scholars whose work is being shackled by overly restrictive formats and antiquated business models. we’d also consider workable areas of the private domain such as the Creative Commons -? works that are progressively licensed so as to allow creative reuse. we’re not necessarily wedded to this idea. what do you think?

candida höfer: the library as museum

The photographs of libraries in “Portugal,” the current exhibition of Candida Höfer at Sonnabend, show libraries as venerable places where precious objects are stored.
hoferlibrary1.jpg
The large format that characterizes Höfer’s photographs of public places, the absence of people, and the angle from which she composes them, invite the viewer “to enter” the rooms and observe. Photography is a silent medium and in Höfer’s libraries this is magnified, creating that feeling of “temple of learning” with which libraries have often been identified. On the other hand, the meticulous attention to detail, hand-painted porcelain markers, ornately carved bookcases, murals, stained glass windows, gilt moldings, and precious tomes are an eloquent representation of libraries as palaces of learning for the privileged. In spite of that, and ever since libraries became public spaces, anyone, in theory, has access to books and the concept of gain or monetary value rarely enters the user’s mind.
hoferlibrary2.jpg
Libraries are a book lover’s paradise, a physical compilation of human knowledge in all its labyrinthine intricacy. With digitization, libraries gain storage capacity and readers gain accessibility, but they lose both silence and awe. Even though in the digital context, the basic concept of the library as a place for the preservation of memory remains, for many “enlightened” readers the realization that human memory and knowledge are handled by for-profit enterprises such as Google, produces a feeling of merchants in the temple, a sense that the public interest has fallen, one more time, into private hands.
As we well know, the truly interesting development in the shift from print to digital is the networked environment and its effects on reading and writing. If, as Umberto Eco says, books “are machines that provoke further thoughts” then the born-digital book is a step toward the open text, and the “library” that eventually will hold it, a bird of different feather.

visual search

I just came across oSkope, a snazzy new “visual search assistant” built by a Zurich/Berlin outfit that allows you to graphically browse items on Amazon, ebay, Flickr or YouTube. More than a demo or prototype, it’s a fully functioning front end to the search engines of the afore-mentioned sites. I played around a bit in Amazon mode… below are some screenshots of a search for “Kafka” in Amazon’s book category. Each search cluster can be displayed in five different configurations (grid, stack, pile, list and graph), re-scaled with a slide bar, or rearranged manually by dragging items around. Click any cover and a small info window pops up with a link to the Amazon page. You can also drag items down into a folder for future reference. Very smooth, very tactile.
Grid:
oskopegrid.jpg
Stack:
oskopestack.jpg
Pile:
oskopepile.jpg
List:
oskopelist.jpg
Graph (arranges items along axes of price and sales rank):
oskopegraph.jpg
A few months back I linked to another visual Amazon browser from TouchGraph that arranges book clusters according to customer purchase patterns (the “people who purchased this also bought…”). I’m still waiting for someone to visualize the connections in the citation indexes: create a cross-referential map that shows the ligatures between texts (as pondered here). Each of these ideas is of course just an incremental step toward more advanced methods of getting the “big picture” view of digital collections.
oSkope, though it could still use some work (Flickr searching was unpredictable and didn’t seem to turn up nearly as much as what I’m sure is in their system, Ebay wasn’t working at all), is a relatively straightforward and useful contribution – ?more than just eye candy. It even helped me stumble upon something wonderful: a recently published study (appropriately, visual) of Kafka, a collab between comic artist R. Crumb and Kafka scholar David Mairowitz.
Browsing graphically is often more engaging than scanning a long list of results, and a crop of new tools – ?LibraryThing, Shelfari, Delicious Library, and now Google Books – ?have recently emerged to address this, all riffing in similar, somewhat nostalgic ways on the experience of shelves (Peter Brantley just blogged another idea in this vein). iTunes too has gotten in on this, its album cover flipper becoming a popular way to sift through one’s music collection.
Perhaps it’s telling, though, that these visual, shelf-inspired browsing tools are focused on old media: books, albums… all bounded objects. You couldn’t simply graft this onto web search and get the same effect (although page previews, of the sort that Snap provides, are becoming increasingly popular). For vast, shifting collections of unbounded, evolving, recombining, and in many cases ephemeral media, different vizualization tools are most likely needed. What might those be?
(oSkope link via Information Aesthetics)

siva podcast on the googlization of libraries

We’re just a couple of days away from launching what promises to be one of our most important projects to date, The Googlization of Everything, a weblog where Siva Vaidhyanathan (who’s a fellow here) will publicly develop his new book, a major critical examination of the Google behemoth. As an appetizer, check out this First Monday podcast conversation with Siva on the subject of Google’s book activities (mp3, transcript).
An excerpt:

Q: So what’s the alternative? Who are the major players, what are the major policy points?
SIVA: I think this is an important enough project where we need to have a nationwide effort. We have to have a publicly funded effort. Guided, perhaps led by the Library of Congress, certainly a consortium of public university libraries could do just as well to do it.
We’re willing to do these sorts of big projects in the sciences. Look at how individual states are rallying billions of dollars to fund stem cell research right now. Look at the ways the United States government, the French government, the Japanese government rallied billions of dollars for the Human Genome Project out of concern that all that essential information was going to be privatized and served in an inefficient and unwieldy way.
So those are the models that I would like to see us pursue. What saddens me about Google’s initiative, is that it’s let so many people off the hook. Essentially we’ve seen so many people say, “Great now we don’t have to do the digital library projects we were planning to do.” And many of these libraries involved in the Google project were in the process of producing their own digital libraries. We don’t have to do that any more because Google will do it for us. We don’t have to worry about things like quality because Google will take care of the quantity.
And so what I would like to see? I would like to see all the major public universities, public research universities, in the country gather together and raise the money or persuade Congress to deliver the money to do this sort of thing because it’s in the public interest, not because it’s in Google’s interest. If it really is this important we should be able to mount a public campaign, a set of arguments and convince the people with the purse strings that this should be done right.

monkeybook3: the desk set

monkeytownsketch.jpg Monkeybook is an occasional series of new media evenings hosted by the Institute for the Future of the Book at Monkey Town, Brooklyn’s premier video salon and A/V sandbox.
Monkeybook 3 (this coming Monday, Aug 27, reservation info below) is presented jointly with the Desk Set, a delightful network of young New York librarians, archivists and weird book people that periodically gets sloshed in northern Brooklyn. (If this rings a bell, it may be because you read about them in the Times last month – ?somehow this became one of the “most emailed” NYT articles of recent months).
I first collided with the Desk Set at a talk I gave at Brooklyn College this June. We resolved to get together later in the summer and eventually this crazy event materialized. The crowd will be primarily librarians and folks from the Desk Set community. Dan and I will be masters of ceremonies, presenting the Institute’s projects and then spinning an eclectic assortment of films and other goodies (including some stuff by Alex Itin, who was the star of Monkeybook 1). It’ll basically be a big librarian party. How can you resist?
Oh, and how cool: we’re an editors’ pick on Going.com!
Monkey Town reservations: http://www.monkeytownhq.com/reservations.html (book soon, it’s already pretty packed!)
Monkey Town: 58 N. 3rd St. (between Wythe and Kent), Williamsburg, Brooklyn

View Larger Map

“the bookish character of books”: how google’s romanticism falls short

tristramgbs.gif
Check out, if you haven’t already, Paul Duguid’s witty and incisive exposé of the pitfalls of searching for Tristram Shandy in Google Book Search, an exercise which puts many of the inadequacies of the world’s leading digitization program into relief. By Duguid’s own admission, Lawrence Sterne’s legendary experimental novel is an idiosyncratic choice, but its many typographic and structural oddities make it a particularly useful lens through which to examine the challenges of migrating books successfully to the digital domain. This follows a similar examination Duguid carried out last year with the same text in Project Gutenberg, an experience which he said revealed the limitations of peer production in generating high quality digital editions (also see Dan’s own take on this in an older if:book post). This study focuses on the problems of inheritance as a mode of quality assurance, in this case the bequeathing of large authoritative collections by elite institutions to the Google digitization enterprise. Does simply digitizing these – ?books, imprimaturs and all – ?automatically result in an authoritative bibliographic resource?
Duguid’s suggests not. The process of migrating analog works to the digital environment in a way that respects the orginals but fully integrates them into the networked world is trickier than simply scanning and dumping into a database. The Shandy study shows in detail how Google’s ambition to organizing the world’s books and making them universally accessible and useful (to slightly adapt Google’s mission statement) is being carried out in a hasty, slipshod manner, leading to a serious deficit in quality in what could eventually become, for better or worse, the world’s library. Duguid is hardly the first to point this out, but the intense focus of his case study is valuable and serves as a useful counterpoint to the technoromantic visions of Google boosters such as Kevin Kelly, who predict a new electronic book culture liberated by search engines in which readers are free to find, remix and recombine texts in various ways. While this networked bibliotopia sounds attractive, it’s conceived primarily from the standpoint of technology and not well grounded in the particulars of books. What works as snappy Web2.0 buzz doesn’t necessarily hold up in practice.
As is so often the case, the devil is in the details, and it is precisely the details that Google seems to have overlooked, or rather sprinted past. Sloppy scanning and the blithe discarding of organizational and metadata schemes meticulously devised through centuries of librarianship, might indeed make the books “universally accessible” (or close to that) but the “and useful” part of the equation could go unrealized. As we build the future, it’s worth pondering what parts of the past we want to hold on to. It’s going to have to be a slower and more painstaking a process than Google (and, ironically, the partner libraries who have rushed headlong into these deals) might be prepared to undertake. Duguid:

The Google Books Project is no doubt an important, in many ways invaluable, project. It is also, on the brief evidence given here, a highly problematic one. Relying on the power of its search tools, Google has ignored elemental metadata, such as volume numbers. The quality of its scanning (and so we may presume its searching) is at times completely inadequate. The editions offered (by search or by sale) are, at best, regrettable. Curiously, this suggests to me that it may be Google’s technicians, and not librarians, who are the great romanticisers of the book. Google Books takes books as a storehouse of wisdom to be opened up with new tools. They fail to see what librarians know: books can be obtuse, obdurate, even obnoxious things. As a group, they don’t submit equally to a standard shelf, a standard scanner, or a standard ontology. Nor are their constraints overcome by scraping the text and developing search algorithms. Such strategies can undoubtedly be helpful, but in trying to do away with fairly simple constraints (like volumes), these strategies underestimate how a book’s rigidities are often simultaneously resources deeply implicated in the ways in which authors and publishers sought to create the content, meaning, and significance that Google now seeks to liberate. Even with some of the best search and scanning technology in the world behind you, it is unwise to ignore the bookish character of books. More generally, transferring any complex communicative artifacts between generations of technology is always likely to be more problematic than automatic.

Also take a look at Peter Brantley’s thoughts on Duguid:

Ultimately, whether or not Google Book Search is a useful tool will hinge in no small part on the ability of its engineers to provoke among themselves a more thorough, and less alchemic, appreciation for the materials they are attempting to transmute from paper to gold.

cornell joins google book search

…offering up to 500,000 items for digitization. From the Cornell library site:

Cornell is the 27th institution to join the Google Book Search Library Project, which digitizes books from major libraries and makes it possible for Internet users to search their collections online. Over the next six years, Cornell will provide Google with public domain and copyrighted holdings from its collections. If a work has no copyright restrictions, the full text will be available for online viewing. For books protected by copyright, users will just get the basic background (such as the book’s title and the author’s name), at most a few lines of text related to their search and information about where they can buy or borrow a book. Cornell University Library will work with Google to choose materials that complement the contributions of the project’s other partners. In addition to making the materials available through its online search service, Google will also provide Cornell with a digital copy of all the materials scanned, which will eventually be incorporated into the university’s own digital library.

audiovisual heritage double play

Two major preservation and access initiatives just reported by Peter Brantley over at O’Reilly Radar (1 and 2):
1. Reframe (set to launch in September ’07)

The Reframe project is a new initiative of Renew Media in partnership with Amazon and with major support from the John D. & Catherine T. MacArthur Foundation, which promises to offer exciting solutions for the dissemination of important media arts and the preservation and accessibility of our visual heritage.
The Reframe project will help connect audiences of independent media to a robust collection of media arts via an integrated, resourceful website. Reframe will aggregate content from individual filmmakers, broadcasters, distributors, public media resources, archives, libraries and other sources of independent and alternative media. Serving as a both an aggregator of content and a powerful marketing tool, Reframe enables content-holders to digitize, disseminate and make available their content to a vast potential audience via a powerful online resource.
Renew Media will create a specialized Reframe website, which will interact with the Amazon storefront, to assist institutions (universities, libraries or museums) and consumers of niche content in browsing, finding, purchasing or renting Reframe content. Reframe website visitors will find it easy to locate relevant content through a rich menu of search and retrieval tools, including conventional search, recommender systems, social networking tools and curated lists. Reframe will allow individual viewers to rate and discuss the films they have seen and to sort titles according to their popularity among users with similar interests.

2. Library of Congress awards to preserve digitized and born-digital works

The Library of Congress, through its National Digital Information Infrastructure and Preservation Program (NDIIPP), today announced eight partnerships as part of its new Preserving Creative America initiative to address the long-term preservation of creative content in digital form. These partners will target preservation issues across a broad range of creative works, including digital photographs, cartoons, motion pictures, sound recordings and even video games. The work will be conducted by a combination of industry trade associations, private sector companies and nonprofits, as well as cultural heritage institutions.
Several of the projects will involve developing standardized approaches to content formats and metadata (the information that makes electronic content discoverable by search engines), which are expected to increase greatly the chances that the digital content of today will survive to become America’s cultural patrimony tomorrow. Although many of the creative content industries have begun to look seriously at what will be needed to sustain digital content over time, the $2.15 million being awarded to the Preserving Creative America projects will provide added impetus for collaborations within and across industries, as well as with libraries and archives.

Partners include the Academy of Motion Picture Arts and Sciences, the American Society of Media Photographers, ARTstor and others. Go here and scroll down part way to see the full list.
One project that caught my and Peter’s eye is an effort by the University of Illinois at Urbana-Champaign to address a particularly vexing problem: how to preserve virtual environments and other complex interactive media:

Interactive media are highly complex and at high risk for loss as technologies rapidly become obsolete. The Preserving Virtual Worlds project will explore methods for preserving digital games and interactive fiction. Major activities will include developing basic standards for metadata and content representation and conducting a series of archiving case studies for early video games, electronic literature and Second Life, an interactive multiplayer game. Second Life content participants include Life to the Second Power, Democracy Island and the International Spaceflight Museum. Partners: University of Maryland, Stanford University, Rochester Institute of Technology and Linden Lab.