Category Archives: Uncategorized

and now we have an ipad

The iPad has arrived, to no one’s surprise: as soon as you use an iPhone, you start wondering what a computer-sized version of the same would be like. (Those interested in how past predictions look now might look at this post by Ben from five years ago.) The iPad is an attractive device and at $500, it seems likely to take off. It seems entirely possible that a tablet could replace laptops and desktops for many computers, to say nothing of Kindles and Nooks. My MacBook Pro suddenly feels rickety. Hardware-wise, it feels like the iPad might finally be Alan Kay’s Dynabook.
And yet: standing on the verge of a potential transformation in how people use computers, I think it’s worth stepping back for a second to think about where we are. I suggest that now might be a useful time to re-read Neal Stephenson’s manifesto, “In the Beginning Was the Command Line”. This is, it needs to be said, a dated piece of writing, as Stephenson has admitted; this is the perpetual curse of writing about technology. Stephenson was writing in 1999, when Microsoft’s monopoly over the computer seemed to be without limit; Apple was then in an interregnum, and Google and Amazon were promising web players in a sea of many other promising web players. “In the Beginning Was the Command Line,” however, is still worth reading because of his understanding of how we use computers. The promise of the open source movement that Stephenson described was that it would give users complete control over their computers: using Linux, you weren’t tied into how corporations thought your computer could be used but were free to change it as you desired. And there’s a deeper question that Stephenson gets at: the problem of how we understand our tools, if at all. We can theoretically understand how an open system works; but a closed system is opaque. Something magic is happening under the hood.
Things have changed since then: the Linux desktop never really took off in the way that seemed possible in 1999. Corporations – Apple and Google – showed that they could use open source in a tremendously profitable way. What made me think about this essay yesterday, however, was the iPad: Apple has created a computer that’s entirely locked down. The only applications that will run on the iPad are those that have been approved by Apple. And this is one of the first computers where the user will be entirely unable to access the file system. I understand why this is possible from a design standpoint: file systems are arcane things, and most people don’t understand them or want to understand them. But this means that Apple has a complete lock on how media gets into your iPad: you’re tied into an Apple-approved mechanism. The user of the iPad, like the user of the iPhone, is directly tied into the Apple economy: your credit card on file with Apple not only lets you buy apps and media, but it will also allow you to buy internet connectivity.
It’s simple – it’s fantastically simple, and it will probably work. But I can’t help but think of how Stephenson metaphorically equates the closed system of Windows 95 to a car with the hood welded shut: you can’t get inside it. Apple’s managed this on a scale that late 90s Microsoft could only dream of. I wonder as well what this means for our understanding of technology: maybe technology’s become something we let others understand for us.

how discourse on the web works

Good weekend reading: Jonathan Dee’s examination of the fall from grace of Charles Johnson’s Little Green Footballs. Internecine fighting on the right isn’t inherently interesting; however, Dee’s piece is as much about how we think now. A few samples:

That concept of the link, in all its permutations, is the key to what happened next, both to Johnson and because of him, and it says something enlightening not just about blogging but also about the nature and prospects of citizen journalism. Whatever you think of him, Johnson is a smart man, a gifted synthesizer of information gathered by other people. But just as for anyone in his position, there is an inevitable limit to what he can learn about places, people, political organizations, etc., without actually encountering them. Instead of causes and effects, motivations and consequences, observation and behavior, his means of intellectual synthesis is, instead, the link: the indiscriminate connection established via search engine.

Johnson evidently fell out with the right-wing blogosphere when he realized that the right wing he was associating with was shading into the historical right wing; in his case, he realized that Vlaams Belang, a Flemish nationalist party in Belgium, weren’t the most savory of consorts:

Regardless of whether Johnson’s view of Vlaams Belang is correct, it is notable that the party is defined for him entirely by the trail it has left on the Internet. This isn’t necessarily unfair – a speech, say, given by Dewinter isn’t any more or less valuable as evidence of his political positions depending on whether you read it (or watch it) on a screen or listen to it in a crowd – but it does have a certain flattening effect in terms of time: that hypothetical speech exists on the Internet in exactly the same way whether it was delivered in 2007 or 1997. The speaker will never put it behind him. (Just as Johnson, despite his very reasonable contention that he later changed his mind, will never be allowed to consign to the past a blog post he wrote in 2004 criticizing that judicial condemnation of Vlaams Belang as “a victory for European Islamic supremacist groups.”) It may be difficult to travel to Belgium and build the case that Filip Dewinter is not just a hateful character but an actual Nazi (and thus that those who can be linked to him are Nazi sympathizers), but sitting at your keyboard, there is no trick to it at all. Not only can the past never really be erased; it co-exists, in cyberspace, with the present, and an important type of context is destroyed. This is one reason that intellectual inflexibility has become such a hallmark of modern political discourse, and why, so often, no distinction is recognized between hypocrisy and changing your mind.

Dee does a fine job of examining exactly how the Internet has changed discourse with political ramifications: it’s a long piece, but it’s worth reading in full.

reading vs writing

Ted Genoways, the editor of the Virginia Quarterly Review, has an essay up at Mother Jones with the alarmist title “The Death of Fiction?”: he points out, to the surprise of nobody, I expect, that the magazine component of the fiction industry is in bad shape right now. He examines the systemic failure that brought us here: part of the problem is the over-supply of reading. The way that we find interesting writing has changed, and we can more easily find interesting content without reading literary journals than was possible twenty years ago. Another is the over-supply of writers: over the past two decades universities have rightly seen adding MFA programs as cash cows, as most students pay full price. When creative writing programs produce, as he suggests, 60,000 new writers a decade, this has the added benefit for the universities of creating a steady stream of writing instructors willing to serve as adjuncts; a huge supply of competition means that they don’t need to be paid very much. There’s a labor problem here: the universities have given their students the misleading idea that writing fiction can be a sustainable career when they have a better chance of supporting themselves by buying scratch tickets. It’s an unfortunate situation; when this is combined with the decline of paying outlets for fiction, it’s easy to project, as he suggests, the death of writing fiction as a paid pursuit.

It’s worth reading this in conjunction with a post on the Virginia Quarterly Review‘s blog, which states the problem more baldly, pointing out an imbalance between readers and writers:

Here at VQR we currently have more than ten times as many submitters each year as we have subscribers. And there’s very, very little overlap. We know–we’ve checked. So there’s an ever-growing number of people writing and submitting fiction, but there’s an ever-dwindling number of people reading the best journals that publish it.

We tend to assume that there’s an economic balance between reading and writing: that reading must necessarily pay for writing, as credits must balance debits in double-entry bookkeeping. Certainly this is what the author of the VQR blog post is doing by suggesting that it’s unnatural for the number of submitters to exceed the number of subscribers. It’s hard to fault them for stating this relationship in financial terms: it’s in a publisher’s interest for a publication to make money. But this is a assumption that’s worth unpacking.

This is a model nominally worked in the print era: ads and subscriber revenue paid enough that journals could be printed and that writers could be paid. This model, however, doesn’t scale to the web: web advertising works for pages that appeal to the lowest common denominator (celebrity gossip, mesothelioma) and can thus attract huge numbers of page views and advertising dollars. It won’t work for something that’s only going to attract 2000 viewers, no matter how influential those visitors are. The economics of the web (as it presently functions) favor linkbait rather than quality: the incentives are different than we had in a print-centered world. And realistically: it’s hard for readers to justify paying for content in a world suffused with free content.

Yet, as the VQR‘s submission numbers indicate, we have an urge to believe that we might be the exception. Writers will write, and hope to get paid for it; like lottery players, they know the odds aren’t in their favor, but they imagine that they might get lucky, that they’re outside the same economic processes in which they participate as readers. State lotteries do have the virtue of funding education; it’s hard to know what good comes of this.

The problem here is that we tend to think of cultural production in economic terms. Historically, the book has been terrific at spreading ideas; however, as a vehicle for getting creators paid, its record isn’t the best. An example: if I go into a book store and buy a copy of Tristram Shandy, we can be sure that Laurence Sterne isn’t making any money off the deal because he’s dead; if I buy a book of T. S. Eliot’s, he won’t get anything as he’s been dead since 1965, but his ancient widow might be paid; if I buy a book by a living author, there’s the off chance that she might be paid, but not if I’m buying it used. In all these cases, my objective as a reader might be fulfilled: I get the book that I want, and, if I’m lucky, I might get something out of that book. The person that I bought the book from is presumably happy because they have my money; if I bought the book new, the publisher is happy because they’re getting money. The author, however, may well be left out: my economic transaction has not been with the author.

There are ways around this: we can, for example, see it as a moral duty to buy books by authors who are still alive and who deserve money new, rather than used. We could buy books directly from authors whenever possible so that they’re getting a more just cut. We need to re-conceptualize how we think about exchange and consumption. Lewis Hyde’s The Gift presents one such way forward: thinking about artistic creation as something outside the economic. But that requires us to think different both as producers and consumers: maybe that’s what the Internet is trying to tell us.

is Google good for history?

Following are Dan Cohen’s prepared remarks for a talk at the American Historical Association Annual Meeting, on January 7, 2010, in San Diego. The panel was entitled “Is Google Good for History?” and also featured talks by Paul Duguid of the University of California, Berkeley and Brandon Badger of Google Books. This is cross-posted from Dan’s blog
Is Google good for history? Of course it is. We historians are searchers and sifters of evidence. Google is probably the most powerful tool in human history for doing just that. It has constructed a deceptively simple way to scan billions of documents instantaneously, and it has spent hundreds of millions of dollars of its own money to allow us to read millions of books in our pajamas. Good? How about Great?
But then we historians, like other humanities scholars, are natural-born critics. We can find fault with virtually anything. And this disposition is unsurprisingly exacerbated when a large company, consisting mostly of better-paid graduates from the other side of campus, muscles into our turf. Had Google spent hundreds of millions of dollars to build the Widener Library at Harvard, surely we would have complained about all those steps up to the front entrance.
Partly out of fear and partly out of envy, it’s easy to take shots at Google. While it seems that an obsessive book about Google comes out every other week, where are the volumes of criticism of ProQuest or Elsevier or other large information companies that serve the academic market in troubling ways? These companies, which also provide search services and digital scans, charge universities exorbitant amounts for the privilege of access. They leech money out of library budgets every year that could be going to other, more productive uses.
Google, on the other hand, has given us Google Scholar, Google Books, newspaper archives, and more, often besting commercial offerings while being freely accessible. In this bigger picture, away from the myopic obsession with the Biggest Tech Company of the Moment (remember similar diatribes against IBM, Microsoft?), Google has been very good for history and historians, and one can only hope that they continue to exert pressure on those who provide costly alternatives.
Of course, like many others who feel a special bond with books and our cultural heritage, I wish that the Google Books project was not under the control of a private entity. For years I have called for a public project, or at least a university consortium, to scan books on the scale Google is attempting. I’m envious of France’s recent announcement to spend a billion dollars on public scanning. In addition, the Center for History and New Media has a strong relationship with the Internet Archive to put content in a non-profit environment that will maximize its utility and distribution and make that content truly free, in all senses of the word. I would much rather see Google’s books at the Internet Archive or the Library of Congress. There is some hope that HathiTrust will be this non-Google champion, but they are still relying mostly on Google’s scans. The likelihood of a publicly funded scanning project in the age of Tea Party reactionaries is slim.

* * *

Long-time readers of my blog know that I have not pulled punches when it comes to Google. To this day the biggest spike in readership on my blog was when, very early in Google’s book scanning project, I casually posted a scan of a human hand I found while looking at an edition of Plato. The post ended up on Digg, and since then it has been one of the many examples used by Google’s detractors to show a lack of quality in their library project.
Let’s discuss the quality issues for a moment, since it is one point of obsession within the academy, an obsession I feel is slightly misplaced. Of course Google has some poor scans–as the saying goes, haste makes waste–but I’ve yet to see a scientific survey of the overall percentage of pages that are unreadable or missing (surely a miniscule fraction in my viewing of scores of Victorian books). Regarding metadata errors, as Jon Orwant of Google Books has noted, when you are dealing with a trillion pieces of metadata, who are likely to have millions of errors in need of correction. Let us also not pretend the bibliographical world beyond Google is perfect. Many of the metadata problems with Google Books come from library partners and others outside of Google.
Moreover, Google likely has remedies for many of these inadequacies. Google is constantly improving its OCR and metadata correction capabilities, often in clever ways. For instance, it recently acquired the reCAPTCHA system from Carnegie Mellon, which uses unwitting humans who are logging into online services to transcribe particularly hard or smudged words from old books. They have added a feedback mechanism for users to report poor scans. Truly bad books can be rescanned or replaced by other libraries’ versions. I find myself nonplussed by quality complaints about Google Books that have engineering solutions. That’s what Google does; it solves engineering problems very well.
Indeed, we should recognize (and not without criticism, as I will note momentarily) that at its heart, Google Books is the outcome, like so many things at Google, of a engineering challenge and a series of mathematical problems: How can you scan tens of million books in a decade? It’s easy to say they should do a better job and get all the details right, but if you do the calculations with those key variables, as I assume Brandon and his team have done, you’ll probably see that getting a nearly perfect library scanning project would take a hundred years rather than ten. (That might be a perfectly fine trade-off, but that’s a different argument or a different project.) As in OCR, getting from 99% to 99.9% accuracy would probably take an order of magnitude longer and be an order of magnitude more expensive. That’s the trade-off they have decided to make, and as a company interested in search, where near-100% accuracy is unnecessary, and considering the possibilities for iterating toward perfection from an imperfect first version, it must have been an easy decision to make.

* * *

Google Books is incredibly useful, even with the flaws. Although I was trained at places with large research libraries of Google Books scale, I’m now at an institution that is far more typical of higher ed, with a mere million volumes and few rare works. At places like Mason, Google Books is a savior, enabling research that could once only be done if you got into the right places. I regularly have students discover new topics to study and write about through searches on Google Books. You can only imagine how historical researchers and all students and scholars feel in even less privileged places. Despite its flaws, it will be the the source of much historical scholarship, from around the globe, over the coming decades. It is a tremendous leveler of access to historical resources.
Google is also good for history in that it challenges age-old assumptions about the way we have done history. Before the dawn of massive digitization projects and their equally important indices, we necessarily had to pick and choose from a sea of analog documents. All of that searching and sifting we did, and the particular documents and evidence we chose to write on, were–let’s admit it–prone to many errors. Read it all, we were told in graduate school. But who ever does? We sift through large archives based on intuition; sometime we even find important evidence by sheer luck. We have sometimes made mountains out of molehills because, well, we only have time to sift through molehills, not mountains. Regardless of our technique, we always leave something out; in an analog world we have rarely been comprehensive.
This widespread problem of anecdotal history, as I have called it, will only get worse. As more documents are scanned and go online, many works of historical scholarship will be exposed as flimsy and haphazard. The existence of modern search technology should push us to improve historical research. It should tell us that our analog, necessarily partial methods have had hidden from us the potential of taking a more comprehensive view, aided by less capricious retrieval mechanisms which, despite what detractors might say, are often more objective than leafing rapidly through paper folios on a time-delimited jaunt to an archive.
In addition, listening to Google may open up new avenues of exploring the past. In my book Equations from God I argued that mathematics was generally considered a divine language in 1800 but was “secularized” in the nineteenth century. Part of my evidence was that mathematical treatises, which often contained religious language in the early nineteenth century, lost such language by the end of the century. By necessity, researching in the pre-Google Books era, my textual evidence was limited–I could only read a certain number of treatises and chose to focus (I’m sure this will sound familiar) on the writings of high-profile mathematicians. The vastness of Google Books for the first time presents the opportunity to do a more comprehensive scan of Victorian mathematical writing for evidence of religious language. This holds true for many historical research projects.
So Google has provided us not only with free research riches but also with a helpful direct challenge to our research methods, for which we should be grateful. Is Google good for history? Of course it is.

* * *

But does that mean that we cannot provide constructive criticism of Google, to make it the best it can be, especially for historians? Of course not. I would like to focus on one serious issue that ripples through many parts of Google Books.
For a company that is a champion of openness, Google remains strangely closed when it comes to Google Books. Google Books seems to operate in ways that are very different from other Google properties, where Google aims to give it all away. For instance, I cannot understand why Google doesn’t make it easier for historians such as myself, who want to do technical analyses of historical books, to download them en masse more easily. If it wanted to, Google could make a portal to download all public domain books tomorrow. I’ve heard the excuses from Googlers: But we’ve spent millions to digitize these books! We’re not going to just give them away! Well, Google has also spent millions on software projects such as Android, Wave, Chrome OS, and the Chrome browser, and they are giving those away. Google’s hesitance with regard to its books project shows that openness goes only so far at Google. I suppose we should understand that; Google is a company, not public library. But that’s not the philanthropic aura they cast around Google Books at its inception or even today, in dramatic op-eds touting the social benefit of Google Books.
In short, complaining about the quality of Google’s scans distracts us from a much larger problem with Google Books. The real problem–especially for those in the digital humanities but increasingly for many others–is that Google Books is only open in the read-a-book-in-my-pajamas way. To be sure, you can download PDFs of many public domain books. But they make it difficult to download the OCRed text from multiple public domain books-what you would need for more sophisticated historical research. And when we move beyond the public domain, Google has pushed for a troubling, restrictive regime for millions of so-called “orphan” books.
I would like to see a settlement that offers greater, not lesser access to those works, in addition to greater availability of what Cliff Lynch has called “computational access” to Google Books, a higher level of access that is less about reading a page image on your computer than applying digital tools to many pages or books at one time to create new knowledge and understanding. This is partially promised in the Google Books settlement, in the form of text-mining research centers, but those centers will be behind a velvet rope and I suspect the casual historian will be unlikely to ever use them. Google has elaborate APIs, or application programming interfaces, for most of its services, yet only the most superficial access to Google Books.
For a company that thrives on openness and the empowerment of users and software developers, Google Books is a puzzlement. With much fanfare, Google has recently launched–evidently out of internal agitation–what it calls a “Data Liberation Front,” to ensure portability of data and openness throughout Google. On dataliberation.org, the website for the front, these Googlers list 25 Google projects and how to maximize their portability and openness–virtually all of the main services at Google. Sadly, Google Books is nowhere to be seen, even though it also includes user-created data, such as the My Library feature, not to mention all of the data–that is, books–that we have all paid for with our tax dollars and tuition. So while the Che Guevaras put up their revolutionary fist on one side of the Googleplex, their colleagues on the other side are working with a circumscribed group of authors and publishers to place messy restrictions onto large swaths of our cultural heritage through a settlement that few in the academy support.
Jon Orwant and Dan Clancy and Brandon Badger have done an admirable job explaining much of the internal process of Google Books. But it still feels removed and alien in way that other Google efforts are not. That is partly because they are lawyered up, and thus hamstrung from responding to some questions academics have, or from instituting more liberal policies and features. The same chutzpah that would lead a company to digitize entire libraries also led it to go too far with in-copyright books, leading to a breakdown with authors and publishers and the flawed settlement we have in front of us today.
We should remember that the reason we are in a settlement now is that Google didn’t have enough chutzpah to take the higher, tougher road–a direct challenge in the courts, the court of public opinion, or the Congress to the intellectual property regime that governs many books and makes them difficult to bring online, even though their authors and publishers are long gone. While Google regularly uses its power to alter markets radically, it has been uncharacteristically meek in attacking head-on this intellectual property tower and its powerful corporate defenders. Had Google taken a stronger stance, historians would have likely been fully behind their efforts, since we too face the annoyances that unbalanced copyright law places on our pedagogical and scholarly use of textual, visual, audio, and video evidence.
I would much rather have historians and Google to work together. While Google as a research tool challenges our traditional historical methods, historians may very well have the ability to challenge and make better what Google does. Historical and humanistic questions are often at the high end of complexity among the engineering challenges Google faces, similar to and even beyond, for instance, machine translation, and Google engineers might learn a great deal from our scholarly practice. Google’s algorithms have been optimized over the last decade to search through the hyperlinked documents of the Web. But those same algorithms falter when faced with the odd challenges of change over centuries and the alienness of the past and old books and documents that historians examine daily.
Because Google Books is the product of engineers, with tremendous talent in computer science but less sense of the history of the book or the book as an object rather than bits, it founders in many respects. Google still has no decent sense of how to rank search results in humanities corpora. Bibliometrics and text mining work poorly on these sources (as opposed to, say, the highly structured scientific papers Google Scholar specializes in). Studying how professional historians rank and sort primary and seconary sources might tell Google a lot, which it could use in turn to help scholars.
Ultimately, the interesting question might not be, Is Google good for history? It might be: Is history good for Google? To both questions, my answer is: Yes.

the other side of the long tail

Jace Clayton quotes an article from The Economist:

A lot of the people who read a bestselling novel, for example, do not read much other fiction. By contrast, the audience for an obscure novel is largely composed of people who read a lot. That means the least popular books are judged by people who have the highest standards, while the most popular are judged by people who literally do not know any better. An American who read just one book this year was disproportionately likely to have read “The Lost Symbol”, by Dan Brown. He almost certainly liked it.

It’s worth paying a look at the original article, “A World of Hits”, which originally came out last November. It’s been borne out since: my holiday and post-holiday discussions of the media were dominated by Avatar, to the exclusion of almost anything else. A lot of people wanted to talk about Avatar, and there’s a fair amount to discuss there: how pretty it is, how it works as mass spectacle, the film’s deeply muddled politics, how ecology and religion are connected. What stands out to me is how rarely this happens any more. It’s not that I don’t have, for example, film people to talk about films with, or movie people to talk about movies with, but there’s always some element of singing to the choir in this. An example: the record I liked the most in 2009 was the favorite record of plenty of other people, but didn’t merit a mention in the New York Times. I don’t think this is necessarily because the Times has especially bad music coverage; there’s simply a lot of media happening, and it’s impossible for any publication to cover everything in depth. The sheer ubiquity of Avatar changed how it could be discussed: something so big can cut across our individual interest groups, enabling broader conversations.
But the inevitable question arises: what does it mean if the only cultural object that everyone can talk about costs $300 million? Producers aren’t going to give $300 million to anyone who has a story that they need to discuss. The great power of the written word – why the word “book” continues to mean so much to us – is its fundamental democracy: that anyone literate can set pen to paper and write something. Technology, the truism goes, is politically neutral; but I wonder if this can be true in a practical sense when the tools of expression are so expensive.

smart thinking from Mitch Ratcliffe

Mitch Ratcliffe posted this very smart piece yesterday and gave permission to cross-post it on if:book.
How to create new reading experiences profitably

Concluding my summary of my recent presentation to a publishing industry group, begun here and continued here, we turn to the question of what to do to revitalize the publishing opportunity.
I wrote a lot about the idea of exploding the limitations of the book in the last installment. Getting beyond the covers. Turning from a distribution model to a reader-centric model. It’s simple to argue that change is needed and to say what needs changing. Here, I offer a few specific ideas about lines of research and development that I would like to see begun by publishers, who, if they wish to remain viable–let alone profitable–must undertake immediately. The change in book publishing will happen at a faster pace than the collapse of newspaper and music publishing did, making a collective effort at research and publication of the results for all to discuss and use, critical during the next 18 months. Think open sourcing the strategy, so that a thousand innovations can bloom.
Making books into e-books is not the challenge facing publishers and authors today. In fact, thinking in terms of merely translating text to a different display interface completely misses the problem of creating a new reading experience. Books have served well as, and will continue to be, containers for moving textual and visual information between places and across generations. They work. They won’t stop working. But when moving to a digital environment, books need to be conceived with an eye firmly set on the interactions that the text/content will inspire. Those interactions happen between the author and work, the reader and the work, the author and reader, among readers and between the work and various services, none of which exist today in e-books, that connect works to one another and readers in the community of one book with those in other book-worlds.
Just as with the Web, where value has emerged out of the connection of documents by publishers and readers–the Web is egalitarian in its connectivity, but still has some hierarchical features in its publishing technologies–books must be conceived of not just as a single work, but a collection of work (chapters, notes, illustrations, even video, if you’re thinking of a “vook”) that must be able to interact internally and with other works with which it can communicate over an IP network. This is not simply the use of social media within the book, though that’s a definite benefit, but making the book accessible for use as a medium of communication. Most communities emerge long after the initial idea that catalyzes them is first published.
These communications “hooks” revolve around traditional bibliographic practices, such as indexing and pagination for making references to a particular location in a book useful, as well as new functionality, such as building meta-books that combine the original text with readers’ notes and annotations, providing verification of texts’ authenticity and completeness, curation (in the sense that, if I buy a book today and invest of myself in it the resulting “version” of the book will be available to others as a public or private publication so that, for instance, I can leave a library to my kids and they can pass it along to their children) and preservation.
Think about how many versions of previously published books, going all the way back to Greek times, when books were sold on scrolls in stalls at Athens, have been lost. We treasure new discoveries of an author’s work. In a time of information abundance, however, we still dismiss all the other contributions that make a book a vital cultural artifact. Instead, we need to recognize that capturing the discussions around a book, providing access (with privacy intact) to the trails that readers have followed on their own or in discussions with others to new interpretations and uses for a text, and the myriad commentaries and marginalia that have made a book important to its readers is the new source of infinite value that can be provided as the experience we call “reading.” Tomorrow’s great literary discovery may not be an original work by a famous author, but a commentary or satire written by others in response to a book (as the many “…with sea monsters and vampires” books out there are beginning to demonstrate today). Community or, for lack of a better way of putting it, collaboration, is the source of emergent value in information. Without people talking about the ideas in a book, the book goes nowhere. This is why Cory Doctorow and Seth Godin’s advice about giving away free e-books makes so much sense, up to a point. Just turning a free ebook into the sale of a paper book leaves so much uninvented value on the table that, frankly, readers will never realize to get if someone doesn’t roll the experience up into a useful service.
The interaction of all the different “versions” of a book is what publishers can facilitate for enhanced revenues. This could also be accomplished by virtually anyone with a budget necessary to support a large community. The fascinating thing about the networked world, of course, is that small communities are what thrive, some growing large, so that anyone could become a “publisher” by growing a small community into a large or multi-book community.
Publishers can, though they may not be needed to, provide the connectivity or, at least, the “switchboards” that connect books. The BookServer project recently announced by The Internet Archive takes only one of several necessary steps toward this vision, though it is an important one: If realized, it will provide searchable indices and access to purchase or borrow any book for consumption in paper or a digital device. That’s big. It’s a huge project, but it leaves out all the content and value added to books by the public, by authors who release new editions (which need to be connected, so that readers want to understand the changes made between editions), and, more widely, the cultural echoes of a work that enhance the enjoyment and importance attributed to a work.
What needs to exist beside the BookServer project is something I would describe as the Infinite Edition Project. This would build on the universal index of books, but add the following features:
Word and phrase-level search access to books, with pagination intact;
A universal pagination of a book based upon a definitive edition or editions, so that the books’ readers can connect through the text from different editions based on a curated mapping of the texts (this would yield one ISBN per book that, coupled with various nascent text indexing standards, could produce an index of a title across media);
Synchronization of bookmarks, notes and highlights across copies of a book;
A reader’s copy of each book that supported full annotation indexing, allowing any reader to markup and comment on any part of a book–the reader’s copy becomes another version of the definitive edition;
An authenticated sharing mechanism that blends reader’s copies of a book based on sharing permissions and, this is key, the ability to sell a text and the annotations separately, with the original author sharing value with annotators;
The ability to pass a book along to someone, including future generations–here, we see where a publisher could sell many copies of their books based on their promise to keep reader’s copies intact and available forever, or a lifetime;
An authentication protocol that validates the content of a book (or a reader’s copy), providing an “audit trail” for attribution of ideas–both authors and creative readers, who would be on a level playing field with authors in this scenario, would benefit, even if the rewards were just social capital and intellectual credit;
Finally, Infinite Edition, by providing a flourishing structure for author-reader and inter-generational collaboration, could be supported by a variety of business models.
One of the notions that everyone, readers and publishers included, have to get over is the idea of a universal format. Text these days is something you store but not something that is useful without metadata and application-layer services. Infinite Edition, as the illustration above shows, is a Web services-based concept that allows the text to move across devices and formats. It would include an API for synchronizing a reader’s copy of a book, as well as for publishers or authors to update or combine versions.
Done right, an Infinite Edition API would let a Kindle user share notes with a Nook user and with a community that was reading the book online; if, at some point, a new version of the book were to be printed, publishers could, with permission, augment the book with contributions of readers. Therein lies a defensible value-added service that, at the same time, is not designed to prevent access to books–it’s time to flip the whole notion of protecting texts on its head and talk about how to augment and extend texts. As I have written elsewhere over the years, this is a feature cryptographic technology is made to solve, but without the stupidity of DRM.
Granted, access to a book would still be on terms laid out by the seller (author, publisher or distributor, such as Amazon, which could sell the text with Kindle access included through its own secure access point to the Infinite Edition Web service) however, it would become an “open” useful document once in the owner’s hands. And all “owners” would be able to do much with their books, including share them and their annotations. Transactions would become a positive event, a way for readers to tap into added services around a book they enjoy.
Publishers must lead this charge, as I wrote last time, because distributors are not focused on the content of books, just the price. A smart publisher will not chase the latest format, instead she will emphasize the quality of the books offered to the market and the wealth of services those books make accessible to customers. This means many additive generations of development will be required of tool makers and device designers who want to capitalize on the functionality embedded in a well-indexed, socially enabled book. It will prevent the wasteful re-ripping of the same content into myriad formats in order to be merely accessible to different reader devices. And the publishers who do invent these services will find distributors as willing as ever to sell them, because a revived revenue model is attractive to everyone.
ePUB would be a fine foundation for an Infinite Edition project. So would plain text with solid pagination metadata. In the end, though, what the page is–paper or digital–will not matter as much as what is on the page. Losing sight of the value proposition in publishing, thinking that the packaging matters more than the content to be distributed, is what has nearly destroyed newspaper publishing. Content is king, but it is also coming from the masses and all those voices have individual value as great as the king. So, help pull these communities of thought and entertainment together to remain a vital contributor to your customers.
This raises the last challenge I presented during my talk. The entire world is moving to a market ideal of getting people what they want or need when they want or need it. Publishing is only one of many industries battling the complex strategic challenge of just-in-time composition of information or products for delivery to an empowered individual customer. This isn’t to say that it is any harder, nor any easier, to be a publisher today compared to say, a consumer electronics manufacturer or auto maker, only that the discipline to recognize what creates wonderful engaging experience is growing more important by the day.
As I intimated in the last posting, this presentation didn’t land me the job I was after. I came in second, which is a fine thing to do in such amazing times. Congratulations to Scott Lubeck, who was named today as executive director of the Book Industry Study Group. I have joined the BISG to be an activist member. I welcome contact from anyone wishing to discuss how the BISG or individual publishers can put these ideas into action, a little at a time or as a concerted effort to transform the marketplace.

the final cut

Julio Cortázar is one of those writers who is mentioned far more often than actually read; most people know that he wrote Hopscotch, a novel often mentioned as a precursor to hypertext fiction, or that he wrote the short stories that became Antonioni’s Blow Up and Godard’s Weekend. I’ve been belatedly making my way through his works over the past year, a pleasurable endeavor that I’d commend to anyone. For all the pleasure of his fiction, there’s a great deal that still bears consideration, especially in the English-speaking world where Cortázar hasn’t been widely read.
“We Love Glenda So Much” is the title story of one of Cortázar’s last collections of short stories, unfortunately out of print in English but easily available online. It’s a short Borgesian piece, told in the first person plural, and the premise is quickly related: the anonymous narrators idolize the movie star Glenda Garson and constitute a secret fan club for the most devoted fans who understand that her films are the only ones that matter. Things quickly escalate: though the members believe Glenda’s work to be perfection, they acknowledge that her films are perhaps not quite perfect. They laboriously gather up all the prints and re-edit them, creating not a director’s cut but a fans’ cut; the recut films are redistributed to an unsuspecting public; the differences between the new films and those released years before are chalked up to the vagaries of memory. All is well in the world; until Glenda decides to return from her retirement, at which point her fans lethally prevent her from sullying the perfection they have helped her achieve.
We Love Glenda So Much was originally published in 1980, just before the VCR became ubiquitous; it’s unclear when the story is set, but it presumes a world where all the copies of Glenda’s films can be gathered in by her hard-working (and conveniently rich) fans. Read 29 years later, it’s a remarkably different text: while fans are still reliably crazy, the way that the entertained interact with entertainers and that the media can be controlled – or not – has been transformed so much as to render Cortázar’s story a strange postcard from a forgotten land. Subcultures no longer exist in the way they did in the way Cortázar describes, even in the way they existed a dozen years ago, when it was necessary to hold fast to your own personal Glenda because of the investment, cultural or economic, that you’d made in loving her above everyone else. With ubiquitous media, everyone is free to be a dilettante.
More importantly, though, there’s been a shift in the center of control. When Cortázar wrote, the producers maintained control tightly; since then, media control has been dispersed to the point where it’s meaningless. (Cortázar’s language suggests a focus on this power shift: the members of the club that loves Glenda, a movement to seize power from Glenda’s producers, announce that think of themselves as a “nucleus,” suggesting a center of force, rather than a simple “club.”) It’s not hard to re-edit any movie you like to your pleasure; however, making your cut – even if you’re the director – the definitive cut is increasingly impossible. I’m not trying to argue that Cortázar was unperceptive and didn’t see the future coming; the story is a fabulation through and through, and I don’t imagine for a minute that Cortázar thought that what he was describing could happen in the real world. But I don’t think that it’s out of line to read Glenda, whose fans know better than her what she should do or not do, as a portrait of the artist himself. This story did appear in his penultimate collection of short stories, released four years before his death, and he may well have had posterity on his mind. Authorial intent is quickly dispensed with after the death of the author: the past year, for example, has seen new editions of the work of Hemingway, Carver, and Nabokov that would almost certainly violate authorial intent.
What makes “We Love Glenda So Much” interesting to me is how it points out this world of transformation that we’ve lived through. When looked at in terms of control, there’s also a distant echo of another transformation – Gutenberg’s invention of the printing press – which similarly dispersed control from the center to the masses. The generally accepted narrative of history has it that the coming of printing led directly to the massive societal upheaval that was the Protestant Reformation: the ability to quickly disperse texts led to a plurality of viewpoints. Gutenberg’s contemporaries aren’t noted for having had any idea how quickly their world would become utterly unrecognizable. Nostradamus, who might have, would arrive a generation later, just in time for the printing press to distribute his prophecies. One wonders how much of an idea we have now.

when we get what we want

It’s the shortest day of the year, New York is under a thick blanket of snow which will soon turn to slush, and it’s hard not to feel let down by the world: when the Democrats gut health care reform in the name of passing something; when Obama fails to accomplish anything meaningful in Copenhagen; as it becomes clear that the war in Afghanistan will soon be marching into its second decade; when bankers still gain million-dollar bonuses for doing, as far as one can tell, nothing useful; when Mexico City can legalize gay marriage before New York State. December, the end of the decade, both summon up the retrospective urge: to look back and wonder what was accomplished, what could have been accomplished. One feels more disappointed when it seems like something could have happened: a year ago it felt like Obama might be a light at the end of a decade-long tunnel.
It’s also the end of a year in which electronic books have received an amount of press that would have been unthinkable five years ago. At the same time, I find it hard to get as excited about anything recently: there’s been a great deal of hubbub about devices, but it seems clear that’s premature: any device will be obsolete in a year or so, when Apple rolls out its long-delayed tablet and a barrage of Google-powered devices follow in its wake. There’s a lot of publishing talk, but it’s not especially interesting for the would-be reader: publishers argue about the dates at which e-books come out relative to hardcovers, and whether to grant exclusive rights to a single distributer. It’s not particularly interesting, especially with the past five years in mind: in the current discussions, the book is simply a commodity, something to be passed off to consumers at the greatest profit the market will bear. (An analogous argument might be made about the social networking world, which seems to reach a logical conclusion in Blippy, a new social networking platform with the genius idea – albeit a recession too late – of replacing the arduousness of even Twitter-length communication with a record of one’s credit card purchases.) The argument could be made that electronic books have finally made it; executives make grand proclamations about how electronic sales will figure as such and such a percentage of future sales. Stephen Covey makes exclusive distribution deals with Amazon. Motoko Rich dutifully reports it all daily in The New York Times.
Part of my problem is personal. The technology world turns over so quickly: the five years that I’ve spent at the Institute is more than enough to make anyone feel wizened and elderly in the amnesiac new media world, if one didn’t come into it feeling that way. It becomes increasingly hard to find anything that seems new or interesting, let alone revelatory. And, as mentioned, it’s the end of the year; it’s dark, there’s a recession going on. But I’m not sure I’m alone in feeling that there’s been something of a longueur in the world of new or social media; I’ve had a number of conversations recently that come back around to this. A post by Whitney Trettien at diapsalmata beautifully identifies a related sort of frustration: the sense of being on the verge of a future that’s dragging its feet. Trettien begins her post with a triad of quotes on the imminence of a new digital reading experience, quotes that the unassuming reader might imagine to be from the present; they turn out to be from 1999.

*     *     *     *     *

Symptoms suggest a diagnosis; for a physician, I’ll suggest St. Teresa of Avila. Answered prayers, she noted, cause more tears than unanswered ones; Truman Capote would have put this on the cover of his unfinished novel about the unsated. At about the same time I was reading Capote, Courtney Love offered a more profane restatement: “I told you from the start just how this would end / when I get what I want I never want it again”. It’s a description that seems apt for the present: we’re living in a world of answered prayers. Obama restored order after eight years of misrule; millions of people are reading digitally from their iPhones and Kindles and whatever else they’re using.
Why tears then? By and large, the current raft of devices and software aren’t particularly innovative or attuned to how reading works; the reading experience itself isn’t substantially different than it was fifteen years ago, when Voyager was offering mass-market books on 3.5” disks for sale in bookstores, though the hardware’s much better. What’s different is the networked aspect: it’s now possible to get a huge number of books near instantaneously. Bored on a Virginia highway over Thanksgiving & feeling the need to re-read Edgar Allan Poe’s Narrative of A. Gordon Pym, I could nearly instantaneously (and without cost) download a copy to my iPhone. This is, it should be said, astonishing. The reading experience was terrible, but that’s beside the point: I could get the text (or an approximation thereof) almost as soon as I could imagine it. This is something new.
I’m interested in this sensation of immediate gratification: it’s one of the hallmarks of the present. Something similar is present when you can buy almost any piece of music from iTunes, or when you can choose from a plethora of movies to watch instantly on Netflix. In a weakened form, it’s the same experience as buying used books on Amazon, something I’m constantly doing: almost any book, no matter how out of print, will pop up on Amazon, and most can be yours for under $10 and inside a week. In all these cases: given enough capital and access to a network, the majority of our desires for media can be solved, increasingly instantaneously.
We don’t think about this very much, but this newfound ability to instantly satisfy our desires is actually a very strange development. So much of human development is a process of learning to deal with desires that are delayed or vexed; so much of the history of the book is a narrative of scarceness. In terms of the market, there was more demand than supply. The move to the digital has changed all that: the supply of a piece of digital content is, for most intents and purposes, infinite, and we find ourselves in a position where supply far exceeds demand. It isn’t just books where this is the case: it might be said that all electronic reading is in this position. If you have an even marginal amount of curiosity, there’s no end of content that could, given the time, be interesting.
But this change in values comes at a price: with this shift in values, it becomes very hard for us to know how to value content. We’re used to the arc of wanting – conceiving a desire, justifying it to ourselves, figuring out a way to get it, receiving it – that serves to convince ourselves of the importance of what we want. When that’s short-circuited, we’re left at a loss. In a previous end of the year piece, I used the example of the rescued Robinson Crusoe in Elizabeth Bishop’s “Crusoe in England”: he looks at his knife, the thing he once valued religiously, and finds that it’s become meaningless, just another knife. “The living soul,” he notes, “has dribbled away.” Maybe that’s where the book is now.
Raymond Aron, from Eighteen Lectures on the Industrial Society: “It is poverty that humanity, as a whole, still suffers from today. Poverty, defined simply by the lack of common measure between the desires of individuals and the means to satisfy them.”

*     *     *     *     *

I’ve been keeping a list this year of the books I’ve finished this year, from a need, I suppose, to keep some kind of record: if books make a man, keeping a list of ingredients might be useful, though for what I’m not sure, yet. I haven’t tallied up my year in reading yet, but I suspect I’ll wind up having read somewhere around a book every three days. (This isn’t particularly difficult if you have a typical New York commute and count on other people to read The New Yorker for you.) This seems desperate: a frantic attempt to stay abreast of a world that moves ever faster. There are more books on my lists of books to read than I have any reasonable chance of completing in my lifetime. This happens.
The most interesting reading experience of my year, however, wasn’t a book that I finished; rather, it was reading at a book that can’t be finished. I’ve been going to the monthly meetings of the New York Finnegans Wake reading group. I’m not, it should be said, the most religious of Joyceans; re-reading Ulysses a decade ago, I had an epiphany and suddenly understood what style meant, but if I had to make a list of my ten favorite writers I’m not entirely sure that Joyce would make the cut. It’s Proust that I go to when I want answers. But there’s something pleasing in the ritual aspect of the group reading of Finnegans Wake: feeling oneself a part of a community, a community devoted to something greater than oneself.
Each month the group makes its way through about two pages in two hours. Wine is drunk, everyone reads a bit of the text aloud. The group has been at this task for a very long time; the median age of the members is twice my age, and plenty of them have been at their reading since before I was born. Close reading is the only real way into Finnegans Wake. In a text so dense, the reader can only begin to understand by listening to others. Different readers find something different in it; some scrutinize the text silently, some sound out words for unvoiced puns, others argue their own idiosyncratic theories about the text. Digressions invariably arise and are followed. But when it works, it feels almost as if the text lifts off the page: out of the cacophony, you begin to hear how Joyce’s overlaid narratives resonate.
One arrives, finally, a bit closer to an understanding of the text. It’s something that’s only really possible through group reading: the individual reader can’t possibly know as much as the group – although something like the experience could certainly be asynchronously recreated if you had enough books and patience. But I’m most interested in the experience of reading this way: because in this community context, reading becomes something more than the interior experience that we usually think of reading as; it’s something entirely outside of the economic context (or even an academic one). I’ve never found anything quite like it on the Web, despite long residencies on a couple of literary mailing lists. Reading like this, one gets the sense that maybe one book might almost be enough for the rest of one’s life – provided, of course, that it was the right book. This, I think, might be one remedy we should be looking for if we’re trying to find way forward for reading: to think about reading as a matter of communal exchange rather than of commodity exchange.