Category Archives: web

the book is reading you

I just noticed that Google Book Search requires users to be logged in on a Google account to view pages of copyrighted works.
google book search account.jpg
They provide the following explanation:

Why do I have to log in to see certain pages?
Because many of the books in Google Book Search are still under copyright, we limit the amount of a book that a user can see. In order to enforce these limits, we make some pages available only after you log in to an existing Google Account (such as a Gmail account) or create a new one. The aim of Google Book Search is to help you discover books, not read them cover to cover, so you may not be able to see every page you’re interested in.

So they’re tracking how much we’ve looked at and capping our number of page views. Presumably a bone tossed to publishers, who I’m sure will continue suing Google all the same (more on this here). There’s also the possibility that publishers have requested information on who’s looking at their books — geographical breakdowns and stats on click-throughs to retailers and libraries. I doubt, though, that Google would share this sort of user data. Substantial privacy issues aside, that’s valuable information they want to keep for themselves.
That’s because “the aim of Google Book Search” is also to discover who you are. It’s capturing your clickstreams, analyzing what you’ve searched and the terms you’ve used to get there. The book is reading you. Substantial privacy issues aside, (it seems more and more that’s where we’ll be leaving them) Google will use this data to refine Google’s search algorithms and, who knows, might even develop some sort of personalized recommendation system similar to Amazon’s — you know, where the computer lists other titles that might interest you based on what you’ve read, bought or browsed in the past (a system that works only if you are logged in). It’s possible Google is thinking of Book Search as the cornerstone of a larger venture that could compete with Amazon.
There are many ways Google could eventually capitalize on its books database — that is, beyond the contextual advertising that is currently its main source of revenue. It might turn the scanned texts into readable editions, hammer out licensing agreements with publishers, and become the world’s biggest ebook store. It could start a print-on-demand service — a Xerox machine on steroids (and the return of Google Print?). It could work out deals with publishers to sell access to complete online editions — a searchable text to go along with the physical book — as Amazon announced it will do with its Upgrade service. Or it could start selling sections of books — individual pages, chapters etc. — as Amazon has also planned to do with its Pages program.
Amazon has long served as a valuable research tool for books in print, so much so that some university library systems are now emulating it. Recent additions to the Search Inside the Book program such as concordances, interlinked citations, and statistically improbable phrases (where distinctive terms in the book act as machine-generated tags) are especially fun to play with. Although first and foremost a retailer, Amazon feels more and more like a search system every day (and its A9 engine, though seemingly always on the back burner, is also developing some interesting features). On the flip side Google, though a search system, could start feeling more like a retailer. In either case, you’ll have to log in first.

insidious tactic #348: charge for web speed

An article in yesterday’s Washington Post — “Executive Wants to Charge for Web Speed” — brings us back to the question of pipes and the future of the internet. fiber optic.jpg The chief technology officer for Bell South says telecoms and cable companies ought to be allowed to offer priority deals to individual sites, charging them extra for faster connections. The Post:

Several big technology firms and public interest groups say that approach would enshrine Internet access providers as online toll booths, favoring certain content and shutting out small companies trying to compete with their offerings.

Among these “big technology firms” are Google, Yahoo!, Amazon and eBay, all of whom have pressed the FCC for strong “network neutrality” provisions in the latest round of updates to the 1996 Telecommunications Act. These would forbid discrimination by internet providers against certain kinds of content and services (i.e. the little guys). BellSouth claims to support the provisions, though the statements of its tech officer suggest otherwise.
Turning speed into a bargaining chip will undoubtedly privilege the richer, more powerful companies and stifle competition — hardly a net-neutral scenario. They claim it’s no different from an airline offering business class — it doesn’t prevent folks from riding coach and reaching their destination. But we all know how cramped and awful coach is. The truth is that the service providers discriminate against everyone on the web. We’re all just freeloaders leeching off their pipes. The only thing that separates Google from the lady blogging about her cat is how much money they can potentially pay for pipe rental. That’s where the “priorities” come in.
Moreover, the web is on its way to merging with cable television, and this, in turn, will increase the demand for faster connections that can handle heavy traffic. So “priority” status with the broadband providers will come at an ever increasing premium. That’s their ideal business model, allowing them to charge the highest tolls for the use of their infrastructure. That’s why the telecos and cablecos want to ensure, through speed-baiting and other screw-tightening tactics, that the net transforms from a messy democratic commons into a streamlined broadcast medium. Alternative media, video blogging, local video artists? These will not be “priorities” in the new internet. Maximum profit for pipe-holders will mean minimum diversity and a one-way web for us.
In a Business Week interview last month, SBC Telecommunications CEO Edward Whitacre expressed what seemed almost like a lust for revenge. Asked, “How concerned are you about Internet upstarts like Google, MSN, Vonage, and others?” he replied:

How do you think they’re going to get to customers? Through a broadband pipe. Cable companies have them. We have them. Now what they would like to do is use my pipes free, but I ain’t going to let them do that because we have spent this capital and we have to have a return on it. So there’s going to have to be some mechanism for these people who use these pipes to pay for the portion they’re using. Why should they be allowed to use my pipes?
The Internet can’t be free in that sense, because we and the cable companies have made an investment and for a Google or Yahoo! or Vonage or anybody to expect to use these pipes [for] free is nuts!

This makes me worry that discussions about “network neutrality” overlook a more fundamental problem: lack of competition. “That’s the voice of someone who doesn’t think he has any competitors,” says Susan Crawford, a cyberlaw and intellectual property professor at Cardozo Law School who blogs eloquently on these issues. She believes the strategy to promote network neutrality will ultimately fail because it accepts a status quo in which a handful of broadband monopolies dominate the market. “We need to find higher ground,” she says.

I think the real fight should be over rights of way and platform competition. There’s a clear lack of competition in the last mile — that’s where choice has to exist, and it doesn’t now. Even the FCC’s own figures reveal that cable modem and DSL providers are responsible for 98% of broadband access in the U.S., and two doesn’t make a pool. If the FCC is getting in the way of cross-platform competition, we need to fix that. In a sense, we need to look down — at the relationship between the provider and the customer — rather than up at the relationship between the provider and the bits it agrees to carry or block…
…Competition in the market for pipes has to be the issue to focus on, not the neutrality of those pipes once they have been installed. We’ll always lose when our argument sounds like asking a regulator to shape the business model of particular companies.

The broadband monopolies have their priorities figured out. Do we?
image: “explosion” (reminded me of fiber optic cable) by The Baboon, via Flickr

google print on deck at radio open source

Open Source, the excellent public radio program (not to be confused with “Open Source Media”) that taps into the blogosphere to generate its shows, has been chatting with me about putting together an hour on the Google library project. Open Source is a unique hybrid, drawing on the best qualities of the blogosphere — community, transparency, collective wisdom — to produce an otherwise traditional program of smart talk radio. As host Christopher Lydon puts it, the show is “fused at the brain stem with the world wide web.” Or better, it “uses the internet to be a show about the world.”
The Google show is set to air live this evening at 7pm (ET) (they also podcast). It’s been fun working with them behind the scenes, trying to figure out the right guests and questions for the ideal discussion on Google and its bookish ambitions. My exchange has been with Brendan Greeley, the Radio Open Source “blogger-in-chief” (he’s kindly linked to us today on their site). We agreed that the show should avoid getting mired in the usual copyright-focused news peg — publishers vs. Google etc. — and focus instead on the bigger questions. At my suggestion, they’ve invited Siva Vaidhyanathan, who wrote the wonderful piece in the Chronicle of Higher Ed. that I talked about yesterday (see bigger questions). I’ve also recommended our favorite blogger-librarian, Karen Schneider (who has appeared on the show before), science historian George Dyson, who recently wrote a fascinating essay on Google and artificial intelligence, and a bunch of cybertext studies people: Matthew G. Kirschenbaum, N. Katherine Hayles, Jerome McGann and Johanna Drucker. If all goes well, this could end up being a very interesting hour of discussion. Stay tuned.
UPDATE: Open Source just got a hold of Nicholas Kristof to do an hour this evening on Genocide in Sudan, so the Google piece will be pushed to next week.

flushing the net down the tubes

Grand theories about upheavals on the internet horizon are in ready supply. Singularities are near. Explosions can be expected in the next six to eight months. Or the whole thing might just get “flushed” down the tubes. This last scenario is described at length in a recent essay in Linux Journal by Doc Searls, which predicts the imminent hijacking of the net by phone and cable companies who will turn it into a top-down, one-way broadcast medium. In other words, the net’s utopian moment, the “read/write” web, may be about to end. Reading Searls’ piece, I couldn’t help thinking about the story of radio and a wonderful essay Brecht wrote on the subject in 1932:
brecht-foto.jpg

Here is a positive suggestion: change this apparatus over from distribution to communication. The radio would be the finest possible communication apparatus in public life, a vast network of pipes. That is to say, it would be if it knew how to receive as well as to transmit, how to let the listener speak as well as hear, how to bring him into a relationship instead of isolating him. On this principle the radio should step out of the supply business and organize its listeners as suppliers….turning the audience not only into pupils but into teachers.

Unless you’re the military, law enforcement, or a short-wave hobbyist, two-way radio never happened. On the mainstream commercial front, radio has always been about broadcast: a one-way funnel. The big FM tower to the many receivers, “prettifying public life,” as Brecht puts it. Radio as an agitation? As an invitation to a debate, rousing families from the dinner table into a critical encounter with their world? Well, that would have been neat.
Now there’s the internet, a two-way, every-which-way medium — a stage of stages — that would have positively staggered a provocateur like Brecht. But although the net may be a virtual place, it’s built on some pretty actual stuff. Copper wire, fiber optic cable, trunks, routers, packets — “the vast network of pipes.” The pipes are owned by the phone and cable companies — the broadband providers — and these guys expect a big return (far bigger than they’re getting now) on the billions they’ve invested in laying down the plumbing. Searls:

The choke points are in the pipes, the permission is coming from the lawmakers and regulators, and the choking will be done….The carriers are going to lobby for the laws and regulations they need, and they’re going to do the deals they need to do. The new system will be theirs, not ours….The new carrier-based Net will work in the same asymmetrical few-to-many, top-down pyramidal way made familiar by TV, radio, newspapers, books, magazines and other Industrial Age media now being sucked into Information Age pipes. Movement still will go from producers to consumers, just like it always did.

If Brecht were around today I’m sure he would have already written (or blogged) to this effect, no doubt reciting the sad fate of radio as a cautionary tale. Watch the pipes, he would say. If companies talk about “broad” as in “broadband,” make sure they’re talking about both ends of the pipe. The way broadband works today, the pipe running into your house dwarfs the one running out. That means more download and less upload, and it’s paving the way for a content delivery platform every bit as powerful as cable on an infinitely broader band. Data storage, domain hosting — anything you put up there — will be increasingly costly, though there will likely remain plenty of chat space and web mail provided for free, anything that allows consumers to fire their enthusiasm for commodities through the synapse chain.
rad30cathedral10.jpg If the net goes the way of radio, that will be the difference (allow me to indulge in a little dystopia). Imagine a classic Philco cathedral radio but with a few little funnel-ended hoses extending from the side that connect you to other listeners. “Tune into this frequency!” “You gotta hear this!” You whisper recommendations through the tube. It’s sending a link. Viral marketing. Yes, the net will remain two-way to the extent that it helps fuel the market. Web browsers, like the old Philco, would essentially be receivers, enabling participation only to the extent that it encouraged others to receive.
You might even get your blog hosted for free if you promote products — a sports shoe with gelatinous heels or a music video that allows you to undress the dancing girls with your mouse. Throw in some political rants in between to blow off some steam, no problem. That’s entrepreneurial consumerism. Make a living out of your appetites and your ability to make them infectious. Hip recommenders can build a cosy little livelihood out of their endorsements. But any non-consumer activity will be more like amateur short-wave radio: a mildly eccentric (and expensive) hobby (and they’ll even make a saccharine movie about a guy communing with his dead firefighter dad through a ghost blog).
Searls sees it as above all a war of language and metaphor. The phone and cable companies will dominate as long as the internet is understood fundamentally as a network of pipes, a kind of information transport system. This places the carriers at the top of the hierarchy — the highway authority setting the rules of the road and collecting the tolls. So far the carriers have managed, through various regulatory wrangling and court rulings, to ensure that the “transport metaphor” has prevailed.
But obviously the net is much more than the sum of its pipes. It’s a public square. It’s a community center. It’s a market. And it’s the biggest publishing system the world has ever known. Searls wants to promote “place metaphors” like these. Sure, unless you’re a lobbyist for Verizon or SBC, you probably already think of it this way. But in the end it’s the lobbyists that will make all the difference. Unless, that is, an enlightened citizens’ lobby begins making some noise. So a broad, broad as in broadband, public conversation should be in order. Far broader than what goes on in the usual progressive online feedback loops — the Linux and open source communities, the creative commies, and the techno-hip blogosphere, that I’m sure are already in agreement about this.
Google also seems to have an eye on the pipes, reportedly having bought thousands of miles of “dark fiber” — pipe that has been laid but is not yet in use. Some predict a nationwide “Googlenet.” But this can of worms is best saved for another post.

explosion

250px-Nuclear_fireball.jpg A Nov. 18 post on Adam Green’s Darwinian Web makes the claim that the web will “explode” (does he mean implode?) over the next year. According to Green, RSS feeds will render many websites obsolete:
The explosion I am talking about is the shifting of a website’s content from internal to external. Instead of a website being a “place” where data “is” and other sites “point” to, a website will be a source of data that is in many external databases, including Google. Why “go” to a website when all of its content has already been absorbed and remixed into the collective datastream.
Does anyone agree with Green? Will feeds bring about the restructuring of “the way content is distributed, valued and consumed?” More on this here.

the times they are a-changin’

Knight Ridder Inc., the second largest newspaper conglomerate in the U.S., is under intense pressure from its more powerful investors to start selling off papers. The New York Times reports that the company is now contemplating “strategic alternatives.” Consider the following in terms of what Bob is saying one post down about time. With the rise of the 24-hour news cycle and the internet, news is adopting a different time signature.

It is unclear who may want to buy Knight Ridder. Newspaper companies, though still immensely profitable, have a murky future that is clouded by a shrinking readership and weak advertising revenue, both of which are being leeched away by the Internet.
…In the six moths that ended in September, newspaper circulation nationally fell 2.6 percent daily and 3.1 percent on Sundays, the biggest decline in any comparable period since 1991, according to the Audit Bureau of Circulations. All in all, 45.2 million people subscribed to 1,457 reporting papers, down from a peak of 63.3 million people and 1,688 newspapers in 1984.
By comparison, 47 million people visited newspaper Web sites, about a third of United States Internet users, according to the circulation bureau.

The time it takes to read the newspaper in print — a massive quilt, chopped up and parceled (I believe Gary Frost said something about this) — you might say it leads to a different sort of understanding of the world around you. It seems to me that the newspapers that will last longest in print are the Sunday editions, aimed at a leisurely audience, taking stock of the week that has just ended and preparing for the one about to commence. On Sundays, the world spreads out before you in print, and perhaps you make a point of taking some time away from the computer (at least, this might be the case for hybrid monkeys like me who are more or less at home with both print and digital). The briskness of discourse on the web and in popular culture does not afford the time to engage with big ideas. Bob talks, not without irony, about “tithing to the church of big ideas.” Set aside the time to engage with world-changing ideas, willfully turn away from the screen.
The persistence of the Sunday print edition, if it comes to pass, might in some way reflect this kind of tithing, this intentional slowing down.

ebr is back

ebr.jpg
ebr is back after a several month hiatus during which time it was overhauled. The site, published by AltX was among the first places where the “technorati meets the literati” and I always found it attractive for its emphasis on sustained analysis of digital artifacts and the occasional pop culture reference. The latest project, first person series, seems to answer a lot of what bob finds attractive in the blogs of juan cole and others. And although I’ve heard ebr called “too linear” (as compared to Vectors, USC’s e-journal) the interface goes a long way toward solving the problem of the scrolling feature of many sites/blogs which privilege what’s new. The interweaving threads with search capabilities seem quite hearty.

a better wikipedia will require a better conversation

There’s an interesting discussion going on right now under Kim’s Wikibooks post about how an open source model might be made to work for the creation of authoritative knowledge — textbooks, encyclopedias etc. A couple of weeks ago there was some dicussion here about an article that, among other things, took some rather cheap shots at Wikipedia, quoting (very selectively) a couple of shoddy passages. Clearly, the wide-open model of Wikipedia presents some problems, but considering the advantages it presents (at least in potential) — never out of date, interconnected, universally accessible, bringing in voices from the margins — critics are wrong to dismiss it out of hand. Holding up specific passages for critique is like shooting fish in a barrel. Even Wikipedia’s directors admit that most of the content right now is of middling quality, some of it downright awful. It doesn’t then follow to say that the whole project is bunk. That’s a bit like expelling an entire kindergarten for poor spelling. Wikipedia is at an early stage of development. Things take time.
Instead we should be talking about possible directions in which it might go, and how it might be improved. Dan for one, is concerned about the market (excerpted from comments):

What I worry about…is that we’re tearing down the old hierarchies and leaving a vacuum in their wake…. The problem with this sort of vacuum, I think, is that capitalism tends to swoop in, simply because there are more resources on that side….
…I’m not entirely sure if the world of knowledge functions analogously, but Wikipedia does presume the same sort of tabula rasa. The world’s not flat: it tilts precariously if you’ve got the cash. There’s something in the back of my mind that suspects that Wikipedia’s not protected against this – it’s kind of in the state right now that the Web as a whole was in 1995 before the corporate world had discovered it. If Wikipedia follows the model of the web, capitalism will be sweeping in shortly.

Unless… the experts swoop in first. Wikipedia is part of a foundation, so it’s not exactly just bobbing in the open seas waiting to be swept away. If enough academics and librarians started knocking on the door saying, hey, we’d like to participate, then perhaps Wikipedia (and Wikibooks) would kick up to the next level. Inevitably, these newcomers would insist on setting up some new vetting mechanisms and a few useful hierarchies that would help ensure quality. What would these be? That’s exactly the kind of thing we should be discussing.
The Guardian ran a nice piece earlier this week in which they asked several “experts” to evaluate a Wikipedia article on their particular subject. They all more or less agreed that, while what’s up there is not insubstantial, there’s still a long way to go. The biggest challenge then, it seems to me, is to get these sorts of folks to give Wikipedia more than just a passing glance. To actually get them involved.
For this to really work, however, another group needs to get involved: the users. That might sound strange, since millions of people write, edit and use Wikipedia, but I would venture that most are not willing to rely on it as a bedrock source. No doubt, it’s incredibly useful to get a basic sense of a subject. Bloggers (including this one) link to it all the time — it’s like the conversational equivalent of a reference work. And for certain subjects, like computer technology and pop culture, it’s actually pretty solid. But that hits on the problem right there. Wikipedia, even at its best, has not gained the confidence of the general reader. And though the Wikimaniacs would be loathe to admit it, this probably has something to do with its core philosophy.
Karen G. Schneider, a librarian who has done a lot of thinking about these questions, puts it nicely:

Wikipedia has a tagline on its main page: “the free-content encyclopedia that anyone can edit.” That’s an intriguing revelation. What are the selling points of Wikipedia? It’s free (free is good, whether you mean no-cost or freely-accessible). That’s an idea librarians can connect with; in this country alone we’ve spent over a century connecting people with ideas.
However, the rest of the tagline demonstrates a problem with Wikipedia. Marketing this tool as a resource “anyone can edit” is a pitch oriented at its creators and maintainers, not the broader world of users. It’s the opposite of Ranganathan’s First Law, “books are for use.” Ranganathan wasn’t writing in the abstract; he was referring to a tendency in some people to fetishize the information source itself and lose sight that ultimately, information does not exist to please and amuse its creators or curators; as a common good, information can only be assessed in context of the needs of its users.

I think we are all in need of a good Wikipedia, since in the long run it might be all we’ve got. And I’m in now way opposed to its spirit of openness and transparency (I think the preservation of version histories is a fascinating element and one which should be explored further — perhaps the encyclopedia of the future can encompass multiple versions of the “the truth”). But that exhilarating throwing open of the doors should be tempered with caution and with an embrace of the parts of the old system that work. Not everything need be thrown away in our rush to explore the new. Some people know more than other people. Some editors have better judgement than others. There is such a thing as a good kind of gatekeeping.
If these two impulses could be brought into constructive dialogue then we might get somewhere. This is exactly the kind of conversation the Wikimedia Foundation should be trying to foster.