Author: Dan Cohen

Open Access Publishing and Scholarly Values

[A contribution to the Hacking the Academy book project. Tom Scheinfeldt and I are crowdsourcing the content of that book in one week.]

In my post The Social Contract of Scholarly Publishing, I noted that there is a supply side and a demand side to scholarly communication:

The supply side is the creation of scholarly works, including writing, peer review, editing, and the form of publication. The demand side is much more elusive—the mental state of the audience that leads them to “buy” what the supply side has produced. In order for the social contract to work, for engaged reading to happen and for credit to be given to the author (or editor of a scholarly collection), both sides need to be aligned properly.

I would now like to analyze and influence that critical mental state of the scholar by appealing to four emotions and values, to try both to increase the supply of open access scholarship and to prod scholars to be more receptive to scholarship that takes place outside of the traditional publishing system.

1. Impartiality

In my second year in college I had one of those late-night discussions where half-baked thoughts are exchanged and everyone tries to impress each other with how smart and hip they are. A sophomoric gabfest, literally and figuratively. The conversation inevitably turned to music. I reeled off the names of bands I thought would get me the most respect. Another, far more mature student then said something that caught everyone off guard: “Well, to be honest, I just like good music.” We all laughed—and then realized how true that statement was. And secretly, we all did like a wide variety of music, from rock to bluegrass to big band jazz.

Upon reflection, many of the best things we discover in scholarship—and life—are found in this way: by disregarding popularity and packaging and approaching creative works without prejudice. We wouldn’t think much of Moby-Dick if Carl Van Doren hadn’t looked past decades of mixed reviews to find the genius in Melville’s writing. Art historians have similarly unearthed talented artists who did their work outside of the royal academies or art schools. As the unpretentious wine writer Alexis Lichine shrewdly said in the face of fancy labels and appeals to mythical “terroir”: “There is no substitute for pulling corks.”

Writing is writing and good is good, no matter the venue of publication or what the crowd thinks. Scholars surely understand that on a deep level, yet many persist in the valuing venue and medium over the content itself. This is especially true at crucial moments, such as promotion and tenure. Surely we can reorient ourselves to our true core value—to honor creativity and quality—which will still guide us to many traditionally published works but will also allow us to consider works in some nontraditional venues such as new open access journals, blogs or articles written and posted on a personal website or institutional repository, or non-narrative digital projects.

2. Passion

Do you get up in the morning wondering what journal you’re going to publish in next or how you’re going to spend your $10 royalty check? Neither do I, nor do most scholars. We wake up with ideas swirling around inside our head about the topic we’re currently thinking about, and the act of writing is a way to satisfy our obsession and communicate our ideas to others. Being a scholar is an affliction of which scholarship is a symptom. If you’re publishing primarily for careerist reasons and don’t deeply care about your subject matter, let me recommend you find another career.

The entire commercial apparatus of the existing publishing system merely leeches on our scholarly passion and the writing that passion inevitably creates. The system is far from perfect for maximizing the spread of our ideas, not to mention the economic bind it has put our institutions in. If you were designing a system of scholarly communication today, in the age of the web, would it look like the one we have today? Disparage bloggers all you like, but they control their communication platform, the outlet for their passion, and most scholars and academic institutions don’t.

3. Shame

This spring Ithaka, the nonprofit that runs JSTOR and that has a research wing to study the transition of academia into the digital age, put out a report based on their survey of faculty in 2009. The report has two major conclusions. First, scholars are increasingly using online resources like Google Books as a starting point for their research rather than the physical library. That is, they have become comfortable in certain respects with “going digital.”

But at the same time the Ithaka report notes that they remain stubbornly wedded to their old ways when it comes to using the digital realm for the composition and communication of their research. In other words, somehow it is finally seeming acceptable to use digital media and technology for parts of our work but to resist it in others.

This divide is striking. The professoriate may be more liberal politically than the most latte-filled ZIP code in San Francisco, but we are an extraordinarily conservative bunch when it comes to scholarly communication. Look carefully at this damning chart from the Ithaka report:

Any faculty member who looks at this chart should feel ashamed. We professors care less about sharing our work—even with underprivileged nations that cannot afford access to gated resources—than with making sure we impress our colleagues. Indeed, there was actually a sharp drop in professors who cared about open access between 2003 and the present.

This would be acceptable, I suppose, if we understood ourselves to be ruthless, bottom-line driven careerists. But that’s not the caring educators we often pretend to be. Humanities scholars in particular have taken pride in the last few decades in uncovering and championing the voices of those who are less privileged and powerful, but here we are in the ivory tower, still preferring to publish in ways that separate our words from those of the unwashed online masses.

We can’t even be bothered to share our old finished articles, already published and our reputation suitably burnished, by putting them in an open institutional repository:

I honestly can’t think of any other way to read these charts than as shameful hypocrisy.

4. Narcissism

The irony of this situation is that in the long run it very well may be better for the narcissistic professor in search of reputation to publish in open access venues. When scholars do the cost-benefit analysis about where to publish, they frequently think about the reputation of the journal or press. That’s the reason many scholars consider open access venues to be inferior, because they do not (yet) have the same reputation as the traditional closed-access publications.

But in their cost-benefit calculus they often forget to factor in the hidden costs of publishing in a closed way. The largest hidden cost is the invisibility of what you publish. When you publish somewhere that is behind gates, or in paper only, you are resigning all of that hard work to invisibility in the age of the open web. You may reach a few peers in your field, but you miss out on the broader dissemination of your work, including to potential other fans.

The dirty little secret about open access publishing is that despite the fact that although you may give up a line in your CV (although not necessarily), your work can be discovered much more easily by other scholars (and the general public), can be fully indexed by search engines, and can be easily linked to from other websites and social media (rather than producing the dreaded “Sorry, this is behind a paywall”).

Let me be utterly narcissistic for a moment. As of this writing this blog has 2,300 subscribers. That’s 2,300 people who have actively decided that they would like to know when I have something new to say. Thousands more read this blog on my website every month, and some of my posts, such as “Is Google Good for History?“, garner tens of thousands of readers. That’s more readers than most academic journals.

I suppose I could have spent a couple of years finding traditional homes for longer pieces such as “Is Google Good for History?” and gotten some supposedly coveted lines on my CV. But I would have lost out on the accumulated reputation from a much larger mass of readers, including many within the academy in a variety of disciplines beyond history.

* * *

When the mathematician Grigori Perelman solved one of the greatest mathematical problems in history, the Poincaré conjecture, he didn’t submit his solution to a traditional journal. He simply posted it to an open access website and let others know about it. For him, just getting the knowledge out there was enough, and the mathematical community responded in kind by recognizing and applauding his work for what it was. Supply and demand intersected; scholarship was disseminated and credited without fuss over venue, and the results could be accessed by anyone with an internet connection.

Is it so hard to imagine this as a more simple—and virtuous—model for the future of scholarly communication?

One Week, One Book: Hacking the Academy

[Reblogged from the THATCamp website. Please note that you don’t need to be a THATCamper to participate. We are soliciting submissions from everyone, worldwide. Join us by writing something in the next week, or if you’ve already written something you think deserves to be included, let us know!]

Tom Scheinfeldt and I have been brewing a proposal for an edited book entitled Hacking the Academy. Let’s write it together, starting at THATCamp this weekend. And let’s do it in one week.

Can an algorithm edit a journal? Can a library exist without books? Can students build and manage their own learning management platforms? Can a conference be held without a program? Can Twitter replace a scholarly society?

As recently as the mid-2000s, questions like these would have been unthinkable. But today serious scholars are asking whether the institutions of the academy as they have existed for decades, even centuries, aren’t becoming obsolete. Every aspect of scholarly infrastructure is being questioned, and even more importantly, being <em>hacked</em>. Sympathetic scholars of traditionally disparate disciplines are cancelling their association memberships and building their own networks on Facebook and Twitter. Journals are being compiled automatically from self-published blog posts. Newly-minted Ph.D.’s are foregoing the tenure track for alternative academic careers that blur the lines between research, teaching, and service. Graduate students are looking beyond the categories of the traditional C.V. and building expansive professional identities and popular followings through social media. Educational technologists are “punking” established technology vendors by rolling their own open source infrastructure.

“Hacking the Academy” will both explore and contribute to ongoing efforts to rebuild scholarly infrastructure for a new millenium. Contributors can write on these topics, which will form chapters:

  • Lectures and classrooms
  • Scholarly societies
  • Conferences and meetings
  • Journals
  • Books and monographs
  • Tenure and academic employment
  • Scholarly Identity and the CV
  • Departments and disciplines
  • Educational technology
  • Libraries

In keeping with the spirit of hacking, the book will itself be an exercise in reimagining the edited volume. Any blog post, video response, or other media created for the volume and tweeted (or tagged) with the hashtag #hackacad will be aggregated at hackingtheacademy.org. The best pieces will go into the published volume (we are currently in talks with a publisher to do an open access version of this final volume). The volume will also include responses such as blog comments and tweets to individual pieces. If you’ve already written something that you would like included, that’s fine too, just be sure to tweet or tag it (or email us the link to where it’s posted).

You have until midnight on May 28, 2010. Ready, set, go!

UPDATE: [5/23/10] 48 hours in, we have 65 contributions to the book. There’s a running list of contributions.

Spring 2010 Roy Rosenzweig Forum on Technology and the Humanities: The Library of Congress Twitter Archive

Beth Dulabahn, Director of Integration Management in the Office of Strategic Initiatives at the Library of Congress, will talk about the donation of the Twitter archive to the library. Beth was involved in the talks with Twitter and will provide some background and insight into this major digital acquisition. There will also be a general discussion of the value of the archive and related issues.

The Spring 2010 Rosenzweig Forum will take place Friday, May 21, at 3 pm, in Rm. 470 (the Center for History and New Media’s lab), in the Research I building, George Mason University (Fairfax Campus). Parking is available in the Sandy Creek Parking deck, right across from Research I. CHNM is on the 4th Floor. Directions to GMU: http://www.gmu.edu/resources/welcome/Directions-to-GMU.html

All are welcome to attend!

Digital Ephemera and the Calculus of Importance

[Thoughts prompted by an invitation to write a piece on the significance of “Notes, Lists, and Everyday Inscriptions” for The New Everyday, an innovative experiment in web publishing sponsored by MediaCommons. Since the editors of this edition of The New Everyday asked for something out of the ordinary for their curated collection, I thought it was time to unveil my Gladwell-esque theory of how criminal profiling and archival priorities share a mathematical foundation.]

How important are small written ephemera such as notes, especially now that we create an almost incalculable number of them on digital services such as Twitter? Ever since the Library of Congress surprised many with its announcement that it would accession the billions of public tweets since 2006, the subject has been one of significant debate. Critics lamented what they felt was a lowering of standards by the library—a trendy, presentist diversion from its national mission of saving historically valuable knowledge. In their minds, Twitter is a mass of worthless and mundane musings by the unimportant, and thus obviously unworthy of an archivist’s attention. The humorist Andy Borowitz summarized this cultural critique in a mocking headline: “Library of Congress to Acquire Entire Twitter Archive; Will Rename Itself ‘Museum of Crap.’

Few readers of this blog will be surprised to find that I take a rather different view of the matter. How could we not want to preserve a vast record of everyday life and thoughts from tens of millions of people, however mundane? (For more on my views of the Twitter/Library of Congress debate, and to inflate my ego, please consult articles from the New York Times, the Washington Post, and Slate.)

As any practicing historian knows, some of the most critical collections of primary sources are ephemera that someone luckily saved for the future. For example, historians of the English Civil War are deeply thankful that Humphrey Bartholomew had the presence of mind to save 50,000 pamphlets (once considered throwaway pieces of hack writing) from the seventeenth century and give them to a library at Oxford. Similarly, I recently discovered during a behind-the-scenes tour of the Cambridge University Library that the library’s off-limits tower, long rumored by undergraduates to be filled with pornography, is actually stocked with old genre fiction such as Edwardian spy novels. (See photographic evidence, below.) Undoubtedly the librarians of 1900 were embarrassed by the stuff; today, social historians and literary scholars can rejoice that they didn’t throw these cheap volumes out. As I have argued in this space, scholars have uses for archives that archivists cannot anticipate.

But let me set aside for a moment my optimistic disposition about the Twitter archive and instead meet the critics halfway. Suppose that we really don’t know if the archive will be useful or not—or worse, perhaps we are relatively sure it will be utterly worthless. Does that necessarily mean that the Library or Congress should not have accessioned it? I was thinking about this fair-minded version of the “What to save?” conundrum recently when I remembered a penetrating article about criminal profiling, which, of all things, helpfully reveals the correct calculus about the importance of digital ephemera such as tweets.

* * *

The act of stopping certain air travelers for additional checks—to give them more costly attention—is a difficult task riven by conflicting theories of whom to check and (as mathematicians know) associated search algorithms. Do utterly random checks work best? Should the extra searches focus on certain groups or certain bits of information (one-way tickets, cash purchases)? Many on the right (which is also home, I suspect, to many of the critics who scoff at the Twitter archive) believe in strong profiling—that is, spending nearly the entire budget and time of the Transportation Security Administration profiling Middle Easterners and Muslims. Many on the left counter that this strong profiling leads to insidious stereotyping.

A more powerful critique of strong profiling was advanced last year by the computational statistician William Press in “Strong Profiling is Not Mathematically Optimal for Discovering Rare Malfeasors” (Proceedings of the National Academy of Sciences, 2009). Press acknowledges that the issue of profiling (whether for terrorists at the airport or for criminals in a traffic stop) has enormous social and political implications. But he seeks to answer a more basic question: does strong profiling actually work? Or is there a more optimal mathematical formula for spending scarce time and resources to achieve the desired outcome?

Press examines two idealized mathematical cases. The first, the “authoritarian” strategy, assumes that we have perfect surveillance of society and precisely know the odds that someone will be a criminal (and thus worthy of additional screening). The second, the “democratic” strategy, assumes that our knowledge of people is messy and incomplete. In that case of imperfect information the mathematics is much more complex, because we can’t assign a reliable probability of criminality to each person and then give them security attention at an intensity commensurate to that value. It turns out that in the democratic case, the fuzzier mathematics strongly suggest a broader range of attention.

Moreover, even beyond the obvious fact that that the democratic model is closest to real life, the democratic algorithm for profiling is better than the authoritarian model, even if that state of omnipotent knowledge was achievable. Even if we had Minority Report-style knowledge, or even if we believed that the universe of potential criminals was entirely a subset of a particular group, it would be unwise to fully rely on this knowledge. To do so would lead to “oversampling,” an inefficient overemphasis on particular individuals. Of course we should pay attention to those with the maximum probability of being a criminal. But we also have to mix into our algorithm some attention to those who are seemingly innocent to achieve the best outcome—to stop the most crimes.

Through some mathematics we need not get into here, Press concludes that the optimal formula for paying attention to subjects is to avoid using the straight probability that each person is a criminal and instead use the square root of that value. For instance, if you feel Person A is 100 times more likely to be a terrorist than Person B, you should spend 10 times, not 100 times, the resources on Person A over Person B. Moreover, as our certainty about potential suspects decreases, the democratic sampling model becomes increasingly more efficient compared to the authoritarian model.

Although couched in the language of crime prevention, what Press is really talking about is the calculus of importance. As Press himself notes, “The idea of sampling by square-root probabilities is quite general and can have many other applications.”

* * *

As it turns out, the calculus of importance is the same for the Transportation Security Administration and for the Library of Congress. Press’s conclusions apply directly to the archivist’s dilemma of how to spend limited resources on saving objects in a digital age. The criminals in our library scenario are people or documents likely to be important to future researchers; innocents are those whom future historians will find uninteresting. Additional screening is the act of archiving—that is, selection for greater attention.

What does this mean for the archiving of digital emphemera such as status updates—those little, seemingly worthless online notes? It means we should continue to expend the majority of resources on those documents and people of most likely future interest, but not to the exclusion of objects and figures that currently seem unimportant.

In other words, if you believe that the notebooks of a known writer are likely to be 100 times more important to future historians and researchers than the blog of a nobody, you should spend 10, not 100, times the resources in preserving those notebooks over the blog. It’s still a considerable gap, but much less than the traditional (authoritarian) model would suggest. The calculus of importance thus implies that libraries and archives should consciously pursue contents such as those in the Cambridge University Library tower, even if they feel it runs counter to common sense.

So even if the skeptics are right and the Twitter archive is a boondoggle for the Library of Congress, it is the correct kind of bet on the future value of digital ephemera, the equivalent of the TSA spending 10% of their budget to examine more closely threats other than those posed by twentysomething Arabs.

The accessioning of the Twitter archive by the Library of Congress is not an expensive affair. Tweets are small digital objects, and even billions of them fit on a few cheap drives. Even with digital asset management, IT labor across time, and electricity costs, storing billions of tweets is economical, especially compared to the cost of storing physical books. University of Michigan Librarian Paul Courant has calculated [Word doc] that the present value of the cost to store a book on library shelves in perpetuity is about $100 (mostly in physical plant costs). An equivalent electronic text costs just $5.

This vast disparity only serves to reinforce the calculus of importance and archival imperatives of institutions such as the Library of Congress. The library and other keepers of our cultural heritage should be doing much more to save the digital ephemera of our age, no matter what we contemporaries think of these scrawls on the web. You never know when a historian will pan a bit of gold out of that seemingly worthless stream.

Roy Prize 2010

This Friday, May 14, 2010, is the deadline for submitting your digital history project for the Roy Rosenzweig Prize for Innovation in Digital History, known around here at the Center for History and New Media (one of the sponsors of the prize, along with the American Historical Association) simply as the Roy Prize. The prize is in honor of Roy Rosenzweig, the founder of CHNM and a pioneer in social history and digital history (and also the nicest person I’ve ever met), who passed away in 2007.

There was a little confusion last year (the inaugural year of the award) engendered by the use of the word “Fellowship” instead of “Prize,” and the AHA and CHNM hope that the revised name will make it crystal clear that the Roy Prize is for the best work in digital history, period. If you have a project that you feel is ready to be nominated for the award, please see the AHA’s page on how to submit your project for the prize.

The Last Digit of Pi – Video of My TEDxNYED Talk and Live Discussion

Here’s the video of my talk “The Last Digit of Pi,” given in New York City on March 6, 2010, at TEDxNYED. I’ll be discussing it live on Friday, May 7, at 3p EDT, on Twitter (follow me there or use the hashtag #tedxnyed to join in the discussion).

[youtube=http://www.youtube.com/watch?v=GoadUAsQTFc&w=425&h=350]

Eliminating the Power Cord

[My live talk at the Shape of Things to Come conference at the University of Virginia, March 27, 2010. It is a riff on a paper that will come out in the proceedings of the conference.]

As I noted in my paper for this conference, what I find interesting about this panel is that we got a chance to compare two projects by Ken Price: the Walt Whitman Archive and Civil War Washington. How their plans and designs differ tell us something about all digital humanities projects. I want to spend my brief time spinning out further what I said in the paper about control, flexibility, creativity, and reuse. It’s a tale of the tension between content creators and content users.

But before I get to Ken’s work, I’d like to start with another technological humanist, Jef Raskin, one of the first employees of Apple Computer and the designer, with Steve Jobs, of the first Macintosh. Just read the principles Raskin lays out in 1979 in “Design Considerations for an Anthropophilic Computer”:

This is an outline for a computer designed for the Person In The Street (or, to abbreviate: the PITS); one that will be truly pleasant to use, that will require the user to do nothing that will threaten his or her perverse delight in being able to say: “I don’t know the first thing about computers.”

You might think that any number of computers have been designed with these criteria in mind, but not so. Any system which requires a user to ever see the interior, for any reason, does not meet these specifications. There must not be additional ROMS, RAMS, boards or accessories except those that can be understood by the PITS as a separate appliance. As a rule of thumb, if an item does not stand on a table by itself, and if it does not have its own case, or if it does not look like a complete consumer item in [and] of itself, then it is taboo.

If the computer must be opened for any reason other than repair (for which our prospective user must be assumed incompetent) even at the dealer’s, then it does not meet our requirements.

Seeing the guts is taboo. Things in sockets is taboo. Billions of keys on the keyboard is taboo. Computerese is taboo. Large manuals, or many of them is taboo.

There must not be a plethora of configurations. It is better to manufacture versions in Early American, Contemporary, and Louis XIV than to have any external wires beyond a power cord.

And you get ten points if you can eliminate the power cord.

Many digital humanities projects implicitly believe strongly in Raskin’s design principle. They take care of what to the content creators and designers seems like hard and annoying work for the end users, freeing those users “to do what they do best.” These editorial projects bring together at once primary sources, middleware, user interfaces, and even tools.

Like the Macintosh, this can be a very good thing. I mostly agree with what Ken has just said, that in the case of Whitman, we probably cannot rely on a loose network of sites to provide canonical texts. Moreover, students new to Walt Whitman can clearly use the contextualization and criticism Ken and his colleagues provide on the Walt Whitman site. Similarly, scholars dipping for the first time into ethnomusicology will appreciate the total research environment provided by EVIA. As Matt Kirschenbaum noted in the last session, good user interfaces can enable new interpretations. I doubt that many scholars would be able to do Hypercities-grade geographical scholarship without a centralized Hypercities site.

But at the same time, like Raskin, sometimes these projects strive too hard to eliminate the power cord.

Raskin thought that the perfect computer would enable creativity at the very surface of the appliance. Access to the guts would not be permitted because to allow so would hinder the capacity of the user to be creative. The computer designers would take care of all of the creativity from the base of the hardware to the interface. But as Bethany Nowviskie discussed this morning, design decisions and user interface embody an argument. And so they also imply control. It’s worth thinking about the level of control the creators assume in each digital humanities project.

I would like to advance this principle: Scholars have uses for edited collections that the editors cannot anticipate. One of the joys of server logs is that we can actually see that principle in action (whereas print editorial projects have no idea how their volumes are being used, except in footnotes many years later). In the September 11 Digital Archive we assumed as historians that all uses of the archive would be related to social history. But we discovered later that many linguists were using the archive to study teen slang at the turn of the century, because it was a large open database that held many stories by teens. Anyone creating resources to serve scholars and scholarship needs to account for these unanticipated uses.

When we think through the principle of unanticipated uses, we begin to realize that there is a push and pull between the scholar and the editor. It is perhaps not a zero sum game, but surely there is a tension between the amount of intellectual work each party gets to do. Editors that put a major intellectual stamp on their collection through data massaging and design and user tools restrict the ability of the scholar to do flexible work on it. Alan Burdette of EVIA was thinking of this when he spoke about his fear of control vs. dynamism this morning.

Are digital humanities projects prepared to separate their interfaces from their primary content? What if Hypercities was just a set of KML files like Phil Ethington’s KML files of LA geography? What about the Grub Street Project? Or Ken’s Civil War Washington? This is a hard question for digital projects—freeing their content for reuse.

I believe Ken’s two projects, one a more traditional editorial project and one a labor of love, struggle with how much intellectual work to cede to the end user. Both projects have rather restrictive terms of use pages and admonishments about U.S. copyright law. Maybe I’m reading something into the terms of use page for Civil War Washington site, but it seems more half-hearted. You can tell that here is a project that isn’t a holding place for fixed perfected primary resources like Whitman’s, but an evolving scholarly discussion that could easily involve others.

Why not then allow for the download of all the data on the site? I don’t think it would detract from Civil War Washington; indeed, it would probably increase the profile of the site. The site would not only have its own interpretations, but allow for other interpretations—off of the site. Why not let others have access to the guts that Raskin wished to cloak? This is the way networked scholarship works. And this is, I believe, what Roger Bagnall was getting at yesterday when he said “we need to think about the death of the [centralized website] project” as the greater success of digital humanities.

Jim Chandler and I have been formulating a rule of thumb for these editorial projects: the more a discipline is secure in its existence, its modes of interpretation, and its methods of creating scholarship, the more likely it is to produce stripped-down, exchangeable data sets. Thus scholars in papyrology just want to get at the raw sources; they would be annoyed by a Mac-like interface or silo.  They have achieved what David Weinberger, in summarizing the optimal form of the web, called “small pieces, loosely joined.”

On the other hand, the newer and less confident disciplines, such as the digital geographic history of Civil War Washington, Hypercities, and Grub Street feel that they need to have a Raskin-like environment—it’s part of the process of justifying their existence. They feel pressure to be judge, jury and executioner. If the Cohen-Chandler law holds true, we will see in the future fewer fancy interfaces and more direct, portable access to humanities materials.

Of course, as I note in my paper, the level of curation apparent in a digital project is related to the question of credit. The Whitman archive feels like a traditional editorial project and thus worthy of credit. If Ken instead produced KML files and raw newspaper scans, he would likely get less credit than a robust, comprehensive site like Civil War Washington.

The irony about the long-suffering debate about credit is that every day humanities scholars deal with complexity, parsing complicated texts, finding meaning in the opaque. And yet somehow when it comes to self-assessment, we are remarkably simple-minded. If we can understand Whitman’s Leaves of Grass, surely we can tease out questions of credit and the intellectual work that goes into, say, complex KML files.

To help spur this transition along, Christine Madsen has made this weekend the important point that the separation of interface and data makes sustainability models easier to imagine (and suggests a new role for libraries). If art is long and life is short, data is longish and user interfaces are fleeting. Just look at how many digital humanities projects that rely on Flash are about to become useless on millions of iPads.

Finally, on sustainability, I made a comparison in my paper between the well-funded Whitman archive and the Civil War Washington site, which was produced through sweat equity. I believe that Ken has a trump card with the latter. Being a labor of love is worth thinking about, because it’s often the way that great scholarship happens. Scholars in the humanities are afflicted with an obsession that makes them wake up in the morning and research and write about topics that drive them and constantly occupy their thoughts. Scholars naturally want to spend their time doing things like Civil War Washington. Being a labor of love is often the best sustainability model.

Regional THATCamps: A Movement

During this year’s Winter Games in Vancouver, I joked that like the Olympics, THATCamp was a movement, not an event. Well, only semi-joked. I did think there was something to trying to spread widely a new participatory and inclusive academic meeting, one that would spread the digital humanities, encourage collaboration and sharing, and disseminate knowledge better than the standard panels-and-lectures scholarly society annual conference. I’m incredibly delighted that the Andrew W. Mellon Foundation has now generously provided funding to support a network of THATCamps worldwide so that the unconference can serve many more people than we can here at the mothership.

THATCamp: The Humanities and Technology Camp was founded in 2008 by the Center for History and New Media’s Dave Lester, Jeremy Boggs, and Tom Scheinfeldt. (Dave is now the assistant director of the Maryland Institute for Technology in the Humanities.) A strange and wonderful thing happened over the last year: THATCamps started to spring up across the country and beyond, many of them hosted and run by those who participated in the yearly event in Fairfax, VA.

I strongly recommend that readers of this blog read Tom’s full post on the grant and regional THATCamps to get a sense of what’s involved (you can also read my prior posts on THATCamp), and hope that you consider hosting your own THATCamp. You don’t have to wait four years to light the torch.

The Last Digit of Pi

[This is a rough transcript of my TEDxNYED talk, delivered on March 6, 2010, in New York City at the Collegiate School. TEDxNYED was an all-day conference “examining the role of new media and technology in shaping the future of education.” For a meta-post about the experience of giving a TED(x) talk, please read “Academic Theater (Reflections on TED & TEDxNYED).” What I actually said and did at TEDxNYED deviated from this transcript; I engaged the audience directly a couple of times, once for fun and once to get their ideas about the subject. I’ll post the video when it’s available.]

I want to tell you a story about a forgotten realm of education and knowledge. It is a cautionary tale, a parable of what happens when the world changes, when tradition is challenged.

Until relatively recently in human history, pi was the much sought-after solution to what was long called the “rectification” or “quadrature” of the circle, fancy words more easily symbolized by the diagram in this slide. How can you transform that circle into the overlaid square? One side of the square would be one-quarter of pi if the diameter of the circle is 1.

Pi was a coveted number for thousands of years, imbued with magical properties. Generations of scholars pursued it doggedly, often considering it the be-all and end-all of geometry.

This is a different pi—pi as we moderns know it:

Well, not all of it, as I’m sure you know. It’s just the first 200 or so digits. The number stretches on forever. I hope you weren’t expecting me to reveal the actual last digit of pi. Because there isn’t one. Strange, no?

Pi wasn’t always this strange. The ancient Egyptians knew better, pegging the ratio of the circumference to the diameter of a circle at 4 over 3 to the 4th power. That’s considerably more definite, and thus much more sensible.

Archimedes knew better, homing in on the value of pi between a couple of very close fractions.

If you are a biblical literalist, pi would seem to be 3, since the Bible clearly describes 30 cubits as encompassing a circle of 10 cubit diameter.

And the solutions kept coming. From ancient mathematicians and philosophers, to medieval scholars, to the Renaissance and the Enlightenment. Everyone seemed capable of finding—with enough effort—the exact value for pi. Squaring the circle was an effort of genius in an ancient science perfectly described centuries ago by Euclid.

But something changed radically in the eighteenth century, just after that book on the right by Joubert de la Rue. A few mathematicians started to take more seriously the nagging feeling that pi didn’t have a perfect solution as a magical fraction. It might not have a last digit after all. This critical number at the center of mathematics might, in fact, be irrational. One mathematician began to reconceptualize pi.

And there he is: the dapper Swiss German mathematician Johann Heinrich Lambert:

He was the son of a tailor, obviously, and was mostly self-taught in mathematics. His brilliant work in the 1760s showed that π/4 could not be a rational number—you could never exactly figure out the value of one side of that square—and thus that pi too was irrational. After Lambert, math textbooks declared the matter solved.

That’s right, problem solved…

Except….circle-squaring kept on going. The world of mathematics had changed with the discoveries of the eighteenth century but somehow the message didn’t get through to many people. John Parker, on the left, came up with my personal favorite solution: pi is precisely 20612/6561. Some circle-squarers, like James Smith on the right, mocked Lambert’s proof as the work of a dilettante.

Things then got testy between the new mathematicians and those who clung to the prior vision of pi. The record of this warfare is as informative as it is humorous. In the 1860s and 70s, James Smith took on Augustus De Morgan, a math professor in London, in a series of short pamphlets, which were the Victorian equivalent of Twitter.

But unsurprisingly, the castigations of professors of mathematics didn’t stop the circle-squarers. Their solutions kept on coming, even in the face of criticism, even after pi had been shown to be transcendental, meaning it couldn’t even be the root of some other number or equation. My favorite book from the turn of the twentieth century had this subtitle on the cover: “The great problem which has baffled the greatest philosophers and the brightest minds of ancient and modern times has now been solved by a humble American citizen of the city of Brooklyn.”

Now, it’s easy to laugh at these misguided circle squarers, especially when they’re from Brooklyn. But if you read circle-squarers seriously, and stop to think about it, they are not so different from you or me. Even in our knowing times, we all persist in doing things that others have long since abandoned as absurd or passé.

History tells us that people are, alas, not very good at seeing the new, and instead are very good at maintaining the past at all costs. This is particularly true in education: Euclid’s Elements, written over 2,000 years ago, was still a standard math textbook well into the 19th century, despite major mathematical advances.

So it’s worth pausing to think about the last digit of pi. Why did so many continue to pursue pi as it was traditionally conceived, and why did they resist the new math?

Think for a moment about the distinction between the old and the new pi. The old was perfect, simple, ordered, divine; the new, seemingly imprecise, prosaic, chaotic, human. So the story of pi is the story, and the psychology, of what happens when the complex and new tries to overtake the simple and traditional.

It’s happening all around us in the digital age. We’re replacing what has been perceived as perfect and ordered with the seemingly imprecise and chaotic.

Look at what has happened, for instance, in the last decade with Wikipedia and the angst about the fate of the traditional Encyclopedia.

Or newspapers in the face of new forms of journalism, such as blogging. A former baseball statistician, Nate Silver of FiveThirtyEight.com, can brazenly decide to analyze elections and economy better than most newspapers? Yes indeed.

Now this audience, hip to the right side of these screens, may want to be as mean as Augustus De Morgan to those still on the left. We may want to leave modern circle-squarers behind, and undoubtedly some of them will be left behind. But for the majority who are unsettled and are caught between the old and the new, we need other methods to convince them and to change the status quo. History tells us it’s not enough to say that people are blind to the future. We have to show precisely what the weaknesses of the old are…

…and we have to show how the new works better than the old.

Knowing pi correctly to the 10th digit is enormously helpful when accurately predicting the movements of heavenly bodies; try using James Smith’s 3 1/8 when tracing the arc of a planet or moon. For some physics, knowing pi accurately to the 40th digit is critical.

Moreover, this modern pi may be strange, but its very strangeness opened up new avenues of research and thought that were just as intellectually challenging and rewarding as squaring the circle. The transcendental nature of pi led mathematicians to ponder infinite sequences of fractions and had an impact on chaos theory. In computer science, coming up with algorithms to reach a billion or trillion digits of pi as quickly as possible advanced the field. And, if you still want an unsolved problem to crack, see if you can figure out if pi is what is called a “normal number,” where the distribution of the digits 0-9 is uniform…

…or is there instead a preponderance of eights. Now that’s a tough problem, related to real issues in modern math. So there are still problems to be solved, more advanced problems. Math didn’t end with the end of the old pi—it just moved in new, more interesting directions.

But to get to that point, mathematicians had to show in a comprehensible way how the new pi created a new order.

Academic Theater (Reflections on TED & TEDxNYED)

This past weekend’s TEDxNYED event in New York took place in the theater of a school just off Broadway. I couldn’t help thinking about the symbolism of that location during the day’s proceedings. TEDx, a spinoff regional program of the billionaires-and-brains edutainment summit in California, TED, pushes speakers like me towards theatrics.

TEDxNYED was enjoyable and I greatly appreciated the opportunity to rub elbows with some digital luminaries and some very smart educators who are doing all the hard work in the trenches while I sit here in the ivory tower blogging. Whatever criticisms may be leveled, TEDxNYED was incredibly well-run and engaging. Before you read my thoughts below, you should first read the wrap-up from Dave Bill, the TEDxNYED “curator,” who gets it exactly right. I’m enormously appreciative of Dave’s hard work and the hard work of his TEDxNYED colleagues.

Back to Broadway: among other things, TEDxNYED gave me a chance to think more about the academic lecture as theater. (It also gave me a welcome chance to summon the vaudevillian genes of my New York Jewish heritage, the effectiveness of which you will be able to assess when the video is posted to the TEDx channel on YouTube in a couple of weeks.)

Take Larry Lessig, the de facto headliner of TEDxNYED. He’s clearly a first-rate legal scholar and influential activist. But after viewing him live, I realized more than ever that he’s also a rather talented performance artist, with crack comedic timing. (Here’s his talk; judge for yourself.)

We professors don’t like to admit it, but comedy and performance are important ingredients in most successful academic lectures, and can spur the pursuit of knowledge and action far better than serious monograph or article. When I was in college nearly everyone interested in history—from any era or place—took Stephen Cohen’s class on Soviet history, mostly because he was entertaining. He even had one lecture consisting entirely of jokes. Sure, it was gimmicky. But I also know several of my classmates who went into careers in diplomacy and history because of the inspiration.

Of course, academic theater can also lead to problems. TED talks are limited to 18 minutes, inevitably leading to reductionism. As I quipped in my talk on the 6,000 year history of π, “Portions have been condensed.” The humanities particularly suffer from this condensation. For instance, as hugely entertaining as Lessig’s talk was, if you watch it I’m sure you’ll pick up that it conflates, quite problematically, two kinds of conservatism: religious conservatism and libertarianism. Just because the Cato Institute can imagine a role for remixes doesn’t mean that those who attend free church potlucks can. Modern conservatism is an extraordinarily complex mix; one need only look at the tension between libertarian and evangelical views of homosexuality. Gina Bianchini, the CEO of Ning, a network of social networks, presented her work as “the joy of connecting optimists from around the world,” leaving out the fact that the history of Ning is far more interesting: it started out as an engine for making web apps, only later turning toward social networking. That’s actually a fascinating, complex business history that I would have liked to hear more about.

TED’s tagline is the catchy “Ideas Worth Spreading.” I’m an intellectual historian and appreciate the emphasis on ideas; as an educator I’m in favor of spreading knowledge. But in my later years I’ve also come to realize that while ideas are important, execution is probably more important. Lessig and Bianchini also know this—Lessig is now working on methods of more effective lobbying and Bianchini is obviously a talented CEO—and it would have helped TEDxNYED if they had explained to the audience the nitty-gritty details of making real change and progress. It doesn’t come from clever sound bites.

The TED spotlight-on-the-stage format also encourages the audience to perceive the speakers as isolated geniuses, coming out to impart wisdom. The host who introduced me credited me as being the solitary creator of several projects and works, all of which were actually broad collaborations. Again, collaboration is more complex than the format allows. Jeff Jarvis decided to blow up the format by getting up on stage with the lights on and ranting about the insanities and inanities of modern education. This was effective in a Lenny Bruce sort of way, but like Bruce, it was the exception that proved the rule that we speakers were bound to a certain form of academic theater. Inspired by Jarvis, I broke the fourth wall and interacted with the audience a couple of times during my talk, but it was perhaps a little superficial.

Regardless of these criticisms—which I give, again, entirely in recognition of the success of the event and with an eye toward improvement for next year—I enjoyed the challenge of doing a TED talk. I’m working on a much more formal Big Lecture at Cambridge University, and TEDxNYED helpfully made me think about the problems with that format as well. Indeed, I’m not blaming TED for the problems of academic theater. I actually believe the fault lies with academics themselves, who have ceded the ground of public intellectualism in the past generation or two, leaving a vacuum that TED and TEDx are happy to fill.

Hopefully—and judging by the tweets and blog posts this is true—the attendees took away more of the advantages than the disadvantages of the format, and will go on from thought to action.

[photo credit: Kevin Jarrett]