Topics/Themes

Hi NUDHL Gang!

When we had our last text-based discussion, I found the submissions by members about themes and ideas they wanted to discuss very helpful (in case you didn’t notice when I combined them and made that list for our discussion).

As you get ready to come over to Kaplan, if anyone has anything they found more relevant or interesting and want to be sure gets attention today, please post–even if it’s list and author style like I did–before the meeting. I think this will help structure our discussion.

My initial contribution is to keep on the table the discussion some of us had after Ben Pauley:

-what constitutes scholarship

-how does DH connect or conflict with the notion of individual scholarship (hactivism could be considered along with this)

And then, from the Gold readings, which are rich with great questions:

-where can we find non-whiteness in DH? Why, as Tara asked, is embedded whiteness present? What are it’s effects? What are ways to deal with systemic white privilege in DH?

-Not only race, as George Williams suggests, but disability, gender, sex, ethnicity, nation–these have strange positionings both in code and at the user interface. How do we acknowledge this? How do we change this?

-Here’s some things I would like to unpack:

What happens when we choose a specific practice–all of which are part (or, as some argue, not part) of DH;

media specific interpretation (I call it textual analysis or close reading) ala McPherson’s UNIX example;

labor/capitalism/neo-liberal presumptions; practices such as performance (hacking in Losh) vs interpretation (i.e. textual analysis);

everyday practices of computing vs constructing the basis and logics of tools functioning (McPherson and Williams, at least);

recognition that DH must do analysis that is media specific, not use old tools of film, etc;

thinking not just about digital divide or images/info collected about non-dominant groups (race, disabilities) but inclusion in logics;

choosing not to code (Posner) because of who one is (embodiment);

rethinking embodiment, self, divisions as suggested via disability).

I’m really excited about our discussion because the encoded race, sex and gender in digital technologies is at the heart of most of what I teach. I’ll share a hardware image to give you an example. It is of male and female ports. Ports are named male and female in order to fit (guess which one is male). The same is true with hard drives. The one that turns on your computer is “master” and other hard drives available to the computer are “slaves.” Apple recently revised this terminology, but it was there until about 2005.

How do hardware stereotypes reflect the DH issues we’ll be talking about today? We are responsible for DH, if not hardware, so this is a question of profound importance to me.

Looking forward to a lively discussion!Ports are called male and female to fit together!

Re-cataloguing Defoe

Hey everyone, it’s a little late, I know, but Michael asked me and I agreed it would be a good idea to write about the first research presentation of 2013 in our NUDHL workshop.

Professor Ben Pauley came from the Eastern Connecticut State University to Northwestern to present a new tool he is now developing. As a member of the Defoe Society, he is developing a tool to catalogue all work that might have been written by Defoe. Just as every author in 18th century, there’s always a mystery involving the writer’s persona, the actual works. Scholars normally use word patterns to claim the authorship of an Unknown/ unsigned text, coming to the question : “If not by (DEFOE), than who?”

If we ignore possibilities of plagiarism or of having influence over other writers of his own time (as many scholars do), the number of texts attributed to Robinson Crusoe’s writer is of astronomic proportions. And that’s what Ben is trying (really, the tool is already in a soft trial, one could say) to gather with his tool. I really wish we had his visual presentation, specially the way he is thinking the cataloguing process to create a better tool. Keywords such as “work” gain a complete new meaning, especially if our reference is the cataloguing “manual” of the library of Congress.

Apart from trying to understand his new categories, we had a vivid discussion about calling the development and launching of this tool a scholarly work/ publication or not. I guess we all have been discussing it for at least a quarter now, right? If other work will be enabled by it, if the scholarship in Defoe’s studies will profit from it, and maybe be developed in ways that would otherwise be impossible, why not? His (Ben’s) first impulse was to say he would not list it as part of his scholarly publications…

Also, as co-founder of Eighteenth-Century Book Tracker (www.easternct.edu/~pauleyb/c18booktracker), an index of freely available facsimiles of eighteenth-century editions, Ben seemed a little skeptical about the future of collaborative platforms. He noted that the contributions were not as abundant as he expected. In our discussions trying to understand why, one of the hypothesis were the search for originality when, again, you want to publish a work. So people would be less willing to share their findings during research, because they can become primary materials. Well, we all know it’s how you read it and use it that counts, but the risk is considered to big. Authorship, originality and our very beloved copyright, again, ladies and gentlemen!

Digital_Humanities

An excellent new monograph is now out by Joanna Drucker & co. titled Digital_Humanities (and, as I hoped when I picked up the title, that underscore IS important). I can’t recommend it enough for those new to DH and those long-affiliated with it (and for the so many of us in between). The best part is that the authors have put their theory to practice and released a beautifully designed, semi-experimental book that also has an open access edition for download here: http://mitpress.mit.edu/books/digitalhumanities-0

I’m about finished with it, and if others are reading it too, maybe a NUDHL spin-off book club may be in order.

 

NUDHL 4: Critiquing the Digital Humanities, Fri, 1/25/13, 12-2pm, AKiH

Please join us for the fourth NUDHL research seminar of the year on Friday, 1/25/13, 12-2pm in the seminar room of the Alice Kaplan Institute for the Humanities.

Here are the details on readings and location.

Hope to see you there!

Martin Mueller on “Morgenstern’s Spectacles or the Importance of Not-Reading”

X-Posted from Martin Mueller’s Scalable Reading Blog:

[I recently stumbled across the draft of a talk I gave at the University of London in 2008. It strikes me as a still relevant reflection on what then I called “Not-Reading” and now prefer to call “Scalable Reading.” I reprint it below with very minor corrections and additions.]

Coming from Homer: the allographic journey of texts and the query potential of the digital surrogate

For the past decade my work has revolved around what I call the ‘allographic journey of texts’ and the ‘query potential of the digital surrogate’. The stuff I am interested in has been around for a long time. I have written a book about the Iliad, another book about the transformation of Greek tragedy by European poets from 1550 to 1800, and I have written a number of essays about Shakespeare that never quite grew into a book.

None of my earlier work required or benefited from a computer. My first book was typeset on a computer, but the work on it was done in the North Library of the British Museum, taking notes by hand. The copy of my book on the Iliad was prepared on a dedicated word processor of the early eighties. Since the mid eighties I have written everything on a personal computer of some sort. Like everybody else, I don’t see how I could possibly do my work without a computer, but do I really write better as a result? If we had to return to pen and paper would we write worse, or even fewer, books?

On the other hand, Nietzsche, when shown an early typewriter, said “Unser Schreibzeug arbeitet auch an unseren Gedanken mit,” and it seems implausible that tool and content exist independently of each other.  The ‘what’ of a thing and the ‘how’ of its creation and reception are likely to be interwoven at some level.

My interest in technology was at first administratively driven. As chair of my English department in the eighties I took a strong interest in using technology to improve what seemed rather antiquated routines for creating or keeping records. We were the first humanities department to have a dedicated word processor, and later we had the first network that allowed faculty to print their stuff on a Laser printer from either a PC or a Mac. Big stuff in those days — at least in the context of English departments.

My scholarly interest in technology grew out of my work on Homer. What is the relationship between orality and literacy in the creation and transmission of the Iliad and Odyssey? I was from the beginning drawn to, and have never strayed from, the hypothesis that these works are hybrids and that their distinctive features result from the productive encounter of two different technologies of the word, the ‘oral’ and the ‘literate’.

The history of the Homeric poems is an ‘allographic journey’. I take the term ‘allographic’ from Nelson Goodman’s Languages of Art, where he distinguishes between ‘autographic’ works (Michelangelo’s David) and ‘allographic’ works, whether a Shakespeare sonnet or the score of Appassionata. The allographic work can always be written differently, and in theory, the rewriting makes no difference. But in practice, there is some difference if only because an allographic change is likely to involve a change in our mode of access to the work.

If we try to imagine Homeric verse in its original setting, the best evidence is probably Odysseus’ account of his adventures (Odyssey 9-12). It is a performance on roughly the same scale as a Shakespeare play or Verdi opera, with  an intermission and a spell-bound audience. That is very different from reading the Venetus A manuscript of the Iliad or the print edition of Venetus A, where the text is surrounded and sometimes  drowned by the marginal scholia. It is different again from reading the Iliad in a Teubner, Budé, or OCT version, where the consistency of format and typography across many authors of a canon facilitates access but also levels difference. You can and should abstract from the accidentals of presentation, and the more standardized the accidentals are the easier it is to perform that abstraction. But over time, the shared accidentals acquire a power of their own: if you spent a lifetime with the blue buckram Oxford Classical Text copies of Homer, Herodotus, and Plato you end up believing at some level that they were written that way, when in fact none of these authors could make head or tail of what they would see on any of those pages.

An interest in the conditions of reception led me to think about the role of surrogates. Our typical encounter with a text is through a surrogate — setting aside whether there is an original in the first place. Assuming that you own a ‘real’ Shakespeare folio and that it is closer to the ‘original’ you would still use a Bevington, Riverside, or Norton text most of time, and not only because you are afraid to damage it. Every surrogate has its own query potential, which for some purposes may exceed that of the original.

Thinking along those lines led me to the Chicago Homer. An oral style is shot through with echoes that the original audiences picked up over a lifetime of listening. The 19th century German school editions of Ameis-Hentze dutifully and comprehensively mark approximate or complete echoes, and with enough patience you can work your way through them. Could we use the Web to create a visual simulation of the network of repetitions and make modern readers ‘see’ what Homer’s audience heard? Starting from that question Craig Berry constructed a database of all repeated phrases, and Bill Parod wrote an interface that allowed you to filter and display repeated phrases while exploring the neural networks of bardic memory. You could also get very quick and accurate answers to questions that are very hard to ask in a print environment, such as “what repetitions are shared by the first and last books of the Iliad but occur nowhere else?”

My experience with the Chicago Homer shaped my view of what digital projects could or should do. In a worthwhile digital project you must do things that allow users to do things that are hard or impossible to do with the source object in its original form.

We are now in the midst of another and deeply consequential change in the allographic journey of texts. Between the late fifteenth and the mid sixteenth century an astonishing percentage of the European cultural heritage moved from circulating in manuscripts to circulating in print. There is a tipping point in changes of this kind. Once enough texts have migrated, the new medium comes to dominate circulation. What exists only in the old medium is increasingly ignored.

I had a striking demonstration of this last summer when I revised my 1984 book on the Iliad for a second edition scheduled to come out next year (2009). I did an online bibliographical search and then asked myself: “What do I miss if I restrict my reading of articles to items that are in Jstor and ignore stuff that exists only in print unless it is referred to repeatedly as importantly but not sufficiently well summarized in reviews or other discussion?” The answer to that is “not very much.” In many fields of the humanities, the allographic migration of journals to a digital medium has clearly gone beyond the tipping point.

With regard to the primary texts that are the focus of attention in the document-centric disciplines in the humanities, the latest phase in their allographic journey raises the question of ‘the query potential of the digital surrogate’. What can you do with the digital text that you cannot do with its printed source? What steps did the digitizers take to maximize its query potential in its new form? What new tools are available to take advantage of a properly digitized text?

A sermon on five texts

In talking about these questions, I’d like to take as my point of departure a handful of quotations that keep running through my mind. In ways that I don’t quite understand myself they set map out the field of my reflections. The first of them is a poem by Christian Morgenstern, an early twentieth century German poet famous for his nonsense poems, many of which bear witness to his philosophical and mystical leanings:

Die Brille The Spectacle
Korf liest gerne schnell und viel;
darum widert ihn das Spiel
all des zwölfmal unerbetnen
Ausgewalzten, Breitgetretnen.
Korf reads avidly and fast.
Therefore he detests the vast
bombast of the repetitious,
twelvefold needless, injudicious.
Meistens ist in sechs bis acht
Wörtern völlig abgemacht,
und in ebensoviel Sätzen
läßt sich Bandwurmweisheit schwätzen.
Most affairs are settled straight
just in seven words or eight;
in as many tapeworm phrases
one can prattle on like blazes.
Es erfindet drum sein Geist
etwas, was ihn dem entreißt:
Brillen, deren Energieen
ihm den Text – zusammenziehen!
Hence he lets his mind invent
a corrective instrument:
Spectacles whose focal strength
shortens texts of any length.
Beispielsweise dies Gedicht
läse, so bebrillt, man – nicht!
Dreiunddreißig seinesgleichen
gäben erst – Ein – – Fragezeichen!!
Thus, a poem such as this,
so beglassed one would just — miss.
Thirty-three of them will spark
nothing but a question mark.

 

second is a quotation by Father Busa, as posted by Willard McCarthy on the Humanist listserv:

“the use of computers in the humanities has as its principal aim the enhancement of the quality, depth and extension of research and not merely the lessening of human effort or time.”

The third is Ranganathan’s fourth law of library science:

Save the time of the reader.

The fourth is a quotation from Douglas Engelbart’s 1962 essay Augmenting Human Intellect, to which John Bradley drew my attention:

You’re probably waiting for something impressive. What I’m trying to prime you for, though, is the realization that the impressive new tricks all are based upon lots of changes in the little things you do. This computerized system is used over and over and over again to help me do little things–where my methods and ways of handling little things are changed until, lo, they’ve added up and suddenly I can do impressive new things. (p.83)

The final quotation comes from Laplace’s Essai philosophique sur les Probabilités:

On voit, par cet Essai, que la théorie des probabilités n’est, au fond, que le bon sens réduit au calcul; elle fait apprécier avec exactitude ce que les esprits justes sentent par une sorte d’instinct, sans qu’ils puissent souvent s’en rendre compte.

One sees in this essay that the theory of probability is at bottom nothing but common sense reduced to calculus” it makes you appreciate with exactitude what judicious mind have sensed through a kind of instinct without often being able to account for it.

Morgenstern and the prospects of a wide document space

Morgenstern’s Spectacles offer a nice way of focusing on the most distinctive query potential of digital text archives: their increasing size and their attendant promise to support analytical operations across far more text than you could possibly read in a lifetime. In the world of business, science, and espionage, the condensing power of digital spectacles is ceaselessly at work, extracting bits of knowledge from vast tailings of bad prose.

Google specializes in what you might want to call Morgenstern’s goggles. It lets you look for very small needles in very large haystacks. If there ar many needles and you are like the Prince of Arragon in the Merchant of Venice Google’s algorithms do a brilliant job of bringing you “what many men desire” and condensing on the first result page of millions of returns the hits that are most likely to be needed right now. There are many occasions in everyday life and scholarly contexts when this shallow but extraordinarily powerful search model works very well. But it is far from a complete model of the kinds of inquiry digital text archives are in principle capable of supporting.

The self-deprecating turn at the end of Morgenstern’s poem may be seen as a prophetic criticism of ‘knowledge extraction’. Why does the poem remain unreadable and in the aggregate yields just one question mark? Is that the fault of the poem, or do some things elude these spectacles or at least the ordinary uses of these spectacles?

Father Busa, Ranganathan and Douglas Engelbart

Let us come back to this question a little later and look at Father Busa’s observation from the perspective of Ranganathan and Douglas Engelbart. It ought to be true that making things easier or faster is not good enough. What is the point for a university to spend money on digital tools and resources if it does not produce better research by extending the scope of materials or methods and deepening the focus of inquiry?

But doubts arise if you look at this statement from the perspectives of Ranganathan’s fourth law of library science and Douglas Engelbart’s famous essay “Augmenting human intellect”. Ranganathan’s fourth law says simply “Save the time of the reader.” Much of what librarians do falls squarely under this heading. If books are neatly catalogued, the catalogue records are kept in filendrawers, the books are kept on clearly labeled shelves, and are reshelved promptly after use, readers can minimize the time cost of locating books and spend the saved time on reading them. In the progressive digitization of research libraries Ranganathan’s fourth law has found many new and powerful applications.

If you are ascetically inclined you might argue that scholarship is like the pinot noir grape and will produce its best results only under adverse conditions, whether in Burgundy or Oregon. There may be a downside to things being too easy, but much good research is hampered by things that take too long, are too expensive, or involve to much hassle. Thus the “mere” “lessening of human effort or time” certainly has the potential for enhancing the “quality, depth and extension of research.” Whether it will necessarily do so is of course another question.

Questioning the distinction between mere convenience and transformative changes is the major point of “Augmenting human intellect” by Douglas Engelbart, the inventor of the computer mouse. His insistence on the cumulative impact of “lots of changes in the little things you do”  is an example of Divide and Conquer at its best. Transformation is the result of incremental and typically minuscule change. When we try to evaluate the impact of digitization on scholarship in the humanities it is important to keep that truth in mind.

Rachel’s salamanders

It is not easy to measure that impact. Everybody uses computers in the ordinary course of university work. In just about all disciplines key generic research activities have gone digital. The bibliographical control of secondary literature and access to the journal literature has largely become an online business. This is of course part of research, but in a stricter sense “research computing” involves ways in which researchers use computers to manipulate their primary objects of attention. More accurately, computers never manipulate objects directly. They manipulate bits of the world translated into the ‘bits’ of binary digits. The key concept here is the “query potential of the digital surrogate.” Disciplines differ remarkably in their use of and dependence on digital surrogates. Some aspects of the actual or potential use of such surrogates in the humanities are well illustrated by a look at evolutionary biology, a discipline that has many structural and historical affinities with philology.

I know a little about the Museum of Vertebrate Zoology at Berkeley because my daughter worked there for a while. You can walk along shelves and shelves of salamander specimens, meticulously prepared and labeled by generations of field biologists going back to the 1800′s. These are, if you will, surrogates of living animals, and the labels‐‐metadata in current jargon‐‐are a minimal representation of their environment. Working with such specimens, with or without a microscope, is not unlike working with books.

There are projects to digitize such collections by creating digital representations ofthe specimens and by harmonizing the metadata across collections so that thespecimens at Berkeley and in the Field Museum exist in a single “document space,” searchable by scientists anywhere anytime.

Such a document space is a new kind of digital surrogate that makes many inquiries more convenient. Whether it enables “new” inquiries is open to question. You could after all think of yourself as imitating Jack Nicholson in Prizzi’s Honor, shuttling by plane between Chicago and Berkeley and rewarding yourself with dinners at Chez Panisse or Charlie Trotter’s on alternate nights. On the other hand, for a graduate student on a modest budget somewhere in Transylvania the query potential opened up by this digital surrogate may be the gateway to a successful career.

As part of her work, my daughter extracted DNA from some of these specimens, fed the DNA sequences into a collaborative gene bank, and used a comparative analysis of sequence to formulate new hypotheses about the descent of certain kinds of salamander families. As a generic research problem, this is a very familiar story to any humanities scholar who has ever traced the affiliations of different representations of the “same” object, whether a text, a score, or an image. But this particular representation of salamanders, and the subsequent manipulation of that representation are impossible without digital technology. You either do it with a computer or you do not do it all.

Over the course of my daughter’s career as a graduate student the time cost of analyzing DNA sequences on a computer dropped from weeks to hours. Ten years earlier, her work would for all practical purposes have been impossible. If you take away the computer you cripple projects like Rachel’s.  Such projects certainly  meet  Father Busa’s requirement that the computer affect the “depth and extension of research.”

It is not clear how much resarch in the humanities either meets the definition of research computing in this stringent sense or depends on the digital manipulation of its primary objects. Thomas Carlyle rewrote his History of the French Revolution after a servant accidentally threw the manuscript into the fire. A modern historian might write a book about the same topic on a word processor, with the chapters carefully backed up on a remote server , the bibliography assembled conveniently with Endnote, the copyediting performed on the author’s digital files, and the book produced from them. Digital technology would not be of the “essence” of this project, however, if one could envisage the scholar producing a very similar book during an extended stay in Paris, taking notes by hand in the Bibliothèque Nationale, composing on a manual typewriter in some small apartment within walking distance, and enjoying the culinary and other pleasures of France during off‐hours. This scenario retains its charm.

One could or course argue that the use of the computer in this hypothetical example illustrates the cumulative power of Engelbart’s “little things.” On the other hand, how different would this research scenario be if our scholar’s digital tool were a mid-eighties portable Compaq with 256 K floppy disks and WordPerfect 4.2? Whichever you look at such projects, they stay within an emulatory paradigm.  There is nothing wrong with this, as long as you recognize that it is not the whole story.

The emulatory paradigm

In  the humanities, and more particularly in Literary Studies,  the use of digital technology remains for the most part tethered to an ‘emulatory model’. Humanist scholars typically encounter the objects of their attention in relatively unmediated forms. The texts, scores, or images they read or look at are rarely the originals, but the surrogates pretend to be close to the originals, and the scholar behaves as if they were. In this regard they differ from scientists or, for that matter, alchemists, who for centuries have dissolved and recombined the targets of their attention for the sake of truth, gold, or both.

Emulation — the same, only better — is a very explicit goal of digital technology in many of its domains. Word processors and digital cameras are obvious examples. The computer screen as a device for reading is an obvious example of failure, at least so far. Much of this failure is due to the low quality of screen displays. Some of it has to do with the deeply engrained readerly habit of seeing the written word in a rectangular that is higher than it is wide. I was very struck by this when I got a monitor that could swivel to portrait mode and emulate the basic layout of a page. All of a sudden it was much easier to read JStor essays on the screen. For reading purposes, the horizontal orientation of the typical computer screen may be the greatest obstacle to emulation. [The last paragraph was written before the Kindel, the iPad, and its many tablet successors.  When it comes to many ordinary forms of reading, these devices have been game changers.)

If you think of the prospects of textually based scholarship in the digital world, the emulatory paradigm is both a blessing and a curse. On the side of blessings, consider Early English Books Online (EEBO). Scholars who work with texts rarely work with the original. They typically work with a surrogate, whether an edition or a facsimile. Sometimes you use the surrogate because it is the only thing you’ve got. More often you use it because it is in various ways more convenient. I own a copy of the Norton facsimile edition of the Shakespeare Folio, which is a closer surrogate of the original than Bevington, Norton, or Riverside, however you define that elusive term, but I rarely use it, and I daresay I am not unusual in that regard.

Microfilm is the least loved of all surrogates, and it may well be the only technology that will be superseded without a retro comeback. For half a century scholars of Early Modern England have had access to microfilm surrogates of their source documents, and they did not “have to” go to the Bodleian or British Library to look at this or that. If you are lucky enough to be associated with a university that has subscribed to EEBO, you now have a much more convenient surrogate at your finger tips. EEBO gives you a surrogate of a surrogate: what you see inside your web browser is a digital image of the microfilm image of the source text. But this surrogate at two removes is much superior in many ways. If somebody has paid the subscription fee for you you can get at it anytime from anywhere, and the fact that the cataloguing ‘metadata’ are also online, makes it a lot easier to find what you need.

Some months ago I asked colleagues what difference digital technology made to their research. I particularly remember a colleague who immediately responded by saying “EEBO has changed everything.” EEBO illustrates the beneficial effects of increasing the velocity of scholarly information. What Father Busa calls the “[mere] lessening of human effort or time” can and often does enhance the “quality, depth and extension of research”.

The curse of emulation

Now to the curse of emulation. If you wanted to take advantage of a word processor a quarter century ago you could not do so without learning something about the ways in which computers process texts, and you needed to familiarize yourself with the ‘command line language’ of some operating system and program. You could not easily learn this in a day, but it was not rocket science, and over the course of a few weeks you you could become quite competent in manipulating textual data in digital form. And you would constantly be reminded of the difference between a printout, a screen image, and the digital facts that underlie both.

There were strong incentives for learning how to process words in such an environment: if you knew how to do it revision became a much simpler task. Moreover, a word processor would automatically renumber your footnotes. I vividly remember conversations in the eighties — the heyday of deconstruction — with job candidates from Yale. They all learned enough to babysit the processing and printing of their dissertations on the university’s mainframe computer because the associated costs and anxieties were far outweighed by the benefits of having your footnotes automatically renumbered. Occasionally, however, footnotes would show up in odd places on the page, and we would joke about ‘ghosts in the margin.’

You don’t know about these things anymore if you use Microsoft Word, and by and large today’s graduate students know a lot less about text processing than the students who used WordPerfect 4.2 or wrote their dissertations on a mainframe. Good riddance in one way, but a loss in another. The skills you acquired and maintained to do word processing in the old-fashioned way were an excellent platform for text analysis. What kinds of useful analytical operations can you perform on a properly structured digital text that are difficult or impossible to do with its print source or with a digital version that is limited to emulatory use? That is the question about the query potential of the digital surrogate. It is a question the implications of which are harder to see intuitively for today’s graduate students than for their precursors of the eighties who by necessity picked up more knowledge about how a computer goes about its business when it processes text.

The prospects of a wide document space

A few years ago I directed a Mellon sponsored project called WordHoard, which we called “an application for the close reading and scholarly analysis of deeply tagged texts.” It is fundamentally a concordance tool and supports the age-old philological task of “going from the word here to the words there.” It contains morphosyntactically tagged texts of Chaucer, Spenser, and Shakespeare. When I wrote to a number of colleagues about this tool, I received a replay from Harold Bloom that read

Dear Mr. Mueller:
I am a throwback and rely entirely on memory in all my teaching and writing.

Harold Bloom

This amused me because I had been telling my students that WordHoard was a useful tool unless you were Harold Bloom and knew Shakespeare by heart. It is probably the case that Harold Bloom remembers more poetry than many scholars have ever read. He might have said the same thing to the monk who showed him his prototype of a concordance. If you have it “by heart” you are the con-cordance. But most of us are grateful to the likes of Douglas Engelbart for the mechanical “tricks” that are “used over and over again to help [us] do little things.” Not even Harold Bloom can read or remember everything, and for most of us the limits of human memory manifest themselves much earlier.

However closely we read individual texts, interpretation is always a form of contextualizing or of putting a particular detail within a wider frame that gives or receives meaning from an act of ‘focalization’. This is a fundamental and recursive procedure that operates from the lowest level of a single sentence through the parts of a work to the level of author, genre, and period.  I recently taught a course on Verdi and Dickens as the melodramatic masters of the 19th century. Traviata and Bleak House, both of them published in 1853, were the central works that highlight the deep shadows haunting 19th century progress. I drew a link between these works and the sudden English interest in Schopenhauer’s work, which was stimulated by John Oxenford’s essay of the same year in the Westminster Review.

This is a very conventional way of looking at 19th century literature, but it shows the progressive contextualization that human readers are very good at. They form large pictures by connecting relatively few dots in striking ways. This virtue is born from necessity. Whether or not the human brain operates like a computer, it is much slower and can perform at most 200 ‘cycles’ per second. To students of artificial intelligence it is a miracle how a person can look across the street and in less than a second can spot a familiar face in a crowd. There appears to be no way in which a computer can perform such an operation in 200 or for that matter 200,000 sequential steps. This enormous capacity of human intelligence to draw useful conclusions from very limited clues is also its greatest weakness. I like to say that paranoia, or the compulsion to connect all dots, is the professional disease of intelligence.

Large digital text archives offer the promise of complementary forms of contextualization. Human readers of the opening words of Emma (“Emma Woodhouse, handsome, clever, and rich”) immediately recognize them as an instance of the “three-adjective rule” that is common in Early Modern prose. Some readers might be curious whether a systematic look at many instances would reveal intersting semantic or prosodic patterns, whether some writers use this trick a lot, and what it might tell you about them. Gathering examples by hand is a very tedious process, and you would noit be likely to undertake it unless you had a strong hunch, backed up by salient detail, that something interesting is going on.

But now imagine that you have access to a very large linguistically annotated literary archive of the kind that corpus linguists have used for half a century. In a ‘linguistic corpus’ texts are not meant to be read by humans but processed by a machine. The texts have certain rudiments of readerly knowledge added to them. Every word declares that it is a plural noun, verb in the past tense — the very things that upset Shakespeare’s rebel peasant Jack Cade in his indictment of Lord Say:

It will be proved to thy face that thou hast men about thee that usually talk of a noun and a verb and such abominable words as no Christian ear can endure to hear

Literary scholars are likely to vary this into something like “talk of a noun and a verb and such tedious words as no literary ear can endure to hear.” Why make explicit what the critic’s innate ‘esprit juste’  perceives  ’par une sorte d’instinct’, to use Laplace’s terms?  The point, of course, is that the machine has no such instinct, but if properly instructed (which is itself a largely automatic process), it can within minutes range across a corpus of 100 million words or more and retrieve all sequences that follow the pattern ‘adjective, adjective, conjunction, adjective.’ If the archive was properly encoded and the access tool is sufficiently flexible, the outcome of such a search might be a list of all the sentences containing this pattern. If there are thousands of them, as there may well be, you may be able to group the sentences by time or the sequence of adjectives. An hour or two spent with such a list may be enough to tell you whether an interesting story is hiding in that list.

I once did this with 19th century fiction and have to confess that the results were rather nugatory. I did, however, discover that Charlotte Bronte was inordinately fond of this pattern. And not much time was spent or lost on this particular wild goose chase.

I have come to use the acronym DATA for such inquiries, which, to quote an IBM executive, put “dumb but fast” machines in the service of “smart but slow” humans.  DATA stands for ‘digitally assisted text analysis’, and the operative word here is ‘assisted’. Like ‘complementary contextualization’, the acronym DATA makes no claim that digitally assisted text analysis marks the end of reading or moves the business of interpretation to another realm. Let me conclude by turning to the  preface to a collection of his essays by the great German classicist Karl Reinhardt. Reinhardt throughout his career was torn between a deep allegiance to the ‘positivistic’ Altertumswissenschaft of Wilamowitz and an equally deep allegiance to the lapsed classicist Nietzsche (‘what use is the authentic text if I don’t understand it?’). The contested word in that continuing dilemma was ‘philology’.

Die Philologie wird sich selbst umso fraglicher, je weniger sie von sich lassen kan. Das heisst nicht, dass sie innerhalb ihres Bereichs an Zuversichtlichkeit verlöre. Heisst auch nicht, dass sie vor der Erweiterung ihrer Grenzen, die die geisteswissenschaftliche Entwicklung ihr gebracht hat, sich verschlösse. Aber es gehört zum philologischen Bewusstsein, mit Erscheinungen zu tun zu haben, die es transzendieren. Wie kann man versuchen wollen, philologisch interpretatorisch an das Herz eines Gedichts zu dringen? Und doch kann man philologisch interpretatorisch sich vor Herzensirrtümern bewahren. Es geht hier um eine andere as die seinerzeit von Gottfried Hermann eingeschärfte ars nesciendi.  Da ging es um Dinge, die durch Zufall, durch die Umstände der Überlieferung, durch die Unzulänglichkeit des Scharfsinns sich nicht  wissen liessen. Hier geht es um etwas notwendigerweise Unerreichliches, dessen Bewusstwerden doch rückwirkt auf das, was es zu erreichen gilt. Es handelt sich nicht um das letzt allgemeine Ignoramus, sondern um jene methodische Bescheidung, die sich klar is, immer etwas ungesagt lassen zu müssen, auch mit allem Merk- und Einfühlungsvermögen an das Eigentlich nicht herandringen zu können, nicht zu dürfen. (Tradition und Geist. Göttingen, 1960, p. 427)

Philology becomes more questionable to itself, the more it does what  it cannot stop doing. This does not mean that it loses its confidence within its own domain. Nor does it mean that it excludes itself from  the expansion of its borders that developments in the humanities have  opened to it.  But it is part of philological awareness that one  deals with with phenomena that transcend it. How can one even try to  approach the heart of a poem with philological interpretation? And  yet, philological interpretation can protect you from errors of the  heart. This is not a matter of the ars nesciendi that Gottfried Hermann insisted on in his day. There it was a matter of things you could not know because of accidents, the circumstances of transmissions, or of inadequate acuity. Here it is a matter of something that is necessarily beyond reach but our awareness of it effects our way of reaching towards it. It is not a matter of the ultimate Ignoramus but of a methodology modesty that is aware of something that must be left unsaid and that with all perceptiveness or intuition you cannot and should not trespass on.

 

Substitute ‘e-philology’,  ’Natural Language Processing’, or ‘quantitative text analysis’ for ‘philology’ and you have a wonderfully pertinent statement of what digitally assisted text analysis can and cannot do in the domain of  Literary Studies.  Notice that Reinhardt’s remarks about the expansion of philological borders comfortably include computers. Indeed there is irony in the fact that his contemporaries would have been enthusiastic adopters of the digital tools that the curent literary academy does not quite know what to do with. They are wonderful tools, as long as you follow Hamlet’s advice to the players and use them with “modesty and cunning.”

X-Post: Digital American Studies

X-post from my Issues in Digital History blog:

reviewing lauren frederica klein’s review, “american studies after the internet.”

Lauren Frederica Klein’s illuminating book review, “American Studies after the Internet,” published in the December 2012 issue of American Quarterly, examines a number of new works related to digital culture in order to ponder what a digital American studies might be. Oddly, Klein spends much time focusing on how we might comprehend, define, historicize, and conceptualize the digital, but she never quite does the same for American studies. On one level, this is fine. After all, she is writing a book review with limited space. But she misses an opportunity to use new digital work to also grapple with recent transformations in American studies itself as it has moved definitely toward a merging with ethnic studies and a more overtly leftwing political agenda in its scholarship. How do the recent changes in American studies themselves connect to the rise of the Internet and the digital?

Klein’s review implies an important, but often overlooked, parallel between digital computing and American studies. They were both born from the political and cultural dynamics of World War II and the Cold War. The development of digital computing, from uses of the Turing machine to the development of ENIAC (by women, as Klein points out, drawing on the work of Jennifer Light), to the Internet itself, received an enormous boost from federal support for the war effort in the United States and from the rise of the military-industrial complex during the decades immediately thereafter. Similarly, American studies, while already developing before the war (as were explorations of computing of course), also took off in the aftermath of World War II. The US government, corporations, and foundations sought out and supported narratives of American exceptionalism to accompany the rise of American global empire. So too, scholars and citizens (and more often than not scholar-citizens) grappled with this situation (for an excellent glimpse at this in both American studies and early British cultural studies, see Joel Pfister’s marvelous book Critique for What? Cultural Studies, American Studies, Left Studies).

So there is a very real historical connection here, one that both Klein and the writers whose books she reviews begin to explore: the digital and American studies have parallel, perhaps even intertwined, historical legacies in the American context of World War II and the Cold War. Perhaps this provides the background for why recent shifts in American studies have occurred at precisely the same time as the rise of the digital humanities. What it means to do American studies and what it means to pursue digital work collide around the struggle in recent United States history to grapple with a post-Cold War world.

The books Klein reviews suggest as much for the digital. All take a historically-informed and theoretically-inquisitive approach to the topic. Nathan Ensmenger’s The Computer Boys Take Over: Computers, Programers, and the Politics of Expertise retraces the ways in which what we might call the gender “codes” of computer coding underwent a considerable change: what began as feminized clerical labor somehow became, in more recent times, the highly masculinized world of adolescent hackers, nerds, developers, macho entrepreneurs, and brainy boys. Wendy Hui Kyong Chun’s Programmed Visions: Software and Memory probes the concept of software as a structuring digital media form that shapes knowledge and power in the contemporary world, creating hidden hierarchies within the very languages it uses, but also offering new opportunities for “intervention, action, and incantation.” Lisa Nakamura and Peter A. Chow-White’s edited volume, Race after the Internet, asks not just how the digital has affected race in a supposedly (but never decisively) postracial age, but also whether race itself has become digitized, a set of codes, networks, and indexes that guide existence in America today. Finally, Matthew Gold’s celebrated essay collection, Debates in the Digital Humanities, which probes this amorphous but emergent field through collaboratively peer-reviewed essays, blog posts, and responses, rounds out the books she reviews.

But what of when and where the digital intersects, or at least runs in parallel, or even at times is directly in tension with American studies?

First the intersections. Today, both fields are dialectically implicated in the rise of certain modes of neoliberal economics. Digital humanities finds itself at once part of the logic of neoliberalism and a field of desires and efforts to oppose neoliberalism’s relentless effort to break down Cold War-era institutions of democratic collectivity (deeply imperfect and flawed in their time, of course, but in retrospect powerfully potential sites of social change). The university, the public school system, the social welfare system, the state itself: the digital is supposed to “transform” these through the pastoral dream of technological solutions to social and political problems (hello Leo Marx?). The political question is whether the digital humanities will merely become a mechanism for further destroying institutions or, alternatively, will it reinvigorate their best aspects? Does the digital humanities’s focus on “data,” for instance, offer deeper paths to quality knowledge, learning, thinking, and living, or does it introduce quantification’s dangerous potential for dehumanization (some digital “humanities” that!)? Will DH’s repeated calls for collaboration make intellectual labor (not to mention the labor that produces the equipment undergirding the so-called Information Economy) even more precarious and undervalued than it is already? Or will DH’s new approaches to knowledge be able to unleash a new vision of social democratic political economies suitable to the cooperative work so many envision the digital enabling? How will the pulsating networks of the digital relate to the traditional social safety net of the welfare state? What should the hierarchies—if any?—be in a world of digital modularity?

The digital is a kind of possibility, but also a problem for humanists (including those who advocate post-humanism, I would argue) as they struggle under difficult conditions to make sense of their own particular disciplines, of academia as a whole, and of the relationship of their intellectual work to the larger political dynamics of the contemporary world, a place in which the Cold War discourses and assumptions that gave rise to the digital as we know it no longer rule.

Similarly, American studies finds itself a central field in which scholars, activists (and scholar-activists) are attempting to piece together the complex ideological, affective, and corporeal relationships among factors of race, gender, sexuality, ethnicity, region, and class as *both* cultural and material factors, but are doing so without a fully clarified geopolitical framework in which to pursue this project of conceptualization. One danger here is that as American studies scholars gain ground on diversity issues, they are merely absorbed into the neoliberal economics of the university. How to conjoin calls for inclusivity to calls for democratic transformation of institutions at their root?

Another challenge in contemporary American studies for the digital age is raised in a book edited by Brian Edwards and Dilip Gaonkar, also reviewed in the December issue of AQ. How, I understand Edwards and Gaonkar asking, might American studies scholars think about America outside the binary between exceptionalist and anti-exceptionalist frameworks, both of which quite sneakily reintroduce the ideological debates of the Cold War era into today’s post-Cold War context? Does the digital, when both historicized and examined in its current ideological logics, appearance as functional tools, and formations as a set of semiotic codes and affective regimes, offer one way in which to offer some responses to the question posed by Edwards and Gaonkar? Perhaps.

“Why isn’t American studies more digital?” Tara McPherson asked at the 2011 American Studies Association conference. As Klein’s review begins to indicate, one answer to that important question may be that it is and always has been more connected to the rise of the digital than we realize. As scholars continue to work back through the intertwined wires (or at least the ones running in historical parallel) of American studies and the digital during World War II and the Cold War, we may also be able to look around now, at the labs and centers, conferencing and “un”-conferencing, the thinking and feeling, the scholarly inquiry and political possibilities, with more clarity as well.

Opportunity: Cultural Studies Association Conference—2013 Digital Humanities Fellowship

Calling all HASTAC Scholars@NUDHL interested in cultural studies, here is an opportunity to participate in this year’s Cultural Studies Association conference in Chicago (don’t forget to x-post on NUDHL blog if you do this!):

The Cultural Studies Association (US) is looking for three HASTAC Scholars to participate in a pilot Digital Humanities Fellowship program. In particular, we are looking for three Scholars to help promote the work of CSA and to foster dialogue between the digital humanities and cultural studies. Using social media networks, DH Fellows will highlight relevant internet content (blog posts, news stories, videos, etc), promote CSA-sponsored events and projects (particularly our annual conference), and increase the organization’s visibility in digital spaces. In addition, DH Fellows will be asked to participate in a roundtable discussion on social networking and cultural studies at the 2013 CSA conference in Chicago, May 23-26, 2013. Registration fees for all Fellows will be waived.

If you would like to be considered for a Digital Humanities Fellowship, please email me (Megan Turner) at M2Turner@ucsd.edu at your earliest convenience. Fellowships will be awarded on a first-come-first-served basis.

Thanks,

Megan Turner

Program Coordinator

http://hastac.org/opportunities/cultural-studies-association-conference—2013-digital-humanities-fellowship

Ben Pauley, Building New Tools for Digital Bibliography @ NUDHL, Fri, 1/11/13, 12-2pm, AKiH

 “Building New Tools for Digital Bibliography: Constructing a Defoe Attributions Database for the Defoe Society”

Dr. Ben Pauley, Associate Professor, Eastern Connecticut State University

Friday, January 11, from 12 to 2 pm in the Alice Kaplan Humanities Institute seminar room, Kresge 2-360.

Lunch served!!

And don’t miss…

Unlocking the English Short Title Catalogue: New Tools for Early Modern and Eighteenth-Century Bibliography and Book History

A Digital Humanities Presentation to Students and Faculty by Ben Pauley, Associate Professor, Eastern Connecticut State University, NU Library Forum Room,
Thursday, January 10, 2013, 3:30 – 5:00 – Refreshments will be served.

The English Short Title Catalogue (ESTC) is the most comprehensive guide in existence to the output to published books in the English-speaking world during the era of handpress printing. With nearly 500,000 bibliographic records and information on more than three million library holdings, it is both the best census that we have of early British and American print and the best available guide to locating extant copies of those items.

Begun in the late 1970s, the ESTC was conceived from the first as an electronic resource, one that would leverage new developments in library technology to facilitate collaboration among scholars and librarians worldwide and one—crucially—that could be continuously revised and refined. In recent years, however, it has become clear that the ESTC is in need of fundamental transformation if it is to keep pace with a scholarly landscape that is being transformed by digitization.

Professor Pauley’s talk will highlight the challenges and opportunities facing the ESTC in its fourth decade, and will present the recommendations of a Mellon-funded planning committee for redesigning the ESTC as a 21st-century research tool. As envisioned, the new ESTC will stand at the intersection of librarianship, bibliography, and the digital Humanities, facilitating new kinds of enquiry in fields such as literary and cultural history, bibliography, and the history of the book.

This event is sponsored by Northwestern University Library’s Center for Scholarly Communication and Digital CurationNUL Special Libraries, and WCAS Department of English

œ

Professor Ben Pauley (Ph.D. Northwestern, 2004) specializes in eighteenth-century literature, with an emphasis on the works of Daniel Defoe. In addition to publishing essays and presenting papers in eighteenth-century literary studies, he has been involved in several digital projects, particularly concerning bibliography. He is the editor and administrator of Eighteenth-Century Book Tracker (www.easternct.edu/~pauleyb/c18booktracker), an index of freely-available facsimiles of eighteenth-century editions. He was co-principal investigator, with Brian Geiger (Director, Center for Bibliographical Studies and Research, University of California-Riverside), of “Early Modern Books Metadata in Google Books,” a recipient of a Google Digital Humanities Research Award for 2010–11 and 2011-12. He is a member of the board of the Defoe Society, serves on the technical review board for 18thConnect, and is an advisor to the recently-launched 18th-Century Common, a public Humanities portal for research in eighteenth-century studies.

 

X-Post: “The Material, Embodied, and Experiential Digital Humanities”

X-post from Issues in Digital History blog:

Bethany Nowviskie on the stakes of the digital humanities in 2013.

Lots of keen insights into the quickly-mutating practice (field?) of digital humanities from Bethany Nowviskie’s remarks (http://nowviskie.org/2013/resistance-in-the-materials/) at the recent MLA, including comments on:

  • making “tacit knowledge exchange” among practitioners more explicit.
  • bringing issues of structural inequality and exclusivity to the surface for continued recognition and discussion.
  • guarding against the “casualization of academic labor,” which Nowviskie argues “begets commodity toolsets, frictionless and uncritical engagement with content, and shallow practices of use.”

But I found most intriguing Nowviskie’s provocations, by way of William Morris, about the striking return of materiality (in all its senses) to digital humanities research:

Momentous cultural and scholarly changes will be brought about not by digitization alone, but by the development of ubiquitous digital-to-physical conversion tools and interfaces. What will humanities research and pedagogy do with consumer-accessible 3d fabrication? With embedded or wearable, responsive and tactile physical computing devices? What will we do with locative and augmented reality technologies that can bring our content off the screen and into our embodied, place-based, mobile lives? Our friends in archaeology and public history, recognizing the potential for students and new humanities audiences, are all over this. Writers and artists have begun to engage, as we can see next door in this year’s e-literature exhibit. And I believe that scholarly editors, paleographers, archivists, and book historians will be the next avid explorers of new digital materialities. But what might other literary scholars do? What new, interpretive research avenues will open up for you, in places of interesting friction and resistance, when you gain access to ?

These strike me as crucial questions, for they bring the digital back to earth, and suggest that when we interact with digital technologies, we are not departing from long-running epistemological and political questions of people, their critical thinking, and the quality of the lives they lead, but rather struggling to confront these issues anew.

Nowviskie proposes that the digital domain matters, in all senses of the word. It is not some separate la-la land, but rather as real as real can be, enfolding—and enfolded by—the very stark, sometimes beautiful, often ugly actual world we work in, with, and try to work through. In other words, as the binary between the virtual and the material gets reconfigured, the digital humanities becomes a key mode for addressing the disorientations that ensue. Thinking through the digital humanities offers a main frame for perceiving continuities and reaching toward, processing, even implementing better iterations of the past.

#MLA13

As many of you are probably aware, the annual MLA conference is currently underway. For the past few years, digital humanities panels and presentations have been on the rise at MLA, and this year the trend continues. Mark Sample has a post called Digital Humanities at MLA 2013 which collects all of the information on the DH-related sessions at MLA this year, and notes that there are a total of 66 (in 2010 there were only 27), or about 8% of all sessions. If you’re not attending, a good way to follow along is via Sample’s post and the #MLA13 hashtag on Twitter.

New issue of the Journal of Digital Humanities

The entire fourth issue of JDH is dedicated to the evaluation of digital humanities projects.

From the Editors:

With this fourth issue we wrap up the first year of the Journal of Digital Humanities, and with it, our first twelve months of attempting to find and promote digital scholarship from the open web using a system of layered review. The importance of assessment and the scholarly vetting process around digital scholarship has been foremost in our minds, as it has in the minds of many others this year. As digital humanities continues to grow and as more scholars and disciplines become invested in its methods and results, institutions and scholars increasingly have been debating how to maintain academic rigor while accepting new genres and the openness that the web promotes.

The entire issue can be accessed in multiple formats here: http://journalofdigitalhumanities.org/