You are looking at content from Sapping Attention, which was my primary blog from 2010 to 2015; I am republishing all items from there on this page, but for the foreseeable future you should be able to read them in their original form at sappingattention.blogspot.com. For current posts, see here.

Why Digital Humanists don't need to understand algorithms, but do need to understand transformations

Jul 20 2016

Debates in the Digital Humanities 2016 is now online, and includes my contribution, Do Digital Humanists Need to Understand Algorithms? (As well as a pretty snazzy cover image) In it I lay out distinction between transformations, which are about states of texts, and algorithms, which are about processes. Put briefly:

Put simply: digital humanists do not need to understand algorithms at all. They do need, however, to understand the transformations that algorithms attempt to bring about. If we do so, our practice will be more effective and more likely to be truly original.

It then moves into one case study; the Jockers-Swafford debate of 2015, large parts of which hung on whether the Fourier transform was a black box and how it its use as a smoothing device might be understood. Its like a lot of whats on this blog, only better thought and edited.

The transformation/algorithm distinction is not a completely firm one, but I have found it extremely useful in a lot of research and teaching problems Ive approached over the last year. So in addition to advertising that article for your consumption/fall syllabi production, I wanted to take the occasion to put on github a tiny little germ of a project to provide one-page, transformation-oriented introductions to basic text-analysis concepts that came out of using this thinking for a workshop on text analysis at the NIH in Bethesda, and describe whats in it. Id love for anyone else to use it, fork it, whatever.

The canonical example of transformation/algorithm I give in the article is sorting. Sortedness is a transformation you can understand; quicksort is a particular algorithm that puts a list into sorted order. If you want to use, say, a concordance to the works of Thomas Aquinas, you must understand sortedness and what it does; but the precise sorting algorithm is unimportant.

Saying we dont need to understand algorithms doesnt mean we shouldnt. There are many cases where we should. But the need to understand the basic transformations attempted is a bare minimum anyone reading an article or performing an early exploratory step should be able to do.

This has shaped the way I introduce algorithmic concepts for humanities audiences.

Take, for example, how to introduce topic modeling. The genre of topic-modeling literature tends to get very quickly into the generative method that underlies LDA (the most widely used topic modeling algorithm). It stops just shy of what a Direchlet distribution is, but puts probability front and center.

The basic transformation of topic modeling, though, is not probabilistic: its about breaking up a series of documents into coherent topics that make them up based on co-occurrence. This framing is much larger than LDA, and need have nothing to do with probabilities; I would (maybe controversially) say that for non-insiders, we should think of matrix operations like latent semantic analysis as topic models as well, just of a different stripe. Its fine if stage 2, or stage 3, of understanding topic models is about the probabilities that drive all effective ones. But for stage 1 we should think about the general goal and simply give the basic advice for who has implement it best. (In this case, Use mallet with hyperparameters.)

Thinking about transformations means spending less time worrying about whether boxes are black or not, and more time thinking about how to create inputs and read outputs. Creating inputs is particularly important for something like topic modeling. One piece of advice that is given much too little in introductions to topic modeling, I think, is: the most important choice is how big your individual documents will be. Should you use paragraphs? Book chapters? Whole books? Unlike setting hyperparameters for LDA, thats immediately addressable by any humanist; and its possible to understand why it matters without knowing anything about the particulars of a topic modeling algorithm.

This has also heavily influenced my thinking about word embedding models. In my 1-page introduction to word embeddings, I say use word2vec and largely leave it at that. If you want to use LSA, though, or Michael Gavins raw vector method based on hand-selected words, thats great. As with topic models or sorting, the things you do with a word embedding model are largely independent of the algorithm that sorted it.

Thats all here. Heres that link to the article again.