“The rise and rise of Pierre Bourdieu in US sociology”

A French sociologist looks at the popularity of Pierre Bourdieu:

Pierre Bourdieu would have turned 85 on 1 August 2015. Thirteen years after his death, the French sociologist remains one of the leading social scientists in the world. His work has been translated into dozens of languages (Sapiro & Bustamante 2009), and he is one of the most cited social theorists worldwide, ahead of major thinkers like Jurgen Habermas, Anthony Giddens, or Irving Goffman (Santoro 2008). That Bourdieu is one of the most prominent social theorists will come as no surprise to those accustomed to the academic scene. A more surprising fact, however, is that he is probably the most cited scholar in the social sciences. In a forthcoming paper on the reception of French sociologists in the United States, Andrew Abbott and I show that, at the turn of this decade, he is referenced in more than 100 sociological articles a year. Important authors like Paul Di Maggio or James Coleman are only cited 60 times, while Mark Granovetter has nearly 50 mentions. Bourdieu is also referenced more often than Émile Durkheim, who for a long time epitomized (French) sociology…

This diversity of topics influenced the reception of Bourdieu’s work abroad. As has been pointed out (see Sallasz and Zavisca, 2008), it was initially read by different (unrelated) groups. Though it happened fairly early, the reception of his work remained confined to local areas for over two decades. In the United States, this situation changed in the late 1980s following a number of efforts to emphasize the systematic character of Bourdieu’s research. The key initiative among these was the 1992 interview book co-authored by Bourdieu and Loïc Wacquant, Invitation to a Reflexive Sociology. Written in English with a US audience in mind, it aims at presenting Bourdieu’s system to a foreign audience. Our data shows that after publication of this book, his work subsequently gained widespread exposure beyond the limited local fields in which it was already popular. Not only were his concepts now used outside of those fields, but references to his work also increasingly pertained to theoretical aspects rather to empirical ones. Starting in the mid-1990s, Bourdieu was regarded as a general social theorist and read across sub-disciplinary lines—as well as across disciplines.

What will happen next? Although prediction and social science don’t square well, several signs indicate that Bourdieu is currently entering the canon of worldwide sociology. In the United States, our study shows that while the number of references to his work continues to increase, scholars’ level of engagement with the text is decreasing. In fact, over the last few years, references to Bourdieu have become more allusive. To measure this change, we hand-coded several hundred references from different periods. The proportion of those extensively citing Bourdieu has decreased steadily since the 2000s. This trait is characteristic of a process of canonization, when an author becomes equated with an idea or a set of ideas (e.g. Foucault and power, Goffman and face-to-face interactions, etc.), and is therefore considered a mandatory reference on the topic. The citation becomes a ritual. In some cases, the author has obviously not read the text in question.

Has Bourdieu become a museum piece? It does not seem so, at least for now. Scholarly interest is still strong and his work is still very much discussed. A good indicator of this is the number of references to an author per article, and comparison with other authors is telling here. Whereas Durkheim is routinely cited but not much debated, and receives an average of one reference per article citing him (fig2a), Bourdieu’s work is still an object of active investment (fig2b). At least 25% of the articles citing Bourdieu make two references to his work, sometimes many more. Bourdieu may well be entering the canon, but his appropriation abroad still fosters debates.

This sort of analysis could be undertaken with any major figure in an academic field: how did their work spread, who was spreading it, when did it peak, and how did the citations coalesce around particular topics or ideas?

The case of Bourdieu is interesting for several reasons. It involves a sociologist from another country who wrote in another language and American sociology can sometimes be provincial. Bourdieu’s work began in the ethnographic realm but hit upon key areas in sociology after the 1960s including social class, culture, and education. His findings have been utilized in multiple disciplines and across countries.

At some point, will there be a Bourdieu backlash or major opposition? The article suggested the next stage is “museum piece” and this seems to imply that his work will remain important but fade into history. Who has the big ideas to replace Bourdieu or significantly tweak his work?

Using algorithms to analyze the literary canon

A new book describes efforts to use algorithms to discover what is in and out of the literary canon:

There’s no single term that captures the range of new, large-scale work currently underway in the literary academy, and that’s probably as it should be. More than a decade ago, the Stanford scholar of world literature Franco Moretti dubbed his quantitative approach to capturing the features and trends of global literary production “distant reading,” a practice that paid particular attention to counting books themselves and owed much to bibliographic and book historical methods. In earlier decades, so-called “humanities computing” joined practitioners of stylometry and authorship attribution, who attempted to quantify the low-level differences between individual texts and writers. More recently, the catchall term “digital humanities” has been used to describe everything from online publishing and new media theory to statistical genre discrimination. In each of these cases, however, the shared recognition — like the impulse behind the earlier turn to cultural theory, albeit with a distinctly quantitative emphasis — has been that there are big gains to be had from looking at literature first as an interlinked, expressive system rather than as something that individual books do well, badly, or typically. At the same time, the gains themselves have as yet been thin on the ground, as much suggestions of future progress as transformative results in their own right. Skeptics could be forgiven for wondering how long the data-driven revolution can remain just around the corner.

Into this uncertain scene comes an important new volume by Matthew Jockers, offering yet another headword (“macroanalysis,” by analogy to macroeconomics) and a range of quantitative studies of 19th-century fiction. Jockers is one of the senior figures in the field, a scholar who has been developing novel ways of digesting large bodies of text for nearly two decades. Despite Jockers’s stature, Macroanalysis is his first book, one that aims to summarize and unify much of his previous research. As such, it covers a lot of ground with varying degrees of technical sophistication. There are chapters devoted to methods as simple as counting the annual number of books published by Irish-American authors and as complex as computational network analysis of literary influence. Aware of this range, Jockers is at pains to draw his material together under the dual headings of literary history and critical method, which is to say that the book aims both to advance a specific argument about the contours of 19th-century literature and to provide a brief in favor of the computational methods that it uses to support such an argument. For some readers, the second half of that pairing — a detailed look into what can be done today with new techniques — will be enough. For others, the book’s success will likely depend on how far they’re persuaded that the literary argument is an important one that can’t be had in the absence of computation…

More practically interesting and ambitious are Jockers’s studies of themes and influence in a larger set of novels from the same period (3,346 of them, to be exact, or about five to 10 percent of those published during the 19th century). These are the only chapters of the book that focus on what we usually understand by the intellectual content of the texts in question, seeking to identify and trace the literary use of meaningful clusters of subject-oriented terms across the corpus. The computational method involved is one known as topic modeling, a statistical approach to identifying such clusters (the topics) in the absence of outside input or training data. What’s exciting about topic modeling is that it can be run quickly over huge swaths of text about which we initially know very little. So instead of developing a hunch about the thematic importance of urban poverty or domestic space or Native Americans in 19th-century fiction and then looking for words that might be associated with those themes — that is, instead of searching Google Books more or less at random on the basis of limited and biased close reading — topic models tell us what groups of words tend to co-occur in statistically improbable ways. These computationally derived word lists are for the most part surprisingly coherent and highly interpretable. Specifically in Jockers’s case, they’re both predictable enough to inspire confidence in the method (there are topics “about” poverty, domesticity, Native Americans, Ireland, sea faring, servants, farming, etc.) and unexpected enough to be worth examining in detail…

The notoriously difficult problem of literary influence finally unites many of the methods in Macroanalysis. The book’s last substantive chapter presents an approach to finding the most central texts among the 3,346 included in the study. To assess the relative influence of any book, Jockers first combines the frequency measures of the roughly 100 most common words used previously for stylistic analysis with the more than 450 topic frequencies used to assess thematic interest. This process generates a broad measure of each book’s position in a very high-dimensional space, allowing him to calculate the “distance” between every pair of books in the corpus. Pairs that are separated by smaller distances are more similar to each other, assuming we’re okay with a definition of similarity that says two books are alike when they use high-frequency words at the same rates and when they consist of equivalent proportions of topic-modeled terms. The most influential books are then the ones — roughly speaking and skipping some mathematical details — that show the shortest average distance to the other texts in the collection. It’s a nifty approach that produces a fascinatingly opaque result: Tristram Shandy, Laurence Sterne’s famously odd 18th-century bildungsroman, is judged to be the most influential member of the collection, followed by George Gissing’s unremarkable The Whirlpool (1897) and Benjamin Disraeli’s decidedly minor romance Venetia (1837). If you can make sense of this result, you’re ahead of Jockers himself, who more or less throws up his hands and ends both the chapter and the analytical portion of the book a paragraph later. It might help if we knew what else of Gissing’s or Disraeli’s was included in the corpus, but that information is provided in neither Macroanalysis nor its online addenda.

Sounds interesting. I wonder if there isn’t a great spot for mixed method analysis: Jockers’ analysis provides the big picture but you also need more intimate and deep knowledge of the smaller groups of texts or individual texts to interpret what the results mean. So, if the data suggests three books are the most influential, you would have to know these books and their context to make sense of what the data says. Additionally, you still want to utilize theories and hypotheses to guide the analysis rather than simply looking for patterns.

This reminds me of the work sociologist Wendy Griswold has done in analyzing whether American novels shared common traits (she argues copyright law was quite influential) or how a reading culture might emerge in a developing nation. Her approach is somewhere between the interpretation of texts and the algorithms described above, relying on more traditional methods in sociology like analyzing samples and conducting interviews.