A growing interest in science cafes in America?

Reuters reports on a supposedly growing trend: science cafes.

Science cafes have sprouted in almost every state including a tapas restaurant near downtown Orlando where Sean Walsh, 27, a graphic designer, describes himself and his friends as some of the laymen in the crowd…

But the typical participant brings at least some college-level education or at least a lively curiosity, said Edward Haddad, executive director of the Florida Academy of Sciences, which helped start up Orlando’s original cafe and organizes the events…

Haddad said the current national push to increase the number of U.S. graduates in science, technology, engineering and math, or the STEM fields, is driving up the number of science cafes…

The U.S. science cafe movement grew out of Cafe Scientifique in the United Kingdom. The first Cafe Scientifique popped up in Leeds in 1998 as a regularly scheduled event where all interested parties could participate in informal forums about the latest in science and technology.

I’m dubious that this is that big of a movement just because “almost every state” now has a science cafe. This is similar to journalists claiming that something is popular because there is a Facebook group devoted to it.

But, this sounds like a fascinating example of a “third place” where Americans can gather between home and work, learn, and interact with others interested in similar topics. In fact, it sounds more like a Parisian salon of the 1800s. However, the article also mentions these cafes are probably more attractive to the NPR crowd and I imagine many Americans would not want to go discuss science in a cafe.

I wonder if the news coverage would be different if Americans were gathering in cafes to talk about other topics. How about The Bachelor? The tea party? Religion? The tone of the article is that it is more unusual for Americans to want to hear about and discuss science when they are not being forced to.

h/t Instapundit

Mapping NFL fandom by county with Facebook likes

Facebook has put their massive data trove to use and examined the geographies of NFL fandom. Here is what they came up with:

The National Football League is one of the most popular sports in America with some incredibly devoted fans. At Facebook we have about 35 million account holders in the United States who have Liked a page for one of the 32 teams in the league, representing one of the most comprehensive samples of sports fanship ever collected. Put another way, more than 1 in 10 Americans have declared their support for an NFL team on Facebook…

While winning seems to matter, NFL teams have local followings that are probably heavily influenced by family ties and/or where a person grew up,  so we were obviously curious to see where the fans for various teams live now. By considering the physical locations of NFL fans, we can construct a map of the top team for each county in the US. It tells an interesting story about the ways that football rivalries and allegiances alternately divide and unite the country, and sometimes even individual states.

In some cases, whole states and even entire regions of the country uniformly support a single team.  For instance the Vikings are easily the only game in town in Minnesota, while New England appears to be comprised of entirely Patriots fans except for a small portion of Connecticut.

There are some states which are divided into regions by teams.  Florida has three teams–the Tampa Bay Bucs, Miami Dolphins, and the Jacksonville Jaguars–and Facebook users there seems fractured in their support, with some counties even defecting to teams from the North. Ohio is another interesting story, with the Cleveland Browns in the North, Cincinatti Bengals in the South, and Pittsburgh Steelers fans occupying the middle of the state.

Some teams, like the Steelers, Cowboys, and Packers, seem to transcend geography, with pockets of fans all over the country. On the other end of the spectrum, the Jets have to share New York with the Giants and are only the most popular team for a single stronghold county in Long Island.

Five quick thoughts:

1. There are few other organizations that could put together such a map without undertaking a major survey (since this is measured at the county level).

2. The best part for Facebook: users voluntarily provided this data.

3. Could Facebook end up being the most important future source for telling us about American society? There are still difficulties: users have to opt in (in this particular case, they had to “like” a NFL team), not everyone is involved (though it seems like pretty close), and not all users are putting everything in their profiles.

4. Is there a way to weight this map with population density? For example, the Cowboys may have a really broad geographic reach but many of those counties have fewer people. In contrast, teams like the Jets or Eagles have smaller reaches yet more people live in those areas.

5. Is there a way to show the percentage of county respondents who liked the dominant team? I imagine there are plenty of counties where one team does not have a strong majority, let alone even much of a plurality. For example, Jets fans barely show up on the map because they are only the top team in one county. Yet, there are plenty of Jets fans.

Bill Gates: we can make progress with goals, data, and a feedback loop

Bill Gates argues in the Wall Street Journal that significant progress can be made around the world if organizations and residents participate in a particular process:

In the past year, I have been struck by how important measurement is to improving the human condition. You can achieve incredible progress if you set a clear goal and find a measure that will drive progress toward that goal—in a feedback loop similar to the one Mr. Rosen describes.

This may seem basic, but it is amazing how often it is not done and how hard it is to get right. Historically, foreign aid has been measured in terms of the total amount of money invested—and during the Cold War, by whether a country stayed on our side—but not by how well it performed in actually helping people. Closer to home, despite innovation in measuring teacher performance world-wide, more than 90% of educators in the U.S. still get zero feedback on how to improve.

An innovation—whether it’s a new vaccine or an improved seed—can’t have an impact unless it reaches the people who will benefit from it. We need innovations in measurement to find new, effective ways to deliver those tools and services to the clinics, family farms and classrooms that need them.

I’ve found many examples of how measurement is making a difference over the past year—from a school in Colorado to a health post in rural Ethiopia. Our foundation is supporting these efforts. But we and others need to do more. As budgets tighten for governments and foundations world-wide, we all need to take the lesson of the steam engine to heart and adapt it to solving the world’s biggest problems.

Gates doesn’t use this term but this sounds like a practical application of the scientific method. Instead of responding to a social problem by going out and trying to “do something,” the process should be more rigorous, involve setting goals, collecting good data, interpreting the data, and then adjusting the process from the beginning. This is related to other points about this process:

1. It is one thing to be able to collect data (and this is often its own complicated process) but it is another to know what to do with it once you have it. Compared to the past, data is relatively easy to obtain today but using it well is another matter.

2. Another broad issue in this kind of feedback loop is developing the measurements and what counts as “success.” Some of this is fairly easy; when Gates praises the UN Millennium Goals, reducing occurrences of disease or boosting incomes has face validity for getting at what matters. But, measuring teacher’s performances or what makes a quality college are a little trickier to define in the first place. Gates calls this developing goals but this could be a lengthy process in itself.

It is interesting that Gates mentions the need for such loops in colleges so that students “could know where they would get the most for their tuition money.” The Gates Foundation has put money into studying public schools and just a few weeks ago released some of their findings:

After a three-year, $45 million research project, the Bill and Melinda Gates Foundation believes it has some answers.

The most reliable way to evaluate teachers is to use a three-pronged approach built on student test scores, classroom observations by multiple reviewers and teacher evaluations from students themselves, the foundation found…

The findings released Tuesday involved an analysis of about 3,000 teachers and their students in Charlotte; Dallas; Denver; Memphis; New York; Pittsburgh; and Hillsborough County, Fla., which includes Tampa. Researchers were drawn from the Educational Testing Service and several universities, including Harvard, Stanford and the University of Virginia…

Researchers videotaped 3,000 participating teachers and experts analyzed their classroom performance. They also ranked the teachers using a statistical model known as value-added modeling, which calculates how much an educator has helped students learn based on their academic performance over time. And finally, the researchers surveyed the students, who turned out to be reliable judges of their teacher’s abilities, Kane said.

All this takes quite a few resources and time. For those interested in quick action, this is not the process to follow. Hopefully, however, the resources and time pay off with better solutions.

The sociology of literature and looking for data and insights in the margins of books

As a big reader, I was interested to see this review of research built on data about readers left behind in books:

Price’s work perches at the leading edge of a growing body of investigations into the history of reading. The field draws from many others, including book history and bibliography, literary criticism and social history, and communication studies. It looks backward to the pre-Gutenberg era, back to the clay tablets and scrolls of ancient civilizations, and forward to current debates about how technology is changing the way we read. Although much of the relevant research has centered on Anglo-American culture of the last three or four centuries, the field has expanded its purview, as scholars uncover the hidden reading histories of cultures many used to dismiss as mostly oral.

It’s a tricky business. A bibliographer works with hard physical evidence—a manuscript, a printed book, a copy of the Times of London. A scholar seeking to pin down the readers of the past often has to read between the lines. Marginalia can be a gold mine of information about a book’s owners and readers, but it’s rare. “Most of the time, most readers historically didn’t, and still don’t, write in their books,” Price explains.

But even a book’s apparent lack of use can be read as evidence. “The John F. Kennedy Library here in Boston owns a copy of Ulysses whose pages—other than a few at the very beginning and very end—are completely uncut,” she says. “This tells us something about the owner of the copy—who happens to be Ernest Hemingway.”…

Since Reading the Romance, the ethnography of reading has taken off among scholars. Radway points to Forgotten Readers, Elizabeth McHenry’s study of African-American literary societies, Ellen Gruber Garvey’s Writing With Scissors, about scrapbooking, and David Henkin’s City Reading, about signage in the urban environment, as strong examples. “People have become very creative about trying to figure out how groups of readers interact with the text as it’s embodied in various forms,” she says.

I have wondered in recent years why more sociologists don’t take up the subject of reading. It seems crucial for understanding the development of modern societies as information moved from a highly regulated environment to a diffuse distribution through books, newspapers, and other printed materials.

I’ve enjoyed the work of sociologist Wendy Griswold who studies reading. I’ve used a few of her pieces in class. Here are some of her fascinating works in the “sociology of literature” that I recommend:

1. Bearing Witness published in 2000. Griswold examines the reading culture in Nigeria and why novels, a common genre in Western society, aren’t prevalent in Nigeria. The short version of the story: it takes a lot of work for a society to be at a level where novels can be easily produced and read.

2. “American Character and the American Novel: An Expansion of Reflection Theory in the Sociology of Literature.” American Journal of Sociology 86(4), 1981. Griswold compares American and European novels in the late 1800s and early 1900s and finds the differences in their content is due more to copyright law than “national characters.”

3. With Terry McDonnell and Nathan Wright. “Reading and the Reading Class in the Twenty-First Century.” Annual Review of Sociology 31, 2005. Here is the abstract:

Sociological research on reading, which formerly focused on literacy, now conceptualizes reading as a social practice. This review examines the current state of knowledge on (a) who reads, i.e., the demographic characteristics of readers; (b) how they read, i.e., reading as a form of social practice; (c) how reading relates to electronic media, especially television and the Internet; and (d) the future of reading. We conclude that a reading class is emerging, restricted in size but disproportionate in influence, and that the Internet is facilitating this development.

Some fascinating stuff about the social forces influencing reading in today’s world.

4. With Nathan Wright. “Wired and Well Read.” In Society Online: The Internet in Context, 2004. If I remember correctly, Griswold and Wright argue the Internet doesn’t compete with reading; rather it enhances reading as those who read before the Internet use the Internet to read more.

Nielsen and Twitter combine to measure Twittering about television

With the rise of Twitter messages about television shows and events, Nielsen and Twitter just announced a new project to measure the connection:

“The Nielsen Twitter TV Rating is a significant step forward for the industry, particularly as programmers develop increasingly captivating live TV and new second-screen experiences, and advertisers create integrated ad campaigns that combine paid and earned media,” said Steve Hasker, President, Global Media Products and Advertiser Solutions at Nielsen. “As a media measurement leader we recognize that Twitter is the preeminent source of real-time television engagement data.”…

The Nielsen Twitter TV Rating will enhance the social TV analytics and metrics available today from SocialGuide by adding the first-ever measurement of the total audience for social TV activity – both those participating in the conversation and those who were exposed to the activity –providing the precise size of the audience and effect of social TV to TV programming.

SocialGuide, recently acquired by Nielsen and NM Incite, currently captures Twitter TV activity for all U.S. programming across 234 TV channels in English and Spanish, and more than 36,000 programs.  Through a sophisticated classification process, SocialGuide matches Tweets to TV programs to offer key social TV metrics including the number of unique Tweets associated with a given program and rankings for the most social TV programs.

This may be interesting in itself but the key may just be translating this into information that TV networks can sell to advertisers:

Brad Adgate, an analyst at Horizon Media, said advertisers will view the Twitter ratings as a useful layer of information about a show’s popularity, but it is “not going to be close to the currency” of existing ratings metrics.

“It lets producers and creative directors know if the storyline is working, like a huge focus group,” Adgate said. “But I don’t think you can translate comments to ratings for a show. Right now I think the bark right now is bigger than its bite.”…

Mark Burnett, executive producer of NBC’s hit “The Voice,” argued that advertisers should value programs that can attract a high level of social media engagement from viewers. Deeply embedded social media elements, such as live Twitter polls, were critical in driving “The Voice” to the top of the Tuesday night ratings among viewers between 18 to 49, Burnett said.

“If you’re an advertiser, wouldn’t you want to know whether people are watching this show passively or if they’re actively engaged in the viewing experience?” Burnett said. “Five years from now this will make traditional television ratings seem archaic.”

In other words, if this metric works well, television networks will be able to charge advertisers more based on increased levels of Twitter engagement or find some way to provide more targeted advertising to Twitter users. What will Twitter engaged TV watchers get out of it? I’m not sure. Will any of this measurement and action based on the data enhance the interactive element of TV watching? Theoretically, if TV networks could get more money for advertising based on social media engagement, they might have more money to put into developing quality programming. But, there are few guarantees there.

I’ll be very interested to see in coming years if Twitter and Facebook continue to remain relatively ad-free or if the need to monetize these experiences to make money takes precedence.

Social psychologist on quest to find researchers who falsify data

The latest Atlantic magazine includes a short piece about a social psychologist who is out to catch other researchers who falsify data. Here is part of the story:

Simonsohn initially targeted not flagrant dishonesty, but loose methodology. In a paper called “False-Positive Psychology,” published in the prestigious journal Psychological Science, he and two colleagues—Leif Nelson, a professor at the University of California at Berkeley, and Wharton’s Joseph Simmons—showed that psychologists could all but guarantee an interesting research finding if they were creative enough with their statistics and procedures.

The three social psychologists set up a test experiment, then played by current academic methodologies and widely permissible statistical rules. By going on what amounted to a fishing expedition (that is, by recording many, many variables but reporting only the results that came out to their liking); by failing to establish in advance the number of human subjects in an experiment; and by analyzing the data as they went, so they could end the experiment when the results suited them, they produced a howler of a result, a truly absurd finding. They then ran a series of computer simulations using other experimental data to show that these methods could increase the odds of a false-positive result—a statistical fluke, basically—to nearly two-thirds.

Just as Simonsohn was thinking about how to follow up on the paper, he came across an article that seemed too good to be true. In it, Lawrence Sanna, a professor who’d recently moved from the University of North Carolina to the University of Michigan, claimed to have found that people with a physically high vantage point—a concert stage instead of an orchestra pit—feel and act more “pro-socially.” (He measured sociability partly by, of all things, someone’s willingness to force fellow research subjects to consume painfully spicy hot sauce.) The size of the effect Sanna reported was “out-of-this-world strong, gravity strong—just super-strong,” Simonsohn told me over Chinese food (heavy on the hot sauce) at a restaurant around the corner from his office. As he read the paper, something else struck him, too: the data didn’t seem to vary as widely as you’d expect real-world results to. Imagine a study that calculated male height: if the average man were 5-foot?10, you wouldn’t expect that in every group of male subjects, the average man would always be precisely 5-foot-10. Yet this was exactly the sort of unlikely pattern Simonsohn detected in Sanna’s data…

Simonsohn stressed that there’s a world of difference between data techniques that generate false positives, and fraud, but he said some academic psychologists have, until recently, been dangerously indifferent to both. Outright fraud is probably rare. Data manipulation is undoubtedly more common—and surely extends to other subjects dependent on statistical study, including biomedicine. Worse, sloppy statistics are “like steroids in baseball”: Throughout the affected fields, researchers who are too intellectually honest to use these tricks will publish less, and may perish. Meanwhile, the less fastidious flourish.

The current research may just provide incentives for researchers to cut corners and end up with false results. Publishing is incredibly important for the career of an academic and there is little systematic oversight of a researcher’s data. I’ve written before about ways that data could be made more open but it would take some work to put these ideas into practice.

What I wouldn’t want to happen is have people read a story like this and conclude that fields like social psychology have nothing to offer because who knows how many of the studies might be flawed. I also wonder about the vigilante edge to this story – it makes a journalistic piece to tell about a social psychologist who is battling his own field but this isn’t how science should work. Simonsohn should be joined by others who should also be concerned by these potential issues. Of course, there may not be many incentives to pursue this work as it might invite criticism from inside and outside the discipline.

NCAA Scholarly Colloquium: ideology versus “In God we trust; everyone else should bring data”

The Chronicle of Higher Education examines how much criticism of the NCAA will be allowed at its upcoming annual Scholarly Colloquium and includes a fascinating quote about how data should be used:

The colloquium was the brainchild of Myles Brand, a former NCAA president and philosopher who saw a need for more serious research on college sports. He and others believed that such an event could foster more open dialogue between the scholars who study sport issues and the people who work in the game.

Mr. Brand emphasized that the colloquium should be data-based and should avoid ideology. “Myles always used to joke: ‘In God we trust; everyone else should bring data,'” said Mr. Renfro, a former top adviser to Mr. Brand.

But as Mr. Renfro watched presentations at last year’s colloquium, which focused on changes the NCAA has made in its academic policies in recent years, he did not see a variety of perspectives.

“I was hearing virtually one voice being sung by a number of people … and it was relatively critical of the NCAA’s academic-reform effort,” he said. “I don’t care whether it was critical or not, but I care about whether there are different perspectives presented.”

This is a classic argument: data versus ideology, facts versus opinions. This short bit about Myles Brand makes it sound like Brand thought bringing more data to the table when discussing the NCAA would be a good thing. Data might blunt opinions and arguments and push people with an agenda to back up their arguments. It could lead to more constructive conversations. But, data is not completely divorced from ideology. Researchers choose what kind of topics to study. Data has to be collected in a good manner. Interpreting data is still an important skill; people can use data incorrectly. And it sounds like an issue here is that people might be able to use data to continue to criticize the NCAA – and this does not make the NCAA happy.

Generally, I’m in favor of bringing more data to the table when discussing issues. However, having data doesn’t necessarily solve problems. As I tell my statistics classes, I don’t want them to be people who blindly believe all data or statistics because it is data and I also don’t want them to be people who dismiss all data or statistics because they can be misused and twisted. It sounds like some of this still needs to be sorted out with the NCAA Scholarly Colloquium.

Census data visualization: metropolitan population change by natural increase, international migration, and domestic migration

The Census regularly puts together new data visualizations to highlight newly collected data. The most recent visualization looks at population change in metropolitan areas between 2010-2011 and breaks down the change by natural increase, international migration, and domestic migration.

Several trends are quickly apparent:

1. Sunbelt growth continues at a higher pace and non-Sunbelt cities tend to lose residents by domestic migration.

2. Population increases by international migration still tends to be larger in New York, Los Angeles, and Miami.

3. There are some differences in natural increases to population. I assume this is basically a measure of birth rates.

However, I have two issues with this visualization. My biggest complaint is that the boxes are not weighted by population. New York has the largest natural increase to the population but it is also the largest metropolitan areas by quite a bit. A second issue is that the box sizes are not all the 50,000 or 10,000 population change as suggested by the key at the top. So while I can see relative population change, it is hard to know the exact figures.

Using GIS to study Gettysburg, the Holocaust, and the American iron industry

Smithsonian takes a look at a historian who uses GIS to get a new perspective on important historical events:

Her principal tool is geographic information systems, or GIS, a name for computer programs that incorporate such data as satellite imagery, paper maps and statistics. Knowles makes GIS sound simple: “It’s a computer software that allows you to map and analyze any information that has a location attached.” But watching her navigate GIS and other applications, it quickly becomes obvious that this isn’t your father’s geography…

What emerges, in the end, is a “map” that’s not just color-coded and crammed with data, but dynamic rather than static—a layered re-creation that Knowles likens to looking at the past through 3-D glasses. The image shifts, changing with a few keystrokes to answer the questions Knowles asks. In this instance, she wants to know what commanders could see of the battlefield on the second day at Gettysburg. A red dot denotes General Lee’s vantage point from the top of the Lutheran Seminary. His field of vision shows as clear ground, with blind spots shaded in deep indigo. Knowles has even factored in the extra inches of sightline afforded by Lee’s boots. “We can’t account for the haze and smoke of battle in GIS, though in theory you could with gaming software,” she says…

Though she’s now been ensconced at Middlebury for a decade, Knowles continues to push boundaries. Her current project is mapping the Holocaust, in collaboration with the U.S. Holocaust Memorial Museum and a team of international scholars. Previously, most maps of the Holocaust simply located sites such as death camps and ghettos. Knowles and her colleagues have used GIS to create a “geography of oppression,” including maps of the growth of concentration camps and the movement of Nazi death squads that accompanied the German Army into the Soviet Union…

Aware of these pitfalls, Knowles is about to publish a book that uses GIS in the service of an overarching historical narrative. Mastering Iron, due out in January, follows the American iron industry from 1800 to 1868. Though the subject matter may not sound as grabby as the Holocaust or Gettysburg, Knowles has blended geographical analysis with more traditional sources to challenge conventional wisdom about the development of American industry.

Sounds pretty interesting. Having detailed geographic data can change one’s perspective. But there are two things that need to happen first before researchers can take advantage of such information:

1. Using GIS well requires a lot of training and then being able to find the right data for the analysis.

2. Using geographic data like this requires a change in mindset from the idea that geography is just a background variable. In sociology, analysis often controls for some geographic variation but doesn’t often consider the location or space as the primary factor.

While GIS is a hot method right now, I think these two issues will hold it back from being widely used for a while.

Using plagiarism detection software to examine anti-Muslim bias in post-9/11 news coverage

A new sociological study suggests mainstream media sources tended to rely on the rhetoric of certain anti-Muslim groups after 9/11:

“The vast majority of organisations competing to shape public discourse about Islam after the September 11 attacks delivered pro-Muslim messages, yet my study shows that journalists were so captivated by a small group of fringe organisations that they came to be perceived as mainstream,” the paper’s author, University of North Carolina assistant professor of sociology Christopher Bail, told Wired.co.uk…

Bail and his team used plagiarism detection software to compare 1,084 press releases produced by 120 different organisations with more than 50,000 television transcripts and newspaper articles produced between 2001 and 2008. The software picked up damning similarities between the releases and stories from news outlets including the New York Times, USA Today, the Washington Times, CBS News, CNN and Fox News Channel.

“We learned the American media almost completely ignored public condemnations of terrorist events by prominent Muslim organisations in the United States,” Bail told Wired.co.uk. “Inattention to these condemnations, combined with the emotional warnings of anti-fringe organisations, has created a very distorted representation of the community of advocacy organisations, think tanks, and religious groups competing to shape the representation of Islam in the American public sphere.”…

Bail’s paper, published in the American Sociological Review, is part of a wider study which will investigate how the influence of these fringe groups has spread beyond media and in to the real world, where doors have been opened to elite conservative social circles and conservative think tanks — the first steps to influencing public policy and national opinion. Bail touched upon this in the current study after analysing publicly available information on the organisations’ membership, which revealed troubling crossovers between fringe and mainstream organisations.

Four quick thoughts:

1. It sounds like there could be some importance influence of social networks. These fringe groups may be on the edges of public discourse but they have connections or means to which to reach more mainstream media sources. How much of this reporting is built on previous personal connections?

2. This sounds like a clever use of plagiarism software. Such software is intended to catch students in using published material incorrectly but it can also be used to track common quotes, phrases, and narratives.

3. In general, how much does the media today rely on press releases and reports from mainstream or fringe groups without interviews, fact-checking, and sorting through all the information?

4. Would a similar study involving elite liberal social circles and think tanks find similar things?