We were delighted to discover that the CCP's study of the Supreme Court's decision in Scott v. Harris made it into New York Times Sunday Magazines Ninth Annual Year in Ideas (standard for selection: "the most clever, important, silly and just plain weird innovations..."). It was especially fitting to share that honor with the Ruppy, the glow-in-the-dark dog, public fears of whom are being investigated in CCP's synthetic biology risk perception project.
The New York Times is reporting on a Supreme Court case about a cross erected in the Mojave National Preserve in 1943. While most of the Court seemed focused on whether the attempt to transfer the land to a private party (and thus avoid establishment issues) was proper, Justice Scalia went right for the establishment question:
The question of the meaning of a cross in the context of a war memorial did give rise to one heated exchange, between Justice Scalia and Peter J. Eliasberg, a lawyer for Mr. Buono with the American Civil Liberties Union Foundation of Southern California.
Mr. Eliasberg said many Jewish war veterans would not wish to be honored by “the predominant symbol of Christianity,” one that “signifies that Jesus is the son of God and died to redeem mankind for our sins.”
Justice Scalia disagreed, saying, “The cross is the most common symbol of the resting place of the dead.”
“What would you have them erect?” Justice Scalia asked. “Some conglomerate of a cross, a Star of David and, you know, a Muslim half moon and star?”
Mr. Eliasberg said he had visited Jewish cemeteries. “There is never a cross on the tombstone of a Jew,” he said, to laughter in the courtroom.
Justice Scalia grew visibly angry. “I don’t think you can leap from that to the conclusion that the only war dead that that cross honors are the Christian war dead,” he said. “I think that’s an outrageous conclusion.”
Stephen Burbank (in an email) points out that this has all the markers of cognitive illiberalism as described by our article in the Harvard Law Review on the Supreme Court's decision in Scott v. Harris:
Because they are not generally aware of their own disposition to form factual beliefs that cohere with their cultural commitments [judges] manifest little uncertainty about their answers to [policy questions turning on issues of disputed fact]. But much worse, because they can see full well the influence that cultural predispositions have on those who disagree with them, participants in policy debates often adopt a dismissive and even contemptuous posture towards their opponents' beliefs....
It may be cognitively difficult for someone with the cultural commitments of Justice Scalia to understand the cross as anything other than a universal symbol profound respect, and to struggle with evidence to the contrary. But struggling with cultural blindspots is something we expect judges to do, particularly in cases involving questions about the establishment clause.
The Chicago Sun-Times and just about every other news source in the country is reporting the Supreme Court decision to hear a challenge to the city of Chicago's ordinance barring handgun ownership (McDonald v. Chicago, No. 08-1521). The debate over the ordinance and the case is ostensibly one about rights, but those rights are, as the majority opinion in Heller indicated, to be balanced with concerns about public safety. Just what public safety requires, though, is precisely what cultural cognition predicts people will disagree over. And, sure enough, as the headline in the Chicago Sun Times (surely intended to generate outrage and rejoicing in different communities) states: "Gun advocates predict drop in crime if gun ban is lifted." McDonald, Heller, and their progeny may strike a compromise that appeals to a broad spectrum of the American public, or they may inflame cultural passions further. Only time will tell. But in the meantime, you can read up on the debate and the role cultural cognition plays in it here:
- Overcoming the Fear of Cultural Politics: Constructing a Better Gun Debate
- More Statistics, Less Persuasion: A Cultural Theory of Gun-Risk Perceptions
- Beyond the Gun Fight: The Aftermath of the Virginia Tech Massacre
- Modeling Facts, Culture and Cognition in the Gun Debate
- Gun Litigation: A Cultural Critique
It’s not clear that the case will ever make it to trial, but if it does, what sort of person would make the best juror for Pittsburgh Steelers quarterback “Big Ben” Roethlisberger in his defense to the civil sexual assault case filed against him? The answer might come as a surprise -- maybe not to Roethlisberger’s lawyers, but probably to many commentators involved in the debate over law and date rape. A study founded on the theory of cultural cognition suggests that Big Ben would likely be judged much more sympathetically by a jury dominated by women who subscribe to highly traditional gender norms than he would by one consisting literally of his “peers.”
Steve Easterbrook has a thoughtful post on his blog about cultural cognition and climate change. In the comments on his post, one of his readers describe a common problem that lay citizens often ascribe to experts: sometimes experts "lie for hire." It's true that members of the public may worry about that, but it's also true that individuals selectively attend to this particular risk when evaluating expert opinion. Dan is in the middle of a study in which he demonstrates precisely this phenomenon, and I'm sure he'll be posting about it here soon!
Dan's post suggests to me that cultural cognition might influence how reassuring individuals find a recent press release by Robotic Technology, Inc., setting the record straight. While the company's Energetically Autonomous Tactical Robot ("EATR" for short), "can find, ingest, and extract energy from biomass in the environment (and other organically-based energy sources)," RTI wants you to know that, "[d]espite the far-reaching reports that this includes “human bodies,” the public can be assured that the engine Cyclone (Cyclone Power Technologies Inc.) has developed to power the EATR runs on fuel no scarier than twigs, grass clippings and wood chips -- small, plant-based items for which RTI’s robotic technology is designed to forage. Desecration of the dead is a war crime under Article 15 of the Geneva Conventions, and is certainly not something sanctioned by DARPA, Cyclone or RTI."
(Thanks to Sarah Lawsky for the link.)
Story today in NY Times on growing concern about the risks posed by artificial intelligence and in particular the possibility that artificially intelligent systems (including ones designed to kill people) will become autonomous. Interesting to consider how this one might play out in cultural terms. Individualism should incline people toward low risk perception, of course. But hierarchy & egalitarianism could go either way, depending on the meanings that AI becomes invested with: if applications are primarily commercial and defense-related and the technology gets lumped in w/ nanotechnology, nuclear, etc., then egalitarians will likely be fearful, and hierarchs not; if AI starts to look like "creation of life" -- akin to synbio -- then expect hierarchs to resist, particularly ones who are highly religious in nature. Wisely, AI stakeholders -- like nanotech & synbio ones-- recognize that the time is *now* to sort out what the likely risk perceptions will be so that they can be managed and steered in way that doesn't distort informed public deliberation:
The meeting on artificial intelligence could be pivotal to the future of the field. Paul Berg, who was the organizer of the 1975 Asilomar meeting and received a Nobel Prize for chemistry in 1980, said it was important for scientific communities to engage the public before alarm and opposition becomes unshakable.
"If you wait too long and the sides become entrenched like with G.M.O.," he said, referring to genetically modified foods, 'then it is very difficult. It’s too complex, and people talk right past each other."
This is a topic ripe for investigation by cultural theorists of risk.
Quite often we'll be developing simulations based on models in which the DV is a 6-point Likert-style response scale. (Usually it's somethign like: strongly disagree / disagree / mildly disagree / mildly agree / agree / strongly agree.) For presentation purposes, it's often useful to reduce this down into two categories: any form of agreement / any form of disagreement. In particular, when graphing, it is much easier to show one cut with confidence intervals than to show five cuts with confidence intervals.
In the past we've done this by converting the DV into a binary variable, then running a logistic regression. But this has numerous drawbacks. First and foremost, it simply throws away all the information about how strongly a person agrees or disagrees. As a result, errors tend to be larger than necessary. Second, and relatedly, the results often aren't as similar to the ologit regressions run against the more information-rich likert DV as one would like. And third, if we want to report both kinds of findings -- binary and likert-style -- this means reporting two separate models that don't always give the same results. In short, it's been a mess, and we've usually just chosen one or the other. But when we've gone with a logit regression, this seems like sad choice to make just to achieve greater simplicity of presentation.
Recently, though, I had coffee with Jeff Lax-- of state-level policy analysis & Gelman Blog fame -- and he suggested something that, in retrospect, reveals that I'm still often trapped in a non-simulation mindset. In essence, he suggested this: "Run simulations on your ologit model & combine the simulations for the agree levels and again for the disagree levels; then take your confidence intervals from those combined simulations." In retrospect, that is so clearly the correct approach that the question is why I didn't see it myself. The answer, I think, is that I was still thinking in terms of the regression model rather than the simulations.
In a recent NY Times Article, Charles Siebert writes about the recent case brought by the National Resources Defense Council, arguing that Naval use of sonar in certain exercises is leading whales to flee to the surface too quickly, suffer the bends, and eventually die:
The question of sonar’s catastrophic effects on whales even reached the Supreme Court last November, in a case pitting the United States Navy against the Natural Resources Defense Council. The council, along with other environmental groups, had secured two landmark victories in the district and appellate courts of California, which ruled to heavily restrict the Navy’s use of sonar devices in its training exercises. The Supreme Court, however, in a 6-to-3 decision widely viewed as a setback for the environmental movement, overturned parts of the lower-court rulings, faulting them for, in the words of Chief Justice John Roberts’s majority opinion, failing “properly to defer to senior Navy officers’ specific, predictive judgments,” thereby jeopardizing the safety of the fleet and sacrificing the public’s interest in military preparedness by “forcing the Navy to deploy an inadequately trained antisubmarine force.” In his decision, Roberts went on to minimize, in a fairly dismissive tone, the issue of harm to marine life: “For the plaintiffs, the most serious possible injury would be harm to an unknown number of the marine animals that they study and observe.”
At the core of the dispute is how serious and plausible the anticipated harm to whales is and how serious and plausible the harm to the Navy would be if it were enjoined from conducting these exercises. While we haven't done any empirical research on this subject in particular, perceptions of risk related to military endeavors tend to be positively correlated with measures of egalitarianism, as are perceptions of environmental risk in general. The combination of environmental risk and military action in this issue makes it a twofer in terms of the cultural cognition of risk.
To kick off our new blog on our new website, what could be better than a little NY Times coverage? Ben Weber has a nice piece about Sonia Sotomayor's nomination. They mention the Harvard Law Review piece covering Scott v. Harris, but there's another piece or two on judicial cognition that you might be interested in.