Holmes again, Holmes again, jiggety jig 
Brian Weatherson links to a recent episode of The Philosopher's Zone, an Australian radio program. They have it on-line both as audio and transcribed.

The host, Alan Saunders, is interviewing Greg Restall. They are discussing the fact that, in classical logic, a contradiction entails everything but, in ordinary reasoning, it does not. Saunders considers an example that I have discussed here recently:
[We] actually allow for the inconsistencies. I mean, let's say it's a Sherlock Holmes story. Now we know that Sherlock Holmes was not the centre of Conan Doyle's interest, he wrote these stories very, very quickly, so if I found an inconsistency like that in a Sherlock Holmes story, I would sort of mentally dissolve it, because I'd think, Oh well, he just wasn't paying attention. If, on the other hand, it were an historical account of a murder and I found [an inconsistency], then I'd say, Well, the evidence isn't all in yet, I'm going to have to suspend judgment on this.
By 'mentally dissolving' the contradiction, we refuse to treat it as evidence about Holmes at all. It provides no constraint on the set of Holmes-worlds. When I discussed inconsistencies in Holmes stories, I suggested something similar.

In reading a historical account, Saunders suggests that we should take opposing claims seriously as evidence-- albeit conflicting evidence-- of how things actually happened. When considering the past, we know that there was some one way that the world was. There is no such assumption about Holmes' world. However, I am not sure that this difference is sufficient to justify Saunders' different handling of these two cases. We can take contradictions in stories as signs that there is some strangeness afoot in them. Why don't we?

[ add comment ] ( 3467 views )   |  [ 0 trackbacks ]   |  permalink
Defer madness 
I wrote most of this entry a couple of weeks ago, after Brian Weatherson pointed to the article in question. Something else came up, so I saved it and moved on. Today I went back, cleaned it up, and posted it.

In a recent paper in Analysis [July 2006, 179-187], Philip Pettit considers the question of whether or not one should acquiesce to the opinion of the majority. He considers three cases, but two are sufficient for the points below.

Case A: Joe is one of many witnesses to an auto accident.* When he saw it, he thought the driver ran a red light. Many other witnesses say that the light was green. Should Joe defer to the majority of reports and conclude that the light was indeed green?

Case B: Joe believes that intelligent design is the best explanation for the existence of order in the universe. The majority of people say otherwise. Should Joe defer to the majority and conclude that intelligent design is hokum?

Pettit suggests that our answer is 'yes' in Case A, but 'no' in Case B. His paper aims to explain and justify the asymmetry.

In a subsequent section, Petit offers a schematic situation. Paraphrasing a bit:
Joe is one of many people who face a given question. Joe and the rest are "equally intelligent, equally informed and equally impartial." Joe disagrees with the answer given by most of the others. Joe knows all of this to be true.
Now, should Joe change his opinion?

Pettit offers the obvious argument for the 'yes' answer: If each person has some independent probability (greater than 1/2) of getting the right answer, then one would be more likely to get the right answer by trusting the majority than the minority. In the limit of large population, probability that the majority will get the right answer approaches one.

This argument gives a 'yes' answer to any instances of the schematic situation. Rather than rebut it, Pettit looks elsewhere for an asymmetry. The difference between Case A and Case B, he suggests, is that the belief that the light was green was closer to the periphery of Joe's web of belief than the belief in intelligent design. Because the latter belief is deeply embedded in his other beliefs, Joe would have to decide if and how to update his other beliefs after deferring to the majority opinion.

Pettit surveys various ways that Joe might try to update his beliefs. If Joe accepts the majority opinion about several questions all at once, then he might end up with inconsistent beliefs. If Joe accepts the majority opinion about one matter, lets that effect his degree of belief, considers the majority opinion regarding a second question, and so on, then the outcome will depend on which question Joe considers first. Since possible inconsistency and path-dependency are to be avoided, Pettit concludes, we should say 'no' in cases where the beliefs are deeply embedded.

Pettit's arguments for inconsistency or path-dependence if Joe defers proceed simply in terms of Joe's beliefs about p, q, and p&q. As such, I suspect that the arguments do not really discriminate between core and peripheral beliefs. Peripheral beliefs can still enter into conjunctions. Admittedly, this suspicion is not an argument; but it does suggest that embeddedness can't explain the asymmetry between Cases A and B.

One real distinction between Case A from Case B is much simpler: Case B is not plausibly an instance of Pettit's general schema, and so the initial argument for a 'yes' answer does not apply. We know that debates about intelligent design do not involve people who are "equally intelligent, equally informed and equally impartial." Both sides would agree on this, although for different reasons; believers in science see the ID crowd as creationist yahoos, and the yahoos portray us as being in the grip of a priori naturalism. Regardless of whatever might be stipulated about Joe and his interlocutors, our background knowledge shapes our intuitions about Case B.

Moreover, path dependence can result if Joe defers in Case A. Suppose there are a dozen witnesses who are evenly divided as to whether the light was red or green. Three of them compare notes before being questioned. Merely as a matter of chance, one of these three will be in the minority. Since the perceptual belief is far from the center of her web, she defers to the other two. Things continue until Joe has a chance to compare notes with his fellow witnesses. By that time, a majority favors one view or the other. This is not quite the same path dependence that worries Pettit; it is not relative to Joe's personal history, but relative to the history of the community. Nevertheless, it is enough to discredit the strategy of deference even for beliefs that are not deeply embedded.

This might seem to be an argument for 'no' in Case A, which would be in tension with the general argument for 'yes' that Pettit begins with. I think that this is just a result of the way the problem is represented. Before Joe discusses the accident with others, he believes that it looked to him that the light was red. Whether he defers to them or not, he should not change his belief about that. Rather, he might change his belief about whether the light was red. Similarly, if he listens to other witness' description of the accident, he will not defer on the basis of their beliefs that the light was green. Rather, he is interested in their reports about whether it looked to them as if the light was red. With this distinction in mind, the cascade to agreement would not occur.

Perhaps there is no such distinction in Case B. For non-perceptual beliefs, one might say, there is no clear distinction between saying how it seemed and judging how it was.** Of course, perceptual beliefs are less embedded in the web of belief. So this would just be Pettit's distinction again.


* He puts the scenarios in the second person, but I have shifted to the third person Joe. It would be presumptuous to stipulate your opinions about intelligent design in Case B.

** I have phrased this as a hypothetical because I am dubious of it. Even if detachment is permitted for some beliefs, it seems like scientific controversies require remembering which evidence one took to be persuasive. As such, Joe should distinguish between having believed a theory on the basis of some evidence and later disbelieving it because his clever friends do. I can distinguish between my sense of a scientific theory based on my (meager) understanding of the evidence and my sense based on what competent scientists tell me; I have beliefs about both, but if asked for a flat-footed judgment about the theory I would probably defer and give the latter.

[ add comment ] ( 6424 views )   |  [ 0 trackbacks ]   |  permalink
Wee-key-pedia guilt 
I have been working on a draft of 'Epistemology and the Wikipedia', a paper which I am going to present next month at the NA-CAP conference. In researching the paper, I have occasionally been struck by an interesting phenomenon. Let's call it Wikipedia Guilt.

The premise of the Wikipedia is that the community will extend and correct it. This is underscored by the rhetoric of Wikipedia's champions. "If you don't like an entry," they say, "write a better one." When I encounter an entry that is shot through with errors, I could do something about it. Often, I do not because-- although I can recognize an article as bunkum-- I lack expertise to replace it with anything authoritative. Yet in researching the paper, I have not even made changes to the entries on philosophical topics that I know a lot about. Although I can rationalize this as critical distance, complaining about solecisms is also a variety of Bad Faith: The entries are hokum partly because I tolerate their persisting as hokum.

To be frank, I try to avoid the Wikipedia when I am not wearing my critical epistemology hat. Nevertheless, I recently succumbed to pangs Wikipedia Guilt and rewrote an entry. This is how it happened:

Cristyn and I were discussing the phrase church key. To me, the canonical church key is made from a single piece of metal with a can punch on one end and a bottle opener on the other. Cristyn was unsure why such a gizmo should be called a church key, except derivatively from earlier gizmos that had looked more like keys.

My first stop for etymology is always the OED, but it just defines 'church-key' as 'the key of the church-door.' No help. On to the web.

Among other pages, my search turned up the Wikipedia entry. This is what it said:
In Medieval Europe, Monks and Nobility were the only brewers. Lagering Cellars in the Monasteries were locked, as the Monks guarded the secrets to their craft. The monks carried keys to these lagering cellars on their cinch - or belts. It was this key from which the "Church Key" opener gets its name.
: Source: Anheuser-Busch Knowledge Base; Internal Dbase
This is a nice story, with an ersatz citation to give it some gravitas. It is almost certainly apocryphal, however, even if the citation is legitimate. The History of the entry revealed that it had been written recently, completely overwriting a more modest entry that merely described what church keys are. This bit of verbal flotsam does not even bother with description; it is all about the bogus etymology.

Feeling a twinge of guilt, I wrote a new entry. Because, you know, I am now an authority on this kind of crap.

[ add comment ] ( 8610 views )   |  [ 0 trackbacks ]   |  permalink
forall x, truth and satisfaction 
Aaron Schiller used forall x for a course he taught in the Spring. A few weeks ago, I had coffee with him and discussed it. He pointed to two weak spots in the chapter on formal semantics, and also relayed his students' desire for more solved problems in the proofs chapter.

These comments jibed with my experience, so I made some changes to the book. I rewrote the section on reasoning about all possible models, breaking it into two parts. I also rewrote the section on truth in QL and the definition of satisfaction. I added two problem sets with solutions to the proofs chapter.

The new version (1.2) is now on-line in several formats. This is the version I will be teaching with in the Fall.

[ 1 comment ] ( 3321 views )   |  [ 0 trackbacks ]   |  permalink
Any publication you can walk away from is a good publication 
As I have mentioned before, I had a summer job in graduate school working with Mike Kalichman on The Responsible Conduct of Research website. After I left, I was credited as a coauthor.

Having just googled my own name, I notice that the whole thing is being reprinted as a serial in the newsletter of the Program in Research Integrity Education at the University of Arizona. The newsletter is subtitled "A Federally Mandated Compliance Education Program," which makes me wonder if anyone actually reads it. You know, as opposed to keeping it around as a kind of federally-mandated talisman.

[ add comment ] ( 3204 views )   |  [ 0 trackbacks ]   |  permalink

<<First <Back | 74 | 75 | 76 | 77 | 78 | 79 | 80 | 81 | 82 | 83 | Next> Last>>