Don’t miss V for Vendetta!

I don’t normally post on the weekends but last night I saw the film V for Vendetta and it blew me away. It is a brilliant political thriller with disturbing parallels to what is currently going on in the US. It kept me completely absorbed.

I’ll write more about it next week but this is just to urge people to see it before it ends its run.

Tiktaalik bridges the gap

As you can imagine, the world of science has been abuzz with the news yesterday of the release of a paper in the prestigious science journal Nature heralding the discovery of a 375-million year old transitional fossil between fish and land animal, which has been named Tiktaalik. The fossil seems to provide evidence of a key evolutionary idea that land animals evolved from fish.
[Read more…]

Precision in language

Some time ago, a commenter to this blog sent me a private email expressing this view:

Have you ever noticed people say “Do you believe in evolution?” just as you would ask “Do you believe in God?” as if both schools of thought have equal footing? I respect others’ religious beliefs as I realize I cannot disprove God just as anyone cannot prove His existence, but given the amount of evidence for evolution, shouldn’t we insist on asking “Do you accept evolution?”

It may just be semantics, but I feel that the latter wording carries an implied affirmation just as “Do you accept that 2+2=4?” carries a different meaning than “Do you believe 2+2=4?”

I guess the point I’m trying to make is that by stating something as a belief, it opens the debate to the possibility that something is untrue. While this may fine for discussions of religion, shouldn’t the scientific community be more insistent that a theory well supported by physical evidence, such as evolution, is not up for debate?

It’s a good point. To be fair, scientists themselves are partly responsible for this confusion because we also say that we “believe” in this or that scientific theory, and one cannot blame the general public from picking up on that terminology. What is important to realize, though, is that the word ‘believe’ is being used by scientists in a different sense from the way it is used in religion.

The late and deeply lamented Douglas Adams, author of The Hitchhiker’s Guide to the Galaxy, who called himself a “radical atheist” puts it nicely (thanks to onegoodmove):

First of all I do not believe-that-there-is-not-a-god. I don’t see what belief has got to do with it. I believe or don’t believe my four-year old daughter when she tells me that she didn’t make that mess on the floor. I believe in justice and fair play (though I don’t know exactly how we achieve them, other than by continually trying against all possible odds of success). I also believe that England should enter the European Monetary Union. I am not remotely enough of an economist to argue the issue vigorously with someone who is, but what little I do know, reinforced with a hefty dollop of gut feeling, strongly suggests to me that it’s the right course. I could very easily turn out to be wrong, and I know that. These seem to me to be legitimate uses for the word believe. As a carapace for the protection of irrational notions from legitimate questions, however, I think that the word has a lot of mischief to answer for. So, I do not believe-that-there-is-no-god. I am, however, convinced that there is no god, which is a totally different stance. . .

There is such a thing as the burden of proof, and in the case of god, as in the case of the composition of the moon, this has shifted radically. God used to be the best explanation we’d got, and we’ve now got vastly better ones. God is no longer an explanation of anything, but has instead become something that would itself need an insurmountable amount of explaining…

Well, in history, even though the understanding of events, of cause and effect, is a matter of interpretation, and even though interpretation is in many ways a matter of opinion, nevertheless those opinions and interpretations are honed to within an inch of their lives in the withering crossfire of argument and counterargument, and those that are still standing are then subjected to a whole new round of challenges of fact and logic from the next generation of historians – and so on. All opinions are not equal. Some are a very great more robust, sophisticated and well supported in logic and argument than others.

When someone says that they believe in god, they mean that they believe something in the absence of, or even counter to, the evidence, and even to reason and logic. When scientists say they believe a particular theory, they mean that they believe that theory because of the evidence and reason and logic, and the more evidence there is, and the better the reasoning behind it, the more strongly they believe it. Scientists use the word ‘belief’ the way Adams says, as a kind of synonym for ‘convinced,’ because we know that no scientific theory can be proven with 100% certainty and so we have to accept things even in the face of this remaining doubt. But the word ‘believe’ definitely does not carry the same meaning in the two contexts.

This can lead to the generation of confusion as warned by the commenter but what can we do about it? One option is, as was suggested, to use different words, with scientists avoiding use of the word ‘believe.’ I would have agreed with this some years ago but I am becoming increasingly doubtful that we can control the way that words are used.

For example, there was a time when I used to be on a crusade against the erroneous use of the word ‘unique’. The Oxford English Dictionary is pretty clear about what this word means:

  • Of which there is only one; one and no other; single, sole, solitary.
  • That is or forms the only one of its kind; having no like or equal; standing alone in comparison with others, freq. by reason of superior excellence; unequalled, unparalleled, unrivalled.
  • Formed or consisting of one or a single thing
  • A thing of which there is only one example, copy, or specimen; esp., in early use, a coin or medal of this class.
  • A thing, fact, or circumstance which by reason of exceptional or special qualities stands alone and is without equal or parallel in its kind.

It means, in short, one of a kind, so something is either unique or it is not. There are no in-betweens. And yet, you often find people saying things like “quite unique” or “very unique” or “almost unique.” I used to try and correct this but have given up. Clearly, people in general think that unique means something like “rare” and I don’t know that we can ever change this even if we all become annoying pedants, correcting people all the time, avoided at parties because of our pursuit of linguistic purity.

Some battles, such as with the word unique are, I believe, lost for good and I expect the OED to add the new meaning of ‘rare’ some time in the near future. It is a pity because then we would then be left with no word with the unique meaning of ‘unique’, but there we are. We would have to say something like ‘absolutely unique’ to convey the meaning once reserved for just ‘unique.’

In science too we often use words with precise operational meanings while the same words are used in everyday language with much looser meanings. For example, in physics the word ‘velocity’ is defined operationally by the situation when you have an object moving along a ruler and, at two points along its motion, you take ruler readings and clock readings, where the clocks are located at the points where the ruler readings are taken, and have been previously synchronized. Then the velocity of the moving object is the number you get when you take the difference between the two ruler readings and divide by the difference between the two clock readings.

Most people (especially sports commentators) have no idea of this precise meaning when they use the word velocity in everyday language, and often use the word synonymously with speed or, even worse, acceleration, although those concepts have different operational meanings. Even students who have taken physics courses find it hard to use the word in its strict operational sense.

Take, for another example, the word ‘theory’. By now, as a result of the intelligent design creationism (IDC) controversy, everyone should be aware that the way this word is used by scientists is quite different from its everyday use. In science, a theory is a powerful explanatory construct. Science depends crucially on its theories because they are the things that give it is predictive power. “There is nothing so practical as a good theory” as Kurt Lewin famously said. But in everyday language, the word theory is used as meaning ‘not factual,’ something that can be false or ignored.

I don’t think that we can solve this problem by putting constraints on how words can be used. English is a wonderful language precisely because it grows and evolves and trying to fix the meanings of words too rigidly would perhaps be stultifying. I now think that we need to change our tactics.

I think that once the meanings of words enter mainstream consciousness we will not be successful in trying to restrict their meanings beyond their generally accepted usage. What we can do is to make people aware that all words have varying meanings depending on the context, and that scientific and other academic contexts tend to require very precise meanings in order to minimize ambiguity.

Heidi Cool has a nice entry where she talks about the importance of being aware of when you are using specialized vocabulary, and the need to know your audience when speaking or writing, so that some of the pitfalls arising from the imprecise use of words can be avoided.

We have to realize though that despite our best efforts, we can never be sure that the meaning that we intend to convey by our words is the same as the meaning constructed in the minds of the reader or listener. Words always contain an inherent ambiguity that allows the ideas expressed by them to be interpreted differently.

I used to be surprised when people read the stuff I wrote and got a different meaning than I had intended. No longer. I now realize that there is always some residual ambiguity in words that cannot be overcome. While we can and should strive for maximum precision, we can never be totally unambiguous.

I agree with philosopher Karl Popper when he said, “It is impossible to speak in such a way that you cannot be misunderstood.” The best we can hope for is to have some sort or negotiated consensus on the meanings of ideas.

On writing-2: Why do we cite other people’s work?

In the previous post on this topic, I discussed the plagiarism case of Ben Domenech, who had lifted entire chunks of other people’s writings and had passed them off as his own.

How could he have done such a thing? After all, all high school and college students get the standard lecture on plagiarism and why it is bad. And even though Domenech was home schooled, it seems unlikely that he thought this was acceptable practice. When he was confronted with his plagiarism, his defense was not one of surprise that it was considered wrong but merely that he had been ‘young’ when he did it or that he had got permission from the author to use their words or that the offending words had been inserted by his editors.

The cautionary lectures that students receive about plagiarism are usually cast in a moralistic way, that plagiarism is a form of stealing, that taking someone else’s words or ideas without proper attribution is as morally reprehensible as taking their money.

What is often overlooked in this kind of approach is that there are many other reasons why writers and academics cite other people’s works when appropriate. By focusing too much on this stealing aspect, we tend to not give students an important insight into how scholarship and research works.

Russ Hunt at St. Thomas University argues that writers cite others for a whole complex of reasons that have little to do with avoiding charges of plagiarism:

[P]ublished scholarly literature is full of examples of writers using the texts, words and ideas of others to serve their own immediate purposes. Here’s an example of the way two researchers opened their discussion of the context of their work in 1984:

To say that listeners attempt to construct points is not, however, to make clear just what sort of thing a ‘point’ actually is. Despite recent interest in the pragmatics of oral stories (Polanyi 1979, 1982; Robinson 1981), conversations (Schank et al. 1982), and narrative discourse generally (Prince 1983), definitions of point are hard to come by. Those that do exist are usually couched in negative terms: apparently it is easier to indicate what a point is not than to be clear about what it is. Perhaps the most memorable (negative) definition of point was that of Labov (1972: 366), who observed that a narrative without one is met with the “withering” rejoinder, “So what?” (Vipond & Hunt, 1984)

It is clear here that the motives of the writers do not include prevention of charges of plagiarism; moreover, it’s equally clear that they are not. . .attempting to “cite every piece of information that is not a) the result of your own research, or b) common knowledge.” What they are doing is more complex. The bouquet of citations offered in this paragraph is informing the reader that the writers know, and are comfortable with, the literature their article is addressing; they are moving to place their argument in an already existing written conversation about the pragmatics of stories; they are advertising to the readers of their article, likely to be interested in psychology or literature, that there is an area of inquiry — the sociology of discourse — that is relevant to studies in the psychology of literature; and they are establishing a tone of comfortable authority in that conversation by the acknowledgement of Labov’s contribution and by using his language –“withering” is picked out of Labov’s article because it is often cited as conveying the power of pointlessness to humiliate (I believe I speak with some authority for the authors’ motives, since I was one of them).

Scholars — writers generally — use citations for many things: they establish their own bona fides and currency, they advertise their alliances, they bring work to the attention of their reader, they assert ties of collegiality, they exemplify contending positions or define nuances of difference among competing theories or ideas. They do not use them to defend themselves against potential allegations of plagiarism.

The clearest difference between the way undergraduate students, writing essays, cite and quote and the way scholars do it in public is this: typically, the scholars are achieving something positive; the students are avoiding something negative. (my italics)

I think that Hunt has hit exactly the right note.

When you cite the works of others, you are strengthening your own argument because you are making them (and their allies) into your allies, and people who challenge what you say have to take on this entire army. When you cite reputable sources or credible authorities for facts or ideas, you become more credible because you are no longer alone and thus not easily dismissed, even if you personally are not famous or a recognized authority.

To be continued. . .

POST SCRIPT: It’s now Daylight Saving Time. Do you know where your spiritual plane is?

It seems like idiotic statements attributing natural events to supernatural causes are not restricted to Christian radical clerics like Pat Robertson. Some Sri Lankan Buddhist clergy are challenging him for the title of Religious Doofus.

Since Sri Lanka sits very close to the equator, the length of the day is the same all year round, not requiring the ‘spring-forward-fall-back’ biannual adjusting of the US. Sri Lankan time used to be 5.5 hours ahead of Universal Time (UT) but in 1996 the government made a one-time shift it to 6.5 hours in order to have sunset arrive later and save energy. But the influential Buddhist clergy were not happy with the change. As a compromise, the clocks were then again adjusted to make it just 6.0 ahead of UT as a compromise. Now the government is thinking of going back to the original 5.5. hours.

Some of the country’s Buddhist clergy are rejoicing at the prospect of a change because they say Sri Lanka’s “old” time fitted better with their rituals.

They believe a decade living in the “wrong” time has upset the country’s natural order with terrible effect.

The Venerable Gnanawimala says the change moved the country to a spiritual plane 500 miles east of where it should be.

“After this change I feel that many troubles have been caused to Sri Lanka. Tsunamis and other natural disasters have been taking place,” he says.

This is what happens when you mix religion and the state. You now have to worry about what your actions are doing to the longitudinal coordinates of your nation’s spiritual plane.

No more daft women!

Evan Hunter, who was the screenwriter on Alfred Hitchcock’s 1963 film The Birds recalled an incident that occurred when he was discussing the screenplay with the director.

I don’t know if you recall the movie. There’s a scene where after this massive bird attack on the house Mitch, the male character, is asleep in a chair and Melanie hears something. She takes a flashlight and she goes up to investigate, and this leads to the big scene in the attic where all the birds attack her. I was telling [Hitchcock] about this scene and he was listening very intently, and then he said, “Let me see if I understand this correctly. There has been a massive attack on the house and they have boarded it up and Mitch is asleep and she hears a sound and she goes to investigate?” I said, “Well, yes,” and he said, “Is she daft? Why doesn’t she wake him up?”

I remembered this story when I was watching the film The Interpreter with Nicole Kidman and Sean Penn. The Kidman character accidentally overhears something at the UN that puts her life at risk. After she complains to government agent Penn that no one seems to be bothered about protecting her from harm, Penn puts her on round-the-clock surveillance. So then what does Kidman do? She sneaks around, giving the slip to the very people assigned to protect her and refuses to tell Penn where she went and to whom she spoke and about what, causing herself and other people to be put at risk and even dying because of her actions. Hitchcock would have said, “Is she daft?”

This is one of my pet peeves about films, where the female character insists on doing something incredibly stupid that puts her and other people at peril. Surely in this day and age we have gone beyond the stale plot device of otherwise smart women behaving stupidly in order to create drama? Surely writers have more imagination than that? Do directors really think that viewers won’t notice how absurd that is?

According to Hunter, Hitchcock was always exploring the motivations of characters, trying to make their actions plausible. Hunter says:

[Hitchcock] would ask surprising questions. I would be in the middle of telling the story so far and he would say, “Has she called her father yet?” I’d say, “What?” “The girl, has she called her father?” And I’d say, “No.” “Well, she’s been away from San Francisco overnight. Does he know where she is? Has she called to tell him she’s staying in this town?” I said, “No.” And he said, “Don’t you think she should call him?” I said, “Yes.” “You know it’s not a difficult thing to have a person pick up the phone.” Questions like that.

(Incidentally, the above link has three screenwriters Arthur Laurents, who wrote Rope (1948), Joseph Stefano, who wrote Psycho (1960), and Evan Hunter reminiscing about working with Hitchcock. It is a fascinating glimpse behind the scenes of how a great director envisages and sets about creating films. The last quote actually reads in the original: “Yes, you know it’s not a difficult thing to have a person pick up the phone.” I changed it because my version makes more sense, and the original is a verbatim transcript of a panel discussion, in which such kinds of punctuation errors can easily occur.)

More generally, I hate it when characters in films and books behave in ways that are unbelievable. The problem is not with an implausible premise, which is often necessary to create a central core for the story. I can even accept the violation of a few laws of physics. For example, I can accept the premise of Superman that a baby with super powers (but susceptible to kryptonite) arrives on Earth from another planet and is adopted by a family and needs to keep his identity secret. I can accept of Batman that a millionaire like Bruce Wayne adopts a secret identity in order to fight crime.

What I cannot stand is when they and the other people act implausibly, when the stories built on this premise have logical holes that you can drive a Batmobile through. The latter, for example, is a flashy vehicle, to say the least, easily picked out in traffic. And yet, nobody in Gotham thinks of following it back to the Batcave, to see who this mysterious hero is. Is the entire population of that city daft?

And how exactly does the Bat-Signal that the Police Commissioner lights up the sky with supposed to work? You don’t need a physics degree to realize that shining a light, however bright, into the sky is not going to create a sharp image there. And what if it’s daytime? And if there are no clouds? (It’s been a long time since I read these comics. Maybe the later editions fixed these problems. But even as a child these things annoyed me.)

And don’t get me started on Spiderman going in and out of his apartment window in a building in the middle of a big city in broad daylight without anyone noticing.

As a fan of films, it really bugs me when filmmakers don’t take the trouble to write plots that make sense, and have characters who don’t behave the way that you would expect normal people to behave. How hard can it be to ensure this, especially when you have the budget to hire writers to create believable characters and a plausible storyline?

If any directors are reading this, I am willing to offer my services to identify and fix plot holes.

So please, no more daft women! No more ditzy damsels in distress! No more Perils of Pauline!

POST SCRIPT: CSA: Confederate States of America

I saw this film last week (see the post script to an earlier posting), just before it ended its very short run in Cleveland. It looks at what history would have been like if the south had won the civil war. Imagine, if you will, an America very much like what we have now except that owning black slaves is as commonplace as owning a dishwasher.

What was troubling is that although this is an imagined alternate history presented in a faux documentary format, much of it is plausible based on what we have now. What was most disturbing for me was seeing in the film racist images and acts that I thought were the over-the-top imaginings of the screenwriter about that might have happened in this alternate history, and then finding out that they actually happened in the real history.

Although the film is a clever satire in the style of This is Spinal Tap, I could not really laugh because the topic itself is so appalling. It is easy to laugh at the preening and pretensions of a rock band. It is hard to laugh at people in shackles.

But the film was well worth seeing, disturbing though it was.

On writing-1: Plagiarism at the Washington Post

If you blinked a couple of weeks ago, you might have missed the meteor that was the rise and fall of the career of Ben Domenech as a blogger for WashingtonPost.com.

This online version of the newspaper is apparently managed independently of the print edition and has its own Executive Editor Jim Brady. For reasons that are not wholly clear, Brady decided that he needed to hire a “conservative” blogger for the website.

The problem with this rationale for the hiring was that no “liberal” counterpart blogger existed at the paper. They did have a popular blogger in Dan Froomkin, someone with a journalistic background, who wrote about politics for the Post and who had on occasion been critical of the Bush White House. As I have written earlier, Glenn Greenwald has pointed out that anything but unswavering loyalty to Bush has become the basis for identifying someone as liberal, and maybe Brady had internalized this critique, prompting him to hire someone who could be counted upon to support Bush in all his actions.

For reasons that are even more obscure, rather than choose someone who had serious journalistic credentials for this new column, Brady selected the untested 24-year old Ben Domenech. It is true that Domenech was something of a boy wonder, at least on paper. He had been home-schooled by his affluent and well-connected Republican family. He then went to William and Mary and wrote for their student newspaper The Flat Hat. He dropped out of college before graduating and co-founded a conservative website called Redstate, where he wrote under the pseudonym Augustine.

His father was a Bush political appointee and his new online column for the Washington Post (called Red America) said in its inaugural posting on March 21 that young Ben “was sworn in as the youngest political appointee of President George W. Bush. Following a year as a speechwriter for HHS Secretary Tommy Thompson and two as the chief speechwriter for Texas Senator John Cornyn, Ben is now a book editor for Regnery Publishing, where he has edited multiple bestsellers and books by Michelle Malkin, Ramesh Ponnuru, and Hugh Hewitt.”

Not bad for a 24-year old without a college degree. And his bio lists even more accomplishments. But getting his own column in WashingtonPost.com was the peak. Soon after that things started going downhill very rapidly.

His decline began when bloggers looked into his writings and found that, as Augustine, he had written a column of the day of Coretta Scott King’s funeral calling her a Communist. This annoyed a lot of people who then started looking more closely at his other writings. It was then that someone discovered that he had plagiarized. And the plagiarism was not subtle. Take for example this excerpt from his review of the film Bringing out the Dead.

Instead of allowing for the incredible nuances that Cage always brings to his performances, the character of Frank sews it all up for him.

But there are those moments that allow Cage to do what he does best. When he’s trying to revive Mary’s father, the man’s family fanned out around him in the living room in frozen semi-circle, he blurts out, “Do you have any music?”

Now compare it with an earlier review posted on Salon.com,

Instead of allowing for the incredible nuance that Cage always brings to his performances, the character of Frank sews it all up for him. . . But there are those moments that allow Cage to do what he does best. When he’s trying to revive Mary’s father, the man’s family fanned out around him in the living room in frozen semi-circle, he blurts out, “Do you have any music?”

Or this sampling from P. J. O’Rourke’s book Modern Manners, which also found its way into Domenech’s columns:

O’Rourke, p.176: Office Christmas parties • Wine-tasting parties • Book-publishing parties • Parties with themes, such as “Las Vegas Nite” or “Waikiki Whoopee” • Parties at which anyone is wearing a blue velvet tuxedo jacket

BenDom: Christmas parties. Wine tasting parties. Book publishing parties. Parties with themes, such as “Las Vegas Nite” or “Waikiki Whoopee.” Parties at which anyone is wearing a blue velvet tuxedo jacket.

O’Rourke: It’s not a real party if it doesn’t end in an orgy or a food fight. • All your friends should still be there when you come to in the morning.

BenDom: It’s not a real party if it doesn’t end in an orgy or a food fight. All your friends should still be there when you come to in the morning.

These are not the kinds of accidental plagiarisms that anyone can fall prey to, where a turn of phrase that appealed to you when you read it a long time ago comes out of you when you are writing and you do not remember that you got it from someone else. These examples are undoubtedly deliberate cut-and-paste jobs.

Once the charges of plagiarism were seen to have some credibility, many people went to Google and the floodgates were opened, Kaloogian-style, with bloggers all over poring over his writings. Within the space of three days a torrent of further examples of plagiarism poured out. These new allegations dated back to his writings at his college newspaper and then later for National Review Online, and Domenech was found to have lifted material from Salon and even from National Review Online, the latter being the same publication for which he was writing, which adds the sin of ingratitude to the dishonesty.

On March 24, just three days after starting his Washington Post column, Ben Domenech resigned under pressure. Soon after, he also resigned as book editor at Regnery.

What can we learn from this? One lesson seemingly is that people can get away with plagiarism for a short while, especially if they are writing in obscurity for little known publications. While he was writing for his college newspaper and even for his own website, no one cared to closely look into his work. Even his future employers at WanshintonPost.com did not seem to have checked him out carefully. Apparently his well-connected family and sterling Bush loyalty was enough to satisfy them that he was a good addition to their masthead.

But as soon as a writer becomes high profile, the chances are very high these days that any plagiarism will come to light.

At one level, this is a familiar cautionary tale to everyone to cite other people’s work when using it. For us in the academic world, where plagiarism is a big no-no, the reasons for citing are not just there are high penalties if you get caught not doing it. The more important reasons arise from the very nature of scholarly academic activity, which I shall look at in a future posting.

To be continued. . .

Changing notions of death-4: Implications for animals

(See part 1, part 2 and part 3 of this series.)

If asked, any one of us would say that we value life, that we consider it precious and not to be taken lightly. While the specific phrase “valuing the culture of life” seems to have been co-opted by those who are specifically opposed to abortion, the general idea that it encapsulates, that life should not be taken casually or at all, is one that all of us would subscribe to.

But of course there are contradictions. People who say they value life often see no problem with supporting the death penalty. Another hypocrisy is with those who support killing in wars, even of civilians, and even in large numbers. We try to rationalize this by saying that civilians are killed inadvertently, but that is a false argument. Civilians are inevitably killed in wars, often deliberately, and we often do nothing to condemn it when it is done by ‘our side.’ To support wars is to support killing and absolve killers, however much we try to sugar coat this unpleasant fact. As Voltaire said, “It is forbidden to kill; therefore all murderers are punished unless they kill in large numbers and to the sound of trumpets.”

In his lecture, Peter Singer pointed out that killing and eating animals, while opposing the withdrawal of life support of those in a persistent vegetative state, poses an ethical problem for people who say that they value a “culture of life.”

He gave as an example the fact that while the 3,000 or so victims of September 11, 2001 were deeply mourned, no one mourned the fact that millions of chickens were killed on that same day and every day before and since. But we do not mourn them the same way. Why not?

If we define death as heart dead or brain dead, then the chickens are as alive as any of us. Even when we lower the bar to thinking of someone in a persistent vegetative state as being ‘effectively dead’, that still does not get us off the hook since, as Singer argued, chickens and other animals have higher levels of consciousness than people in a persistent vegetative state. Free range chickens seem to show signs of happiness, curiosity, anxiety, fear, and the sense of self-awareness that, if present in humans, would definitely bar us from killing them. If that is the case, then if we oppose the withdrawal of life support systems even from those in a persistent vegetative state, then how can we justify killing chickens, or any other animal for that matter?

He posed the question of why the killing of human beings is deplored but that of chickens is not. He said that appealing to species chauvinism (“We are human, and so are justified in valuing human life over non-human animal life.”) was not really an ethically justifiable defense, though many people used it.

After all, if we allowed that particular chauvinist line of defense, where do we draw the line? What if I say that because I am male, I am justified in thinking that the lives of women are worth less than that of men? We would reject that line of argument as rank sexism. What if I say that because I am brown skinned, I am justified in treating non-brown people as inferior? We would reject that argument as rank racism. So why should we think that the argument “I am human so I am justified in valuing human life over animal life?” is acceptable?

Singer’s point was that as soon as we shift our definition of death from that defined by the complete lack of heart or brain function, and to judgments about the nature or level of the consciousness involved, we have gone into ethically tricky territory for those non-vegetarians who argue that because of belief in a “culture of life,” human beings must be kept alive at all costs. Because you cannot argue that people in a persistent vegetative state should be kept on life support while arguing that perfectly healthy animals can be killed.

People of certain religious traditions (like Christians, Jews, and Muslims) perhaps can find justification for this discrepant behavior by appealing to their religious beliefs that include species chauvinism as part of their doctrines. In the view of these religions, humans are specially favored by god and thus fundamentally different from, and superior to, other animals so valuing human life and disregarding non-human animal life is allowable. It is noteworthy that Buddhism and Hinduism do not assert such a species chauvinistic attitude. They seem to treat human and non-human animals on an equal footing and vegetarianism is advocated by both religions.

But if we leave out religious sanction and argue on strictly ethical grounds, it becomes hard to justify opposing the withdrawal of life support systems to people who are in a persistent vegetative state on the grounds that such people are still ‘alive’, and square it with the killing of healthy animals for food, as we routinely do.

Singer made a cogent argument that none of us can really ethically justify the killing of animals for food, when it is not necessary for survival. Singer himself is a vegetarian.

I am not sure if Singer was able to resolve some of the ethical issues of what constitutes death by the end of his talk, after I had left. But his ideas were very thought provoking.

POST SCRIPT: Juggling

Good jugglers are amazing. For a fine example of this art, go here and then click on “Watch Chris Bliss.”

Internet sleuthing

I am a firm believer in cooperative learning. The combined efforts of many people can produce results that would be impossible for a single person. And the internet is a wonderful mechanism for enabling collective action.

What follows is a modern-day detective story that illustrates what is possible when the collective strength of people working together, sharing information and ideas, and building on each others’ ideas, combined with the speed of communication and resources available on the internet.

Howard Kaloogian is a Republican candidate in San Diego’s special congressional election to replace disgraced Republican Duke Cunningham, who pleaded guilty to bribery, resigned his seat, and is now in jail. In his campaign, Kaloogian tried to propagate the White House meme that things are just peachy in Iraq and that the media is deliberately sabotaging the war effort, painting a dark picture by reporting only the daily bombings, beheadings, kidnappings, extortions, etc.
[Read more…]

Changing notions of death-3: Doctors versus guardians

In part 1 and part 2 of this series of posts, we saw how the idea of when someone had died had shifted to the point where people in a persistent vegetative state could have their life support systems removed because they are considered to be ‘effectively’ dead. But even if the family is agreed on what action should be taken with a family member in a persistent vegetative state, there are already moves under way to shift the bar even lower. The question now being raised is as to what should be done if the doctors determine that further treatment is futile but the family does not want to remove life support.

In a recent case in England, doctors had recommended that an 18-month old infant (identified as just MB) who suffers from the severest form of spinal muscular atrophy – an incurable and progressively worsening condition leading to complete paralysis – be allowed to die. The parents objected and the matter went to trial.

On March 15, 2006, the judge ruled in the parents’ favor, refusing to declare that it would be lawful to withdraw life-sustaining ventilation.

A momentary look of wistfulness passed over the face of MB’s mother as the judge listed five possible options, one of which was to allow the child to die peacefully in his parents’ arms – the one favoured by the paediatricians. The parents have fought long and hard against the received medical wisdom of the case, even though, as the judge said, they may be deluding themselves that their son has a future.

At long last, Mr Justice Holman gave his ruling that the boy shall live, if not perhaps for long.

In that case, the judge in England did not shift the goal posts on what constitutes death or the conditions under which people are ‘allowed to die.’

But people might be surprised to know that a similar situation had occurred in the US and that doctors and hospitals were allowed to override the family’s will. Remarkably, this little noticed event took place on March 15, 2006 during the high point of the events surrounding Terri Schiavo.

While Americans were riveted by dramatic events unfolding in Pinellas Park, Fla., a five-month-old Houston baby took his last breath after a hospital let him die despite his mother’s objections.

Sun Hudson was born Sept. 25 with thanatophoric dysplasia, an incurable and fatal form of dwarfism. Doctors said his tiny lungs would never fully grow and that he would never breathe on his own.

Hudson’s mother, Wanda, put up a fight when doctors advised removing Sun from a respirator. She said she did not believe in sickness or death. (my italics)

This was the first time that life support was removed over the objections of the legal guardian and without any advance directives from the patient, such as a living will. Perhaps the ultimate irony, if not outright hypocrisy, was that this Texas law was signed in 1999 by then Governor George W. Bush. The baby Sun Hudson was allowed to die in Texas against the wishes of his mother because of a state law then-Governor Bush signed, on the very same day that now-President Bush dramatically cut short his vacation and flew back to Washington to sign the federal law that supported the parents’ right to keep life support continuing for Terri Schiavo.

The doctors were able to override the mother’s wishes on March 15, 2005 because the case took place in Texas and that state has a law that authorizes doctors and hospitals to override the wishes of the patient’s families. The hospital took this action under the The Texas Advance Directives Act (1999), also known as the Texas Futile Care Law, which according to Wikipedia, “describes certain provisions that are now Chapter 166 of the Texas Health & Safety Code. Controversy over these provisions mainly centers on Section 166.046, Subsection (e), which allows a health care facility to discontinue life-sustaining treatment against the wishes of the patient or guardian ten days after giving written notice.”

As with the case of shifting the definition of death from heart dead to brain dead, serious ethical issues are raised by this act. There are concerns that this law was passed because hospitals did not want to shoulder the cost of maintaining life support for patients who cannot pay for it. Although the law (as I read it) does not explicitly say that the inability to pay for life support can be a reason for termination of services, it is easy to see that financial considerations are going to come into play.

It is unlikely that patients who have rich families who can pay the bills are going to have their wishes overridden and life support removed. But one can see why hospitals, which have become businesses, would not like the prospect of indefinitely providing expensive life support care if they have no hope of being reimbursed. What adds further suspicion to the view that commercial concerns are significant is that if another hospital is willing to accept the patient, then the patient can be shifted there. But it is unlikely that another hospital is going to accept a new patient who requires extensive life support when that patient is unable to pay.

This blatant hypocrisy and contradiction between Bush’s behavior as governor of Texas and as President later did not go completely unnoticed, though it did not get the attention it warranted. In an editorial on March 22, 2005, the Concord Monitor voiced concern over the implications of the Texas law:

On the same day President Bush interrupted his vacation and rushed to Washington to sign the Schiavo bill, a Texas hospital removed the breathing tube keeping 6-month-old Sun Hudson alive. According to The Houston Chronicle, the hospital’s action, the first of its kind, was made possible by a 1999 bill signed into law by Bush, then Texas’s governor.

That law allows hospitals to discontinue life-sustaining care even when doing so runs counter to the wishes of the patient’s guardians. Before ending the patient’s life under the law Bush signed, however, two conditions must be met. Doctors must deem that there is no chance for recovery and the patient must be unable to pay the hospital bill for continuing care. (my italics)

Added John Paris, a medical ethicist at Boston College, told Newsday “The Texas statute that Bush signed authorized the ending of the life, even over the parents’ protest. And what he’s doing here is saying, ‘The parents are protesting. You shouldn’t stop [treatment]'”

Apart from this being another example of Bush subordinating principle to political expediency, it also clearly shows that society is steadily lowering the bar on death, first making it a judgment of whether someone is ‘effectively dead’ and who gets to make that decision, and now coming down to the question of whether someone is worth keeping alive and putting that decision (at least in Texas) in the hands of doctors and hospitals and not parents and guardians. While the judgment that further treatment is futile may be a medical and scientific judgment, the decision to withdraw life support will undoubtedly be also driven by financial considerations as to whether the patients and their families can pay the cost of continued treatment.

To be continued. . .

POST SCRIPT: Canned bird hunts

When not shooting old friends in the face, ‘Deadeye Dick’ Cheney kills birds for fun, and has killed up to 70 pheasants in just one shooting session. What is more, the birds he shot were bred in captivity to make them easy targets and one wonders what kind of fascination he finds in personally slaughtering such a large number of tame birds.

The comic strip Doonesbury suggests one reason, and Nate Corddry from The Daily Show tries to find out what the thrill is by going on one such canned quail hunt and bringing back a report.

Changing notions of death-2: Persistent vegetative state

The next stage in the evolution of when death occurs (see part 1 on this topic) came with the tragic case of Nancy Cruzan.

In 1983, 25-year old Nancy Cruzan careened off the road, flipped over and was thrown from her car into a ditch. Nancy hadn’t breathed for at least 15 minutes before paramedics found and revived her – a triumph of modern medicine launching her family’s seven-year crusade to free Nancy from a persistent vegetative state.

Nancy Cruzan’s sad fate launched a fresh examination of death, centering around whether a person in a particular kind of coma, known as a persistent vegetative state, could be considered to be ‘effectively dead’ even if they did not meet the legal conditions of being heart dead or brain dead.
[Read more…]