Posts Tagged ‘debate’
is belief in god irrational? – that is not the question
Debates between theists and atheists have become commonplace over the past few years, for better or worse, and the topic has often been vague enough to allow the protagonists plenty of leeway to espouse their views. True or false, rational or irrational, these are the oppositional terms most often used. These debates are often quite arid, with both parties firing from fixed positions and very carefully concealing from observers any palpable hits they’ve received from the other side. Whether they’ve contributed to the continued rise of the nones is hard to say.
I heard another one recently, bearing the title Is belief in God irrational? It was hosted on the Reasonable Doubts podcast, one that I recommend to those interested in the claims of Christianity in particular, as these ‘doubtcasters’ know their Bible pretty well and are well up on Christian politics, particularly in the US. The debaters were Chris Hallquist (atheist) and Randal Rauser (theist), and it was pretty hard to listen to at times, with much squabbling and point-scoring over the definition of rationality, and obscure issues of epistemology. I found the theist in particular to be shrill and often quite unpleasant in his faux-contempt for the other side, but then I’m probably biased.
I found myself, as I very often do, arguing or speculating my way through the topic from a very different standpoint, and here are my always provisional thoughts.
Let me begin by more or less rejecting two of the terms of the debate, ‘God’ and ‘irrational’. I’m not particularly interested in God, that’s to say the Judeo-Christian god, and I strongly object to designating that particular amalgam of Canaanite, Ugaritic and other Semitic deities as capital G God, as if one can, through a piece of semantic legerdemain, magick away the thousands of other deities that people have worshipped and adhered to over the centuries. It’s as if the Apple company chose to name their next Ipad ‘Tablet’, thereby rendering irrelevant all the other tablets produced by competing companies. Of course we have marketing regulations that prevent that sort of manipulation, but not so in religion.
So I will refer henceforth to gods, or supernatural entities and supernatural agency, with all their various and sometimes contradictory qualities, rather than to God, as defined by Aquinas and others. It is supernatural agency of any kind that I call into question.
More important for me, though, is the question of rationality. I’m not a philosopher, but I’ve certainly dipped into philosophy many times over the past 40 years or so, and I’ve even been obsessed with it at times. And rationalism has long been a major theme of philosophers, but I’ve never found a satisfactory way to define it. In the context of this debate, I would prefer the term ‘reasonable’ to ‘rational’. Being reasonable has a more sociable quality to it, it lacks the hard edge of rationality. So, for my purposes I’ll re-jig the topic to – Is belief in supernatural entities reasonable?
But I want to say more about rationality, to illustrate my difficulties with the term. Hume famously or perhaps notoriously wrote that reason can never be more than the slave of the emotions. This raises the question – what are these emotions that have such primacy and why are they so dominant? I have no doubt that a modern-day Hume – and Hume was always interested in the science of his day – would write differently about the factors that dominate and guide our reason. He would write about evolved instincts as much as about emotions. Above all the survival instinct, which we appear to share with every other living creature. Let me give some examples, which might bring some of our fonder notions of rationality into question.
A large volume of psychometric data in recent years has told us that we generally have a distorted view of ourselves and our competence. In assessing our physical attractiveness, our driving ability, our generosity to others and just about everything else, we take a more flattering view of ourselves than others take of us. What’s more, this is seen as no bad thing. In terms of surviving and thriving in a competitive environment, there’s a pay-off in being over-confident about your attractiveness, as a romantic partner, a business partner, or your nation’s Prime Minister. Of course, if you’re too over-confident, if the distortion between reality and self-perception becomes too great, it will act to your detriment. But does this mean that having a clear-eyed, non-distorted view of your qualities is rational, by that very fact, or irrational, because it puts you at a disadvantage vis-à-vis others? To put it another way, does rationality mean conformity to strict observation and logic, or is it behaviour that contributes to success in terms of well-being and thriving (within the constraints of our profoundly social existence)?
I don’t have any (rational?) answer to that conundrum, but I suppose my preference for the term ‘reasonable’ puts me in the second camp. So my answer to my own question, ‘Is it reasonable to believe in supernatural entities’ is that it depends on the circumstances.
Let’s look at belief in Santa, an eminently supernatural entity. He is, at least on Christmas Eve, endowed with omnipresence, being able to enter hundreds of millions of houses laden with gifts in an impossibly limited time-period. He’s even able to enter all these houses through the chimney in spite of the fact that 99.99% of them don’t have chimneys. What’s more, he’s omniscient, ‘He knows if you’ve been bad or good’, according to the sacred hymn ‘Santa Claus is coming to town’, ostensibly written by J F Coots and Haven Gillespie, but they were really just conduits for the Word of Santa. We consider it perfectly reasonable for three- and four-year-olds to believe in Santa, and, apart from some ultra-rationalist atheists and more than a few cultish Christians and adherents of rival deities, we generally encourage the belief. Clearly, we believe it does no harm and might even do some good. An avuncular, convivial figure with a definite fleshly representation, he’s also remote and mysterious with his supernatural powers and his distant home at the North Pole, which to a preschooler might as well be Mars, or Heaven. As an extra parent, he increases the quotient of love, security and belonging. To be watched over like Santa watches over kids might seem a bit creepy as you get older, but three-year-olds would have no such concerns, they’d accept it as their due, and would no doubt find his magical powers as well as his total jollity, knowledge and insight thoroughly inspiring as well as comforting. From a parent’s perspective, it’s all good, pretty much.
Of course, if your darling 23-year-old believes in Santa, that’s a problem. We expect our kids to grow out of this belief, and they rarely disappoint. They don’t need much encouragement. Children are bombarded with TV Santas, department store Santas, skinny Santas, bad Santas, Santas that look just like their Uncle Bill, etc, and they usually go through a period of jollying their parents along before making their big apostate announcement. Santas are human, all too human.
Santa belief is, it would seem, a harmless and perhaps positive massaging of a child’s vivid imagination, but when a child’s ready for school, she’s expected to put away childish things, little by little.
And isn’t that what many atheists say about the deities of the Big Monotheisms? Yes, but too many atheists underestimate the hurdles that need to be overcome. Most of these atheists either already live in highly secularised societies, such as here in Australia, or other English-speaking or European countries. Even the USA has many more atheists in it than the entire population of Australia, if we make the conservative assumption that 10% of its citizens are non-believers. Atheists are learning to club together but the religious have been doing it for centuries, and you’re likely to lose a lot of club benefits if you declare yourself a non-believer in a region of fervent or even routine belief. Or worse – I just read today of a Filipino lad who was murdered by his schoolmates after coming out as an atheist on a networking site. So just from a self-preservation point of view it might be reasonable to at least pretend to believe, in certain circumstances.
But there are many other situations in which it’s surely reasonable to believe – I mean really believe – in the supernatural agent or agents of your culture. The first of these is that supernatural agency explains things more satisfactorily to more people than any other available explanation. This might sound strange coming from a non-believer like myself, but it’s undoubtedly true. Bear in mind that I’m talking about satisfactory explanations, not true ones, and that I’m talking about most, not all, people.
Why was belief in supernatural agency virtually universal in the long ago? I don’t think that’s hard to understand. As human populations grew and became more successful in terms of harnessing of resources and domination of the landscape, they came to realise that they were prey to forces well beyond their control, forces that threatened them more seriously than any earthly predator. Famine, disease, earthquakes, storms, the seemingly arbitrary deaths of new-borns, sudden outbreaks of warfare between once-neighbourly tribes – all of these were unforeseen and demanded an explanation. Thoughts tended to converge on one common theme: someone, some force was out to get them, someone was angry with them, or disapproved somehow. Some unseen, perhaps unseeable agent.
Psychologists have done a lot of work on agency in recent years. They’ve found that we can create convincing agents for ourselves with the most basic computer-generated or pen-and-paper images. Give them some animation, have one chasing another, and we’re ready to attribute all sorts of motives and purposes. Recognising, or just suspecting, agency behind the movement of a bush, the flying through the air of a rock, or an unfamiliar sound in the distance, has been a useful mindset for our ancestors as they sought to survive against the hazards of life. ‘If in doubt, it’s an agent’ might have been humanity’s first slogan, though of course humanity didn’t come up with it, they got it from their own mammalian ancestors. My pet cat’s reaction to thunder and lightning clearly indicates her view that someone’s out to get her.
But what about the supernatural part of supernatural agency? That, too, is very basic to our nature, and it’s another feature of our thinking that has been brought to light by psychologists in recent times. I won’t go into the ingenious experiments they’ve conducted on children here – look up the work of Justin Barrett, Paul Harris and others – but they show conclusively that very young children assume that the adults around them, those towering, confident, competent and purposive figures, are omniscient, omnipotent and immortal, until experience tells them otherwise. As children we think more in terms of absolutes. Good and evil are palpably real to us, as ‘bad’ and ‘good’ are some of the first categories we ever learn from the god-like beings, our parents, who protect us and are obsessed with us (if we’re lucky in our choice of parents), and who have created us in their image.
Given all this, we might come to understand the naturalness of religion, and its near-universality. But what about the argument, which some of these psychological findings might support, that religion is a form of childishness that we should grow out of, like belief in Santa? It’a common argument among atheists, which to some degree I share, but I also feel, along with the psychologists who have shed such light on the default thinking of children, that ‘childish’ thinking is something we need to learn from rather than dismissing it with contempt. This kind of thinking is far more ingrained in us than we often like to admit, and it’ll always be more natural to us than the kind of reasoning that produces our scientific theories and technology. Creationism is easy – a supernatural agent did it – but evolution – the theory of natural selection from random variation – is much harder. The idea that we’re the special creation of a supernatural agent who’s obsessed with our welfare is far more comforting than the idea that we’re the product of purposeless selection from variation, existing by apparent chance on one insignificant planet in an insignificant galaxy amongst billions of others. In terms of appeal to our most basic needs, for protection, belonging and significance in the scheme of things, religious belief has an awful lot going for it.
So belief in a supernatural being, for whom we are special, is eminently reasonable. And yet… I don’t believe in such a being, and an increasing number of people are abandoning such a belief, especially in ‘the west’, and especially amongst the intelligentsia, which I’ll broadly define as those who make their living through their brainpower, such as scientists, academics, doctors, lawyers, teachers, journalists, writers and artists. New Scientist, in its fascinating recent issue on the Big Questions, features a graph of the world’s religious belief systems. I can’t vouch for its accuracy, but it claims 2.2 billion Christians, 1.6 billion Moslems, 900 million Hindus, and 750 million in the category ‘secular/non-religious/atheist/agnostic’. These are the top four religious categories. I find that fourth figure truly extraordinary, especially considering that it was only really recognised and counted as a category from the mid-twentieth century, or even later. In Australia, where religious belief is counted in the national census every five years, this optional question was first put in 1971. In that year the percentage of people who professed to having no religion was minuscule – about 5%. Since then, the category of the ‘nones’ has been by far the fastest growing category, and if trends continue, the non-religious will be in the majority by mid-century.
So, while I recognise that religious belief is quite reasonable, it’s clear that, in some parts of the world, a growing number find non-belief more reasonable, and I’m not even going to explore here the reasons why. You can work those out for yourself. It’s clear though that we’re entering a new era with regard to religious belief.
food irradiation and the organic food movement
Food irradiation is a well-known process for preserving food and eliminating or reducing bacteria. It’s used for much the same purpose that pressure cooking of tinned food is used, or the pasteurization of milk. All food used by NASA astronauts in space is irradiated, to reduce the possibility of food-borne illness.
advantages and disadvantages of irradiation
According to the US Department of Health’s Center for Disease Control and Prevention (CDC), irradiation, if applied correctly, has been clearly shown to reduce or eliminate food pathogens, without reducing the nutritional value of the food. It should be noted that irradiation doesn’t make food radioactive. I’ll look at the science of irradiation shortly.
Of course it’s not a cure-all. For example, it doesn’t halt the ageing process, and can make older fruit look fresher than it is. The reduction in nutritional value of the food, caused by the ageing process, can be masked by irradiation. It can also kill off bacteria that produce an odour that alerts you that the food is going off. Also, it doesn’t get rid of neurotoxins like those produced by Clostridium botulinum. Irradiation will kill off the bacteria, but not the toxins produced by the bacteria prior to irradiation.
how does food irradiation work?
Three different types of irradiation technology are used, using gamma rays (cobalt-60), electron beams and x-rays. The idea is the same with each, the use of ionising radiation to break chemical bonds in molecules within bacteria and other microbes, leading to their death or greatly inhibiting their growth. The amount of ionising radiation is carefully measured, and the radiation takes place in a special room or chamber for a specified duration.
When radioactive cobalt 60 is the energy source, it’s contained in two stainless steel tubes, one inside the other, called ‘source pencils’. They’re kept on a rack in an underground water chamber, and raised out of the water when required. The water isn’t radioactive. Food products move along a conveyor belt into a room where they’re exposed to the rack containing the source pencils. Gamma rays (photons) pass through the tubes and treat the food. The cobalt 60 process is generally used in the USA.
An Electron-beam Linear Accelerator generates, concentrates and accelerates electrons to up to 99% of light-speed.These electron beams are scanned over the product. The machine uses energy levels of 5, 7.5 or 10 MeV (million electron volts). Again the product is usually guided under the beam by a conveyor system at a predetermined speed to obtain the appropriate dosage. This will clearly vary with product type and thickness.
The X-ray process starts with an electron beam accelerator targeting electrons on a metal plate. The energy that isn’t absorbed is converted into x-rays, which, like gamma rays, can penetrate food containers more than 40cms thick. Shipping containers, for example.
Most of the radiation used in these processes passes through the food without being absorbed. It’s the absorbed radiation, of course, that has the effect, destroying microbes and so extending shelf life, and slowing down the ripening of fruits and vegetables. The potential is there for food irradiation to replace chemical fumigants and fungicides used after harvest. It also has the potential, through the use of higher doses, to kill contaminating bacteria in meat, such as Salmonella.
Food irradiation is a cold treatment. It doesn’t significantly raise the temperature of the food, and this minimises nutrient loss or changes in texture, colour and flavour. The energy it uses is too low to cause food to become radioactive. It has been compared to light passing through a window. Food irradiation uses the same principle as pasteurization, and can be described as pasteurization by energy instead of heat, or cold pasteurization..
the use of food irradiation in Australia
Due largely to fears about irradiation having to do with radioactivity and nuclear energy, the process isn’t used as widely in Australia (or indeed the USA) as it could be. Irradiation is used in some 50 countries, but the level of usage varies for each country, from very limited in Austria and other EU countries, to a very widespread usage in Brazil. Food Standards Australia New Zealand (FSANZ) summarises our situation thus:
In Australia and New Zealand, only herbs and spices, herbal infusions, tomatoes, capsicums and some tropical fruits can be irradiated.
FSANZ has established that there is a technological need to irradiate these foods, and that there are no safety concerns or significant loss of nutrients when irradiating these foods.
Irradiated food or ingredients must be labelled clearly as having been treated by ionising radiation.
food irradiation, health and safety
Since 1950 hundreds of studies have been carried out on animals fed with irradiated products, including multi-generational studies. On the basis of these studies, food irradiation has been approved by the World Health Organization, the American Dietetic Association, the Scientific Committee of the European Union and many other national and international monitoring bodies. Of course this hasn’t stopped many individuals and organisations from complaining and campaigning against the practice. Concerns include: chemical changes harmful to the consumer; impairment of flavour; the destruction of more ‘good’ than ‘bad’ bacteria; and that it’s an unnecessary process which runs counter to the movement towards regional product, seasonality and real freshness. I’ve already mentioned other problems, such as that it can mask spoiled food, and that it doesn’t destroy toxins already released by bacteria.
opposition from the organic food movement
Food products must be irradiation-free if they are to certified as ‘organic’, in Australia and elsewhere. Now, I’ve fairly regularly expressed irritation with the ‘organic’ food ideology, most particularly in this post, but I recognise that it appeals to a very diverse set of people, with perhaps a majority simply believing, on faith, that ‘organic’ food will be more nutritious, safer and better for the environment than conventional food. Most of those people wouldn’t know much about food irradiation, but hey, it sounds dodgy, so why not avoid it? I’ve no great argument to make with such people, apart from the old ‘knowledge is power’ arguments, but there are a few individuals and organisations trying to get food irradiation banned, based on what they claim to be evidence. Unsurprisingly, most of these critics are also ‘organic’ food proponents. I’ll look at some criticisms from Eden Organic Foods, a US outfit, which admittedly represents the extreme end of the spectrum (nature before the fall?).
Firstly, in their ‘factsheet’ on irradiation, linked to above (and reprinted verbatim here by another alarmist organisation, the Center for Food Safety), they waste no time in informing us that the beams used are ‘millions of times more powerful than standard medical x-rays’. This sounds pretty scary, but it’s a bogus comparison. Irradiation is designed to kill bugs and bacteria, whereas medical x-rays are for making visible what is invisible to the naked eye. Clearly, the first and foremost concern in testing and studying the technology is to make sure that the chemical changes it induces are safe for humans. Comparisons with medical x-rays are more than irrelevant to this concern, as the author of this factsheet well knows.
Next comes this disturbing claim:
Radiation can do strange things to food, by creating substances called “unique radiolytic products.” These irradiation byproducts include a variety of mutagens – substances that can cause gene mutations, polyploidy (an abnormal condition in which cells contain more than two sets of chromosomes), chromosome aberrations (often associated with cancerous cells), and dominant lethal mutations (a change in a cell that prevents it from reproducing) in human cells. Making matters worse, many mutagens are also carcinogens
Wow. So much for the poor people of Brazil – they’re obviously done for. But how is it that the world’s top scientific agencies missed all these mutagens and carcinogens? Let’s take a closer look.
The term ‘radiolytic products’ simply means the products created by chemical changes that occur when food is irradiated. Similarly, the products created by heat treatment, or simply cooking, might be called ‘thermolytic products’. These are not ‘strange’, they’re quite predictable, for irradiation would be totally ineffective if it didn’t bring about some chemical changes. One of the differences is that radiolytic products are generally undetectable and produce only minor changes in the food compared to the major operation we call cooking. It is, of course, precisely these products that the scientific community scrutinises when determining the safety of irradiated foods.
Interestingly, in an article, dating back to 1999, called ‘Scientific answers to irradiation bugaboos’, for 21st Century Science & Technology magazine, Marjorie Mazel Hecht has this to say:
The July 1986 report of the Council for Agricultural Science and Technology (CAST), which reviewed all the research work on food irradiation, defined unique radiolytic products “as compounds that are formed by treating foods with ionizing energy, but are not found normally in any untreated foods and are not formed by other accepted methods of food processing.”
The report states that “on the basis of this definition no unique radiolytic compounds have been found in 30 years of research. Compounds produced in specific foods by ionizing energy have always been found in the same foods when processed by other accepted methods or in other foods” (Vol. 1, p. 15).
This slightly contradicts the factsheet put out by Idaho University’s Radiation Information Network, which acknowledges the existence of such products while insisting on their nugatory nature:
Scientists find the changes in food created by irradiation minor to those created by cooking. The products created by cooking are so significant that consumers can smell and taste them, whereas only a chemist with extremely sensitive lab equipment may be able to detect radiolytic products.
Needless to say, alarmists thrive on these contradictions. So what evidence is there of mutagenic irradiation byproducts? Well, there are radiolytic byproducts of fatty acids in meat, called alkylcyclobutanones (2-ACBs), first detected a few decades ago, and the research done on them seems to be so far inconclusive. A book entitled Food Irradiation Research and Technology, the second edition of which was published last year, states that ‘knowledge about the toxicological properties of 2-ACBs is still scarce’, and that ‘it may be prudent to collect more knowledge on the toxicological and metabolic properties of 2-ACBs in order to quantify a possible risk – albeit minimal.’ The book describes a number of studies on rats and humans, going into more detail than I can comprehend, but the results have been difficult to interpret and generally not easily replicable in other studies, indicating very minute and hard-to-measure effects. No doubt such studies will be ongoing. As far as I know, 2-ACBs are the only products about which there is any concern.
What is obvious though, in looking at the research material available online, is the difference between the caution, skepticism and uncertainty of researchers compared to the adamantine certainty of such critics as the Center for Food Safety.
But what about polyploidy? Polyploid cells contain more than two paired sets of chromosomes. Eukaryotic cells, those of multicellular creatures, are diploid (two sets), and prokaryotic, bacterial cells are haploid (one set). Polyploidy is regarded as a chromosomal aberration, common in many plants and some invertebrates, but relatively rare in humans. However it is present in humans, and the percentage varies from individual to individual, and within individuals from day to day and week to week, depending on a range of factors including diet, age, and even circadian rhythms. Levels of up to 3-4% in human lymphocytes have been found in healthy individuals, though some researchers have claimed much higher percentages, in liver cells. The overall finding so far is that fluctuations in polyploidy are the norm, and no clear correlation has been found so far between these fluctuations and health profiles. It seems that the biological significance of polyploidy isn’t known.
Critics of irradiation have been going on about polyploidy and other mutations supposedly caused by irradiation for decades, and unsurprisingly, some are fanatically obsessed with the issue, accompanying their rants with long reference lists, mostly from like-minded activists. However, the text Safety of irradiated foods, 2nd edition discusses polyploidy in some detail, with particular reference to a study of malnourished Indian children fed irradiated wheat, a study regularly cited by anti-irradiation activists. It turns out that there were many problems with the study. First, not enough cells were counted to validly pinpoint an effect, such as a change in diet. Secondly, polyploidy is notoriously difficult to detect – superimposed diploid cells can be easily mistaken for polyploid cells under a microscope (in fact when two independent observers looked at the same microscope slides, one found 34 polyploid cells, the other found 9). Further, the study only gave group results rather than individual results, so it wasn’t possible to know whether the polyploidy was restricted to one or two individuals rather than spread over the group. Another problem was that the reference or control group was found to have no polyploidy at all, a very strange finding given that other researchers always found some degree of polyploidy in their subjects, regardless of irradiation or other effects. In fact, the study was so poorly written up that it’s impossible to replicate – for example the exact diet given the children wasn’t described. How was the wheat fed to the children?. Presumably it was prepared in some way, but how? The omission is crucial. The study also didn’t take into account the effect of malnutrition itself on chromosomal abnormalities. And so on.
You get the picture, and it’s the same with other claims about mutations and carcinogens. Every time you look into the claims you find the same problems that no doubt other scientific watchdog organisations have found – poorly conducted studies that either can’t be replicated or haven’t survived replication. That, of course is no reason for complacency, and at least the activists can assist, in their sometimes muddle-headed ways, in improving our knowledge of 2-ACBs, polyploidy and other biological effects, just as the creationists who bang on about a lack of transitional forms, or ‘irreducible complexity’, help us to focus on refutations, clarifications and further evidence.
Finally, food irradiation, while clearly not the zappo-horrorshow that activists are determined to make it, doesn’t replace proper handling techniques and a good instinct about food quality. The fact is, though, that it does increase shelf life, and is a useful tool in our increasingly global economy, where food is shipped from here to there and everywhere, in season and out. If you prefer to eat locally, with fresh and seasonal produce, fine, and we can argue about the sustainability of that approach on a worldwide scale, but let’s none of us pretend that food irradiation is other than what it is. Let the evidence, properly evaluated, be your guide.
on cowardice, courage and the abuse of language
In this post I want to try to avoid politics, and to focus on the English language, its use and abuse. If you google the word ‘coward’, followed by the word ‘meaning’ (I often ask my NESB students to do this with words they don’t know), you’ll come up first with this definition: a person who is contemptibly lacking in the courage to do or endure dangerous or unpleasant things. Second comes this: [a person who is] excessively afraid of danger or pain.
These are, to me, bog-standard, uncontroversial definitions of the word ‘coward’. To be a coward is to be nothing more and nothing less than what these definitions describe.
So, as a person who cares about language, it disturbs and aggravates me that the word ‘coward’ is now regularly used by the media and by commentators of all kinds, from world leaders to pub philosophers, to refer to suicide bombers, mass shooters, Wikileakers and terrorists of every description. I would ask you to pause for a moment, and think of these categories of people, and the people themselves, if you can bear it. Think of, say, Thenmozhi Rajaratnam, a member of the Tamil Tigers and the suicide killer of Rajiv Ghandi and 14 others beside herself in 1991. Or Reem Riyashi, the wealthy Palestinian mother of two and Hamas operative who killed herself and 4 Israelis at the Erez Crossing in Gaza in 2004. Think of Martin Bryant, the murderer of 35 people at Port Arthur in Tasmania in 1996, or Anders Behring Breivik, killer of 77 people by bombing and gunfire in Norway in 2011. Think again of Bradley (now Chelsea) Manning, who leaked large quantities of classified US information to Wikileaks in early 2010, or Edward Snowden, recent leaker of classified documents from the USA’s National Security Agency to various media outlets. Now think finally of Mohamed Atta, a principal player in the September 11 2001 attacks in the USA, and pilot of the plane that crashed into the North Tower of the world trade centre, or Noordin Top, mastermind of several fatal bombings in Indonesia, and indefatigable recruiter and indoctrinator for the Jihadist organisation Jemaah Islamiya.
No doubt these characters will awaken many diverse thoughts, but it’s unlikely that cowardice would be part of your description of any of them, especially after having been primed with the definitions of a coward at the top of this post. It seems like describing any of these various characters as cowards would simply be what philosophers call a ‘category mistake’, so far from cowardly, in the bog-standard sense of that term, have been the actions that have made them notorious.
So what is going on here? Intellectual laziness? Overblown rhetoric? Well, yes and no. To dismiss this rhetoric of cowardice as just plain ignorant or lazy would be to miss the point of it, for there is method in this apparent madness, intended or not. The real point of describing any or all of these people as cowards is to remove them as far as possible from any association with another word, more or less directly opposed to cowardice: courage.
Courage is seen as positive of course. It’s seen as a virtue, yet when we delve further into it, as Socrates and his interlocutors did in the Laches, we find it be a more slippery concept than at first glance. It rather sticks in our craw, to say the least, to claim that Mohamad Atta was courageous in carrying out his mission to fly an unfamiliar Boeing 767 into the World Trade Centre, or that Thenmozhi Rajaratnam showed amazing courage in blowing herself up with Rajiv Ghandi and many others, or that Anders Brehvik displayed steely resolve and courage in carrying out his long-planned slaughter of scores of innocent children. The actions of Manning and Snowden have naturally received more mixed responses, with some feeling that the term ‘courageous’ is singularly apt in describing them, while others would baulk at the term.
So let’s perform the same operation on ‘courage’ as we did on the word ‘coward’. Here’s the very straightforward result:
Courage: the ability to do something that frightens one.
Now, it’s worth noting that this bog-standard definition, as with that of ‘coward’, has nothing whatever to say about the moral implications of the action or actions that the brave person engages in and the coward avoids. That action might be the slaughter, or the rescue, of thousands. This is key: the moral implications or the consequences of the actions are irrelevant to the definition. For some, it seems, this point is hard, if not impossible, to swallow. That’s the problem; because of the negative load that the term ‘coward’ carries, some people are determined to describe any action that they consider has negative consequences as cowardly. But to try to extend the meaning of the term from the bog-standard, more limited definition quoted at the top means moving away from consensus into a field of contestation that enormously diminishes the coherence and so the usefulness of the term.
I was prompted to write this piece because a recent editorial in a major Australian newspaper, attacking Edward Snowden as a coward, was brought to my attention. It was the last straw, you might say. I admit I haven’t read the editorial, and I can’t recall the newspaper, but really, you don’t need to read the detail – and I may well be convinced by the newspaper editor’s views of the implications of Snowden’s actions – to know that the application of the term ‘coward’ to Snowden’s leaking of classified information is just wrong, by the definition of terms.
This sort of thing should matter to those who respect language and its value as an effective communicative tool. By the bog-standard consensus definitions given above, we need to admit that the actions of Atta and Rajaratnam, for example, were courageous. As people in full possession of their faculties, as I assume they were, they would have had to overcome enormous fear and anxiety to perform their suicidal actions. Of course we can and should condemn their actions on a whole host of ethical grounds, but to call them cowardly doesn’t add anything to the ethical debate, it just muddies definitions (while allowing us to let off steam, to vent our indignation and disgust). It’s just name-calling.
Some thoughts on morality and its origins
I remember, quite a few Christmases ago now, a slightly acrimonious discussion breaking out about religion and morality. I simply observed – it wasn’t my family. It never is.
A born-again religious woman asked her sister – ‘where do you get your morality from if not from religion?’ She responded tartly, ‘From my mum’. This response pleased one of those present, at least! But as to the implicit claim that we get our morality from religion, my silent response was ‘how does that happen?’
Religion, at least in its monotheistic versions, implies a supernatural being, from whom all morality flows. But if you ask believers whether their cherished supernatural entity talks to them and advises them regularly about the moral decisions they face in their daily lives, you would get, well, a variety of responses, from ‘yes, he does actually’, to something like ‘you miss the point completely’. The second response might lead on to – well, theology. We were given free will, the deity’s ways are mysterious but Good, he communicates with us indirectly, you need to read the signs etc etc. But you’ll be relieved I hope to hear that this won’t be an essay on religion, which you should realise by now I find interminably boring when it tries to connect itself with morality – which is most of the time.
I’m more interested here in trying, inter alia, to define human morality, to determine whether it’s objective, or universal, and if those two terms are synonymous. And as I generally do, I’ll start with a rough and ready, semi-ignorant or uninformed definition, and then try to smarten it up – possibly overturning the original definition in the process.
So, roughly, I consider human morality to be an emergent property of our socially wired brains, something which is, therefore, evolving. I don’t consider it to be objective, because that suggests something outside ourselves, like objective reality. We can talk about it being ‘universal’, as in ‘universal human rights’, which may be agreed upon by consensus, but that’s a convenient fiction, as there’s no true consensus, as, for example, the Cairo Declaration (on human rights in Islam) reveals. Not that we shouldn’t strive for consensus, based on our current understanding of human interests and human thriving. I’m a strong believer in human rights. I suppose what I’m saying here is that my ‘universality’, far from being a metaphysical construction, is a pragmatic term about what we can generally agree on as being what we need in terms of basic liberties, and limitations to those liberties, in order to best thrive, as a thoroughly social species (deeply connected with other species).
So with this rough and ready definition, I want to look at some controversial contributions to the debate, and to add my reflections on them. I read The Moral Landscape, by Sam Harris, a while back, and found it generally agreeable, and was surprised at the apparent backlash against it, though I didn’t try to follow the controversy. However, when philosophers like Patricia Churchland and Simon Blackburn get up and respectfully disagree, finding Harris ‘naive’ and misguided and so forth, I feel it’s probably long overdue for me to get my own views clear.
The difficulty that many see with Harris’s view is encapsulated in the subtitle of his book, ‘How science can determine human values’. I recognised that this claim was asking for trouble, being ‘scientistic’ and all, but I felt sympathetic in that it seemed to me that our increasing knowledge of the world has deeply informed our values. We don’t call Australian Aboriginals or Tierra del Fuegans or Native Americans savages anymore, and we don’t describe women as infantile or prone to hysteria, or homosexuals as insane or unnatural, or children as spoilt by the sparing of the rod, because our knowledge of the human species has greatly advanced, to the point where we feel embarrassed by quite recent history in terms of its ethics. But there’s a big difference between science informing human values, and enriching them, and science being the determinant of human values. Or is there?
What Harris is saying is, forget consensus, forget agreements, morality is about facts, arrived at by reason. He brings this up early on in The Moral Landscape:
… truth has nothing, in principle, to do with consensus: one person can be right, and everyone else can be wrong. Consensus is a guide to discovering what is going on in the world, but that is all that it is. Its presence or absence in no way constrains what may or may not be true.
Clearly one of Harris’s targets, in taking such an uncompromising stance on morality being about truth or facts rather than values, is moral relativism, which he regularly attacks. Yet the most cogent critics of his views aren’t moral relativists, they’re people, like Blackburn, who question whether the moral realm can ever be seen as a branch of science, however broadly defined (and Harris defines it very broadly for his purposes). One of the points of dispute – but there are many others – is the claim that you can’t derive values from facts. For example, no amount of information about genetic variation within human groups can actually determine what you ought to do in terms of discrimination based on perceived racial differences. Such information can and should inform decisions, but they can’t determine them, because they are facts, while values – what you should do with those facts – are categorically different.
It seems to me that Harris often chooses clear-cut issues to highlight morality-as-fact, such as that a secure, healthful, well-educated life is better than one in which you get beaten up on a daily basis. Presumably he imagines that all the gradations in between can be measured precisely as to their truth-value in contributing to well-being. But surely it’s in these difficult areas that questions of value seem to be most ‘subjective’. Can we make an objective moral claim, say, about vegetarianism, true for all people everywhere? What about veganism? I very much doubt it. Yet we also need to look skeptically at those values he sees as clear-cut. Take this example from The Moral Landscape:
In his wonderful book The Blank Slate, Steven Pinker includes a quotation from the anthropologist Donald Symons that captures the problem of multiculturalism very well:
If only one person in the world held down a terrified, struggling screaming little girl, cut off her genitals with a septic blade, and sewed her back up, leaving only a tiny hole for urine and menstrual flow, the only question would be how severely that person should be punished, and whether the death penalty would be a sufficiently severe sanction. But when millions of people do this, instead of the enormity being magnified millions-fold, suddenly it becomes ”culture”, and thereby magically becomes less, rather than more, horrible, and is even defended by some Western “moral thinkers”, including feminists.
Now, as a card-carrying humanist, and someone generally quite comfortable with the values that, over time, have emerged in my part of the western world, namely Australia, I’m implacably opposed to the practice described here by Symons. But even so, I see a number of problems with this description. And ‘description’ is an important term to think about here, because the way we describe things is an essential indicator of our understanding of the world. The description here is of a ‘procedure’, and it is brief and clinical, leaving aside the depiction of the ‘terrified struggling screaming little girl’. It isn’t a description likely to have much resonance for those who subject their daughters and nieces to this practice. After all, this is a traditional cultural practice, however horrific. It is still practiced regularly in many African countries, and in proximate countries such as Yemen. Clearly the practice aligns with rigid attitudes about the role and place of women in those cultures, attitudes that go back a long way – the first reference to female circumcision, on an Egyptian sarcophagus, dates back almost 4000 years, but it’s likely that it goes back a lot further than that. As Wikipedia puts it, ‘Practitioners see the circumcision rituals as joyful occasions that reinforce community values and ethnic boundaries, and the procedure as an essential element in raising a girl.’
Now, Symons (and presumably Pinker, and Harris) take the view that this is clearly a criminal practice, and that culture should not be used as an excuse. It’s a view backed up by most of the nations in which it occurs, who have instituted laws against it, and in 2012 the UN General Assembly unanimously voted to take all necessary steps to end it, but these national and international good intentions face a long, uphill battle. However, if you look at some of the first descriptions of this practice, by outsiders such as Strabo or Philo of Alexandria, both writing in the time of Christ, you won’t find any censoriousness, nor would you expect to. It was well accepted in the Graeco-Roman world that customs varied widely, and that many foreign customs were weird, wild and wonderful. It’s likely that observers from the dominant culture felt morally superior, as is always the case, but there was no attempt to suppress other cultural practices – any more than there was only 200 hundred years ago, in Australia, with respect to the native inhabitants. The ‘mother country’ sent out clear and regular messages at the time about treating the natives with respect, and non-interference with their cultural practices (though it would no doubt have considered them barbaric and savage as a matter of course). It’s really only in recent times that, as a result of our growing confidence in a universal approach to morality or ‘well-being’, we (the dominant culture) have spoken out against what we now unabashedly call female genital mutilation, as well as other practices such as purdah and witch-hunting.
From all this, you might guess that I’m ambivalent about Harris’s confident approach to moral value. Well, yes and no, he said ambivalently. I can’t tell you how mightily glad I am that I live in a part of the world in which purdah and infibulation aren’t prevalent. However, I can’t step outside of my space and time, and I don’t know what it would be like to live in a world where these practices were standard. And living in such a world doesn’t mean being being transported to it ‘suddenly’, it means being steeped in its values. After all, my own Anglo-Australian culture was one that, less than 200 years ago, transported homeless boys, in danger of ‘going to the bad’, to Australia where they often ended up being worked to death on chain gangs, and this was considered perfectly normal. I would have considered it perfectly normal, for I’m not so arrogant as to imagine I could transcend the moral values of my culture as it was in the 1830s.
So, to return to the passage from The Moral Landscape quoted above. It isn’t a factual passage, it’s a description, with interpretive and speculative features. It describes, first the actions of ‘one person’, engaged in what seems to us an insane surgical procedure, then we’re asked to multiply this act by millions, and ‘suddenly’ consider it culture. But this strikes me as a deliberately manipulative putting of the cart before the horse. The real motive seems to be to ask us to dismiss culture altogether. After all, any human product that can be called into being ‘suddenly’, and which ‘magically’ blights our moral understanding of the world cannot surely be taken seriously. Harris, as I recall, used similar arguments against religion, perhaps in The End of Faith (which I haven’t read), but certainly in some of his talks on the subject. A practice or belief which we might lock someone up for, ‘suddenly’ becomes acceptable when engaged in by millions and called ‘religion’.
This strikes me as a glib and naive argument, which could only appeal to historically uninformed (or indifferent) ‘rationalists’. Cultural and religious beliefs and practices, weird, wild, wonderful and occasionally horrifying though they might be, are far too widespread, and too deeply woven into the identity of individuals and social groups, to be set aside in this way.
This is a very very complex issue, one that, dare I say, middle-class intellectuals like Harris and Pinker tend to skate over, even with a degree of contempt. For myself, I deal with these cultural issues with a mixture of fear – ‘don’t provoke the culturally wounded, they’ll just get angry and dangerous’ – and concern – ‘if you take away these people’s cultural/religious identity, how will they cope?’. Perhaps I’m being arrogant about the power of western secular values, but it seems to me that much of the world’s turmoil comes from resentment at old cultural and religious certainties being undermined.
So I believe in cultural sensitivity, for strategic purposes but also because we are all culturally embedded, no matter how scientifically enlightened we claim to be. However, I don’t think all cultures are, or all culture is, equally valuable or equally healthy. How I measure that, though, is a big question since I can’t step outside of my own culture. Perhaps therein lies the difficulty about getting all ‘scientific’ about morality. Science itself is hardly culture-free – a dangerous point to make in some circles.
So I don’t think I’ve gotten much further as to where morality comes from. To say that it comes from culture requires a thorough definition and understanding of that concept, otherwise we’re just deferring any real explanation, but clearly that is the way to go. But I prefer to look at this connection with culture, and with other more fundamental aspects of our social nature, from a humanist perspective. Western secular humanism tends to wear its culture lightly, and to value skepticism, reflection and analysis as – possibly cultural – tools for dismantling or at least loosening the overly heavy and oppressive armour that cultural beliefs and practices can become.
what does curiosity actually mean?
You might say that Philip Ball has performed a curious task with his book, Curiosity. He’s taken this term, which we moderns might take for granted, and examined what intellectuals and the public have made of it down through the ages – with a particular focus on that wobbly symbol of the seventeenth century British scientific enlightenment, the Royal Society. I’ve been spending a bit of time in the seventeenth century lately, what with Dava Sobel’s book on the struggle to measure longitude, Matthew Cobb’s book on the untangling of the problem of eggs and sperm and conception, not to mention Bill Bryson’s lively treatment of Hooke, Leeuwenhoek and cells and protozoa in A Short History of Nearly Everything.
That century, with some of its most interesting actors, including Francis Bacon, René Descartes, William Harvey, Jan Swammerdam, Nicolas Steno, Johann Komensky (aka Comenius), Samuel Butler, Thomas Hobbes, Robert Hooke, Robert Boyle, Antonie van Leeuwenhoek, Thomas Shadwell, Margaret Cavendish and Isaac Newton, represented a great testing period for science and its reception by the public. Curiosity has always had its enemies, and still does, as evidenced by some Papal pronouncements of recent years, but in earlier, more universally religious times, knowledge and its pursuit were treated with great wariness and suspicion, a suspicion sanctioned by the Biblical tale of the fall. The Catholic Church had risen to a position of great power in the west, though the revolting Lutherans, Anglicans, Calvinists and their ilk had spoiled the party somewhat, and England in particular, having grown in pride and prosperity during the Elizabethan period, was flexing its muscles and exercising its grey matter in exciting new ways. The sense of renovation was captured by the versatile Bacon, with works like the Novum Organum (New Method), The New Atlantis and The Advancement of Learning.
In the past I’ve described curiosity and scepticism as the twin pillars of the scientific mindset, but they’re really more like a pair of essential forces that interact and modify each other. Scepticism without curiosity is just pure negativity and nihilism, curiosity without scepticism is directionless and naive.
But perhaps that’s overly glib. What, if any, are the limits of curiosity, and when is it a bad thing? It killed the cat, after all.
The word derives from the Latin ‘cura’, meaning care. Think of the word ‘curator’. However, if you think of one of the most curious works of the ancients, Pliny the Elder’s Natural History, you’d have to say, from a modern perspective, that little care was taken to separate truth from fiction in his massive and sometimes bizarre collection of curios. This sort of unfiltered inclusivity in collecting ‘facts’ and stories goes back at least to Herodotus, the ‘father of lies’ as well as of history, and it goes forward to medieval bestiaries and herbaria. These collections of the weird and wonderful were, of course, not intended to be scientific in the modern sense. The term ‘science’ wasn’t in currency and no clear scientific methodologies had been elaborated. As to curiosity, it certainly wasn’t a fixed term, and after the political establishment of Christianity, it was more often than not seen in a negative light. ‘We want no curious disputation after possessing [i.e. accepting the truth of] Jesus Christ’, wrote Tertullian in the early Christian days. Another early Christian, Lactantius [c240-c320], explained that the reason Adam and Eve were created last was so that they’d remain forever ignorant of how their god created everything else. That was how it was intended to be. Modern creationists follow this tradition – God did it, we don’t know how and we don’t really care.
Fast forward to Francis Bacon, who still, in the early 17th century, had to contend with the view of curiosity as a sinful extravagance, a view that had dominated Europe for almost a millennium and a half. Bacon had quite a pragmatic, almost business-like view of curiosity as a tool to benefit humanity. The ‘cabinet of curiosities’ was becoming well established in his time, and Bacon advised all monarchs, indeed all rich and powerful men, to maintain one, well sorted and labelled, as if to do so would be magically empowering. The problem with these cabinets, though, was that there was little understanding about the relations between entities and articles. That’s to say, there was little that was modernly scientific about them. Their objects were largely unrelated rarities and oddities, having only one thing in common, that they were ‘curious’. Bacon recognised that this wouldn’t quite do, and tried to point a way forward. He didn’t entirely succeed, but – small steps.
Ball’s book is at pains to correct, or at least provide nuance to, the standard view of Bacon as initiator of and father-figure to the British scientific enlightenment. In fact, Bacon may have been a Rosicrucian, and his utopian New Atlantis describes a more or less priestly caste of technical experts, living and working in Solomon’s House, and keeping their arts and knowledge largely under wraps, like the alchemists and mages of earlier generations. Bacon, with his government connections and his obvious ambition to be benefited by as well as benefiting the state, was concerned to harness knowledge to productivity and profit, and those who see science largely as a coercion of nature have cursed him for it ever since. Mining and metallurgy, engineering and manufacturing were his first subjects, but he also imagined great changes in agriculture – the breeding of plants, fruits and flowers, as well as animals, to create ‘super-organisms’, in and out of season, for our benefit and delight. The art and science of the kitchens of Solomon’s House produces superior dishes, as well as wines and other beverages, and printing and textiles have advanced greatly, with new fabrics, papers, dyes and machinery. Even the weather is subject to manipulation, with rain, snow and sunshine under the control of the savants. The details of all these advancements are kept vague of course, (and here’s where Bacon’s insistence on ‘secret knowledge’ plays to his advantage, a point not sufficiently noted by Ball in his need to connect Bacon with the the alchemist-magicians of the past) but what is represented here is promise, a faith in human ingenuity to improve on the products of the natural world.
In focusing on all these benefits, Bacon manages largely to sidestep the religious aversion to curiosity as a form of intellectual avarice. However, Bacon and his more curious compatriots were never too far from the magical dark arts. Few intellectuals of this period, for example, would have dismissed alchemy out of hand, in spite of Chaucer’s delicious mockery of it over 200 years before, or Ben Jonson’s more contemporaneous take in The Alchemist. What differentiated Bacon was an interest in system, however vaguely adumbrated, and a harnessing of this system to the interests of the state.
Bacon tried to interest James I in a state sponsored proto-scientific institution, but this got nowhere, largely because he couldn’t devise anything like a practical program for such an entity, but a generation or two after his death, after a civil war, a brief republic and a restoration, the Royal Society was formed under the more or less indifferent patronage of Charles II. Bacon was seen as its guiding spirit, and there was an expectation, or hope, that its members would be virtuosi, a term then in currency. As Ball explains:
The virtuoso was ‘a rational artist in all things’… meaning the arts as well as the sciences, pursued methodically with a scientist’s understanding of perspective, anatomy and so forth. (It is after all in the arts that the epithet ‘virtuoso’ survives today.) The virtuoso was permitted, indeed expected, to indulge pure curiosity: to pry into any aspect of nature or art, no matter how trivial, for the sake of knowing. There was no sense that this impulse need be harnessed and disciplined by anything resembling a systematic program, or by an attempt to generalise from particulars to overarching theories.
Charles II, in spite of having some scientific pretensions, paid scant attention to his own Society, and neglected to fund it. What was perhaps worse for the Society was his amused approval of a hit play of the time, Thomas Shadwell’s The Virtuoso, which satirized the Society through its central character, Sir Nicholas Gimcrack. The play, as well as many criticisms of the Society’s practices by the likes of the philosopher Thomas Hobbes and the aristocratic Margaret Cavendish (Duchess of Newcastle-upon-Tyne), presented another kind of negativity vis-a-vis unbridled curiosity, more modern, if not more pointed than the old religious objections.
The play-goer first encounters Sir Nicholas Gimcrack lying on a table making swimming motions. He tells his visitors that he’s learning to swim, but they are dubious about his method. His response:
I content myself with the speculative part of swimming; I care not for the practick. I seldom bring anything to use; tis not my way. Knowledge is my ultimate end.
This was the updated criticism. Pointless observations and experiments, leading nowhere and of no practical use. Gimcrack appears to have been based on Robert Hooke, one of the Royal Society’s most brilliant members, who was suitably enraged on viewing the play. Shadwell mocked Hooke’s prized invention, the air pump, intended to create a vacuum for the purpose of observing objects inserted into it, and he presented a jaundiced view of Gimcrack, through the dialogue of his niece, as ‘a sot that has spent two thousand pounds in microscopes to find out the nature of eels in vinegar, mites in a cheese, and the blue of plums.’ These were all examined in Hooke’s ground-breaking and breath-taking work Micrographia.
Most of Shadwell’s mockery hasn’t stood the test of time, but he was far from the only one who targeted the practices and the approach of the Society and of ‘virtuosi’, sometimes with humour, sometimes with indignation. Their criticisms are worth examining, both for what they reveal of the era, and for their occasional relevance today. Many of them seem totally misplaced – mocking the ‘weighing of air’, which they naturally saw as the weighing of nothing, or the examining, through the newish tool the microscope, of a gnat’s leg. It should be recalled that Hooke, through his microscopic investigations, was the first to highlight and to name the individual cell. Yet it was a common criticism of the era, due largely to the ignorance of the interconnectedness of all things that the scientifically literate now take for granted, that these explorations were simply time-wasting dilettantism. The philosophical curmudgeon Thomas Hobbes, for example, firmly believed that experiments couldn’t produce significant truths about the world. It seems that the general public, who didn’t have access to such things, saw microscopes and telescopes as magical devices which didn’t so much reveal new worlds as to create them. If they couldn’t be verified with one’s own eyes, how could these visions be trusted? And there was the old religious argument that we weren’t meant to see them, that we should keep to our god-given limitations.
Generally speaking, as Ball describes it, though the criticisms and misgivings weren’t so clearly religious as they had been, they centred on a suspicion about unrestrained curiosity and questioning, which might lead to an undermining of the social order (a big issue after the recent upheavals in England), and to atheism (they were on the money with that one). They had a big impact on the Royal Society, which struggled to survive in the late seventeenth and early eighteenth centuries. It’s worth noting too, that the later eighteenth century Enlightenment on the continent was much more political and social than scientific.
But rather than try to analyse these criticisms, I’ll provide a rich sample of them, without comment. None of them are ‘representative’, but together they give a flavour of the times, or of the more conservative feeling of the time.
[Is there] anything more Absurd and Impertinent than a Man who has so great a concern upon his Hands as the Preparing for Eternity, all busy and taken up with Quadrants, and Telescopes, Furnaces, Syphons and Air-pumps?
John Norris, Reflections on the conduct of human life, 1690
Through worlds unnumber’d though the God be known,
‘Tis ours to trace him only in our own….
The bliss of man (could pride that blessing find)
Is not to act or think beyond mankind;
No powers of body or of soul to share,
But what his nature and his state can bear.
Why has not a man a microscopic eye?
For this plain reason, man is not a fly.
Say what the use, were finer optics giv’n,
T’inspect a mite, not comprehend the heav’n? …
Then say not man’s imperfect, Heav’n in fault;
Say rather, man’s as perfect as he ought:
His knowledge measur’d to his state and place,
His time a moment, and a point his space.
Alexander Pope, An Essay on Man
There are some men whose heads are so oddly turned this way, that though they are utter strangers to the common occurrences of life, they are able to discover the sex of a cockle, or describe the generation of a mite, in all its circumstances. They are so little versed in the world, that they scarce know a horse from an ox; but at the same time will tell you, with a great deal of gravity, that a flea is a rhinoceros, and a snail an hermaphrodite.
… the mind of man… is capable of much higher contemplations [and] should not be altogether fixed upon such mean and disproportionate objects.
Joseph Addison, The Tatler, 1710
But could Experimental Philosophers find out more beneficial Arts then our Fore-fathers have done, either for the better increase of Vegetables and brute Animals to nourish our bodies, or better and commodious contrivances in the Art of Architecture to build us houses… it would not onely be worth their labour, but of as much praise as could be given to them: But as Boys that play with watry Bubbles, or fling Dust into each others Eyes, or make a Hobby-horse of Snow, are worthy of reproof rather then praise, for wasting their time with useless sports; so those that addict themselves to unprofitable Arts, spend more time then they reap benefit thereby… they will never be able to spin Silk, Thred, or Wool, &c. from loose Atomes; neither will Weavers weave a Web of Light from the Sun’s Rays, nor an Architect build an House of the bubbles of Water and Air… and if a Painter should draw a Lowse as big as a Crab, and of that shape as the Microscope presents, can any body imagine that a Beggar would believe it to be true? but if he did, what advantage would it be to the Beggar? for it doth neither instruct him how to avoid breeding them, or how to catch them, or to hinder them from biting.
[Inventors of telescopes etc] have done the world more injury than benefit; for this art has intoxicated so many men’s brains, and wholly employed their thoughts and bodily actions about phenomena, or the exterior figures of objects, as all better arts and studies are laid aside.
Margaret Cavendish, Observations upon Experimental Philosophy, 1666
[A virtuoso is one who] has abandoned the society of men for that of Insects, Worms, Grubbs, Maggots, Flies, Moths, Locusts, Beetles, Spiders, Grasshoppers, Snails, Lizards and Tortoises….
To what purpose is it, that these Gentlemen ransack all Parts both of Earth and Sea to procure these Triffles?… I know that the desire of knowledge, and the discovery of things yet unknown is the pretence; but what Knowledge is it? What Discoveries do we owe to their Labours? It is only the discovery of some few unheeded Varieties of Plants, Shells, or Insects, unheeded only because useless; and the knowledge, they boast so much of, is no more than a Register of their Names and Marks of Distinction only.
Mary Astell, The character of a virtuoso, 1696
There are many other such comments, very various, some attempting to be witty, others indignant or contemptuous, and some quite astute – the Royal Society did have more than its share of dabblers and dilettantes, and was far from being simply ‘open to talents’ – but for the most parts the criticisms haven’t dated well. You won’t see The Virtuoso in your local playhouse in the near future. Wide-ranging curiosity, mixed with a big dose of scepticism and critical analysis of what the contemporary knowledge provides, has proved itself many times over in the development of scientific theory and an ever-expanding world view, taking us very far from the supposedly ‘better arts and studies’ the seventeenth century pundits thought we should be occupied by, but also making us realize that the science that has flowed from curiosity has mightily informed those ‘better arts and studies’, which can be perhaps best summarized by the four Kantian questions, Who are we? What do we know? What should we do? and What can we hope for?
on transcendental constructions: a critique of Scott Atran
Some years ago, when watching some of the talks and debates in the first ‘Beyond Belief’ conference at the Salk Institute, I noted some tension between Sam Harris and his critique of religion generally and Islam in particular, and Scott Atran, an anthropologist, who appeared to be quite contemptuous of Harris’s views. Beyond noting the tension, I didn’t pay too much attention to it at the time, but I’ve decided now to look at this issue more closely because I’ve just read Ayaan Hirsi Ali’s powerful book Infidel, which gives an insider’s informed and critical view of Islam, particularly from a woman’s perspective, and I’ve also listened to Chris Mooney’s Point of Inquiry interview with Atran back in April, shortly after the Boston marathon bombing.
The interview, called ‘What makes a terrorist?’ was mainly about the psychology of the more recent batch of terrorists, but in the latter half, Atran responded to a question about the role of Islam specifically in recent terrorist behaviour. It’s this response I want to examine, not so much in the light of Sam Harris’s contrasting views, but in comparison to those of Hirsi Ali.
In bringing up the role of Islam in terrorism, Chris Mooney cites Sam Harris as pointing out that ‘there’s something about Islam today that is more violent’. Atran’s immediate response is that ‘this is such a complex and confused issue’, then he says that ‘religions are fairly neutral vessels’. This idea that religions, especially those that survive over time, have a degree of neutrality to them, has some truth, and in fact it served as the basis for my critique of Melvyn Bragg’s absurd claims that Christianity and the KJV Bible were largely responsible for feminism, democracy and the anti-slavery movement. But there is a limit to this ‘neutrality’. Religions are clearly not so ‘neutral’, morally or culturally, that they’re interchangeable with each other. Fundamentalist, or ultra-orthodox, or ultra-conservative Judaism is not the same as its Islamic or Christian counterparts. In fact, far from it. And yet these three religions ostensibly share the same deity.
The interaction between religion and culture is almost impenetrably complex. I wrote about this years ago in an essay about traditional Australian Aboriginal religion/culture, in which it’s reasonable to say that religion is culture and culture is religion. In such a setting, apostasy would be meaningless or impossible – essentially a denial of one’s own identity. Having said that, if your religion, via one of its principal texts, tells you that apostasy is punishable by death, you’ve already got a yawning separation between religion and cultural identity – the very reason for the excessive threat of punishment is to desperately try to plug that gap. It’s like the desperate cry of a father – ‘you’ll never amount to anything without me!’ – as the son walks out the door for the last time.
These major religions – Judaism, Islam and Christianity – are embedded in texts that are embedded in culture. Different, varied texts interacting complexly – reinforcing, challenging, altering the culture from whence they sprung. Differently. Judaism’s major text, always arguably, is the Torah. Christianity’s is the New Testament, or is it the gospels? Islamic scholars – but also those believers who rarely ever read the sacred texts – will argue about which texts are most important and why. Nevertheless, Judaism, Christianity and Islam all have a different feel to them from each other, even given the enormous variation within each religion. Judaism is profoundly insular, with its chosen people uniquely flayed by their demanding, unforgiving god. Christianity is profoundly other-worldly with its obsession with the saviour, the saved, the end of days, the kingdom to come, the soul struggling for release, not to mention sin sin sin. Islam, a harsh, desert religion, somehow even more than the other two, is about denial, control, submission, and jihad in all its complex and contradictory manifestations and interpretations. The status of women in each religion, in a general sense, is different. Christianity gives women the most ‘wriggle-room’ from the start, but its interaction with the different cultures captured by the religion can sometimes open up that space, or close it down. The New Testament presents a patriarchal culture of course, but in the gospels women aren’t given too bad a rap. Paul of Tarsus notoriously displays some misogyny elsewhere in the NT, but it isn’t particularly specific and no detailed restrictions on women’s freedom are presented. More importantly, the dynamism of western culture has blown away many attempts to maintain the restrictions on women’s freedom dictated by Christian dogma – pace the Catholic Church. In any case, Christianity has no equivalent to Sharia Law, with its deity-given restrictions and overall fearfulness of the freedom and power of women. And neither Christianity nor Islam has the obsession with ritual and with interpretation of the deity’s very peculiar requirements that orthodox Judaism has.
To return, though, to Atran. He argues that the reason the big religions survive and thrive is precisely due to their lack of fixed propositions – which is why, he says, that we need sermons to continually update and modernise the interpretations of texts, parables, suras and the like. I’m not sure if the Khutbas of Moslem Imams serve the same purpose as priests’ sermons, but I generally agree with Atran here. The point, of course, is that though there is much leeway for interpretation, there are still boundaries, and the boundaries are different for Islam compared to Christianity, etc.
What follows is my analysis of what Atran has to say about what are, in fact, very complex and contentious matters relating to religion and social existence. Whole books could be, and of course are, devoted to this, so I’ll try not to get too bogged down. I’m using my own transcript of Atran’s interview with Mooney, slightly edited. Occasionally I can’t quite make out what Atran is saying, as he sometimes talks softly and rapidly, but I’ll do my best.
So, after his slightly over-simplified claim that these big religions are ‘neutral vessels’, Atran goes on with his definition. These religions are:
… moral frameworks that provide a transcendental moral foundation for large groups coalescing – for how else do you get genetic relatives to form large co-operative groups? They don’t have to be necessarily religious today, but it involves transcendental ideas. Take human rights, for example, that’s a crazy idea. Two hundred and fifty years ago a bunch of intellectuals in Europe decided that providence or nature made all human beings equal, endowed by their creator with rights to liberty and happiness, when the history of 200,000 years of human life had been mostly cannibalism, infanticide, murder, the suppression of minorities and women, and so [through the wars?] and social engineering, they took this crackpot idea and made it real.
I have a few not so minor quibbles to make here. Presumably Atran is using the term ‘transcendental’ in the way that I would use the term “over-arching’ – a much more neutral, and if you like, secular term. The trouble is – and he uses this term often throughout the interview – Atran uses ‘transcendental’ with deliberate rhetorical intent, taking advantage of its massive semantic load to undercut various secular concepts, in this case the ‘crackpot’ concept of human rights.
This isn’t to say that Atran objects to human rights. My guess is that he regards it as a somewhat arbitrary and unlikely concept, invented by a bunch of European intellectuals in the Enlightenment era, that just happened to catch on, and a good thing too. That’s not how I see it. It’s just much much more complex than that. So much so that I hesitate to even begin to explore it here. The germ of the concept goes back at least as far as Aristotle, and it involves the increasingly systematic study of human history, and human psychology. It involves the science of evolution, and it involves pragmatic global developments in commerce and diplomacy. Eighteenth century Enlightenment ideas had a catalytic effect, as did many developments of the scientific enlightenment of the previous century, as did the growth of democratic ideas and the concept of systematic universal education and health-care in the nineteenth century, in the west.
My point is that, though I have no problems with calling human rights a convenient fiction – nobody ‘really’ has rights as such – it’s based on a this-worldly (i.e. non-transcendental) understanding of how both individuals and societies flourish and thrive, in terms of the contract or compromise between them.
Atran goes on:
But, in general, societies that have unfalsifiable and unverifiable transcendental constructions win out over those that don’t – I mean, Darwin talked about it as moral virtue, and said that this is responsible for the kind of patriotism, sympathy and loyalty that makes certain tribes win out over other tribes in […] competition for dominance and survival, and again, without these transcendental ideas people can’t really be blinded to [exit strategies], I mean, societies that are based on social contracts, no matter how good they are, the idea that there’s always a better deal down the line makes them liable to collapse, while these societies are much less prone to that. And there are all sorts of other things associated with these sorts of unverifiable propositions.
Presumably these ‘unfalsifiable and unverifiable transcendental constructions’ are religions, and I’ve no great objection to that characterisation, but I’m not so convinced about the positive value for ‘dominance and survival’ of these constructions. One could argue that my kind of scepticism can only flourish in a secure environment such as we have in the west, where such ‘undermining’ values as anti-nationalism and atheism can’t threaten the social cohesion of our collective prosperity and sense of superiority to non-western notions. There are just no ‘better deals down the line’, except maybe more health, wealth and happiness, commitment to which requires the very opposite of an ‘exit strategy’. In other words, western ‘social contract’ societies, in which religious belief is rapidly diminishing (outside the US), are showing no sign of collapsing, because there is no meaningful exit strategy, unless a delusional one. There is no desire or motivation to exit. We’re largely facing our demons and rejecting overly ‘idealistic’ solutions.
Perhaps my meaning will be clearer when we look at more of Atran’s remarks:
So now, the propositions, these things themselves can be interpreted, however, depending on the political and social climate of the age. Islam has been interpreted in ways that were extremely progressive at one time, and at least parts of it are extremely retrogressive, especially as concerns science for example, the position of women in the world, especially parts of it in many countries it’s extremely retrograde. But, Islam itself, I mean does it have some essence that encourages this kind of crazy violence? No, not at all – that truly is absurd, and just false.
Atran’s becoming a bit incoherent here, and maybe he expresses himself better elsewhere, but his base argument is that there’s no ‘essence’ to Islam which renders it more violent than other religions, or transcendental constructions (eg communism or fascism) for that matter. He overplays his hand, I think, when he claims that this is ‘absurd’ and obviously false. We could call this ‘the argument from petulance’. Islam does have some essential differences, I think, which makes it more able to act against women and against scientific ideas, though I agree that this is a matter of degree, and that it’s very complex. For example, the growth of Catholicism in Africa has combined with certain aspects of tribal culture and patriarchy to make African Catholic spokesmen very outspoken against homosexuality – and a recent local television program had a Moslem leader speaking up in favour of gay marriage. So, yes, there is nothing fixed in stone about Islam or Christianity with respect to human values.
The thing is that, for writers like Ayaan Hirsi Ali, and I suspect Sam Harris too, the question of ‘essentialism’ is largely academic, for right here and right now people are being targeted by Moslems (under the pressure of cultural connections or disconnections), because they are apostates, or critics, or women trying to get an education, or women dressing too ‘immodestly’, and this is causing great tension, even to the point of death and destruction here and there. In fact, Hirsi Ali, in calling for an enlightenment in the Moslem world, is backing a non-essentialist view. It’s the culture that has to change, but of course religion, with its transcendentalist, eternalist underpinnings, acts as a strong brake against cultural transformation. To engage in the battle for moderation is to battle for this-wordly, evidence-based thinking on human flourishing, against transcendentalist ideas of all kinds.
Atran, I think, relies too heavily on his notion of ‘transcendental constructions’, which he uses too widely and sweepingly, even with a degree of smugness. Let me provide one more quote from his interview, with some final comments.
But again, I don’t see anything about Islam itself… you need some kind of transcendental ideal to get people to sacrifice for genetic strangers, for these large groups. Religion is the best thing that human history has come up with, but there are other competing transcendental notions of which democratic liberalism, human rights, communism, fascism, are others, and right now the democratic-liberal-human rights thing is predominant in a large part of the world and it’s a salvation [……..] and people don’t want that or feel left in the driftwood of globalisation, they are looking for something else to give them equal power and significance.
Methinks Atran might’ve been spending too much time in the study of religious/transcendental ideas – he’s seeing everything though that perspective. I myself have written about democracy, in its various manifestations, from a sceptical perspective many times, and I’ve been critical of the over-use of the concept of rights, and so forth. It’s true enough that people can take these concepts, along with fascism or communism, to a transcendental level, making of them an unquestionable given for ‘right living’ or ‘a decent society’, but they can also be taken pragmatically and realistically, reasonably, as the most serviceable approaches to a well-functioning social order. Social evolution is moving quickly, and we can make sacrifices for genetic strangers, based on our growing understanding, as humans, of our common genetic inheritance. We’re not so much genetic strangers, perhaps, as we once thought ourselves to be. Indeed, it’s this growing understanding, a product of science, that is expanding our circle of connection beyond even the human. We need to promote this understanding as much as we can, in the teeth of transcendentalist, eternalist, other-worldly ideas about submission to deities, heavenly rewards and spiritual superiority.
‘organic’ food – the greatest scam in the west
[there is a] fashion to talk as if art were something different from nature, so that things artificial should be separated from things natural, as differing totally in kind… Things artificial differ from things natural not in form or essence, but only in the efficient.
Francis Bacon, De Augmentis Scientiarum, 1623
Someone at work offers me some food, from a cooking class she teaches. She describes it as very healthful, ‘and organic too’, she proudly confides. ‘Well of course, it’s organic, it’s food,’ I mildly reply. ‘Well, yes, but you know what I mean,’ she says.
Unfortunately, I did know what she meant. She meant ‘organic’ in the cheap, shallow, duplicitous, marketing way, not in the deep, scientific way.
And so I begin a piece that is long overdue, and which, I’m sure, will not win me any friends, assuming anybody reads it at all.
‘Organic’ food has been getting my goat for a few years now, and it’s time I laid out my objections based on the evidence I’ve accumulated over the years, while at the same time looking again at the evidence, just in case there’s something redeeming about this labelling and marketing practice that I’ve missed.
First, though, I’ll talk about marketing, which is the real focus of my ire. The term ‘organic food’, as so many people have pointed out, is tautologous. All food is organic, that is an unarguable, scientific fact. So it takes a deal of hubris, and, I reluctantly admit, a deal of marketing genius, to be able to sell a product and a process intended, quite deliberately, to cast doubt on the health and nutritional value of 99.999% of the food we eat. This is the scam of all scams, and what’s more, it has been entirely successful. Usually when we think of scams, we think of those who got caught – the Bernie Madoffs and Jeff Skillings of the world, the bad and the blameworthy who make us feel better for not being like them. Their scams are over, lessons learnt, systems tightened, vigilance heightened, but there’s no end in sight for the organic food scam. It’ll be with us for as long as the words ‘toxic chemicals’ have currency, and that’ll be around the twelfth of forever. What’s more, there’s no ‘body’ to blame, no obvious perpetrator or mastermind. In that way, and in more than a few others, it’s a bit like religion.
I note that most people I know who swear by ‘organic’ food are also opposed to GMOs, suspicious of mainstream medicine, and dabblers in various pseudo-scientific approaches to health and well-being. They certainly place more value in ‘the fruits of the earth’ than the products of the lab. This article of faith has been labelled ‘the naturalistic fallacy’ by sceptics, though philosophers might quibble about that – as would I, having struggled over many years with that particular concept, introduced by philosopher George Moore more than a century ago. Probably better to label this way of thinking as ‘the appeal to nature’. In any case, it’s certainly an example of fallacious reasoning, as the insightful Francis Bacon was one of the first to point out.
My many qualms about the ‘organic’ food movement have been reinforced by a listen to the ever-reliable Brian Dunning’s Skeptoid piece on the subject, and I’ll use that as the basis, or at least the starting point, for this post. In fact, you might well be better off listening to Dunning’s analysis, which will doubtless be more comprehensive and concise than mine. I’m mainly writing this to get the information and the understanding of the issues more clearly lodged in my head.
The generally understood scientific term for an organic compound is one that’s produced by living entities. Chemically, it’s a carbon-based molecule with a carbon-hydrogen bond. Coal is an organic compound, and so, interestingly, is plastic. If the term ‘organic’ is used in any other way, you should be sceptical. My scepticism compels me to use the term ‘organic’ food, with scare quotes, to highlight this dubious use.
In order to be certified ‘organic’, food and agricultural products must be produced under a set of guidelines which vary from country to country, and which are regulated in different ways in different places. This Wikipedia article provides some of the guidelines common to most western countries:
- no human sewage sludge fertilizer used in cultivation of plants or feed of animals[1]
- avoidance of synthetic chemical inputs not on the National List of Allowed and Prohibited Substances (e.g. fertilizer, pesticides, antibiotics, food additives, etc.), genetically modified organisms, irradiation, and the use of sewage sludge;
- use of farmland that has been free from prohibited synthetic chemicals for a number of years (often, three or more);
- keeping detailed written production and sales records (audit trail);
- maintaining strict physical separation of organic products from non-certified products;
- undergoing periodic on-site inspections.
So let’s look at the first three of these, which, presumably, are key to producing goods superior to, or healthier (and tastier) than goods that don’t earn the ‘organic’ label.
The issue of sludge fertiliser and its potential dangers isn’t really an ‘organic’ food issue, it’s one for any agricultural product. If you use sludge fertiliser, and it causes contamination to humans or animals, obviously there will be consequences for your business and yourself, whether you’re trying to produce ‘organic’ food or not. The ‘freedom from sewage sludge’ label that ‘organic’ foods are presumably entitled to display appears to be meaningless unless non-‘organic’ producers are all using the stuff. And even if they were, the issue is one of contaminants, not sewage sludge per se. I don’t know if this is an issue in Australia, but there is no evidence, out of the USA, that anyone is being contaminated by non-‘organic’ foods. No matter what the complexities of applying sludge in farming – organic or inorganic, treatment methods, etc – it is irrelevant to the ‘organic’ food issue. It appears to be used for ideological reasons, to hint that, somehow, somewhere, the use of untreated or improperly treated sludge is slowly killing us.
The second guideline, which for some reason incorporates the first guideline, rendering that guideline superfluous, is the key guideline to understanding the psychology of ‘organic’ food, and the ‘appeal to nature’ fallacy upon which it’s based. ‘Organic’ food producers must not use ‘synthetic’ fertilisers or pesticides ‘not on the National List of Allowed and Prohibited Substances’, in other words nothing synthetic of any kind.
‘Organic’ producers and marketers like to promote their products as fertilizer and pesticide free. This is complete bullshit. Virtually all agricultural products are subject to pest infestation and this needs to be dealt with, one way or another. Methods also need to be employed to enrich the soil, to render it more fertile. The only difference between ‘organic’ producers and the rest is that ‘organic’ producers are constrained by their anti-science ideology. Synthetic fertilizer, for example, involves the production of the key nutrients for plant growth – nitrogen, potassium and phosphorus – on a commercial scale. ‘Organic’ farming involves the same nutrients, but delivered the hard way, through fish and bone meal, earthworm castings and the like. The only difference is that these materials are more costly and less efficient, as they deliver a much lesser and more variable load of the nutrient per volume, and are thus less straightforward to use accurately and systematically, and are far costlier to transport. The use of synthetic fertilisers, as I’ve pointed out many times, has, with the improvement through hybridisation of particular grains and fruits, increased crop yields by ten times and more, and has saved the lives of millions since their introduction in the sixties.
But the real point here is the duplicity of labelling synthetic fertilizer, which is able to isolate and concentrate the required nutrients in the most efficient way, as ‘chemical’ (with the implication that it just might be toxic), as if the fertilizers used in ‘organic’ farming are somehow free from chemistry.
The third guideline mentioned above really shows how committed the ‘organic’ marketers are to scaring people about conventional farming. There is no need to keep conventional farmland free from ‘prohibited synthetical chemicals’ in order to use it for ‘organic’ farming. I wonder what is meant by ‘prohibited’ here? If they’re prohibited by government authorities, then of course you shouldn’t use the land – but then why would any farmer use such substances, thus poisoning her own produce? If they’re prohibited solely by ‘organic’ regulations, then they’re simply ideologically driven, arbitrary, and a product of the ‘appeal to nature’ fallacy.
As Dunning points out, ‘organic’ products are perfectly healthy and safe, but there’s no reason to believe they’re healthier and safer than non-‘organic’ products. I personally prefer to avoid eating too many processed food products because I think it’s better for our bodies to expend energy on the process of digestion, and because many processed foods have added sugar which our bodies don’t need and which can cause problems. I think there’s a fair amount of good scientific evidence for this approach to diet. I’ve not as yet encountered any scientific evidence for the benefits of ‘organic’ foods, except that they’re generally unprocessed and vegetarian, which is mostly good (don’t forget, though that a diet of potato chips is also vegetarian).
A perhaps more subtle, and superficially more cogent argument for ‘organic’ foods is the environmental argument. Okay, so conventional food isn’t poisoning us or giving us cancer or heart disease, but you gotta admit that it’s unsustainable. ‘Organic’ food really cares for the soil, it’s based on a deep connection with nature, a respect for the land, it gives as good as it takes, it’s about long-term sustainability. Conventional farming is, by contrast, instrumentalist, exploitative, impersonal, short-term, destructive etc etc.
This is a simplistic and ideological claim, not evidence-based. Firstly, let’s look at how conventional farming obtains its three key nutrients for enriching the soil. Nitrogen is, of course, freely available from the atmosphere and infinitely sustainable. Phosphorus is mined from phosphate rock, of which we have reserves to last centuries. Potassium comes from ancient ocean deposits, of which we have millenia of reserves. Of course these reserves are finite, so seawater extraction is considered a viable alternative, for both potassium and phosphorus. As Dunning points out, this creates a sustainable cycle as plant matter and farm runoff returns to the oceans, but ‘organic’ certification, at least in the US, doesn’t allow sustainable atmospheric and seawater extraction. ‘Organic’ chemical fertiliser can only be sourced from animal waste and other recycled resources, using criteria which are ideological rather than scientific, and so more or less arbitrary. Further, these resources can’t be marshalled in sufficiently commercial quantities to feed large populations, especially in developing countries where there just isn’t the infrastructure to make fertilisation under ‘organic’ guidelines viable on a commercial scale. ‘Organic’ farming is a distinctly western, middle class ideology.
It’s also insulting to conventional farmers to suggest that they’re more exploitative and short-term in their use of their own land. This goes as much for multinational agricultural concerns as for individual farmers. Both groups are interested in long-term viability, for obvious reasons. Crop rotation and other forms of long-term soil management have long been practised by conventional farmers, who must naturally balance these with other production concerns. This is surely grist for the mill for all agriculturalists, as they would wish to reduce the cost of applying fertiliser or herbicides wherever possible.
Returning to the pesticide/herbicide issue, it’s often harder, and more expensive, for ‘organic’ farmers to find ‘natural’ or plant-based chemicals to use instead of synthetic products, and these costs must needs be passed on to consumers. The synthetic products have, of course, been passing health and safety checks for decades. One such chemical, rotenone, a colourless, odourless ketone found in the seeds and stems of a number of plants including the jicama vine, has in recent years been all but abandoned, due to connections found between its use and the incidence of Parkinson’s Disease among farm workers.
I could go on, but hopefully you get the picture. I’ll end, as I began, with the use of language. There are plenty of organic entities, to use the word in its right way, that are poisonous to humans – be they berries or bugs, frogs or sea creatures. A fine example is the fugu fish, with its deadly poison, tetrododoxin, of which quantities are found in the skin, the skeleton, the intestines, the ovaries, and above all the liver. Eaten usually as sashimi (ie raw), it must be prepared by rigorously trained chefs, and even then you can never be sure – which seems to be essential to its charm as a delicacy. To quote from this travel advisor:
Tetrododoxin does not cross the blood-brain barrier, so the victims remain fully conscious while their central nervous system gradually shuts down, first producing dizziness and incoherent speech, then paralysing the muscles. This can lead to asphyxia, and possibly death. (There is no antidote for fugu poisoning).
meanderings – weight loss, Ayaan Hirsi Ali, a befuddled reverend gentleman, and a plethora of choices
Listening to various podcasts and interviews and debates and documentaries, absorbing, absorbing, as well as reading texts about complex ideas which don’t easily mesh with each other, which spin me off in different directions and which of course I only discuss with my undemanding self, all this befuddles my brain a bit, making it hard to know what to start writing about, I’ve forced myself to sit down here, as I must do on a daily basis, and see if I can pin down an idea or two.
Montaigne-like I always feel its safest to start with myself and last night I caught the last few minutes of another Michael Mosley documentary on diet, lifestyle and health, in which he ended up being sceptical of the effort required to give yourself the best shot at living to 200, which I felt was fair enough, but I was alerted to my situation weight-wise by the observation that, as you lose weight, your metabolism slows down, which is one of the main reasons it’s so hard to keep the weight off, and a good example of that is that yesterday was my birthday so I indulged myself a little, culinarily speaking, but not really a lot, but I paid for it with my weight going up by one whole kilogram on the day before, and experience tells me that it’ll take about three days of food deprivation to get that kilo off again, so is there a way to get the metabolism to speed up, if in fact that’s the problem? We’re in the depths of winter here, though winters here are mild, but I still use it as an excuse not to go out in the cold for exercise, and poverty has prevented me from buying an exercise bike and getting my metabolism up through the HIT program, which would surely be my best option. Excuses excuses in short.
I’ve just started reading Ayaan Hirsi Ali’s Infidel, having read The Caged Virgin a few years back, and I’m still pissed off by the silly remark about her by Reza Aslan, a supposedly liberal Moslem. The use of a pared-down, low-key style when dealing with horrors is always effective. And even when not dealing with horrors. Also, the courage of this woman – even with the in-your-Moslem-face title of the book – is inspiring. She’s a proud catch-me-if-you-can infidel, whose affront to a particular religion is a challenge to our liberal secular values, a challenge we surely must meet. Also, she’s a human being who sees clearly the damage to human values, and particularly the value of the female species, wrought by many religious and cultural practices.
Interestingly I just happened to watch a debate video yesterday, filmed in 2007 between Christopher Hitchens and one Al Sharpton, a reverend gentleman of some note in the US, apparently. Hitchens was quite genial, for Hitchens, and Sharpton seemed far from sharp, so it was all a bit ho-hum. But I was delighted to find that Hirsi Ali was in the audience, and she asked the last question of the video, addressing it to Sharpton. It was a good question too, just what I would’ve asked him. Throughout the debate, Sharpton complained that Hitchens was avoiding the key point of the debate, whether ‘God is great’ or not, as per Hitchens’ book. But Sharpton was nowhere consistent, claiming in his first speaking opportunity (after Hitchens had kicked off the debate) to be annoyed that Hitchens was focusing on the putative criminal behaviour of the Christian god, and the questionable morality of the gospels, etc, but wasn’t focusing on the existence of this god, and then suddenly switching from the ontological issue to the ethical issue, by claiming that because people used the god in this or that way, or described or interpreted him thus and so, that made no difference to the god himself – as if the Bible wasn’t a written history of that god, and the proper way for that god to be known according to most theologians. In fact Sharpton’s various harpings on this issue amounted to a complete rejection of the Bible, it seemed to me, in favour of a personal relationship with this god in which believers can make of him (or her) what they will. An extraordinarily flexible theology which would seem to render a common morality, derived somehow from this being, completely impossible.
Of course Sharpton tried to make other points, including a variant of the watch/watchmaker shibboleth, all of them equally specious. The only thing going for him was that he seemed a genuinely likeable, kindly gentleman. And to be fair, Hitchens makes some specious arguments too, not about gods but about Iraq.
But to return to Hirsi Ali, her question was multi-faceted, concerning, inter alia, Sharpton’s god’s existence, and how she/he/it came into being, but the last facet was ‘Isn’t it odd that you carry a Christian title and that you refuse, even for once tonight, to defend the church, and the content of the Bible?’ My admiration for her and devotion to her shone in that moment, and of course Sharpton responded by presenting again this relativistic notion of an ahistorical, unparticularised god who is ‘the same’, but of course quite different, whether you’re a Christian, a Moslem or a follower of no organised religion. This conception, which conveniently leaves out the central figure of Jesus, seems a complete denial of the particularity of Christianity, and simply highlights the questioning of Sharpton’s Christian title. However, I really do believe that the poor befuddled gentleman just didn’t get it.
Speaking of befuddlement, while absorbing and forgetting all these ideas and positions and facts I’m reading or hearing about, I get torn about what I should focus and write on – for example if there’s anything to this system 1 and system 2 thinking and the idea that religious thinking is ‘more natural’ than scientific thinking (words must be chosen carefully here), or if it’s true that there’s an asymmetry between liberals and conservatives in their cognitive biases and motivated reasoning, or should I launch into a critique of the marketing scam that is ‘organic’ food, or should I try to elaborate in my own words the reason why the Copernican theory wasn’t considered heretical until Galileo came along quite a bit later? Or should I just give up and watch the Tour de France to its seemingly inevitable conclusion? Help me, oh goddess.
Anatol Rapoport’s rules: a tough ask
‘Anatol Rapoport…..once promulgated a list of rules for how to write a successful critical commentary on an opponent’s work. First, he said, you must attempt to re-express your opponent’s position so clearly, vividly and fairly that your opponent says “Thanks, I wish I’d thought of putting it that way.” Then, you should list any points of agreement (especially if they are not matters of general or widespread agreement), and third, you should mention anything you have learned from your opponent. Only then are you permitted to say so much as a word of rebuttal or criticism.’
These rules are useful for everyday spoken arguments too, of the more or less intellectual sort. They can be adapted for the spoken, to keep your interlocutor onside, (along with smiles, nods, effective impressions of interest, appropriate turn-taking and the suppression of irritated grimaces). I think I’m reasonably good at following Rapoport’s rules in writing – or trying to keep to them. In conversation it’s much harder, especially after a few drinks!
I once watched former High Court Justice Michael Kirby, whom I already greatly admired, in a television exchange with a Pauline Hanson supporter. He listened politely to the man’s bluster, then spent some time painstakingly and seemingly sympathetically presenting the Hanson view on immigration, and then he described in details its flaws while defending its validity as a position, at least on an emotional level. It was an extraordinary performance, full of discipline, grace and poise. I use the memory to punish myself when I perform disappointingly in an argument – sometimes ‘winning’ the argument, or at least seeming to, but losing the respect or interest of the opponent.
two quotes
From Reza Aslan’s book on cosmic warriors and how we might deal with them:
If secularism arose in the eighteenth and nineteenth centuries in response to the erosion of religious certainties, perhaps the rise of religious identities can be blamed on the growing disillusionment with secularism.
… as pious young men and women hectored and humiliated by pseudointellectuals such as Ayaan Hirsi Ali, Oriana Fallaci and Brigitte Gabriel, who make a living fanning the flames of racism and Islamophobia.
Unfortunately for you, Reza, secularism is continuing to rise, throughout Europe, in Britain, in Australasia, in Canada, in Japan, and even in the USA, as every census and reputable survey conducted in every one of those countries in the last thirty years clearly indicates. But of course it’s not surprising that religious people have a problem with facts. Here in Australia, where the rise in secularism is accelerating, as recent censuses clearly show, we have been judged by the OECD as ‘the happiest country in the world’ – top of the Better Life Index – for the third year in a row. Now what was that about growing disillusionment Reza?
Reza Aslan is a liberal Muslim who reminds me of Jack Miles, a liberal Christian (who highly recommends Reza’s book). Both have interesting things to say, and knowledge to impart, in their vastly different fields of interest and expertise, but both just can’t help having a go at those they haven’t the mind to get a handle on – atheists. Or secularists, or nonbelievers or whatever. They just can’t help having their infantile moments.
I’ve not read Fallaci or Gabriel, mea culpa, but Aayan Hirsi Ali writes about the questionable value of Islam as I have written about the questionable value of Christianity, the religion I know most intimately. I suspect we share a concern about people who readily accept a thoroughly male supernatural being with a shady past as their lord and master. It just doesn’t seem quite right. It’s a concern we share with such other pseudointellectuals as Albert Einstein, Bertrand Russell, Richard Feynman and Francis Crick, to name but an exalted few. However, I also suspect, or rather I’m certain, that Hirsi Ali has suffered for her apostasy quite a bit more than I have, and I admire her bravery. Is Aslan describing her as racist? Oh no, Allah forbid, he’s just describing her as cynically fostering racism and Islamophobia to sell her books. It goes to show, or at least suggest, that even the most liberal religious people can’t pull their heads out of their arses for long enough to see that most secular thinking and writing is sincere and passionate as well as true.
A shame, for I mostly agree with Reza Aslan’s approach to dealing with the cosmic religious warriors described in his book.













