Should democratically elected leaders in more or less secular countries ever say that this or that religion is essentially good or essentially bad? The dilemma is especially acute, perhaps, if the religion that they want to speak about is one which they don’t happen to practise, and presumably don’t know about in any depth. But ever since September 2001, and especially over the last few weeks of intensifying conflict with Islamic State, it has been a question that Western heads of government cannot completely duck. The West is at war with an adversary which claims to be acting in the name of Islam. Does that mean that the West is, in any sense whatever, at war with Islam?