About me
Bookshop

Get new posts by email.

About me

Hazards and risks

When thinking about risk assessment, it is often useful to separate the hazard (the bad thing that might happen) from the risk (the hazard plus the likelihood of it occurring). Things which are quite hazardous (wild tigers) can be low risk if they’re unlikely to cause harm (there aren’t any wild tigers in the UK). And things which can seem non-hazardous (coconuts) can be high risk if very likely to cause harm (perilously dangling above someone’s head on a windy day).


In a conversation recently, someone commented in passing that it wasn’t possible to meaningfully assess risk when a hazard is so large as to be unquantifiable: say the end of the earth, a world war, or a global pandemic.

I entirely disagree.

Firstly, no hazard is unquantifiable large: of the examples cited, a pandemic is less hazardous than the apocalypse.

But even ignoring that point, it’s self-evident that an activity with a low likelihood of causing the apocalypse is lower risk than one with a high likelihood of causing the apocalypse.

“Okay,” you might say, “but what I really meant is that you can’t compare with between hazards because the apocalyptic one will always win out.”

But that’s nonsense too. A threat with a negligible likelihood of causing a cataclysmic event is self-evidently lower risk than a threat with an extremely high likelihood of killing someone. Firing a gun in a crowded place is riskier than allowing visitors to tour nuclear power plants, even though there is an infinitesimally small chance of the latter being the start of a chain of events that leads to a nuclear disaster.

One could spend a lifetime trying to derive where the lines lie: and, indeed, as a society, we do just that. Through time, democracy, effort and research, we try to reach a societal consensus on where the balance of risks lies. We end up taking actions that have potentially world-ending consequences (say, building nuclear weapons) because we believe it’s the least risky approach.


In reality, risk assessments are generally much more complex than this implies: the theoretical balancing of risks is often easier than understanding the likely effects of each course of action. Any given action (or inaction) has myriad effects, only a fraction of which are pertinent to the specific risk under consideration.

In medicine, giving antibiotics might reduce the risk of a bacterial infection ending someone’s life. It may also cause side effects for an individual, including death. When applied as a general rule in guidance, it will also have extensive wider societal implications: financial cost, the opportunity cost of choosing to prescribe antibiotics rather than spending time doing something else, antimicrobial resistance, and so on. Not giving antibiotics is also very likely to have a whole host of implications, which take effort to foresee. And both courses of action will almost certainly have unforeseeable consequences, too.

The process of working out the implications of each course of action can become enormously complicated, and can often be extremely uncertain. But it must be done because we must make a decision.


In medicine, at least, well-written and considered guidelines constantly try to take a reasoned, explained judgement as to which path is most likely to lead to net benefit in most circumstances. NICE, for example, is typically great at explaining its committees’ thinking on these things—and also great at changing guidance when the real world implications of implementation turn out to differ from predictions.

But sadly, not all guidelines are well-written and considered. Astonishingly, I still come across newly published guidance which reports that intervention X will reduce the risk of disease Y, with no consideration even of side effects for individual patients, let alone wider societal consequences. The guidance vacuously recommends X based on its impact on Y alone.

If Y is common but mild and self-limiting, and X is extraordinarily expensive, then prescribing X will rarely be justified.

If the risks are such that you’d need to prescribe 3,000 doses of X to prevent one case of disease Y, and X has common side effects, then prescribing X may not be justifiable.

It shames my profession that some of this faulty guidance is public health guidance—the part of the medical profession that ought to be most attuned to accounting for costs and unintended consequences.


Balancing risks can be very hard, but it is always possible and indeed always necessary, especially in medicine.

This post was filed under: Health, Post-a-day 2023.

Recently published posts

‘Sinterlation’ by Ian Randall / 21 November 2024

‘Landed’ by Les Johnson / 20 November 2024

A incurious ego / 18 November 2024

‘Types of Happiness’ / 17 November 2024

Herring gull / 16 November 2024

‘Small Things Like These’ / 15 November 2024




Random posts from the archive





The content of this site is copyright protected by a Creative Commons License, with some rights reserved. All trademarks, images and logos remain the property of their respective owners. The accuracy of information on this site is in no way guaranteed. Opinions expressed are solely those of the author. No responsibility can be accepted for any loss or damage caused by reliance on the information provided by this site. Information about cookies and the handling of emails submitted for the 'new posts by email' service can be found in the privacy policy. This site uses affiliate links: if you buy something via a link on this site, I might get a small percentage in commission. Here's hoping.