donderdag 12 juni 2014

Mind The Risk

Last weekend I decided to take all of this to a new level. I've started a new website, and eventually company of my own: Mind The Risk.


As a consequence I won't be posting on the HEACH-blog, but future postings will be found on Mind The Risk. Additionally most of the material from this blog will be moved there in the near future (all reviews and summaries of professional literature is already there).

Find the new site here. And thanks for visiting us!

dinsdag 13 mei 2014

Indexed

Was just reminded of what once was one of my favourite blogs to follow and started following it again. Great observations here. Take for example this: http://thisisindexed.com/2014/05/back-back-beasts/

woensdag 7 mei 2014

Trapping Safety Into Rules

In the picture below you'll find a scan of the review I wrote of this book for the magazine of the Dutch Society of Safety Science, NVVK Info. This was published there in the May 2014 issue.


The English translation goes roughly like this:

Trapping Safety Into Rules

edited by Corinne Bieder and Mathilde Bourrier (ISBN 978-1-4094-5226-3)

My current employer has a very fine internal library that contains a decent selection of books on safety. The fine ladies in the library also alert professionals to new releases that aren’t in the collection yet, but may be wanted in future. The title of this book immediately caught my attention when it was mentioned in a newsletter and that very day an order for a personal copy went out.

On 278 pages the book presents a collection of 15 articles, plus an introduction and epilogue by the editors. The contributions are based upon experiences within various branches: (mainly Norwegian) offshore, air traffic control, railways and patient safety, and there are also some more theoretically oriented contributions. The fact that the book is a collection of articles slightly affects the coherence and flow (even though I’m pleasantly surprised by the fact that the contributors refer to each other’s contributions quite frequently) and in some cases also the readability suffers. Some authors just seem to have a hard time ditching the academic approach to writing that they forget that books should be easy to read. A listing of references that fills an entire paragraph is just nonsense and should have been caught and removed by the editors.

The book’s title might suggest that the book has a critical stance with regard to (safety)rules, but as the subtitle ‘How desirable or avoidable is proceduralization?’ indicates the book highlights both sides and discusses both the pros and cons of safety rules in four different parts with the following headers:

1: Where do we stand on the bureaucratic path towards safety?
2: Contrasting approaches to safety rules
3: Practical attempts to reach beyond proceduralization: The magic tools illusion
4: standing back to move forward

The subjects discussed in the book are too varied to cover in a full overview here, so let me pick out some highlights. In part 1 we find among others an interesting discussion of ‘No rule, no use’. This is an attitude that almost any safety professional will have met at some time (or even regularly). A safety measure may be wise to implement, but never the less the question arises: “Where does it say that we have to do this?”. Ergo, quite often without a rule saying so, one does not implement a safety measure.

In the second part especially Kenneth Pettersen’s discussion about abductive thinking appealed to me. Me discusses this from a background in airplane maintenance. Mechanics often manage in doing their jobs and dealing with unexpected situations because they deviate from rules. The standard rules often aren’t helpful to solve unique problems, so one has to deviate. This is one of the most interesting contributions in the collection and more or less mandatory reading.

Part 3 is the most voluminous part of the book. It discusses a number of attempts and experiences with possible alternatives to traditional rules and regulations. The first two contributions are from medical care and are about the use of checklists in operation rooms -what I’m missing here is the nuance that a (good) checklist in fact is little more than a very smartly written procedure - and about crew resource management in hospitals. After that we get two little exciting articles about safety management (systems) and their influence on procedures in air traffic and railways.

Much more exciting are no less than three articles about regulations and safety culture with a background in Norwegian offshore where elements of safety culture are implemented in laws and regulations over the last decades. This phenomenon is regarded with a critical eye amd among other things the authors discuss how far it is possible to steer and regulate safety culture. Especially chapter 13 is interesting with its argumentation why regulators should leave safety culture alone and rather stick to rules.

The final part contains only two articles of which the first deals with the paradox role that procedures apparently have in safety management, because they on one side are a means to transfer knowledge and prescribe a safe way of doing a job, but on the other hand they may also cause rigidity and ‘auto pilot behaviour’. The final article by Todd La Porte has some interesting view on the differences between administrative and operational safety rules. Of special interest is the anecdote about the aborted exercise on an aircraft carrier and the praise that the “culprit” (the sailor who had made a mistake, but reported this immediately upon discovery) received from the highest ranking officer.

In the epilogue both editors stress once more that rules are affected by the way they are created (preferably bottom-up, although also this is no guarantee for a good rule) and that one before creating a new rule must ask oneself the question what the added value of a rule will be. One of the leasons to be learned from this book.


Published by Ashgate (www.ashgate.com)

donderdag 24 april 2014

All Accidents Are Preventable?

Really?

I think for the statement "all accidents are preventable" to become REMOTELY true (ish) we will need to add some words. For example:

All accidents are preventable... with the benefit of hindsight.

or

All accidents are preventable... in theory.

or

All accidents are preventable... given unlimited knowledge, resources, perfect prediction (and quite some luck).

All of which, regrettably makes it a rather useless statment in my everyday job. And besides, do we really want to prevent absolutely everything? Really??

dinsdag 1 april 2014

Gerd Gigerenzer - Gut Feelings - Short Cuts To Better Decision Making

This book was on my “things to do” list for quite a while (triggered by Gladwell’s “Blink” and Gigerenzer’s “Risk” book) and went substantially up to the top of the list after “Antifragile” and Taleb’s promotion of heuristics. It’s very compact, almost 230 sides of easy to read (and often even slightly humorous) text.

The first part of the book is titled “Unconscious Intelligence”. We think of intelligence as deliberate, conscious activities guided by the laws of logic. Much of our mental activities, however, are unconscious and driven by gut feelings and intuitions without involvement of formal logic. Our mind often relies on the unconscious and uses rules of thumb in order to adapt and economize. One reason it needs to do so is because there is only so much information we can digest at one time.

One advantage is that these simple rules are less prone to estimation and calculation error and are intuitively transparent as opposed to complex models. So, an intuitive shortcut, or heuristic, often gets us where we want without a smaller chance of big errors, and with less effort. Gigerenzer says therefore that it’s not a question if but when we can trust our intuitions.

One example for this that an experienced chess player will usually generate the best solution first and he will not do better with more time to reflect and reconsider - rather in contrary. Inexperienced players on the other hand will most of the time benefit from long deliberation. So, stop thinking when you are skilled. Thinking too much about processes we master (expertly) will usually slow down or disrupt performance as everyone who has tried to think about going down the stairs can confirm. These things run better outside our conscious awareness, so more is not always better.

Heuristics try to react on the most important information, ignore the rest and lead to fast action. Heuristics are a result of evolved capacities of our brain: simple rules are developed over time and thanks to practice we are able to perform an action really quickly and well - effective and efficient. The selection of the applicable rules is unconscious. Gigerenzer thus views our mind as an adaptive toolbox with rules of thumb that can be transferred culturally or genetically and also developed by ourselves, or adapted from existing rules.

One function of intuition is also to help us master one of our main challenges: we don’t have all the information, so we have to go beyond the information that is given to us. Out brain often “sees” more than our eyes by “inventing” things in addition to what in fact is seen, like depth (a ‘third dimension’) in a (2D) drawing.

If a gut feeling will have the wanted/correct outcome depends upon the context it’s used in. So a heuristic is neither good nor bad, this depends upon the environmental structures. Selection of rules of thumbs can be triggered by the environment (‘automatic rules’) or selected after a quick evaluation; conscious or (often) not. If the first chosen of the latter category ‘flexible rules’ doesn’t work another is selected. Gut feelings may seem simplistic, but their underlying intelligence is selecting the right rule of thumb for the right situation (depending on circumstances, environment, etc).

It’s perceived wisdom that complex problems demand complex solutions. In fact, however, in unpredictable situations the opposite is true. As things are, our world has limited predictability. Keeping that in mind one may consider that we should spend less resources and money on making complex predictions (and on consultants who make them for us). In hindsight a lot of information may help to explain things and it’s easy to fit information to past events. In order to predict the future, however, much of the information one gets is not helpful in predicting and thus it’s important to ignore the information that is not of value. The art of intuition is to ignore everything except the best clue that has a good chance on hitting that useful information. Psychological research suggests that people often (but not always!) base intuitive judgments on one single good reason. (do check pages 150 and 151!).

The final chapter of part one discusses intuition and logic. It starts with the famous ‘Linda problem’. Criticizing Kahneman et.al. Gigerenzer argues that calling the intuitive solution of most people a fallacy is not correct because gut feelings (or humans in general) are not governed by the rules of mathematical logic. He says that the human brain has to operate in a uncertain world, not in the artificial certainty of a logical system. Our brain has to make sense of information given and go beyond it. It zooms in on certain parts that seem particularly relevant within the context and it views words (like ‘probable’) in their common use and conversational meaning rather than in a formal academic/logical sense. Gigerenzer sees heuristics as a way to success rather than as a cause for error.

I find that there are arguments for both ‘sides’. Humans are bad with probabilities and numbers (something which Gigerenzer addresses in his other book, by the way) but humans don’t operate necessarily to the rules of formal logic - as anyone can confirm who has seen humans in action. An interesting thing to think about and keep in the back of your mind.

Gigerenzer concludes that logical arguments may conflict with intuition, but that intuition is often the better guide in the real world. Nevertheless many psychologists treat formal logic as the basis of cognition and many economists use logic as the basis for ‘rational’ action. This isn’t how the world works, however, and logic is blind to content and culture and it ignores environmental structures and evolved capacities. Gigerenzer closes by stating that good intuition must go beyond information given and therefore beyond logic.

Part two is called “Gut feelings in action” and discusses a couple of real-life examples of intuitions. It starts with the functions of recognition (a very strong memory function of our brain) and recall (not so strong, especially with age). Recognition helps us to separate the old from the new. The recognition heuristic can help us making intuitive judgments, like when someone who knows nothing about football is asked what team will win, he will probably pick the best known, and more often than not be right. This means that in some cases ignorance even can be beneficial because more knowledge may mean that one cannot rely on this heuristic anymore. One remarkable fact is that during tests people who relied on the recognition heuristic made snap decisions which appeared to impress people with greater knowledge who needed time to reflect.

The recognition heuristic is an example of a flexible rule which is chosen by our unconscious intelligence after an evaluation process. It can be overridden consciously in several ways. Another example of ‘one reason’ decisions is the way that political preferences are commonly ranged on a left-right scale, even if the subject at hand is totally unrelated to the left-right opposite.

Not always do we rely on just one reason, often we make intuitive judgments based on evaluation of a sequence of cues that are evaluated, but again only one determines the final decision - so called sequential decision making. We go through a series of cues (most important first, second next, etc) and evaluate options. As long as options ‘score equal’ we continue evaluating, but at the first cue where one option is best we stop. We don’t evaluate all pros and cons to find the optimal solution rather we choose the ‘first best’. Sequential decision making based on the first good reason is very efficient, transparent and often more robust and accurate than complex models.

One tool for sequential decision making is a ‘fast and frugal tree’ which through a couple of yes/no questions leads to a quick decision, rather than working through a huge, complex complete decision tree. Fast and frugal trees have three building blocks: 1) a search rule that looks up factors in order of importance, 2) a stop rule that stops looking further if a factor allows so, and 3) decision rule that classifies an object. Among others medical services use these ‘developed intuitions’ for making diagnoses. These simple, transparent empirically informed rules help making better decisions.

The last two chapters deal with moral behavior and social instincts (e.g. imitation and trust). People do morally unbelievable things (like the example of a WW II mass murder illustrates) because of their reluctance to break group order. Peer group pressure (consciously or not) is enormous and it may even overrule deeply rooted moral instincts like “You shall not kill”.

One important heuristic to understand is that people will usually opt for the default (chose by the environment, or ‘system’) instead of making a conscious choice - as is illustrated by the differences in percentages of organ donors between various countries. By understanding the process and framing instructions or requests well, one may be able to steer things in a desired (and preferably morally just) direction.

This summary/review obviously only scratches the surface of themes treated in this most recommended book. Hope I tickled your interest, now go and read it for yourself. 

I’ve read the Penguin pocket version ISBN 978-0-141-01591-0

donderdag 13 maart 2014

Antifragile - a review (or two, sort of)

In the picture below you'll find a scan of the review I wrote of Taleb's latest book, Antifragile, for the magazine of the Dutch Society of Safety Experts, NVVK Info and was published there in this month's issue.


The English translation goes roughly like this:

Antifragile - Things That Gain from Disorder, by Nassim Nicholas Taleb 
(ISBN 9780141038223)

While I was reading this book - during the days around Christmas 2013 - a major storm ravaged northern and western Europe. At one point my wife asked “What is that book about, anyway” and the perfect illustration of my answer presented itself in all newspapers the next morning: hundreds of thousands of households in Norway, Sweden, The Netherlands, Germany, France and The United Kingdom were without electricity. How much easier could you want for an example to explain the concept of a fragile system? Some systems are so much optimalised that they work smoothly in the normal situation (in Taleb’s vocabulary ‘mediocristan’), but as soon as an event of certain impact (unexpectedly, or dismissed as not probable) manifests itself, the vulnerable system collapses as a house of cards. The system is fragile and breaks, often with catastrophic consequences.

Those who have read Taleb’s last book, the bestseller and masterpiece “The Black Swan”, will notice that “Antifragile” picks up things where Taleb stopped the last time. The new book is the closing part of a trilogy that started about ten years ago with “Fooled By Randomness”. In “Antifragile” Taleb tries primarily to provide some strategies that enable us to cope with the fragility of systems - on one hand by assessing if something is fragile, or is made fragile by decisions that are made, and on the other side by discussing possibilities to make things (more) antifragile.

Taleb use a triad, or three different states, to illustrate the point. The first state is fragile: things break as a consequence of shocks, disorder or time. One example is an e-book that depends if your reader has enough power. The second state is robust: things are insensible to shocks (until a certain level), like an ordinary book that usually “works”. Finally there’s antifragile: things that do profit from disorder, like an oral delivery which goes from mouth to mouth and often only wins fantasy. For some this may seem strange; after all in common understanding it’s robust that is used as the opposite of fragile, but Taleb aims at another kind of opposite: not the difference between braking and not breaking, but between being harmed by and benefiting from something.

It’s likely that the reader has to get used to Taleb’s use of language and his jargon. He had a couple of specific “own” words (for example ‘fragalista’ which he uses as a label for people who advance fragility - often named, like the president of the Federal Reserve) or terms which he uses to characterize a certain phenomenon (“soccer mom”). This also applies to his style of writing. Taleb likes to stroll, and it shows in his books. This is no scientific academic report or paper (luckily it isn’t!) but a continuous discussion and reflection where the author switches from one subject to another. Sometimes he returns to a subject, sometimes he doesn’t. The contents may vary from seriously technical or mathematical stuff to fantastic or just every day reflections.

The language is flowery and at times outrageous. Taleb is erudite with a preference for the Classics and we’ll know this! All of this is coupled to an enormous waywardness and a certain arrogance (he doesn’t shy back of kicking seriously against the established order within among others economy en science). This may be a certain threshold for some readers. With that in mind it’s a relief that one doesn’t need to read everything. Some parts of the book can be skipped without any problem and without missing part of the message. In fact, on various occasion the author even indicates that a part can be skipped.

One may question how relevant this book is for safety professionals since there is only a limited discussion of ‘real’ safety subjects in the book (Fukushima as one example of fragility is a rare exception). My opnion: definitely relevant. Taleb’s books are about uncertainty and risk, about the thinking about risk and about decision making (with all kinds of consequences). This alone makes the book (and definitely its predecessor) almost mandatory reading.

Do note that this is by no means a ‘Do Book’ that will lead you to an antifragile state (presuming that this is possible and desirable) within 7 Steps. Reading the book can, however, help to create awareness and start a process. To this end it is necessary that each reader picks out relevant elements for himself and translates those to his own every day practice (domain blindness is one returning theme, by the way). Examples are fragility enhancing things like naive interventionism (the urge to implement measures that are likely to be unnecessary which may cause unwanted side-effects that in fact cause a worse situation) and agency problems (decisions taken or advocated by people who will profit in some way from that decision).
On the positive side Taleb discusses a number of tools that may help to enhance robustness or even gain antifragility. To name but a few: the principle of redundancy (preferably applied aggressively because in a time of scarcity one can gain a lot from your own abundance) even though it’s despised by economists and bean counters, optionality (making use of opportunities when they present themselves), flexibility, using heuristics (practical mental rules of thumb that make the everyday life easier) and allowing some hormensis in our lives and organizations (some disorder and stress is necessary to make things stronger, like not always washing your hands before diner can enforce your immune system and is thus healthy...).


One other thought that struck me while reading the book is that Taleb’s fragile/antifragile thinking unwillingly manages to combine two safety theories that are often regarded as opposites within the same framework. I’m obviously talking about Perrow’s Normal Accident Theory (fragile, Black Swans) and the antifragile concepts of High Reliability Organisations (Weick & co) and Resilience (Hollnagel, Woods etc.). For safety professionals it may be confusing that Taleb on several occasions says that antifragile and resilience (to him synonymous to robust) aren’t the same thing, while elements that “we” (safety professionals) range within the resilience framework are seen by Taleb as antifragile elements. A matter of definition, I think, and don’t let that stand in the way of gathering some useful lessons from this book.

woensdag 26 februari 2014

Black Swan workshop 20 February 2014

Petromaks - Black Swan Workshop,
Stavanger, 20 February 2014

This workshop was part of a new (?) project that aims to develop concepts, approaches and methods for proper understanding of risk for Norwegian petroleum activities, with due attention to the knowledge dimension and surprises. And so over 100 participants gathered in Stavanger, coming from mainly oil and gas companies, universities, some students and a few ‘outsiders’ as well.

Chairman of the day, Terje Aven, opened the day, introducing the research program and the goal of the day: to have an exchange of ideas about the Black Swan concept and how to meet the challenges they pose. Knowledge is one of the central elements because we need to know where to put resources. Not in finding solutions right now, but having guidance on the direction.

Aven defined a Black Swan as “a surprising extreme event relative to present knowledge or beliefs”. This means that a Black Swan can be a surprise for some, but not for others, it’s knowledge dependent.
Three different types of Black Swans: 1) unknown unknowns, 2) unknown knowns, and 3) known but not believed to be likely because probability is judged to be too low.

Since knowledge is a key, Aven thought that risk assessment may have an important role in dealing with Black Swans - but we have to see beyond current risk assessment practices since traditional assessment practices may ignore the knowledge dimension about uncertainties and surprises.


A couple of approaches for improved risk assessment were presented which included RedTeaming (playing ‘devil’s advocate’), scenario analysis/system thinking, challenge analysis (providing arguments for events to occur), and anticipatory failure determination based on theory of inventive problem solving. Aven also addressed different types of risk problems caused by knowledge uncertainties which put weight on the use of the cautionary and precautionary principles.

The program before lunch featured four academic speakers. Andrew Stirling from the University of Sussex was the first of these. He warned that, since Black Swans are not objective, one should not try to bury subjectivity under analysis. An argument in favour of precaution followed and the interesting observation that defence of scientific rationality (against application of the precautionary principle) is often surprisingly emotional.

According to Stirling, uncertainty requires deliberation about action - you cannot analyze your way out of it. Deliberation will produce more robust knowledge than probabilistic analysis. Some examples were used to illustrate that evidence-based research often includes so large uncertainties that they often can be used as argument for pretty much any decision.

A matrix was presented in which knowledge about likelihood and knowledge about possibilities were presented against an problematic/unproblematic scale. As was demonstrated risk assessment is (according to the definition of Knight) only applicable in one quadrant (good knowledge about both possibilities and likelihood), yet we are forced to use this instrument a.o. through regulations and political forces in a desire for ‘evidence based policies’. The other quadrants were labeled ‘uncertainty’ (’risk’ with bad knowledge on likehood), ‘ambiguity’ (bad knowledge about possibilities, but good about likehood) and ‘ignorance’ (bad knowledge about both). A number of tools were proposed to get out of the “risk corner” and have a wider view. Tools included the ones that Aven had mentioned before. One way is to be more humble and not be caught in ‘swan-alysis’…


Concluding Stirling said that the point is not putting things in boxes (e.g. what type of Black Swan we’re dealing with) but rather what to do things for. Critical deliberation is more important than analysis. One problem with the Black Swan metaphor may be the suggestion of white (= good, many of those) and black (= bad, only a few - so we’re good), but things are definitely not that binary!!

Ortwin Renn had an inspiring Powerpoint-free speech about different Black Swans, the role of (risk) analysis and risk management (methods). Some Black Swans are in the long tails of the normal distribution, others are about problems in knowledge or knowledge transfer. There often is a disparity of knowledge within or between companies - transfer of knowledge may be beneficial. And there are the “real” surprises, the rare events which have never been seen, or there is no pattern that could predict them.

One reason for the popularity of the Black Swan is that we humans experience many random unique events in the course of our lives. Our memory builds on unique events, not on routine. But… risk assessment works the other way… and builds on a very formal approach.

How to deal with these challenges? We can’t say about probability of Black Swan events, but we can say something about the vulnerability of our system! This we can do by analysis, but a different kind of analysis.
Resilience is about the ability of a system to cope with stress without losing its function. Highly efficient systems are usually NOT resilient. There is an unavoidable trade-off between efficiency and resilience. This trade-off is not a mathematical decision. It’s about risk appetite, compensation, politics, our attitude to precaution. It’s an open question we need to deliberate about!

ESRA chairman Enrico Zio had a presentation that reflected his search for answers around the theme of uncertainty. We try to predict by modeling, but there is a difference between ‘real risk’, assumed risk and expected protection. So there is a multitude of risk profiles and various analyses don’t give the same answers. One solution might be to combine deterministic and probabilistic safety assessments.


Zio continued by discussing two different approaches to safety which Hollnagel calls Safety I (look at things that go wrong) and Safety II (look at things that go well - resilience thinking). Improving the ration between Safety I (decrease) and Safety II (increase) may be one way to decrease Black Swan risks. Observability and controllability are two important issues related to Safety II.

Concluding Zio warned about trying to solve today’s (complex) problems with the mindset of yesterday.

After a number of critical remarks with regard to knowledge and probability Seth Guikema from John Hopkins University talked about why we cannot discard historical data and expertise. Guikema underlined that risk assessment helps to understand problems, but does not make the decisions. People do!

Historical data can be useful as it says something about the frequency and extent of the events that have happened, but it cannot give information about eventual Black Swans in the future. While you cannot determine the inputs themselves, you can use models from things that have happened and run these with BIG events and see how the system responds to these. Guikema illustrated this with example from the impact of historical hurricanes on power grids in the USA and how the recent hurricane that ravaged the Philippines would affect the USA. This variation on ‘red teaming’ proves to be a useful way to assess vulnerability. One ironic comment was that people tend to forget the lessons learned from previous storms.

Again: data driven models cannot help initiating events, but can be used to assess the impact and more. An iterative ongoing process was presented as shown in the picture below.


After lunch there was possibility of participants to present views or discuss practical examples. Regrettably only a few used this opportunity - despite the rather large number of participants.

First up was Conoco Phillips who presented two different cases. Kjell Sandve discussed some challenges within risk management especially the question if increasing costs were ‘caused’ by HSE requirements. One main problem according to Sandve was a challenge to reduce requirements once they had been applied one place (but maybe weren’t quite necessary everywhere). He asked also if we actually have the right tools and if risk assessment supports good decisions. His appeal to the project was: Don’t make more complex tools/methods - rather the opposite!

Malene Sandøy starting talking about a 1970s’ decision problem, i.c. how high should jackets for facilities be build and the Black Swan that one met some years later when it turned out that the seabed subsided and platforms were “sinking”. Related Black Swans were stricter regulations with regard to wave height, new design requirements and not in the least a much longer life time for the structure than originally anticipated.
After a rather technical discussion of design loads and how these analyses could be used Sandøy ended up with a clear improvement potential: From calculations (analyses) that “no one” understands to broad discussion in the organization of scenarios to design for.

Third speaker during the audience participation was yours truly who brought some views from outside the oil & gas breaking through some barriers of domain dependence. The presentation included a retrospective on some Black swans that have affected Norwegian Railways in the past years, both negative (the 22/7 terror attack) and positive (the national reaction to 22/7, the ash cloud and, interestingly, the financial crisis). Safety related accidents are rarely Black Swans, but as Stirling said one shouldn’t be too stuck up with putting things in boxes and a more ‘relaxed’ approach to the definition will give several ‘Dark Fowl’ events. One example was the Sjursøya accident in March 2010 which was discussed. A quick assessment of the system where the accident originated (the Alnabru freight terminal) leads to the conclusion that this was a Fragile system. Measures taken after the accident were discussed related to the Fragile - Robust - Antifragile triad from Taleb’s latest book.


Ullrika Sahlin from Lund University came from an environmental background which was interesting because she related Black Swans not so much to events (as safety folks tend to do), but rather to continuous exposure from certain processes. Sahlin presented a couple of thoughts about the subject and expectations from the project.

Her presentation included discussions around various perspectives on risk (a.o. traditional statistical and Bayesian), the quality of our knowledge and the processes we use to produce the knowledge we use for evidence based management and not in the least assumptions (to which one member from the audience remarked that “If you assume, you make an ASS of U and ME”).

One advice from Sahlin was that we should communicate our estimates with humility and communicate our uncertainties with confidence.


Igor Kozine from the University of Danmark was the last participant who had a relatively spontaneous short presentation around a 2003 Financial Times article that told about a Pentagon initiative to have people placing bets on anticipated terror attacks as a way of risk assessment. This project was discontinued (at least officially) because of public outrage.

After a coffee break with engaged mingling and discussions there was a concluding session with discussion and reflections. Themes that came along included:
  • In line with Safety I/II: should we focus on failures or on success? Is focusing on success a way to manage Black Swans? Regulators may have a problem with those approaches… Rather than either/or one should compare alternative strategies. There aren’t too many error recovery studies yet.
  • What about resilience on a personal level? How much must one know? Dilemma between compliance and using your head in situations that are not described in the rules - see Piper Alpha.
  • Taleb’s Antifragility and Resilience: safety people may have a different understanding about Resilience than Taleb, and (according to Rasmus from the University from Danmark) also original literature differentiates between Robust (bounces back to normal) and Resilient (bounces back to a NEW normal).
  • Stirling summarized that one important dilemma is the question: What is the analysis for? Shall it help making a decision, or shall it describe knowledge about some risks? One solution may be to look at the recipient of the analysis, not just at the object to be analyzed. What are the power structures that are working on us?
  • The last question was if we really need new methods, or rather a good way to use existing ones and their output? Aven concluded that risk assessment in the broad sense of the term has a role to play. Knowledge is important as is the communication of knowledge.

vrijdag 17 januari 2014

Ulrich Beck - Weltrisikogesellschaft

(ISBN 978-3-518-41425-5)

This is definitely one of the most difficult to read books I’ve encountered in the past years, due to the way it’s written. As it’s in German (and I’m not sure if this particular version has an English translation) this may be a bridge too far for some anyway, but even for people who are used to reading both German and academic papers this one may be prove challenging. Anyway, it was for me and I considered a few times to just give up, but I decided to persist and at least try to squeeze the main essence from it. This wasn’t easy due to the flowery and circular language that is frequented by words that common mortals will avoid at any cost in every day conversations (or writing, for that matter). It’s probably a typical example of a book written by an academic for other academics…

Okay, having given that warning let’s proceed to the contents. The book is a follow-up of his 1986 classic “Risikogesellschaft” and expands on the ideas presented there and an English book from 1999. It deals with risk on a global level and how is dealt with it, or should be dealt with is. Pretty much of the discussion is fairly philosophical in character and grounded in sociological theories (Beck is a Professor in Sociology, after all) which I’m not all that familiar with, so that may very well be one reason that much of the writing doesn’t appeal all that much to me. If I try to summarize some of the most important points these would include:
  • The “success” of modern society - and not its failure - has created some major risks (often as unintended side effects). These include economical/financial, ecological/environmental and terrorism, and related, some kinds of warfare.
  • These major risks have impact that transcend national borders and often have global impact (hence the title of the book: World Risk Society).
  • Trying to solve these problems/risks with common approaches on national levels is futile (trying to have an individual environment-friendly life-style is nice, but as long as not all the major countries change course rather pointless).
  • Risks are no longer dealt with only by governments and regulators, but also by bottom-up transnational organisations (e.g. Greenpeace) or top-down supranational organisations. Or even by ad hoc cooperation of for example customers who decide to boycott a certain brand and thereby effectively cause a change in policy.
  • The effects of risks aren’t necessary dependable on the fact that anticipated consequences actually materialize; if presented “well” the possibility becomes the reality which is best illustrated by the experiences of everyone who tries to travel by plane after 9/11.
  • The perception of risk does not necessarily depend on ‘scientific’ analysis and rationality, but also on presentation, knowledge and ignorance, information and disinformation, culture and religion.
  • And let’s not forget the role of massmedia.
  • The distribution of risk is in many cases asymmetrical and often in favour of the decision maker, but not necessarily so - many global risks will affect everyone in a most democratic manner. Even so, risk export happens. Often.
This only scratches the surface, of course, but that’s what I took out of it and remembered after a first difficult read through it. And I should of course mention one brilliant quote that every safety pro should take note of:
Risiko Ergo Sum.

As a whole an interesting, but very demanding, book which gives a slightly different view on risk from a totally different perspective than most safety professionals usually deal with. Some clear links to Taleb’s work too. But I’d say start with Taleb and you will have covered a good deal of Beck’s points too, in a (IMHO) more pleasantly readable way. The Wikipedia pages on Beck and his work are very informative, by the way.