[SystemSafety] NTSB report on Boeing 787 APU battery fire at Boston Logan
Matthew Squair
mattsquair at gmail.com
Sun Dec 7 07:01:17 CET 2014
John Downer (whose on the list I think) coined the phrase 'epistemic
accident' to cover accidents which are due to our knowledge being
contingent as much on on theories, and assumptions, as facts. Said theories
which may then prove to be not quite good enough. Apologies if I mangle the
definition a bit.
I also noticed the NTSB homed in on the need to surface assumptions and
make them explicit. Assumptions of course being the epistemic equivalent of
whoopy cushions in engineering. :)
Matthew Squair
MIEAust, CPEng
Mob: +61 488770655
Email; Mattsquair at gmail.com
Web: http://criticaluncertainties.com
Matthew Squair
MIEAust, CPEng
Mob: +61 488770655
Email; Mattsquair at gmail.com
Web: http://criticaluncertainties.com
On 6 Dec 2014, at 2:30 am, Mike Ellims <michael.ellims at tesco.net> wrote:
In the Guardian Gawande states..
" There was an essay that I read two decades ago that I think has influenced
almost every bit of writing and research I've done ever since. It was by two
philosophers - Samuel Gorovitz and Alasdair MacIntyre - and their subject
was the nature of human fallibility. They wondered why human beings fail at
anything that we set out to do. Why, for example, would a meteorologist fail
to correctly predict where a hurricane was going to make landfall, or why
might a doctor fail to figure out what was going on inside my son and fix
it? They argued that there are two primary reasons why we might fail. The
first is ignorance: we have only a limited understanding of all of the
relevant physical laws and conditions that apply to any given problem or
circumstance. The second reason, however, they called "ineptitude", meaning
that the knowledge exists but an individual or a group of individuals fail
to apply that knowledge correctly."
However I think that Gorovitz and MacIntyre argue something very different,
the following is I believe the essence of their argument. I have edited it
because the paper is very long and not the easiest of reads.
{First they discuss where our traditional views of error come from i.e.
the natural sciences}
For on this view all scientific error will arise either from the
limitations of the present state of natural science-that is, from
ignorance or from the willfulness or negligence of the natural
scientist-that is, from ineptitude. This classification is treated as
exhaustive.
<snip>
This view of ignorance and ineptitude as the only sources of error has been
transmitted from the pure to the applied sciences, and hence, more
specifically, from medical science to medical practice viewed as the
application of what is learned by medical science.
<snip>
{they then go on to look at the issue that doctors - and engineers face
dealing }
{ with particular situations, EMPHISIS ADDED below }
Precisely because our understanding and expectations of particulars
cannot be fully spelled out merely in terms of law like
generalizations and initial conditions, the best possible judgment may
always turn out to be erroneous, and erroneous not merely because
our science has not yet progressed far enough or because the
scientist has been either willful or negligent, but because of the
necessary fallibility of our knowledge of particulars.
<snip>
The recognition of this element of necessary fallibility IMMEDIATELY
DISPOSES OF THAT TWOFOLD CLASSIFICATION of the sources of error which
we have seen both to inform natural scientists' understanding of
their own practices and to be rooted in the epistemology that
underlies that understanding. Error may indeed arise from the
present state of scientific ignorance or from willfulness or
negligence. But it may also arise precisely from this third factor,
which we have called necessary fallibility in respect to particulars.
-----Original Message-----
From: systemsafety-bounces at lists.techfak.uni-bielefeld.de
[mailto:systemsafety-bounces at lists.techfak.uni-bielefeld.de
<systemsafety-bounces at lists.techfak.uni-bielefeld.de>] On Behalf Of
Peter Bernard Ladkin
Sent: 05 December 2014 11:58
To: systemsafety at lists.techfak.uni-bielefeld.de
Subject: Re: [SystemSafety] NTSB report on Boeing 787 APU battery fire at
Boston Logan
On 2014-12-05 12:36 , Martin Lloyd wrote:
On 05/12/2014 10:52, Mike Ellims wrote:
Interestingly research suggests surgeons who expect things to go
wrong and plan for failure have much higher success rates.
Does anyone have a reference to these research results?
Atul Gawande is giving the Reith Lectures at the moment on a closely related
topic, namely how to improve the success rate of/avoid avoidable failures in
medicine. A summary of the first is
http://www.theguardian.com/news/2014/dec/02/-sp-why-doctors-fail-reith-lectu
re-atul-gawande The BBC page is
http://www.bbc.co.uk/programmes/articles/6F2X8TpsxrJpnsq82hggHW/dr-atul-gawa
nde-2014-reith-lectures
PBL
Prof. Peter Bernard Ladkin, Faculty of Technology, University of Bielefeld,
33594 Bielefeld, Germany
Tel+msg +49 (0)521 880 7319 www.rvs.uni-bielefeld.de
_______________________________________________
The System Safety Mailing List
systemsafety at TechFak.Uni-Bielefeld.DE
---
This email is free from viruses and malware because avast! Antivirus
protection is active.
http://www.avast.com
_______________________________________________
The System Safety Mailing List
systemsafety at TechFak.Uni-Bielefeld.DE
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.techfak.uni-bielefeld.de/mailman/private/systemsafety/attachments/20141207/57308f77/attachment.html>
More information about the systemsafety
mailing list