The Epistemic Value of Cautionary Tales
Twice in NASA history, the agency embarked on a slippery slope that resulted in catastrophe. Each decision, taken by itself, seemed correct, routine, and indeed, insignificant and unremarkable. Yet in retrospect, the cumulative effect was stunning. In both pre-accident periods, events unfolded over...
Gespeichert in:
Veröffentlicht in: | The Journal of technology studies 2006-04, Vol.32 (2), p.90-94 |
---|---|
1. Verfasser: | |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Twice in NASA history, the agency embarked on a slippery slope that resulted in
catastrophe. Each decision, taken by itself, seemed correct, routine, and indeed,
insignificant and unremarkable. Yet in retrospect, the cumulative effect was stunning.
In both pre-accident periods, events unfolded over a long time and in small increments
rather than in sudden and dramatic occurrences. NASA’s challenge is to design systems
that maximize the clarity of signals, amplify weak signals so they can be tracked, and
account for missing signals. For both accidents there were moments when management
definitions of risk might have been reversed were it not for the many missing signals –
an absence of trend analysis, imagery data not obtained, concerns not voiced,
information overlooked or dropped from briefings. A safety team must have equal and
independent representation so that managers are not again lulled into complacency by
shifting definitions of risk . . . Because ill-structured problems are less visible and
therefore invite the normalization of deviance, they may be the most risky of all. –
Vol. I, Section 8.5, Report of the Columbia Accident Investigation Board (August
2003). |
---|---|
ISSN: | 1071-6084 1541-9258 |
DOI: | 10.21061/jots.v32i2.a.4 |