Lecture 10: Aviation System Safety Analysis

0.0(0)
Studied by 0 people
call kaiCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/38

encourage image

There's no tags or description

Looks like no tags are added yet.

Last updated 1:25 AM on 4/11/26
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No analytics yet

Send a link to your students to track their progress

39 Terms

1
New cards

Aircraft Systems Overview

  • System: A collection of components intended to function together to perform a task.

  • An aircraft is a system comprising structure, propulsion, controls, etc. to carry people or cargo from A to B.

  • Systems are usually composed of sub-systems (and eventually, components)

  • Flight controls (actuators), hydraulics (pumps), electrical power generation (circuit breakers, connectors), landing gear (brakes, doors), Pitot-static (heated pitot tube, air data computer)

  • All of these are subsystems of the aircraft system

Systems interact with each other; system definitions must include the interfaces.

2
New cards

Essential/Mission Systems

  • For airworthiness - consider only systems essential for flight.

    • Flight controls, propulsion, navigation, etc.

  • Mission systems - required for the intended function of the aircraft but not essential for safe flight

    • In flight entertainment (IFE), weapons, research sensors, NRC helicopter research fly-by-wire flight control systems

Mission systems must at least be assessed to ensure they “do no harm” to essential systems nor introduce new hazards!

3
New cards

Systems Context - Simple v. Complex

  • This presentation mainly intended to discuss sophisticated, complex aircraft

  • Many cert rules are based on basic lessons learned long ago on small, simple aircraft, and they still apply because there has been no change to the technology

  • The need for predictable handling qualities such as prescribed “stick force gradient”

4
New cards

Case: Turkish Airlines Flight 1951 B737 Amsterdam - 25 Feb 2009

  • Incorrect autopilot mode selection and one faulty radio altimeter (undetected) caused the auto-throttles to decrease thrust to idle early approach (flare for landing)

  • The crew noticed low airspeed too late to increase thrust and prevent the aircraft from aerodynamic stall

  • Crashed 1.5 km north of runway 18R killing 9 passengers and crew including all 3 pilots

Crew error due to lack of “automation mode awareness” is a system safety issue.

5
New cards

Boeing 737 Max 8

Lion Air 610, Indonesia Oct 29, 2018

  • Dove into sea 13 minutes after T-O, 189 fatalities

  • Newly delivered airplane (2 mths, 800 hrs in service)

  • Experienced crew

Ethiopian Airlines 302, Ethiopia March 10, 2019

  • Dove into ground 6 minutes after T-O, 157 fatalities

  • Also a new airplane (4 months in service)

  • Experienced captain, inexperienced first officer

Single failure in a system caused the loss of 346 lives and 2 aircraft.

6
New cards

System Safety Assessment - General

  • Much of system safety analysis is imagining how systems can fail

  • Need to analyze potential failures in a rigorous and systematic way

  • Must include potential human factor errors (due to maintenance and flight crew actions)

  • Think about possible cascading failures

7
New cards

System Safety Assessment - “What If?”

The airworthiness mindset is to always ask: What can possibly go wrong?

  • Failure Mode (or Failure Condition) - In what ways can this system fail to function as intended; or function when it should not!

  • Failure Effect - What are the consequences of each failure condition?

  • Probability - How often will each failure condition occur?

  • Mitigation - What (if anything) should be done about them?

8
New cards

Safety/Risk

  • The level of safety is the inverse of the level of risk

  • For any given hazard there is:

    • Some probability that the hazard will result in an unsafe event

      • For components & systems “failure probability” is the inverse of “reliability”

    • Some level of severity associated with the unsafe event

    • Risk is the product of these two

RISK = severity x probability

9
New cards

Regulations = “Thirteen-O-Nine” pt a

CS 25.1309 Equipment, systems, and installations

(a) The aeroplane equipment and systems must be designed and installed so that:

(1) Those required for type certification or by operating rules, or whose improper functioning would reduce safety, perform as intended under the aeroplane operating and environmental conditions.

(2) Other equipment and systems are not a source of danger in themselves and do not adversely affect the proper functioning of those covered by subpara a1

TL;DR:

  • Required equipment must function as intended under all conditions - not necessarily the same as “function as designed”

  • Non-required systems must “do not harm”

10
New cards

Regulations = “Thirteen-O-Nine” pt b

CS 25.1309 Equipment, systems, and installations

(b) The aeroplane systems and associated components, considered separately and in relation to other systems, must be designed so that -

(1) Any catastrophic failure condition

(i) is extremely improbable; and

(ii) does not result from a single failure; and

(2) Any hazardous failure condition is extremely remote; and

(3) Any major failure condition is remote

TL;DR:

  • No single failure shall lead to a catastrophic condition

  • Failure conditions of increasing severity must be increasingly unlikely

11
New cards

Regulations = “Thirteen-O-Nine” pt c

CS 25.1309 Equipment, systems, and installations

(c) Information concerning unsafe system operating conditions must be provided to the crew to enable them to take appropriate corrective action. A warning indication must be provided if immediate corrective action is required.

Systems and controls, including indications and annunciations must be designed to minimize crew errors, which would create additional hazards.

TL;DR:

  • Crew must be informed/annunciations for warnings and cautions

  • Design to ensure crew takes the correct action in response

Examples: consistent use of colours (red = warning, amber = caution)

12
New cards

Regulations = “Thirteen-O-Nine” pt d-f

CS 25.1309 Equipment, systems, and installations

(d) Compliance with the requirements of para b must be shown by analysis, and where necessary, by appropriate ground, flight, or simulator tests. The analysis must consider:

(1) Possible modes of failure, including malfunctions and damage from external sources.

(2) The probability of multiple failures and undetected failures.

(3) The resulting effects on the aeroplane and occupants, considering the stage of flight and operating conditions, and

(4) The crew warning cues, corrective action required, and the capability of detecting faults.

(e) In showing compliance with paragraphs (a) and (b) of this section with regard to the electrical system and equipment design and installation, critical environmental conditions must be considered. For electrical generation, distribution, and utilization equipment required by or used in complying with this manual, except […] the ability to provide continuous, safe service under foreseeable environmental conditions may be shown by environmental tests, design analysis, or reference to previous comparable service experience on other aircraft.

(f) Electrical Wiring Interconnection Systems (EWIS) must be assessed in accordance with the requirements of 525.1709

TL;DR:

  • Analysis and/or testing is required - methods are not prescriptive but must be acceptable to the cert authority

  • Must consider external sources (birds, lightning, HIRF)

  • Consider environmental conditions

13
New cards

Probability and Severity

For increasing severity decreasing probability allowed

14
New cards

Qualitative Probability (for a fleet of airplanes of one type)

Probable failure conditions… anticipated to occur one or more times during the entire operational life of each aeroplane.

Remote… unlikely to occur to each aeroplane during its total life… may occur several times when considering the total operational life of a number aeroplanes of the type.

Extremely remote… not anticipated to occur to each aeroplane during its total life… may occur a few times in whole fleet.

Extremely improbable… not anticipated to occur during the entire operational life of all aeroplanes of one type.

15
New cards

Where do the quantitative values come from?

  • Historical serious accident rate 1 per million hours of flight (1 accident per 1×10^6 hours)

  • About 10% due to systems Failure Conditions (FC)

  • Safety target for new a/c: probability of a serious accident from all system FC < 10 ^-7 per flight hour (at least as good as histroical record)

  • Arbitrary assumption of 100 catastrophic FCs per airplane

Safety target for Catastrophic FC for any system is therefore < 1 × 10^-9 per flight hour.

16
New cards

Fail Safe Design Concept

  • Failure of any single system element should be assumed regardless of probability and must not be catastrophic

  • Subsequent failures during same flight, should also be assumed, unless their joint probability with the first failure is shown to be extremely improbable (< 10^-9 per hr)

  • Apply design principles to ensure acceptable LOS w probabilities no greater than:

    • REMOTE; <10^-5 per hour for MAJOR FCs

    • EXTREMELY REMOTE; <10^-7 per hour for HAZARDOUS FCs

    • EXTREMELY IMPROBABLE; < 10 ^-9 per hour for CATASTROPHIC FCs

17
New cards

Failsafe Design Principles pt I

Designed Integrity and Quality, including life limits

  • Ensure intended function and prevent failures (simple, robust, but may need…)

Redundancy or backup systems

  • to enable continued function after any single failure; eg two or more engines, hydraulic systems, flight control systems, etc. (adds complexity)

Isolation and/or segregation of systems, components, and elements

  • So failure of one does not cause the failure of another (ensure independence but requires space/weight)

Proven reliability

  • Well understood designs with known low failure rates/high reliability (slows innovation!)

18
New cards

Failsafe Design Principles pt II

Failure warning or indication

  • to provide detection (risks crew overload)

Flight crew procedures

  • Corrective action after failure detection (difficult to quantify human performance)

Checkability

  • ability to check a component’s condition (flight or maintenance crew workload)

Design failure effect limits

  • ability to sustain damage; limit safety impact (analysis and testing effort/cost)

19
New cards

Failsafe Design Principles pt III

Design failure path

  • to control and direct the effects of failure in a way that limits its safety impact (analysis and testing effort/cost)

Margins or factors of safety

  • To allow for undefined/unforeseeable adverse conditions (usually adds weight)

Error-tolerance

  • Consider effects of errors during design, test, manufacture, operation, and maintenance (requires strong human factors knowledge)

20
New cards

Simplified Process Overview

FHA’s ——Validation of Safety Targets———>

PSSA ——-Verification of Safety Targets——> SSA

SSA: System Safety Analysis

PSSA: Preliminary SSA
FHA: Functional Hazard Assessments

  • Aircraft and system levels

21
New cards

Definition - Validation

The determination that the requirements for a product are sufficiently correct and complete.

Top down from aircraft to system to sub-system to component.

22
New cards

Definition - Verification

The evaluation of an implementation to determine that applicable requirements are met.

Bottom up from component to sub-system to system.

23
New cards

General Heat Systems Configuration

How the (EPS) functions:

  • Pilot commands motor torque via “throttle” thrust lever

  • Pilot commands motor RPM via propeller control lever

  • Battery provides HV DC electricity

  • Motor controller takes HV DC and torque command and outputs HV AC to motor

  • Battery Management System (BMS) monitors battery and makes HV connection to motor controller

  • Displays and annunciators show status and warnings to pilot

24
New cards

Functional Hazard Assessment (FHA) At Aircraft and System Levels

  • High level, qualitative assessment of basic aircraft functions

  • Identify each Failure Condition (FC) and the rationale for its severity classification

  • System FHA iterative as systems evolve; consider interactions of systems

  • Generate top level events for subsequent system analysis using Fault Tree Analysis (FTA) during Preliminary System Safety Analysis (PSSA)

  • Derive Development Assurance Levels (DALs) for h/w and s/w based on system criticality

25
New cards

Clarification regarding back-up and redundant systems

A back-up system only operates when the primary systems cannot; it replaces the capability of the primary system

A redundant system consists of multiple components operating simultaneously; the failure of any one does not cause an overall system failure due to the capacity of the other components

26
New cards

Preliminary System Safety Analysis (PSSA)

  • Iterative analysis embedded in overall development process down to “item level”

  • Completes the FC list including those due to latent faults and common causes

  • Derives system & item safety requirements

  • Identify protective strategies such as built-in test, dissimilarity, monitoring, flight, or maintenance crew tasks/intervals

  • Include h/w and s/w errors to assist in establishing DALs

  • Define verification strategies to be used during SSA - ie qualitative vs quantitative and specific test/analysis methods to show compliance

27
New cards

P/SSA tools and methods from ARP4761

Fault Tree Analysis (FTA)

  • Boolean logic (AND, OR, NOT, etc.) as introduced in prior examples

Dependence Diagrams (DD)

  • Uses paths to depict the relationship of failure effects;

  • Parallel paths for AND gates; series paths for OR gates

Markov Analysis (MA)

  • Statistical method

Failure Modes and Effects Analysis (FMEA)

28
New cards

Top Down Methods: Fault Tree Analysis (FTA)/Dependence Diagrams (DD)

  • Quantify probability of occurrence for the top event vs. safety target

  • Establish reliability budgets and Development Assurance Levels (DALs) for hardware and software

  • Assess the impact of a design modification

  • Identify the need for a design modification and/or identify unique situations that require special attention

  • Show compliance with qualitative and/or quantitative safety objectives as part of the SSA

  • Establish crew/maintenance tasks and intervals necessary to meet the requirements of the safety assessment

29
New cards

Failure Modes and Effects Analysis (FMEA) - Bottom Up Method

  • FMEA is a systematic bottom up method of identifying failure modes and determining effects on the next higher level

  • Generally used to address effects resulting from single failures

  • Assess for each phase of flight and include “failure detection/indication” provisions

  • Generates lengthy document - collect failure modes of similar effect in summary document (FMES)

30
New cards

Common Cause Analysis - CCA

  • CCAs consider possible violations of failure condition independence

    • Three types of analysis - Common Mode Analysis (CMA); Particular Risk Analysis (PRA); Zonal Safety Analysis (ZSA)

  • Adding apparent redundancy introduces complexity; difficult to make such systems truly independent

  • Even truly independent systems may be affected by a common event due to their physical proximity or response to a common external condition

    • Egs: fuel starvation or contamination affecting all engines; icing affecting all Pitot tubes, bird strikes, etc.

31
New cards

Common Cause Analysis - CMA

Common Mode Analysis (CMA) - assess independence of faults due to:

  • s/w errors (not possible to test for all possibilities; use multiple code implementations)

  • Complex h/w failure (use multiple dissimilar computers, etc.)

  • Production/repair flaw (factor of safety; high production/process QA)

  • Installation/maintenance errors (dissimilar connections; eg hydraulic hoses and electric cables)

  • Cascading faults (auto load shedding)

Eg: dual electrical generating system analyzed as though the two systems are completely independent. Either system can provide more than enough electrical power to systems essential for flight but neither system can provide enough power to sustain all loads under all conditions. One system fails for a reason that has been anticipated and designed for; second system continues to function provided crew sheds some electrical load. If load is not shed, the second system may fail due to overheating, etc.

Therefore, the two systems were not truly independent. (highlights the importance of human factors - required crew actions in this case - may drive design change to automatic load shedding).

32
New cards

Common Cause Analysis - PRA

Particular Risk Analysis (PRA) - due to factors external to the system:

  • Fire

  • High energy devices

  • Hail, ice, snow

  • Bird strike

  • Lightning

  • HIRF

  • Leaking fluids

  • Flailing shafts

  • Tire tread separation or wheel rim release

  • New particular risks for electric propulsion systems due to external events: battery thermal runaway, high voltage hazards to occupants

Assign a probability of “1” to each of these for analysis

33
New cards

Common Cause Analysis - ZSA

Zonal Safety Analysis (ZSA)

  • Observance of basic installation rules such as segregation of:

    • Power and signal wiring;

    • electrical and fluid lines

  • Interference between systems

  • Potential for maintenance errors

Requires knowledge of physical installation (not just schematic)

34
New cards

Common Cause Analysis - ZSA cont’d

  • Requires experienced engineers and maintenance staff to do this properly! Knowledge of where problems have occurred in the past is essential

  • Maintenance error potential is particularly important:

    • Can a valve be installed in reverse?

    • Can a pipe or connector be physically connected to more than one other pipe/connector in the same bay?

    • Does removal of an item for servicing require removal of another item?

  • Use of 3D CAD models for this purpose; waiting until the actual aircraft is available can be very expensive (ie need to reroute wiring or hydraulics, etc.)

35
New cards

Common Cause Mitigations and Examples

36
New cards

Heat EPS Common Cause Analysis

  • Common Mode (CMA) - independence of DC-DC converter failure conditions?

  • Particular Risk (PRA) - due to external factors (high ambient temperatures - risk of TR?)

  • Zonal Safety (ZSA) - electric power and coolant lines in close proximity?

*Requires knowledge of phsyical installation (not just schematic)

37
New cards

Software and Complex Hardware - Development Assurance Levels (DALS)

DAL Assigned based on s/w criticality

Top level failure condition severity classification → DAL for s/w

  • Catastrophic → A

  • Hazardous → B

  • Major → C

  • Minor → D

  • No safety effect → E

Not possible to test every aspect of s/w and complex electronic h/w

Level A, B very expensive - limited to safety critical applications

38
New cards

Human Factors

Actions of Flight Crews & Maintenance Personnel must be accounted for in system design:

Potential for errors due to:

  • Misleading or conflicting information

  • Insufficient or excessive information

  • High workload (eg some phases of flight)

  • Poor ergonomics (eg control function)

Some crew procedures are required to meet safety targets (ie pre-flight checks for latent/dormant faults)

39
New cards

Other Topics - MEL & CMR

Minimum Equipment List (MEL)

  • A list of equipment/functions which need not be operative for flight

  • based on stated compensating precautions (eg operational or time limitations, flight crew procedures, or ground crew checks)

Certification Maintenance Requirements (CMR)

  • Periodic maintenance or flight crew check required to achieve compliance with CS 25.1309(b) for Hazardous and Catastrophic Failure Conditions

  • Where such checks cannot be accepted as basic servicing or airmanship they become Certification Maintenance Requirements

  • A CMR is identified as an operating limitation on the Type Certificate for the aeroplane