航空论坛_航空翻译_民航英语翻译_飞行翻译

 找回密码
 注册
搜索
查看: 1001|回复: 0
打印 上一主题 下一主题

The Application of System Safety To the Commercial Launch Industry [复制链接]

Rank: 9Rank: 9Rank: 9

跳转到指定楼层
1#
发表于 2008-12-21 21:04:54 |只看该作者 |倒序浏览
FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 Chapter 13: The Application of System Safety to the Commercial Launch Industry This chapter is intended for use as a pull-out handbook, separate from the FAA System Safety Handbook. 13.1 INTRODUCTION...............................................................................................................................1 13.2 OFFICE OF COMMERCIAL SPACE TRANSPORTATION (AST)..........................................................1 13.3 LICENSING PROCESS ................................................................................................................2 13.4 SYSTEM SAFETY ENGINEERING PROCESS .........................................................................5 13.5 SOFTWARE SAFETY ................................................................................................................15 FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 -1 13.0 The Application of System Safety To the Commercial Launch Industry 13.1 Introduction The office of the Associate Administrator for Commercial Space Transportation (AST), under Title 49, U.S. Code, Subtitle IX, Sections 70101-70119 (formerly the Commercial Space Launch Act), exercises the FAA’s responsibility to: regulate the commercial space transportation industry, only to the extent necessary to ensure compliance with international obligations of the United State and to protect the public health and safety, safety of property, and national security and foreign policy interest of the United States, …encourage, facilitate, and promote commercial space launches by the private sector, recommend appropriate changes in Federal statutes, treaties, regulations, policies, plans, and procedures, and facilitate the strengthening and expansion of the United States space transportation infrastructure. [emphasis added] The mandated mission of the AST is “…to protect the public health and safety and the safety of property….” AST has issued licenses for commercial launches of both sub-orbital sounding rockets and orbital expendable launch vehicles. These launches have taken place from Cape Canaveral Air Station (CCAS), Florida, Vandenburg Air Force Base (VAFB), California, White Sands Missile Range (WSMR), New Mexico, Wallops Flight Facility (WFF), Wallops Island, Virginia, overseas, and the Pacific Ocean. AST has also issued launch site operator licenses to Space Systems International (SSI) of California, the Spaceport Florida Authority (SFA), the Virginia Commercial Space Flight Authority (VCSFA), and the Alaska Aerospace Development Corporation (AADC). SSI operates the California Spaceport located on VAFB; SFA the Florida Space Port located on CCAS; VCSFA the Virginia Space Flight Center located on WFF; and AADC the Kodiak Launch Complex, located on Kodiak Island, Alaska. 13.2 Office of Commercial Space Transportation (AST) AST is divided into three functional components, the office of the Associate Administrator (AST-1), the Space Systems Development Division (SSDD), and the Licensing and Safety Division (LASD). 13.2.1 The office of the Associate Administrator (AST-1) AST-1 establishes policy, provides overall direction and guidance to ensures that the divisions function efficiently and effectively relative to the mandated mission “…to protect the public health and safety and the safety of property….” 13.2.2 The Space Systems Development Division (SSDD) The SSDD assess new and improved launch vehicle technology and their impacts upon both the existing and planned space launch infrastructures. SSDD works with the FAA and DOD Air Traffic Services to ensure full integration of space transportation flights into the Space and Air Traffic Management System. SSDD is AST’s interface with the Office of Science and Technology Policy (OSTP), other Government agencies, and the aerospace industry working to create a shared 2010 space launch operations vision and in the development of the Global Positioning (Satellite) System (GPS) for the guidance of launch vehicles and tracking at ranges. SSDD is also engaged in analyzes of orbital debris and its impact to current and future space launch missions and the commercialization of outer space. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 -2 13.2.3 The Licensing and Safety Division (LASD) LASD’s primary objective is to carry out AST’s responsibility to ensure public health and safety through the licensing of commercial space launches and launch site operations, licensing the operation of non- Federal space launch sites, and determining insurance or other financial responsibility requirements for commercial launch activities. AST/LASD looks to ensure protection of public health and safety and the safety of property through its licensing and compliance monitoring processes. 13.3 LICENSING PROCESS The components of the licensing process include a pre-licensing consultation period, policy review, payload review, safety evaluation, financial responsibility determination, and an environmental review. The licensing process components most concerned with the application of system safety methodologies are the safety evaluation, financial responsibility determination, and environmental determination. A space launch vehicle requires the expenditure of enormous amounts of energy to develop the thrust and velocity necessary to put a payload into orbit. The accidental or inadvertent release of that energy could have equally enormous and catastrophic consequences, both near and far. 13.3.1 Safety Evaluation It is the applicant’s responsibility to demonstrate that they understand all hazards and risks posed by their launch operations and how they plan to mitigate them. Hazard mitigation may take the form of safety devices, protective systems, warning devices, or special procedures. There are a number of technical analyses; some quantitative and some qualitative, that the applicant may perform in order to demonstrate that their commercial launch operations will pose no unacceptable threat to the public. The quantitative analyses tend to focus on 1) the reliability and functions of critical safety systems, and 2) the hazards associated with the hardware, and the risk those hazards pose to public property and individuals near the launch site and along the flight path, to satellites and other on-orbit spacecraft. The most common hazard analyses used for this purpose are Fault Tree Analysis, Failure Modes and Effects Analysis, and Over-flight Risk and On-Orbit Collision Risk analyses using the Poisson Probability Distribution. The qualitative analyses focus on the organizational attributes of the applicant such as launch safety policies and procedures, communications, qualifications of key individuals, and critical internal and external interfaces. It is AST/LASD’s responsibility to ensure that the hazard analyses presented by the applicant demonstrates effective management of accident risks by identifying and controlling the implicit as well as explicit hazards inherent in the launch vehicle and proposed mission. LASD must evaluate the applicant’s safety data and safety related hardware/software elements and operations to ascertain that the demonstrations provided by the applicant are adequate and valid. Specifically, the LASD evaluation is designed to determine if the applicant has: · Identified all energy and toxic sources and implemented controls to preclude accidental or inadvertent release. · Evaluated safety critical aspects, potential safety problems, and accident risk factors. · Identified potential hazardous environments or events, and assessed their causes, possible effects and probable frequency of occurrence. · Implemented effective hazard elimination, prevention or mitigation measures or techniques to minimize accident risk to acceptable levels. · Specified the means by which hazard controls or mitigation methodology can be verified and validated. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 -3 13.3.2 Financial Responsibility Determination Section 70112 of the Act requires that all commercial licensees demonstrate financial responsibility to compensate for the maximum probable loss from claims by: · A third party for death, bodily injury, or property damage or loss resulting from an activity carried out under the license; and · The U.S. Government against a person for damage or loss to government property resulting from an activity carried out under the license. Section 70112 also requires that the Department of Transportation set the amounts of financial responsibility required of the licensee. The licensee can then elect to meet this requirement by: · Proving it has financial reserves equal to or exceeding the amount specified, or · Placing the required amount in escrow, or · Purchasing liability insurance equal to the amount specified. The most common and preferred method is via the purchase of liability insurance. The methodology developed for setting financial responsibility requirements for commercial launch activities is called Maximum Probable Loss (MPL) analysis 1 . MPL analysis was developed to protect launch participants from the maximum probable loss due to claims by third parties and the loss of government property during commercial launch activities. Note that this is maximum probable loss, not maximum possible loss. Generally speaking, MPL is determined by identifying all possible accident scenarios, examining those with the highest potential losses for both government property and third party, and then estimating the level of loss that would not be exceeded at a given probability threshold. If the launch is to take place from a private licensed range and no government property is at risk, no government property financial responsibility requirement will be issued. An integral part of, and critical input to the MPL, is the Facility Damage and Personnel (DAMP) Injury Analysis 2 : DAMP uses information about launch vehicles, trajectories, failure responses, facilities and populations in the launch area to estimate the risk and casualty expectations from impacting inert debris, secondary debris and overpressures from impact explosions. Together, the MPL and DAMP analyses are used to determine the financial responsibility determinations necessary to insure compensation for losses resulting from an activity carried out under the commercial license. 13.3.3 Environmental Determination The environmental determination ensures that proposed commercial space launch activities pose no threat to the natural environment. The National Environmental Policy Act (NEPA) of 1969, as amended, requires that: Federal agencies consider the environmental consequences of major Federal actions; take actions that protect, restore, and enhance the environment; and ensure that environmental information is available to public officials and citizens before making decisions and taking action. The licensing of commercial space launch activities, either for a launch or launch site, is considered a major Federal action. Consequently, AST is responsible for analyzing the environmental impacts associated with proposed commercial space launch activities. AST is also responsible for the assessing the applicant’s preparation and submittal of Environmental Assessments and Environmental Impact Statements to ensure compliance with the NEPA. 1 Futron Corporation developed the MPL Analysis methodology employed by AST. 2 Research Triangle Institute developed the DAMP Analysis methodology employed by AST. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 -4 Figure 13-3 System Safety, Software Acquisition and Systematic Software Acquisition Process SRR – System Requirements Review SDR – System Design Review SSR – Software Specification Review PDR – Preliminary Design Review CDR – Critical Design Review TRR – Test Readiness Review CSU – Computer Software Unit CSCI – Computer Software Configuration Item PD – Preliminary Design DD – Detailed Design Concept Research & Development Design Deployment/ Operations Disposition Preliminary Hazard List System Safety Program Plan (Initiated Conceptual Phase – updated remainder of system life cycle Preliminary Hazard Analysis (PHA) System Hazard Analysis (SHA) Subsystem Hazard Analysis (SSHA) System Safety Process Software Acquisition Process Software Requirements Analysis PD DD Code, CSCI & CSU Test System Integration Test Operations Maintenance & Upgrade SRR SDR SSR CDR PDR Configuration Management Software Safety Program Management Development of Function Hazard List Tailor Generic Software Safety Requirements Preliminary Software Hazards Analysis (PSHA) Derive System Specific Safety Critical Requirements Software Safety Architectural Design Hazard Analysis (SSADHA) Software Detailed Design Subsystem Hazard Analysis (SSDDHA) Software Safety Integration Test Planning Software Safety Integration Testing & Analysis Validation of Software Operations & Support Requirements Systematic Software Safety Process TRR - Link hazard causal factors to design - Identify hazards that cross subsystem boundaries - Ensure hazards are mitigated in interfacing subsystems or external systems - Identify unresolved interface safety issues Software Safety Planning FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 -5 13.4 SYSTEM SAFETY ENGINEERING PROCESS 13.4.1 Overview The System Safety Engineering Process is the structured application of system safety engineering and management principles, criteria, and techniques to address safety within the constraints of operational effectiveness, time, and cost throughout all phases of a system’s life cycle. The intent of the System Safety Engineering Process is to identify, eliminate, or control hazards to acceptable levels of risk throughout a system’s life cycle. This process is performed by the vehicle developer/operator. Because of the complexity and variety of vehicle concepts and operations, such a process can help ensure that all elements affecting public safety are considered and addressed. Without such a process, very detailed requirements would have to be imposed on all systems and operations, to ensure that all hazards have been addressed which could have the undesired effect of restricting design alternatives and innovation or could effectively dictate design and operations concepts. The process (as described in Mil Std 882C) includes a System Safety Program Plan (SSPP). The SSPP (or its equivalent) provides a description of the strategy by which recognized and accepted safety standards and requirements, including organizational responsibilities, resources, methods of accomplishment, milestones, and levels of effort, are to be tailored and integrated with other system engineering functions. The SSPP lays out a disciplined, systematic methodology that ensures all risks – all events and system failures (probability and consequence) that contribute to expected casualty – are identified and eliminated, or that their probability of occurrence is reduced to acceptable levels of risk. The SSPP should indicate the methods employed for identifying hazards, such as Preliminary Hazards Analysis (PHA), Subsystem Hazard Analysis (SSHA), Failure Mode and Effects Analysis (FMEA), Fault Tree Analysis. Risk Mitigation Measures are likewise identified in the plan. These include avoidance, design/redesign, process/procedures and operational rules and constraints. The System Safety Engineering Process identifies the safety critical systems. Safety critical systems are defined as any system or subsystem whose performance or reliability can affect public health and safety and safety of property. Such systems, whether they directly or indirectly affect the flight of the vehicle, may or may not be critical depending on other factors such as flight path and vehicle ability to reach populated areas. For this reason, it is important to analyze each system for each phase of the vehicle mission from ground operations and launch through reentry and landing operations. Examples of potentially safety critical systems that may be identified through the system safety analysis process using PHA or other hazard analysis techniques may include, but are not limited to: · Structure/integrity of main structure · Thermal Protection System (e.g., ablative coating) · Temperature Control System (if needed to control environment for other critical systems) · Main Propulsion System · Propellant Tanks · Power Systems · Propellant Dumping System FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 -6 · Landing Systems · Reentry Propulsion System · Guidance, Navigation and Control System(s), Critical Avionics (Hardware and Software) - includes Attitude, Thrust and Aerodynamic Control Systems · Health Monitoring System (hardware and software) · Flight Safety System (FSS) · Flight Dynamics (ascent and reentry) for stability (including separation dynamics) and maneuverability · Ground Based Flight Safety Systems (if any) including telemetry, tracking and command and control systems · Depending on the concept, additional “systems” might include pilot and life support systems and landing systems if they materially affect public health and safety · Others identified through hazard analysis 13.4.2 Validation of Safety Critical Systems Through the system safety process, the applicant demonstrates that the proposed vehicle design and operations satisfy regulatory requirements and that the system is capable of surviving and performing safely in all operating environments including launch, orbit, reentry and recovery. Documentation must show adequate design, proper assembly, and vehicle control during all flight phases. Documentation is expected to consist of design information and drawings, analyses, test reports, previous program experience, and quality assurance plans and records. AST uses a pre-application consultation process to help a potential applicant to understand what must be documented and to help identify potential issues with an applicant’s proposed activities that could preclude its obtaining a license. The pre-application process should be initiated by the applicant early in their system development (if possible during the operations concept definition phase) and maintained until their formal license application is completed. This pre-application process should be used to provide AST with an understanding of the safety processes to be used, the safety critical systems identified, analysis and test plan development, analysis and test results, operations planning and flight rules development. Analyses may be acceptable as the primary validation methodology in those instances where the flight regime cannot be simulated by tests, provided there is appropriate technical rationale and justification. Qualification tests, as referenced in the safety demonstration process and the System Safety Program Plan, are normally conducted to environments higher than expected. For example, expendable launch vehicle (ELV) Flight Safety Systems (FSS) are qualified to environments a factor of two or higher than expected. (See Figure 13-2) These tests are conducted to demonstrate performance and adequate design margins and may be in the form of multi-environmental ground tests, tests to failure, and special flight tests. Such tests are normally preceded with detailed test plans and followed by test reports. 3 3 Test plans are important elements of the ground and flight test programs. Such plans define, in advance, the nature of the test (what is being tested and what the test is intended to demonstrate with respect to system functioning, system performance and system reliability). The test plan should be consistent with the claims and purpose of the test and wherever appropriate, depending on the purpose of the test, clearly defined criteria for pass and fail should be identified. A well-defined test plan and accompanying test report may replace observation by the FAA. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 -7 Figure 13-2: Relationship of Use Environment to Qualification Test Use Environment Qualification Test Environment Vibration T e m p e r a t u r e In addition, Quality assurance (QA) records are useful in establishing verification of both design adequacy and vehicle assembly and checkout (workmanship). Table 13-1, Validation Acceptance Matrix, identifies sample approaches that may be employed to validate acceptance for critical systems. Examples of types of analyses, ground tests, and flight tests are provided following this matrix. (Note: Quality Assurance programs and associated records are essential where analysis or testing, covering all critical systems, are involved.) FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 8 Table 13-1: Validation Acceptance Matrix Candidate Critical System Analyses Ground Flight Test Test Structure/Integrity of Main Structure X X P Thermal Protection X P P Environmental Control (temp, humidity) X X X Propulsion: Main, Auxiliary and Reentry (de-orbit) X P P Propellant Tank Pressurization X X P GN&C, Critical Avionics *; includes de-orbit targeting (e.g., star-tracker, GPS) X X X Health Monitoring * X X X Flight Safety System (FSS)* X X X Recovery and Landing* X P P Ordnance* (other than Safety) X X X Electrical and Power* X X X Telemetry and Tracking and Command* X X X Flight Control (ascent, separation, reentry) * X X X FSS Ground Support Equipment (if any) * X X N/A P - partial; cannot satisfy all aspects X - If in sufficient detail when combined with test results or selected analyses * - Includes both hardware and software FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 9 13.4.3 Analyses There are various types of analyses that may be appropriate to help validate the viability of a critical system or component. The following provides examples of some types of critical systems analysis methodologies and tools. · Mechanical Structures and Components (Vehicle Structure, Pressurization, Propulsion System including engine frame thrust points, Ground Support Equipment) · Types of Analyses: Structural Loads, Thermal, Fracture Mechanics, Fatigue, Form Fit & Function · Software Tools for Analyses: Nastran, Algor, Computational Fluid Dynamics codes, CAD/CAM · Thermal Protection System · Types of Analyses (for TPS and Bonding Material): Transient and Steady State Temperature Analyses, Heat Load, and Heating and Ablative Analyses. · Software Tools for Analyses: SINDA by Network Analysis Inc. · Electrical/Electronic Systems & Components (Electrical, Guidance, Tracking, Telemetry, Navigation, Communication, FSS, Ordnance, Flight Control and Recovery) · Types of Analyses: Reliability, FMEA, Single Failure Point, Sneak Circuit, Fault Tree, Functional Analysis, Plume effects · Software Tools for Analyses: MathCad, Relex, and FaultrEase · Propulsion Systems (Propulsion, FSS, Ordnance, Flight Control) · Types of Analyses: Analytical Simulation of nominal launch and abort sequences for Main Engines, Orbital Maneuvering System (including restart for reentry-burn) and Attitude Control System; capacity analysis for consumables; Plume Flow Field Modeling · Software Tools for Analyses: Nastran, Algor, SPF-III, and SINDA · Aerodynamics (Structure, Thermal, Recovery) · Types of Analyses: Lift, Drag, Stability, Heating, Performance, Dispersion, Plume effects · Software Tools for Analyses: Post 3/6 DOF, Computational Fluid Dynamics Codes Monte Carlo Simulation Codes · Software (Guidance, Tracking & Telemetry & Command, FSS, Flight Control and Recovery) · Types of Analyses: Fault Tree, Fault Tolerance, Software Safety (including abort logic), Voting Protocol Dead Code, Loops, and Unnecessary Code · Validation Methodologies, such as ISO 9000-3 4 4 ISO 9000-3 is used in the design, development, and maintenance of software. Its purpose is to help produce software products that meet the customers' needs and expectations. It does so by explaining how to control the quality of both products and the processes that produce these products. For software product quality, the standard highlights four measures: specification, code reviews, software testing and measurements. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 10 13.4.4 Ground Test Ground tests include all testing and inspections performed by the applicant prior to flight, including qualification, acceptance and system testing. It is anticipated that an applicant will perform various types of ground tests to validate the capability of critical systems and components. The following provides examples of some types of critical systems validation ground tests. Again these are only examples and should not be construed as the only types of ground tests which may be used to validate a specific system for a specific operational environment, nor should it be interpreted that all of these example ground tests will be necessary to validate a specific system. Mechanical Systems and Components (Vehicle Structure, Pressurization, Propulsion System including engine frame thrust points, Ground Support Equipment) Types of Tests: Load, Vibration (dynamic and modal), Shock, Thermal, Acoustic, Hydro-static, Pressure, Leak, Fatigue, X-ray, Center of Gravity, Mass Properties, Moment of Inertia, Static Firing, Bruceton Ordnance, Balance, Test to Failure (simulating non-nominal flight conditions), Non-Destructive Inspections Electrical/Electronic Systems (Electrical, Guidance, Tracking, Telemetry and Command, Flight Safety System (FSS), Ordnance, Flight Control and Recovery) Types of Tests: Functional, Power/Frequency Deviation, Thermal Vacuum, Vibration, Shock, Acceleration, X-ray, recovery under component failures, abort simulations, TDRSS integration testing (up to and including pre-launch testing with flight vehicle) Propulsion Systems (Propulsion, FSS, Ordnance, Flight Control) Types of Tests: Simulation of nominal launch and abort sequences for engines (including restart, if applicable), Orbital Maneuvering System (including restart for reentry-burn) and Attitude Control System; Environmental testing (Thermal, Vibration, Shock, etc.) Thermal Protection System Types of Tests (for TPS and bonding material): Thermal, Vibration, Humidity, Vacuum, Shock Aerodynamics (Structure, Thermal, Recovery) Types of Tests: Wind Tunnel, Arc Jet, Drop Tests (Landing Systems) Software (Electrical, Guidance, Tracking, Telemetry, Command, FSS, Ordnance, Flight Control and Recovery) Types of Tests: Functional, Fault Tolerance, Cycle Time, Simulation, Fault Response, Independent Verification and Validation, Timing, Voting Protocol, Abort sequences (flight and in-orbit) under non-nominal conditions with multiple system failures, Integrated Systems Tests 13.4.5 Flight Tests If an applicant’s System Safety Plan includes a flight test program, then a considerable amount of planning is needed to define the flight test program that will establish the performance capabilities of the vehicle for routine and repetitive commercial operations. When flight testing is indicated, a flight test plan will be needed to demonstrate that the vehicle’s proposed method of operations is acceptable and will not be a hazard to the public health and safety, and safety of property. The purpose of flight-testing is to verify the system performance, validate the design, identify system deficiencies, and demonstrate safe operations. Experience repeatedly shows that while necessary and important, analyses and ground tests cannot and do not uncover all potential safety issues associated with new launch systems. Even in circumstances where all known/identified safety critical functions can be FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 11 exercised and validated on the ground, there is still the remaining concern with unrecognized or unknown interactions (“the unknown unknowns”). The structure of the test program will identify the flight test framework and test objectives, establish the duration and extent of testing; identify the vehicle’s critical systems, identify the data to be collected, and detail planned responses to nominal and unsatisfactory test results. Test flight information includes verification of stability, controllability, and the proper functioning of the vehicle components throughout the planned sequence of events for the flight. All critical flight parameters should be recorded during flight. A post-flight comparative analysis of predicted versus actual test flight data is a crucial tool in validating safety critical performance. Below are examples of items from each test flight that may be needed to verify the safety of a reusable launch vehicle. Listed with each item are examples of what test-flight data should be monitored or recorded during the flight and assessed postflight: Vehicle/stage launch phase: Stability and controllability during powered phase of flight. · Vehicle stage individual rocket motor ignition timing, updates on propellant flow rates, chamber temperature, chamber pressure, and burn duration, mixture ratio, thrust, specific impulse (ISP) · Vehicle stage trajectory data (vehicle position, velocity, altitudes and attitude rates, roll, pitch, yaw attitudes) · Vehicle stage Attitude, Guidance and Control system activities · Functional performance of the Vehicle Health Monitoring System · Functional performance of the Flight Safety System/Safe Abort System · Electrical power, and other critical consumables, usage and reserves (i.e. gases, fluids, etc…) · Actual thermal and vibroacoustic environment · Actual structural loads environment Staging/separation phase of boost and upper stages: Stable shutdown of engines, and nominal separation of the booster & upper stages. · Separation activity (timestamp, i.e., separation shock loads, and dynamics between stamps) · Functional performance of the Vehicle Health Monitoring System · Electrical power, and other critical consumables, usage and reserves (i.e. gases, fluids, etc…) · Functional performance of the Flight Safety System/Safe Abort System Booster stage turn-around (re-orientation) or “loft” maneuver phase (if applicable): · Rocket motor re-start (if applicable): timing, updates on propellant flow rates, chamber temperature, chamber pressure, burn duration, mixture ratio, thrust, ISP · Attitude, Guidance and Control system activities · Actual structural loads environment · Actual thermal and vibroacoustic environment FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 12 · Functional performance of the Flight Safety System/Safe Abort System Booster stage flyback phase (if applicable): Flyback engine cut-off, fuel dump or vent (if required), nominal descent to the planned impact area, proper functioning and reliability of the RLV landing systems. · Booster stage post-separation (flyback) trajectory data · Electrical power usage and reserves · Booster stage landing system deployment activity (timestamp) · Actual thermal and vibroacoustic environment · Actual structural loads environment · Functional performance of the Vehicle Health Monitoring System · Functional performance of the Flight Safety System/Safe Abort System · Attitude, Guidance and Control system activities Vehicle stage ascent phase (if multistage): nominal ignition of the stage’s engine, stability and controllability of the stage during engine operation, orbital insertion – simulated (for suborbital) or actual – of the vehicle. · Vehicle individual rocket motor ignition timing, updates on propellant flow rates, chamber temperature, chamber pressure, and burn duration · Vehicle circularization and phasing burn activities (ignition timing, updates on propellant flow rates, chamber temperature, chamber pressure, and burn duration) · Vehicle trajectory data (vehicle position, altitude, velocity, roll, pitch, yaw attitudes at a minimum) · Attitude, guidance and control system activities · Functional performance of the Vehicle Health Monitoring System · Functional performance of the Flight Safety System/Safe Abort System · Electrical power, and other critical consumables, usage and reserves (i.e. gases, fluids, etc…) · Actual structural loads environment · Actual thermal and vibroacoustic environment Vehicle descent (including vehicle’s de-orbit burn targeting and execution phases): Function of the programmed flight of the vehicle/upper stage to maintain the capability to land (if reusable) at the planned landing site, or to reenter for disposal (if expendable), assurance of fuel dump or depletion, and proper descent and navigation to the planned or alternate landing site. · Vehicle pre-deorbit burn trajectory data · Vehicle deorbit burn data (ignition timing, updates on propellant flow rate, chamber temperature, chamber pressure, and burn duration) · Vehicle descent trajectory data (position, velocity, and attitude) · Attitude, Guidance and Control system activities FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 13 · Actual thermal and vibroacoustic environment · Actual structural loads environment · Functional performance of the Vehicle Health Monitoring System · Functional performance of the Flight Safety System/Safe Abort System · Electrical power and other critical consumables usage and reserves (i.e. gases, fluids, etc…) · Vehicle landing system deployment activity (timestamp) 13.4.6 Performance and Reliability Data Performance and reliability data may be supported by flight history on other vehicles with similar or comparable safety critical systems, sub-systems, and components, and by conducting both analyses and tests, at the respective levels. A flight history could mean extensive documentation might not be required if it can be shown through test results, analyses, or empirical data, that the flight regimes experienced are similar to the proposed flight regime. The degree of applicability of data depends on the degree of similarity to environmental conditions and how environmental conditions compare to the history and anticipated reactions of this system. Even when the same system, sub-system, or component is known to have an extensive (and favorable) flight history in the same or more severe environments, interfaces and integration with other systems must still be examined and tested. Another method of acquiring data is through estimating system, sub-system, and component 3-sigma performance and reliability numbers from testing evaluations and (where applicable) flight data. The use of similarity is not new to launch operations. EWR 127-1, Paragraph. 4.14.1.2, states: as required, qualification by similarity analysis shall be performed; if qualification by similarity is not approved, then qualification testing shall be performed. For example, if component A is to be considered as a candidate for qualification by similarity to a component B that has already been qualified for use, component A shall have to be a minor variation of component B. Dissimilarities shall require understanding and evaluation in terms of weight, mechanical configuration, thermal effects, and dynamic response. Also, the environments encountered by component B during its qualification or flight history shall have to be equal to or more severe than the qualification environments intended for component A. 13.4.7 Operational Controls There is an interrelationship between the system design capabilities and the systems operational limitations. Figure 2 depicts the relationship between the vehicle systems and the scope of operations within which the vehicle is operated. What constitutes a safety critical system may depend on the scope and nature of the vehicle design and its proposed operations. Intended operational requirements affect the proposed vehicle design requirements and vehicle capabilities/limitations and also establish the operational system constraints necessary to protect public health and safety. For example, reusable launch vehicle landing sites may have to be within some minimum cross-range distance from the orbital ground trace because of cross-range limitations of the vehicle. A vehicle operator may choose, or be required, to mitigate certain vehicle limitations through the use of operational controls rather than relieving vehicle limitations through design changes. Test parameters and analytic assumptions will further define the limits of flight operations. The scope of the analyses and environmental tests, for example, will constitute the dimensions of the applicant’s demonstration process and therefore define the limits of approved operations if a license is issued. Such testing limits, identified system and subsystem limits, and analyses also are expected to be reflected in FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 14 mission monitoring and mission rules addressing such aspects as commit to launch, flight abort, and commit to reentry. Vehicle capabilities/limitations and operational factors such as launch location and flight path each affect public risk. The completion of system operation demonstrations, such as flight simulations and controlled flight tests, provide additional confidence in the vehicle systems and performance capabilities. As confidence in the systems overall operational safety performance increases, key operational constraints such as restrictions on overflight of populated areas may be relaxed. The following are examples of the types of operations-related considerations that may need to be addressed by the applicant when establishing their operations scenarios. Launch commit criteria/rules Human override capability to initiate safe abort during launch and reentry System monitoring, inspection and checkout procedures For re-flight: inspection and maintenance Selected primary and alternate landing sites for each stage Surveillance/control of landing areas Standard limits on weather Coordination with appropriate air space authorities Limits on flight regime (ties in with analysis, testing and demonstrating confidence in system performance and reliability) Limits on over-fight of populated areas Others identified through hazard analysis FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 15 Figure 13-2: Interrelationship between Safety Critical Systems and Safety Critical Operations Safety Critical Systems • Design Standards for systems • Analysis, Tests, Inspections Safety Critical Operations • Operations Standards for Systems • Analysis, Tests, Rehearsals, Simulations, Controlled Flight Tests Vehicle Capabilities/Limitations Operational Capabilities/ Limitations Public Risk 13.4.8 Determination of Risk to the Public Expected casualty is used in the space transportation industry as a measure of risk to public safety. Expected casualty is the expected average number of human casualties per mission. Human casualty is defined as a fatality or serious injury. The application of the expected casualty analysis to determine public risk is further defined in FAA Advisory Circular 431-02. 13.4.9 Determination of Need for Additional Risk Mitigation The results of the expected casualty analysis may identify the need for additional risk mitigation measures that need to be employed. These measures may include additional operational controls or may require the redesign of certain safety critical systems. These additional risk mitigation measures would be evaluated within the System Safety Process and the resultant risk to the public would be determined. 13.5 SOFTWARE SAFETY 13.5.1 Safety Critical Software Safety-critical software plays an ever-increasing role in Commercial Space Transportation (CST) computer systems. To preserve CST flight integrity, software-based hazards must be identified and eliminated or reduced to acceptable levels of risk. Particular concern surrounds potential softwareinduced accidents occurring during CST launch and reentry. Due to mission complexity, software failures manifested at these critical times can cause serious accidents. Populated areas would suffer major harm if defective software were to permit CST vehicles to violate their defined safety launch limits. Safetycritical software, relative to CST launch vehicles, payloads and ground support equipment is inherently defined as any software within a control system containing one or more hazardous or safety critical functions. Safety critical functions are usually but not always associated with safety-critical systems. Therefore, the following definition for safety –critical systems may also be applied to safety-critical functions. A safety-critical system (or function) has been inherently defined as any system or subsystem FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 16 (or function) whose performance or reliability can affect (i.e. malfunction or failure will endanger) public health, safety and safety of property. 5 13.5.2 Systematic Software Safety Process Introduction The Systematic Software Safety Process (SSSP) encompasses the application of an organized periodic review and assessment of safety-critical software and software associated with safety-critical system, subsystems and functions. The Systematic Software Safety Process consist primarily of the following elements: · Software safety planning · The software safety organization · A software safety team · Application of the software safety process during all life cycle phases · Identification and application of life cycle phase-independent software safety activities · Identification of special provisions · Software safety documentation Software Safety Planning Software system safety planning is deemed essential early in the software life cycle. Most importantly, planning should impose provisions for accommodating safety well before each of the software design, coding, testing, deployment and maintenance phases starts in the cycle. Moreover, these provisions are to be planned carefully to impact minimally the software development process. The software system safety plan should contain provisions assuring that: · Software safety organization is properly chartered and a safety team is commissioned in time. · Acceptable levels of software risk are defined consistently with risks defined for the entire system. · Interfaces between software and the rest of the system’s functions are clearly delineated and understood. · Software application concepts are examined to identify safety-critical software functions for hazards. · Requirements and specifications are examined for safety hazards (e.g. identification of hazardous commands, processing limits, sequence of events, timing constraints, failure tolerance, etc.) · Design and implementation is properly incorporated into the software safety requirements. · Appropriate verification and validation requirements are established to assure proper implementation of software system safety requirements. · Test plans and procedures can achieve the intent of the software safety verification requirements. 5 Reference D. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 17 · Results of software safety verification efforts are satisfactory. The Institute of Electrical and Electronic Engineering (IEEE) offers a comprehensive standard (Standard for Software Safety Plans) focusing solely on planning. The Standard articulates in sufficient detail both software safety management and supporting analyses. The Standard’s annex describes the kind of analyses to be performed during the software requirements, design, code, test and change phases of the traditional life cycle. Similar planning models are provided by the Department of Defense (DOD) Defense Standard 00-55-Annex B. Software Safety Organization Safety oversight consists of a staff function crossing several organizational boundaries. By its nature, it is meant to interact with other staff and line functions, including program or project management, software system design, quality assurance, programming, reliability, testing, human factors, and operations. Accountability-wise, the ultimate responsibility for the development and operation of a safe software system(s) rests with the CST applicant or licensed operator. Thus, the applicant’s or operator’s top management should be committed to supporting the software safety process across all these staff and line functions. A software safety organization can take one of many shapes, depending on the needs of the applicant or licensed operator. However, the following requisites are recommended: · Centralized authority and responsibility dedicated to the safety initiatives · Safety team independence, and · High enough safety team status relative to the rest of the organization. Centralization allows a single organization to focus entirely on hazards and their resolutions during any life cycle phase, be it design, coding or testing. Independence prevents bias and conflicts of interest during organizationally sensitive hazard assessment and management. A high status empowers the team to conduct its mission with sufficient visibility and importance. By endorsing these requisites, CST applicants and operators will indicate they are attentive to the safety aspects of their project or mission. Software Safety Team Safety planning also calls for creating a software safety team. Team size and shape depends commensurately on mission size and importance. To be effective, the team should consist of analytical individuals with a sufficient system engineering background. Military Standard (MIL STD) 882C provides a comprehensive matrix of minimum qualifications for key system safety personnel. It can apply to software system safety as well, provided professional backgrounds include sufficient experience with software development (software requirements, design, coding, testing, etc.) Several typical activities expected of the team range from identifying software-based hazards to tracing safety requirements and limitations in the actual code, to developing software safety test plans and reviewing test results for their compliance with safety requirements. Software Safety During Life Cycle Phases The SSSP should support a structured program life cycle model that incorporates both the system design and engineering, and software acquisition process. Prominent software life cycle models include the FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 18 waterfall and spiral methodologies. Although different models may carry different lifecycle emphasis, the adopted model should not affect the SSSP itself. For discussion purposes only, this enclosure adopts a waterfall model (subject to IEEE/IEA Standard for Information Technology-software life cycle processes No. 12207.) For brevity, only some phases (development, operation, maintenance and support) of the Standard are addressed in terms of their relationship to software safety activities. This relationship is summarized in Table 13-2 The table’s contents partly reflect some of the guidance offered by the National Aeronautics and Space Administration (NASA) Standard 8719.13A and NASA Guidebook GB-1740.13- 96. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 19 Table 13-2: Software Safety Activities Relative to the Software Life Cycle Life Cycle Phase Corresponding Safety Activity Inputs Expected Results Milestones To Be Met Concept/ Requirements/ Specifications -Review software concept for safety provisions -Derive generic and systemspecific software safety requirements. -Analyze software requirements for hazards. -Identify potential software/system interface hazards -Develop Functional Hazards List (FHL) -Develop initial Preliminary Software Hazard Analysis (PSHA) -Preliminary Hazard Analysis (PHA) [from system safety analysis] -Generic and system-wide safety specs. PSHA Report PHL Software Concept Review (SCR) Software Requirements Review (SRR) and Software Specification Review (SSR) Architecture/ Preliminary Software Design At high design level: -Identify Safety Critical Computer Software Components (SCCSCs) -Verify correctness & completeness of architecture -Ensure test coverage of software safety requirements. PSHA Software Safety Architectural Design Hazard Analysis (SSADHA) Report Preliminary Design Review (PDR) Detailed Design At the low design(unit) level: -Focus on SCCSCs at the unit level. -Verify correctness/ completeness of detail. Design PSHA SSADHA Software Safety Detailed Design Hazard Analysis (SSDDHA) Report Critical Design Review (CDR) Implementation Coding -Examine correctness & completeness of code from safety requirements. -Identify possibly unsafe code. -Walk-through/audit the code PSHA, SSADHA, SSDDHA Software Safety Implementation Hazard Analysis (SSIHA) report Test Readiness Review (TRR) Integration and Testing -Ensure test coverage of software safety requirements. -Review test documents and results for safety requirements. -Final SSHA Test documents -Software Safety Integration Testing (SSIT) Report -Final SSHA report Acceptance Operations and Maintenance -Operating and Support Hazard Analysis (O&SHA) All of the above plus all incidents reports O&SHA Report(s), as required Deployment FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 20 Figure 3 provides a composite overview of the entire safety process. The figure consists of three parts. The top part reflects the broader System Safety Process described in draft Advisory Circular 431.35-2. The middle part illustrates a typical waterfall Software Acquisition Process life cycle. The bottom part also partly corresponds to the Systematic Software Safety Process. In Figure 3, all processes shown in horizontal bars are subject to a hypothetical schedule with time duration not drawn to any scale. Phase-independent software safety activities NASA’s Software Safety Standard 8719.13A mentions activities not tied to specific phases. The Standard lists the following ones meant to occur throughout the life cycle: · Tracing safety requirements keeping track of the software safety requirements during design, coding and testing, including the correspondence between these requirements and the system hazard information. · Tracking discrepancies between the safety and development aspects of the software. · Tracking changes made to the software to see if they impact the safety process. · Conducting safety program reviews to verify if safety controls are being implemented to minimize hazards. Special Provisions Commercial Off the Shelf (COTS): COTS software targets a broad range of applications, with no specific one envisioned ahead of time. Therefore, care must be taken to ensure COTS software presence minimizes risk when it becomes embedded or coupled to specific applications. Consideration ought to be given to designing the system such that COTS software remains isolated from safety-critical functions. If isolation is not possible, then safeguards and oversight should be applied. Software Reuse: Reusable software originates from a previous or different application. Usually, developers intend to apply it to their current system, integrating it “as is” or with some minor modifications. The Software Safety Team verification/validation plan, etc.) Annex B should serve as a general model for preparing software safety documents The results of most of the safety analyses activities usually require preparing several hazard analysis reports documenting the findings of the safety team. The team has also the responsibility of presenting their findings to decision-making management at critical milestones, like the Software Requirements Review (SRR), Preliminary Design Review (SDR), Critical Design Review (CDR), etc. Towards this end, DOD Defense Standard 00-55-Annex E describes how to prepare a software safety “case”. The Standard defines a case as “a well-organized and reasoned justification, based on objective evidence, that the software does or will satisfy the safety aspects of the Software Requirement”. 13.5.3 Software Safety Documentation Numerous documents are to be prepared and distributed during the SSSP. To track them, a comprehensive checklist, such as that cited in MIL STD 882 may be applied. Two highly recommended documents are the Safety Assessment Report (SAR) and the System Safety Hazard Analysis Report (SSHA). DOD Defense Standard 00-55-Annex B offers a detailed outline of the contents of numerous software safety documents (software safety plan, case or report, records log, audit plan, audit report, quality plan, risk management plan, verification/validation plan, etc.) Annex B should serve as a general model for preparing software safety documents. The results of most of the safety analyses activities usually require preparing several hazard analysis reports documenting the findings of the safety team. The team also has the responsibility of presenting their findings to decision-making management at critical milestones, like the Software Requirements Review (SRR), Preliminary Design Review (SDR), FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 21 Critical Design Review (CDR), etc. Towards this end, DOD Defense Standard 00-55-Annex E describes how to prepare a software safety “case”. The Standard defines a case as “a well-organized and reasoned justification, based on objective evidence, that the software does or will satisfy the safety aspects of the Software Requirement”. 13.5.4 Safety Critical Software functions Software can be labeled defective if it does not perform as expected. Major classes of defects are: · Software not executing · Software executing too late, too early, suddenly or out of sequence, or · Software executing but producing wrong information. In turn, defective software can be labeled hazardous if it consists of safety-critical functions that command, control and monitor sensitive CST systems. Some typical software functions considered safety-critical include: · Ignition Control: any function that controls or directly influences the pre-arming, arming, release, launch, or detonation of a CST launch system. · Flight Control: any function that determines, controls, or directs the instantaneous flight path of a CST vehicle. · Navigation: any function that determines and controls the navigational direction of a CST vehicle. · Monitoring: any function that monitors the state of CST systems for purposes of ensuring its safety. · Hazard Sensing: any function that senses hazards and/or displays information concerning the protection of the CST system. · Energy Control: any function that controls or regulates energy sources in the CST system. · Fault Detection: any function that detects, prioritizes, or restores software faults with corrective logic. · Interrupt Processing: any function that provides interrupt priority schemes and routines to enable or disable software-processing interrupts. · Autonomous Control: any function that has autonomous control over safety-critical hardware. · Safety Information Display: any function that generates and displays the status of safetycritical hardware or software systems. · Computation: any function that computes safety-critical data. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 22 13.5.5 Software Safety Risk and Hazard Analysis Risk Assessment A key element in system safety program planning is the identification of the acceptable level of risk for the system. The basis for this is the identification of hazards. Various methodologies used in the identification of hazards are addressed in Sections 2.3 & 2.4 of draft AC 431.35-2. Once the hazards and risks are identified, they need to be prioritized and categorize so that resources can be allocated to the functional areas having an unacceptable risk potential. Risk assessment and the use of a Hazard Risk Index (HRI) Matrix as a standardized means with which to group hazards by risk are described in Attachment 2, Sections 6.1 & 6.2 of draft AC 431.35-2. This section presents specialized methods of analyzing hazards, which possess software influence or causal factors and supplements the HRI presented in draft AC 431.35-2. The Hazard Risk Index presented in draft AC 431.35-2 is predicated on the probability of hazard occurrence and the ability to obtain component reliability information from engineering sources. Hardware reliability modeling of a system is well established; however, there is no uniform, accurate or practical approach to predicting and measuring the software reliability portion of the system. Since software does not fail in the same manner as hardware, in that it is not a physical entity, it does not wear out, break, or degrade over time; software problems are referred to as a software error. Software errors general occur due to implementation or human failure mechanisms (such as documentation errors, coding errors, incorrect interpretation of design requirements, specification oversight, etc.) or requirement errors (failure to anticipate a set of conditions that lead to a hazard). Unlike hardware, software has many more failure paths than hardware, making it difficult to test all paths. Thus the ultimate goal of software system safety is to find and eliminate the built-in unintended and undesired hazardous functions driven by software in a CST system. Classification of Software Safety Risk There are two basic steps in classifying safety risk for software. The first being the establishment of severity within the context of the CST system and then applying an acceptable methodology for determining the software’s influence on system level risks. Refer to Figures 13-4 and 13-5. Regardless of the contributory factors (hardware, software, or human error) the severity of risk as present in draft AC 431.35-2 Attachment 2, Section 6.1.2, Figure 6.1.2, remain applicable criteria for the determination of hazard criticality for those risks possessing software contributory factors. The second half of the equation for the classification of risk is applying an acceptable methodology for determining the software’s influence on system level hazards. The probability factors contained in draft AC 431.35-2 has been determined for hardware based upon historical “best” practices. Data for the assignment of accurate probabilities to software error has not matured. Thus alternate methods for determining probability propagated by software causal factors need to be used. Numerous methods of determining software effects on hardware have been developed and two of the most commonly used are presented in MIL-STD 882C and RTCA DO-178 and are shown in Figure 4. These methods address the software’s “control capability” within the context of the software casual factors. An applicant Software System Safety Team should review these lists and tailor them to meet the objectives of their CST system and integrated software development program. This activity of categorizing software causal factors is for determining both likelihood, and the design, coding, and test activities required to mitigate the potential software contributor. A Software Hazard FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 23 Criticality (SHC) Matrix, similar to the hazard risk index (HRI) 6 matrix is used to determine the acceptability of risk for software hazards. Figure 3 shows an example of a typical SHC matrix using the control categories of MIL-STD 882C [Mil882C]. The SHC matrix can assist the software system safety team in allocating software safety requirements against resources and in the prioritization of software design and programming tasks. Software Hazard Analysis/Risk Mitigation Fault tree analysis (FTA) may be used to trace system-specific software safety-critical functional hazards 7 . The hazard software causal factor nodes are then traced to the appropriate mitigating CST System Requirement, design segment, and coding segment. The hazard should be tracked through the test procedure development; to assure the test procedures have been written sufficiently to demonstrate the hazard is adequately controlled (mitigated). Software Safety Analysis Methods and Tools 8 The following is not intended to be an all-inclusive or exhaustive list of software safety analysis methods and tools; nor does it represent an explicit or implicit AST recommendation thereof. 6 See Attachment 2, Section 6.2 of AC 431.35-2 for discussion and illustration of HRI. 7 The actual analysis techniques used to identify hazards, their causes and effects, hazard elimination, or risk reduction requirements and how they should be met should be addressed in the applicant’s System Safety Program Plan. The System Safety Society’s System Safety Handbook identifies additional system safety analysis techniques that can be used. 8 Reference E FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 24 MIL-STD 882C RTCA-DO-178B (I) Software exercises autonomous control over potentially hazardous hardware systems, subsystems or components without the possibility of intervention to preclude the occurrence of a hazard. Failure of the software or a failure to prevent an event leads directly to a hazard’s occurrence. II(a) Software exercises control over potentially hazardous hardware systems, subsystems, or components allowing time for intervention by independent safety systems to mitigate the hazard. However, these systems by themselves are not considered adequate. II(b) Software item displays information requiring immediate operator action to mitigate a hazard. Software failure will allow or fail to prevent the hazard’s occurrence. III(a) Software items issues commands over potentially hazardous hardware systems, subsystem, or components requiring human action to complete the control function. There are several, redundant, independent safety measures for each hazardous event. III(b) Software generates information of a safety critical nature used to make safety critical decisions. There are several, redundant, independent safety measures for each hazardous event. (IV) Software does not control safety critical hardware systems, subsystems, or components and does not provide safety critical information. (A) Software whose anomalous behavior, as shown by the system safety assessment process, would cause or contribute to a failure of system function resulting in a catastrophic failure condition for the vehicle. (B) Software whose anomalous behavior, as shown by the system safety assessment process, would cause or contribute to a failure of system function resulting in a hazardous/severe major failure condition of the vehicle. (C) Software whose anomalous behavior, as shown by the system safety assessment process, would cause or contribute to a failure of system function resulting in a major failure condition for the vehicle. (D) Software whose anomalous behavior, as shown by the system safety assessment process, would cause or contribute to a failure of system function resulting in a minor failure condition for the aircraft. (E) Software whose anomalous behavior, as shown by the system safety assessment process, would cause or contribute to a failure of function with no effect on vehicle operational capability or pilot workload. Once software has been confirmed as level E by the certification authority, no further guidelines of this document apply. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 25 Figure 13-3: Software Hazard Criticality Matrix* (I) S/W without possibility of intervention– leads directly to hazard occurrence (IIa) S/W with time for intervention– can not stand alone (IIb) S/W displays information but requires operator to mitigate hazard - allow or fail to prevent hazard occurrence. (IIIa) S/W issues commands requiring human action to complete control function– several redundant, independent measures for each event. (IIIb) S/W generate information of a safety critical nature to make safety critical decisions - several redundant, independent measures for each event. (IV) S/W does not control safety critical H/W systems or provide safetycritical information CONTROL CATEGORY CATASTROPHIC CRITICAL MARGINAL NEGLIGIBLE 1 1 1 2 2 3 1 2 2 3 3 4 3 4 4 5 5 5 5 5 5 5 5 5 1 High Risk Significant Analyses and Testing Resources 2 Medium Risk Requirements and Design Analysis and Dept Test Required 3 Moderate Risk High Levels of Analysis and Testing Acceptable with Managing Activity Approval 4 Moderate Risk High Levels of Analysis and Testing Acceptable with Managing Activity Approval 5 Low Risk Acceptable *Extracted from MIL-STD 882C FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 26 It is intended to provide a limited representative sampling of those software safety analysis methods and tools available to the CST licensee or operator. General systems safety analysis have been omitted in that they are addressed in Paragraph 4.3. It is the licensee or operator’s responsibility to assess the applicability and viability of a particular analysis method or tool to their CST, methods of operations, and organizational capabilities. · Code Inspection: a formal review process during which a safety team checks the actual code, comparing it stepwise to a list of hazard concerns. · Hardware/Software Safety Analysis 9 : this analysis is a derivative of the system PHA 10 . The PHA when integrated with the requirements leveled upon the software will identify those programs, routines, or modules that are critical to system safety and must be examined in depth. · Software Failure Modes and Effects Analysis (SFMEA) 11 : identifies software related design deficiencies through analysis of process flow-charting. It also identifies interest areas for verification /validation and test and evaluation. Technique is used during and after the development of software specifications. The results of the PHA and SSHA, if complete, can be used ass a guide for focusing the analysis. · Software Fault Tree Analysis (SFTA) 12 : used to identify the root cause(s) of a “top” undesired event. When a branch of the hardware FTA leads to the software of the system, the SFTA is applied to that portion of software controlling that branch of the hardware FTA. The outputs from the SFMEA, Software Requirements Hazard Analysis (SRHA), Interface Analysis, and Human Factors/Man-Machine Interface Analysis can provide inputs to the SFTA. SFTA can be performed at any or all levels of system design and development. · Software Hazard Analysis (SHA) 13 : used to identify, evaluate, and eliminate or mitigate software hazards by means of a structured analytical approach that is integrated into the software development process. · Software Sneak Circuit Analysis (SSCA) 14 : is used to uncover program logic that could cause undesired program outputs or inhibits, or incorrect sequencing/timing. When software controls a safety critical event, an SSCA can help detect a condition that would cause a catastrophic mishap if the cause were an inadvertent enabling condition. Generic Software Safety Provisions Two recommended sources for the applicant of generic software safety provisions used in the design and development of CST systems that have safety-critical applications are the Joint Software System Safety Committee Software System Safety Handbook and Eastern and Western Range Safety Requirements, (EWR 127-1). Using the generic software safety provision previously discussed and other available software safety “best practices” the applicant should be able to develop system software safety requirements. This should be done early in the software engineering process, in order for software design features to be specified that will eliminate, mitigate, or control hazards/risks at an acceptable level with minimal program impact. 9 Alternate Names: Software Hazard Analysis (SHA) and Follow-On Software Hazard Analysis. 10 See Paragraph 4.3. 11 Alternate Names: Also knows as Software Fault Hazard Analysis (SFHA) and Software Hazardous Effects Analysis (SHEA). 12 Alternate Name: Also know as Soft Tree Analysis (STA). 13 Alternate Name: Software Safety Analysis (SSA). 14 Should be cross-referenced to system SCA. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 27 Design and Development Process Guidelines The following guidelines should be applied to the software design and development process: · A software quality assurance program should be established for systems having safety-critical functions. · At least two people should be thoroughly familiar with the design, coding, testing and operation of each software module in the CST system. · The software should be analyzed throughout the design, development, and maintenance processes by a software system safety team to verify and validate the safety design requirements have been correctly and completely implemented. · The processes as described in the software development plan should be enforceable and auditable. Specific coding standards or testing strategies should be enforced and they should be independently audited. · Desk audits, peer reviews, static and dynamic analysis tools and techniques, and debugging tools should be used to verify implementation of identified safety-critical computing system functions. System Design Requirements and Guidelines The following system design requirements and guidelines should apply: · The CST system should have at least one safe state identified for each operation phase. · Software should return hardware systems under the control of software to a designed safe state when unsafe conditions are detected. · Where practical, safety-critical functions should be performed on a standalone computer. If this is not practical, safety-critical functions should be isolated to the maximum extent practical from non-critical functions. · Personnel not associated with the original design team should design the CST system and its software for ease of maintenance. · The software should be designed to detect safety-critical failures in external hardware input or output hardware devices and revert to a safe state upon their occurrence. · The software should make provisions for logging all system errors detected. · Software control of safety-critical functions should have feedback mechanisms that give positive indications of the function’s occurrence. · The system and software should be designed to ensure that design safety requirements are not violated under peak load conditions. · Applicant should clearly identify an overall policy for error handling. Specific error detection and recovery situations should be identified. · When redundancy is used to reduce the vulnerability of a software system to a single mechanical or logic failure, the additional failure modes from the redundancy scheme should be identified and mitigated. · The CST system should be designed to ensure that the system is in a safe state during powerup. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 28 · The CST system should not enter an unsafe or hazardous state after an intermittent power transient or fluctuation. · The CST system should gracefully degrade to a secondary mode of operation or shutdown in the event of a total power loss so that potentially unsafe states are not created. · The CST system should be designed such that a failure of the primary control computer will be detected and the CST system returned to a safe state. · The software should be designed to perform a system level check at power-up to verify that the system is safe and functioning properly prior to application of power to safety-critical functions. · When read-only memories are used, positive measures, such as operational software instructions, should be taken to ensure that the data is not corrupted or destroyed. · Periodic checks of memory, instruction, and data buss(es) should be performed. · Fault detection and isolation programs should be written for safety-critical subsystems of the computing system. · Operational checks of testable safety-critical system elements should be made immediately prior to performance of a related safety-critical operation. · The software should be designed to prevent unauthorized system or subsystem interaction from initiating or sustaining a safety-critical sequence. · The system design should prevent unauthorized or inadvertent access to or modification of the software and object coding. · The executive program or operating system should ensure the integrity of data or programs loaded into memory prior to their execution. · The executive program or operating system should ensure the integrity of data and program during operational reconfiguration. · Safety-critical computing system functions and their interfaces to safety-critical hardware should be controlled at all times. The interfaces should be monitored to ensure that erroneous or spurious data does not adversely affect the system, that interface failures are detected, and that the state of the interface is safe during power-up, power fluctuations & interruptions, in the event of system errors or hardware failure. · Safety-critical operator display legends and other interface functions should be clear, concise and unambiguous and, where possible, be duplicated using separate display devices. · The software should be capable of detecting improper operator entries or sequences of entries or operations and prevent execution of safety-critical functions as a result. · The system should alert the operator to an erroneous entry or operation. · Alerts should be designed such that routine alerts are readily distinguished from safetycritical alerts. · Safety-critical computing system functions should have one and only one possible path leading to their execution. · Files used to store safety-critical data should be unique and should have a single purpose. FAA System Safety Handbook, Chapter 13: Launch Safety December 30, 2000 13 - 29 · The software should be annotated, designed, and documented for ease of analysis, maintenance, and testing of future changes to the software. Safety-critical variables should be identified in such a manner that they can be readily distinguished from non-safety-critical variables. Configuration Control The overall System Configuration Management Plan should provide for the establishment of a Software Configuration Control Board (SCCB) prior to the establishment of the initial baseline. The SCCB should review and approve all software changes (modifications and updates) occurring after the initial baseline is been established. The software system safety program plan should provide for a thorough configuration management process that includes version identification, access control, change audits, and the ability to restore previous revisions of the system. Modified software or firmware should be clearly identified with the version of the modification, including configuration control information. Both physical and electronic “fingerprinting” of the version are encouraged. Testing Systematic and thorough testing should provide evidence for critical software assurance. Software test results should be analyzed to identify potential safety anomalies that may occur. The applicant should use independent test planning, execution, and review for critical software. Software system testing should exercise a realistic sample of expected operational inputs. Software testing should include boundary, outof-bounds and boundary crossing test conditions. At a minimum, software testing should include minimum and maximum input data rates in worst case configurations to determine the system capabilities and responses to these conditions. Software testing should include duration stress testing. The stress test time should be continued for at least the maximum expected operation time for the system. Testing should be conducted under simulated operational environments. Software qualification and acceptance testing should be conducted for safety-critical functions. References: AST Licensing And Safety Division Directive No. 001, Licensing Process and Procedures dated March 15, 1996. FAA Advisory Circular AC 431-01, Reusable Launch Vehicle System Safety Process, dated April 1999 (Draft) Code of Federal Regulations, Commercial Space Transportation, Department of Transportation Title 14, Federal Aviation Administration, Chapter III, Part 415 – Launch Licenses, and Part 431 – Launch and Reentry of a Reusable Launch Vehicle (RLV) FAA Advisory Circular AC 431-03, Software System Safety (Draft) System Safety Society, System Safety Handbook, 2 nd Edition, dated July 1997 Joint Software System Safety Committee Software System Safety Handbook Eastern and Western Range Safety Requirements, EWR 127-1. The Application of System Safety to the Commercial Launch Industry Licensing Process, FAA/ASY Safety Risk Assessment News Reports No. 97-4 and 97-5
您需要登录后才可以回帖 登录 | 注册


Archiver|航空论坛 ( 渝ICP备10008336号 )

GMT+8, 2024-11-23 23:19 , Processed in 0.029002 second(s), 9 queries .

Powered by Discuz! X2

© 2001-2011 MinHang.CC.

回顶部