Capturing Interactions and Emergent Failure Behavior in Complex Engineered Systems at Multiple Scales

Author(s):  
Nikolaos Papakonstantinou ◽  
Seppo Sierla ◽  
David C. Jensen ◽  
Irem Y. Tumer

Large complex systems exhibit complex nominal and failure behavior and understanding that behavior is critical to the accurate assessment of risk. However, this assessment is difficult to accomplish in the early design stage. Multiple subsystem interactions and emergent behavior further complicate early design risk analysis. The goal of this paper is to demonstrate necessary modifications of an existing function-based failure assessment tool for application to the large complex system design domain. Specifically, this paper demonstrates how specific adaptations to this early, qualitative approach to system behavioral simulation and analysis help overcome some of the challenges to large complex system design. In this paper, a boiling water nuclear reactor design serves as a motivating case study for showing how this approach can capture complex subsystem interactions, identify emergent behavior trends, and assess failures at both the component and system level.

2015 ◽  
Vol 138 (1) ◽  
Author(s):  
Jesse Austin-Breneman ◽  
Bo Yang Yu ◽  
Maria C. Yang

During the early stage design of large-scale engineering systems, design teams are challenged to balance a complex set of considerations. The established structured approaches for optimizing complex system designs offer strategies for achieving optimal solutions, but in practice suboptimal system-level results are often reached due to factors such as satisficing, ill-defined problems, or other project constraints. Twelve subsystem and system-level practitioners at a large aerospace organization were interviewed to understand the ways in which they integrate subsystems in their own work. Responses showed subsystem team members often presented conservative, worst-case scenarios to other subsystems when negotiating a tradeoff as a way of hedging against their own future needs. This practice of biased information passing, referred to informally by the practitioners as adding “margins,” is modeled in this paper with a series of optimization simulations. Three “bias” conditions were tested: no bias, a constant bias, and a bias which decreases with time. Results from the simulations show that biased information passing negatively affects both the number of iterations needed and the Pareto optimality of system-level solutions. Results are also compared to the interview responses and highlight several themes with respect to complex system design practice.


Author(s):  
Joseph R. Piacenza ◽  
Kenneth John Faller ◽  
Mir Abbas Bozorgirad ◽  
Eduardo Cotilla-Sanchez ◽  
Christopher Hoyle ◽  
...  

Abstract Robust design strategies continue to be relevant during concept-stage complex system design to minimize the impact of uncertainty in system performance due to uncontrollable external failure events. Historical system failures such as the 2003 North American blackout and the 2011 Arizona-Southern California Outages show that decision making, during a cascading failure, can significantly contribute to a failure's magnitude. In this paper, a scalable, model-based design approach is presented to optimize the quantity and location of decision-making agents in a complex system, to minimize performance loss variability after a cascading failure, regardless of where the fault originated in the system. The result is a computational model that enables designers to explore concept-stage design tradeoffs based on individual risk attitudes (RA) for system performance and performance variability, after a failure. The IEEE RTS-96 power system test case is used to evaluate this method, and the results reveal key topological locations vulnerable to cascading failures, that should not be associated with critical operations. This work illustrates the importance of considering decision making when evaluating system level tradeoffs, supporting robust design.


Author(s):  
Jesse Austin-Breneman ◽  
Bo Yang Yu ◽  
Maria C. Yang

The early stage design of large-scale engineering systems challenges design teams to balance a complex set of considerations. Established structured approaches for optimizing complex system designs offer strategies for achieving optimal solutions, but in practice sub-optimal system-level results are often reached due to factors such as satisficing, ill-defined problems or other project constraints. Twelve sub-system and system-level practitioners at a large aerospace organization were interviewed to understand the ways in which they integrate sub-systems. Responses showed sub-system team members often presented conservative, worst-case scenarios to other sub-systems when negotiating a trade-off as a way of hedging their own future needs. This practice of biased information passing, referred to informally by the practitioners as adding “margins,” is modeled with a series of optimization simulations. Three “bias” conditions were tested: no bias, a constant bias and a bias which decreases with time. Results from the simulations show that biased information passing negatively affects both the number of iterations needed to reach and the Pareto optimality of system-level solutions. Results are also compared to the interview responses and highlight several themes with respect to complex system design practice.


Author(s):  
Jesse Austin-Breneman ◽  
Bo Yang Yu ◽  
Maria C. Yang

Complex system design requires managing competing objectives between many subsystems. Previous field research has demonstrated that subsystem designers may use biased information passing as a negotiation tactic and thereby reach sub-optimal system-level results due to local optimization behavior. One strategy to combat the focus on local optimization is an incentive structure that promotes system-level optimization. This paper presents a new subsystem incentive structure based on Multi-disciplinary Optimization (MDO) techniques for improving robustness of the design process to such biased information passing strategies. Results from simulations of different utility functions for a test suite of multi-objective problems quantify the system robustness to biased information passing strategies. Results show that incentivizing subsystems with this new weighted structure may decrease the error resulting from biased information passing.


2004 ◽  
Vol 127 (4) ◽  
pp. 536-544 ◽  
Author(s):  
H. Mahmoud ◽  
P. Kabamba ◽  
A. G. Ulsoy ◽  
G. Brusher

The problem of setting, balancing, and determining priorities of design targets among the subsystems constituting an engineering system, i.e., managing the targets, is addressed. A new norm-based benchmarking approach is proposed to relate the system-level design objectives to subsystem design targets. The proposed approach provides a systematic means of setting and balancing subsystem design targets to deliver the desired system performance and ranks the priorities of the subsystem targets. Furthermore, the use of system norms, rather than output signal norms, to quantify system and subsystem performance reduces the number of design targets in multi-input multi-output (MIMO) systems. The approach is illustrated on a vehicle example, consisting of a frame, body, and body mounts as the subsystems.


2012 ◽  
Vol 134 (12) ◽  
Author(s):  
Jesse Austin-Breneman ◽  
Tomonori Honda ◽  
Maria C. Yang

Large-scale engineering systems require design teams to balance complex sets of considerations using a wide range of design and decision-making skills. Formal, computational approaches for optimizing complex systems offer strategies for arriving at optimal solutions in situations where system integration and design optimization are well-formulated. However, observation of design practice suggests engineers may be poorly prepared for this type of design. Four graduate student teams completed a distributed, complex system design task. Analysis of the teams' design histories suggests three categories of suboptimal approaches: global rather than local searches, optimizing individual design parameters separately, and sequential rather than concurrent optimization strategies. Teams focused strongly on individual subsystems rather than system-level optimization, and did not use the provided system gradient indicator to understand how changes in individual subsystems impacted the overall system. This suggests the need for curriculum to teach engineering students how to appropriately integrate systems as a whole.


Author(s):  
David C. Jensen ◽  
Christopher Hoyle ◽  
Irem Y. Tumer

For complex, safety-critical systems failures due to component faults and system interactions can be catastrophic. One aspect of ensuring a safe system design is the analysis of the impact and risk of potential faults early in the system design process. This early design-stage analysis can be accomplished through function-based reasoning on a qualitative behavior simulation of the system. Reasoning on the functional effect of failures provides designers with the information needed to understand the potential impact of faults. This paper proposes three different methods for evaluating and grouping the results of a function failure analysis and their use in design decision-making. Specifically, a method of clustering failure analysis results based on consequence is presented to identify groups of critical failures. A method of clustering using Latent Class Analysis provides characterization of high-level, emergent system failure behavior. Finally, a method of identifying functional similarity provides lists of similar and identical functional effects to a system state of interest. These three methods are applied to the function-based failure analysis results of 677 single and multiple fault scenarios in an electrical power system. The risk-based clustering found three distinct levels of scenario functional impact. The Latent Class Analysis identified five separate failure modes of the system. Finally, the similarity grouping identified different groups of scenarios with identical and similar functional impact to specific scenarios of interest. The overall goal of this work is to provide a framework for making design decisions that decrease system risks.


Author(s):  
Lukman Irshad ◽  
Salman Ahmed ◽  
Onan Demirel ◽  
Irem Y. Tumer

Detection of potential failures and human error and their propagation over time at an early design stage will help prevent system failures and adverse accidents. Hence, there is a need for a failure analysis technique that will assess potential functional/component failures, human errors, and how they propagate to affect the system overall. Prior work has introduced FFIP (Functional Failure Identification and Propagation), which considers both human error and mechanical failures and their propagation at a system level at early design stages. However, it fails to consider the specific human actions (expected or unexpected) that contributed towards the human error. In this paper, we propose a method to expand FFIP to include human action/error propagation during failure analysis so a designer can address the human errors using human factors engineering principals at early design stages. To explore the capabilities of the proposed method, it is applied to a hold-up tank example and the results are coupled with Digital Human Modeling to demonstrate how designers can use these tools to make better design decisions before any design commitments are made.


Author(s):  
David C. Jensen ◽  
Irem Y. Tumer ◽  
Tolga Kurtoglu

Software-driven hardware configurations account for the majority of modern complex systems. The often costly failures of such systems can be attributed to software specific, hardware specific, or software/hardware interaction failures. The understanding of the propagation of failures in a complex system is critical because, while a software component may not fail in terms of loss of function, a software operational state can cause an associated hardware failure. The least expensive phase of the product life cycle to address failures is during the design stage. This results in a need to evaluate how a combined software/hardware system behaves and how failures propagate from a design stage analysis framework. Historical approaches to modeling the reliability of these systems have analyzed the software and hardware components separately. As a result significant work has been done to model and analyze the reliability of either component individually. Research into interfacing failures between hardware and software has been largely on the software side in modeling the behavior of software operating on failed hardware. This paper proposes the use of high-level system modeling approaches to model failure propagation in combined software/hardware system. Specifically, this paper presents the use of the Function-Failure Identification and Propagation (FFIP) framework for system level analysis. This framework is applied to evaluate nonlinear failure propagation within the Reaction Control System Jet Selection of the NASA space shuttle, specifically, for the redundancy management system. The redundancy management software is a subset of the larger data processing software and is involved in jet selection, warning systems, and pilot control. The software component that monitors for leaks does so by evaluating temperature data from the fuel and oxidizer injectors and flags a jet as having a failure by leak if the temperature data is out of bounds for three or more cycles. The end goal is to identify the most likely and highest cost paths for fault propagation in a complex system as an effective way to enhance the reliability of a system. Through the defining of functional failure propagation modes and path evaluation, a complex system designer can evaluate the effectiveness of system monitors and comparing design configurations.


Sign in / Sign up

Export Citation Format

Share Document