Morris method

In applied statistics, the Morris method for global sensitivity analysis is a so-called one-step-at-a-time method, meaning that in each run only one input parameter is given a new value. It facilitates a global sensitivity analysis by making a number r of local changes at different points x(1  r) of the possible range of input values.

Method's details

Elementary effects' distribution

The finite distribution of elementary effects associated with the ith input factor, is obtained by randomly sampling different x from Ω, and is denoted by Fi[1]

Variations

In the original work of Morris the two sensitivity measures proposed were respectively the mean, μ, and the standard deviation, σ, of Fi. However, choosing Morris has the drawback that, if the distribution, Fi, contains negative elements, which occurs when the model is non-monotonic, when computing the mean some effects may cancel each other out. Thus, the measure μ on its own is not reliable for ranking factors in order of importance. It is necessary to consider at the same time the values of μ and σ, as a factor with elementary effects of different signs (that cancel each other out) would have a low value of μ but a considerable value of σ that avoids underestimating the factors.[1]

μ*

If the distribution, Fi, contains negative elements, which occurs when the model is non-monotonic, when computing the mean some effects may cancel each other out. When the goal is to rank factors in order of importance by making use of a single sensitivity measure, scientific advice is to use μ∗,which by making use of the absolute value, avoids the occurrence of effects of opposite signs.[1]

In Revised Morris method μ* is used to detect input factors with an important overall influence on the output. σ is used to detect factors involved in interaction with other factors or whose effect is non-linear.[1]

Method's steps

The method starts by sampling a set of start values within the defined ranges of possible values for all input variables and calculating the subsequent model outcome. The second step changes the values for one variable (all other inputs remaining at their start values) and calculates the resulting change in model outcome compared to the first run. Next, the values for another variable are changed (the previous variable is kept at its changed value and all other ones kept at their start values) and the resulting change in model outcome compared to the second run is calculated. This goes on until all input variables are changed. This procedure is repeated r times (where r is usually taken between 5 and 15), each time with a different set of start values, which leads to a number of r(k + 1) runs, where k is the number of input variables. Such number is very efficient compared to more demanding methods for sensitivity analysis.[2]

A sensitivity analysis method widely used to screen factors in models of large dimensionality is the design proposed by Morris.[3] The Morris method deals efficiently with models containing hundreds of input factors without relying on strict assumptions about the model, such as for instance additivity or monotonicity of the model input-output relationship. The Morris method is simple to understand and implement, and its results are easily interpreted. Furthermore, it is economic in the sense that it requires a number of model evaluations that is linear in the number of model factors. The method can be regarded as global as the final measure is obtained by averaging a number of local measures (the elementary effects), computed at different points of the input space.[2]

See also

References

  1. Andrea Saltelli; Stefano Tarantola; Francesca Campolongo; Marco Ratto (2004). Sensitivity analysis in practice: a guide to assessing scientific models. John Willy & Sons, Ltd. pp. 94–120. ISBN 9780470870938.
  2. Campolongo, F.; Cariboni, J.; Saltelli, A. (2003). "Sensitivity analysis: the Morris method versus the variance based measures" (PDF).
  3. Morris, M.D. (1991). "Factorial Sampling Plans for Preliminary Computational Experiments" (PDF). Technometrics. 33 (2): 161–174. CiteSeerX 10.1.1.584.521. doi:10.2307/1269043. JSTOR 1269043.
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.