Stochastic Sensitivity Analysis for Models with Parameter

Transcription

Stochastic Sensitivity Analysis for Models with Parameter
Journal of Information & Computational Science 12:2 (2015) 657–672
Available at http://www.joics.com
January 20, 2015
Stochastic Sensitivity Analysis for Models with Parameter
Correlation and Model Approximation Method ?
Xian Dong a , Zhan Wang a,b , Jianrong Pan a,b,∗
a Department
b State
of Civil Engineering, South China University of Technology, Guangzhou 510640, China
Key Laboratory of Subtropical Building Science, South China University of Technology
Guangzhou 510640, China
Abstract
In the stochastic sensitivity analysis, a large number of simulation models lead to low computational
efficiency. With the dimension increasing in assigned problems, the accurate is difficult to achieve
by popular regression methodologies. Without considering the correlations of parameters, inaccurate
sensitivity coefficients would be calculated by analyzing the effect of parametric variables on structures,
moreover, the existing methods only calculate the local gradient as sensitivity. According to these
problems, approximate model and global sensitivity method are employed for design sensitivity analysis
of structures. The approximate model is constructed by the hybrid neural network which possesses
significant learning capacity and generalization capability with a small amount of information. The
improved chaotic particle swarm is applied to optimize the neural network so as to improve the
computational efficiency and accuracy. The proposed sensitivity analysis method values the global
response of the outputs by varying all the input parameters at a time with the correlations of parameters.
Uniform design and Latin hypercube sampling are used to sample points. Numerical analysis show that
the proposed method can successfully measure the actual sensitivity.
Keywords: Approximate Model; Sensitivity Coefficient; Correlation between Parameters; Improved
Chaotic Particle Swarm; Stochastic
1
Introduction
System optimization or design reinforcement require estimating the influence of various parameters on structures, that is sensitivity analysis. The actual effect of engineering structures caused
by mechanical properties of material, stochastic external loads and stochastic section errors would
be inadvertent through deterministic parameters models, therefore the uncertainty and stochastic
of parameters [1, 2] need to be considered in complicated structures. In recent years, the methods
?
Project supported by the National Nature Science Foundation of China (No. 51178192, 51378009, 51378219),
and Key Laboratory of Subtropical Building Science Foundation, South China University of Technology
(No. 2012ZA05).
∗
Corresponding author.
Email address: [email protected] (Jianrong Pan).
1548–7741 / Copyright © 2015 Binary Information Press
DOI: 10.12733/jics20105305
658
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
of uncertainty and sensitivity analysis for structures have been developed by domestic and foreign scholars including Monte Carlo simulation [3], stochastic finite element method [4], response
surface method [5, 6] and so on.
The major advantage of Monte Carlo simulation is simplicity in formulation, however there
is a consequence of this method that too many complex models cause high computational cost.
The polynomial functions for inputs and outputs would be established based on response surface
method which is convenient for compiling calculation programs, however it is nearly impossible
to achieve the satisfactory precision when the parameters of function having highly nonlinear
relation. Stochastic finite element method which including perturbation stochastic finite element
method and spectral stochastic finite element method has complex random theories. Variation
coefficients would be limited in a special range conducting analysis by the method of perturbation
stochastic finite element.
Analysis of structure stochastic sensitivity needs for a large amount of simulations, computational cost is higher for the structure is more complex, so far, the most effective way is to replace
the physical models by establishing the approximate models [7, 8]. The sensitivity techniques have
been classified into two groups: local sensitivity analysis methods and global sensitivity analysis
methods. The local sensitivity analysis methods reflect the local response of the output(s) by
varying input parameters one at a time then holding other parameters at determined values, the
global sensitivity analysis methods reflect the global response of the output(s) in a finite region
[9]. The local sensitivity analysis technique has been widely used for easily implement, but the
effect of global parameters would be covered for physical parameters with different dimension
magnitudes. The correlation of parameters as an essential component effects the output results
in complex structure systems, in addition, the inaccurate results would be obtained by neglecting
the parameter relativity. By now the sensitivity analysis methods have evolved as a major area
of research, but only a few studies have considered the correlations between input parameters in
the sensitivity analysis of structures.
In this paper, the analysis of stochastic parameters sensitivity can be divided into two procedures: approximate model representation and stochastic sensitivity analysis. In Section 2, we
use hybrid neural network (optimized by the method of improved Chaotic particle swarm) to
approximate the structure models so as to improve the computational efficiency, avoid computing
large finite element models. As an application of improved chaotic particle swarm to optimize the
neural network, the high magnitudes nonlinear function of complex structure would be built by
limited samples. Section 3 transforms the correlated variables to independent normal variables
equivalently by nataf model, Section 4 develops a stochastic sensitivity method to estimate the
global effect of numerous parameters based on varying input parameters at the same time. Section
5 tests some representative examples and shows that the proposed procedure is useful in solving
the problems that neural network is easily to fall into local minimum and slow convergence etc.
2
2.1
Method of Establishing Approximate Model
Algorithm of Feed-forward Neural Network with Back-propagation
Learning
Neural network has the behaviors of self-learning, self-organization and simulate nonlinear
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
659
relationship, it will simulate the complicated nonlinearly relation between inputs and outputs
based on the sample space divided by continuous hypersurface which is constituted of bounded
continuous differentiable functions. The fundamental of feed-forward neural network is it adjusts
the weights and threshold value by error back propagation, which takes use of errors caused by
the forward propagation of input layer, hidden layer and output layer feeding back to neurons of
each layer.
(1) The mean square error is the sum of each neuron error when the output has been derived
from the network, one sample error E is given by:
n1 −1
1X
E=
(dl − yl )2
2 l=0
(1)
The sum error of all the samples is given by:
n
Ec =
n −1
1
1XX
(dl − yl )2
2 n=1 l=0
(2)
The network weight is adjusted by gradient descent method
∂Ec
4wik = −
∂wik
(3)
where dl is the sample output, yl is the network output.
(2) When the output layers have iterated t times the equations are given by:
∂Ec
0
∂wkl
0
p
p
p
X
X
X
0
0
∂Ec
∂E p
∂E p ∂ylp ∂ulp
0
=
=
=−
(dpl − ylp )f (ulp )xkp
0
0
0
0p
p
∂yl ∂ul ∂wkl
∂wkl
∂wkl
p=1
p=1
p=1
0
0
wkl (t + 1) = wkl (t) − η
=−
p
X
0
(dpl − ylp )ylp (1 − ylp )xkp
p=1
=
− ylp )ylp (1
t
X
0
0
0
∂ulp =
wkl xkp
k=0
0
p
yl = f (ulp )
p
δkl
(dpl
(4)
(5)
− ylp )
(6)
(7)
(8)
0
only affects yl .
where p is the number of samples, E p is a function about y0 , y1 , . . . , ym−1 , wkl
(3) Hidden layers neurons are trained by incremental learning rules such as:
p
X
∂Ec
p p
δik
xi
=−
∂wik
p=1
wik (t + 1) = wik (t) − η
p
=
δik
m−1
X
l=0
0
0
(9)
∂Ec
∂wik
0
p
wik xip (1 − xip )
δik
(10)
(11)
660
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
wik (t + 1) = wik (t) + η
p
X
p p
δik
xi
(12)
p=0
The input layers, hidden layers and output layers are connected by the neural network with the
map of n-dimension space to m-dimension space. The number of hidden neurons is n1 , wik is the
weight of input layers to hidden layers, bk is the threshold value of input layers to hidden layers,
0
0
wkl is the weight of hidden layers to output layers, bl is the threshold value of hidden layers to
output layers, f is a transfer function. The calculation process is expressed as:

n−1
P

 x0k = f (
wik xi − bk )
k = 0, 1, 2, . . . , n1 − 1

i=0
(13)
n−1
P 0 0
0


l = 0, 1, 2, . . . , m − 1
wkl xk − bl )
 xl = f (
k=0
The learning rate, momentum coefficient and other parameters have a serious impact on the
feedforward network. The network will fall into local extremum and could not obtain the optimal
weight distribution for the improper values of these parameters, therefore, in this paper the
network will be optimized by chaos particle swarm algorithm.
2.2
The Algorithm of Traditional Chaotic Particle Optimization
Particle swarm optimization originated from the behavior of biotic population and has widely used
to solve optimization problems. In this algorithm, any candidate solution of optimization can be
analogy to a particle in v-dimensional search space. These particles moved in a v-dimensional
search space with a certain speed which adjusted by the positions and velocities of its own
and companions’. The qualities of all the particles are evaluated by a fitness function which is
determined by the optimization goal. Each particle is influenced by the fitness function moving
toward to its best local position, therefore the swarm moves toward to the best global position in
an identified space.
A biotic population modal is represented as X = (x1 , x2 , ..., xk )T consisting of k particles in vdimensional search space, the ith particle’s position is xi = (xi,1 , xi,2 , ..., xi,v )T and the velocity is
vi = (vi,1 , xi,2 , ..., xi,v )T . The equations describing the process of the particle swarm optimization
are given by:
j+1
j
vi,v
= wvi,v
+ m1 rand()(pbestji,v − xji,v )m2 rand()(gbestji,v − xji,v )
(14)
j
j+1
xj+1
i,v = xi,v + vi,v
(15)
where rand() is a random number in interval (0, 1), m1 , m2 are acceleration constants. Where
v
denotes the ith0 s velocity in v-dimensional space which iterated in the jth time, xji,v is the
vi,j
ith0 s position in the vth dimensional space which is iterated in the jth time; pbestji,v denotes the
best position of the particle i in the vth dimensional space, gbestji,v is the global best position of
the swarm in the vth dimensional space. w is a decline inertial weight decreasing from 0.9 to 0.5
[10] with the increasing iteration times and w can be expressed as:



 m1 rand() ∈ [0, 2]
m2 rand() ∈ [0, 2]


k
 w(k) = −0.5
+ 0.9
maxnumber
(16)
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
661
Chaos is a motion state, that it is seemingly irregular but determined and unpredictable ubiquitous in certainty nonlinear dynamical systems. The characteristic of chaotic system is extremely
sensitive to infinitely-small initial changes and perturbation. It will deviate from the original
direction of evolution after prolonged exercise even by a condition of tiny disturbance. The
unpredictability of the chaotic system derives from unstable state belonging to the system.
One of typical chaos system is logistic mapping model. Suppose that xn is a parameter related
to the individual number of an insect in the nth year and xn+1 is the number of the (n + 1)th
year, both xn and xn+1 are all integers. The equation about the parameters xn and xn+1 can be
described as v = f (xn ), n = 1, 2, . . ., where n is a nonnegative integer. Logistic model can be
given as:
xn+1 = µxn (1 − xn ), n = 1, 2, . . . ,
(17)
where µ is a control parameter deciding the state of the system, when the value rang of µ is [3, 4],
the dynamic action of the system including two states period-doubling and chaos. If µ takes the
value range [3.56994567, 4], the system will get into chaotic state, if µ = 4, the system will be
totally in chaotic state, the algorithm which making the swarm chaos with µ = 4 can be carried
out as following steps:
(1) Map xi,v to the interval [0, 1.0], xi,v is the location of particle xi in the vth dimension, [av , bv ]
is the domain of xi in the vth dimension.
hxi,v =
xi,v − av
b v − av
(18)
= 4hxji (1−hxji )
(2) Generate a chaotic sequence h1x,v , h2x,v , . . . , hJx,v by iterating an equation hxj+1
i
for j times;
(3) Map particles of the chaotic sequence to the primary searching space following the function
xji,v = av + hji,v (bv − av ), chaos column points xji = (hji,1 , hji,2 , . . . , hji,n ) in each dimensional
chaotic sequence can be derived from Tent mapping.
2.3
Improved Chaotic Particle Swarm Optimization
The phenomenon of particle swarm premature convergence is increasingly outstanding when solving complex optimization problems. So far the main method to solve the problem is increasing
the size of the particle swarm, but this approach could not solve this problem fundamentally.
Chaotic variables which are stochastic motion state variables obtained by deterministic equations. Using chaotic variables to optimize the searching process can jump out of local minima,
keep the diversity of population, improve global optimization performance [11].
There will be no significance in the process of searching the optimal solution if the new particles
which generated by chaotification always in the state of chaos. The algorithm can be constantly
close to the optimal solution only under the alternate states of chaos and stability [12].
Fitness functions according to the objective function are established to estimate whether the
particles are in premature convergence or not. In this paper, we set the relative errors of the
network outputs and the actual response as the fitness functions based on the algorithm of back
662
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
propagation. To insure the functions are nonnegative, the expressions are represented as follows:
v
u n
n
.X
uX
2
t
u(x) =
(yi − f (xi ))
f (xi )
i=1
yi = f
n
³X
0
0
0
´
wki xk − bi
k=0
0
xk = f
m
³X
(19)
i=1
(20)
´
wlk xl − bk
(21)
l=0
where u(x) is the judgment function of premature, yi is the network output, xi is the network
input, x0k is hidden layer neuron.
The status of particle swarms that it is in steady state or chaos is determined by the movement
distance of current particle, the equation about it can be given as:
smove = abs(xi (t) − xi (t − 1))
(22)
where t is a number of the particle evolutionary generations. It is considered that algorithm has
reached to the local optimization when smove or u(x) is less than a limit value, and the variables
will get into chaotization by chaotic searching. The concrete steps are represented in Section 2.2.
2.4
Optimization of Neural Network by Chaotic Particle Swarm Algorithm
Encode matrixes of every particle by coding strategy, the matrixes are consisted of two kinds
particles, the first kind particles are the weights of input layers to the hidden layers and hidden
layers to the output layers, the second kind particles are the threshold values of input layers to
hidden layers and hidden layers to the output layers. These matrixes are expressed as:
P article(1) = [W1 , W2 ]


w1,1
w1,2
. . . w1,n

 w2,1 w2,2 . . .

W1 =  .
..
..
 ..
.
.

wm,1 wm,2 . . .

w
wn+1,2
 n+1,1
 wn+2,1 wn+2,2

W2 = 
..
..

.
.

wn+k,1 wn+k,2
w2,n
..
.






wm,n
. . . wn+1,m
. . . wn+2,m
..
..
.
.
. . . wn+k,m
(23)







(24)
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
P article(2) = [Z1 , Z2 ]

663

z1,1

 z2,1

Z1 =  .
 ..

zm,1

z
 1,k+1
 z2,k+1

Z2 = 
..

.

zm,k+1
z1,2
. . . z1,k
z2,2
..
.
. . . z2,k
..
..
.
.






zm,2 . . . zm,k
z1,k+2
. . . z1,k+g
z2,k+2
..
.
. . . z2,k+g
..
..
.
.
(25)







(26)
zm,k+2 . . . zm,k+g
where n is the dimension of input layer variables, m is the number of hidden layer neurons, k is the
dimension of output layers, g is the population size. The computational process of optimization
hybrid neural network are described as follows:
(1) Make parameters initialized. Assuming that the size of particle population is G, inertia weight
is w, learning factors are m1 and m2 , the maximum value of particle swarm evolutionary
generations is T , the energy function is E, the iteration number of chaotic map is Q, the time
of function initial changed is f = 0, the maximum number of energy changing times of PSO
is fmax , the judgment of premature convergence is u, the judgment of steady state is s. Set
the value of current evolutionary generation is t = 1, the weights and the threshold values of
the network are separately represented by two kinds of particles which randomly generated
in feasible region.
(2) The energy functions are constructed by the fitness function and the movement distance
function of particles. The global best position of initial particles swarm which is expressed
as gbest would be selected by the initial fitness value. The best position of current particles
swarm is expressed as pbest.
(3) The position and velocity of the new particle are updated according to the equations14 and
15. After estimating the energy values of the new generated particles, go straight to step 4
for the ith particle value is higher than value u and s, if the energy value of the ith particle
is lower than value u or s, selects the position of the particles’ lowest energy value to be the
new position that the particles are q chaotic column points which have generated by logistic
map or tent map.
(4) Evaluate the energy values of the new particles are estimated, if the energy values are lower
than the values which are in before the update, the value of pbest and the corresponding
position will be updated, if the lowest energy value is lower than gbest, the value of gbest and
the corresponding position will be updated and then f = 0, otherwise, f = f + 1.
(5) If f ≥ fmax , the error back propagation is used to have a local search near the gbest. The
searching result will instead of pbest and update the corresponding position for it is better
than gbest, otherwise, the searching result will instead of the worst particle of pbest, f = 0,
t = t + 1. The process is finished when t ≥ T , otherwise, go back to step 3.
(6) Save the results and optimize the network.
664
3
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
Transformation between Correlated Variables and Independent Variables
In order to achieve the equivalent transformation between correlated variables and independent
variables, there are three problems to be solved, the first is Joint probability density functions of
correlated random variables could not be determined; the second is correlation coefficients would
have changed with the normalization of random variables; the third is changing the correlation
coefficients to zero is not equivalent to transform correlated variables to independent variables,
in order to conquer these problems, Section 3.1 and 3.2 would discuss in detail.
3.1
Transformation between Correlated Non Normal Variables and
Correlated Normal Variables
That normal variables are independent when the correlation coefficients are zero is a mathematical
theorem, so firstly, converts non-normal random variables to normal random variables by the
functions of equal probability transformation. Consider a random vector X = [X1 X2 . . . X3] and
a standard normal distribution vector Y = [Y1 Y2 . . . Y3 ]. Then the relationship between random
distribution and normal distribution are given by:
(
Fi (xi ) = Φ(yi )
i = 1, 2, . . . , n
(27)
xi = Fi−1 [Φ(yi )]
Joint probability density functions deriving from the rules of implicit function derivation are
expressed as follows:
f1 (x)f2 (x) . . . fn (x)
fx (x) =
φn (y, ρ0 )
(28)
φ(y1 )φ(y2 ) . . . φ(yn )
1 T −1
1
e( 2 y ρ0 y)
φn (y, ρ0 ) = p
(29)
n
(2π) det(ρ0 )
where Fi (·) is a marginal probability distribution function, Fi−1 [·] is the inverse function of Fi (·),
Φ(·) is a standard normal distribution function, ρ0 is the equivalent correlation coefficient matrix
of Y , fi (·), i = 1, 2, . . . , n is the marginal probability density function corresponding with Fi (·),
φ(·) is the probability density function of standard normal variables, det(·) represents the matrix
determinant, (·)−1 is the function of matrix inversion. The correlation coefficient ρ0xy of standard
normal random vector y can be derived from the following equation:
Z +∞ Z +∞ ³
xi − µxi ´³ xj − µxj ´
fXi ,Xj (xi , xj ) dxi dxj
ρij =
σxi
σxj
−∞
−∞
(30)
Z +∞ Z +∞ ³ −1
−1
Fi (Φ(yi )) − µxi ´³ Fj (Φ(yj )) − µxj ´
=
φ2 (yi , yj , ρ0ij ) dxi dxj
σxi
σxj
−∞
−∞
where ρij is the correlation coefficient between xi and xj of random vector X, ρ0ij is the correlation
coefficient matrix component of standard normal random vector y, fXi Xj (xi , xj ) is the joint probability density function between the ith and jth variables. The matrix ρ0 could be determined
by the nonlinear function (30) for the marginal probability density function of random variables
and ρ have been calculated.
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
3.2
665
Transformation between Correlated Normal Variables and Independent Normal Variables
Suppose that the correlation coefficient matrix ρ0 have been obtained by Section 3.1, for an
intermediate variable M is represented as:
√
M =V · Λ
(31)
where Λ is a diagonal matrix composed by the eigenvalues λ of the correlation coefficient matrix
ρ0 , V is a matrix composed by the eigenvectors v corresponding with the eigenvalues λ. The
relationship between independent standard normal variables U and correlated standard normal
variables Y can be represented as follows:
U = M −1 · Y
4
(32)
Stochastic Sensitivity Estimation for Structure Systems
Using Approximate Model and Monte Carlo
To analyze stochastic response of structure by Monte Carlo method needs large number of simulations. In this paper, hybrid neural network is used to establish a reasonable network between
random variables and structure response with selecting small-scale certain samples, in addition,
the hybrid neural network is optimized by the algorithm of improved chaos particle swarm, moreover enormous workload would be avoided by the approximate model.
4.1
4.1.1
The Sampling Strategy of the Approximate Model
Uniform Design Sampling Method for Neural Network
Uniform design [13] is a method that setting the experiment points uniformly, it’s algorithm is
the combination between number theory and multivariate statistical. More information about
the experiment model can be obtained from the fewer experiments by uniform design, so it is
suitable for the multifactorial experiments and the model system parameters which are completely
unknown. The regression model is given by:
J = E(x1 , x2 , . . . , xn ) + ε
(33)
where E is a known function class, the value intervals of xn are C n ⊂ [0, 1]n .
The uniformity of the design points is measured by deviation. Assume that γ = {xi | i =
1, 2, . . ., n} are the n test points of C n , g is an arbitrary function of E satisfy the conditions of
g ∈ C n , N (g, ) is the number of points which are in the range g ∈ C n , the deviation of point set
y in C n is
¯
¯
¯ N (g, γ)
¯
e(n, γ) = sup ¯¯
− (ν[0, g])¯¯
(34)
n
(g∈C n )
where v([o, g]) denotes the rectangular volume which consists of n points, sup denotes the upper
bound. In this section, we adopt good lattice point method to construct an algorithm which
666
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
making the points distributing uniformly in the range of experiment. (1) Make an assumption
that n is the experiment times, c is a number that it is less than n, C is a vector consist of
positive integers c that the greatest common divisor between n and c is 1. (2) The jth column
of uniform design table is consist of uij = icj [modn], if icj is more than n, subtracting h which
is an appropriate time of n from icj and the difference between icj and h is in interval [1, n], the
uij are given by:
 j

u1 = c j


(


uij + cj
uij + cj ≤ n
ui+1,j =

uij + cj − n
uij + cj ≥ n




i = 1, . . . , n − 1
4.1.2
(35)
Process of Random Sampling
Latin Hypercube Sampling (LHS) is a method for generating a sequence of points from multidimensional distributions [14]. The distinction between Latin hypercube sampling and direct
sampling is that taking advantage of LHS to avoid sampling repeatedly. For an assumption that
the range of each variables is divided into n equal probability intervals, F (xi ) is a probability
distribution function of random variables. The n + 1 boundary points which have been generated
in domain are given by:
³k´
(k)
k = 1, 2, . . . , n + 1
(36)
bi = Fi−1
n
The samples in each equal probability interval are expressed as:
(k)
xi
=
Fi−1
³ k − 0.5 ´
n
k = 1, 2, . . . , n
(37)
The sampling process of considering the correlation between parameters are listed as below
(1) Generate n-dimensional random variables which follow normal distribution, assume a computational accuracy θ, n-dimensional random variables X;
(2) Calculate the correlation coefficient matrix of normal distribution random variables;
(3) Transform correlated random normal variables to random independent variables by the algorithm of Section 3;
(4) Calculate the correlation coefficient matrix ρ of the original model, obtain the equivalent
correlation coefficient matrix ρ0 according to the literature [15];
(5) Generate random variables following normal distribution with the correlation coefficient ρ0
according to the inverse transformation of Section 3, calculate the absolute values of the
differences between correlation coefficient matrix of the generated variables and ρ0 , if all the
differences are less than θ, go to step 6, otherwise, go back to step 2;
(6) Transform the correlated random normal variables correlation coefficient matrix ρ0 to the
correlated random non-normal variables based on the equal probability transformation.
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
4.2
667
Calculation of Structure Sensitivity Coefficients
The sensitivity of each parameter is determined by the output variations corresponding to the
range of each input, the outputs are the predicted results based on neural network and the inputs
are the random samples according to probability density function. Considering the impact of each
random variable on structure response, correlation coefficients are used to measure the correlation
between random variables and structure response, the coefficients of variation are brought in to
eliminate the effect of the random variable dimensions and measurement scales. The equations
are given by:
n
P
(xi − x)(f (xi ) − f (xi ))
i=1
Rr = r n
(38)
P
2
2
(xi − x) (f (xi ) − f (xi ))
i=1
v
u³ u
X Rr ´
Rr u
Sr =
/t
γr
γ
r=1 r
(39)
where Rr is the Pearson correlation coefficients, xi is random variable, n is Monte Carlo sampling
times, f (xi ) is the network output according to xi , x is the mean value of xi , f (xi ) is the mean
value of f (xi ); γr is the coefficient of variation according to Monte Carlo random samples, u is a
number represented the random variables total dimensions, Sr is the sensitivity value of random
variables, the absolute value of coefficient reflects the influence of random variables on structure
response, positive and negative reflect the degree of correlation between random variables and
structure response. The flow chart of structure stochastic sensitivity analysis based on Section 2,
Section 3 and Section 4 is shown in Fig. 1.
5
Examples and Analysis
In this section, comparing approximate models which are established by the method of Section
2 with Monte Carlo to verify the accuracy and then two engineering examples are selected to
calculate sensitivity values by stochastic analysis. To assess the accuracy of the meta-model,
the values of 10000 additional test samples which obtained by Monte Carlo are employed approximately as exact solutions. It will demonstrate the superiority of this algorithm since the
performance functions of engineering structures are complex implicit functions. The Monte Carlo
results considering relativity are abbreviated to MC, the results of hybrid neural network are
abbreviated to HNN.
5.1
5.1.1
Verify the Approximate Model
Example 1
Considering a non-linear function with three uniform random variables, f (x) = x21 + x22 + x23 +
x1 x2 x3 , x1 ∈ [10, 20], x2 ∈ [15, 25], x3 ∈ [5, 15]. The obtained results are presented in Table 1,
the frequency histogram are shown in Fig. 4.
668
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
Determine
sampling number
Particle swarm
initialization
Not
satisfy
Sample according to
the variable distributions
Not
satisfy
Random analysis of the
structure by Monte Carlo
Save the network
structure
Satisfy
The operation of
particle swarm (a)
Judge the accuracy
Stochastic sensitivity
analysis
Judge premature
and steady
Output the stucture
response
Y
Generate chaotic
sequences
Construct the network
of structure
Calculate the structure
response
The operation of
particle swarm (b)
Optimize the neural
network
Sample according to
the sampling strategy
Output the final
particles
Determine the statistical
parameters and distributions
Judge limit
value of iteration
N
Satisfy
Fig. 1: Stochastic sensitivity analysis flow char
Table 1: Analysis results of the structure response (example 1)
Statistics
Sample number
Mean value (mm)
Standard deviation
Compute time (min)
MC
10000
3778.9
1287.6
1.67
HNN
50
3780.9
1286.9
0.18
Relative error
/
0.053%
0.054%
/
5.1.2
Example 2
A frame structure whose beam and column are all I shaped cross section is acted upon by horizontal load P (Fig. 2). Assuming that the parameters of I shaped cross section and load are
random input variables with uniform distributions and the distributions of these parameters are
shown in Table 2, transverse displacement of the joint 3 is the output variable. The structure
stochastic is analyzed based on these parameters. The analysis results of the structure response
are showed in Table 3 and the frequency histogram is shown in Fig. 5.
Table 2: Distribution parameters of random variables (example 2)
—
Value interval
Beam height
Beam width
Beam flange
Beam web
Load
(mm)
(mm)
thickness (mm)
thickness (mm)
(KN)
200-300
120-180
8-12
8-12
14-20
669
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
1-1
P
2
3
10
10
250
4m
150
1-1
1
4
Fig. 2: Schematic of truss structure
Table 3: Analysis results of the structure response (example 2)
Statistics
Sample number
Mean value (mm)
Standard deviation
Compute time (min)
MC
10000
23.7
9.8
16.67
HNN
50
23.51
9.61
0.1
Relative error
/
0.8%
1.94%
/
5.1.3
Example 3
Joint is the basic component of frame structure. According to the mechanics of joints, all kinds
of joints have the characteristic of semi-rigid. The prerequisite of researching the semi-rigid joints
is to explore the impact of each component of joint. So a welded steel joint is utilized as the
model. The connection of joint is fully welded, the flange connection is slop butt weld and the
web connection is twin fillet welding seam (Fig. 3). Assume that the beam height, beam width,
beam flange thickness, beam web thickness and column flange thickness are input variables with
uniform distribution and the distributions of these parameters are shown in Table 4, the joint
rotation is output variable. The analysis results of the structure response are shown in Table 5
and the frequency histogram is shown in Fig. 6. By comparing the results of HNN with the results
of MC in Table 1, 3 and 5, it is evident that the results of HNN meet requirement of the accuracy,
while requiring much fewer model runs. Such as in example1, the relative error of the mean
value between MC and HNN is 0.053%, the relative error of the standard deviation is 0.054%; in
example 3, the compute time of HNN is 0.92 hour while the compute time of MC is 166.67 hour.
The frequency histograms of Fig. 4, Fig. 5 and Fig. 6 indicate that the approximations are close
to the true distributions, moreover HNN proves useful for the sensitivity analysis of complicated
models, the more complex the model, the more obvious advantages of the algorithm.
Table 4: Distribution parameters of random variables (example 3)
—
Value interval
Beam height
Beam width
Beam flange
Beam web
Load
(mm)
(mm)
thickness (mm)
thickness (mm)
(KN)
200-300
120-180
8-14
6-14
14-20
670
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
Table 5: Analysis results of the structure response (example 3)
Statistics
Sample number
Mean value (mm)
Standard deviation
Compute time (min)
MC
10000
5.06
0.68
166.7
HNN
50
5.08
0.72
0.92
Relative error
/
0.8%
1.94%
/
16
300
16
800
30 140 30
MC calculations
Probability frequency
700
5
Beam
5
8
Joint region
234
8
Column
HNN calculations
600
500
400
300
200
100
0
1000 2000 3000 4000 5000 6000 7000 8000 9000
V
Fig. 3: Schematic diagram of joint
Fig. 4: Frequency histogram (example 1)
700
800
MC calculations
MC calculations
600
HNN calculations
Probability frequency
Probability frequency
700
600
500
400
300
200
10
20
30
40
50
60
70
Displacements of the third joint (mm)
Fig. 5: Frequency histogram (example 2)
5.2
500
400
300
200
100
100
0
0
HNN calculations
80
0
3.5
4.0
4.5
5.0
5.5 6.0
Rotation (1e-4)
6.5
7.0
7.5
Fig. 6: Frequency histogram (example 3)
Stochastic Sensitivity
Calculate the sensitivity coefficients of example 2 and example 3 by the algorithm of section
the correlation coefficients between beam height and beam width are both 0.8 in example 2 and
example 3. The sensitivity results are shown in Fig. 7 and Fig. 8. BW is the beam width, BH is
the beam height, BFT is the beam flange thickness, BWT is the beam web thickness, CFT is the
column flange thickness, P is the load.
The sensitivity coefficients in Fig. 7 reflect the positive and negative correlation between pa-
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
0.4
0.4
0.2
0.2
0
0
−0.2
−0.2
−0.4
−0.4
−0.6
Sensitivity without considering
parameters correlation
Sensitivity considering parameters
correlation
−0.8
−1.0
BW
BH
BFT
BWT
P
Fig. 7: Sensitivity coefficients (example 2)
−0.6
Sensitivity without considering
parameters correlation
Sensitivity considering parameters
correlation
−0.8
−1.0
671
BH
BW
BFT
BWT
CFT
Fig. 8: Sensitivity coefficients (example 3)
rameters and structure response. When the parameters correlation is not considered, beam height
and beam web thickness are positively correlated with joint 3 displacement, that is to say, the
displacement of joint 3 is increased with an increase of beam height and web thickness. The
sensitivity coefficients considering parameters correlation are −0.7215, −0.5315, −0.334, −0.106
and 0.2722 for BW, BH, BFT, BWT and P in turn. However this conclusion is obviously consistent with the actual situation that all the parameters are negatively correlated with displacement
of joint 3 except the external load with considering parameters correlation. Fig. 8 reflects the
differences between considering correlation and without considering correlation. In the condition
of considering parameters correlation, BH is the most important sensitive parameter for rotation,
other sensitive parameters are BW, CFT, BWT and BFT in order of importance. The sensitivity
coefficient of BW is 0.2 under the condition of without considering parameters correlation, that
is mean the rotation is increased with increasing the beam width and beam flange thickness. It
is obvious false for without considering parameters correlation. However all the parameters are
negatively correlated with the rotation for considering parameters correlation is accordance with
the actual form mechanics.
6
Conclusions
This study suggests hybrid neural network as the approximate model and global sensitivity technique to stochastic sensitivity problems. The conclusions are as follows:
(1) The algorithm of chaos-particle swarm has been improved by optimizing the particles alternately between steady and chaos. The proposed technique that hybrid neural network was
optimized by the improved chaos-particle swarm provided a satisfactory approach for highly
non-linear implicit functions.
(2) Agreement appears much closer between these results which were calculated by hybrid neural
network and the Monte Carlo results. In addition, these results of approximate models could
be used for sensitivity analysis in probability design.
(3) Compared with Monte Carlo which would obtain complex structure responses by finite element, numerical analysis has demonstrated that the algorithm in this paper shows great
672
X. Dong et al. / Journal of Information & Computational Science 12:2 (2015) 657–672
superiority in computational efficiency.
(4) The effect of the random variable dimensions and measurement scales on structure sensitivity
can be removed by the combination of global sensitivity algorithm and stochastic analysis.
The parameter sensitivities are then obtained numerically using the global method. The calculations of the sensitivity coefficients considering parameters correlation are more consistent
with the actual.
References
[1]
J. Choi, J. W. Harvey, M. H. Conklin, Use of multi-parameter sensitivity analysis to determine
relative importance of factors influencing natural attenuation of mining contaminants [C], Proceedings of the Technical Meeting, Charleston, South Carolina, March, 1999, 8-12
[2]
M. Horsen, J. C. Refsgaard, S. Hansen, Assessment of uncertainty in simulation of nit rate leaching
to aquifers at catchment scale [J], Journal of Hydrology, 242(3), 2001, 210-227
[3]
Phelim Boyle, Mark Broadieb, Paul Glasserman, Monte Carlo methods for security pricing [J],
Journal of Economic Dynamics and Control, 21(8), 1997, 1267-1321
[4]
Daojin Lin, Quan Qin, Reliability stochastic finite element analysis of in-plane buckling of an
existing arch bridge [J], Engineering Mechanics, 22(6), 2005, 122-126
[5]
Ran Hu, Dianqing Li, Chuangbing Zhou et al., Structural reliability analysis using stochastic
response surface method [J], Engineering Mechanics, 27(9), 2010, 192-200
[6]
Lufeng Yang, Xianfeng Yang, Bo Yu et al., Improved stochastic response surface method based on
stepwise regression analysis [J], Chinese Journal of Computational Mechanics, 30(1), 2013, 88-93
[7]
Long Tang, Guangyao Li, Hu Wang, Kriging-HDMR meramodeling technique for nonlinear problems [J], Chinese Journal of Theoretical and Applied Mechanics, 4, 2011, 780-784
[8]
Hu Wang, Long Tang, G. Y. Li, Adaptive MLS-HDMR metamodeling techniques for high dimensional problems [J], Expert Systems with Applications, 38(11), 2011, 14117-14126
[9]
Chonggang Xu, Gertner, George Zdzislaw, Uncertainty and sensitivity analysis for models with
correlated parameters [J], Reliability Engineering and System Safety, 93(10), 2008, 1563-1573
[10] Y. Shi, R. C. Eberhart, Empirical study of particle swarm optimization [C], Proceedings of the
1999 IEEE Congress on Evolutionary Computation, CEC 99, 1999
[11] Fei Gao, Hengqing Tong, Parameter estimation for system based on particle swarm optimization
[J], Acta Phys., 55(2), 2006, 577-582
[12] Xiaobo Xu, Kangfeng Zheng, Dan Li et al., New chaos-particle swarm optimization algorithm [J],
Journal on Communication, 33(1), 2012, 24-29
[13] Kaitai Fang, Changxing Ma, Orthogonal Design and Uniform Design [M], Beijing: Science Press,
2001
[14] W. Clifford Hansen, C. Jon Helton, J. C´edric Sallaberry, Use of replicated Latin hypercube sampling to estimate sampling variance in uncertainty and sensitivity analysis results for the geologic
disposal of radioactive waste [J], Procedia-Social and Behavioral Sciences, 2(6), 2010, 7674-7675
[15] A. Der Kiureghian, P. L. Liu, Structural reliability under incomplete probability information [J],
Journal of Engineering Mechanics, 112(1), 1986, 85-104