TY - JOUR
T1 - Natural gradient algorithm for stochastic distribution systems with output feedback
AU - Zhang, Zhenning
AU - Sun, Huafei
AU - Peng, Linyu
N1 - Funding Information:
This subject is supported by the National Natural Science Foundations of China , Nos. 10871218 , 10932002 , 61179031 .
PY - 2013/10
Y1 - 2013/10
N2 - In this paper, we use natural gradient algorithm to control the shape of the conditional output probability density function for the stochastic distribution systems from the viewpoint of information geometry. The considered system here is of multi-input and single output with an output feedback and a stochastic noise. Based on the assumption that the probability density function of the stochastic noise is known, we obtain the conditional output probability density function whose shape is only determined by the control input vector under the condition that the output feedback is known at any sample time. The set of all the conditional output probability density functions forms a statistical manifold (M), and the control input vector and the output feedback are considered as the coordinate system. The Kullback divergence acts as the distance between the conditional output probability density function and the target probability density function. Thus, an iterative formula for the control input vector is proposed in the sense of information geometry. Meanwhile, we consider the convergence of the presented algorithm. At last, an illustrative example is utilized to demonstrate the effectiveness of the algorithm.
AB - In this paper, we use natural gradient algorithm to control the shape of the conditional output probability density function for the stochastic distribution systems from the viewpoint of information geometry. The considered system here is of multi-input and single output with an output feedback and a stochastic noise. Based on the assumption that the probability density function of the stochastic noise is known, we obtain the conditional output probability density function whose shape is only determined by the control input vector under the condition that the output feedback is known at any sample time. The set of all the conditional output probability density functions forms a statistical manifold (M), and the control input vector and the output feedback are considered as the coordinate system. The Kullback divergence acts as the distance between the conditional output probability density function and the target probability density function. Thus, an iterative formula for the control input vector is proposed in the sense of information geometry. Meanwhile, we consider the convergence of the presented algorithm. At last, an illustrative example is utilized to demonstrate the effectiveness of the algorithm.
KW - Information geometry
KW - Kullback divergence
KW - Natural gradient algorithm
KW - Stochastic distribution system
UR - http://www.scopus.com/inward/record.url?scp=84881000247&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84881000247&partnerID=8YFLogxK
U2 - 10.1016/j.difgeo.2013.07.004
DO - 10.1016/j.difgeo.2013.07.004
M3 - Article
AN - SCOPUS:84881000247
SN - 0926-2245
VL - 31
SP - 682
EP - 690
JO - Differential Geometry and its Applications
JF - Differential Geometry and its Applications
IS - 5
ER -