📄 199-201.html
字号:
<HTML>
<HEAD>
<META name=vsisbn content="0849398010">
<META name=vstitle content="Industrial Applications of Genetic Algorithms">
<META name=vsauthor content="Charles Karr; L. Michael Freeman">
<META name=vsimprint content="CRC Press">
<META name=vspublisher content="CRC Press LLC">
<META name=vspubdate content="12/01/98">
<META name=vscategory content="Web and Software Development: Artificial Intelligence: Other">
<TITLE>Industrial Applications of Genetic Algorithms:Space Shuttle Main Engine Condition Monitoring Using Genetic Algorithms and Radial Basis Function Neural Network</TITLE>
<!-- HEADER -->
<STYLE type="text/css">
<!--
A:hover {
color : Red;
}
-->
</STYLE>
<META NAME="ROBOTS" CONTENT="NOINDEX, NOFOLLOW">
<!--ISBN=0849398010//-->
<!--TITLE=Industrial Applications of Genetic Algorithms//-->
<!--AUTHOR=Charles Karr//-->
<!--AUTHOR=L. Michael Freeman//-->
<!--PUBLISHER=CRC Press LLC//-->
<!--IMPRINT=CRC Press//-->
<!--CHAPTER=10//-->
<!--PAGES=199-201//-->
<!--UNASSIGNED1//-->
<!--UNASSIGNED2//-->
<CENTER>
<TABLE BORDER>
<TR>
<TD><A HREF="197-198.html">Previous</A></TD>
<TD><A HREF="../ewtoc.html">Table of Contents</A></TD>
<TD><A HREF="202-204.html">Next</A></TD>
</TR>
</TABLE>
</CENTER>
<P><BR></P>
<P><FONT SIZE="+1"><B>BACKGROUND</B></FONT></P>
<P>As discussed in the introduction, the objective of most neural networks is to estimate a function <IMG SRC="images/10-01i.jpg"> from a training set of representative input/output pairings. Qualitatively, the RBF network does this by forming <I>localized</I> “bumps” or response regions within the input space. The superposition of the these local response regions forms a response surface that is of an order higher than the dimension of the input vector and spans the space covered by the input training patterns.</P>
<P>By definition, a radial basis function is one which decreases (or increases) monotonically away from a central point, thereby giving it an inherent bump form. Classic kernel functions (or in the case of the RBFNN, “neurons”) that exhibit this propensity are the Gaussian, Cauchy, and the Inverse Multiquadric. These forms can be written generally as [5,6]:</P>
<DL>
<DD><B>A.</B> Cauchy function,
<P ALIGN="CENTER"><IMG SRC="images/10-02d.jpg"></P>
<DD><B>B.</B> Inverse Multiquadric,
<P ALIGN="CENTER"><IMG SRC="images/10-03d.jpg"></P>
<DD><B>C.</B> Gaussian function,
<P ALIGN="CENTER"><IMG SRC="images/10-04d.jpg"></P>
</DL>
<P>The form of z determines the type of radial scaling, or equivalently, the extent of the region influenced by the RBF.
</P>
<P>Figure 10.1 delineates the Gaussian response function. To reiterate, the RBFNN positions a collection of these RBFs (in this case, Gaussians) through-out the space covered by the input training patterns. The parameter <IMG SRC="images/10-02i.jpg"> specifies the location of the RBF within the input space (<IMG SRC="images/10-03i.jpg"> has the same dimension as the input vector <IMG SRC="images/10-04i.jpg">) and the parameter <IMG SRC="images/10-05i.jpg"> determines the width of the local function. Thus, a given RBF will be centered at μ<SUB><SMALL>j</SMALL></SUB> within the input space and have a “receptive field” which is proportional to <IMG SRC="images/10-06i.jpg">. Moreover, it will give a maximum response for input vectors, <IMG SRC="images/10-07i.jpg">, which are nearest the RBF center, <IMG SRC="images/10-08i.jpg">.</P>
<P><A NAME="Fig1"></A><A HREF="javascript:displayWindow('images/10-01.jpg',300,218)"><IMG SRC="images/10-01t.jpg"></A>
<BR><A HREF="javascript:displayWindow('images/10-01.jpg',300,218)"><FONT COLOR="#000077"><B>Figure 10.1</B></FONT></A> Gaussian function.</P>
<P>By arranging an assortment of these receptive fields, response areas are created which sufficiently cover the input space; sufficient in the sense that the RBFNN can approximate the underlying function to within some pre-defined error criteria. More specifically, a complex decision hypersurface is constructed through the overlapping of the localized kernel regions. With a developed approximation surface, the RBFNN estimates an output, for an incoming input case, by first evaluating each of the kernel functions (in other words, determining where the input vector lies on the hypersurface) and then forming a weighted linear summation of their responses. The difficulty arises not from the logical evaluation of an input, but rather the establishment of the network parameters for the hypersurface construction, namely: center positions (μ<SUB><SMALL>j</SMALL></SUB>), kernel widths (σ<SUB><SMALL>i</SMALL></SUB>), and the weighting coefficients for the summation of the individual kernel responses.</P>
<P>The development of an RBFNN is done in a two-part learning scheme known as hybrid learning (Figure 10.2). The initial forward connections of the network contain the RBF centers μ<SUB><SMALL>j</SMALL></SUB>, obtained through unsupervised assimilation, followed by an output layer of weighting parameters, formed through supervised instruction. Training in the unsupervised mode is done without a pre-defined learning goal; input categorization and learning must be done using correlations within the input training data in contrast to feedback from a teacher or critic. For the RBFNN, the learning scheme essentially clusters the training inputs vectors and specifies where to position the RBF centers so that the desired response coverage is obtained. Thus, via unsupervised learning, the RBF center positions (the forward connections of Figure 10.2, μ<SUB><SMALL>j</SMALL></SUB>) are chosen a <I>priori</I> and remain fixed throughout the establishment of the weighting coefficients (<I>w</I><SUB><SMALL>i</SMALL></SUB>).</P>
<P>The rearward connections, comprising the output layer of the network in Figure 10.2, specify the weighting (or regression) coefficients which are trained in a supervised fashion. Supervised means that the learning is based on comparison of the network output with the known “correct” answers.</P>
<P>For an RBFNN with a single layer of kernel functions, given that the basis function centers (μ<SUB><SMALL>j</SMALL></SUB>) are fixed, the optimal weight array for the output connections which gives the best functional mapping can be found using the least squares normal equation developed in multiple linear regression theory. [5]</P>
<P>With all the parameters set, the fundamental mapping can then be written as:</P>
<P ALIGN="CENTER"><IMG SRC="images/10-05d.jpg"></P>
<P>Thus, for an input vector <IMG SRC="images/10-09i.jpg"> the solution f(<IMG SRC="images/10-10i.jpg">) is a weighted linear summation of each RBF’s response to <IMG SRC="images/10-11i.jpg">. Kernel functions that have centers within the region of <IMG SRC="images/10-12i.jpg"> will give the largest responses, whereas those farthest away will give negligible contributions to the series formed by Equation (10.4). Moreover, the kernel function responses (<I>g</I><SUB><SMALL>i</SMALL></SUB><I>(x)</I>) will be bounded between 0 and 1 with the assigned weights (<I>w</I><SUB><SMALL>i</SMALL></SUB>) specifying the neuron’s heights.</P>
<P>The following sections detail the procedures of RBF center selection and width estimation.</P>
<P><A NAME="Fig2"></A><A HREF="javascript:displayWindow('images/10-02.jpg',400,325)"><IMG SRC="images/10-02t.jpg"></A>
<BR><A HREF="javascript:displayWindow('images/10-02.jpg',400,325)"><FONT COLOR="#000077"><B>Figure 10.2</B></FONT></A> RBFNN architecture.<P><BR></P>
<CENTER>
<TABLE BORDER>
<TR>
<TD><A HREF="197-198.html">Previous</A></TD>
<TD><A HREF="../ewtoc.html">Table of Contents</A></TD>
<TD><A HREF="202-204.html">Next</A></TD>
</TR>
</TABLE>
</CENTER>
<hr width="90%" size="1" noshade>
<div align="center">
<font face="Verdana,sans-serif" size="1">Copyright © <a href="/reference/crc00001.html">CRC Press LLC</a></font>
</div>
<!-- all of the reference materials (books) have the footer and subfoot reveresed -->
<!-- reference_subfoot = footer -->
<!-- reference_footer = subfoot -->
</BODY>
</HTML>
<!-- END FOOTER -->
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -