📄 202-204.html
字号:
<HTML>
<HEAD>
<META name=vsisbn content="0849398010">
<META name=vstitle content="Industrial Applications of Genetic Algorithms">
<META name=vsauthor content="Charles Karr; L. Michael Freeman">
<META name=vsimprint content="CRC Press">
<META name=vspublisher content="CRC Press LLC">
<META name=vspubdate content="12/01/98">
<META name=vscategory content="Web and Software Development: Artificial Intelligence: Other">
<TITLE>Industrial Applications of Genetic Algorithms:Space Shuttle Main Engine Condition Monitoring Using Genetic Algorithms and Radial Basis Function Neural Network</TITLE>
<!-- HEADER -->
<STYLE type="text/css">
<!--
A:hover {
color : Red;
}
-->
</STYLE>
<META NAME="ROBOTS" CONTENT="NOINDEX, NOFOLLOW">
<!--ISBN=0849398010//-->
<!--TITLE=Industrial Applications of Genetic Algorithms//-->
<!--AUTHOR=Charles Karr//-->
<!--AUTHOR=L. Michael Freeman//-->
<!--PUBLISHER=CRC Press LLC//-->
<!--IMPRINT=CRC Press//-->
<!--CHAPTER=10//-->
<!--PAGES=202-204//-->
<!--UNASSIGNED1//-->
<!--UNASSIGNED2//-->
<CENTER>
<TABLE BORDER>
<TR>
<TD><A HREF="199-201.html">Previous</A></TD>
<TD><A HREF="../ewtoc.html">Table of Contents</A></TD>
<TD><A HREF="204-206.html">Next</A></TD>
</TR>
</TABLE>
</CENTER>
<P><BR></P>
</P>
<P><FONT SIZE="+1"><B><I>Clustering and Center Selection</I></B></FONT></P>
<P>RBFNNs fix the center positions (μ<SUB><SMALL>j</SMALL></SUB>’s) and use simple multi-linear regression to obtain the optimal output weight array. To give this linear RBFNN more flexibility, a subset of RBF centers should be chosen which can best explain the variance in the dependent (or output) variables; in other words, centers are chosen so that the input space is adequately covered. Considering this, it seems natural to simply place a center over every input training point, μ<SUB><SMALL>j</SMALL></SUB> = <B>x</B><SUB><SMALL>j</SMALL></SUB> for j=1 to j=p (p designates the number of training cases). This choice would result in a network that “memorizes” all the training pairs and on occasion an added impropriety of exhibiting extremely poor generalization ability. The large number of neurons provides too many free parameters for the linear regression phase, thereby making the network oversensitive to the training set details. Furthermore, if the quantity of available training sets is large, then additional computational speed and memory problems will result, further making the RBF infeasible.</P>
<P>To counter these effects, the smallest subset of kernel function centers is selected that can perform the necessary mapping with a sufficient amount of accuracy. There have been many heuristics developed that can accomplish this subset selection process. Some start by adding one neuron at a time until a pre-specified error goal is met. Others start with centers at all points in the initial input data and then make predictions on what subset will prove to be most efficacious.</P>
<P><FONT SIZE="+1"><B><I>RBF Width Estimation</I></B></FONT></P>
<P>The width parameter (σ in Figure 10.1) doesn’t just control the shape of the RBF, it also specifies where and how much it overlaps with other RBFs spaced throughout the hypersurface. Physically, it determines the <I>shape of the response surface</I> (or the neuron’s receptive field). Thus, if an extremely nonlinear problem is encountered, a greater amount of flexibility built into the surface construction will lead to a better performing network.</P>
<P>Recall the RBFs detailed in Equations 10.1-10.3. The parameter <I>z</I> introduces this “flexibility” by determining the type radial spread. Assuming that the network input vectors are n-dimensional and there are in neurons in the network, then z can take on the following forms (taken from [6]).</P>
<DL>
<DD><B>A.</B> Each neuron has the same spherical size if r is constant,
<P ALIGN="CENTER"><IMG SRC="images/10-06d.jpg"></P>
<DD><B>B.</B> Also, each neuron can have a different spherical size if r = [r<SUB><SMALL>1</SMALL></SUB> r<SUB><SMALL>2</SMALL></SUB> … r<SUB><SMALL>m</SMALL></SUB>]<SUP><SMALL>T</SMALL></SUP>, <P ALIGN="CENTER"><IMG SRC="images/10-07d.jpg"></P>
<DD><B>C.</B> r can also be a vector in n-dimensional space giving each center the same ellipsoidal size with axes coincident with the basis vectors,
<P ALIGN="CENTER"><IMG SRC="images/10-08d.jpg"></P>
<DD><B>D.</B> If r is an n x n matrix <B>R</B>, then each center will have the same ellipsoidal size but with arbitrary axes,
<P ALIGN="CENTER"><IMG SRC="images/10-09d.jpg"></P>
<DD><B>E.</B> Finally, r can be a matrix of size n x m as r = [<B>r</B><SUB><SMALL>1</SMALL></SUB> <B>r</B><SUB><SMALL>2</SMALL></SUB> … <B>r</B><SUB><SMALL>m</SMALL></SUB>] giving each center a different ellipsoidal size, with axes coincident with the basis vectors,
<P ALIGN="CENTER"><IMG SRC="images/10-10d.jpg"></P>
</DL>
<P>The choice of r for each neuron is usually left to random decision. There are some methods which allow for local “tuning” of the parameter, like those found in [4], but none account for the simultaneous evolution of the hypersurface. The next section provides an overview of what has been done in the past for these parameter estimations, with the following section outlining the genetic algorithm approach specific to the problem stated at the outset of the chapter.
</P>
<P><FONT SIZE="+1"><B>PREVIOUS RBF PARAMETER SELECTION APPROACHES</B></FONT></P>
<P><FONT SIZE="+1"><B><I>Non-genetic Approaches</I></B></FONT></P>
<P>Non-genetic approaches to the problem of center (μ<SUB><SMALL>j</SMALL></SUB>) selection typically separate a subset of centers from the training input patterns (x). A very crude method is to evolve one hidden unit at a time during the training processes, with the center being randomly selected from the input training vectors. The addition of hidden units would continue until some training error goal has been met. A more elegant approach uses a k-means clustering algorithm that groups together input patterns belonging to the same class region [4,7]. The vector mean positions of these regions are then used as the RBF center subset.</P>
<P>References [5,6] make use of statistical error prediction methods. For example, Bayesian Information Criterion and Generalized Cross Validation can be used to select input patterns from the training set that have the most probability of reducing the mapping error during the supervised training phase. This is done by looping through the training set and assigning scores to each input pattern, with the best score being selected. The selection process continues until the error prediction heuristic reaches a minimum. In other words, the addition of further RBF centers will not assist in the mapping error reduction.</P>
<P>Cascade-correlation can also be used with an RBF architecture for the evolution of one hidden unit at a time [8]. The process is actually a gradient descent procedure that places the RBF center in a position that reduces the current training error the most.</P>
<P>Non-genetic RBF width estimation procedures are limited. Most rely on a random search, sort of “brute force,” method for width establishment. Reference [4] does provide local tuning procedures, but it doesn’t consider the evolution of a whole network.</P><P><BR></P>
<CENTER>
<TABLE BORDER>
<TR>
<TD><A HREF="199-201.html">Previous</A></TD>
<TD><A HREF="../ewtoc.html">Table of Contents</A></TD>
<TD><A HREF="204-206.html">Next</A></TD>
</TR>
</TABLE>
</CENTER>
<hr width="90%" size="1" noshade>
<div align="center">
<font face="Verdana,sans-serif" size="1">Copyright © <a href="/reference/crc00001.html">CRC Press LLC</a></font>
</div>
<!-- all of the reference materials (books) have the footer and subfoot reveresed -->
<!-- reference_subfoot = footer -->
<!-- reference_footer = subfoot -->
</BODY>
</HTML>
<!-- END FOOTER -->
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -