📄 sv_user.html
字号:
depend on the type of SV machine you have chosen, if you choseregression estimation though, the following menu would appear :<P><PRE> General parameters ================== 1. Bound on Alphas (C) 0 2. Scaling off 3. Chunking off 4. Epsilon accuracy 0 6. Number of Machines 1 0. Exit Please enter your choice:</PRE><P>This is identical to the options for the Pattern Recognition case,except that Pattern Recognition does not have epsilon as a parameter,so that option is not included. For the moment, leave chunking off(you only need chunking for large data sets i.e. those > 700examples), and leave the number of machines set to 1. This leavesalpha bounds, epsilon accuracy, and scaling.<P>The bound on the alphas is simply the value of C which you wish touse. A value of 0 will set C to infinity, which corresponds torequiring linear separability in the pattern recognition case.Another way of thinking of it is that the Value of C represents howmuch you punish errors (i.e. examples that lie on the wrong side ofthe hyperplane). The larger value of C, the less errors you areallowing the decision rule to have.<P>Epsilon accuracy is simply a floating point number, which controls thesize of the epsilon insensitive zone.<P>Scaling is used when the data in the input file is not properlyscaled. Scaling can be done either globally (i.e. all values arescaled the same) or locally (each individual attribute is scaledindependently). As a guideline you may wish to think of it this way;if the attributes are all of the same type (e.g. pixel values) thenscale globally, if they are of different types (e,g, age, height,weight) then scale locally. When you select the scaling option, theprogram first asks if you want to scale the data, then it asks if allattributes are to be scaled with the same factor. Answering Ycorresponds to global scaling and N corresponds to local scaling. Youare then asked to specifiy the lower and upper bounds for the scaleddata e.g. -1 and 1, or 0 and 1.<P><H2><A NAME="SECTION00024000000000000000">Setting Kernel Specific Paramters</A></H2><P>The final option on the SVM paramter menu is used to set the kernelspecific paramters. This is obviously greatly dependent on the kernelyou have chosen to use. Some kernels such as Inifinite DimensionalSplines do not have any paramters, most kernels do though. As anexample, if you are using the 2 layer Neural Network kernel (seesection 2.2), then selecting option 4 would result in the SVM programprompting you to enter the values for the kernel's two paramters - band c.<P>Once again, there is no rule for which values of paramters are goodfor different kernels on different data sets. You simply have to trydifferent values and try to develop an understanding of how the kernelworks.<P><H2><A NAME="SECTION00025000000000000000">Setting the Expert Parameters</A></H2><P>Don't. There you have it in one word, unless you know exactly whatyou want to achieve then its not a good idea to mess with thesesettings. There is (like everything else) one exception though. Youcan choose which optimiser to use. As a suggestion, you may want touse LOQO for pattern recognition, and MINOS for regressionestimation. Do not use the third option BOTTOU as that optimiser hasnot been fully implemented yet.<P><H2><A NAME="SECTION00026000000000000000">Life After Parameter Setting</A></H2><P>After you have chosen the parameters, you are now ready to try them.If you reached the parameter menu by running the SVM program withoutspecifying a parameter file, then when you select exit the SV machinewill run on the data with the paramters you have just specified (andthese parameters will be saved in tmp.params).If you are using paragen, then you will be given the option to savethe parameters (at which point you have to specify a filename),display the parameters, or change them if you are not happy with yourselection (You will also have an option ``special commands'' whichallows you to add parameter sets, but its best to ignore this for thetime being). You can now run the SV machine with your parameter filespecified on the command line, and see what results you get.<P><H1><A NAME="SECTION00030000000000000000">Output from the SVM Program</A></H1><P>Generating parameter files and running the SVM program on various setsof data may be fun, but it isn't of much practical use unless you canunderstand some of the output which the support vector machine gives.<P>So here's some example output (for pattern recognition), and we'll gothrough it step by step :<P><PRE>rm: fort.9: No such file or directoryFri Jan 2 11:27:28 WET 1998 SV Machine parameters ===================== Pattern Recognition Full polynomial Alphas unbounded Input values will be globally scaled between 0 and 1. Training data will not be posh chunked. Training data will not be sporty chunked. Number of SVM: 1 Degree of polynomial: 2. Kernel Scale Factor: 256. Kernel Threshold: 1.--------------------------------------------------------------------------------Positive SVs: 12 13 16 41 111 114 157 161 Negative SVs: 8 36 126 138 155 165 There are 14 SVs (8 positive and 6 negative).Max alpha size: 3.78932B0 is -1.87909Objective = 9.71091Fri Jan 2 11:27:33 WET 1998Training set: Total samples: 200Positive samples: 100of which errors: 0Negative samples: 100of which errors: 0----------------------------------------Test set: Total samples: 50Positive samples: 25of which errors: 0Negative samples: 25of which errors: 0There are 1 lagrangian multipliers per support vector.No. alpha(0) Deviation 8 1.86799 3.77646e-08 12 0.057789 6.97745e-08 13 2.75386 16 0.889041 -1.63897e-08 36 1.53568 5.93671e-08 41 0.730079 3.91323e-09 111 0.359107 -1.38041e-07 114 0.88427 -9.06655e-08 126 0.561356 1.15792e-07 138 2.558 2.0497e-08 155 2.50095 -1.24475e-08 157 0.247452 -1.35147e-07 161 3.78932 -1.2767e-07 165 0.686947 1.03066e-07Finished checking support vector accuracy.Total deviation is 9.30535e-07 No. of SVs: 14Average deviation is 6.64668e-08Minimum alpha is 0.057789Maximum alpha is 3.78932</PRE><P>The SVM program uses a different type of optimiser to construct therule, depending on which one you selected when setting the parameters.When using LOQO the optimiser gives no output, but MINOSS, the otheroptimiser, gives an output of the following form :<P><PRE> ============================== M I N O S 5.4 (Dec 1992) ============================== Begin SV_TEST hi!OPTIMAL SOLUTION FOUND (0)----------------------------------------</PRE><P>In this case, the optimiser signals that an optimal solution wasfound. If the data is scaled badly, or the data is inseperable (andthe bound on the alphas is infinite), then an error may occur here.Therefore, you will have to ensure the scaling options are setcorrectly, and you may have to change the bound on the alpha values(the value of C).<P>The next section informs the user how many support vectors there are,and lists the example numbers of those examples which were supportvectors. This section also indicates the largest alpha value, and thevalue of b0.<P>Next is important information as to how the SVMperformed on both the training set and the test set. In the case ofpattern recognition (as shown above), the ouput indicates the numberof positive and negative samples, and the number of those which weremisclassified in both the training and the test set. For instance, inthe example above, all of the examples in the training set were classified correctly. Obviously, in regression estimation, you arenot interested in classification. When running the SVM program toperform regression estimation, various measures of error are displayedhere. The user is given the average (absolute) error on the trainingset. Also, the totals and averages are displayed for both absoluteand squared error on the training set.<P>Following the performance statistics, a list of the values of thealphas (lagrange multipliers) for each support vector is given, alongwith its deviation (how far away the support vector is from theboundary of the margin). Finally some statistics are given, indicatingthe minimum and maximum alpha values.<P><H1><A NAME="SECTION00040000000000000000">Further Information</A></H1><P>There is an on-line version of the support vector machine which hasbeen developed in the department. The web site has a graphicalinterface which allows you to plot a few points and see what decisionboundary is produced. Be warned that when the site has many users, itcan be a little slow. It is sometimes useful to see what the SVMprogram actually does though (at least in two dimensions), and thepage also provides links to other SVM sites. The web address of the page is :<P>http://svm.cs.rhbnc.ac.uk<P><H1><A NAME="SECTION00050000000000000000"> About this document ... </A></H1><P> <STRONG>Support Vector Machine - User Manual</STRONG><P>This document was generated using the <A HREF="http://www-dsed.llnl.gov/files/programs/unix/latex2html/manual/"><STRONG>LaTeX</STRONG>2<tt>HTML</tt></A> translator Version 96.1-h (September 30, 1996) Copyright © 1993, 1994, 1995, 1996, <A HREF="http://cbl.leeds.ac.uk/nikos/personal.html">Nikos Drakos</A>, Computer Based Learning Unit, University of Leeds. <P> The command line arguments were: <BR><STRONG>latex2html</STRONG> <tt>-split 0 -html_version 3.0 sv_user</tt>. <P>The translation was initiated by Mark O. Stitson on Wed Jun 3 14:45:37 BST 1998<HR><IMG WIDTH=37 HEIGHT=24 ALIGN=BOTTOM ALT="next" SRC="file:/usr/lib/latex2html/icons/next_motif.gif"> <IMG WIDTH=26 HEIGHT=24 ALIGN=BOTTOM ALT="up" SRC="file:/usr/lib/latex2html/icons/up_motif_gr.gif"> <IMG WIDTH=63 HEIGHT=24 ALIGN=BOTTOM ALT="previous" SRC="file:/usr/lib/latex2html/icons/previous_motif.gif"> <BR><P><ADDRESS><I>Mark O. Stitson <BR>Wed Jun 3 14:45:37 BST 1998</I></ADDRESS></BODY></HTML>
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -