代码搜索:classifier
找到约 4,824 项符合「classifier」的源代码
代码结果 4,824
www.eeworm.com/read/351797/10609647
m train.m
function net = train(net, tutor, varargin)
% TRAIN
%
% Train a support vector classifier network using the specified tutor.
%
% load data/iris x y;
%
% C = 100;
% kernel = r
www.eeworm.com/read/421949/10676546
m bayescln.m
function [I,Pkx]=bayescln(X,MI,SIGMA,Pk)
% BAYESCLN Bayes classifier for Gaussian distributiuon.
% [I,Pkx]=bayescln(X,MI,SIGMA,Pk)
%
% This function classifies into the class according to the
%
www.eeworm.com/read/421949/10676564
m~ bayescln.m~
function [I,Pkx]=bayescln(X,MI,SIGMA,Pk)
% BAYESCLN Bayes classifier for Gaussian distributiuon.
% [I,Pkx]=bayescln(X,MI,SIGMA,Pk)
%
% This function classifies into the class according to the
%
www.eeworm.com/read/418755/10928180
m demo.m
%
% DEMONSTRATION OF ADABOOST_tr and ADABOOST_te
%
% Just type "demo" to run the demo.
%
% Using adaboost with linear threshold classifier
% for a two class classification problem.
%
% Bug Reporting:
www.eeworm.com/read/418695/10935744
m testd.m
%TESTD Classification error estimate
%
% [e,j,k,l] = testd(A,W,r,iter)
%
% Test of dataset A on the classifier defined by W. Returns:
% e - the fraction of A that is incorrectly classified by W.
%
www.eeworm.com/read/467949/6997137
m demo.m
%
% DEMONSTRATION OF ADABOOST_tr and ADABOOST_te
%
% Just type "demo" to run the demo.
%
% Using adaboost with linear threshold classifier
% for a two class classification problem.
%
% Bug Reporting:
www.eeworm.com/read/466142/7039596
m do_naive_bayes.m
function do_naive_bayes(config_file)
%% Function that runs the Naive Bayes classifier on histograms of
%% vector-quantized image regions. Based on the paper:
%%
%% Visual categorization with ba
www.eeworm.com/read/299984/7140004
m rnnc.m
%RNNC Random Neural Net classifier
%
% W = RNNC(A,N,S)
%
% INPUT
% A Input dataset
% N Number of neurons in the hidden layer
% S Standard deviation of weights in an input layer (default: 1
www.eeworm.com/read/299984/7140350
m mogc.m
%MOGC Mixture of Gaussian classifier
%
% W = MOGC(A,N)
% W = A*MOGC([],N);
%
% INPUT
% A Dataset
% N Number of mixtures (optional; default 2)
% R,S Regularization parameters, 0
www.eeworm.com/read/299984/7140542
m lssvc.m
function W = lssvc(A, TYPE, PAR, C)
%LSSVC Least-Squares Support Vector Classifier
%
% W = lssvc(A,TYPE,PAR,C);
%
% INPUT
% A dataset
% TYPE Type of the kernel (optional; default: '