代码搜索:Gradient

找到约 2,951 项符合「Gradient」的源代码

代码结果 2,951
www.eeworm.com/read/396844/2406671

m olgd.m

function [net, options, errlog, pointlog] = olgd(net, options, x, t) %OLGD On-line gradient descent optimization. % % Description % [NET, OPTIONS, ERRLOG, POINTLOG] = OLGD(NET, OPTIONS, X, T) uses on
www.eeworm.com/read/396844/2406690

m gpgrad.m

function g = gpgrad(net, x, t) %GPGRAD Evaluate error gradient for Gaussian Process. % % Description % G = GPGRAD(NET, X, T) takes a Gaussian Process data structure NET % together with a matrix X of
www.eeworm.com/read/396844/2406705

m glmgrad.m

function [g, gdata, gprior] = glmgrad(net, x, t) %GLMGRAD Evaluate gradient of error function for generalized linear model. % % Description % G = GLMGRAD(NET, X, T) takes a generalized linear model da
www.eeworm.com/read/396844/2407027

m maximize_params.m

function CPD = maximize_params(CPD, temp) % MAXIMIZE_PARAMS Find ML params of an MLP using Scaled Conjugated Gradient (SCG) % CPD = maximize_params(CPD, temperature) % temperature parameter is igno
www.eeworm.com/read/393163/2487838

m mixexp_graddesc.m

%%%%%%%%%% function [theta, eta] = mixture_of_experts(q, data, num_iter, theta, eta) % MIXTURE_OF_EXPERTS Fit a piecewise linear regression model using stochastic gradient descent. % [theta, eta] =
www.eeworm.com/read/393163/2488217

m maximize_params.m

function CPD = maximize_params(CPD, temp) % MAXIMIZE_PARAMS Find ML params of an MLP using Scaled Conjugated Gradient (SCG) % CPD = maximize_params(CPD, temperature) % temperature parameter is igno
www.eeworm.com/read/369958/2788110

m gradlfixed.m

function [grad] = gradlfixed(Sigma,indsup,Alpsup,C,Xapp,yapp,pow); %GRADLFIXED Computes the gradient of an upper bound on SVM loss wrt SIGMA^POW % GRAD = GRADLFIXED(SIGMA,INDSUP,ALPSUP,C,XAPP,YAPP,
www.eeworm.com/read/369958/2788112

m gradwfixed.m

function [grad] = gradwfixed(Sigma,indsup,Alpsup,C,Xapp,yapp,Sigmaold,pow); %GRADWFIXED Computes the gradient of an upper bound on SVM loss wrt SIGMA^POW % GRAD = GRADWFIXED(SIGMA,INDSUP,ALPSUP,C,X
www.eeworm.com/read/359369/2978470

m olgd.m

function [net, options, errlog, pointlog] = olgd(net, options, x, t) %OLGD On-line gradient descent optimization. % % Description % [NET, OPTIONS, ERRLOG, POINTLOG] = OLGD(NET, OPTIONS, X, T) uses
www.eeworm.com/read/359369/2978542

m rbfgrad.m

function [g, gdata, gprior] = rbfgrad(net, x, t) %RBFGRAD Evaluate gradient of error function for RBF network. % % Description % G = RBFGRAD(NET, X, T) takes a network data structure NET together