代码搜索:Gradient
找到约 2,951 项符合「Gradient」的源代码
代码结果 2,951
www.eeworm.com/read/396844/2406671
m olgd.m
function [net, options, errlog, pointlog] = olgd(net, options, x, t)
%OLGD On-line gradient descent optimization.
%
% Description
% [NET, OPTIONS, ERRLOG, POINTLOG] = OLGD(NET, OPTIONS, X, T) uses on
www.eeworm.com/read/396844/2406690
m gpgrad.m
function g = gpgrad(net, x, t)
%GPGRAD Evaluate error gradient for Gaussian Process.
%
% Description
% G = GPGRAD(NET, X, T) takes a Gaussian Process data structure NET
% together with a matrix X of
www.eeworm.com/read/396844/2406705
m glmgrad.m
function [g, gdata, gprior] = glmgrad(net, x, t)
%GLMGRAD Evaluate gradient of error function for generalized linear model.
%
% Description
% G = GLMGRAD(NET, X, T) takes a generalized linear model da
www.eeworm.com/read/396844/2407027
m maximize_params.m
function CPD = maximize_params(CPD, temp)
% MAXIMIZE_PARAMS Find ML params of an MLP using Scaled Conjugated Gradient (SCG)
% CPD = maximize_params(CPD, temperature)
% temperature parameter is igno
www.eeworm.com/read/393163/2487838
m mixexp_graddesc.m
%%%%%%%%%%
function [theta, eta] = mixture_of_experts(q, data, num_iter, theta, eta)
% MIXTURE_OF_EXPERTS Fit a piecewise linear regression model using stochastic gradient descent.
% [theta, eta] =
www.eeworm.com/read/393163/2488217
m maximize_params.m
function CPD = maximize_params(CPD, temp)
% MAXIMIZE_PARAMS Find ML params of an MLP using Scaled Conjugated Gradient (SCG)
% CPD = maximize_params(CPD, temperature)
% temperature parameter is igno
www.eeworm.com/read/369958/2788110
m gradlfixed.m
function [grad] = gradlfixed(Sigma,indsup,Alpsup,C,Xapp,yapp,pow);
%GRADLFIXED Computes the gradient of an upper bound on SVM loss wrt SIGMA^POW
% GRAD = GRADLFIXED(SIGMA,INDSUP,ALPSUP,C,XAPP,YAPP,
www.eeworm.com/read/369958/2788112
m gradwfixed.m
function [grad] = gradwfixed(Sigma,indsup,Alpsup,C,Xapp,yapp,Sigmaold,pow);
%GRADWFIXED Computes the gradient of an upper bound on SVM loss wrt SIGMA^POW
% GRAD = GRADWFIXED(SIGMA,INDSUP,ALPSUP,C,X
www.eeworm.com/read/359369/2978470
m olgd.m
function [net, options, errlog, pointlog] = olgd(net, options, x, t)
%OLGD On-line gradient descent optimization.
%
% Description
% [NET, OPTIONS, ERRLOG, POINTLOG] = OLGD(NET, OPTIONS, X, T) uses
www.eeworm.com/read/359369/2978542
m rbfgrad.m
function [g, gdata, gprior] = rbfgrad(net, x, t)
%RBFGRAD Evaluate gradient of error function for RBF network.
%
% Description
% G = RBFGRAD(NET, X, T) takes a network data structure NET together