In machine learning, a loss function is a mathematical function that must be minimized in order to achieve convergence. Choosing the proper loss function is an important step in designing your neural network. The IDLmllfHuber (Huber) loss function is implemented with the following formula:
where x is the calculated output of the model and y is the predicted output or truth.
Example
Compile_opt idl2
LossFunction = IDLmllfHuber(0.5)
Print, LossFunction(Findgen(10)/9.0, Fltarr(10))
Typically, you will pass an object of this class to a neural network model definition:
Classifier = IDLmlFeedForwardNeuralNetwork([3, 7, 1],
LOSS_FUNCTION=IDLmllfHuber(0.25)
Syntax
Kernel = IDLmllfHuber(Delta)
Arguments
Delta
Specify a threshold value that determines how small the error has to be to make it quadratic.
Keywords
None
Version History
See Also
IDLmllfCrossEntropy, IDLmllfLogCosh, IDLmllfMeanAbsoluteError, IDLmllfMeanSquaredError