SmoothedHinge Class
Some of the trainers accept a loss parameter that will be used for training. It is also known as loss function, objective function, or optimization score function.
- Inheritance
-
builtins.objectSmoothedHinge
Constructor
SmoothedHinge(smoothing_const=1.0)
Parameters
Name | Description |
---|---|
smoothing_const
|
Smoothing constant |
Examples
###############################################################################
# Smoothed Hinge Loss
from nimbusml.linear_model import FastLinearBinaryClassifier
# can also use loss class instead of string
from nimbusml.loss import SmoothedHinge
# specifying the loss function as a string keyword
trainer1 = FastLinearBinaryClassifier(loss='smoothed_hinge')
# equivalent to loss='smoothed_hinge'
trainer2 = FastLinearBinaryClassifier(loss=SmoothedHinge())
trainer3 = FastLinearBinaryClassifier(loss=SmoothedHinge(smoothing_const=0.5))
Remarks
Losses can be specified either as a string or a loss object. When loss is specified as one of these strings, the default values are used for the loss parameters. To change the default parameters, a loss object should be used, as seen in examples below.
Each trainer supports only a subset of the losses mentioned above. To get the supported losses and the default loss, please refer to the documentation page for the specific trainer.
The Smoothed hinge loss
for classification. Its string name is 'smoothed_hinge'
.
It can be used for AveragedPerceptronBinaryClassifier, FastLinearBinaryClassifier, FastLinearClassifier, SgdBinaryClassifier.