Minimize an objective function using a stochastic approximation of gradient descent.

Inputs

• Data: input dataset

• Preprocessor: preprocessing method(s)

Outputs

• Learner: stochastic gradient descent learning algorithm

• Model: trained model

The Stochastic Gradient Descent widget uses [stochastic gradient descent](https://en.wikipedia.org/wiki/Stochastic_gradient_descent) that minimizes a chosen loss function with a linear function. The algorithm approximates a true gradient by considering one sample at a time, and simultaneously updates the model based on the gradient of the loss function. For regression, it returns predictors as minimizers of the sum, i.e. M-estimators, and is especially useful for large-scale and sparse datasets.

1. Specify the name of the model. The default name is “SGD”.

2. Algorithm parameters:
3. Regularization norms to prevent overfitting: - None. - [Lasso (L1)](https://en.wikipedia.org/wiki/Taxicab_geometry) (L1 leading to sparse solutions) - [Ridge (L2)](https://en.wikipedia.org/wiki/Norm_(mathematics)#p-norm) (L2, standard regularizer) - [Elastic net](https://en.wikipedia.org/wiki/Elastic_net_regularization) (mixing both penalty norms).

Regularization strength defines how much regularization will be applied (the less we regularize, the more we allow the model to fit the data) and the mixing parameter what the ratio between L1 and L2 loss will be (if set to 0 then the loss is L2, if set to 1 then it is L1).

4. Learning parameters. - Learning rate:

• Initial learning rate.

• Inverse scaling exponent: learning rate decay.

• Number of iterations: the number of passes through the training data.

• If Shuffle data after each iteration is on, the order of data instances is mixed after each pass.

• If Fixed seed for random shuffling is on, the algorithm will use a fixed random seed and enable replicating the results.

5. Produce a report.

6. Press Apply to commit changes. Alternatively, tick the box on the left side of the Apply button and changes will be communicated automatically.

Examples¶

For the classification task, we will use iris dataset and test two models on it. We connected [Stochastic Gradient Descent](../model/stochasticgradient.md) and [Tree](../model/tree.md) to [Test & Score](../evaluate/testandscore.md). We also connected [File](../data/file.md) to Test & Score and observed model performance in the widget.