Neural Network

A multi-layer perceptron (MLP) algorithm with backpropagation.


  • Data: input dataset

  • Preprocessor: preprocessing method(s)


  • Learner: multi-layer perceptron learning algorithm

  • Model: trained model

The Neural Network widget uses sklearn’s Multi-layer Perceptron algorithm that can learn non-linear models as well as linear.


  1. A name under which it will appear in other widgets. The default name is “Neural Network”.

  2. Set model parameters:

    • Neurons per hidden layer: defined as the ith element represents the number of neurons in the ith hidden layer. E.g. a neural network with 3 layers can be defined as 2, 3, 2.

    • Activation function for the hidden layer:

      • Identity: no-op activation, useful to implement linear bottleneck

      • Logistic: the logistic sigmoid function

      • tanh: the hyperbolic tan function

      • ReLu: the rectified linear unit function

    • Solver for weight optimization:

      • L-BFGS-B: an optimizer in the family of quasi-Newton methods

      • SGD: stochastic gradient descent

      • Adam: stochastic gradient-based optimizer

    • Alpha: L2 penalty (regularization term) parameter

    • Max iterations: maximum number of iterations

    Other parameters are set to sklearn’s defaults.

  3. Produce a report.

  4. When the box is ticked (Apply Automatically), the widget will communicate changes automatically. Alternatively, click Apply.


The first example is a classification task on iris dataset. We compare the results of Neural Network with the Logistic Regression.


The second example is a prediction task, still using the iris data. This workflow shows how to use the Learner output. We input the Neural Network prediction model into Predictions and observe the predicted values.