A fully connected feedforward neural network trained via backpropagation for classification and regression.
A multilayer perceptron (MLP) is a class of feedforward artificial neural network consisting of at least three layers of nodes: an input layer, one or more hidden layers, and an output layer. Every neuron in each layer is connected to every neuron in the adjacent layers — a topology called full connectivity — and each non-input neuron applies a nonlinear activation function (such as sigmoid, tanh, or ReLU) to a weighted sum of its inputs. This nonlinearity is what allows MLPs to model complex, non-linear relationships that simpler linear models cannot capture.
Training an MLP relies on the backpropagation algorithm combined with gradient descent. During a forward pass, inputs propagate through the network to produce a prediction; the error between that prediction and the true label is then computed via a loss function. Backpropagation efficiently calculates the gradient of this loss with respect to every weight in the network by applying the chain rule layer by layer, and those gradients are used to update weights in the direction that reduces error. Repeated over many training examples, this process allows the network to learn rich internal representations of the data.
MLPs became a cornerstone of practical machine learning after Rumelhart, Hinton, and Williams demonstrated the effectiveness of backpropagation in 1986, enabling networks with hidden layers to be trained reliably for the first time at scale. They remain widely used today as baseline models for tabular data, and as components within larger architectures. While deep convolutional and transformer-based models have surpassed MLPs on tasks like image recognition and natural language processing, the MLP's simplicity, interpretability, and versatility keep it central to both research and production machine learning pipelines.