A model whose behavior is governed by learnable numerical values called parameters.
A parameterized model is any function or system whose outputs are determined by a set of adjustable numerical values — the parameters. In machine learning, these parameters are not hand-coded by engineers but instead learned from data. The model begins with some initial parameter configuration, makes predictions, and then updates those values based on how wrong its predictions were. This feedback loop, driven by optimization algorithms like gradient descent, gradually shapes the parameters into a configuration that captures meaningful patterns in the training data.
The mechanics of parameterization vary by model type, but the principle is universal. In a linear regression model, the parameters are slope and intercept values. In a neural network, they are the weights and biases associated with each connection and neuron across potentially billions of nodes. In large language models, parameter counts now reach into the hundreds of billions, with each value encoding some fragment of learned knowledge about language, reasoning, or world structure. The sheer scale of modern parameterized models is one reason they can generalize across such a wide range of tasks.
Parameterization matters because it defines the boundary between what a model can and cannot represent. A model with too few parameters may be too rigid to capture complex patterns — a problem called underfitting. One with too many parameters relative to the available data may memorize noise rather than learn generalizable structure — a problem called overfitting. Choosing the right parameterization, including the number of parameters and how they are structured, is one of the central design decisions in building effective machine learning systems.
The concept also underpins transfer learning and fine-tuning, where a model pre-trained on one large dataset has its parameters further adjusted for a specific downstream task. Rather than learning from scratch, the model inherits a strong parameter initialization, dramatically reducing the data and compute needed for specialization. This reusability of learned parameters is a key reason why large pre-trained models have become the dominant paradigm in modern AI development.