A linear model that makes predictions using only a small subset of input features.
A Sparse Linear Model (SLM) is a predictive framework that constrains most of its learned coefficients to zero, effectively selecting a compact subset of input features to explain the target variable. Rather than weighting every available feature, the model is trained with a sparsity-inducing penalty—most commonly an L1 regularization term—that drives irrelevant or redundant coefficients toward zero during optimization. The result is a model that depends on far fewer features than are present in the original data, making it especially well-suited for high-dimensional settings where the number of predictors can vastly exceed the number of observations.
The mechanics of sparsity are typically enforced through regularization techniques such as Lasso (Least Absolute Shrinkage and Selection Operator), Elastic Net, or basis pursuit. During training, the optimization objective balances fitting the data well against a penalty proportional to the sum of absolute coefficient values. This trade-off encourages the model to zero out features that contribute little predictive signal, performing simultaneous variable selection and parameter estimation in a single step. Variants like group Lasso extend this idea to structured sparsity, zeroing out entire groups of related features at once.
Sparse linear models matter for several practical reasons. First, they are highly interpretable: a model that uses ten features out of ten thousand is far easier to audit, explain, and trust than a dense alternative. Second, they generalize better in data-scarce regimes by reducing overfitting through implicit dimensionality reduction. Third, they are computationally efficient at inference time, since predictions require only a handful of multiplications. These properties have made SLMs a foundational tool in genomics, finance, natural language processing, and any domain where practitioners need both predictive accuracy and transparent feature attribution.