A regression metric that penalizes large prediction errors by squaring residuals before averaging.
Root Mean Squared Error (RMSE) is a standard evaluation metric for regression models that quantifies the average magnitude of prediction errors. It is computed by taking each difference between a predicted value and its corresponding observed value, squaring those differences, averaging them across all samples, and then taking the square root. This final square root step returns the error to the same units as the original target variable, making RMSE directly interpretable — an RMSE of 5 degrees Celsius, for instance, means predictions deviate from true values by roughly 5 degrees on average.
The squaring step is what distinguishes RMSE from simpler alternatives like Mean Absolute Error (MAE). Because squaring amplifies larger residuals disproportionately, RMSE is more sensitive to outliers and big mistakes than MAE. This property is often desirable: in many real-world applications, a single large error is far more costly than several small ones, and RMSE naturally reflects that asymmetry. Conversely, when outliers are noise rather than meaningful signal, MAE may be preferred precisely because RMSE would overweight them.
In machine learning practice, RMSE is ubiquitous as both a training loss function and a held-out evaluation metric. It appears across domains including time-series forecasting, demand prediction, recommendation systems, and computer vision tasks like depth estimation. During model selection and hyperparameter tuning, practitioners compare RMSE values across candidate models on a validation set to identify the best-performing configuration. It is also the basis of the Mean Squared Error (MSE) loss, which is mathematically equivalent but lacks the square root — MSE is often preferred during gradient-based optimization because it is differentiable and computationally simpler, while RMSE is reported for human interpretability.
One important limitation of RMSE is that it is scale-dependent, making it unsuitable for comparing models trained on targets with different units or magnitudes. Normalized variants such as NRMSE (normalized by the range or standard deviation of the target) address this. Despite its limitations, RMSE remains one of the most widely reported metrics in regression benchmarks, offering a concise, interpretable summary of a model's predictive accuracy.