NGBoost
This topic explains how the NGBoost model works
NGBoost
The NGBoost model provides Stanford ML Group's NGBoost. NGBoost enables predictive uncertainty estimation with Gradient Boosting through probabilistic predictions (including real valued outputs). With the use of Natural Gradients, NGBoost overcomes technical challenges that make generic probabilistic prediction hard with gradient boosting.
Probability Distributions
Myst's implementation of NGBoost contains four probability distributions - Normal, Cauchy, Laplace, and Lognormal. Each model outputs scale and loc parameters that can be used to determine the upper and lower bounds of the associated probability distribution. The loc parameter in the lognormal is returned in log space.
Backtest Results
In backtests MAPE, MSE, and MAE are reported for symmetric distributions (Normal, Cauchy, Laplace). For asymmetric distributions (Lognormal), only mean-negative-log-likelihood.
Parameters
Parameter | Description | Default Value |
---|---|---|
distribution | The assumed distributional form of the endogenous data, given the exogenous data. | Normal |
n_estimators | The number of boosting iterations to fit. | 500 |
learning_rate | The learning rate | 0.01 |
max_depth | The maximum depth of the base learner. | 3 |
minibatch_frac | The percent subsample of rows to use in each boosting iteration. | 1 |
col_sample | The percent subsample of columns to use in each boosting iteration. | 1 |
Tutorial
See our Create a Price Forecast with Probabilistic Prediction tutorial for examples using NGBoost!
Updated over 2 years ago