Motivation and Setting
Previous sections focused on finite collections of random probability measures, typically corresponding to discrete groups that are either exchangeable or follow a simple ordering. In many real-world applications, however, predictors vary continuously, and it is more natural to consider an uncountable collection of random probability measures indexed by predictors:
where is a vector of predictors, is the conditional distribution of the response given , and denotes the predictor space.
The goal of density regression is to model the entire conditional density flexibly, allowing its shape, skewness, and tail behavior to vary smoothly with predictors, rather than modeling only conditional means or quantiles.
Mixture-of-Experts Models
A classical approach to density regression is the hierarchical mixture-of-experts (MoE) model:
where:
- each mixture component corresponds to a linear regression,
- the regression parameters and precisions are component-specific,
- the mixture weights vary with predictors.
In machine learning, the predictor-dependent weights are often modeled using tree-based structures or logistic regressions. While flexible, these approaches require specifying a finite number of components and may struggle to adapt to complex conditional density shapes.
Nonparametric Bayesian Density Regression
A fully nonparametric Bayesian alternative models the conditional density as a predictor-dependent mixture:
where is a collection of random mixing measures. The prior on this collection, denoted , must induce dependence across predictor values while allowing sufficient flexibility.
A desirable prior should:
- be centered on a reasonable parametric model,
- allow local deviations from that model,
- use only a small number of mixture components where appropriate,
- adapt automatically as more data are observed.
Predictor-Dependent Stick-Breaking Processes
A natural construction uses predictor-dependent stick-breaking representations:
where:
- are global mixture atoms,
- are predictor-dependent stick-breaking weights.
For computational and conceptual simplicity, the atoms are often assumed predictor-independent, with only the weights depending on . This yields the infinite mixture regression model
which generalizes finite mixture-of-experts models to an infinite setting.
Kernel Stick-Breaking Processes
One effective construction is the kernel stick-breaking process (KSBP):
where:
- is a random location in predictor space,
- is a bandwidth parameter,
- is a kernel function attaining its maximum at .
This construction implies:
- mixture component has the largest influence near ,
- weights decay smoothly as moves away from ,
- the number of effectively active components is small in any local region of predictor space.
When the kernels are flat, the model reduces to a standard Dirichlet process mixture as a limiting case.
Probit Stick-Breaking Processes (PSBP)
An alternative with strong computational advantages is the probit stick-breaking process:
where:
- is the standard normal cumulative distribution function,
- ,
- is a regression function over predictors.
This approach has several key properties:
- when predictors are absent and , the model reduces to a Dirichlet process with precision 1,
- the parameter controls how quickly weights decay with , analogous to the DP concentration parameter,
- predictor dependence is introduced naturally through ,
- Gaussian latent variable representations enable efficient Gibbs sampling.
Logistic stick-breaking processes are similar but use a logistic link instead of a probit link.
Example: Glucose Tolerance Prediction
The probit stick-breaking process was applied to an epidemiological study of glucose tolerance in patients. The response variable was 2-hour plasma glucose level, with predictors including insulin sensitivity, age, waist-to-hip ratio, body mass index, and blood pressure measures.
Key findings include:
- the glucose distribution is strongly right-skewed,
- the shape of the distribution changes markedly with insulin sensitivity,
- mean or median regression models fail to capture these changes.
Bayesian density regression revealed that:
- insulin sensitivity and age have posterior inclusion probabilities close to 1,
- other predictors have very low inclusion probabilities and can be discarded,
- the right tail of the glucose distribution disappears as insulin sensitivity increases,
- aging amplifies the heavy right tail, especially for individuals with low insulin sensitivity.
These results demonstrate the ability of density regression to uncover predictor-dependent changes in distributional shape, not merely changes in central tendency.
Summary
Density regression using predictor-dependent stick-breaking processes provides:
- flexible modeling of full conditional densities,
- automatic adaptation of the number of mixture components,
- local modeling behavior in predictor space,
- principled Bayesian uncertainty quantification.
Compared with classical regression and finite mixture models, these approaches are especially valuable when distributional features such as skewness, multimodality, or tail behavior change with predictors, as commonly observed in biomedical and epidemiological data.
