Figure 4 shows the structure for the hierarchical nonlinear variance (HNV) model. It utilises variance neurons and nonlinearities in building a hierarchical model for both the means and variances. Without the variance neurons the model would correspond to a multi-layer perceptron with latent variables at hidden neurons. Note that computation nodes as hidden neurons would result in multiple paths from upper layer latent variables to the observations. This type of structure was used in [5] and it has a quadratic as opposed to linear computational complexity.