Article

Machine Learning: Bayesian Inference

中使用的 Computational Vision calculations, Bayesian Inference is a method to update model hypotheses z following observations on data x.

It utilizes Bayes’ theorem, that - 

The prior hypothesis on the latent variables z, p (z), quantifies our belief before a new batch of data x 是观察到的. We assume that data x are generated from a constant data-generating process, p(x) and that we can quantify the likelihood of observing x given our hypothesis z through the conditional distribution p (x | z). Despite being a somewhat abstract concept, Bayesian Inference has many real-world applications and is a powerful tool that underpins many machine learning techniques. A non-exhaustive list of applications includes:

  • Probabilistic and generative modelling – for example image standardization and Variational AutoEncoders
  • Uncertainty quantification for parametric models – for example Bayesian neural networks

One of the most ubiquitous applications for Bayesian Inference is uncertainty quantification. In parametric regression for example, our observed data are in fact pairs of datum (x,y) and our task is to predict labels y on new data points x, given that we have a closed-form expression for the posterior distribution p (z | {x, y}). Note that in this setting, the likelihood has a fixed-form of p (y | x, z). Uncertainty is then quantified by the posterior predictive distribution,

which has a closed-form expression for a limited number of scenarios, for example when p (y | x, z) is a normal distribution with isotropic standard deviation and whose expected value is a linear model of z. Otherwise, samples of the latent variable z must be generated from the posterior distribution p (z | {x, y}) and the expected value and variance of p (y│x, {x, y}) must be evaluated numerically.

It is worth concluding this introduction to Bayesian Inference by noting that the application of Bayes’ theorem as described here, assumes that we have a closed-form expression for evaluating the posterior distribution p (z│x). This is not always the case and approximations of p (z│x) must often be made in order for model predictions to be used, for example via the posterior predictive distribution. A large branch of machine learning called approximate inference, in fact exists to achieve just this.

At im体育APP, we use the Bayesian Inference method within our Digital 工程 operations. im体育APP的 Digital 工程 team for more information. 

找到相关的 资源

MAKING CERTAIN FOR NEARLY 190 YEARS

Digital 工程 Services

Our team of over 9,000 Engaged Experts in North America, 欧洲, The Middle East, 澳大利亚, Asia and Africa are ready to help you.