Machine Learning: Bayesian Inference
中使用的 Computational Vision calculations, Bayesian Inference is a method to update model hypotheses z following observations on data x.
It utilizes Bayes’ theorem, that -
The prior hypothesis on the latent variables z, p (z), quantifies our belief before a new batch of data x 是观察到的. We assume that data x are generated from a constant data-generating process, p(x) and that we can quantify the likelihood of observing x given our hypothesis z through the conditional distribution p (x | z). Despite being a somewhat abstract concept, Bayesian Inference has many real-world applications and is a powerful tool that underpins many machine learning techniques. A non-exhaustive list of applications includes:
- Probabilistic and generative modelling – for example image standardization and Variational AutoEncoders
- Uncertainty quantification for parametric models – for example Bayesian neural networks
One of the most ubiquitous applications for Bayesian Inference is uncertainty quantification. In parametric regression for example, our observed data are in fact pairs of datum (x,y) and our task is to predict labels y on new data points x, given that we have a closed-form expression for the posterior distribution p (z | {x, y}). Note that in this setting, the likelihood has a fixed-form of p (y | x, z). Uncertainty is then quantified by the posterior predictive distribution,
which has a closed-form expression for a limited number of scenarios, for example when p (y | x, z) is a normal distribution with isotropic standard deviation and whose expected value is a linear model of z. Otherwise, samples of the latent variable z must be generated from the posterior distribution p (z | {x, y}) and the expected value and variance of p (y│x, {x, y}) must be evaluated numerically.
It is worth concluding this introduction to Bayesian Inference by noting that the application of Bayes’ theorem as described here, assumes that we have a closed-form expression for evaluating the posterior distribution p (z│x). This is not always the case and approximations of p (z│x) must often be made in order for model predictions to be used, for example via the posterior predictive distribution. A large branch of machine learning called approximate inference, in fact exists to achieve just this.
At im体育APP, we use the Bayesian Inference method within our Digital 工程 operations. im体育APP的 Digital 工程 team for more information.
找到相关的 资源
Digital 工程 Services
建模 and 模拟
We specialize in modeling and simulation to accelerate your research and development initiatives, optimize your designs and support safe, efficient operations.
Data Science and Machine Learning
Learn how we combine data science and machine learning methods, higher-level statistics, and scientific programming with expertise in the physical sciences to increase productivity and growth.
Discrete im体育APP Method (DEM)
Our Discrete im体育APP Method (DEM) services deliver innovative solutions that have helped companies across a wide range of industries.