이 글은 Generative model 에 대한 필자의 이해를 높이고자 작성된 글입니다. 참고자료는 자료1 입니다.
Latent variable model(LVM)
- defines a distribution over observation x by using a (vector) latent variable z(: as an explanation for the observation)and specifying:
- The prior distribution p(z) for the latent variable
- The likelihood p(x|z) that connects the latent variable to the observation
- The joint distribution p(x,z)=p(x|z)p(z)
We are interested in computing the marginal likelihood p(x) and the posterior posterior distribution p(z|x). This comes up with generating a latent variable which contains the observation from the model, and it follows:
z∼p(z),x∼p(x|z)
From going from observations to latent values p(z|x), which is called inference.
p(z|x)=p(x,z)p(x)=p(x,z)∫p(x,z)dz
So we have to compute marginal likelihood p(x)=∫p(x,z)dz, since we cannot evaluate this anatically, we take a detour
p(x|z)p(z)=p(x,z)=p(z|x)p(x)
As computing p(z|x) as inverting p(x|z) probabilistically. But this is intractable, to avoid intractable inference:
- Invertible model/Normalizing flows
- Designing models for which inferece is tractable
- Key idea: approximate the data distribution by transforming the prior distribution using an invertible function
- Simpler training but less expressive models
- Using approximate inference
- Markov Chain monte Carlo: generate samples from the exact posterior using a Markov Chain
- Variational Inference: approximate the posterior pθ(z|x) with a tractable distributionqϕ(z|x)
'AI, Deep Learning Basics > Methodology' 카테고리의 다른 글
[Diffusion Model] DDPM, DDIM (0) | 2023.01.28 |
---|---|
(작성중) [Generative Model] Generative Adversarial Networks(GAN) (0) | 2022.03.17 |
Uncertainty (0) | 2022.02.20 |