"Classical" evidence lower bound approaches allows researcher to perform a simplified Bayesian inference over quite complex models, like deep learning models. This approach involves MC-like sampling at each optimization iteration. Alternative approach is to consider parameters W as a sample from unknown distribution that changes under action of optimization operator (like SGD) at each optimization step. From the researcher perspective, this approach is useufl because doesn't need to change the optimization at all.
Stochastic Gradient Fisher Scoring from paper
Constant SGD as Variational EM from paper
- Bylinkin Dmitry (Project wrapping, Final demo, Algorithms)
- Semenov Andrei (Tests writing, Library writing, Algorithms)
- Shestakov Alexander (Project planning, Blog post, Algorithms)
- Solodkin Vladimir (Basic code writing, Documentation writing, Algorithms)
The main folder for the library is swaglib. It has the following structure:
(TODO)
swaglib/
├── __init__.py
├── helpers.py
├── optimizers/
├── __init__.py
├── optimizer.py
├── elbo/
├── sgfs/