Note: we are not actively developing this library anymore, but we are still maintaining it. We recommend instead looking at Aboleth, which has similar functionality and is implemented on top of TensorFlow.
A library of scalable Bayesian generalized linear models with fancy features
revrand is a python (2 and 3) supervised machine learning library that contains implementations of various Bayesian linear and generalized linear models (i.e. Bayesian linear regression and Bayesian generalized linear regression).
revrand can be used for large scale approximate Gaussian process regression, like GPflow and GPy, but it uses random basis kernel approximations (see , , ) as opposed to inducing point approximations.
A few features of this library are:
- Random Basis functions that can be used to approximate Gaussian processes with shift invariant covariance functions (e.g. Matern) when used with linear models , , .
- A fancy basis functions/feature composition framework for combining basis functions like those above and radial basis functions, sigmoidal basis functions, polynomial basis functions etc with basis function parameter learning.
- Non-Gaussian likelihoods with Bayesian generalized linear models (GLMs). We infer all of the parameters in the GLMs using stochastic variational inference , and we approximate the posterior over the weights with a mixture of Gaussians, like .
- Large scale learning using stochastic gradients (Adam, AdaDelta and more).
- Scikit Learn compatibility, i.e. usable with pipelines.
- A host of decorators for scipy.optimize.minimize and stochastic gradients that enhance the functionality of these optimisers.
Here is an example of approximating a Matern 3/2 kernel with some of our basis functions,
here is an example of the algorithms in revrand approximating a Gaussian Process,
and here is an example of running using our Bayesian GLM with a Poisson likelihood and integer observations,
Have a look at some of the demo notebooks for how we generated these plots, and more!
To install, you can use
$ pip install revrand
or simply run
setup.py in the location where you have cloned or
downloaded this repository:
$ python setup.py install
Now have a look at our quickstart guide to get up and running quickly!
- Home Page
- Report on the algorithms in revrand
- Issue tracking
Bugs & Feedback
For bugs, questions and discussions, please use Github Issues.
|||(1, 2) Yang, Z., Smola, A. J., Song, L., & Wilson, A. G. "A la Carte -- Learning Fast Kernels". Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, pp. 1098-1106, 2015.|
|||(1, 2) Le, Q., Sarlos, T., & Smola, A. "Fastfood-approximating kernel expansions in loglinear time." Proceedings of the international conference on machine learning. 2013.|
|||(1, 2) Rahimi, A., & Recht, B. "Random features for large-scale kernel machines". Advances in neural information processing systems. 2007.|
|||Kingma, D. P., & Welling, M. "Auto-encoding variational Bayes". Proceedings of the 2nd International Conference on Learning Representations (ICLR). 2014.|
|||Gershman, S., Hoffman, M., & Blei, D. "Nonparametric variational inference". Proceedings of the international conference on machine learning. 2012.|
Copyright & License
Copyright 2015 National ICT Australia.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.