Summary
The purpose of this chapter is to make the reader aware of some of the software packages available that can implement probability models connected with medical informatics. The modelling techniques considered are logistic regression, neural networks, Bayesian networks, class-probability trees, and hidden Markov models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
J.H. Badsberg. An Environment for Graphical Models. PhD dissertation, Department of Mathematical Sciences, Aalborg University, 1995.
J.M.G. Barahona, P.D.H. Quiros, and T. Bollinger. A brief history of free software and open source. IEEE Software, 16(1):32–33, 1999.
R.A. Becker, J.M. Chambers, and A.R. Wilks. The New S Language. Wadworth & Brooks/Cole, Pacific Grove, CA, 1988.
C.M. Bishop. Neural Networks for Pattern Recognition. Clarendon Press, Oxford, 1995.
C.M. Bishop, D. Spiegelhalter, and J. Winn. VIBES: A variational inference engine for Bayesian networks. In Advances in Neural Information Processing Systems 9, MIT Press, Cambridge, MA, 2003.
S.G. Bottcher and C. Dethlefsen. Deal: A package for learning Bayesian networks. Technical report, Department of Mathematical Sciences, Aalborg University, 2003.
L. Breiman, J.H. Friedman, R.A. Olshen, and C.J. Stone. Classification and Regression Trees. Chapman & Hall, New York, 1984.
W. Buntine. A Theory of Learning Classification Rules. PhD dissertation, School of Computing Science, University of Technology, Sydney, February 1990.
S.K. Card, J.D. Mackinlay, and B. Shneiderman. Readings in Information Visualization: Using Vision to Think. Morgan Kaufmann, San Fransisco, CA, 1999.
J.M. Chambers and T.J. Hastie. Statistical Models in S. Wadsworth & Brooks/Cole Advanced Books & Software, Pacific Grove, CA, 1992.
J. Cheng and R. Greiner. Learning Bayesian belief network classifiers: Algorithms and systems. In E. Stroulia and S. Matwin, editors, Proceedings of the 14th Canadian Conference on Artificial Intelligence, Lecture Notes in Computer Science, pages 141–151, Springer-Verlag, New York, 2001.
D. Collett. Modelling Binary Data. Chapman & Hall, London, 1991.
G.F Cooper and E. Herskovits. A Bayesian method for the induction of probabilistic networks from data. Machine Learning, 9:309–347, 1992.
F.G. Cozman. The JavaBayes system. The ISBA Bulletin, 7(4):16–21, 2001.
A.C. Davidson and D.V. Hinkley. Bootstrap Methods and Their Applications. Cambridge University Press, Cambridge, 1997.
K.A. De Jong. Evolutionary Computation: A Unified Approach. MIT Press, Cambridge, MA, 2003.
D. Edwards. Introduction to Graphical Modelling. Springer-Verlag, New York, 2nd edition, 2000.
B. Efron and R.J. Tibshirani. An Introduction to the Bootstrap. Chapman & Hall, New York, 1993.
D.B. Fogel. Evolutionary Computation: Toward a New Philosophy of Machine Intelligence. IEEE Press, New York, 1995.
K. Fogel. Open Source Development with CVS. Coriolis, Scottsdale, AZ, 1999.
N. Friedman. The Bayesian structural EM algorithm. In G.F. Cooper and S. Moral, editors, Proceedings of the 14th Conference on Uncertainty in Artificial Intelligence, pages 129–138, Morgan Kaufmann, San Francisco, CA, 1998.
M.N. Gibbs. Bayesian Gaussian Processes for Regression and Classification. PhD dissertation, Department fo Computing Science, University of Cambridge, 1997.
W.R. Gilks, S. Richardson, and D.J. Spiegelhalter, editors. Markov Chain Monte Carlo in Practice. Chapman & Hall, London, 1996.
M.T. Hagan, H.B. Demuth, and M. Beale. Neural Network Design. PWS Publishing, Boston, 1996.
F.E. Harrell. Regression Modeling Strategies. Springer, New York, 2001.
E. Horvitz, D. Hovel, and C. Kadie. MSBNx: A component-centric toolkit for modeling and inference with Bayesian networks. Technical Report MSR-TR-2001-67, Microsoft Research, Redmond, WA, July 2001.
B.R. Hunt, R.L. Lipsman, and J.M. Rosenberg. A Guide to MATLAB: For Beginners and Experienced Users. Cambridge University Press, Cambridge, 2001.
R. Ihaka and R. Gentleman. R: a language for data analysis and graphics. Journal of Computational and Graphical Statistics, 5:299–314, 1996.
H. James. Editorial. Neural Computing and Applications, 5:129–130, 1997.
K.B. Korb and A.E. Nicholson. Bayesian Artificial Intelligence. CRC Press, London, 2003.
S.L. Lauritzen. gRaphical models in R. R News, 3(2):39, 2002.
K.P. Murphy. The Bayes Net Toolbox for Matlab. Computing Science and Statistics, 33:331–350, 2001, The Interface Foundation of North America.
I.T. Nabney. NETLAB: Algorithms for Pattern Recognition. Springer, London, 2002.
G.M. Nielson, H. Hagan, and H. Müller. Scientific Visualization: Overviews, Methodologies, and Techniques. IEEE Computer Society, Los Alamitos, CA, 1997.
R.C. O’Reilly and Y. Munakata. Computational Explorations in Cognitive Neuroscience. MIT Press, Cambridge, MA, 2000.
R.C. Pavlicek. Embracing Insanity: Open Source Software Development. SAMS, Indianapolis, IN, 2000.
B. Perens. The Open Source definition. In C. DiBona and S. Ockman, editors, Open Sources: Voices From the Open Source Revolution, pages 171–188. O’Reilly & Associates, Sebastopol, CA, 1999.
J.C. Principe, N.R. Euliano, and W.C. Lefebvre. Neural and Adaptive Systems. John Wiley, New York, 2000.
J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufman, San Mateo, CA, 1993.
A.E. Raftery and C. Volinsky, 1999. Bayesian Model Averaging Home Page [WWW]. Available from: http://www.research.att.com/volinsky/bma.html [accessed 9 July 1999].
M. Ramoni and P. Sebastiani. Learning Bayesian networks from incomplete databases. Technical Report KMI-TR-43, Knowledge Media Institute, Open University, February 1997.
D.K. Rosenberg. Open Source: The Unauthorized White Papers. M & T Books, Foster City, CA, 2000.
R. Scheines, P. Spirtes, C. Glymour, and C. Meek. TETRAD II: Tools for Discovery. Lawrence Erlbaum Associates, Hillsdale, NJ, 1994.
D. Spiegelhalter, A. Thomas, N. Best, and W. Gilks. BUGS: Bayesian inference Using Gibbs Sampling. MRC Biostatistics Unit, Cambridge, 1996.
D.J. Spiegelhalter, A. Thomas, and N.G. Best. WinBUGS Version 1.2 User Manual. MRC Biostatistics Unit, Cambridge, 1999.
P. Spirtes, C. Glymour, and R. Scheines. Causation, Prediction, and Search. MIT Press, Cambridge, MA, 2nd edition, 2001.
R. Stallman. The GNU Operating System and the Free Software Movement. In C. DiBona and S. Ockman, editors, Open Sources: Voices From the Open Source Revolution, pages 53–70. O’Reilly & Associates, Sebastopol, CA, 1999.
M. Svensén. GTM: The Generative Topographic Mapping. PhD dissertation, Neural Computing Research Group, Aston University, April 1998.
W.N. Venables and B.D. Ripley. Modern Applied Statistics with S-Plus. Springer, New York, 3rd edition, 1999.
W.N. Venables and B.D. Ripley. S Programming. Springer, New York, 2000.
G.R. Warnes. HYDRA: a Java library for Markov chain Monte Carlo. Journal of Statistical Software, 7: issue 4, 2002.
M. Welsh, M.K. Dalheimer, and L. Kaufman. Running Linux. O’Reilly & Associates, Sebastopol, CA, 3rd edition, 1999.
D.J. Wilkinson and S.K.H. Yeung. A sparse matrix approach to Bayesian computation in large linear models. Computational Statistics and Data Analysis, 44:423–516, 2004.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag London Limited
About this chapter
Cite this chapter
Dybowski, R. (2005). Software for Probability Models in Medical Informatics. In: Husmeier, D., Dybowski, R., Roberts, S. (eds) Probabilistic Modeling in Bioinformatics and Medical Informatics. Advanced Information and Knowledge Processing. Springer, London. https://doi.org/10.1007/1-84628-119-9_16
Download citation
DOI: https://doi.org/10.1007/1-84628-119-9_16
Publisher Name: Springer, London
Print ISBN: 978-1-85233-778-0
Online ISBN: 978-1-84628-119-8
eBook Packages: Computer ScienceComputer Science (R0)