Skip to main content

Part of the book series: Advanced Information and Knowledge Processing ((AI&KP))

Summary

The purpose of this chapter is to introduce the reader to a type of artificial neural network called a multi-layer perceptron. The intention is not to present a detailed, comprehensive treatise on the subject; instead, we provide a brief, informal, tutorial in the spirit of Chapter 1.

We start with the historical background and then introduce the concept of regression. This concept holds for both continuous-valued and binary-valued response variables; however, when applied to the latter, probabilistic classification models are created.

In the context of medicine, probabilistic classification models are usually obtained using logistic regression analysis, but if logistic functions are nested to produce multi-layer perceptrons, the high flexibility of the resulting models enables them to handle complex classification tasks.

We discuss some important points that should be kept in mind when applying multi-layer perceptrons to classification problems. Finally, we end the tutorial with some recommended reading.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. I. Aleksander and H. Morton. An Introduction to Neural Computing. Chapman & Hall, London, 1990.

    Google Scholar 

  2. E.B. Baum and D. Haussler. What size net gives valid generalization? Neural Computation, 1(1):151–215, 1989.

    Google Scholar 

  3. R. Beale and T. Jackson. Neural Computing: An Introduction. IOP Publishing, Bristol, 1990.

    Google Scholar 

  4. C.M. Bishop. Novelty detection and neural network validation. IEE Proceedings: Vision, Image & Signal Processing, 141:217–222, 1994.

    Article  Google Scholar 

  5. C.M. Bishop. Neural Networks for Pattern Recognition. Clarendon Press, Oxford, 1995.

    Google Scholar 

  6. A. Blum and P. Langley. Selection of relevant features and examples in machine learning. Artificial Intelligence, 97(1–2):245–271, 1997.

    Article  MathSciNet  Google Scholar 

  7. S.E. Fahlman and C. Lebiere. The cascade-correlation learning architecture. In D.S. Touretzky, editor, Advances in Neural Information Processing Systems 2, pages 524–532, Los Altos, CA, 1990. Morgan Kaufmann.

    Google Scholar 

  8. K. Gurney. An Introduction to Neural Networks. UCL Press, London, 1997.

    Google Scholar 

  9. D.J. Hand. Construction and Assessment of Classification Rules. John Wiley, Chichester, 1997.

    Google Scholar 

  10. D. Hebb. Organization of Behaviour. Wiley, New York, 1949.

    Google Scholar 

  11. K. Hornik. Approximation capabilities of multilayer feedforward networks. Neural Networks, 4(2):251–257, 1991.

    Article  Google Scholar 

  12. R. Kohavi and G. John. Wrappers for feature selection. Artificial Intelligence, 97(1–2):273–324, 1997.

    Article  Google Scholar 

  13. W.J. Krzanowski. Principles of Multivariate Analysis. Oxford University Press, Oxford, 1988.

    Google Scholar 

  14. D.J.C. MacKay. A practical Bayesian framework for back-propagation networks. Neural Computation, 4(3):448–472, 1992.

    Google Scholar 

  15. T. Masters. Practical Neural Network Recipes in C++. Academic Press, London, 1993.

    Google Scholar 

  16. W. McCulloch and W. Pitts. A logical calculus of the ideas immanent in nervous activity. Bulletin of Mathematical Biophysics, 5:115–133, 1943.

    MathSciNet  Google Scholar 

  17. M.L. Minsky and S.A. Papert. Perceptrons. MIT Press, Cambridge, 1969.

    Google Scholar 

  18. I.T. Nabney. NETLAB: Algorithms for Pattern Recognition. Springer, London, 2002.

    Google Scholar 

  19. R.M. Neal. Bayesian Learning for Neural Networks. PhD thesis, University of Toronto, 1994.

    Google Scholar 

  20. N. Rashevsky. Topology and life: In search of general mathematical principles in biology and sociology. Bulletin of Mathematical Biophysics, 16:317–348, 1954.

    MathSciNet  Google Scholar 

  21. B.D. Ripley. Pattern Recognition and Neural Networks. Cambridge University Press, Cambridge, 1996.

    Google Scholar 

  22. F. Rosenblatt. The perceptron: A probabilistic model for information storage and organization in the brain. Psychological Review, 65:386–408, 1958.

    MathSciNet  Google Scholar 

  23. F. Rosenblatt. On the convergence of reinforcement procedures in simple perceptrons. Technical report VG-1196-G-4, Cornell Aeronautical Laboratory, Buffalo, NY, 1960.

    Google Scholar 

  24. L. Tarassenko. A Guide to Neural Computing Applications. Arnold, London, 1998.

    Google Scholar 

  25. L. Tarassenko, P. Hayton, N. Cerneaz, and M. Brady. Novelty detection for the identification of masses in mammograms. In Proceedings of the 4th IEE International Conference on Artificial Neural Networks, pages 442–447, Cambridge, 1995. Cambridge University Press.

    Google Scholar 

  26. P.J. Werbos. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. PhD thesis, Harvard University, Cambridge, MA, 1974.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag London Limited

About this chapter

Cite this chapter

Dybowski, R. (2005). A Casual View of Multi-Layer Perceptrons as Probability Models. In: Husmeier, D., Dybowski, R., Roberts, S. (eds) Probabilistic Modeling in Bioinformatics and Medical Informatics. Advanced Information and Knowledge Processing. Springer, London. https://doi.org/10.1007/1-84628-119-9_3

Download citation

  • DOI: https://doi.org/10.1007/1-84628-119-9_3

  • Publisher Name: Springer, London

  • Print ISBN: 978-1-85233-778-0

  • Online ISBN: 978-1-84628-119-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics