Abstract
The present chapter describes applications of error entropy (and entropyinspired) risks, in a variety of classification tasks performed by more sophisticated machines than those considered in the preceding chapters. These include multi-layer perceptrons (MLPs), recurrent neural networks (RNNs), complex-valued neural networks (CVNNs), modular neural networks (MNNs), and decision trees. We also present a clustering algorithm using a MEE-like concept, LEGClust, which is used in building MNNs. Besides implementation issues, an extensive set of experimental results and comparisons to non-EE approaches are presented. Since the respective learning algorithms use the empirical versions of the risks, the corresponding acronyms (MSE, CE, SEE, and so forth) labeling tables and graphs of results refer from now on to the empirical versions of the risks.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2013 Springer Berlin Heidelberg
About this chapter
Cite this chapter
Marques de Sá, J.P., Silva, L.M.A., Santos, J.M.F., Alexandre, L.A. (2013). Applications. In: Minimum Error Entropy Classification. Studies in Computational Intelligence, vol 420. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-29029-9_6
Download citation
DOI: https://doi.org/10.1007/978-3-642-29029-9_6
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-29028-2
Online ISBN: 978-3-642-29029-9
eBook Packages: EngineeringEngineering (R0)