Abstract
Finding suitable jobs for US Navy sailors from time to time is an important and ever-changing process. An Intelligent Distribution Agent and particularly its constraint satisfaction module take up the challenge to automate the process. The constraint satisfaction module’s main task is to assign sailors to new jobs in order to maximize Navy and sailor happiness. We present various neural network techniques combined with several statistical criteria to optimize the module’s performance and to make decisions in general. The data was taken from Navy databases and from surveys of Navy experts. Such indeterminate subjective component makes the optimization of the constraint satisfaction a very sophisticated task. Single-Layer Perceptron with logistic regression, Multilayer Perceptron with different structures and algorithms and Support Vector Machine with Adatron algorithm are presented for achieving best performance. Multilayer Perceptron neural network and Support Vector Machine with Adatron algorithm produced highly accurate classification and encouraging prediction.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Baars, B. J.: A Cognitive Theory of Consciousness, Cambridge University Press, Cambridge, 1988.
Baars, B. J.: In the Theater of Consciousness, Oxford University Press, Oxford, 1997.
Franklin, S.: Artificial Minds, Cambridge, Mass., MIT Press, 1995.
Franklin, S., Kelemen, A., McCauley, L.: IDA: A cognitive agent architecture, In the proceedings of IEEE International Conference on Systems, Man, and Cybernetics ‘88, IEEE Press, pp. 2646–2651, 1998.
Kondadadi, R., Dasgupta, D., Franklin, S.: An Evolutionary Approach for Job Assignment, Proceedings of International Conference on Intelligent Systems-2000, Louisville, Kentucky, pp. 139–142, 2000.
Macs, P.: How to Do the Right Thing, Connection Science, 1: 3, pp. 291–323, 1990.
Liang T. T., Thompson, T. J.: Applications and Implementation — A large-scale personnel assignment model for the Navy, The Journal For The Decisions Sciences Institute, Volume 18, No. 2, pp. 234–249, Spring, 1987.
Liang, T. T., Buclatin, B.B.: Improving the utilization of training resources through optimal personnel assignment in the U.S. Navy, European Journal of Operational Research 33, pp. 183–190, North-Holland, 1988.
Gale D., Shapley, L. S.: College Admissions and stability of marriage, The American Mathematical monthly, Vol 60, No 1, pp. 9–15, 1962.
Bishop, C. M.: Neural Networks for Pattern Recognition, Oxford University Press, 1995.
Haykin, S.: Neural Networks, Prentice Hall Upper Saddle River, New Jersey, 1999.
Girosi, F., Jones, M., Poggio, T.: Regularization theory and neural networks architectures, Neural Computation, 7, pp. 219–269, 1995.
Biganzoli, E., Boracchi, P., Mariani, L., Marubini, E.: Feed Forward Neural Networks for the Analysis of Censored Survival Data: A Partial Logistic Regression Approach, Statistics in Medicine 17, pp. 1169–1186, 1998.
Akaike, H.: A new look at the statistical model identification, IEEE Trans. Automatic Control, Vol. 19, No. 6, pp. 716–723, 1974.
Rissanen, J.: Modeling by shortest data description. Automat., Vol. 14, pp. 465–471, 1978.
Ripley, B.D.: Statistical aspects of neural networks, in Barndorff-Nielsen, O. E., Jensen, J. L. and Kendall, W. S. (eds), Networks and Chaos — Statistical and Probabilistic Aspects, Chapman and Hall, London, pp. 40–123, 1993.
Ripley, B. D.: Statistical ideas for selecting network architectures, in Kappen, B. and Gielen, S. (eds), Neural Networks: Artificial Intelligence and Industrial Applications, Springer, London, pp. 183–190, 1995.
Stone, M.: Cross-validatory choice and assessment of statistical predictions (with discussion), Journal of the Royal Statistical Society, Series B, 36, pp. 111–147, 1974.
Vapnik, V. N.: Statistical Learning Theory, Springer, 1998.
Cortes C., Vapnik, V.: Support vector machines, Machine Learning, 20, pp. 273–297, 1995.
Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines (and other kernel-based learning methods), Cambridge University Press, 2000.
Brown, M. P. S., Grundy, W. N., Lin, D., Cristianini, N., Sugnet, C., Ares, M., Haussier, D.: Support Vector Machine Classification of Microarray Gene Expression Data, University of California Technical Report USCC-CRL-99–09, 1999.
Friess T. T., Cristianini N., Campbell C., The Kernel-Adatron: a Fast and Simple Learning Procedure for Support Vector Machines. In Shavlik, J. (ed), Proceedings of the Fifteenth International Conference on Machine Learning, pp. 188–196, 1998.
Anlauf, J. K., Bich!, M.: The Adatron: an adaptive perceptron algorithm, Europhysics Letters, 10 (7), pp. 687–692, 1989.
Boser, B., Guyon, I., Vapnik, V.: A Training Algorithm for Optimal Margin Classifiers, In Proceedings of the Fifth Workshop on Computational Learning Theory, pp. 144–152, 1992.
Schumacher, M., Rossner, R., Vach, W.: Neural networks and logistic regression: Part I’, Computational Statistics and Data Analysis, 21, pp. 661–682, 1996.
Principe, J., Lefebvre, C., Lynn, G., Fancourt C., Wooten D.: Neurosolutions User’s Manual, Release 4. 10, Gainesville, FL: NeuroDimension, Inc., 2001.
Matlab User Manual, Release 6. 0, Natick, MA: MathWorks, Inc., 2000.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Kelemen, A., Liang, Y., Kozma, R., Franklin, S. (2003). Optimizing Intelligent Agent’s Constraint Satisfaction with Neural Networks. In: Abraham, A., Jain, L.C., Kacprzyk, J. (eds) Recent Advances in Intelligent Paradigms and Applications. Studies in Fuzziness and Soft Computing, vol 113. Physica, Heidelberg. https://doi.org/10.1007/978-3-7908-1770-6_12
Download citation
DOI: https://doi.org/10.1007/978-3-7908-1770-6_12
Publisher Name: Physica, Heidelberg
Print ISBN: 978-3-7908-2521-3
Online ISBN: 978-3-7908-1770-6
eBook Packages: Springer Book Archive