Abstract
In this paper, we demonstrate that cabinet approval ratings can automatically be inferred with good performance by a neural network technique, that is, information-theoretic competitive learning. Because cabinet approval rating estimation is an extremely complex process with much non-linearity, neural networks may give much better performance than conventional statistical methods. Though an attempt to infer public opinions seem to be a challenging topic for machine learning, little attempts have been made to infer approval ratings to our best knowledge. In this context, we try to apply information-theoretic competitive learning to the problem of cabinet approval ratings. Information-theoretic competitive learning has been developed so as to simulate competitive processes of neurons. One of the main characteristics of the method is that it is a very soft-type of competitive learning in which conventional competitive learning is only a special case. Though the method seems to be promising due to its general property, we have had a few experimental results to show better performance. Experimental results show that without any teacher information neural networks can appropriately infer the rise and fall of approval ratings through a process of information maximization. This experiment result surely opens up new perspectives for neural networks as well as mass communication studies.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Miller, M.M., Denham, B.: Horserace, issue coverage in prestige newspapers during 1988, 1992 elections. Newspaper Research Journal 15(4), 20–28 (1994)
Miller, M.M., Andsager, J.L., Riechert, B.P.: Framing the candidates in presidential primaries: issues and images in press releases and news coverage. Journalism and Mass communication quarterly 75(2), 312–324 (1998)
Domke, D., Fan, D.P., Michael, S., Smith, D.V.S., Watts, M.D.: News media, candidates and issues, and public opinion in the 1996 presidential campaign. Journalism and Mass Communication Quaterly 74(4), 718–737 (1996)
Fan, D.: Computer content analysis of press coverage and prediction of public opinion for the 1995 sovereignty referendum in quebec. Journalism and Mass Communication Quaterly 74(4), 351–366 (1996)
Watts, M.D., Domke, D., Shah, D.V., Fan, D.P.: Elite cues and media bias in presidential campaigns. Communication Research 26(2), 144–175 (1999)
Danowski, J.A., Rebecca, A.: Linking gender language in news about presidential candidates to gender gaps in polls: a time-series analysis of the 1996 campaign. Ablex Publishing, Westport (1996)
Yoshida, F.: Main features of tex-ray, a software for analyzing japanese sentences, and its applications: an attempt to predict poll support ratings for koizumi cabinet from editorial content of four major newspapers (in japanese). Journal of mass communication studies 68, 80–96 (2006)
Linsker, R.: How to generate ordered maps by maximizing the mutual information between input and output. Neural Computation 1, 402–411 (1989)
Atick, J.J., Redlich, A.N.: Toward a theory of early visual processing. Neural Computation 2, 308–320 (1990)
Becker, S.: Mutual information maximization: models of cortical self-organization. Network: Computation in Neural Systems 7, 7–31 (1996)
Becker, S., Hinton, G.E.: Learning mixture models of spatial coherence. Neural Computation 5, 267–277 (1993)
Kamimura, R., Kamimura, T., Shultz, T.R.: Information theoretic competitive learning and linguistic rule acquistion. Transactions of the Japanese Society for Artificial Intelligence 16(2), 287–298 (2001)
Kamimura, R., Kamimura, T., Uchida, O.: Flexible feature discovery and structural information. Connection Science 13(4), 323–347 (2001)
Kamimura, R.: Information theoretic competitive learning in self-adaptive multi-layered networks. Connection Science 13(4), 323–347 (2003)
Kamimura, R.: Teacher-directed learning: information-theoretic competitive learning in supervised multi-layered networks. Connection Science 15, 117–140 (2003)
Kamimura, R.: Information-theoretic competitive learning with inverse euclidean distance. Neural Processing Letters 18, 163–184 (2003)
Kamimura, R.: Unifying cost and information in information-theoretic competitive learning. Neural Networks 18, 711–718 (2006)
Kamimura, R.: Improving information-theoretic competitive learning by accentuated information maximization. International Journal of General Systems 34(3), 219–233 (2006)
Rumelhart, D.E., Zipser, D.: Feature discovery by competitive learning. In: Rumelhart, D.E., G.E.H., et al. (eds.) Parallel Distributed Processing, vol. 1, pp. 151–193. MIT Press, Cambridge (1986)
Grossberg, S.: Competitive learning: from interactive activation to adaptive resonance. Cognitive Science 11, 23–63 (1987)
DeSieno, D.: Adding a conscience to competitive learning. In: Proceedings of IEEE International Conference on Neural Networks, San Diego, pp. 117–124. IEEE, Los Alamitos (1988)
Ahalt, S.C., Krishnamurthy, A.K., Chen, P., Melton, D.E.: Competitive learning algorithms for vector quantization. Neural Networks 3, 277–290 (1990)
Xu, L.: Rival penalized competitive learning for clustering analysis, RBF net, and curve detection. IEEE Transaction on Neural Networks 4(4), 636–649 (1993)
Luk, A., Lien, S.: Properties of the generalized lotto-type competitive learning. In: Proceedings of International conference on neural information processing, San Mateo, pp. 1180–1185. Morgan Kaufmann Publishers, San Francisco (2000)
Hulle, M.M.V.: The formation of topographic maps that maximize the average mutual information of the output responses to noiseless input signals. Neural Computation 9(3), 595–606 (1997)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kamimura, R., Yoshida, F. (2006). Automatic Inference of Cabinet Approval Ratings by Information-Theoretic Competitive Learning. In: King, I., Wang, J., Chan, LW., Wang, D. (eds) Neural Information Processing. ICONIP 2006. Lecture Notes in Computer Science, vol 4233. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11893257_99
Download citation
DOI: https://doi.org/10.1007/11893257_99
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-46481-5
Online ISBN: 978-3-540-46482-2
eBook Packages: Computer ScienceComputer Science (R0)