Automatic Inference of Cabinet Approval Ratings by Information-Theoretic Competitive Learning

  • Ryotaro Kamimura
  • Fumihiko Yoshida
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 4233)


In this paper, we demonstrate that cabinet approval ratings can automatically be inferred with good performance by a neural network technique, that is, information-theoretic competitive learning. Because cabinet approval rating estimation is an extremely complex process with much non-linearity, neural networks may give much better performance than conventional statistical methods. Though an attempt to infer public opinions seem to be a challenging topic for machine learning, little attempts have been made to infer approval ratings to our best knowledge. In this context, we try to apply information-theoretic competitive learning to the problem of cabinet approval ratings. Information-theoretic competitive learning has been developed so as to simulate competitive processes of neurons. One of the main characteristics of the method is that it is a very soft-type of competitive learning in which conventional competitive learning is only a special case. Though the method seems to be promising due to its general property, we have had a few experimental results to show better performance. Experimental results show that without any teacher information neural networks can appropriately infer the rise and fall of approval ratings through a process of information maximization. This experiment result surely opens up new perspectives for neural networks as well as mass communication studies.


Input Pattern Modality Pattern Negative Word Generalization Error Input Unit 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Miller, M.M., Denham, B.: Horserace, issue coverage in prestige newspapers during 1988, 1992 elections. Newspaper Research Journal 15(4), 20–28 (1994)Google Scholar
  2. 2.
    Miller, M.M., Andsager, J.L., Riechert, B.P.: Framing the candidates in presidential primaries: issues and images in press releases and news coverage. Journalism and Mass communication quarterly 75(2), 312–324 (1998)Google Scholar
  3. 3.
    Domke, D., Fan, D.P., Michael, S., Smith, D.V.S., Watts, M.D.: News media, candidates and issues, and public opinion in the 1996 presidential campaign. Journalism and Mass Communication Quaterly 74(4), 718–737 (1996)Google Scholar
  4. 4.
    Fan, D.: Computer content analysis of press coverage and prediction of public opinion for the 1995 sovereignty referendum in quebec. Journalism and Mass Communication Quaterly 74(4), 351–366 (1996)Google Scholar
  5. 5.
    Watts, M.D., Domke, D., Shah, D.V., Fan, D.P.: Elite cues and media bias in presidential campaigns. Communication Research 26(2), 144–175 (1999)CrossRefGoogle Scholar
  6. 6.
    Danowski, J.A., Rebecca, A.: Linking gender language in news about presidential candidates to gender gaps in polls: a time-series analysis of the 1996 campaign. Ablex Publishing, Westport (1996)Google Scholar
  7. 7.
    Yoshida, F.: Main features of tex-ray, a software for analyzing japanese sentences, and its applications: an attempt to predict poll support ratings for koizumi cabinet from editorial content of four major newspapers (in japanese). Journal of mass communication studies 68, 80–96 (2006)Google Scholar
  8. 8.
    Linsker, R.: How to generate ordered maps by maximizing the mutual information between input and output. Neural Computation 1, 402–411 (1989)CrossRefGoogle Scholar
  9. 9.
    Atick, J.J., Redlich, A.N.: Toward a theory of early visual processing. Neural Computation 2, 308–320 (1990)CrossRefGoogle Scholar
  10. 10.
    Becker, S.: Mutual information maximization: models of cortical self-organization. Network: Computation in Neural Systems 7, 7–31 (1996)MATHCrossRefGoogle Scholar
  11. 11.
    Becker, S., Hinton, G.E.: Learning mixture models of spatial coherence. Neural Computation 5, 267–277 (1993)CrossRefGoogle Scholar
  12. 12.
    Kamimura, R., Kamimura, T., Shultz, T.R.: Information theoretic competitive learning and linguistic rule acquistion. Transactions of the Japanese Society for Artificial Intelligence 16(2), 287–298 (2001)CrossRefGoogle Scholar
  13. 13.
    Kamimura, R., Kamimura, T., Uchida, O.: Flexible feature discovery and structural information. Connection Science 13(4), 323–347 (2001)CrossRefGoogle Scholar
  14. 14.
    Kamimura, R.: Information theoretic competitive learning in self-adaptive multi-layered networks. Connection Science 13(4), 323–347 (2003)CrossRefMathSciNetGoogle Scholar
  15. 15.
    Kamimura, R.: Teacher-directed learning: information-theoretic competitive learning in supervised multi-layered networks. Connection Science 15, 117–140 (2003)CrossRefGoogle Scholar
  16. 16.
    Kamimura, R.: Information-theoretic competitive learning with inverse euclidean distance. Neural Processing Letters 18, 163–184 (2003)CrossRefGoogle Scholar
  17. 17.
    Kamimura, R.: Unifying cost and information in information-theoretic competitive learning. Neural Networks 18, 711–718 (2006)CrossRefGoogle Scholar
  18. 18.
    Kamimura, R.: Improving information-theoretic competitive learning by accentuated information maximization. International Journal of General Systems 34(3), 219–233 (2006)CrossRefGoogle Scholar
  19. 19.
    Rumelhart, D.E., Zipser, D.: Feature discovery by competitive learning. In: Rumelhart, D.E., G.E.H., et al. (eds.) Parallel Distributed Processing, vol. 1, pp. 151–193. MIT Press, Cambridge (1986)Google Scholar
  20. 20.
    Grossberg, S.: Competitive learning: from interactive activation to adaptive resonance. Cognitive Science 11, 23–63 (1987)CrossRefGoogle Scholar
  21. 21.
    DeSieno, D.: Adding a conscience to competitive learning. In: Proceedings of IEEE International Conference on Neural Networks, San Diego, pp. 117–124. IEEE, Los Alamitos (1988)CrossRefGoogle Scholar
  22. 22.
    Ahalt, S.C., Krishnamurthy, A.K., Chen, P., Melton, D.E.: Competitive learning algorithms for vector quantization. Neural Networks 3, 277–290 (1990)CrossRefGoogle Scholar
  23. 23.
    Xu, L.: Rival penalized competitive learning for clustering analysis, RBF net, and curve detection. IEEE Transaction on Neural Networks 4(4), 636–649 (1993)CrossRefGoogle Scholar
  24. 24.
    Luk, A., Lien, S.: Properties of the generalized lotto-type competitive learning. In: Proceedings of International conference on neural information processing, San Mateo, pp. 1180–1185. Morgan Kaufmann Publishers, San Francisco (2000)Google Scholar
  25. 25.
    Hulle, M.M.V.: The formation of topographic maps that maximize the average mutual information of the output responses to noiseless input signals. Neural Computation 9(3), 595–606 (1997)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2006

Authors and Affiliations

  • Ryotaro Kamimura
    • 1
  • Fumihiko Yoshida
    • 2
  1. 1.Information Science Laboratory 
  2. 2.Department of Media StudiesTokai UniversityHiratsuka KanagawaJapan

Personalised recommendations