Advertisement

Soft Computing

, Volume 22, Issue 5, pp 1569–1576 | Cite as

Outgoing call recommendation using neural network

  • Seokhoon Kang
Focus

Abstract

Personalized data collected from smartphones and similar devices reflect time dependency. Currently, research using such data as a basis for prediction of future data is ongoing. This suggests that not only is time dependency reflected in past data but also in newly produced data, and that time-dependent weights also are reflected therein. This paper analyzed the most prominent feature of personalized data, the call log. The random forest method was used to find highly correlated data between call data and new outgoing calls. This information was then learned through a neural network. Data were collected from 10 subjects, and 80% was learned, while 20% was used for prediction. The results showed that for six suggested numbers, the Hit Ratio was 77% while for nine suggested numbers, it was 90%. This indicates that when nine numbers are recommended, the probability that one of them will be called is high. “Recent call volume” and “changing data” showed a high correlation, and experiments were conducted so that time change could be adequately reflected. The problem of repetitions was addressed while maintaining the Hit Ratio.

Keywords

Smartphone Call prediction Intention analysis RNN Scheduling Communication 

Notes

Acknowledgements

This study was funded by Incheon National University. (Grant Number 2012 0066).

Compliance with ethical standards

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

This article does not contain any studies with human participants or animals performed by any of the authors.

References

  1. Aragam B, Zhou Q (2015) Concave penalized estimation of sparse Gaussian Bayesian networks. J Mach Learn Res 16:2273–2328MathSciNetzbMATHGoogle Scholar
  2. Connor JT et al (1994) Recurrent neural networks and robust time series prediction. IEEE Trans Neural Netw 5(2):240–254CrossRefGoogle Scholar
  3. He YB, Geng Z (2008) Active learning of causal networks with intervention experiments and optimal designs. J Mach Learn Res 9:2523–2547MathSciNetzbMATHGoogle Scholar
  4. Heckerman D, Geiger D, Chickering DM (1995) Learning Bayesian networks: the combination of knowledge and statistical data. Mach Learn 20(3):197–243zbMATHGoogle Scholar
  5. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9:1735–1780CrossRefGoogle Scholar
  6. Krizhevsky A, Sutskever I, Hinton G (2012)ImageNet classification with deep convolutional neural networks. In: Proceedings of advances in neural information processing systems 25, pp 1090–1098Google Scholar
  7. Long Y, Wang S, Wu W, Yang X, Jeon G, Liu K (2015) Structured-light-assisted wireless digital optical communications. Opt Commun 355:406–410CrossRefGoogle Scholar
  8. Petrosian A et al (2000) Recurrent neural network based prediction of epileptic seizures in intra- and extracranial EEG. Neurocomputing 30:201–218CrossRefGoogle Scholar
  9. Roth M et al (2016) Driver and pedestrian awareness-based collision risk analysis. In: 2016 IEEE Intelligent vehicles symposium (IV), Gothenburg, Sweden, June 19–22, pp 454–459Google Scholar
  10. Weigend AS, Neil A G et al (1994) Time series prediction. Addison Wesley, BostonGoogle Scholar
  11. Wu W, Yang X, Pang Y, Peng J, Jeon G (2013) A multifocus image fusion method by using hidden Markov model. Opt Commun 287:63–72CrossRefGoogle Scholar
  12. Wu J, Anisetti M, Wu W, Damiani E, Jeon G (2016) Bayer demosaicking with polynomial interpolation. IEEE Trans Image Process 25(11):5369–5382MathSciNetCrossRefGoogle Scholar

Copyright information

© Springer-Verlag GmbH Germany, part of Springer Nature 2017

Authors and Affiliations

  1. 1.Department of Embedded Systems EngineeringUniversity of IncheonIncheonRepublic of Korea

Personalised recommendations