Advertisement

Data Mining and Information Systems: Quo Vadis?

  • Robert Stahlbock
  • Stefan Lessmann
  • Sven F. Crone
Chapter
Part of the Annals of Information Systems book series (AOIS, volume 8)

Abstract

Information and communication technology has been a steady source of innovations which have considerably impacted the way companies conduct business in the digital as well as the physical world. Today, information systems (IS) holistically support virtually all aspects of corporations and nonprofit institutions, along internal processes from purchasing and operationsmanagement toward sales, marketing, and eventually the customer (horizontally along the supply chain), from these operational functions toward finance, accounting, and upper management activities (vertically across the hierarchy) and externally to collaborate with external partners, suppliers, or customers. The holistic support of internal business processes and external relationships by means of IS has, in turn, led to the vast growth of internal and external data being stored and processed within corporate environments.

Keywords

Data Mining Partial Little Square Random Forest Information System Support Vector Regression 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Notes

Acknowledgments

We would like to thank all authors who submitted their work for consideration to this focused issue. Their contributions made this special issue possible. We would like to thank especially the reviewers for their time and their thoughtful reviews. Finally, we would like to thank the two series editors, Ramesh Sharda and Stefan Voß for their valuable advice and encouragement, and the editorial staff for their support in the production of this special issue (Hamburg, June 2009).

References

  1. 1.
    Agrawal, R. and Srikant, R. Fast algorithms for mining association rules in large databases. In: Bocca, J. B., Jarke, M., and Zaniolo, C. (eds.), Proc. of the 20th Intern. Conf. on Very Large Databases (VLDB’94), pp. 487–499, Santiago de Chile, Chile, 1994. Morgan Kaufmann.Google Scholar
  2. 2.
    Ayres, I. Super Crunchers: Why Thinking-By-Numbers Is the New Way to Be Smart. Bantam Dell, New York, 2007.Google Scholar
  3. 3.
    Berry, M. J. A. and Linoff, G. Data Mining Techniques: For Marketing, Sales and Customer Relationship Management. Wiley, New York, 2nd ed., 2004.Google Scholar
  4. 4.
    Bose, I. and Xi, C. Quantitative models for direct marketing: A review from systems perspective. European Journal of Operational Research, 195 (1): 1–16, 2009.CrossRefGoogle Scholar
  5. 5.
    Boylu, F., Aytug, H., and Köhler, G. J. Induction over strategic agents. Information Systems Research, forthcoming.Google Scholar
  6. 6.
    Breiman, L. Random forests. Machine Learning, 45 (1): 5–32, 2001.CrossRefGoogle Scholar
  7. 7.
    Cabena, P., Hadjnian, P., Stadler, R., Verhees, J., and Zanasi, A. Discovering Data Mining: From Concept to Implementation. Prentice Hall, MLondon, 1997.Google Scholar
  8. 8.
    Crook, J. N., Edelman, D. B., and Thomas, L. C. Recent developments in consumer credit risk assessment. European Journal of Operational Research, 183 (3): 1447–1465, 2007.CrossRefGoogle Scholar
  9. 9.
    Davis, F. D. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 13 (3): 319–340, 1989.CrossRefGoogle Scholar
  10. 10.
    Fayyad, U., Piatetsky-Shapiro, G., and Smyth, P. From data mining to knowledge discovery in databases: An overview. AI Magazine, 17 (3): 37–54, 1996.Google Scholar
  11. 11.
    Felici, G., Simeone, B., and Spinelli, V. Classification techniques and error control in logic mining. Annals of Information Systems, in this issue.Google Scholar
  12. 12.
    Figueroa, C. J. Predicting customer loyalty labels in a large retail database: A case study in Chile. Annals of Information Systems, in this issue.Google Scholar
  13. 13.
    Fildes, R., Nikolopoulos, K., Crone, S. F., and Syntetos, A. A. Forecasting and operational research: A review. Journal of the Operational Research Society, 59: 1150–1172, 2006.Google Scholar
  14. 14.
    Freitas, A. On rule interestingness measures. Knowledge-Based Systems, 12 (5–6): 309–315, October 1999. URL http://www.cs.kent.ac.uk/pubs/1999/1407.
  15. 15.
    Friedman, J. H. Recent advances in predictive (machine) learning. Journal of Classification, 23 (2): 175–197, 2006.CrossRefGoogle Scholar
  16. 16.
    Geczy, P., Izumi, N., Akaho, S., and Hasida, K. Behaviorally founded recommendation algorithm for browsing assistance systems. Annals of Information Systems, in this issue.Google Scholar
  17. 17.
    Geng, L. and Hamilton, H. J. Interestingness measures for data mining: A survey. ACM Computing Surveys, 38 (3): Article No. 9, 2006.Google Scholar
  18. 18.
    Gijsberts, A., Metta, G., and Rothkrantz, L. Evolutionary optimization of least-squares support vector machines. Annals of Information Systems, in this issue.Google Scholar
  19. 19.
    Guyon, I., Weston, J., Barnhill, S., and Vapnik, V. Gene selection for cancer classification using support vector machines. Machine Learning, 46 (1-3): 389–422, 2002.CrossRefGoogle Scholar
  20. 20.
    Han, J. and Kamber, M. Data mining: Concepts and Techniques. The Morgan Kaufmann series in data management systems. Morgan Kaufmann, San Francisco, 7th ed., 2004.Google Scholar
  21. 21.
    Hand, D. J. Data mining: Statistics and more? American Statistician, 52 (2): 112–118, 1998.CrossRefGoogle Scholar
  22. 22.
    Hand, D. J. Statistics and data mining: Intersecting disciplines. ACM SIGKDD Explorations Newsletter, 1 (1): 16–19, 1999.CrossRefGoogle Scholar
  23. 23.
    Hand, D. J. Classifier technology and the illusion of progress. Statistical Science, 21 (1): 1–14, 2006.CrossRefGoogle Scholar
  24. 24.
    Hand, D. J., Mannila, H., and Smyth, P. Principles of Data Mining. Adaptive computation and machine learning. MIT Press, Cambridge, London, 2001.Google Scholar
  25. 25.
    Hastie, T., Tibshirani, R., and Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. Springer, New York, 2002.Google Scholar
  26. 26.
    Japkowicz, N. and Stephen, S. The class imbalance problem: A systematic study. Intelligent Data Analysis, 6 (5): 429–450, 2002.Google Scholar
  27. 27.
    Joachims, T. Text categorization with support vector machines: Learning with many relevant features. In: Nedellec, C. and Rouveirol, C. (eds.), Proc. of the 10th European Conf. on Machine Learning, vol.1398 of Lecture Notes in Computer Science, pp. 137–142, Chemnitz, Germany, 1998. Springer.Google Scholar
  28. 28.
    Joachims, T. Making large-scale SVM learning practical. In: Schölkopf, B., Burges, C. J. C., and Smola, A. J. (eds.), Advances in Kernel Methods: Support Vector Learning, pp. 169–184. MIT Press, Cambridge, 1999.Google Scholar
  29. 29.
    Johansson, U., König, R., and Niklasson, L. Genetically evolved kNN ensembles. Annals of Information Systems, in this issue.Google Scholar
  30. 30.
    Karamitopoulos, L., Evangelidis, G., and Dervos, D. PCA-based time series similarity search. Annals of Information Systems, in this issue.Google Scholar
  31. 31.
    Le Bras, Y., Lenca, P., and Lallich, S. Mining interesting rules without support requirement: A general universal existential upward closure property. Annals of Information Systems, in this issue.Google Scholar
  32. 32.
    Lemmond, T. D., Chen, B. Y., Hatch, A. O., and Hanley, W. G. An extended study of the discriminant random forest. Annals of Information Systems, in this issue.Google Scholar
  33. 33.
    Liu, A., Martin, C., La Cour, B., and Ghosh, J. Effects of oversampling versus cost-sensitive learning for Bayesian and SVM classifiers. Annals of Information Systems, in this issue.Google Scholar
  34. 34.
    Liu, B., Hsu, W., Chen, S., and Ma, Y. Analyzing the subjective interestingness of association rules. IEEE Intelligent Systems, 15 (5): 47–55, 2000.CrossRefGoogle Scholar
  35. 35.
    Mangasarian, O. L. and Wild, E. W. Privacy-preserving random kernel classification of checkerboard partitioned data. Annals of Information Systems, in this issue.Google Scholar
  36. 36.
    Martens, D. and Baesens, B. Building acceptable classification models. Annals of Information Systems, in this issue.Google Scholar
  37. 37.
    Narayanan, A. and Shmatikov, V. How to break anonymity of the Netflix prize dataset, 2006. URL http://www.citebase.org/abstract?id=oai:arXiv.org:cs/0610105.
  38. 38.
    Olafsson, S. Introduction to operations research and data mining. Computers and Operations Research, 33 (11): 3067–3069, 2006.CrossRefGoogle Scholar
  39. 39.
    Olafsson, S., Li, X., and Wu, S. Operations research and data mining. European Journal of Operational Research, 187 (3): 1429–1448, 2008.Google Scholar
  40. 40.
    Özöğür-Akyüz, S., Hussain, Z., and Shawe-Taylor, J. Prediction with the SVM using test point margins. Annals of Information Systems, in this issue.Google Scholar
  41. 41.
    Ringle, C. M., Sarstedt, M., and Mooi, E. A. Repose-based segmentation using finite mixture partial least squares. Annals of Information Systems, in this issue.Google Scholar
  42. 42.
    Ryan, S. and Hamel, L. Using web text mining to predict future events: A test of the wisdom of crowds hypothesis. Annals of Information Systems, in this issue.Google Scholar
  43. 43.
    Saar-Tsechansky, M. and Provost, F. Decision-centric active learning of binary-outcome models. Information Systems Research, 18 (1): 4–22, 2007.CrossRefGoogle Scholar
  44. 44.
    Truta, T. M. and Campan, A. Avoiding attribute disclosure with the (extended) p-sensitive k-anonymity model. Annals of Information Systems, in this issue.Google Scholar
  45. 45.
    Vapnik, V. N. Estimation of Dependences Based on Empirical Data. Springer, New York, 1982.Google Scholar
  46. 46.
    Vapnik, V. N. The Nature of Statistical Learning Theory. Springer, New York, 1995.Google Scholar
  47. 47.
    Vapnik, V. N. Statistical Learning Theory. Wiley, New York, 1998.Google Scholar
  48. 48.
    Voβ, S. Meta-heuristics: The state of the art. In: Nareyek, A. (ed.), Local Search for Planning and Scheduling, vol. 2148 of Lecture Notes in Artificial Intelligence, pp. 1–23. Springer, Berlin, 2001.Google Scholar
  49. 49.
    Weiss, G. M. The impact of small disjuncts on classifier learning. Annals of Information Systems, in this issue.Google Scholar
  50. 50.
    Weiss, G. M. Mining with rarity: A unifying framework. ACM SIGKDD Explorations Newsletter, 6 (1): 7–19, 2004.CrossRefGoogle Scholar
  51. 51.
    Weiss, G. M., Zadrozny, B., and Saar-Tsechansky, M. Guest editorial: special issue on utility-based data mining. Data Mining and Knowledge Discovery, 17 (2): 129–135, 2008.CrossRefGoogle Scholar
  52. 52.
    Wu, X., Kumar, V., Ross Quinlan, J., Ghosh, J., Yang, Q., Motoda, H., McLachlan, G., Ng, A., Liu, B., Yu, P., Zhou, Z.-H., Steinbach, M., Hand, D., and Steinberg, D. Top 10 algorithms in data mining. Knowledge and Information Systems, 14 (1): 1–37, 2008.CrossRefGoogle Scholar
  53. 53.
    Yu, P. (ed.). Proc. of the 2007 Intern. Workshop on Domain Driven Data Mining. ACM, New York, 2007.Google Scholar

Copyright information

© Springer Science+Business Media, LLC 2010

Authors and Affiliations

  1. 1.Institute of Information SystemsUniversity of HamburgHamburgGermany
  2. 2.Lecturer at the FOM University of Applied SciencesEssen/HamburgGermany
  3. 3.Centre for ForecastingLancaster University Management SchoolLancasterUK

Personalised recommendations