Abstract
To analyze Hadoop techniques like MapReduce, which will help to process the data faster and in efficient way to detect dementia. For given voluminous dementia dataset, current solution uses different data partitioning strategies which experiences large communication cost and expensive mining process due to duplicate and unnecessary transactions transferred among computing nodes. To clear this issues proposed algorithm uses data partitioning techniques such as Min-Hash and Locality Sensitive Hashing which will reduce processing time and improve efficiency of final result. We are taking help of MapReduce programming model of Hadoop [3]. We implement this technique on a Hadoop platform. For pattern matching we use FPgrowth algorithm. Finally we shows that the proposed system requires less time to finding frequent item sets. The idea behind research is to adopt to cope with the special requirement of health domain related with patients.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Xun Y, Zhang J, Qin X, Zhao X (2017) Fidoop-DP: data partitioning in frequent itemset mining on hadoop cluster. IEEE Trans Parallel Distrib Syst 28(1):101–114
Agarwal R, Singh S, Vats S (2016) Implementation of an improved algorithm for frequent itemset mining using hadoop. In: International conference on computing, communication and automation (ICCCA2016)
Xie J, et al (2010) Improving mapreduce performance through data placement in heterogeneous hadoop clusters. IEEE
Thusoo A, et al (2010) Hive-a petabyte scale data warehouse using hadoop. IEEE
Zhang J, et al (2015) Research on multiple files input programming method based on MapReduce. Springer
Singh N, Agrawal S (2015) A review of research on MapReduce scheduling algorithms in Hadoop. In: ICCCA
Gates AF, et al (2009) Building a high-level dataflow system on top of Map-Reduce: the Pig experience. In: Proceedings of the VLDB endowment
Qi J, Yu Y, Wang L, Liu J (2016) K*-means: an effective and efficient K-means clustering algorithm. In: IEEE International conferences on big data and cloud computing (BDCloud), social computing and networking
Gawwad MA et al (2017) Frequent itemset mining for big data using greatest common divisor technique. Data Sci J 16(25):1–10. https://doi.org/10.5334/dsj-2017-02
Waghere S, Rajarajeswari P (2017) Parallel frequent dataset mining and feature subset selection for high dimensional data on hadoop using map-reduce. Int J Appl Eng Res 12(18):7783–7789 ISSN 0973-4562
Dementia Guide, National Health Services (NHS). http://www.nhs.uk/Conditions/dementia-guide/Pages/about-dementia.aspx
Dementia Including Alzheimer’s Disease, Office of Disease Prevention andHealthPromotion(ODPHP). http://www.healthypeople.gov/2020/topics-objectives/topic/dementias-includingalzheimers-diseas
Liu H, Motoda H (1998) Feature selection for knowledge discovery and data mining. In: The springer international series in engineering and computer science, vol 454, pp 73–95
Waghere SS, Rajarajeswari P (2017) A survey on achieving best knowledge from frequent item set mining using fidoop. Int J Comput Appl (0975 – 8887) 171(9)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Waghere, S., RajaRajeswari, P., Ganesan, V. (2021). Design and Implementation of System Which Efficiently Retrieve Useful Data for Detection of Dementia Disease. In: Kumar, A., Mozar, S. (eds) ICCCE 2020. Lecture Notes in Electrical Engineering, vol 698. Springer, Singapore. https://doi.org/10.1007/978-981-15-7961-5_144
Download citation
DOI: https://doi.org/10.1007/978-981-15-7961-5_144
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-7960-8
Online ISBN: 978-981-15-7961-5
eBook Packages: EngineeringEngineering (R0)